A Multifocal Graph-Based Neural Network Scheme for Topic Event Extraction

被引:0
|
作者
Wan, Qizhi [1 ,2 ]
Wan, Changxuan [1 ,2 ]
Xiao, Keli [3 ]
Hu, Rong [1 ,2 ]
Liu, Dexi [1 ,2 ]
Liao, Guoqiong [1 ]
Liu, Xiping [1 ,2 ]
Shuai, Yuxin [1 ,2 ]
机构
[1] Jiangxi Univ Finance & Econ, Sch Comp & Artificial Intelligence, Nanchang, Peoples R China
[2] Jiangxi Key Lab Data & Knowledge Engn, Nanchang, Peoples R China
[3] SUNY Stony Brook, Coll Business, Stony Brook, NY 11794 USA
基金
中国国家自然科学基金;
关键词
Event Topic; topic event extraction; event graphs; subgraph; graph neural network;
D O I
10.1145/3696353
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Event extraction is a long-standing and challenging task in natural language processing, and existing studies mainly focus on extracting events within sentences. However, a significant problem that has not been carefully investigated is whether an "event topic" can be identified to represent the main aspects of extracted events. This article formulates the "topic event" extraction problem, aiming to identify a representative event from extracted ones. Specifically, after defining the topic event, we develop a multifocal graph-based framework to handle the extraction task. To enrich the associations of events and their tokens, we construct four event graphs, including the event subgraph and three event-associated graphs (i.e., event dependency parsing graph, event organization graph, and event share token graph), that reflect the internal and external structures of events, respectively. Subsequently, we design a multi-attention event-graph neural network to capture these event graph structures and improve event subgraph embedding. Finally, the output embeddings in the last layer of each channel are concatenated and fed into a fully connected network for topic event recognition. Extensive experiments validate the effectiveness of our method, and the results confirm its superiority over state-of-the-art baselines. In-depth analyses explore the essential factors (e.g., graph structures, attentions, feature generation method, etc.) determining the extraction performance.
引用
收藏
页数:36
相关论文
共 50 条
  • [41] Graph-Based Analysis in Network Security
    Collins, M. Patrick
    2011 - MILCOM 2011 MILITARY COMMUNICATIONS CONFERENCE, 2011, : 1333 - 1337
  • [42] GRAPH-BASED KINSHIP REASONING NETWORK
    Li, Wanhua
    Zhang, Yingqiang
    Lv, Kangchen
    Lu, Jiwen
    Feng, Jianjiang
    Zhou, Jie
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,
  • [43] Graph-based network analysis in schizophrenia
    Micheloyannis, Sifis
    WORLD JOURNAL OF PSYCHIATRY, 2012, 2 (01): : 1 - 12
  • [44] A Graph-Based Approach to Automatic Convolutional Neural Network Construction for Image Classification
    Yuan, Gonglin
    Xue, Bing
    Zhang, Mengjie
    2020 35TH INTERNATIONAL CONFERENCE ON IMAGE AND VISION COMPUTING NEW ZEALAND (IVCNZ), 2020,
  • [45] A Graph-Based Approach to Topic Clustering of Tourist Attraction Reviews
    Sirilertworakul, Nuttha
    Yimwadsana, Boonsit
    INFORMATION AND SOFTWARE TECHNOLOGIES, ICIST 2019, 2019, 1078 : 343 - 354
  • [46] Automatic Labeling of Topic Models Using Graph-Based Ranking
    He, Dongbin
    Wang, Minjuan
    Khattak, Abdul Mateen
    Zhang, Li
    Gao, Wanlin
    IEEE ACCESS, 2019, 7 : 131593 - 131608
  • [47] A Graph-based Topic Modeling Approach to Detection of Irrelevant Citations
    Phu Pham
    Hieu Le
    Nguyen Thanh Tam
    Quang-Dieu Tran
    VIETNAM JOURNAL OF COMPUTER SCIENCE, 2023, 10 (02) : 197 - 216
  • [48] Graph-based topic models for trajectory clustering in crowd videos
    Al Ghamdi, Manal
    Gotoh, Yoshihiko
    MACHINE VISION AND APPLICATIONS, 2020, 31 (05)
  • [49] Graph-based topic models for trajectory clustering in crowd videos
    Manal Al Ghamdi
    Yoshihiko Gotoh
    Machine Vision and Applications, 2020, 31
  • [50] A graph-based convolutional neural network stock price prediction with leading indicators
    Wu, Jimmy Ming-Tai
    Li, Zhongcui
    Srivastava, Gautam
    Tasi, Meng-Hsiun
    Lin, Jerry Chun-Wei
    SOFTWARE-PRACTICE & EXPERIENCE, 2021, 51 (03): : 628 - 644