Modeling Dense Cross-Modal Interactions for Joint Entity-Relation Extraction

被引:0
|
作者
Zhao, Shan [1 ]
Hu, Minghao [2 ]
Cai, Zhiping [1 ]
Liu, Fang [3 ]
机构
[1] Natl Univ Def Technol, Coll Comp, Changsha, Peoples R China
[2] PLA Acad Mil Sci, Beijing, Peoples R China
[3] Hunan Univ, Sch Design, Changsha, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Joint extraction of entities and their relations benefits from the close interaction between named entities and their relation information. Therefore, how to effectively model such cross-modal interactions is critical for the final performance. Previous works have used simple methods such as label-feature concatenation to perform coarse-grained semantic fusion among cross-modal instances, but fail to capture fine-grained correlations over token and label spaces, resulting in insufficient interactions. In this paper, we propose a deep Cross-Modal Attention Network (CMAN) for joint entity and relation extraction. The network is carefully constructed by stacking multiple attention units in depth to fully model dense interactions over token-label spaces, in which two basic attention units are proposed to explicitly capture fine-grained correlations across different modalities (e.g., token-to-token and label-to-token). Experiment results on CoNLL04 dataset show that our model obtains state-of-the-art results by achieving 90.62% F1 on entity recognition and 72.97% F1 on relation classification. In ADE dataset, our model surpasses existing approaches by more than 1.9% F1 on relation classification. Extensive analyses further confirm the effectiveness of our approach.
引用
收藏
页码:4032 / 4038
页数:7
相关论文
共 50 条
  • [41] Cross-modal interactions in discrete bimanual movements
    Kagerer, Florian A.
    JOURNAL OF SPORT & EXERCISE PSYCHOLOGY, 2014, 36 : S34 - S34
  • [42] Cross-modal interactions in the perception of musical performance
    Vines, Bradley W.
    Krumhansl, Carol L.
    Wanderley, Marcelo M.
    Levitin, Daniel J.
    COGNITION, 2006, 101 (01) : 80 - 113
  • [43] A Triple Relation Network for Joint Entity and Relation Extraction
    Wang, Zixiang
    Yang, Liqun
    Yang, Jian
    Li, Tongliang
    He, Longtao
    Li, Zhoujun
    ELECTRONICS, 2022, 11 (10)
  • [44] Dual Encoding Integrating Key Frame Extraction for Video-text Cross-modal Entity Resolution
    Zeng Z.
    Cao J.
    Weng N.
    Jiang G.
    Fan Q.
    Binggong Xuebao/Acta Armamentarii, 2022, 43 (05): : 1107 - 1116
  • [45] Entity-Relation Extraction-A Novel and Lightweight Method Based on a Gate Linear Mechanism
    Peng, Guangming
    Chen, Xiong
    ELECTRONICS, 2020, 9 (10) : 1 - 15
  • [46] An Entity-Relation Extraction Method Based on the Mixture-of-Experts Model and Dependency Parsing
    Li, Yuanxi
    Wang, Haiyan
    Zhang, Dong
    APPLIED SCIENCES-BASEL, 2025, 15 (04):
  • [47] Effective Modeling of Encoder-Decoder Architecture for Joint Entity and Relation Extraction
    Nayak, Tapas
    Ng, Hwee Tou
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 8528 - 8535
  • [48] Cross-Modal and Hierarchical Modeling of Video and Text
    Zhang, Bowen
    Hu, Hexiang
    Sha, Fei
    COMPUTER VISION - ECCV 2018, PT XIII, 2018, 11217 : 385 - 401
  • [49] Joint feature fusion hashing for cross-modal retrieval
    Cao, Yuxia
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 6149 - 6162
  • [50] Joint Entity and Relation Extraction for Long Text
    Cheng, Dong
    Song, Hui
    He, Xianglong
    Xu, Bo
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, KSEM 2021, PT II, 2021, 12816 : 152 - 162