A comprehensive exploration of semantic relation extraction via pre-trained CNNs

被引:35
|
作者
Li, Qing [1 ]
Li, Lili [2 ]
Wang, Weinan [3 ]
Li, Qi [4 ]
Zhong, Jiang [1 ,5 ]
机构
[1] Chongqing Univ, Coll Comp Sci, Chongqing, Peoples R China
[2] Chongqing Univ, Sch Civil Engn, Chongqing, Peoples R China
[3] Peking Univ, Sch Math Sci, Beijing, Peoples R China
[4] Shaoxing Univ, Dept Comp Sci & Engn, Shaoxing, Peoples R China
[5] Chongqing Univ, Key Lab Dependable Serv Comp Cyber Phys Soc, Chongqing, Peoples R China
关键词
Relation extraction; Semantic relation; Natural language processing; Convolutional neural networks;
D O I
10.1016/j.knosys.2020.105488
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Semantic relation extraction between entity pairs is a crucial task in information extraction from text. In this paper, we propose a new pre-trained network architecture for this task, and it is called the XM-CNN. The XM-CNN utilizes word embedding and position embedding information. It is designed to reinforce the contextual output from the MT-DNNKD pre-trained model. Our model effectively utilized an entity-aware attention mechanisms to detected the features and also adopts and applies more relation-specific pooling attention mechanisms applied to it. The experimental results show that the XM-CNN achieves state-of-the-art results on the SemEval-2010 task 8, and a thorough evaluation of the method is conducted. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:7
相关论文
共 50 条
  • [41] Towards unifying pre-trained language models for semantic text exchange
    Miao, Jingyuan
    Zhang, Yuqi
    Jiang, Nan
    Wen, Jie
    Pei, Kanglu
    Wan, Yue
    Wan, Tao
    Chen, Honglong
    WIRELESS NETWORKS, 2024, 30 (07) : 6385 - 6398
  • [42] Span-aware pre-trained network with deep information bottleneck for scientific entity relation extraction
    Wang, Youwei
    Cao, Peisong
    Fang, Haichuan
    Ye, Yangdong
    NEURAL NETWORKS, 2025, 186
  • [43] Pre-processing Effects of the Tuberculosis Chest X-Ray Images on Pre-trained CNNs: An Investigation
    Tasci, Erdal
    ARTIFICIAL INTELLIGENCE AND APPLIED MATHEMATICS IN ENGINEERING PROBLEMS, 2020, 43 : 589 - 596
  • [44] AStitchInLanguageModels: Dataset and Methods for the Exploration of Idiomaticity in Pre-Trained Language Models
    Madabushi, Harish Tayyar
    Gow-Smith, Edward
    Scarton, Carolina
    Villavicencio, Aline
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 3464 - 3477
  • [45] Iris Recognition Using an Enhanced Pre-Trained Backbone Based on Anti-Aliased CNNs
    Zambrano, Jorge E.
    Pilataxi, Jhon I.
    Perez, Claudio A.
    Bowyer, Kevin W.
    IEEE ACCESS, 2024, 12 : 94570 - 94583
  • [46] Implementation of CNNs for Crop Diseases Classification: A Comparison of Pre-trained Model and Training from Scratch
    Sahu, Priyanka
    Chug, Anuradha
    Singh, Amit Prakash
    Singh, Dinesh
    Singh, Ravinder Pal
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2020, 20 (10): : 206 - 215
  • [47] Evaluation of transfer learning of pre-trained CNNs applied to breast cancer detection on infrared images
    Chaves, Esdras
    Goncalves, Caroline B.
    Albertini, Marcelo K.
    Lee, Soojeong
    Jeon, Gwanggil
    Fernandes, Henrique C.
    APPLIED OPTICS, 2020, 59 (17) : E23 - E28
  • [48] Surface-aware Mesh Texture Synthesis with Pre-trained 2D CNNs
    Kovacs, Aron Samuel
    Hermosilla, Pedro
    Raidou, Renata G.
    COMPUTER GRAPHICS FORUM, 2024, 43 (02)
  • [49] Parallel Corpus Filtering via Pre-trained Language Models
    DiDi Labs
    arXiv, 2020,
  • [50] Compression of Generative Pre-trained Language Models via Quantization
    Tao, Chaofan
    Hou, Lu
    Zhang, Wei
    Shang, Lifeng
    Jiang, Xin
    Liu, Qun
    Luo, Ping
    Wong, Ngai
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4821 - 4836