Improving Open Information Extraction with Distant Supervision Learning

被引:3
|
作者
Han, Jiabao [1 ]
Wang, Hongzhi [1 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
关键词
Distant supervision learning; Open information extraction; Neural network; Sequence-to-sequence model;
D O I
10.1007/s11063-021-10548-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Open information extraction (Open IE), as one of the essential applications in the area of Natural Language Processing (NLP), has gained great attention in recent years. As a critical technology for building Knowledge Bases (KBs), it converts unstructured natural language sentences into structured representations, usually expressed in the form of triples. Most conventional open information extraction approaches leverage a series of manual pre-defined extraction patterns or learn patterns from labeled training examples, which requires a large number of human resources. Additionally, many Natural Language Processing tools are involved, which leads to error accumulation and propagation. With the rapid development of neural networks, neural-based models can minimize the error propagation problem, but it also faces the problem of data-hungry in supervised learning. Especially, they leverage existing Open IE tools to generate training data, and it causes data quality issues. In this paper, we employ a distant supervision learning approach to improve the Open IE task. We conduct extensive experiments by employing two popular sequence-to-sequence models (RNN and Transformer) and a large benchmark data set to demonstrate the performance of our approach.
引用
收藏
页码:3287 / 3306
页数:20
相关论文
共 50 条
  • [1] Improving Open Information Extraction with Distant Supervision Learning
    Jiabao Han
    Hongzhi Wang
    Neural Processing Letters, 2021, 53 : 3287 - 3306
  • [2] Improving distant supervision using inference learning
    Roller, Roland
    Agirre, Eneko
    Soroa, Aitor
    Stevenson, Mark
    PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL) AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (IJCNLP), VOL 2, 2015, : 273 - 278
  • [3] Curriculum learning for distant supervision relation extraction
    Liu Qiongxin
    Wang Peng
    Wang Jiasheng
    Ma Jing
    JOURNAL OF WEB SEMANTICS, 2020, 61-62 (61-62):
  • [4] Improving Distant Supervision of Relation Extraction with Unsupervised Methods
    Peng, Min
    Huang, Jimin
    Sun, Zhaoyu
    Wang, Shizhong
    Wang, Hua
    Zhuo, Guangping
    Tian, Gang
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2016, PT I, 2016, 10041 : 561 - 568
  • [5] Information Extraction Using Distant Supervision and Semantic Similarities
    Park, Youngmin
    Kang, Sangwoo
    Seo, Jungyun
    ADVANCES IN ELECTRICAL AND COMPUTER ENGINEERING, 2016, 16 (01) : 11 - 18
  • [6] Enhanced Distant Supervision with State-Change Information for Relation Extraction
    Shah, Jui
    Zhang, Dongxu
    Brody, Sam
    McCallum, Andrew
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 5573 - 5579
  • [7] Improving distant supervision relation extraction with entity-guided enhancement feature
    Haixu Wen
    Xinhua Zhu
    Lanfang Zhang
    Neural Computing and Applications, 2023, 35 : 7547 - 7560
  • [8] Improving distant supervision relation extraction with entity-guided enhancement feature
    Wen, Haixu
    Zhu, Xinhua
    Zhang, Lanfang
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (10): : 7547 - 7560
  • [9] Robust Distant Supervision Relation Extraction via Deep Reinforcement Learning
    Qin, Pengda
    Xu, Weiran
    Wang, William Yang
    PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, 2018, : 2137 - 2147
  • [10] Revisiting Distant Supervision for Relation Extraction
    Jiang, Tingsong
    Liu, Jing
    Lin, Chin-Yew
    Sui, Zhifang
    PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018), 2018, : 3580 - 3585