Distantly-Supervised Named Entity Recognition with Noise-Robust Learning and Language Model Augmented Self-Training

被引:0
|
作者
Meng, Yu [1 ]
Zhang, Yunyi [1 ]
Huang, Jiaxin [1 ]
Wang, Xuan [1 ]
Zhang, Yu [1 ]
Ji, Heng [1 ]
Han, Jiawei [1 ]
机构
[1] Univ Illinois, Champaign, IL 61820 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the problem of training named entity recognition (NER) models using only distantly-labeled data, which can be automatically obtained by matching entity mentions in the raw text with entity types in a knowledge base. The biggest challenge of distantlysupervised NER is that the distant supervision may induce incomplete and noisy labels, rendering the straightforward application of supervised learning ineffective. In this paper, we propose (1) a noise-robust learning scheme comprised of a new loss function and a noisy label removal step, for training NER models on distantly-labeled data, and (2) a self-training method that uses contextualized augmentations created by pre-trained language models to improve the generalization ability of the NER model. On three benchmark datasets, our method achieves superior performance, outperforming existing distantlysupervised NER models by significant margins(1).
引用
收藏
页码:10367 / 10378
页数:12
相关论文
共 50 条
  • [31] Language model based on deep learning network for biomedical named entity recognition
    Hou, Guan
    Jian, Yuhao
    Zhao, Qingqing
    Quan, Xiongwen
    Zhang, Han
    METHODS, 2024, 226 : 71 - 77
  • [32] Knowledge-Augmented Language Model and Its Application to Unsupervised Named-Entity Recognition
    Liu, Angli
    Du, Jingfei
    Stoyanov, Veselin
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 1142 - 1150
  • [33] Revisiting Self-Training for Few-Shot Learning of Language Model
    Chen, Yiming
    Zhang, Yan
    Zhang, Chen
    Lee, Grandee
    Cheng, Ran
    Li, Haizhou
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9125 - 9135
  • [34] A CNN-Based Semi-supervised Self-training Method for Robust Underwater Fish Recognition
    Li, Tanqing
    Zhao, Zhili
    Zhang, Hengyu
    Li, Kun
    Lv, Wenjun
    PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON ELECTRONIC INFORMATION TECHNOLOGY AND COMPUTER ENGINEERING, EITCE 2023, 2023, : 1553 - 1559
  • [35] Noise-Robust Vision-Language Pre-Training With Positive-Negative Learning
    Huang, Zhenyu
    Yang, Mouxing
    Xiao, Xinyan
    Hu, Peng
    Peng, Xi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (01) : 338 - 350
  • [36] Language Model Pre-training Method in Machine Translation Based on Named Entity Recognition
    Li, Zhen
    Qu, Dan
    Xie, Chaojie
    Zhang, Wenlin
    Li, Yanxia
    INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2020, 29 (7-8)
  • [37] Research-based-named Entity Recognition Learning Text Biomedical Extraction by Adoption of Training Bidirectional Language Model (BiLM)
    Abed, Alshreef
    Jingling, Yuan
    Li, Lin
    Journal of Computers (Taiwan), 2020, 31 (04) : 157 - 173
  • [38] DeepSpacy-NER: an efficient deep learning model for named entity recognition for Punjabi language
    Navdeep Singh
    Munish Kumar
    Bavalpreet Singh
    Jaskaran Singh
    Evolving Systems, 2023, 14 : 673 - 683
  • [39] DeepSpacy-NER: an efficient deep learning model for named entity recognition for Punjabi language
    Singh, Navdeep
    Kumar, Munish
    Singh, Bavalpreet
    Singh, Jaskaran
    EVOLVING SYSTEMS, 2023, 14 (04) : 673 - 683
  • [40] A Robust Semi-Supervised Broad Learning System Guided by Ensemble-Based Self-Training
    Guo, Jifeng
    Chen, C. L. Philip
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (11) : 6410 - 6422