GeoNER:Geological Named Entity Recognition with Enriched Domain Pre-Training Model and Adversarial Training

被引:0
|
作者
MA Kai [1 ,2 ]
HU Xinxin [1 ,2 ]
TIAN Miao [3 ]
TAN Yongjian [1 ,2 ]
ZHENG Shuai [1 ,2 ]
TAO Liufeng [3 ,4 ,5 ]
QIU Qinjun [3 ,4 ,5 ]
机构
[1] Hubei Key Laboratory of Intelligent Vision Based Monitoring for Hydroelectric Engineering, China Three Gorges University
[2] College of Computer and Information Technology, China Three Gorges University
[3] Key Laboratory of Geological Survey and Evaluation of Ministry of Education, China University of Geosciences
[4] School of Computer Science, China University of Geosciences
[5] Key Laboratory of Quantitative Resource Evaluation and Information Engineering, Ministry of Natural Resources, China University of
关键词
D O I
暂无
中图分类号
TP391.1 [文字信息处理]; P628 [数学勘探];
学科分类号
摘要
As important geological data, a geological report contains rich expert and geological knowledge, but the challenge facing current research into geological knowledge extraction and mining is how to render accurate understanding of geological reports guided by domain knowledge. While generic named entity recognition models/tools can be utilized for the processing of geoscience reports/documents, their effectiveness is hampered by a dearth of domain-specific knowledge, which in turn leads to a pronounced decline in recognition accuracy. This study summarizes six types of typical geological entities, with reference to the ontological system of geological domains and builds a high quality corpus for the task of geological named entity recognition(GNER). In addition, Geo Wo BERT-adv BGP(Geological Word-base BERTadversarial training Bi-directional Long Short-Term Memory Global Pointer) is proposed to address the issues of ambiguity, diversity and nested entities for the geological entities. The model first uses the fine-tuned word granularitybased pre-training model Geo Wo BERT(Geological Word-base BERT) and combines the text features that are extracted using the Bi LSTM(Bi-directional Long Short-Term Memory), followed by an adversarial training algorithm to improve the robustness of the model and enhance its resistance to interference, the decoding finally being performed using a global association pointer algorithm. The experimental results show that the proposed model for the constructed dataset achieves high performance and is capable of mining the rich geological information.
引用
收藏
页码:1404 / 1417
页数:14
相关论文
共 50 条
  • [31] Poster: Boosting Adversarial Robustness by Adversarial Pre-training
    Xu, Xiaoyun
    Picek, Stjepan
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 3540 - 3542
  • [32] Improved OOD Generalization via Adversarial Training and Pre-training
    Yi, Mingyangi
    Hou, Lu
    Sun, Jiacheng
    Shang, Lifeng
    Jiang, Xin
    Liu, Qun
    Ma, Zhi-Ming
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [33] Entity recognition in the field of coal mine construction safety based on a pre-training language model
    Xu, Na
    Liang, Yanxiang
    Guo, Chaoran
    Meng, Bo
    Zhou, Xueqing
    Hu, Yuting
    Zhang, Bo
    ENGINEERING CONSTRUCTION AND ARCHITECTURAL MANAGEMENT, 2023,
  • [34] Context-aware Adversarial Training for Name Regularity Bias in Named Entity Recognition
    Ghaddar, Abbas
    Langlais, Philippe
    Rashid, Ahmad
    Rezagholizadeh, Mehdi
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2021, 9 : 586 - 604
  • [35] Adversarial momentum-contrastive pre-training
    Xu, Cong
    Li, Dan
    Yang, Min
    PATTERN RECOGNITION LETTERS, 2022, 160 : 172 - 179
  • [36] Robust Pre-Training by Adversarial Contrastive Learning
    Jiang, Ziyu
    Chen, Tianlong
    Chen, Ting
    Wang, Zhangyang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [37] ResiAdvNet: A named entity recognition model for potato diseases and pests based on progressive residual structures and adversarial training
    Tang, Wentao
    Wen, Xianhuan
    Li, Miao
    Chen, Yuqi
    Hu, Zelin
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2024, 227
  • [38] Better Representations via Adversarial Training in Pre-Training: A Theoretical Perspective
    Xing, Yue
    Lin, Xiaofeng
    Song, Qifan
    Xu, Yi
    Zeng, Belinda
    Cheng, Guang
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [39] Improved Attention Mechanism and Adversarial Training for Respiratory Infectious Disease Text Named Entity Recognition
    Liu, Junhong
    Wei, Wenxue
    Zhang, Yukun
    Liang, Lei
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT V, 2023, 14258 : 103 - 114
  • [40] Subset selection for domain adaptive pre-training of language model
    Hwang, Junha
    Lee, Seungdong
    Kim, Haneul
    Jeong, Young-Seob
    SCIENTIFIC REPORTS, 2025, 15 (01):