Ensemble pretrained language models to extract biomedical knowledge from literature

被引:5
|
作者
Li, Zhao [1 ]
Wei, Qiang [1 ]
Huang, Liang-Chin [1 ]
Li, Jianfu [1 ]
Hu, Yan [1 ]
Chuang, Yao-Shun [1 ]
He, Jianping [1 ]
Das, Avisha [1 ]
Keloth, Vipina Kuttichi [2 ]
Yang, Yuntao [1 ]
Diala, Chiamaka S. [1 ]
Roberts, Kirk E. [1 ]
Tao, Cui [1 ]
Jiang, Xiaoqian [1 ]
Zheng, W. Jim [1 ]
Xu, Hua [2 ,3 ]
机构
[1] Univ Texas Hlth Sci Ctr Houston, McWilliams Sch Biomed Informat, Houston, TX 77030 USA
[2] Yale Univ, Sch Med, Sect Biomed Informat & Data Sci, New Haven, CT 06510 USA
[3] Yale Univ, Sch Med, Sect Biomed Informat & Data Sci, 100 Coll St, New Haven, CT 06510 USA
基金
美国国家卫生研究院;
关键词
named entity recognition; relation extraction; large language model; ensemble learning; knowledge base; RECOGNITION; NAME;
D O I
10.1093/jamia/ocae061
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objectives The rapid expansion of biomedical literature necessitates automated techniques to discern relationships between biomedical concepts from extensive free text. Such techniques facilitate the development of detailed knowledge bases and highlight research deficiencies. The LitCoin Natural Language Processing (NLP) challenge, organized by the National Center for Advancing Translational Science, aims to evaluate such potential and provides a manually annotated corpus for methodology development and benchmarking.Materials and Methods For the named entity recognition (NER) task, we utilized ensemble learning to merge predictions from three domain-specific models, namely BioBERT, PubMedBERT, and BioM-ELECTRA, devised a rule-driven detection method for cell line and taxonomy names and annotated 70 more abstracts as additional corpus. We further finetuned the T0pp model, with 11 billion parameters, to boost the performance on relation extraction and leveraged entites' location information (eg, title, background) to enhance novelty prediction performance in relation extraction (RE).Results Our pioneering NLP system designed for this challenge secured first place in Phase I-NER and second place in Phase II-relation extraction and novelty prediction, outpacing over 200 teams. We tested OpenAI ChatGPT 3.5 and ChatGPT 4 in a Zero-Shot setting using the same test set, revealing that our finetuned model considerably surpasses these broad-spectrum large language models.Discussion and Conclusion Our outcomes depict a robust NLP system excelling in NER and RE across various biomedical entities, emphasizing that task-specific models remain superior to generic large ones. Such insights are valuable for endeavors like knowledge graph development and hypothesis formulation in biomedical research.
引用
收藏
页码:1904 / 1911
页数:8
相关论文
共 50 条
  • [31] Extracting Latent Steering Vectors from Pretrained Language Models
    Subramani, Nishant
    Suresh, Nivedita
    Peters, Matthew E.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 566 - 581
  • [32] Natural language processing: algorithms and tools to extract computable information from EHRs and from the biomedical literature
    Ohno-Machado, Lucila
    Nadkarni, Prakash
    Johnson, Kevin
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2013, 20 (05) : 805 - 805
  • [33] Knowledge Graphs and Pretrained Language Models Enhanced Representation Learning for Conversational Recommender Systems
    Qiu, Zhangchi
    Tao, Ye
    Pan, Shirui
    Liew, Alan Wee-Chung
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 15
  • [34] DLAMA: A Framework for Curating Culturally Diverse Facts for Probing the Knowledge of Pretrained Language Models
    Keleg, Amr
    Magdy, Walid
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 6245 - 6266
  • [35] KNOWLEDGE TRANSFER FROM LARGE-SCALE PRETRAINED LANGUAGE MODELS TO END-TO-END SPEECH RECOGNIZERS
    Kubo, Yotaro
    Karita, Shigeki
    Bacchiani, Michiel
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8512 - 8516
  • [36] ReGen: Reinforcement Learning for Text and Knowledge Base Generation using Pretrained Language Models
    Dognin, Pierre L.
    Padhi, Inkit
    Melnyk, Igor
    Das, Payel
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1084 - 1099
  • [37] Few-shot Knowledge Graph-to-Text Generation with Pretrained Language Models
    Li, Junyi
    Tang, Tianyi
    Zhao, Wayne Xin
    Wei, Zhicheng
    Yuan, Nicholas Jing
    Wen, Ji-Rong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1558 - 1568
  • [38] Knowledge Graph Completion for Power Grid Main Equipment Using Pretrained Language Models
    Lin, Chenxiang
    Zheng, Zhou
    Cai, Shitao
    Fu, Li
    Xie, Wei
    Ma, Teng
    Zhang, Zhihong
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT IV, 2023, 14089 : 828 - 838
  • [39] Measuring and Improving Consistency in Pretrained Language Models
    Elazar, Yanai
    Kassner, Nora
    Ravfogel, Shauli
    Ravichander, Abhilasha
    Hovy, Eduard
    Schutze, Hinrich
    Goldberg, Yoav
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2021, 9 : 1012 - 1031
  • [40] An Embarrassingly Simple Approach for Transfer Learning from Pretrained Language Models
    Chronopoulou, Alexandra
    Baziotis, Christos
    Potamianos, Alexandros
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 2089 - 2095