KBioXLM: A Knowledge-anchored Biomedical Multilingual Pretrained Language Model

被引:0
|
作者
Geng, Lei [1 ]
Yan, Xu [1 ]
Cao, Ziqiang [1 ]
Li, Juntao [1 ]
Li, Wenjie [3 ]
Li, Sujian [2 ]
Zhou, Xinjie [4 ]
Yang, Yang [4 ]
Zhang, Jun [5 ]
机构
[1] Soochow Univ, Inst Artificial Intelligence, Suzhou, Peoples R China
[2] Peking Univ, Beijing, Peoples R China
[3] Hong Kong Polytech Univ, Hong Kong, Peoples R China
[4] Pharmcube, Beijing, Peoples R China
[5] Changping Lab, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
CORPUS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most biomedical pretrained language models are monolingual and cannot handle the growing cross-lingual requirements. The scarcity of non-English domain corpora, not to mention parallel data, poses a significant hurdle in training multilingual biomedical models. Since knowledge forms the core of domain-specific corpora and can be translated into various languages accurately, we propose a model called KBioXLM, which transforms the multilingual pretrained model XLM-R into the biomedical domain using a knowledge-anchored approach. We achieve a biomedical multilingual corpus by incorporating three granularity knowledge alignments (entity, fact, and passage levels) into monolingual corpora. Then we design three corresponding training tasks (entity masking, relation masking, and passage relation prediction) and continue training on top of the XLM-R model to enhance its domain crosslingual ability. To validate the effectiveness of our model, we translate the English benchmarks of multiple tasks into Chinese. Experimental results demonstrate that our model significantly outperforms monolingual and multilingual pretrained models in cross-lingual zero-shot and few-shot scenarios, achieving improvements of up to 10+ points. Our code is publicly available at https://github.com/ ngwlh-gl/KBioXLM.
引用
收藏
页码:11239 / 11250
页数:12
相关论文
共 50 条
  • [31] Automatic Taxonomy Classification by Pretrained Language Model
    Kuwana, Ayato
    Oba, Atsushi
    Sawai, Ranto
    Paik, Incheon
    ELECTRONICS, 2021, 10 (21)
  • [32] Insights Into Multilingual Students to Capitalize on Their Language Knowledge
    Ittner, Anne C.
    Jennerjohn, Anna
    Helman, Lori
    READING TEACHER, 2025,
  • [33] Language model for multilingual natural language generation
    Zhang, Dongmo
    Ge, Yong
    Yao, Tianfang
    Shanghai Jiaotong Daxue Xuebao/Journal of Shanghai Jiaotong University, 2000, 34 (07): : 944 - 947
  • [34] IIITT at CASE 2021 Task 1: Leveraging Pretrained Language Models for Multilingual Protest Detection
    Jada, Pawan Kalyan
    Reddy, Duddukunta Sashidhar
    Hande, Adeep
    Priyadharshini, Ruba
    Sakuntharaj, Ratnasingam
    Chakravarthi, Bharathi Raja
    CASE 2021: THE 4TH WORKSHOP ON CHALLENGES AND APPLICATIONS OF AUTOMATED EXTRACTION OF SOCIO-POLITICAL EVENTS FROM TEXT (CASE), 2021, : 98 - 104
  • [35] Translate to Disambiguate: Zero-shot Multilingual Word Sense Disambiguation with Pretrained Language Models
    Kang, Haoqiang
    Blevins, Terra
    Zettlemoyer, Luke
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 1562 - 1575
  • [36] An Empirical Study on Adaptive Inference for Pretrained Language Model
    Liu, Weijie
    Zhao, Xin
    Zhao, Zhe
    Ju, Qi
    Yang, Xuefeng
    Lu, Wei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4321 - 4331
  • [37] Parameter-efficient online knowledge distillation for pretrained language models
    Wang, Yukun
    Wang, Jin
    Zhang, Xuejie
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
  • [38] DagoBERT: Generating Derivational Morphology with a Pretrained Language Model
    Hofmannt, Valentin
    Pierrehumbertt, Janet B.
    Schiitzet, Hinrich
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3848 - 3861
  • [39] INTEGRATING PRETRAINED LANGUAGE MODEL FOR DIALOGUE POLICY EVALUATION
    Wang, Hongru
    Wang, Huimin
    Wang, Zezhong
    Wong, Kam-Fai
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6692 - 6696
  • [40] Enhancing pretrained language models with structured commonsense knowledge for textual inference
    Du, Li
    Ding, Xiao
    Xiong, Kai
    Liu, Ting
    Qin, Bing
    KNOWLEDGE-BASED SYSTEMS, 2022, 254