Cross-lingual distillation for domain knowledge transfer with sentence transformers

被引:0
|
作者
Piperno, Ruben [1 ,2 ]
Bacco, Luca [1 ,3 ]
Dell'Orletta, Felice [1 ]
Merone, Mario [2 ]
Pecchia, Leandro [2 ,4 ]
机构
[1] Inst Computat Linguist Antonio Zampolli, Natl Res Council, ItaliaNLP Lab, Via Giuseppe Moruzzi, 1, I-56124 Pisa, Italy
[2] Univ Campus Biomed Roma, Dept Engn, Res Unit Intelligent Technol Hlth & Wellbeing, Via Alvaro Portillo 21, I-00128 Rome, Italy
[3] Univ Campus Biomed Roma, Dept Engn, Res Unit Comp Syst & Bioinformat, Via Alvaro Portillo 21, I-00128 Rome, Italy
[4] Fdn Policlin Univ Campus Biomed Roma, Via Alvaro del Portillo 200, I-00128 Rome, Italy
关键词
Cross-lingual learning; Knowledge distillation; Sentence transformers; Biomedical domain; Domain adaptation;
D O I
10.1016/j.knosys.2025.113079
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advancements in Natural Language Processing (NLP) have substantially enhanced language understanding. However, non-English languages, especially in specialized and low-resource domains like biomedicine, remain largely underrepresented. Bridging this gap is essential for promoting inclusivity and expanding the global applicability of NLP technologies. This study presents a cross-lingual knowledge distillation framework that utilizes sentence transformers to improve domain-specific NLP capabilities in non-English languages. Specifically, the framework focuses on biomedical text classification tasks. By aligning sentence embeddings between a teacher model trained on English biomedical corpora and a multilingual student model, the proposed method effectively transfers both domain-specific and task-specific knowledge. This alignment allows the student model to efficiently process and adapt to biomedical texts in Spanish, French, and German, particularly in low-resource settings with limited tuning data. Extensive experiments with domain-adapted models like BioBERT and multilingual BERT with machine-translated text pairs demonstrate substantial performance improvements in downstream biomedical NLP tasks. The proposed framework proves highly effective in scenarios characterized by limited training data availability. The results highlight the scalability and effectiveness of this approach, facilitating the development of robust multilingual models tailored to the biomedical domain, thus advancing global accessibility and impact in biomedical NLP applications.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Aligning Cross-lingual Sentence Representations with Dual Momentum Contrast
    Wang, Liang
    Zhao, Wei
    Liu, Jingming
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3807 - 3815
  • [42] Domain Mismatch Doesn't Always Prevent Cross-Lingual Transfer Learning
    Edmiston, Daniel
    Keung, Phillip
    Smith, Noah A.
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 892 - 899
  • [43] Cross-lingual font style transfer with full-domain convolutional attention
    Zhao, Hui-huang
    Ji, Tian-le
    Rosin, Paul L.
    Lai, Yu-Kun
    Meng, Wei-liang
    Wang, Yao-nan
    PATTERN RECOGNITION, 2024, 155
  • [44] ABSent: Cross-Lingual Sentence Representation Mapping with Bidirectional GANs
    Fu, Zuohui
    Xian, Yikun
    Geng, Shijie
    Ge, Yingqiang
    Wang, Yuting
    Dong, Xin
    Wang, Guang
    de Melo, Gerard
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7756 - 7763
  • [45] Cross-lingual thesaurus for multilingual knowledge management
    Yang, Christopher C.
    Wei, Chih-Ping
    Li, K. W.
    DECISION SUPPORT SYSTEMS, 2008, 45 (03) : 596 - 605
  • [46] Cross-lingual sentiment transfer with limited resources
    Rasooli, Mohammad Sadegh
    Farra, Noura
    Radeva, Axinia
    Yu, Tao
    McKeown, Kathleen
    MACHINE TRANSLATION, 2018, 32 (1-2) : 143 - 165
  • [47] Translation Artifacts in Cross-lingual Transfer Learning
    Artetxe, Mikel
    Labaka, Gorka
    Agirre, Eneko
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7674 - 7684
  • [48] Distillation Language Adversarial Network for Cross-lingual Sentiment Analysis
    Wang, Deheng
    Yang, Aimin
    Zhou, Yongmei
    Xie, Fenfang
    Ouyang, Zhouhao
    Peng, Sancheng
    2022 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2022), 2022, : 45 - 50
  • [49] Choosing Transfer Languages for Cross-Lingual Learning
    Lin, Yu-Hsiang
    Chen, Chian-Yu
    Lee, Jean
    Li, Zirui
    Zhang, Yuyan
    Xia, Mengzhou
    Rijhwani, Shruti
    He, Junxian
    Zhang, Zhisong
    Ma, Xuezhe
    Anastasopoulos, Antonios
    Littell, Patrick
    Neubig, Graham
    Anastasopoulos, Antonios
    Littell, Patrick
    Neubig, Graham
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3125 - 3135
  • [50] Multilingual Knowledge Graph Embeddings for Cross-lingual Knowledge Alignment
    Chen, Muhao
    Tian, Yingtao
    Yang, Mohan
    Zaniolo, Carlo
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1511 - 1517