Cross-lingual distillation for domain knowledge transfer with sentence transformers

被引:0
|
作者
Piperno, Ruben [1 ,2 ]
Bacco, Luca [1 ,3 ]
Dell'Orletta, Felice [1 ]
Merone, Mario [2 ]
Pecchia, Leandro [2 ,4 ]
机构
[1] Inst Computat Linguist Antonio Zampolli, Natl Res Council, ItaliaNLP Lab, Via Giuseppe Moruzzi, 1, I-56124 Pisa, Italy
[2] Univ Campus Biomed Roma, Dept Engn, Res Unit Intelligent Technol Hlth & Wellbeing, Via Alvaro Portillo 21, I-00128 Rome, Italy
[3] Univ Campus Biomed Roma, Dept Engn, Res Unit Comp Syst & Bioinformat, Via Alvaro Portillo 21, I-00128 Rome, Italy
[4] Fdn Policlin Univ Campus Biomed Roma, Via Alvaro del Portillo 200, I-00128 Rome, Italy
关键词
Cross-lingual learning; Knowledge distillation; Sentence transformers; Biomedical domain; Domain adaptation;
D O I
10.1016/j.knosys.2025.113079
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advancements in Natural Language Processing (NLP) have substantially enhanced language understanding. However, non-English languages, especially in specialized and low-resource domains like biomedicine, remain largely underrepresented. Bridging this gap is essential for promoting inclusivity and expanding the global applicability of NLP technologies. This study presents a cross-lingual knowledge distillation framework that utilizes sentence transformers to improve domain-specific NLP capabilities in non-English languages. Specifically, the framework focuses on biomedical text classification tasks. By aligning sentence embeddings between a teacher model trained on English biomedical corpora and a multilingual student model, the proposed method effectively transfers both domain-specific and task-specific knowledge. This alignment allows the student model to efficiently process and adapt to biomedical texts in Spanish, French, and German, particularly in low-resource settings with limited tuning data. Extensive experiments with domain-adapted models like BioBERT and multilingual BERT with machine-translated text pairs demonstrate substantial performance improvements in downstream biomedical NLP tasks. The proposed framework proves highly effective in scenarios characterized by limited training data availability. The results highlight the scalability and effectiveness of this approach, facilitating the development of robust multilingual models tailored to the biomedical domain, thus advancing global accessibility and impact in biomedical NLP applications.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Cross-Lingual Sentence Extraction for Information Distillation
    Singla, Adish Kumar
    Hakkani-Tuer, Dilek
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 2707 - 2710
  • [2] Cross-Lingual Knowledge Distillation for Answer Sentence Selection in Low-Resource Languages
    Gupta, Shivanshu
    Matsubara, Yoshitomo
    Chadha, Ankit
    Moschitti, Alessandro
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 14078 - 14092
  • [3] Cross-Lingual Knowledge Distillation for Chinese Video Captioning
    Hou J.-Y.
    Qi Y.-Y.
    Wu X.-X.
    Jia Y.-D.
    Jisuanji Xuebao/Chinese Journal of Computers, 2021, 44 (09): : 1907 - 1921
  • [4] Cross-Lingual Knowledge Transfer for Clinical Phenotyping
    Papaioannou, Jens-Michalis
    Grundmann, Paul
    van Aken, Betty
    Samaras, Athanasios
    Kyparissidis, Ilias
    Giannakoulas, George
    Gers, Felix
    Loeser, Alexander
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 900 - 909
  • [5] Zero-Shot Cross-Lingual Knowledge Transfer in VQA via Multimodal Distillation
    Weng, Yu
    Dong, Jun
    He, Wenbin
    Chaomurilige
    Liu, Xuan
    Liu, Zheng
    Gao, Honghao
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, : 1 - 11
  • [6] Knowledge Distillation Based Training of Universal ASR Source Models for Cross-lingual Transfer
    Fukuda, Takashi
    Thomas, Samuel
    INTERSPEECH 2021, 2021, : 3450 - 3454
  • [7] Reinforced Transformer with Cross-Lingual Distillation for Cross-Lingual Aspect Sentiment Classification
    Wu, Hanqian
    Wang, Zhike
    Qing, Feng
    Li, Shoushan
    ELECTRONICS, 2021, 10 (03) : 1 - 14
  • [8] Probing Cross-Lingual Lexical Knowledge from Multilingual Sentence Encoders
    Vulic, Ivan
    Glavas, Goran
    Liu, Fangyu
    Collier, Nigel
    Ponti, Edoardo Maria
    Korhonen, Anna
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 2089 - 2105
  • [9] Robust cross-lingual knowledge base question answering via knowledge distillation
    Wang, Shaofei
    Dang, Depeng
    DATA TECHNOLOGIES AND APPLICATIONS, 2021, 55 (05) : 661 - 681
  • [10] Monolingual and Cross-Lingual Knowledge Transfer for Topic Classification
    D. Karpov
    M. Burtsev
    Journal of Mathematical Sciences, 2024, 285 (1) : 36 - 48