Cross-lingual distillation for domain knowledge transfer with sentence transformers

被引:0
|
作者
Piperno, Ruben [1 ,2 ]
Bacco, Luca [1 ,3 ]
Dell'Orletta, Felice [1 ]
Merone, Mario [2 ]
Pecchia, Leandro [2 ,4 ]
机构
[1] Inst Computat Linguist Antonio Zampolli, Natl Res Council, ItaliaNLP Lab, Via Giuseppe Moruzzi, 1, I-56124 Pisa, Italy
[2] Univ Campus Biomed Roma, Dept Engn, Res Unit Intelligent Technol Hlth & Wellbeing, Via Alvaro Portillo 21, I-00128 Rome, Italy
[3] Univ Campus Biomed Roma, Dept Engn, Res Unit Comp Syst & Bioinformat, Via Alvaro Portillo 21, I-00128 Rome, Italy
[4] Fdn Policlin Univ Campus Biomed Roma, Via Alvaro del Portillo 200, I-00128 Rome, Italy
关键词
Cross-lingual learning; Knowledge distillation; Sentence transformers; Biomedical domain; Domain adaptation;
D O I
10.1016/j.knosys.2025.113079
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advancements in Natural Language Processing (NLP) have substantially enhanced language understanding. However, non-English languages, especially in specialized and low-resource domains like biomedicine, remain largely underrepresented. Bridging this gap is essential for promoting inclusivity and expanding the global applicability of NLP technologies. This study presents a cross-lingual knowledge distillation framework that utilizes sentence transformers to improve domain-specific NLP capabilities in non-English languages. Specifically, the framework focuses on biomedical text classification tasks. By aligning sentence embeddings between a teacher model trained on English biomedical corpora and a multilingual student model, the proposed method effectively transfers both domain-specific and task-specific knowledge. This alignment allows the student model to efficiently process and adapt to biomedical texts in Spanish, French, and German, particularly in low-resource settings with limited tuning data. Extensive experiments with domain-adapted models like BioBERT and multilingual BERT with machine-translated text pairs demonstrate substantial performance improvements in downstream biomedical NLP tasks. The proposed framework proves highly effective in scenarios characterized by limited training data availability. The results highlight the scalability and effectiveness of this approach, facilitating the development of robust multilingual models tailored to the biomedical domain, thus advancing global accessibility and impact in biomedical NLP applications.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] cViL: Cross-Lingual Training of Vision-Language Models using Knowledge Distillation
    Gupta, Kshitij
    Gautam, Devansh
    Mamidi, Radhika
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 1734 - 1741
  • [32] cViL: Cross-Lingual Training of Vision-Language Models using Knowledge Distillation
    Gupta, Kshitij
    Gautam, Devansh
    Mamidi, Radhika
    Proceedings - International Conference on Pattern Recognition, 2022, 2022-August : 1734 - 1741
  • [33] Analyzing the Evaluation of Cross-Lingual Knowledge Transfer in Multilingual Language Models
    Rajaee, Sara
    Monz, Christof
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 2895 - 2914
  • [34] A Multi-media Approach to Cross-lingual Entity Knowledge Transfer
    Lu, Di
    Pan, Xiaoman
    Pourdamghani, Nima
    Chang, Shih-Fu
    Ji, Heng
    Knight, Kevin
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 54 - 65
  • [35] Chinese-Vietnamese cross-lingual event retrieval method based on knowledge distillation
    Gao S.
    He Z.
    Yu Z.
    Zhu E.
    Wu S.
    Journal of Intelligent and Fuzzy Systems, 2024, 46 (04): : 8461 - 8475
  • [36] Cross-lingual Transfer of Monolingual Models
    Gogoulou, Evangelia
    Ekgren, Ariel
    Isbister, Tim
    Sahlgren, Magnus
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 948 - 955
  • [37] Model Selection for Cross-Lingual Transfer
    Chen, Yang
    Ritter, Alan
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 5675 - 5687
  • [38] Can Cross-Domain Term Extraction Benefit from Cross-lingual Transfer?
    Tran, Hanh Thi Hong
    Martinc, Matej
    Doucet, Antoine
    Pollak, Senja
    DISCOVERY SCIENCE (DS 2022), 2022, 13601 : 363 - 378
  • [39] A Cross-Lingual Sentence Similarity Calculation Method With Multifeature Fusion
    Wang, Lingxin
    Liu, Shengquan
    Qiao, Longye
    Sun, Weiwei
    Sun, Qi
    Cheng, Huaqing
    IEEE ACCESS, 2022, 10 : 30666 - 30675
  • [40] PHONETIC NAME MATCHING FOR CROSS-LINGUAL SPOKEN SENTENCE RETRIEVAL
    Ji, Heng
    Grishman, Ralph
    Wang, Wen
    2008 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY: SLT 2008, PROCEEDINGS, 2008, : 281 - +