A two-stage fine-tuning method for low-resource cross-lingual summarization

被引:3
|
作者
Zhang, Kaixiong [1 ,2 ]
Zhang, Yongbing [1 ,2 ]
Yu, Zhengtao [1 ,2 ]
Huang, Yuxin [1 ,2 ]
Tan, Kaiwen [1 ,2 ]
机构
[1] Kunming Univ Sci & Technol, Fac Informat Engn & Automat, Kunming 650500, Peoples R China
[2] Kunming Univ Sci & Technol, Yunnan Key Lab Artificial Intelligence, Kunming 650500, Peoples R China
基金
中国国家自然科学基金;
关键词
cross-lingual; low-resource; summarization; fine-tuning;
D O I
10.3934/mbe.2024047
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Cross-lingual summarization (CLS) is the task of condensing lengthy source language text into a concise summary in a target language. This presents a dual challenge, demanding both cross-language semantic understanding (i.e., semantic alignment) and effective information compression capabilities. Traditionally, researchers have tackled these challenges using two types of methods: pipeline methods (e.g., translate-then-summarize) and end-to-end methods. The former is intuitive but prone to error propagation, particularly for low-resource languages. The later has shown an impressive performance, due to multilingual pre-trained models (mPTMs). However, mPTMs (e.g., mBART) are primarily trained on resource-rich languages, thereby limiting their semantic alignment capabilities for low-resource languages. To address these issues, this paper integrates the intuitiveness of pipeline methods and the effectiveness of mPTMs, and then proposes a two-stage fine-tuning method for low-resource cross-lingual summarization (TFLCLS). In the first stage, by recognizing the deficiency in the semantic alignment for low-resource languages in mPTMs, a semantic alignment fine-tuning method is employed to enhance the mPTMs' understanding of such languages. In the second stage, while considering that mPTMs are not originally tailored for information compression and CLS demands the model to simultaneously align and compress, an adaptive joint fine-tuning method is introduced. This method further enhances the semantic alignment and information compression abilities of mPTMs that were trained in the first stage. To evaluate the performance of TFLCLS, a low-resource CLS dataset, named Vi2ZhLow, is constructed from scratch; moreover, two additional low-resource CLS datasets, En2ZhLow and Zh2EnLow, are synthesized from widely used large-scale CLS datasets. Experimental results show that TFCLS outperforms state-of-the-art methods by 18.88%, 12.71% and 16.91% in ROUGE-2 on the three datasets, respectively, even when limited with only 5,000 training samples.
引用
收藏
页码:1125 / 1143
页数:19
相关论文
共 50 条
  • [1] Cross-Lingual Lemmatization and Morphology Tagging with Two-Stage Multilingual BERT Fine-Tuning
    Kondratyuk, Dan
    16TH SIGMORPHON WORKSHOP ON COMPUTATIONAL RESEARCH IN PHONETICS PHONOLOGY, AND MORPHOLOGY (SIGMORPHON 2019), 2019, : 12 - 18
  • [2] Consistency Regularization for Cross-Lingual Fine-Tuning
    Zheng, Bo
    Dong, Li
    Huang, Shaohan
    Wang, Wenhui
    Chi, Zewen
    Singhal, Saksham
    Che, Wanxiang
    Liu, Ting
    Song, Xia
    Wei, Furu
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 3403 - 3417
  • [3] Composable Sparse Fine-Tuning for Cross-Lingual Transfer
    Ansell, Alan
    Ponti, Edoardo Maria
    Korhonen, Anna
    Vulic, Ivan
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1778 - 1796
  • [4] Effective Fine-tuning Methods for Cross-lingual Adaptation
    Yu, Tao
    Joty, Shafiq
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 8492 - 8501
  • [5] Enhancing low-resource cross-lingual summarization from noisy data with fine-grained reinforcement learning
    Huang, Yuxin
    Gu, Huailing
    Yu, Zhengtao
    Gao, Yumeng
    Pan, Tong
    Xu, Jialong
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2024, 25 (01) : 121 - 134
  • [6] Cross-Lingual Summarization Method Based on Joint Training and Self-Training in Low-Resource Scenarios
    Cheng, Shaohuan
    Tang, Yujia
    Liu, Qiao
    Chen, Wenyu
    Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2024, 53 (05): : 762 - 770
  • [7] Augmenting Low-Resource Cross-Lingual Summarization with Progression-Grounded Training and Prompting
    Ma, Jiu Shun
    Huang, Yuxin
    Wang, Linqin
    Huang, Xiang
    Peng, Hao
    Yu, Zhengtao
    Yu, Philip
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (09)
  • [8] Cross-lingual embedding for cross-lingual question retrieval in low-resource community question answering
    HajiAminShirazi, Shahrzad
    Momtazi, Saeedeh
    MACHINE TRANSLATION, 2020, 34 (04) : 287 - 303
  • [9] Cross-Lingual Morphological Tagging for Low-Resource Languages
    Buys, Jan
    Botha, Jan A.
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2016, : 1954 - 1964
  • [10] Cross-lingual Intermediate Fine-tuning improves Dialogue State Tracking
    Moghe, Nikita
    Steedman, Mark
    Birch, Alexandra
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1137 - 1150