Mixed-Lingual Pre-training for Cross-lingual Summarization

被引:0
|
作者
Xu, Ruochen [1 ]
Zhu, Chenguang [1 ]
Shi, Yu [1 ]
Zeng, Michael [1 ]
Huang, Xuedong [1 ]
机构
[1] Microsoft Cognit Serv Res Grp, Redmond, WA 98052 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-lingual Summarization (CLS) aims at producing a summary in the target language for an article in the source language. Traditional solutions employ a twostep approach, i.e. translate!summarize or summarize!translate. Recently, end-to-end models have achieved better results, but these approaches are mostly limited by their dependence on large-scale labeled data. We propose a solution based on mixed-lingual pretraining that leverages both cross-lingual tasks such as translation and monolingual tasks like masked language models. Thus, our model can leverage the massive monolingual data to enhance its modeling of language. Moreover, the architecture has no task-specific components, which saves memory and increases optimization efficiency. We show in experiments that this pre-training scheme can effectively boost the performance of cross-lingual summarization. In Neural Cross-Lingual Summarization (NCLS) (Zhu et al., 2019b) dataset, our model achieves an improvement of 2.82 (English to Chinese) and 1.15 (Chinese to English) ROUGE-1 scores over state-of-the-art results.
引用
收藏
页码:536 / 541
页数:6
相关论文
共 50 条
  • [41] clstk: The Cross-Lingual Summarization Tool-Kit
    Jhaveri, Nisarg
    Gupta, Manish
    Varma, Vasudeva
    PROCEEDINGS OF THE TWELFTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING (WSDM'19), 2019, : 766 - 769
  • [42] Med-UniC: Unifying Cross-Lingual Medical Vision-Language Pre-Training by Diminishing Bias
    Wan, Zhongwei
    Liu, Che
    Zhang, Mi
    Fu, Jie
    Wang, Benyou
    Cheng, Sibo
    Ma, Lei
    Quilodran-Casas, Cesar
    Arcucci, Rossella
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [43] Multilingual Multimodal Pre-training for Zero-Shot Cross-Lingual Transfer of Vision-Language Models
    Huang, Po-Yao
    Patrick, Mandela
    Hu, Junjie
    Neubig, Graham
    Metze, Florian
    Hauptmann, Alexander
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 2443 - 2459
  • [44] EMMA- X: An EM-like Multilingual Pre-training Algorithm for Cross-lingual Representation Learning
    Guo, Ping
    Wei, Xiangpeng
    Hu, Yue
    Yang, Baosong
    Liu, Dayiheng
    Huang, Fei
    Xie, Jun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [45] Unified Training for Cross-Lingual Abstractive Summarization by Aligning Parallel Machine Translation Pairs
    Cheng, Shaohuan
    Chen, Wenyu
    Tang, Yujia
    Fu, Mingsheng
    Qu, Hong
    MATHEMATICS, 2024, 12 (13)
  • [46] Cross-lingual Cross-temporal Summarization: Dataset, Models, Evaluation
    Zhang, Ran
    Ouni, Jihed
    Eger, Steffen
    COMPUTATIONAL LINGUISTICS, 2024, 50 (03) : 1001 - 1047
  • [47] Cross-Lingual Training for Automatic Question Generation
    Kumar, Vishwajeet
    Joshi, Nitish
    Mukherjee, Arijit
    Ramakrishnan, Ganesh
    Jyothi, Preethi
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 4863 - 4872
  • [48] Multi-Task Learning for Cross-Lingual Abstractive Summarization
    Takase, Sho
    Okazaki, Naoaki
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 3008 - 3016
  • [49] A Comprehensive Survey and Prospect of Cross-Lingual Summarization Method Research
    Wang, Jing-Dong
    Chang, Duo
    Meng, Fan-Qi
    Qu, Guangqiang
    Journal of Network Intelligence, 2024, 9 (01): : 384 - 412
  • [50] Multi-Task Learning for Cross-Lingual Abstractive Summarization
    Takase, Sho
    Okazaki, Naoaki
    2022 Language Resources and Evaluation Conference, LREC 2022, 2022, : 3008 - 3016