Mixed-Lingual Pre-training for Cross-lingual Summarization

被引:0
|
作者
Xu, Ruochen [1 ]
Zhu, Chenguang [1 ]
Shi, Yu [1 ]
Zeng, Michael [1 ]
Huang, Xuedong [1 ]
机构
[1] Microsoft Cognit Serv Res Grp, Redmond, WA 98052 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-lingual Summarization (CLS) aims at producing a summary in the target language for an article in the source language. Traditional solutions employ a twostep approach, i.e. translate!summarize or summarize!translate. Recently, end-to-end models have achieved better results, but these approaches are mostly limited by their dependence on large-scale labeled data. We propose a solution based on mixed-lingual pretraining that leverages both cross-lingual tasks such as translation and monolingual tasks like masked language models. Thus, our model can leverage the massive monolingual data to enhance its modeling of language. Moreover, the architecture has no task-specific components, which saves memory and increases optimization efficiency. We show in experiments that this pre-training scheme can effectively boost the performance of cross-lingual summarization. In Neural Cross-Lingual Summarization (NCLS) (Zhu et al., 2019b) dataset, our model achieves an improvement of 2.82 (English to Chinese) and 1.15 (Chinese to English) ROUGE-1 scores over state-of-the-art results.
引用
收藏
页码:536 / 541
页数:6
相关论文
共 50 条
  • [21] Product-oriented Machine Translation with Cross-modal Cross-lingual Pre-training
    Song, Yuqing
    Chen, Shizhe
    Jin, Qin
    Luo, Wei
    Xie, Jun
    Huang, Fei
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2843 - 2852
  • [22] XLM-E: Cross-lingual Language Model Pre-training via ELECTRA
    Chi, Zewen
    Huang, Shaohan
    Dong, Li
    Ma, Shuming
    Zheng, Bo
    Singhal, Saksham
    Bajaj, Payal
    Song, Xia
    Mao, Xian-Ling
    Huang, Heyan
    Wei, Furu
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 6170 - 6182
  • [23] INFOXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training
    Chi, Zewen
    Dong, Li
    Wei, Furu
    Yang, Nan
    Singhal, Saksham
    Wang, Wenhui
    Song, Xia
    Mao, Xian-Ling
    Huang, Heyan
    Zhou, Ming
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3576 - 3588
  • [24] PTEKC: pre-training with event knowledge of ConceptNet for cross-lingual event causality identification
    Zhu, Enchang
    Yu, Zhengtao
    Huang, Yuxin
    Gao, Shengxiang
    Xian, Yantuan
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2025, 16 (03) : 1859 - 1872
  • [25] A Cross-Lingual Summarization method based on cross-lingual Fact-relationship Graph Generation
    Zhang, Yongbing
    Gao, Shengxiang
    Huang, Yuxin
    Tan, Kaiwen
    Yu, Zhengtao
    PATTERN RECOGNITION, 2024, 146
  • [26] Cross-Lingual Speech-to-Text Summarization
    Pontes, Elvys Linhares
    Gonzalez-Gallardo, Carlos-Emiliano
    Torres-Moreno, Juan-Manuel
    Huet, Stephane
    MULTIMEDIA AND NETWORK INFORMATION SYSTEMS, 2019, 833 : 385 - 395
  • [27] Cross-lingual extreme summarization of scholarly documents
    Takeshita, Sotaro
    Green, Tommaso
    Friedrich, Niklas
    Eckert, Kai
    Ponzetto, Simone Paolo
    INTERNATIONAL JOURNAL ON DIGITAL LIBRARIES, 2024, 25 (02) : 249 - 271
  • [28] A Robust Abstractive System for Cross-Lingual Summarization
    Ouyang, Jessica
    Song, Boya
    McKeown, Kathleen
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 2025 - 2031
  • [29] Few-Shot Cross-Lingual Stance Detection with Sentiment-Based Pre-training
    Hardalov, Momchil
    Arora, Arnav
    Nakov, Preslav
    Augenstein, Isabelle
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10729 - 10737
  • [30] XLM-K: Improving Cross-Lingual Language Model Pre-training with Multilingual Knowledge
    Jiang, Xiaoze
    Liang, Yaobo
    Chen, Weizhu
    Duan, Nan
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10840 - 10848