DIONYSUS: A Pre-trained Model for Low-Resource Dialogue Summarization

被引:0
|
作者
Li, Yu [1 ,2 ]
Peng, Baolin [2 ]
He, Pengcheng [2 ]
Galley, Michel [2 ]
Yu, Zhou [1 ]
Gao, Jianfeng [2 ]
机构
[1] Columbia Univ, New York, NY 10027 USA
[2] Microsoft Res, Redmond, WA 98052 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dialogue summarization has recently garnered significant attention due to its wide range of applications. However, existing methods for summarizing dialogues have limitations because they do not take into account the inherent structure of dialogue and rely heavily on labeled data, which can lead to poor performance in new domains. In this work, we propose DIONYSUS (dynamic input optimization in pre-training for dialogue summarization), a pre-trained encoder-decoder model for summarizing dialogues in any new domain. To pretrain DIONYSUS, we create two pseudo summaries for each dialogue example: one from a fine-tuned summarization model and the other from important dialogue turns. We then choose one of these pseudo summaries based on information distribution differences in different types of dialogues. This selected pseudo summary serves as the objective for pre-training DIONYSUS using a self-supervised approach on a large dialogue corpus. Our experiments show that DIONYSUS outperforms existing methods on six datasets, as demonstrated by its ROUGE scores in zero-shot and few-shot settings.
引用
收藏
页码:1368 / 1386
页数:19
相关论文
共 50 条
  • [31] A Pre-Trained Language Model Based on LED for Tibetan Long Text Summarization
    Ouyang, Xinpeng
    Yan, Xiaodong
    Hao, Minghui
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 992 - 997
  • [32] Vision Enhanced Generative Pre-trained Language Model for Multimodal Sentence Summarization
    Jing, Liqiang
    Li, Yiren
    Xu, Junhao
    Yu, Yongcan
    Shen, Pei
    Song, Xuemeng
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (02) : 289 - 298
  • [33] Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization
    Chernyshev, Daniil
    Dobrov, Boris
    IEEE ACCESS, 2024, 12 : 47219 - 47230
  • [34] EFFICIENT UTILIZATION OF LARGE PRE-TRAINED MODELS FOR LOW RESOURCE ASR
    Vieting, Peter
    Luescher, Christoph
    Dierkes, Julian
    Schlueter, Ralf
    Ney, Hermann
    2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [35] Low-Resource Dialogue Summarization with Domain-Agnostic Multi-Source Pretraining
    Zou, Yicheng
    Zhu, Bolin
    Hu, Xingwu
    Gui, Tao
    Zhang, Qi
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 80 - 91
  • [36] Pre-trained Personalized Review Summarization with Effective Salience Estimation
    Xu, Hongyan
    Liu, Hongtao
    Lv, Zhepeng
    Yang, Qing
    Wang, Wenjun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 10743 - 10754
  • [37] Modeling Content Importance for Summarization with Pre-trained Language Models
    Xiao, Liqiang
    Lu Wang
    Hao He
    Jin, Yaohui
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3606 - 3611
  • [38] Pre-trained Word Embedding based Parallel Text Augmentation Technique for Low-Resource NMT in Favor of Morphologically Rich Languages
    Hailu, Tulu Tilahun
    Yu, Junqing
    Fantaye, Tessfu Geteye
    PROCEEDINGS OF THE THIRD INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND APPLICATION ENGINEERING (CSAE2019), 2019,
  • [39] Adapting Generative Pre-trained Language Model for Open-domain Multimodal Sentence Summarization
    Lin, Dengtian
    Jing, Liqiang
    Song, Xuemeng
    Liu, Meng
    Sun, Teng
    Nie, Liqiang
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 195 - 204
  • [40] Grounding Dialogue History: Strengths and Weaknesses of Pre-trained Transformers
    Greco, Claudio
    Testoni, Alberto
    Bernardi, Raffaella
    AIXIA 2020 - ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 12414 : 263 - 279