Enhanced Seq2Seq Autoencoder via Contrastive Learning for Abstractive Text Summarization

被引:7
|
作者
Zheng, Chujie [1 ]
Zhang, Kunpeng [2 ]
Wang, Harry Jiannan [1 ]
Fan, Ling [3 ,4 ]
Wang, Zhe [4 ]
机构
[1] Univ Delaware, Newark, DE 19716 USA
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Tongji Univ, Shanghai, Peoples R China
[4] Tezign Com, Shanghai, Peoples R China
关键词
Abstractive Text Summarization; Contrastive Learning; Data Augmentation; Seq2seq;
D O I
10.1109/BigData52589.2021.9671819
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a denoising sequence-to-sequence (seq2seq) autoencoder via contrastive learning for abstractive text summarization. Our model adopts a standard Transformer-based architecture with a multi-layer bi-directional encoder and an auto-regressive decoder. To enhance its denoising ability, we incorporate self-supervised contrastive learning along with various sentence-level document augmentation. These two components, seq2seq autoencoder and contrastive learning, are jointly trained through fine-tuning, w hich i mproves t he performance of text summarization with regard to ROUGE scores and human evaluation. We conduct experiments on two datasets and demonstrate that our model outperforms many existing benchmarks and even achieves comparable performance to the state-of-the-art abstractive systems trained with more complex architecture and extensive computation resources.
引用
收藏
页码:1764 / 1771
页数:8
相关论文
共 50 条
  • [41] History-based attention in Seq2Seq model for multi-label text classification
    Xiao, Yaoqiang
    Li, Yi
    Yuan, Jin
    Guo, Songrui
    Xiao, Yi
    Li, Zhiyong
    KNOWLEDGE-BASED SYSTEMS, 2021, 224
  • [42] Residual Seq2Seq model for Building energy management
    Kim, Marie
    Kim, Nae-soo
    Song, YuJin
    Pyo, Cheol Sig
    2019 10TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY CONVERGENCE (ICTC): ICT CONVERGENCE LEADING THE AUTONOMOUS FUTURE, 2019, : 1126 - 1128
  • [43] Automatic Generation of Pseudocode with Attention Seq2seq Model
    Xu, Shaofeng
    Xiong, Yun
    2018 25TH ASIA-PACIFIC SOFTWARE ENGINEERING CONFERENCE (APSEC 2018), 2018, : 711 - 712
  • [44] Map Matching Based on Seq2Seq with Topology Information
    Bai, Yulong
    Li, Guolian
    Lu, Tianxiu
    Wu, Yadong
    Zhang, Weihan
    Feng, Yidan
    APPLIED SCIENCES-BASEL, 2023, 13 (23):
  • [45] Seq2Seq模型的短期水位预测
    刘艳
    张婷
    康爱卿
    李建柱
    雷晓辉
    水利水电科技进展, 2022, 42 (03) : 57 - 63
  • [46] Smoothing and Shrinking the Sparse Seq2Seq Search Space
    Peters, Ben
    Martins, Andre F. T.
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 2642 - 2654
  • [47] Sliding Window Seq2seq Modeling for Engagement Estimation
    Yu, Jun
    Lu, Keda
    Jing, Mohan
    Liang, Ziqi
    Zhang, Bingyuan
    Sun, Jianqing
    Liang, Jiaen
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 9496 - 9500
  • [48] SparQL Query Prediction Based on Seq2Seq Model
    Yang D.-H.
    Zou K.-F.
    Wang H.-Z.
    Wang J.-B.
    Ruan Jian Xue Bao/Journal of Software, 2021, 32 (03): : 805 - 817
  • [49] Untargeted Code Authorship Evasion with Seq2Seq Transformation
    Choi, Soohyeon
    Jang, Rhongho
    Nyang, DaeHun
    Mohaisen, David
    arXiv, 2023,
  • [50] Exaggerated Portrait Caricatures Generation Based On Seq2Seq
    Xu, Kun
    Tang, Chenwei
    Lv, Jiancheng
    He, Zhenan
    2019 9TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND TECHNOLOGY (ICIST2019), 2019, : 36 - 44