Enhanced Seq2Seq Autoencoder via Contrastive Learning for Abstractive Text Summarization

被引:7
|
作者
Zheng, Chujie [1 ]
Zhang, Kunpeng [2 ]
Wang, Harry Jiannan [1 ]
Fan, Ling [3 ,4 ]
Wang, Zhe [4 ]
机构
[1] Univ Delaware, Newark, DE 19716 USA
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Tongji Univ, Shanghai, Peoples R China
[4] Tezign Com, Shanghai, Peoples R China
关键词
Abstractive Text Summarization; Contrastive Learning; Data Augmentation; Seq2seq;
D O I
10.1109/BigData52589.2021.9671819
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we present a denoising sequence-to-sequence (seq2seq) autoencoder via contrastive learning for abstractive text summarization. Our model adopts a standard Transformer-based architecture with a multi-layer bi-directional encoder and an auto-regressive decoder. To enhance its denoising ability, we incorporate self-supervised contrastive learning along with various sentence-level document augmentation. These two components, seq2seq autoencoder and contrastive learning, are jointly trained through fine-tuning, w hich i mproves t he performance of text summarization with regard to ROUGE scores and human evaluation. We conduct experiments on two datasets and demonstrate that our model outperforms many existing benchmarks and even achieves comparable performance to the state-of-the-art abstractive systems trained with more complex architecture and extensive computation resources.
引用
收藏
页码:1764 / 1771
页数:8
相关论文
共 50 条
  • [21] CoLRP: A Contrastive Learning Abstractive Text Summarization Method with ROUGE Penalty
    Tan, Caidong
    Sun, Xiao
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [22] Open-Domain Table-to-Text Generation based on Seq2seq
    Cao, Juan
    Gong, Junpeng
    Zhang, Pengzhou
    2018 INTERNATIONAL CONFERENCE ON ALGORITHMS, COMPUTING AND ARTIFICIAL INTELLIGENCE (ACAI 2018), 2018,
  • [23] Sparsing and Smoothing for the seq2seq Models
    Zhao S.
    Liang Z.
    Wen J.
    Chen J.
    IEEE Transactions on Artificial Intelligence, 2023, 4 (03): : 464 - 472
  • [24] Fine Grained Named Entity Recognition via Seq2seq Framework
    Zhu, Huiming
    He, Chunhui
    Fang, Yang
    Xiao, Weidong
    IEEE ACCESS, 2020, 8 : 53953 - 53961
  • [25] Energy Investment Risk Assessment for Nations Via Seq2seq Model
    Liang, Panning
    Yu, Mingyang
    Jiang, Lincheng
    FRONTIERS IN ENVIRONMENTAL SCIENCE, 2021, 9
  • [26] Improving Seq2Seq Grammatical Error Correction via Decoding Interventions
    Zhou, Houquan
    Liu, Yumeng
    Li, Zhenghua
    Zhang, Min
    Zhang, Bo
    Li, Chen
    Zhang, Ji
    Huang, Fei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 7393 - 7405
  • [27] Unsupervised Abstractive Text Summarization with Length Controlled Autoencoder
    Dugar, Abhinav
    Singh, Gaurav
    Navyasree, B.
    Kumar, Anand M.
    2022 IEEE 19TH INDIA COUNCIL INTERNATIONAL CONFERENCE, INDICON, 2022,
  • [28] Seq2Seq Imitation Learning for Tactile Feedback-based Manipulation
    Yang, Wenyan
    Angleraud, Alexandre
    Pieters, Roel S.
    Pajarinen, Joni
    Kamarainen, Joni-Kristian
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 5829 - 5836
  • [29] Linguistic Descriptions of Human Motion with Generative Adversarial Seq2Seq Learning
    Goutsu, Yusuke
    Inamura, Tetsunari
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 4281 - 4287
  • [30] Powering Predictive Treatment Planning by a Seq2seq Deep Learning Predictor
    Lee, D.
    Hu, Y.
    Kuo, L.
    Alam, S.
    Yorke, E.
    Rimner, A.
    Zhang, P.
    MEDICAL PHYSICS, 2021, 48 (06)