A Hierarchical Encoding-Decoding Scheme for Abstractive Multi-document Summarization

被引:0
|
作者
Shen, Chenhui [1 ,2 ]
Cheng, Liying [1 ,3 ]
Xuan-Phi Nguyen [1 ,3 ]
You, Yang [2 ]
Bing, Lidong [1 ,3 ]
机构
[1] Alibaba Grp, DAMO Acad, Singapore, Singapore
[2] Natl Univ Singapore, Singapore, Singapore
[3] Hupan Lab, Hangzhou 310023, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pre-trained language models (PLMs) have achieved outstanding achievements in abstractive single-document summarization (SDS). However, such benefits may not fully extend to multi-document summarization (MDS), where the handling of cross-document information is more complex. Previous works either design new MDS architectures or apply PLMs bluntly with concatenated source documents as a reformulated SDS task. While the former does not utilize previous pre-training efforts and may not generalize well across different domains, the latter may not sufficiently attend to the intricate cross-document relationships unique to MDS tasks. Instead, we enforce hierarchy on both the encoder and decoder to better utilize a PLM to facilitate multi-document interactions for the MDS task. Across 10 MDS benchmarks from various domains, our method outperforms or is competitive with the previous best models, including those with additional MDS pre-training or with more parameters. It outperforms its corresponding PLM backbone by up to 3 ROUGEL and is favored by humans.(1)
引用
收藏
页码:5872 / 5887
页数:16
相关论文
共 50 条
  • [31] Improving Abstractive Multi-document Summarization with Predicate-Argument Structure Extraction
    Cheng, Huangfei
    Wu, Jiawei
    Li, Tiantian
    Cao, Bin
    Fan, Jing
    PRICAI 2022: TRENDS IN ARTIFICIAL INTELLIGENCE, PT II, 2022, 13630 : 268 - 282
  • [32] A Hybrid Solution To Abstractive Multi-Document Summarization Using Supervised and Unsupervised Learning
    Bhagchandani, Gaurav
    Bodra, Deep
    Gangan, Abhishek
    Mulla, Nikahat
    PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND CONTROL SYSTEMS (ICCS), 2019, : 566 - 570
  • [33] ViMs: a high-quality Vietnamese dataset for abstractive multi-document summarization
    Tran, Nhi-Thao
    Nghiem, Minh-Quoc
    Nguyen, Nhung T. H.
    Nguyen, Ngan Luu-Thuy
    Van Chi, Nam
    Dinh, Dien
    LANGUAGE RESOURCES AND EVALUATION, 2020, 54 (04) : 893 - 920
  • [34] ViMs: a high-quality Vietnamese dataset for abstractive multi-document summarization
    Nhi-Thao Tran
    Minh-Quoc Nghiem
    Nhung T. H. Nguyen
    Ngan Luu-Thuy Nguyen
    Nam Van Chi
    Dien Dinh
    Language Resources and Evaluation, 2020, 54 : 893 - 920
  • [35] A Multi-Document Coverage Reward for RELAXed Multi-Document Summarization
    Parnell, Jacob
    Unanue, Inigo Jauregi
    Piccardi, Massimo
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5112 - 5128
  • [36] Automatic Multi-Document Summarization for Indonesian Documents Using Hybrid Abstractive-Extractive Summarization Technique
    Yapinus, Glorian
    Erwin, Alva
    Galinium, Maulahikmah
    Muliady, Wahyu
    2014 6TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY AND ELECTRICAL ENGINEERING (ICITEE), 2014, : 39 - 43
  • [37] Boosting multi-document summarization with hierarchical graph convolutional networks
    Song, Yingjie
    Yang, Li
    Luo, Wenming
    Xiao, Xiong
    Tang, Zhuo
    NEUROCOMPUTING, 2025, 614
  • [38] MULTI-DOCUMENT VIDEO SUMMARIZATION
    Wang, Feng
    Merialdo, Bernard
    ICME: 2009 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-3, 2009, : 1326 - 1329
  • [39] Multi-Document Abstractive Summarization using Chunk-graph and Recurrent Neural Network
    Niu, Jianwei
    Chen, Huan
    Zhao, Qingjuan
    Sun, Limin
    Atiquzzaman, Mohammed
    2017 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2017,
  • [40] On redundancy in multi-document summarization
    Calvo, Hiram
    Carrillo-Mendoza, Pabel
    Gelbukh, Alexander
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2018, 34 (05) : 3245 - 3255