On Extractive and Abstractive Neural Document Summarization with Transformer Language Models

被引:0
|
作者
Pilault, Jonathan [1 ,2 ,3 ]
Li, Raymond [1 ]
Subramanian, Sandeep [1 ,2 ,4 ]
Pal, Christopher [1 ,2 ,3 ,4 ,5 ]
机构
[1] Element AI, Montreal, PQ, Canada
[2] Mila, Montreal, PQ, Canada
[3] Polytech Montreal, Montreal, PQ, Canada
[4] Univ Montreal, Montreal, PQ, Canada
[5] Canada CIFAR AI Chair, Montreal, PQ, Canada
来源
PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP) | 2020年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a method to produce abstractive summaries of long documents that exceed several thousand words via neural abstractive summarization. We perform a simple extractive step before generating a summary, which is then used to condition the transformer language model on relevant information before being tasked with generating a summary. We also show that this approach produces more abstractive summaries compared to prior work that employs a copy mechanism while still achieving higher ROUGE scores. We provide extensive comparisons with strong baseline methods, prior state of the art work as well as multiple variants of our approach including those using only transformers, only extractive techniques and combinations of the two. We examine these models using four different summarization tasks and datasets: arXiv papers, PubMed papers, the Newsroom and BigPatent datasets. We find that transformer based methods produce summaries with fewer n-gram copies, leading to n-grain copying statistics that are more similar to human generated abstracts. We include a human evaluation, finding that transformers are ranked highly for coherence and fluency, but purely extractive methods score higher for informativeness and relevance. We hope that these architectures and experiments may serve as strong points of comparison for future work.
引用
收藏
页码:9308 / 9319
页数:12
相关论文
共 50 条
  • [31] Assessing Abstractive and Extractive Methods for Automatic News Summarization
    Oliveira, Hilario
    Lins, Rafael Dueire
    PROCEEDINGS OF THE 2024 ACM SYMPOSIUM ON DOCUMENT ENGINEERING, DOCENG 2024, 2024,
  • [32] Multi-Task Learning for Abstractive and Extractive Summarization
    Chen, Yangbin
    Ma, Yun
    Mao, Xudong
    Li, Qing
    DATA SCIENCE AND ENGINEERING, 2019, 4 (01) : 14 - 23
  • [33] Extractive Elementary Discourse Units for Improving Abstractive Summarization
    Xiong, Ye
    Racharak, Teeradaj
    Minh Le Nguyen
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2675 - 2679
  • [34] Word topical mixture models for extractive spoken document summarization
    Chen, Berlin
    Chen, Yi-Ting
    2007 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-5, 2007, : 52 - 55
  • [35] Neural sentence fusion for diversity driven abstractive multi-document summarization
    Fuad, Tanvir Ahmed
    Nayeem, Mir Tafseer
    Mahmud, Asif
    Chali, Yllias
    COMPUTER SPEECH AND LANGUAGE, 2019, 58 : 216 - 230
  • [36] An Extractive-and-Abstractive Framework for Source Code Summarization
    Sun, Weisong
    Fang, Chunrong
    Chen, Yuchen
    Zhang, Quanjun
    Tao, Guanhong
    You, Yudu
    Han, Tingxu
    Ge, Yifei
    Hu, Yuling
    Luo, Bin
    Chen, Zhenyu
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (03)
  • [37] Abstractive vs. Extractive Summarization: An Experimental Review
    Giarelis, Nikolaos
    Mastrokostas, Charalampos
    Karacapilidis, Nikos
    APPLIED SCIENCES-BASEL, 2023, 13 (13):
  • [38] Abstractive Document Summarization via Bidirectional Decoder
    Wan, Xin
    Li, Chen
    Wang, Ruijia
    Xiao, Ding
    Shi, Chuan
    ADVANCED DATA MINING AND APPLICATIONS, ADMA 2018, 2018, 11323 : 364 - 377
  • [39] Attention based Abstractive Summarization of Malayalam Document
    Nambiar, Sindhya K.
    Peter, David S.
    Idicula, Sumam Mary
    AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 250 - 257
  • [40] Key phrase aware transformer for abstractive summarization
    Liu, Shuaiqi
    Cao, Jiannong
    Yang, Ruosong
    Wen, Zhiyuan
    INFORMATION PROCESSING & MANAGEMENT, 2022, 59 (03)