On Extractive and Abstractive Neural Document Summarization with Transformer Language Models

被引:0
|
作者
Pilault, Jonathan [1 ,2 ,3 ]
Li, Raymond [1 ]
Subramanian, Sandeep [1 ,2 ,4 ]
Pal, Christopher [1 ,2 ,3 ,4 ,5 ]
机构
[1] Element AI, Montreal, PQ, Canada
[2] Mila, Montreal, PQ, Canada
[3] Polytech Montreal, Montreal, PQ, Canada
[4] Univ Montreal, Montreal, PQ, Canada
[5] Canada CIFAR AI Chair, Montreal, PQ, Canada
来源
PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP) | 2020年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a method to produce abstractive summaries of long documents that exceed several thousand words via neural abstractive summarization. We perform a simple extractive step before generating a summary, which is then used to condition the transformer language model on relevant information before being tasked with generating a summary. We also show that this approach produces more abstractive summaries compared to prior work that employs a copy mechanism while still achieving higher ROUGE scores. We provide extensive comparisons with strong baseline methods, prior state of the art work as well as multiple variants of our approach including those using only transformers, only extractive techniques and combinations of the two. We examine these models using four different summarization tasks and datasets: arXiv papers, PubMed papers, the Newsroom and BigPatent datasets. We find that transformer based methods produce summaries with fewer n-gram copies, leading to n-grain copying statistics that are more similar to human generated abstracts. We include a human evaluation, finding that transformers are ranked highly for coherence and fluency, but purely extractive methods score higher for informativeness and relevance. We hope that these architectures and experiments may serve as strong points of comparison for future work.
引用
收藏
页码:9308 / 9319
页数:12
相关论文
共 50 条
  • [21] Diverse Decoding for Abstractive Document Summarization
    Han, Xu-Wang
    Zheng, Hai-Tao
    Chen, Jin-Yuan
    Zhao, Cong-Zhi
    APPLIED SCIENCES-BASEL, 2019, 9 (03):
  • [22] Graphs in clusters: a hybrid approach to unsupervised extractive long document summarization using language models
    Gokhan, Tuba
    Price, Malcolm James
    Lee, Mark
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (07)
  • [23] Investigating Hallucinations in Pruned Large Language Models for Abstractive Summarization
    Chrysostomou, George
    Zhao, Zhixue
    Williams, Miles
    Aletras, Nikolaos
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 1163 - 1181
  • [24] Effectiveness of French Language Models on Abstractive Dialogue Summarization Task
    Zhou, Yongxin
    Portet, Francois
    Ringeval, Fabien
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 3571 - 3581
  • [25] Integrating Topic-Aware Heterogeneous Graph Neural Network With Transformer Model for Medical Scientific Document Abstractive Summarization
    Khaliq, Ayesha
    Khan, Atif
    Awan, Salman Afsar
    Jan, Salman
    Umair, Muhammad
    Zuhairi, Megat F.
    IEEE ACCESS, 2024, 12 : 113855 - 113866
  • [26] A Survey of Abstractive Text Summarization Utilising Pretrained Language Models
    Syed, Ayesha Ayub
    Gaol, Ford Lumban
    Boediman, Alfred
    Matsuo, Tokuro
    Budiharto, Widodo
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2022, PT I, 2022, 13757 : 532 - 544
  • [27] Neural Abstractive Text Summarization with Sequence-to-Sequence Models
    Shi, Tian
    Keneshloo, Yaser
    Ramakrishnan, Naren
    Reddy, Chandan K.
    ACM/IMS Transactions on Data Science, 2021, 2 (01):
  • [28] A Review on Neural network based Abstractive Text Summarization models
    Tandel, Jinal
    Mistree, Kinjal
    Shah, Parth
    2019 IEEE 5TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2019,
  • [29] A Survey of Extractive and Abstractive Automatic Text Summarization Techniques
    Dalal, Vipul
    Malik, Latesh
    2013 Sixth International Conference on Emerging Trends in Engineering and Technology (ICETET 2013), 2013, : 109 - 110
  • [30] Multi-Task Learning for Abstractive and Extractive Summarization
    Yangbin Chen
    Yun Ma
    Xudong Mao
    Qing Li
    Data Science and Engineering, 2019, 4 (1) : 14 - 23