Exploring Abstractive Text Summarization: Methods, Dataset, Evaluation, and Emerging Challenges

被引:0
|
作者
Sunusi, Yusuf [1 ]
Omar, Nazlia [1 ]
Zakaria, Lailatul Qadri [1 ]
机构
[1] Univ Kebangsaan Malaysia, Ctr Artificial Intelligence Technol, Bangi 43600, Malaysia
关键词
Abstractive text summarization; systematic literature review; natural language processing; evaluation metrics; dataset; computation linguistics; MODEL; RNN;
D O I
10.14569/IJACSA.2024.01507130
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
-The latest advanced models for abstractive summarization, which utilize encoder-decoder frameworks, produce exactly one summary for each source text. This systematic literature review (SLR) comprehensively examines the recent advancements in abstractive text summarization (ATS), a pivotal area in natural language processing (NLP) that aims to generate concise and coherent summaries from extensive text sources. We delve into the evolution of ATS, focusing on key aspects such as encoder-decoder architectures, innovative mechanisms like attention and pointer-generator models, training and optimization methods, datasets, and evaluation metrics. Our review analyzes a wide range of studies, highlighting the transition from traditional sequence-to-sequence models to more advanced approaches like Transformer-based architectures. We explore the integration of mechanisms such as attention, which enhances model interpretability and effectiveness, and pointer-generator networks, which adeptly balance between copying and generating text. The review also addresses the challenges in training these models, including issues related to dataset quality and diversity, particularly in low-resource languages. A critical analysis of evaluation metrics reveals a heavy reliance on ROUGE scores, prompting a discussion on the need for more nuanced evaluation methods that align closely with human judgment. Additionally, we identify and discuss emerging research gaps, such as the need for effective summary length control and the handling of model hallucination, which are crucial for the practical application of ATS. This SLR not only synthesizes current research trends and methodologies in ATS, but also provides insights into future directions, underscoring the importance of continuous innovation in model development, dataset enhancement, and evaluation strategies. Our findings aim to guide researchers and practitioners in navigating the evolving landscape of abstractive text summarization and in identifying areas ripe for future exploration and development.
引用
收藏
页码:1340 / 1357
页数:18
相关论文
共 50 条
  • [31] Generative Adversarial Network for Abstractive Text Summarization
    Liu, Linqing
    Lu, Yao
    Yang, Min
    Qu, Qiang
    Zhu, Jia
    Li, Hongyan
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 8109 - 8110
  • [32] Evaluating the Factual Consistency of Abstractive Text Summarization
    Kryscinski, Wojciech
    McCann, Bryan
    Xiong, Caiming
    Socher, Richard
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 9332 - 9346
  • [33] Highlighted Word Encoding for Abstractive Text Summarization
    Lal, Daisy Monika
    Singh, Krishna Pratap
    Tiwary, Uma Shanker
    INTELLIGENT HUMAN COMPUTER INTERACTION (IHCI 2019), 2020, 11886 : 77 - 86
  • [34] Abstractive Text Summarization by Incorporating Reader Comments
    Gao, Shen
    Chen, Xiuying
    Li, Piji
    Ren, Zhaochun
    Bing, Lidong
    Zhao, Dongyan
    Yan, Rui
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 6399 - 6406
  • [35] Improving Abstractive Text Summarization with History Aggregation
    Liao, Pengcheng
    Zhang, Chuang
    Chen, Xiaojun
    Zhou, Xiaofei
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [36] Data Selection Curriculum for Abstractive Text Summarization
    Sun, Shichao
    Yuan, Ruifeng
    He, Jianfei
    Cao, Ziqiang
    Li, Wenjie
    Jia, Xiaohua
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 7990 - 7995
  • [37] CivilSum: A Dataset for Abstractive Summarization of Indian Court Decisions
    Malik, Manuj
    Zhao, Zheng
    Fonseca, Marcio
    Rao, Shrisha
    Cohen, Shay B.
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2241 - 2250
  • [38] Review on text summarization evaluation methods
    Indu, M.
    Kavitha, K., V
    2016 INTERNATIONAL CONFERENCE ON RESEARCH ADVANCES IN INTEGRATED NAVIGATION SYSTEMS (RAINS), 2016,
  • [39] Exploring Explainable Selection to Control Abstractive Summarization
    Wang Haonan
    Gao Yang
    Bai Yu
    Lapata, Mirella
    Huang Heyan
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 13933 - 13941
  • [40] ATSSI: Abstractive Text Summarization using Sentiment Infusion
    Bhargava, Rupal
    Sharma, Yashvardhan
    Sharma, Gargi
    TWELFTH INTERNATIONAL CONFERENCE ON COMMUNICATION NETWORKS, ICCN 2016 / TWELFTH INTERNATIONAL CONFERENCE ON DATA MINING AND WAREHOUSING, ICDMW 2016 / TWELFTH INTERNATIONAL CONFERENCE ON IMAGE AND SIGNAL PROCESSING, ICISP 2016, 2016, 89 : 404 - 411