Exploring Abstractive Text Summarization: Methods, Dataset, Evaluation, and Emerging Challenges

被引:0
|
作者
Sunusi, Yusuf [1 ]
Omar, Nazlia [1 ]
Zakaria, Lailatul Qadri [1 ]
机构
[1] Univ Kebangsaan Malaysia, Ctr Artificial Intelligence Technol, Bangi 43600, Malaysia
关键词
Abstractive text summarization; systematic literature review; natural language processing; evaluation metrics; dataset; computation linguistics; MODEL; RNN;
D O I
10.14569/IJACSA.2024.01507130
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
-The latest advanced models for abstractive summarization, which utilize encoder-decoder frameworks, produce exactly one summary for each source text. This systematic literature review (SLR) comprehensively examines the recent advancements in abstractive text summarization (ATS), a pivotal area in natural language processing (NLP) that aims to generate concise and coherent summaries from extensive text sources. We delve into the evolution of ATS, focusing on key aspects such as encoder-decoder architectures, innovative mechanisms like attention and pointer-generator models, training and optimization methods, datasets, and evaluation metrics. Our review analyzes a wide range of studies, highlighting the transition from traditional sequence-to-sequence models to more advanced approaches like Transformer-based architectures. We explore the integration of mechanisms such as attention, which enhances model interpretability and effectiveness, and pointer-generator networks, which adeptly balance between copying and generating text. The review also addresses the challenges in training these models, including issues related to dataset quality and diversity, particularly in low-resource languages. A critical analysis of evaluation metrics reveals a heavy reliance on ROUGE scores, prompting a discussion on the need for more nuanced evaluation methods that align closely with human judgment. Additionally, we identify and discuss emerging research gaps, such as the need for effective summary length control and the handling of model hallucination, which are crucial for the practical application of ATS. This SLR not only synthesizes current research trends and methodologies in ATS, but also provides insights into future directions, underscoring the importance of continuous innovation in model development, dataset enhancement, and evaluation strategies. Our findings aim to guide researchers and practitioners in navigating the evolving landscape of abstractive text summarization and in identifying areas ripe for future exploration and development.
引用
收藏
页码:1340 / 1357
页数:18
相关论文
共 50 条
  • [1] End to End Urdu Abstractive Text Summarization With Dataset and Improvement in Evaluation Metric
    Raza, Hassan
    Shahzad, Waseem
    IEEE ACCESS, 2024, 12 : 40311 - 40324
  • [2] Abstractive text summarization: State of the art, challenges, and improvements
    Shakil, Hassan
    Farooq, Ahmad
    Kalita, Jugal
    NEUROCOMPUTING, 2024, 603
  • [3] Abstractive text summarization using deep learning with a new Turkish summarization benchmark dataset
    Ertam, Fatih
    Aydin, Galip
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (09):
  • [4] Social-sum-Mal: A Dataset for Abstractive Text Summarization in Malayalam
    Raj, M. rahul
    Pankaj, Dhanya S.
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (11)
  • [5] Abstractive Text Summarization for the Urdu Language: Data and Methods
    Awais, Muhammad
    Muhammad Adeel Nawab, Rao
    IEEE ACCESS, 2024, 12 : 61198 - 61210
  • [6] CATAMARAN: A Cross-lingual Long Text Abstractive Summarization Dataset
    Chen, Zheng
    Lin, Hongyu
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 6932 - 6937
  • [7] Deep Learning Based Abstractive Text Summarization: Approaches, Datasets, Evaluation Measures, and Challenges
    Suleiman, Dima
    Awajan, Arafat
    Suleiman, Dima (d.suleiman@psut.edu.jo), 1600, Hindawi Limited, 410 Park Avenue, 15th Floor, 287 pmb, New York, NY 10022, United States (2020):
  • [8] Deep Learning Based Abstractive Text Summarization: Approaches, Datasets, Evaluation Measures, and Challenges
    Suleiman, Dima
    Awajan, Arafat
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2020, 2020
  • [9] A Dataset for Exploring Gaze Behaviors in Text Summarization
    Yi, Kun
    Guo, Yu
    Jiang, Weifeng
    Wang, Zhi
    Sun, Lifeng
    MMSYS'20: PROCEEDINGS OF THE 2020 MULTIMEDIA SYSTEMS CONFERENCE, 2020, : 243 - 248
  • [10] An approach to Abstractive Text Summarization
    Huong Thanh Le
    Tien Manh Le
    2013 INTERNATIONAL CONFERENCE OF SOFT COMPUTING AND PATTERN RECOGNITION (SOCPAR), 2013, : 371 - 376