ATTSUM: A Deep Attention-Based Summarization Model for Bug Report Title Generation

被引:9
|
作者
Ma, Xiaoxue [1 ]
Keung, Jacky Wai [1 ]
Yu, Xiao [2 ,3 ]
Zou, Huiqi [1 ]
Zhang, Jingyu [1 ]
Li, Yishu
机构
[1] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
[2] Wuhan Univ Technol, Sanya Sci & Educ Innovat Pk, Sanya 572024, Peoples R China
[3] Wuhan Univ Technol, Sch Comp Sci & Artificial Intelligence, Wuhan 430062, Peoples R China
关键词
Computer bugs; Decoding; Transformers; Semantics; Training; Software; Vocabulary; Bug reports; deep learning; text summarization; title generation; transformers;
D O I
10.1109/TR.2023.3236404
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Concise and precise bug report titles help software developers to capture the highlights of the bug report quickly. Unfortunately, it is common that bug reporters do not create high-quality bug report titles. Recent long short-term memory (LSTM)-based sequence-to-sequence models such as iTAPE were proposed to generate bug report titles automatically, but the text representation method and LSTM employed in such model are difficult to capture the accurate semantic information and draw the global dependencies among tokens effectively. This article proposes a deep attention-based summarization model (i.e., AttSum) to generate high-quality bug report titles. Specifically, the AttSum model employs the encoder.decoder framework, which utilizes the robustly optimized bidirectional-encoder-representations-from-transformers approach to encode the bug report bodies to capture contextual semantic information better, the stacked transformer decoder to automatically generate titles, and the copy mechanism to handle the rare token problem. To validate the effectiveness of AttSum, we conduct automatic and manual evaluations on 333563 "< body, title>" pairs of bug reports and perform a practical analysis of its ability to improve low-quality titles. The result shows that AttSum is superior to the state-of-the-art baselines by a substantial margin both on automatic evaluation metrics (e.g., by 3.4%-58.8% and 7.7%-42.3% in terms of recall-oriented understudy for gisting evaluation in F1 and bilingual evaluation understudy, separately) and three human-set modalities (e.g., by 1.9%-57.5%). Moreover, we analyze the impact of the training data size on AttSum and the results imply that our approach is robust enough to generate much better titles.
引用
收藏
页码:1663 / 1677
页数:15
相关论文
共 50 条
  • [21] Attention-Based Image Caption Generation
    Manasa, M.
    Sowmya, D.
    Reddy, Y. Supriya
    Sreedevi, Pogula
    PROCEEDINGS OF THE 5TH INTERNATIONAL CONFERENCE ON DATA SCIENCE, MACHINE LEARNING AND APPLICATIONS, VOL 1, ICDSMLA 2023, 2025, 1273 : 364 - 369
  • [22] Deep Attention-Based Alignment Network for Melody Generation from Incomplete Lyrics
    Reddy, M. Gurunath
    Zhang, Zhe
    Yu, Yi
    Harscoet, Florian
    Canales, Simon
    Tang, Suhua
    2022 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM), 2022, : 236 - 239
  • [23] KeyTitle: towards better bug report title generation by keywords planning
    Meng, Qianshuang
    Zou, Weiqin
    Cai, Biyu
    Zhang, Jingxuan
    SOFTWARE QUALITY JOURNAL, 2024, 32 (04) : 1655 - 1682
  • [24] A Weighted PageRank-Based Bug Report Summarization Method Using Bug Report Relationships
    Kim, Beomjun
    Kang, Sungwon
    Lee, Seonah
    APPLIED SCIENCES-BASEL, 2019, 9 (24):
  • [25] Attention-Based Deep Learning Model for Image Desaturation of SDO/AIA
    Zhang, Xinze
    Xu, Long
    Ren, Zhixiang
    Yu, Xuexin
    Li, Jia
    RESEARCH IN ASTRONOMY AND ASTROPHYSICS, 2023, 23 (08)
  • [26] An attention-based deep learning model for citywide traffic flow forecasting
    Zhou, Tao
    Huang, Bo
    Li, Rongrong
    Liu, Xiaoqian
    Huang, Zhihui
    INTERNATIONAL JOURNAL OF DIGITAL EARTH, 2022, 15 (01) : 323 - 344
  • [27] An attention-based hybrid deep learning model for EEG emotion recognition
    Yong Zhang
    Yidie Zhang
    Shuai Wang
    Signal, Image and Video Processing, 2023, 17 : 2305 - 2313
  • [28] An attention-based hybrid deep learning model for EEG emotion recognition
    Zhang, Yong
    Zhang, Yidie
    Wang, Shuai
    SIGNAL IMAGE AND VIDEO PROCESSING, 2023, 17 (05) : 2305 - 2313
  • [29] An attention-based deep learning model for multiple pedestrian attributes recognition
    Yaghoubi, Ehsan
    Borza, Diana
    Neves, Joao
    Kumar, Aruna
    Proenca, Hugo
    IMAGE AND VISION COMPUTING, 2020, 102 (102)
  • [30] Attention-Based Distributed Deep Learning Model for Air Quality Forecasting
    Mengara, Axel Gedeon Mengara
    Park, Eunyoung
    Jang, Jinho
    Yoo, Younghwan
    SUSTAINABILITY, 2022, 14 (06)