ATTSUM: A Deep Attention-Based Summarization Model for Bug Report Title Generation

被引:9
|
作者
Ma, Xiaoxue [1 ]
Keung, Jacky Wai [1 ]
Yu, Xiao [2 ,3 ]
Zou, Huiqi [1 ]
Zhang, Jingyu [1 ]
Li, Yishu
机构
[1] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
[2] Wuhan Univ Technol, Sanya Sci & Educ Innovat Pk, Sanya 572024, Peoples R China
[3] Wuhan Univ Technol, Sch Comp Sci & Artificial Intelligence, Wuhan 430062, Peoples R China
关键词
Computer bugs; Decoding; Transformers; Semantics; Training; Software; Vocabulary; Bug reports; deep learning; text summarization; title generation; transformers;
D O I
10.1109/TR.2023.3236404
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Concise and precise bug report titles help software developers to capture the highlights of the bug report quickly. Unfortunately, it is common that bug reporters do not create high-quality bug report titles. Recent long short-term memory (LSTM)-based sequence-to-sequence models such as iTAPE were proposed to generate bug report titles automatically, but the text representation method and LSTM employed in such model are difficult to capture the accurate semantic information and draw the global dependencies among tokens effectively. This article proposes a deep attention-based summarization model (i.e., AttSum) to generate high-quality bug report titles. Specifically, the AttSum model employs the encoder.decoder framework, which utilizes the robustly optimized bidirectional-encoder-representations-from-transformers approach to encode the bug report bodies to capture contextual semantic information better, the stacked transformer decoder to automatically generate titles, and the copy mechanism to handle the rare token problem. To validate the effectiveness of AttSum, we conduct automatic and manual evaluations on 333563 "< body, title>" pairs of bug reports and perform a practical analysis of its ability to improve low-quality titles. The result shows that AttSum is superior to the state-of-the-art baselines by a substantial margin both on automatic evaluation metrics (e.g., by 3.4%-58.8% and 7.7%-42.3% in terms of recall-oriented understudy for gisting evaluation in F1 and bilingual evaluation understudy, separately) and three human-set modalities (e.g., by 1.9%-57.5%). Moreover, we analyze the impact of the training data size on AttSum and the results imply that our approach is robust enough to generate much better titles.
引用
收藏
页码:1663 / 1677
页数:15
相关论文
共 50 条
  • [1] The effect of using video title in attention-based video summarization
    Li, Changwei
    Yeh, Zhi-Ting
    Gunuganti, Jeshmitha
    Chang, Jia-Bin
    Norouzi, Mehdi
    2024 2ND ASIA CONFERENCE ON COMPUTER VISION, IMAGE PROCESSING AND PATTERN RECOGNITION, CVIPPR 2024, 2024,
  • [2] Duplicate Bug Report Detection Using an Attention-Based Neural Language Model
    Ben Messaoud, Montassar
    Miladi, Asma
    Jenhani, Ilyes
    Mkaouer, Mohamed Wiem
    Ghadhab, Lobna
    IEEE TRANSACTIONS ON RELIABILITY, 2023, 72 (02) : 846 - 858
  • [3] Unsupervised Deep Bug Report Summarization
    Li, Xiaochen
    Jiang, He
    Liu, Dong
    Ren, Zhilei
    Li, Ge
    2018 IEEE/ACM 26TH INTERNATIONAL CONFERENCE ON PROGRAM COMPREHENSION (ICPC 2018), 2018, : 144 - 155
  • [4] Attention-based Clinical Note Summarization
    Kanwal, Neel
    Rizzo, Giuseppe
    37TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, 2022, : 813 - 820
  • [5] BugSum: Deep Context Understanding for Bug Report Summarization
    Liu, Haoran
    Yu, Yue
    Li, Shanshan
    Guo, Yong
    Wang, Deze
    Mao, Xiaoguang
    2020 IEEE/ACM 28TH INTERNATIONAL CONFERENCE ON PROGRAM COMPREHENSION, ICPC, 2020, : 94 - 105
  • [6] Attention-Based Deep Recurrent Model for Survival Prediction
    Sun Z.
    Dong W.
    Shi J.
    He K.
    Huang Z.
    ACM Transactions on Computing for Healthcare, 2021, 2 (04):
  • [7] Attention-based Deep Learning Model for Text Readability Evaluation
    Sun, Yuxuan
    Chen, Keying
    Sun, Lin
    Hu, Chenlu
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [8] Deep Learning-Based Bug Report Summarization Using Sentence Significance Factors
    Koh, Youngji
    Kang, Sungwon
    Lee, Seonah
    APPLIED SCIENCES-BASEL, 2022, 12 (12):
  • [9] Bug-Transformer: Automated Program Repair Using Attention-Based Deep Neural Network
    Yao, Jie
    Rao, Bingbing
    Xing, Weiwei
    Wang, Liqiang
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2022, 31 (12)
  • [10] aDFR: An Attention-Based Deep Learning Model for Flight Ranking
    Yi, Yuan
    Cao, Jian
    Tan, YuDong
    Nie, QiangQiang
    Lu, XiaoXi
    WEB INFORMATION SYSTEMS ENGINEERING, WISE 2020, PT II, 2020, 12343 : 548 - 562