Finding the best learning to rank algorithms for effort-aware defect prediction

被引:14
|
作者
Yu, Xiao [1 ,2 ]
Dai, Heng [3 ]
Li, Li [4 ]
Gu, Xiaodong [5 ]
Keung, Jacky Wai [6 ]
Bennin, Kwabena Ebo [7 ]
Li, Fuyang [1 ]
Liu, Jin [8 ]
机构
[1] Wuhan Univ Technol, Sch Comp Sci & Artificial Intelligence, Wuhan, Peoples R China
[2] Wuhan Univ Technol, Sanya Sci & Educ Innovat Pk, Sanya, Peoples R China
[3] Wuhan Qingchuan Univ, Sch Mech & Elect Engn, Wuhan, Peoples R China
[4] Beihang Univ, Sch Software, Beijing, Peoples R China
[5] Shanghai Jiao Tong Univ, Sch Software, Shanghai, Peoples R China
[6] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
[7] Wageningen Univ & Res, Informat Technol Grp, Wageningen, Netherlands
[8] Wuhan Univ, Sch Comp Sci, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
Software defect prediction; Empirical study; Learning to rank; Ranking instability; REGRESSION; RETRIEVAL; PRONENESS; MODELS; RIDGE;
D O I
10.1016/j.infsof.2023.107165
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Context: Effort-Aware Defect Prediction (EADP) ranks software modules or changes based on their predicted number of defects (i.e., considering modules or changes as effort) or defect density (i.e., considering LOC as effort) by using learning to rank algorithms. Ranking instability refers to the inconsistent conclusions produced by existing empirical studies of EADP. The major reason is the poor experimental design, such as comparison of few learning to rank algorithms, the use of small number of datasets or datasets without indicating numbers of defects, and evaluation with inappropriate or few metrics.Objective: To find a stable ranking of learning to rank algorithms to investigate the best ones for EADP,Method: We examine the practical effects of 34 algorithms on 49 datasets for EADP. We measure the performance of these algorithms using 7 module-based and 7 LOC-based metrics and run experiments under cross-release and cross-project settings, respectively. Finally, we obtain the ranking of these algorithms by performing the Scott-Knott ESD test.Results: When module is used as effort, random forest regression performs the best under cross-release setting, and linear regression performs the best under cross-project setting among the learning to rank algorithms; (2) when LOC is used as effort, LTR-linear (Learning-to-Rank with the linear model) performs the best under cross-release setting, and Ranking SVM performs the best under cross-project setting.Conclusion: This comprehensive experimental procedure allows us to discover a stable ranking of the studied algorithms to select the best ones according to the requirement of software projects.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Learning to rank software modules for effort-aware defect prediction
    Rao, Jiqing
    Yu, Xiao
    Zhang, Chen
    Zhou, Junwei
    Xiang, Jianwen
    2021 21ST INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY AND SECURITY COMPANION (QRS-C 2021), 2021, : 372 - 380
  • [2] An Empirical Study of Learning to Rank Techniques for Effort-Aware Defect Prediction
    Yu, Xiao
    Bennin, Kwabena Ebo
    Liu, Jin
    Keung, Jacky Wai
    Yin, Xiaofei
    Xu, Zhou
    2019 IEEE 26TH INTERNATIONAL CONFERENCE ON SOFTWARE ANALYSIS, EVOLUTION AND REENGINEERING (SANER), 2019, : 298 - 309
  • [3] Improving effort-aware defect prediction by directly learning to rank software modules
    Yu, Xiao
    Rao, Jiqing
    Liu, Lei
    Lin, Guancheng
    Hu, Wenhua
    Keung, Jacky Wai
    Zhou, Junwei
    Xiang, Jianwen
    INFORMATION AND SOFTWARE TECHNOLOGY, 2024, 165
  • [4] On effort-aware metrics for defect prediction
    Carka, Jonida
    Esposito, Matteo
    Falessi, Davide
    EMPIRICAL SOFTWARE ENGINEERING, 2022, 27 (06)
  • [5] Effort-Aware Defect Prediction Models
    Mende, Thilo
    Koschke, Rainer
    14TH EUROPEAN CONFERENCE ON SOFTWARE MAINTENANCE AND REENGINEERING (CSMR 2010), 2010, : 107 - 116
  • [6] On effort-aware metrics for defect prediction
    Jonida Çarka
    Matteo Esposito
    Davide Falessi
    Empirical Software Engineering, 2022, 27
  • [8] On the relative value of clustering techniques for Unsupervised Effort-Aware Defect Prediction
    Yang, Peixin
    Zhu, Lin
    Zhang, Yanjiao
    Ma, Chuanxiang
    Liu, Liming
    Yu, Xiao
    Hu, Wenhua
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 245
  • [9] Effort-aware and just-in-time defect prediction with neural network
    Qiao, Lei
    Wang, Yan
    PLOS ONE, 2019, 14 (02):
  • [10] A Novel Effort Measure Method for Effort-Aware Just-in-Time Software Defect Prediction
    Chen, Liqiong
    Song, Shilong
    Wang, Can
    INTERNATIONAL JOURNAL OF SOFTWARE ENGINEERING AND KNOWLEDGE ENGINEERING, 2021, 31 (08) : 1145 - 1169