Security Development Lifecycle-Based Adaptive Reward Mechanism for Reinforcement Learning in Continuous Integration Testing Optimization

被引:0
|
作者
Yang, Yang [1 ]
Wang, Weiwei [2 ]
Li, Zheng [3 ]
Zhang, Lieshan [1 ]
Pan, Chaoyue [3 ]
机构
[1] Zhejiang Sci Tech Univ, Sch Informat Sci & Engn, Hangzhou 310018, Zhejiang, Peoples R China
[2] Beijing Inst Petrochem Technol, Sch Informat Engn, Beijing 100095, Peoples R China
[3] Beijing Univ Chem Technol, Coll Informat Sci & Technol, Beijing 100029, Peoples R China
关键词
Continuous integration testing; reinforcement learning; test case prioritization; reward mechanism; PRIORITIZATION;
D O I
10.1142/S0218194024500244
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continuous automated testing throughout each cycle can ensure the security of the continuous integration (CI) development lifecycle. Test case prioritization (TCP) is a critical factor in optimizing automated testing, which prioritizes potentially failed test cases and improves the efficiency of automated testing. In CI automated testing, the TCP is a continuous decision-making process that can be solved with reinforcement learning (RL). RL-based CITCP can continuously generate a TCP strategy for each CI development lifecycle, with the reward mechanism as the core. The reward mechanism consists of the reward function and the reward strategy. However, there are new challenges to RL-based CITCP in real-industry CI testing. With high-frequency iteration, the reward function is often calculated with a fixed length of historical information, ignoring the spatial characteristics of the current cycle. Therefore, the dynamic time window (DTW)-based reward function is proposed to perform the reward calculation, which adaptively adjusts the recent historical information range based on the integration cycle. Moreover, with low-failure testing, the reward strategy usually only rewards failure test cases, which creates a sparse reward problem in RL. To address this issue, the similarity-based reward strategy is proposed, which increases the reward objects of some passed test cases, similar to the failure test cases. The DTW-based reward function and the similarity-based reward strategy together constitute the proposed adaptive reward mechanism in RL-based CITCP. To validate the effectiveness of the adaptive reward mechanism, experimental verification is carried out on 13 industrial data sets. The experimental results show that the adaptive reward mechanism can improve the TCP effect, where the average NAPFD is maximally improved by 7.29%, the average Recall is maximally improved by 6.04% and the average TTF is improved by 6.81 positions with a maximum of 63.77.
引用
收藏
页码:1457 / 1483
页数:27
相关论文
共 50 条
  • [21] Robust Adaptive Scaffolding with Inverse Reinforcement Learning-Based Reward Design
    Fahid, Fahmid Morshed
    Rowe, Jonathan P.
    Spain, Randall D.
    Goldberg, Benjamin S.
    Pokorny, Robert
    Lester, James
    ARTIFICIAL INTELLIGENCE IN EDUCATION: POSTERS AND LATE BREAKING RESULTS, WORKSHOPS AND TUTORIALS, INDUSTRY AND INNOVATION TRACKS, PRACTITIONERS AND DOCTORAL CONSORTIUM, PT II, 2022, 13356 : 204 - 207
  • [22] A learning algorithm for optimizing continuous integration development and testing practice
    Marijan, Dusica
    Gotlieb, Arnaud
    Liaaen, Marius
    SOFTWARE-PRACTICE & EXPERIENCE, 2019, 49 (02): : 192 - 213
  • [23] Reinforcement Learning based Scheduling Optimization Mechanism on Switches
    Lu, Xijia
    Wang, Xingwei
    Jia, Jie
    Wang, Xue
    Huang, Min
    2022 IEEE 28TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, ICPADS, 2022, : 554 - 561
  • [24] Test Case Prioritization Technique in Continuous Integration Based on Reinforcement Learning
    Zhao Y.-F.
    Hao D.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (06): : 2708 - 2726
  • [25] Reinforcement Learning based Interconnection Routing for Adaptive Traffic Optimization
    Kao, Sheng-Chun
    Yang, Chao-Han Huck
    Chen, Pin-Yu
    Ma, Xiaoli
    Krishna, Tushar
    PROCEEDINGS OF THE 13TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON NETWORKS-ON-CHIP (NOCS'19), 2019,
  • [26] Adaptive Water Environment Optimization Strategy Based on Reinforcement Learning
    Dang T.
    Liu J.
    Computer-Aided Design and Applications, 2024, 21 (S23): : 1 - 18
  • [27] Adaptive reward shaping based reinforcement learning for docking control of autonomous underwater vehicles
    Chu, Shuguang
    Lin, Mingwei
    Li, Dejun
    Lin, Ri
    Xiao, Sa
    OCEAN ENGINEERING, 2025, 318
  • [28] A SDN Routing Optimization Mechanism Based on Deep Reinforcement Learning
    Lan J.
    Yu C.
    Hu Y.
    Li Z.
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2019, 41 (11): : 2669 - 2674
  • [29] A SDN Routing Optimization Mechanism Based on Deep Reinforcement Learning
    Lan Julong
    Yu Changhe
    Hu Yuxiang
    Li Ziyong
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2019, 41 (11) : 2669 - 2674
  • [30] Computerized Adaptive Testing Method Based on Reinforcement Learning for Series Diagnosis
    Liu Z.
    Wu J.
    Yao F.
    Liu Q.
    Chen E.
    Sha J.
    Wang S.
    Su Y.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2024, 37 (01): : 13 - 26