Security Development Lifecycle-Based Adaptive Reward Mechanism for Reinforcement Learning in Continuous Integration Testing Optimization

被引:0
|
作者
Yang, Yang [1 ]
Wang, Weiwei [2 ]
Li, Zheng [3 ]
Zhang, Lieshan [1 ]
Pan, Chaoyue [3 ]
机构
[1] Zhejiang Sci Tech Univ, Sch Informat Sci & Engn, Hangzhou 310018, Zhejiang, Peoples R China
[2] Beijing Inst Petrochem Technol, Sch Informat Engn, Beijing 100095, Peoples R China
[3] Beijing Univ Chem Technol, Coll Informat Sci & Technol, Beijing 100029, Peoples R China
关键词
Continuous integration testing; reinforcement learning; test case prioritization; reward mechanism; PRIORITIZATION;
D O I
10.1142/S0218194024500244
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continuous automated testing throughout each cycle can ensure the security of the continuous integration (CI) development lifecycle. Test case prioritization (TCP) is a critical factor in optimizing automated testing, which prioritizes potentially failed test cases and improves the efficiency of automated testing. In CI automated testing, the TCP is a continuous decision-making process that can be solved with reinforcement learning (RL). RL-based CITCP can continuously generate a TCP strategy for each CI development lifecycle, with the reward mechanism as the core. The reward mechanism consists of the reward function and the reward strategy. However, there are new challenges to RL-based CITCP in real-industry CI testing. With high-frequency iteration, the reward function is often calculated with a fixed length of historical information, ignoring the spatial characteristics of the current cycle. Therefore, the dynamic time window (DTW)-based reward function is proposed to perform the reward calculation, which adaptively adjusts the recent historical information range based on the integration cycle. Moreover, with low-failure testing, the reward strategy usually only rewards failure test cases, which creates a sparse reward problem in RL. To address this issue, the similarity-based reward strategy is proposed, which increases the reward objects of some passed test cases, similar to the failure test cases. The DTW-based reward function and the similarity-based reward strategy together constitute the proposed adaptive reward mechanism in RL-based CITCP. To validate the effectiveness of the adaptive reward mechanism, experimental verification is carried out on 13 industrial data sets. The experimental results show that the adaptive reward mechanism can improve the TCP effect, where the average NAPFD is maximally improved by 7.29%, the average Recall is maximally improved by 6.04% and the average TTF is improved by 6.81 positions with a maximum of 63.77.
引用
收藏
页码:1457 / 1483
页数:27
相关论文
共 50 条
  • [31] Reward Mechanism Design for Deep Reinforcement Learning-Based Microgrid Energy Management
    Hu, Mingjie
    Han, Baohui
    Lv, Shilin
    Bao, Zhejing
    Lu, Lingxia
    Yu, Miao
    2023 6TH INTERNATIONAL CONFERENCE ON RENEWABLE ENERGY AND POWER ENGINEERING, REPE 2023, 2023, : 201 - 205
  • [32] A train trajectory optimization method based on the safety reinforcement learning with a relaxed dynamic reward
    Cheng, Ligang
    Cao, Jie
    Yang, Xiaofeng
    Wang, Wenxian
    Zhou, Zijian
    DISCOVER APPLIED SCIENCES, 2024, 6 (09)
  • [33] Reward design for multi-agent reinforcement learning with a penalty based on the payment mechanism
    Matsunami N.
    Okuhara S.
    Ito T.
    Transactions of the Japanese Society for Artificial Intelligence, 2021, 36 (05)
  • [34] Reinforcement learning based optimization of process chromatography for continuous processing of biopharmaceuticals
    Nikita, Saxena
    Tiwari, Anamika
    Sonawat, Deepak
    Kodamana, Hariprasad
    Rathore, Anurag S.
    CHEMICAL ENGINEERING SCIENCE, 2021, 230
  • [35] An Agent-based Self-Adaptive Mechanism with Reinforcement Learning
    Yu, Danni
    Li, Qingshan
    Wang, Lu
    Lin, Yishuai
    IEEE 39TH ANNUAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE WORKSHOPS (COMPSAC 2015), VOL 3, 2015, : 582 - 585
  • [36] Adaptive Stabilizing Control of Smart Transformer Based on Reinforcement Learning Optimization
    Tang, Jian
    Zou, Zhixiang
    Yang, Jiajun
    Buticchi, Giampaolo
    Hua, Wei
    IEEE TRANSACTIONS ON INDUSTRY APPLICATIONS, 2024, 60 (03) : 4324 - 4337
  • [37] RLBS: An Adaptive Backtracking Strategy Based on Reinforcement Learning for Combinatorial Optimization
    Bachiri, Ilyess
    Gaudreault, Jonathan
    Quimper, Claude-Guy
    Chaib-draa, Brahim
    2015 IEEE 27TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2015), 2015, : 936 - 942
  • [38] Dynamic Adaptive Checkpoint Mechanism for Streaming Applications Based on Reinforcement Learning
    Zhang, Zhan
    Liu, Tianming
    Shu, Yanjun
    Chen, Siyuan
    Liu, Xian
    2022 IEEE 28TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, ICPADS, 2022, : 538 - 545
  • [39] Multi-agent Reinforcement Learning Based on Adaptive State Approximation in Sparse Reward Scenarios
    Fang, Baofu
    Yu, Tingting
    Wang, Hao
    Wang, Zaijun
    Jiqiren/Robot, 2024, 46 (06): : 663 - 671
  • [40] Hierarchical optimization of reservoir development strategy based on reinforcement learning
    Wang, Haochen
    Zhang, Kai
    Chen, Nancy
    Wang, Zhongzheng
    Xin, Guojing
    Liu, Chengcheng
    Zhang, Liming
    Zhou, Wensheng
    Liu, Chen
    GEOENERGY SCIENCE AND ENGINEERING, 2023, 226