Security Development Lifecycle-Based Adaptive Reward Mechanism for Reinforcement Learning in Continuous Integration Testing Optimization

被引:0
|
作者
Yang, Yang [1 ]
Wang, Weiwei [2 ]
Li, Zheng [3 ]
Zhang, Lieshan [1 ]
Pan, Chaoyue [3 ]
机构
[1] Zhejiang Sci Tech Univ, Sch Informat Sci & Engn, Hangzhou 310018, Zhejiang, Peoples R China
[2] Beijing Inst Petrochem Technol, Sch Informat Engn, Beijing 100095, Peoples R China
[3] Beijing Univ Chem Technol, Coll Informat Sci & Technol, Beijing 100029, Peoples R China
关键词
Continuous integration testing; reinforcement learning; test case prioritization; reward mechanism; PRIORITIZATION;
D O I
10.1142/S0218194024500244
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continuous automated testing throughout each cycle can ensure the security of the continuous integration (CI) development lifecycle. Test case prioritization (TCP) is a critical factor in optimizing automated testing, which prioritizes potentially failed test cases and improves the efficiency of automated testing. In CI automated testing, the TCP is a continuous decision-making process that can be solved with reinforcement learning (RL). RL-based CITCP can continuously generate a TCP strategy for each CI development lifecycle, with the reward mechanism as the core. The reward mechanism consists of the reward function and the reward strategy. However, there are new challenges to RL-based CITCP in real-industry CI testing. With high-frequency iteration, the reward function is often calculated with a fixed length of historical information, ignoring the spatial characteristics of the current cycle. Therefore, the dynamic time window (DTW)-based reward function is proposed to perform the reward calculation, which adaptively adjusts the recent historical information range based on the integration cycle. Moreover, with low-failure testing, the reward strategy usually only rewards failure test cases, which creates a sparse reward problem in RL. To address this issue, the similarity-based reward strategy is proposed, which increases the reward objects of some passed test cases, similar to the failure test cases. The DTW-based reward function and the similarity-based reward strategy together constitute the proposed adaptive reward mechanism in RL-based CITCP. To validate the effectiveness of the adaptive reward mechanism, experimental verification is carried out on 13 industrial data sets. The experimental results show that the adaptive reward mechanism can improve the TCP effect, where the average NAPFD is maximally improved by 7.29%, the average Recall is maximally improved by 6.04% and the average TTF is improved by 6.81 positions with a maximum of 63.77.
引用
收藏
页码:1457 / 1483
页数:27
相关论文
共 50 条
  • [41] Power Integration Mall Recommendation Model Based on Reverse Reward Feedback Learning Optimization
    Xiong Lei
    Lv Xinke
    Guan Yongming
    Liu Zhiyong
    PROCEEDINGS OF 2019 2ND INTERNATIONAL CONFERENCE ON BIG DATA TECHNOLOGIES (ICBDT 2019), 2019, : 101 - 108
  • [42] Modified Atom Search Optimization Based on Immunologic Mechanism and Reinforcement Learning
    Fu, Yanming
    Li, Zhuohang
    Qu, Chiwen
    Chen, Haiqiang
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2020, 2020
  • [43] Navigation of Mobile Robots Based on Deep Reinforcement Learning: Reward Function Optimization and Knowledge Transfer
    Li, Weijie
    Yue, Ming
    Shangguan, Jinyong
    Jin, Ye
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2023, 21 (02) : 563 - 574
  • [44] Deep Reinforcement Learning Visual Target Navigation Method Based on Attention Mechanism and Reward Shaping
    Meng, Yiyue
    Guo, Chi
    Liu, Jingnan
    Wuhan Daxue Xuebao (Xinxi Kexue Ban)/Geomatics and Information Science of Wuhan University, 2024, 49 (07): : 1100 - 1108
  • [45] DYNAMIC SECURITY RULE OPTIMIZATION BASED ON DEEP LEARNING AND ADAPTIVE ALGORITHMS
    HANG F.
    XIE L.
    ZHANG Z.
    LIU Y.
    HU J.I.A.N.
    Scalable Computing, 2024, 25 (04): : 2603 - 2613
  • [46] Navigation of Mobile Robots Based on Deep Reinforcement Learning: Reward Function Optimization and Knowledge Transfer
    Weijie Li
    Ming Yue
    Jinyong Shangguan
    Ye Jin
    International Journal of Control, Automation and Systems, 2023, 21 : 563 - 574
  • [47] A review on modeling tumor dynamics and agent reward functions in reinforcement learning based therapy optimization
    Almasy, Marton Gyorgy
    Horompo, Andras
    Kiss, Daniel
    Kertesz, Gabor
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 43 (06) : 6939 - 6946
  • [48] Optimization framework of laser oscillation welding based on a deep predictive reward reinforcement learning net
    Tian, Wenhao
    Hu, Peipei
    Zhang, Chen
    JOURNAL OF INTELLIGENT MANUFACTURING, 2024,
  • [49] GNN-RL: Dynamic Reward Mechanism for Connected Vehicle Security using Graph Neural Networks and Reinforcement Learning
    Rathore, Heena
    Griffith, Henry
    2023 IEEE INTERNATIONAL CONFERENCE ON SMART COMPUTING, SMARTCOMP, 2023, : 201 - 203
  • [50] A Hybrid ACO Algorithm Based on Bayesian Factorizations and Reinforcement Learning for Continuous Optimization
    Liu, Qishuai
    Hui, Qing
    2016 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2016, : 4236 - 4243