Reinforcement Learning-based Adversarial Attacks on Object Detectors using Reward Shaping

被引:1
|
作者
Shi, Zhenbo [1 ]
Yang, Wei [2 ]
Xu, Zhenbo [3 ]
Yu, Zhidong [1 ]
Huang, Liusheng [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] Univ Sci & Technol China, Hefei Natl Lab, Hefei, Peoples R China
[3] Beihang Univ, Hangzhou Innovat Inst, Hangzhou, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Object Detection; Reinforcement Learning; Adversarial Attack;
D O I
10.1145/3581783.3612304
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the field of object detector attacks, previous methods primarily rely on fixed gradient optimization or patch-based cover techniques, often leading to suboptimal attack performance and excessive distortions. To address these limitations, we propose a novel attack method, Interactive Reinforcement-based Sparse Attack (IRSA), which employs Reinforcement Learning (RL) to discover the vulnerabilities of object detectors and systematically generate erroneous results. Specifically, we formulate the process of seeking optimal margins for adversarial examples as a Markov Decision Process (MDP). We tackle the RL convergence difficulty through innovative reward functions and a composite optimization method for effective and efficient policy training. Moreover, the perturbations generated by IRSA are more subtle and difficult to detect while requiring less computational effort. Our method also demonstrates strong generalization capabilities against various object detectors. In summary, IRSA is a refined, efficient, and scalable interactive, iterative, end-to-end algorithm.
引用
收藏
页码:8424 / 8432
页数:9
相关论文
共 50 条
  • [1] Adversarial Attacks and Detection on Reinforcement Learning-Based Interactive Recommender Systems
    Cao, Yuanjiang
    Chen, Xiaocong
    Yao, Lina
    Wang, Xianzhi
    Zhang, Wei Emma
    PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 1669 - 1672
  • [2] Adversarial Attacks Against Reinforcement Learning-Based Portfolio Management Strategy
    Chen, Yu-Ying
    Chen, Chiao-Ting
    Sang, Chuan-Yun
    Yang, Yao-Chun
    Huang, Szu-Hao
    IEEE ACCESS, 2021, 9 : 50667 - 50685
  • [3] Countering Evasion Attacks for Smart Grid Reinforcement Learning-Based Detectors
    El-Toukhy, Ahmed T.
    Mahmoud, Mohamed M. E. A.
    Bondok, Atef H.
    Fouda, Mostafa M.
    Alsabaan, Maazen
    IEEE ACCESS, 2023, 11 : 97373 - 97390
  • [4] Evading Deep Reinforcement Learning-based Network Intrusion Detection with Adversarial Attacks
    Merzouk, Mohamed Amine
    Delas, Josephine
    Neal, Christopher
    Cuppens, Frederic
    Boulahia-Cuppens, Nora
    Yaich, Reda
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AVAILABILITY, RELIABILITY AND SECURITY, ARES 2022, 2022,
  • [5] Reward Shaping Based Federated Reinforcement Learning
    Hu, Yiqiu
    Hua, Yun
    Liu, Wenyan
    Zhu, Jun
    IEEE ACCESS, 2021, 9 : 67259 - 67267
  • [6] Realistic Adversarial Attacks on Object Detectors Using Generative Models
    D. Shelepneva
    K. Arkhipenko
    Journal of Mathematical Sciences, 2024, 285 (2) : 245 - 254
  • [7] Using Natural Language for Reward Shaping in Reinforcement Learning
    Goyal, Prasoon
    Niekum, Scott
    Mooney, Raymond J.
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2385 - 2391
  • [8] ADVERSARIAL ATTACKS ON OBJECT DETECTORS WITH LIMITED PERTURBATIONS
    Shi, Zhenbo
    Yang, Wei
    Xu, Zhenbo
    Chen, Zhi
    Li, Yingjie
    Zhu, Haoran
    Huang, Liusheng
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 1375 - 1379
  • [9] Plan-based Reward Shaping for Reinforcement Learning
    Grzes, Marek
    Kudenko, Daniel
    2008 4TH INTERNATIONAL IEEE CONFERENCE INTELLIGENT SYSTEMS, VOLS 1 AND 2, 2008, : 416 - 423
  • [10] Potential Based Reward Shaping for Hierarchical Reinforcement Learning
    Gao, Yang
    Toni, Francesca
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3504 - 3510