Reinforcement Learning-based Adversarial Attacks on Object Detectors using Reward Shaping

被引:1
|
作者
Shi, Zhenbo [1 ]
Yang, Wei [2 ]
Xu, Zhenbo [3 ]
Yu, Zhidong [1 ]
Huang, Liusheng [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] Univ Sci & Technol China, Hefei Natl Lab, Hefei, Peoples R China
[3] Beihang Univ, Hangzhou Innovat Inst, Hangzhou, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Object Detection; Reinforcement Learning; Adversarial Attack;
D O I
10.1145/3581783.3612304
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the field of object detector attacks, previous methods primarily rely on fixed gradient optimization or patch-based cover techniques, often leading to suboptimal attack performance and excessive distortions. To address these limitations, we propose a novel attack method, Interactive Reinforcement-based Sparse Attack (IRSA), which employs Reinforcement Learning (RL) to discover the vulnerabilities of object detectors and systematically generate erroneous results. Specifically, we formulate the process of seeking optimal margins for adversarial examples as a Markov Decision Process (MDP). We tackle the RL convergence difficulty through innovative reward functions and a composite optimization method for effective and efficient policy training. Moreover, the perturbations generated by IRSA are more subtle and difficult to detect while requiring less computational effort. Our method also demonstrates strong generalization capabilities against various object detectors. In summary, IRSA is a refined, efficient, and scalable interactive, iterative, end-to-end algorithm.
引用
收藏
页码:8424 / 8432
页数:9
相关论文
共 50 条
  • [21] Adversarial Attacks in a Deep Reinforcement Learning based Cluster Scheduler
    Zhang, Shaojun
    Wang, Chen
    Zomaya, Albert Y.
    2020 IEEE 28TH INTERNATIONAL SYMPOSIUM ON MODELING, ANALYSIS, AND SIMULATION OF COMPUTER AND TELECOMMUNICATION SYSTEMS (MASCOTS 2020), 2020, : 1 - 8
  • [22] XSS adversarial example attacks based on deep reinforcement learning
    Chen, Li
    Tang, Cong
    He, Junjiang
    Zhao, Hui
    Lan, Xiaolong
    Li, Tao
    COMPUTERS & SECURITY, 2022, 120
  • [23] Distributed Control using Reinforcement Learning with Temporal-Logic-Based Reward Shaping
    Zhang, Ningyuan
    Liu, Wenliang
    Belta, Calin
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 168, 2022, 168
  • [24] Bayesian Inverse Reinforcement Learning-based Reward Learning for Automated Driving
    Zeng, Di
    Zheng, Ling
    Li, Yinong
    Yang, Xiantong
    Jixie Gongcheng Xuebao/Journal of Mechanical Engineering, 2024, 60 (10): : 245 - 260
  • [25] Adversarial Attacks on Deep Learning-Based UAV Navigation Systems
    Mynuddin, Mohammed
    Khan, Sultan Uddin
    Mahmoud, Nabil Mahmoud
    Alsharif, Ahmad
    2023 IEEE CONFERENCE ON COMMUNICATIONS AND NETWORK SECURITY, CNS, 2023,
  • [26] Learning-Based Robust Anomaly Detection in the Presence of Adversarial Attacks
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2022 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2022, : 1206 - 1211
  • [27] Adversarial Attacks to Machine Learning-Based Smart Healthcare Systems
    Newaz, A. K. M. Iqtidar
    Haque, Nur Imtiazul
    Sikder, Amit Kumar
    Rahman, Mohammad Ashiqur
    Uluagac, A. Selcuk
    2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [28] Optimizing Reinforcement Learning Agents in Games Using Curriculum Learning and Reward Shaping
    Khan, Adil
    Muhammad, Muhammad
    Naeem, Muhammad
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2025, 36 (01)
  • [29] Towards a Practical Defense Against Adversarial Attacks on Deep Learning-Based Malware Detectors via Randomized Smoothing
    Gibert, Daniel
    Zizzo, Giulio
    Le, Quan
    COMPUTER SECURITY. ESORICS 2023 INTERNATIONAL WORKSHOPS, CPS4CIP, PT II, 2024, 14399 : 683 - 699
  • [30] Reward Shaping for Reinforcement Learning by Emotion Expressions
    Hwang, K. S.
    Ling, J. L.
    Chen, Yu-Ying
    Wang, Wei-Han
    2014 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2014, : 1288 - 1293