Using Reinforcement Learning to Escape Automatic Filter-based Adversarial Example Defense

被引:0
|
作者
Li, Yantao [1 ]
Dan, Kaijian [1 ]
Lei, Xinyu [2 ]
Qin, Huafeng [3 ]
Deng, Shaojiang [1 ]
Zhou, Gang [4 ]
机构
[1] Chongqing Univ, Coll Comp Sci, Chongqing, Peoples R China
[2] Michigan Technol Univ, Dept Comp Sci, Houghton, MI USA
[3] Chongqing Technol & Business Univ, Sch Comp Sci & Informat Engn, Chongqing, Peoples R China
[4] William & Mary, Comp Sci Dept, Williamsburg, VA USA
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Adversarial examples; image classification; reinforcement learning; filter;
D O I
10.1145/3688847
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks can be easily fooled by the adversarial example, which is a specially crafted example with subtle and intentional perturbations. A plethora of papers have proposed to use filters to effectively defend against adversarial example attacks. However, we demonstrate that the automatic filter-based defenses may not be reliable. In this article, we present URL2AED, Using a Reinforcement Learning scheme TO escape the automatic filter-based Adversarial Example Defenses. Specifically, URL2AED uses a specially crafted policy gradient reinforcement learning (RL) algorithm to generate adversarial examples (AEs) that can escape automatic filter-based AE defenses. In particular, we properly design reward functions in policy-gradient RL for targeted attacks and non-targeted attacks, respectively. Furthermore, we customize training algorithms to reduce the possible action space in policy-gradient RL to accelerate URL2AED training while still ensuring that URL2AED generates successful AEs. To demonstrate the performance of the proposed URL2AED, we conduct extensive experiments on three public datasets in terms of different perturbation degrees of parameter, different filter parameters, transferability, and time consumption. The experimental results show that URL2AED achieves high attack success rates for automatic filter-based defenses and good cross-model transferability.
引用
收藏
页数:26
相关论文
共 50 条
  • [31] Filter-based Online Neuro-Fuzzy Model Learning using Noisy Measurements
    Gu, Wen
    Lan, Jianglin
    Mason, Byron
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [32] A Gaussian Filter-Based Feature Learning Approach Using Genetic Programming to Image Classification
    Bi, Ying
    Xue, Bing
    Zhang, Mengjie
    AI 2018: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, 11320 : 251 - 257
  • [33] Image Captioning using Adversarial Networks and Reinforcement Learning
    Yan, Shiyang
    Wu, Fangyu
    Smith, Jeremy S.
    Lu, Wenjin
    Zhang, Bailing
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 248 - 253
  • [34] Identification of Adversarial Android Intents using Reinforcement Learning
    Rathore, Hemant
    Nikam, Piyush
    Sahay, Sanjay K.
    Sewak, Mohit
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [35] Automatic berthing using supervised learning and reinforcement learning
    Shimizu, Shoma
    Nishihara, Kenta
    Miyauchi, Yoshiki
    Wakita, Kouki
    Suyama, Rin
    Maki, Atsuo
    Shirakawa, Shinichi
    OCEAN ENGINEERING, 2022, 265
  • [36] Infrared Adversarial Patch Generation Based on Reinforcement Learning
    Zhou, Shuangju
    Li, Yang
    Tan, Wenyi
    Zhao, Chenxing
    Zhou, Xin
    Pan, Quan
    MATHEMATICS, 2024, 12 (21)
  • [37] An adversarial reinforcement learning based system for cyber security
    Xia, Song
    Qiu, Meikang
    Jiang, Hao
    4TH IEEE INTERNATIONAL CONFERENCE ON SMART CLOUD (SMARTCLOUD 2019) / 3RD INTERNATIONAL SYMPOSIUM ON REINFORCEMENT LEARNING (ISRL 2019), 2019, : 227 - 230
  • [38] Fault detection method based on adversarial reinforcement learning
    Li, Junhuai
    Wu, Yunwen
    Wang, Huaijun
    Xu, Jiang
    FRONTIERS IN COMPUTER SCIENCE, 2023, 4
  • [39] AUTOMATIC PROGRAMMING OF BEHAVIOR-BASED ROBOTS USING REINFORCEMENT LEARNING
    MAHADEVAN, S
    CONNELL, J
    ARTIFICIAL INTELLIGENCE, 1992, 55 (2-3) : 311 - 365
  • [40] Catheter Tracking: Filter-Based vs. Learning-Based
    Brost, Alexander
    Wimmer, Andreas
    Liao, Rui
    Hornegger, Joachim
    Strobel, Norbert
    PATTERN RECOGNITION, 2010, 6376 : 293 - +