Using Reinforcement Learning to Escape Automatic Filter-based Adversarial Example Defense

被引:0
|
作者
Li, Yantao [1 ]
Dan, Kaijian [1 ]
Lei, Xinyu [2 ]
Qin, Huafeng [3 ]
Deng, Shaojiang [1 ]
Zhou, Gang [4 ]
机构
[1] Chongqing Univ, Coll Comp Sci, Chongqing, Peoples R China
[2] Michigan Technol Univ, Dept Comp Sci, Houghton, MI USA
[3] Chongqing Technol & Business Univ, Sch Comp Sci & Informat Engn, Chongqing, Peoples R China
[4] William & Mary, Comp Sci Dept, Williamsburg, VA USA
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Adversarial examples; image classification; reinforcement learning; filter;
D O I
10.1145/3688847
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks can be easily fooled by the adversarial example, which is a specially crafted example with subtle and intentional perturbations. A plethora of papers have proposed to use filters to effectively defend against adversarial example attacks. However, we demonstrate that the automatic filter-based defenses may not be reliable. In this article, we present URL2AED, Using a Reinforcement Learning scheme TO escape the automatic filter-based Adversarial Example Defenses. Specifically, URL2AED uses a specially crafted policy gradient reinforcement learning (RL) algorithm to generate adversarial examples (AEs) that can escape automatic filter-based AE defenses. In particular, we properly design reward functions in policy-gradient RL for targeted attacks and non-targeted attacks, respectively. Furthermore, we customize training algorithms to reduce the possible action space in policy-gradient RL to accelerate URL2AED training while still ensuring that URL2AED generates successful AEs. To demonstrate the performance of the proposed URL2AED, we conduct extensive experiments on three public datasets in terms of different perturbation degrees of parameter, different filter parameters, transferability, and time consumption. The experimental results show that URL2AED achieves high attack success rates for automatic filter-based defenses and good cross-model transferability.
引用
收藏
页数:26
相关论文
共 50 条
  • [21] Adversarial Example Defense Method Based on Inverse Perturbation Fusing Generative Adversarial Network
    Zhang S.-H.
    Zhang X.-W.
    Song D.-D.
    Yang Y.-L.
    Zuo D.-X.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2023, 51 (04): : 879 - 884
  • [22] Adversarial attack and defense in reinforcement learning-from AI security view
    Chen, Tong
    Liu, Jiqiang
    Xiang, Yingxiao
    Niu, Wenjia
    Tong, Endong
    Han, Zhen
    CYBERSECURITY, 2019, 2 (01)
  • [23] Adversarial attack and defense in reinforcement learning-from AI security view
    Tong Chen
    Jiqiang Liu
    Yingxiao Xiang
    Wenjia Niu
    Endong Tong
    Zhen Han
    Cybersecurity, 2
  • [24] Defense Strategies Against Adversarial Jamming Attacks via Deep Reinforcement Learning
    Wang, Feng
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2020 54TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2020, : 336 - 341
  • [25] Automatic Object Recoloring Using Adversarial Learning
    Khodadadeh, Siavash
    Motiian, Saeid
    Lin, Zhe
    Boloni, Ladislau
    Ghadar, Shabnam
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 1487 - 1495
  • [26] Using Collision Momentum in Deep Reinforcement Learning based Adversarial Pedestrian Modeling
    Chen, Dianwei
    Yurtsever, Ekim
    Redmill, Keith A.
    Ozguner, Umit
    2023 IEEE INTELLIGENT VEHICLES SYMPOSIUM, IV, 2023,
  • [27] iDensity: An automatic Gabor filter-based algorithm for breast density assessment
    Gandomkar, Ziba
    Tay, Kevin
    Ryder, Will
    Brennan, Patrick C.
    Mello-Thoms, Claudia
    MEDICAL IMAGING 2015: IMAGE PERCEPTION, OBSERVER PERFORMANCE, AND TECHNOLOGY ASSESSMENT, 2015, 9416
  • [28] Forming Adversarial Example Attacks Against Deep Neural Networks With Reinforcement Learning
    Akers, Matthew
    Barton, Armon
    COMPUTER, 2024, 57 (01) : 88 - 99
  • [29] Adversarial attacks on cooperative multi-agent deep reinforcement learning: a dynamic group-based adversarial example transferability method
    Zan, Lixia
    Zhu, Xiangbin
    Hu, Zhao-Long
    COMPLEX & INTELLIGENT SYSTEMS, 2023, 9 (06) : 7439 - 7450
  • [30] Adversarial attacks on cooperative multi-agent deep reinforcement learning: a dynamic group-based adversarial example transferability method
    Lixia Zan
    Xiangbin Zhu
    Zhao-Long Hu
    Complex & Intelligent Systems, 2023, 9 : 7439 - 7450