Using Reinforcement Learning to Escape Automatic Filter-based Adversarial Example Defense

被引:0
|
作者
Li, Yantao [1 ]
Dan, Kaijian [1 ]
Lei, Xinyu [2 ]
Qin, Huafeng [3 ]
Deng, Shaojiang [1 ]
Zhou, Gang [4 ]
机构
[1] Chongqing Univ, Coll Comp Sci, Chongqing, Peoples R China
[2] Michigan Technol Univ, Dept Comp Sci, Houghton, MI USA
[3] Chongqing Technol & Business Univ, Sch Comp Sci & Informat Engn, Chongqing, Peoples R China
[4] William & Mary, Comp Sci Dept, Williamsburg, VA USA
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Adversarial examples; image classification; reinforcement learning; filter;
D O I
10.1145/3688847
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks can be easily fooled by the adversarial example, which is a specially crafted example with subtle and intentional perturbations. A plethora of papers have proposed to use filters to effectively defend against adversarial example attacks. However, we demonstrate that the automatic filter-based defenses may not be reliable. In this article, we present URL2AED, Using a Reinforcement Learning scheme TO escape the automatic filter-based Adversarial Example Defenses. Specifically, URL2AED uses a specially crafted policy gradient reinforcement learning (RL) algorithm to generate adversarial examples (AEs) that can escape automatic filter-based AE defenses. In particular, we properly design reward functions in policy-gradient RL for targeted attacks and non-targeted attacks, respectively. Furthermore, we customize training algorithms to reduce the possible action space in policy-gradient RL to accelerate URL2AED training while still ensuring that URL2AED generates successful AEs. To demonstrate the performance of the proposed URL2AED, we conduct extensive experiments on three public datasets in terms of different perturbation degrees of parameter, different filter parameters, transferability, and time consumption. The experimental results show that URL2AED achieves high attack success rates for automatic filter-based defenses and good cross-model transferability.
引用
收藏
页数:26
相关论文
共 50 条
  • [41] Author Correction: A multi-agent reinforcement learning based approach for automatic filter pruning
    Zhemin Li
    Xiaojing Zuo
    Yiping Song
    Dong Liang
    Zheng Xie
    Scientific Reports, 15 (1)
  • [42] Kalman Filter-Based Differential Privacy Federated Learning Method
    Yang, Xiaohui
    Dong, Zijian
    APPLIED SCIENCES-BASEL, 2022, 12 (15):
  • [43] RL-VAEGAN: Adversarial defense for reinforcement learning agents via style transfer
    Hu, Yueyue
    Sun, Shiliang
    KNOWLEDGE-BASED SYSTEMS, 2021, 221
  • [44] AN ANALYSIS OF SAMPLING FOR FILTER-BASED FEATURE EXTRACTION AND ADABOOST LEARNING
    Haselhoff, Anselm
    Kummert, Anton
    VISAPP 2009: PROCEEDINGS OF THE FOURTH INTERNATIONAL CONFERENCE ON COMPUTER VISION THEORY AND APPLICATIONS, VOL 2, 2009, : 180 - 185
  • [45] A Kalman Filter-based Actor-Critic Learning Approach
    Wang, Bin
    Zhao, Dongbin
    PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2014, : 3657 - 3662
  • [46] Reinforcement Learning-based Adversarial Attacks on Object Detectors using Reward Shaping
    Shi, Zhenbo
    Yang, Wei
    Xu, Zhenbo
    Yu, Zhidong
    Huang, Liusheng
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 8424 - 8432
  • [47] Automatic core design using reinforcement learning
    Kobayashi, Y
    Aiyoshi, E
    PROCEEDINGS OF THE 2004 AMERICAN CONTROL CONFERENCE, VOLS 1-6, 2004, : 5784 - 5789
  • [48] A New Gabor Filter-Based Method for Automatic Recognition of Hatched Residential Areas
    Wu, Jianhua
    Wei, Pengjie
    Yuan, Xiaofang
    Shu, Zhigang
    Chiang, Yao-Yi
    Fu, Zhongliang
    Deng, Min
    IEEE ACCESS, 2019, 7 : 40649 - 40662
  • [49] A novel Kalman filter-based navigation using beacons
    Doraiswami, R
    IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 1996, 32 (02) : 830 - 840
  • [50] Scalable and Autonomous Network Defense Using Reinforcement Learning
    Campbell, Robert G.
    Eirinaki, Magdalini
    Park, Younghee
    IEEE ACCESS, 2024, 12 : 92919 - 92930