Generating transferable adversarial examples based on perceptually-aligned perturbation

被引:0
|
作者
Hongqiao Chen
Keda Lu
Xianmin Wang
Jin Li
机构
[1] Guangzhou University,Institute of Artificial Intelligence and Blockchain
[2] Chinese Academy of Sciences,State Key Laboratory of Information Security
关键词
Adversarial example; Transferability; Robust model; Perceptually-aligned perturbation;
D O I
暂无
中图分类号
学科分类号
摘要
Neural networks (NNs) are known to be susceptible to adversarial examples (AEs), which are intentionally designed to deceive a target classifier by adding small perturbations to the inputs. And interestingly, AEs crafted for one NN can mislead another model. Such a property is referred to as transferability, which is often leveraged to perform attacks in black-box settings. To mitigate the transferability of AEs, many approaches are explored to enhance the NN’s robustness. Especially, adversarial training (AT) and its variants are shown be the strongest defense to resist such transferable AEs. To boost the transferability of AEs against the robust models that have undergone AT, a novel AE generating method is proposed in this paper. The motivation of our method is based on the observation that robust models with AT is more sensitive to the perceptually-relevant gradients, hence it is reasonable to synthesize the AEs by the perturbations that have the perceptually-aligned features. The detailed process of the proposed method is given as below. First, by optimizing the loss function over an ensemble of random noised inputs, we obtain perceptually-aligned perturbations that have the noise-invariant property. Second, we employ Perona–Malik (P–M) filter to smooth the derived adversarial perturbations, such that the perceptually-relevant feature of the perturbation is significantly reinforced and the local oscillation of the perturbation is substantially suppressed. Our method can be generally applied to any gradient-based attack method. We carry out extensive experiments under ImageNet dataset for various robust and non-robust models, and the experimental results demonstrate the effectiveness of our method. Particularly, by combining our method with diverse inputs method and momentum iterative fast gradient sign method, we can achieve state-of-the-art performance in terms of fooling the robust models.
引用
收藏
页码:3295 / 3307
页数:12
相关论文
共 50 条
  • [41] Feature Space Perturbations Yield More Transferable Adversarial Examples
    Inkawhich, Nathan
    Wen, Wei
    Li, Hai
    Chen, Yiran
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 7059 - 7067
  • [42] Efficient and Transferable Adversarial Examples from Bayesian Neural Networks
    Gubri, Martin
    Cordy, Maxime
    Papadakis, Mike
    Le Traon, Yves
    Sen, Koushik
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 180, 2022, 180 : 738 - 748
  • [43] Generating adversarial examples with collaborative generative models
    Xu, Lei
    Zhai, Junhai
    INTERNATIONAL JOURNAL OF INFORMATION SECURITY, 2024, 23 (02) : 1077 - 1091
  • [44] An efficient framework for generating robust adversarial examples
    Zhang, Lili
    Wang, Xiaoping
    Lu, Kai
    Peng, Shaoliang
    Wang, Xiaodong
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2020, 35 (09) : 1433 - 1449
  • [45] Generating adversarial examples with collaborative generative models
    Lei Xu
    Junhai Zhai
    International Journal of Information Security, 2024, 23 : 1077 - 1091
  • [46] Generating adversarial examples with input significance indicator
    Qiu, Xiaofeng
    Zhou, Shuya
    NEUROCOMPUTING, 2020, 394 : 1 - 12
  • [47] Generating Fluent Adversarial Examples for Natural Languages
    Zhang, Huangzhao
    Zhou, Hao
    Miao, Ning
    Li, Lei
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5564 - 5569
  • [48] A Transferable Adversarial Belief Attack With Salient Region Perturbation Restriction
    Zhang, Shihui
    Zuo, Dongxu
    Yang, Yongliang
    Zhang, Xiaowei
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4296 - 4306
  • [49] Generating Adversarial Examples by Adversarial Networks for Semi-supervised Learning
    Ma, Yun
    Mao, Xudong
    Chen, Yangbin
    Li, Qing
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2019, 2019, 11881 : 115 - 129
  • [50] Defending against and generating adversarial examples together with generative adversarial networks
    Ying Wang
    Xiao Liao
    Wei Cui
    Yang Yang
    Scientific Reports, 15 (1)