Generating transferable adversarial examples based on perceptually-aligned perturbation

被引:0
|
作者
Hongqiao Chen
Keda Lu
Xianmin Wang
Jin Li
机构
[1] Guangzhou University,Institute of Artificial Intelligence and Blockchain
[2] Chinese Academy of Sciences,State Key Laboratory of Information Security
关键词
Adversarial example; Transferability; Robust model; Perceptually-aligned perturbation;
D O I
暂无
中图分类号
学科分类号
摘要
Neural networks (NNs) are known to be susceptible to adversarial examples (AEs), which are intentionally designed to deceive a target classifier by adding small perturbations to the inputs. And interestingly, AEs crafted for one NN can mislead another model. Such a property is referred to as transferability, which is often leveraged to perform attacks in black-box settings. To mitigate the transferability of AEs, many approaches are explored to enhance the NN’s robustness. Especially, adversarial training (AT) and its variants are shown be the strongest defense to resist such transferable AEs. To boost the transferability of AEs against the robust models that have undergone AT, a novel AE generating method is proposed in this paper. The motivation of our method is based on the observation that robust models with AT is more sensitive to the perceptually-relevant gradients, hence it is reasonable to synthesize the AEs by the perturbations that have the perceptually-aligned features. The detailed process of the proposed method is given as below. First, by optimizing the loss function over an ensemble of random noised inputs, we obtain perceptually-aligned perturbations that have the noise-invariant property. Second, we employ Perona–Malik (P–M) filter to smooth the derived adversarial perturbations, such that the perceptually-relevant feature of the perturbation is significantly reinforced and the local oscillation of the perturbation is substantially suppressed. Our method can be generally applied to any gradient-based attack method. We carry out extensive experiments under ImageNet dataset for various robust and non-robust models, and the experimental results demonstrate the effectiveness of our method. Particularly, by combining our method with diverse inputs method and momentum iterative fast gradient sign method, we can achieve state-of-the-art performance in terms of fooling the robust models.
引用
收藏
页码:3295 / 3307
页数:12
相关论文
共 50 条
  • [31] Push & Pull: Transferable Adversarial Examples With Attentive Attack
    Gao, Lianli
    Huang, Zijie
    Song, Jingkuan
    Yang, Yang
    Shen, Heng Tao
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 2329 - 2338
  • [32] Generative Transferable Universal Adversarial Perturbation for Combating Deepfakes
    Guo, Yuchen
    Wang, Xi
    Fu, Xiaomeng
    Li, Jin
    Li, Zhaoxing
    Chai, Yesheng
    Hao, Jizhong
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 1980 - 1985
  • [33] Generating universal adversarial perturbation with ResNet
    Xu, Jian
    Liu, Heng
    Wu, Dexin
    Zhou, Fucai
    Gao, Chong-zhi
    Jiang, Linzhi
    INFORMATION SCIENCES, 2020, 537 (537) : 302 - 312
  • [34] Structure Matters: Towards Generating Transferable Adversarial Images
    Peng, Dan
    Zheng, Zizhan
    Luo, Linhao
    Zhang, Xiaofeng
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1419 - 1426
  • [35] Adversarial transformation network with adaptive perturbations for generating adversarial examples
    Zhang, Guoyin
    Da, Qingan
    Li, Sizhao
    Sun, Jianguo
    Wang, Wenshan
    Hu, Qing
    Lu, Jiashuai
    INTERNATIONAL JOURNAL OF BIO-INSPIRED COMPUTATION, 2022, 20 (02) : 94 - 103
  • [36] Generating Adversarial Examples With Distance Constrained Adversarial Imitation Networks
    Tang, Pengfei
    Wang, Wenjie
    Lou, Jian
    Xiong, Li
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2022, 19 (06) : 4145 - 4155
  • [37] WBA: A Warping-based Approach to Generating Imperceptible Adversarial Examples
    Hua, Chengyao
    Zhang, Shigeng
    Wang, Weiping
    Li, Zhankai
    Zhang, Jian
    2022 IEEE INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, 2022, : 361 - 368
  • [38] A Region-Adaptive Local Perturbation-Based Method for Generating Adversarial Examples in Synthetic Aperture Radar Object Detection
    Duan, Jiale
    Qiu, Linyao
    He, Guangjun
    Zhao, Ling
    Zhang, Zhenshi
    Li, Haifeng
    REMOTE SENSING, 2024, 16 (06)
  • [39] Transferable adversarial examples can efficiently fool topic models
    Wang, Zhen
    Zheng, Yitao
    Zhu, Hai
    Yang, Chang
    Chen, Tianyi
    COMPUTERS & SECURITY, 2022, 118
  • [40] Dynamic loss yielding more transferable targeted adversarial examples
    Zhang, Ming
    Chen, Yongkang
    Li, Hu
    Qian, Cheng
    Kuang, Xiaohui
    NEUROCOMPUTING, 2024, 590