Generating transferable adversarial examples based on perceptually-aligned perturbation

被引:0
|
作者
Hongqiao Chen
Keda Lu
Xianmin Wang
Jin Li
机构
[1] Guangzhou University,Institute of Artificial Intelligence and Blockchain
[2] Chinese Academy of Sciences,State Key Laboratory of Information Security
关键词
Adversarial example; Transferability; Robust model; Perceptually-aligned perturbation;
D O I
暂无
中图分类号
学科分类号
摘要
Neural networks (NNs) are known to be susceptible to adversarial examples (AEs), which are intentionally designed to deceive a target classifier by adding small perturbations to the inputs. And interestingly, AEs crafted for one NN can mislead another model. Such a property is referred to as transferability, which is often leveraged to perform attacks in black-box settings. To mitigate the transferability of AEs, many approaches are explored to enhance the NN’s robustness. Especially, adversarial training (AT) and its variants are shown be the strongest defense to resist such transferable AEs. To boost the transferability of AEs against the robust models that have undergone AT, a novel AE generating method is proposed in this paper. The motivation of our method is based on the observation that robust models with AT is more sensitive to the perceptually-relevant gradients, hence it is reasonable to synthesize the AEs by the perturbations that have the perceptually-aligned features. The detailed process of the proposed method is given as below. First, by optimizing the loss function over an ensemble of random noised inputs, we obtain perceptually-aligned perturbations that have the noise-invariant property. Second, we employ Perona–Malik (P–M) filter to smooth the derived adversarial perturbations, such that the perceptually-relevant feature of the perturbation is significantly reinforced and the local oscillation of the perturbation is substantially suppressed. Our method can be generally applied to any gradient-based attack method. We carry out extensive experiments under ImageNet dataset for various robust and non-robust models, and the experimental results demonstrate the effectiveness of our method. Particularly, by combining our method with diverse inputs method and momentum iterative fast gradient sign method, we can achieve state-of-the-art performance in terms of fooling the robust models.
引用
收藏
页码:3295 / 3307
页数:12
相关论文
共 50 条
  • [21] Feature-Based Adversarial Training for Deep Learning Models Resistant to Transferable Adversarial Examples
    Ryu, Gwonsang
    Choi, Daeseon
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (05) : 1039 - 1049
  • [22] Generating Adversarial Examples With Conditional Generative Adversarial Net
    Yu, Ping
    Song, Kaitao
    Lu, Jianfeng
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 676 - 681
  • [23] Towards Transferable Adversarial Examples Using Meta Learning
    Fan, Mingyuan
    Yin, Jia-Li
    Liu, Ximeng
    Guo, Wenzhong
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2021, PT I, 2022, 13155 : 178 - 192
  • [24] Generating facial expression adversarial examples based on saliency map
    Sun, Yudao
    Yin, Juan
    Wu, Chunhua
    Zheng, KangFeng
    Niu, XinXin
    IMAGE AND VISION COMPUTING, 2021, 116
  • [25] Learning Transferable Adversarial Examples via Ghost Networks
    Li, Yingwei
    Bai, Song
    Zhou, Yuyin
    Xie, Cihang
    Zhang, Zhishuai
    Yuille, Alan
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11458 - 11465
  • [26] Generating Natural Language Adversarial Examples
    Alzantot, Moustafa
    Sharma, Yash
    Elgohary, Ahmed
    Ho, Bo-Jhang
    Srivastava, Mani B.
    Chang, Kai-Wei
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 2890 - 2896
  • [27] Generating Adversarial Examples With Shadow Model
    Zhang, Rui
    Xia, Hui
    Hu, Chunqiang
    Zhang, Cheng
    Liu, Chao
    Xiao, Fu
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (09) : 6283 - 6289
  • [28] Generating Watermarked Speech Adversarial Examples
    Wang, Yumin
    Ye, Jingyu
    Wu, Hanzhou
    PROCEEDINGS OF ACM TURING AWARD CELEBRATION CONFERENCE, ACM TURC 2021, 2021, : 254 - 260
  • [29] Towards Transferable Unrestricted Adversarial Examples with Minimum Changes
    Liu, Fangcheng
    Zhang, Chao
    Zhang, Hongyang
    2023 IEEE CONFERENCE ON SECURE AND TRUSTWORTHY MACHINE LEARNING, SATML, 2023, : 327 - 338
  • [30] Direction-aggregated Attack for Transferable Adversarial Examples
    Huang, Tianjin
    Menkovski, Vlado
    Pei, Yulong
    Wang, Yuhao
    Pechenizkiy, Mykola
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2022, 18 (03)