Generating transferable adversarial examples based on perceptually-aligned perturbation

被引:0
|
作者
Hongqiao Chen
Keda Lu
Xianmin Wang
Jin Li
机构
[1] Guangzhou University,Institute of Artificial Intelligence and Blockchain
[2] Chinese Academy of Sciences,State Key Laboratory of Information Security
关键词
Adversarial example; Transferability; Robust model; Perceptually-aligned perturbation;
D O I
暂无
中图分类号
学科分类号
摘要
Neural networks (NNs) are known to be susceptible to adversarial examples (AEs), which are intentionally designed to deceive a target classifier by adding small perturbations to the inputs. And interestingly, AEs crafted for one NN can mislead another model. Such a property is referred to as transferability, which is often leveraged to perform attacks in black-box settings. To mitigate the transferability of AEs, many approaches are explored to enhance the NN’s robustness. Especially, adversarial training (AT) and its variants are shown be the strongest defense to resist such transferable AEs. To boost the transferability of AEs against the robust models that have undergone AT, a novel AE generating method is proposed in this paper. The motivation of our method is based on the observation that robust models with AT is more sensitive to the perceptually-relevant gradients, hence it is reasonable to synthesize the AEs by the perturbations that have the perceptually-aligned features. The detailed process of the proposed method is given as below. First, by optimizing the loss function over an ensemble of random noised inputs, we obtain perceptually-aligned perturbations that have the noise-invariant property. Second, we employ Perona–Malik (P–M) filter to smooth the derived adversarial perturbations, such that the perceptually-relevant feature of the perturbation is significantly reinforced and the local oscillation of the perturbation is substantially suppressed. Our method can be generally applied to any gradient-based attack method. We carry out extensive experiments under ImageNet dataset for various robust and non-robust models, and the experimental results demonstrate the effectiveness of our method. Particularly, by combining our method with diverse inputs method and momentum iterative fast gradient sign method, we can achieve state-of-the-art performance in terms of fooling the robust models.
引用
收藏
页码:3295 / 3307
页数:12
相关论文
共 50 条
  • [1] Generating transferable adversarial examples based on perceptually-aligned perturbation
    Chen, Hongqiao
    Lu, Keda
    Wang, Xianmin
    Li, Jin
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (11) : 3295 - 3307
  • [2] Generating Transferable Adversarial Examples for Speech Classification
    Kim, Hoki
    Park, Jinseong
    Lee, Jaewook
    PATTERN RECOGNITION, 2023, 137
  • [3] Improving the invisibility of adversarial examples with perceptually adaptive perturbation
    Zhang, Yaoyuan
    Tan, Yu-an
    Sun, Haipeng
    Zhao, Yuhang
    Zhang, Quanxing
    Li, Yuanzhang
    INFORMATION SCIENCES, 2023, 635 : 126 - 137
  • [4] Common knowledge learning for generating transferable adversarial examples
    Yang, Ruijie
    Guo, Yuanfang
    Wang, Junfu
    Zhou, Jiantao
    Wang, Yunhong
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (10)
  • [5] Generating Transferable Adversarial Examples against Vision Transformers
    Wang, Yuxuan
    Wang, Jiakai
    Yin, Zinxin
    Gong, Ruihao
    Wang, Jingyi
    Liu, Aishan
    Liu, Xianglong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5181 - 5190
  • [6] A hypothetical defenses-based training framework for generating transferable adversarial examples
    Hao, Lingguang
    Hao, Kuangrong
    Jin, Yaochu
    Zhao, Hongzhi
    KNOWLEDGE-BASED SYSTEMS, 2024, 305
  • [7] Generating Transferable Adversarial Examples From the Perspective of Ensemble and Distribution
    Zhang, Huangyi
    Liu, Ximeng
    PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CYBER SECURITY, ARTIFICIAL INTELLIGENCE AND DIGITAL ECONOMY, CSAIDE 2024, 2024, : 173 - 177
  • [8] Generating Adversarial Examples with Image-To-Perturbation Network
    Wang, Desheng
    Jin, Weidong
    Wu, Yunpu
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 7055 - 7060
  • [9] Efficient Adversarial Training with Transferable Adversarial Examples
    Zheng, Haizhong
    Zhang, Ziqi
    Gu, Juncheng
    Lee, Honglak
    Prakash, Atul
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 1178 - 1187
  • [10] Transferable adversarial examples based on global smooth perturbations
    Liu, Yujia
    Jiang, Ming
    Jiang, Tingting
    COMPUTERS & SECURITY, 2022, 121