GNP ATTACK: TRANSFERABLE ADVERSARIAL EXAMPLES VIA GRADIENT NORM PENALTY

被引:1
|
作者
Wu, Tao [1 ]
Luo, Tie [1 ]
Wunsch, Donald C. [2 ]
机构
[1] Missouri Univ Sci & Technol, Dept Comp Sci, Rolla, MO 65409 USA
[2] Missouri Univ Sci & Technol, Dept Elect & Comp Engn, Rolla, MO USA
关键词
Adversarial machine learning; Transferability; Deep neural networks; Input gradient regularization;
D O I
10.1109/ICIP49359.2023.10223158
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial examples (AE) with good transferability enable practical black-box attacks on diverse target models, where insider knowledge about the target models is not required. Previous methods often generate AE with no or very limited transferability; that is, they easily overfit to the particular architecture and feature representation of the source, white-box model and the generated AE barely work for target, blackbox models. In this paper, we propose a novel approach to enhance AE transferability using Gradient Norm Penalty (GNP). It drives the loss function optimization procedure to converge to a flat region of local optima in the loss landscape. By attacking 11 state-of-the-art (SOTA) deep learning models and 6 advanced defense methods, we empirically show that GNP is very effective in generating AE with high transferability. We also demonstrate that it is very flexible in that it can be easily integrated with other gradient based methods for stronger transfer-based attacks.
引用
收藏
页码:3110 / 3114
页数:5
相关论文
共 50 条
  • [21] Generating Transferable Adversarial Examples for Speech Classification
    Kim, Hoki
    Park, Jinseong
    Lee, Jaewook
    PATTERN RECOGNITION, 2023, 137
  • [22] Making Adversarial Examples More Transferable and Indistinguishable
    Zou, Junhua
    Duan, Yexin
    Li, Boyu
    Zhang, Wu
    Pan, Yu
    Pan, Zhisong
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3662 - 3670
  • [23] Towards the Transferable Reversible Adversarial Example via Distribution-Relevant Attack
    Tian, Zhuo
    Zhou, Xiaoyi
    Xing, Fan
    Zhao, Ruiyang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XI, 2025, 15041 : 292 - 305
  • [24] Transferable Structural Sparse Adversarial Attack Via Exact Group Sparsity Training
    Di Ming
    Ren, Peng
    Wang, Yunlong
    Feng, Xin
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 24696 - 24705
  • [25] Transferable adversarial attack on image tampering localization
    Cao, Gang
    Wang, Yuqi
    Zhu, Haochen
    Lou, Zijie
    Yu, Lifang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 102
  • [26] Diffusion Models for Imperceptible and Transferable Adversarial Attack
    Chen, Jianqi
    Chen, Hao
    Chen, Keyan
    Zhang, Yilan
    Zou, Zhengxia
    Shi, Zhenwei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (02) : 961 - 977
  • [27] AdvCheck: Characterizing adversarial examples via local gradient checking
    Chen, Ruoxi
    Jin, Haibo
    Chen, Jinyin
    Zheng, Haibin
    Zheng, Shilian
    Yang, Xiaoniu
    Yang, Xing
    COMPUTERS & SECURITY, 2024, 136
  • [28] Attack Agnostic Detection of Adversarial Examples via Random Subspace Analysis
    Drenkow, Nathan
    Fendley, Neil
    Burlina, Philippe
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 2815 - 2825
  • [29] UNIVERSAL ADVERSARIAL ATTACK VIA ENHANCED PROJECTED GRADIENT DESCENT
    Deng, Yingpeng
    Karam, Lina J.
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 1241 - 1245
  • [30] Transferable adversarial examples based on global smooth perturbations
    Liu, Yujia
    Jiang, Ming
    Jiang, Tingting
    COMPUTERS & SECURITY, 2022, 121