Push & Pull: Transferable Adversarial Examples With Attentive Attack

被引:29
|
作者
Gao, Lianli [1 ,2 ,3 ]
Huang, Zijie [2 ,3 ]
Song, Jingkuan [1 ]
Yang, Yang [2 ,3 ]
Shen, Heng Tao [2 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Inst Neurol, Sichuan Prov Peoples Hosp, Chengdu 611731, Peoples R China
[2] Univ Elect Sci & Technol China, Future Media Ctr, Chengdu 611731, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Feature extraction; Computational modeling; Task analysis; Predictive models; Neural networks; Iterative methods; Image classification; adversarial attack; transferability; targeted attack;
D O I
10.1109/TMM.2021.3079723
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Targeted attack aims to mislead the classification model to a specific class, and it can be further divided into black-box and white-box targeted attack depending on whether the classification model is known. A growing number of approaches rely on disrupting the image representations to craft adversarial examples. However, this type of methods often suffer from either low white-box targeted attack success rate or poor black-box targeted attack transferability. To address these problems, we propose a Transferable Attentive Attack (TAA) method which adds perturbation to clean images based on the attended regions and features. This is motivated by one important observation that deep-learning based classification models (or even shallow-learning based models like SIFT) make the prediction mainly based on the informative and discriminative regions of an image. Specifically, the corresponding features of the informative regions are firstly extracted, and the anchor image's features are iteratively "pushed" away from the source class and simultaneously "pulled" closer to the target class along with attacking. Moreover, we introduce a new strategy that the attack selects the centroids of source and target class cluster as the input of triplet loss to achieve high transferability. Experimental results demonstrate that our method improves the transferability of adversarial example, while maintaining higher success rate for white-box targeted attacks compared with the state-of-the-arts. In particular, TAA attacks on image-representation based task like VQA also result in a significant performance drop in terms of accuracy.
引用
收藏
页码:2329 / 2338
页数:10
相关论文
共 50 条
  • [41] Speckle-Variant Attack: Toward Transferable Adversarial Attack to SAR Target Recognition
    Peng, Bowen
    Peng, Bo
    Zhou, Jie
    Xia, Jingyuan
    Liu, Li
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [42] Attack as Detection: Using Adversarial Attack Methods to Detect Abnormal Examples
    Zhao, Zhe
    Chen, Guangke
    Liu, Tong
    Li, Taishan
    Song, Fu
    Wang, Jingyi
    Sun, Jun
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2024, 33 (03)
  • [43] Greedy attack and gumbel attack: Generating adversarial examples for discrete data
    Yang, Puyudi
    Chen, Jianbo
    Hsieh, Cho-Jui
    Wang, Jane-Ling
    Jordan, Michael I.
    Journal of Machine Learning Research, 2020, 21
  • [44] Feature-Based Adversarial Training for Deep Learning Models Resistant to Transferable Adversarial Examples
    Ryu, Gwonsang
    Choi, Daeseon
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2022, E105D (05) : 1039 - 1049
  • [45] Crafting transferable adversarial examples via contaminating the salient feature variance
    Ren, Yuchen
    Zhu, Hegui
    Sui, Xiaoyan
    Liu, Chong
    INFORMATION SCIENCES, 2023, 644
  • [46] Generating transferable adversarial examples based on perceptually-aligned perturbation
    Chen, Hongqiao
    Lu, Keda
    Wang, Xianmin
    Li, Jin
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2021, 12 (11) : 3295 - 3307
  • [47] Rethinking the optimization objective for transferable adversarial examples from a fuzzy perspective
    Yang, Xiangyuan
    Lin, Jie
    Zhang, Hanlin
    Zhao, Peng
    NEURAL NETWORKS, 2025, 184
  • [48] Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks
    Dong, Yinpeng
    Pang, Tianyu
    Su, Hang
    Zhu, Jun
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4307 - 4316
  • [49] UCG: A Universal Cross-Domain Generator for Transferable Adversarial Examples
    Li, Zhankai
    Wang, Weiping
    Li, Jie
    Chen, Kai
    Zhang, Shigeng
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 3023 - 3037
  • [50] Generating transferable adversarial examples based on perceptually-aligned perturbation
    Hongqiao Chen
    Keda Lu
    Xianmin Wang
    Jin Li
    International Journal of Machine Learning and Cybernetics, 2021, 12 : 3295 - 3307