Push & Pull: Transferable Adversarial Examples With Attentive Attack

被引:29
|
作者
Gao, Lianli [1 ,2 ,3 ]
Huang, Zijie [2 ,3 ]
Song, Jingkuan [1 ]
Yang, Yang [2 ,3 ]
Shen, Heng Tao [2 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Inst Neurol, Sichuan Prov Peoples Hosp, Chengdu 611731, Peoples R China
[2] Univ Elect Sci & Technol China, Future Media Ctr, Chengdu 611731, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
基金
中国国家自然科学基金;
关键词
Perturbation methods; Feature extraction; Computational modeling; Task analysis; Predictive models; Neural networks; Iterative methods; Image classification; adversarial attack; transferability; targeted attack;
D O I
10.1109/TMM.2021.3079723
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Targeted attack aims to mislead the classification model to a specific class, and it can be further divided into black-box and white-box targeted attack depending on whether the classification model is known. A growing number of approaches rely on disrupting the image representations to craft adversarial examples. However, this type of methods often suffer from either low white-box targeted attack success rate or poor black-box targeted attack transferability. To address these problems, we propose a Transferable Attentive Attack (TAA) method which adds perturbation to clean images based on the attended regions and features. This is motivated by one important observation that deep-learning based classification models (or even shallow-learning based models like SIFT) make the prediction mainly based on the informative and discriminative regions of an image. Specifically, the corresponding features of the informative regions are firstly extracted, and the anchor image's features are iteratively "pushed" away from the source class and simultaneously "pulled" closer to the target class along with attacking. Moreover, we introduce a new strategy that the attack selects the centroids of source and target class cluster as the input of triplet loss to achieve high transferability. Experimental results demonstrate that our method improves the transferability of adversarial example, while maintaining higher success rate for white-box targeted attacks compared with the state-of-the-arts. In particular, TAA attacks on image-representation based task like VQA also result in a significant performance drop in terms of accuracy.
引用
收藏
页码:2329 / 2338
页数:10
相关论文
共 50 条
  • [21] Transferable adversarial examples can efficiently fool topic models
    Wang, Zhen
    Zheng, Yitao
    Zhu, Hai
    Yang, Chang
    Chen, Tianyi
    COMPUTERS & SECURITY, 2022, 118
  • [22] Dynamic loss yielding more transferable targeted adversarial examples
    Zhang, Ming
    Chen, Yongkang
    Li, Hu
    Qian, Cheng
    Kuang, Xiaohui
    NEUROCOMPUTING, 2024, 590
  • [23] Feature Space Perturbations Yield More Transferable Adversarial Examples
    Inkawhich, Nathan
    Wen, Wei
    Li, Hai
    Chen, Yiran
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 7059 - 7067
  • [24] Efficient and Transferable Adversarial Examples from Bayesian Neural Networks
    Gubri, Martin
    Cordy, Maxime
    Papadakis, Mike
    Le Traon, Yves
    Sen, Koushik
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, VOL 180, 2022, 180 : 738 - 748
  • [25] Transferable Sparse Adversarial Attack on Modulation Recognition With Generative Networks
    Jiang, Zenghui
    Zeng, Weijun
    Zhou, Xingyu
    Chen, Pu
    Yin, Shenqian
    IEEE COMMUNICATIONS LETTERS, 2024, 28 (05) : 999 - 1003
  • [26] Intermediate-Layer Transferable Adversarial Attack With DNN Attention
    Yang, Shanshan
    Yang, Yu
    Zhou, Linna
    Zhan, Rui
    Man, Yufei
    IEEE ACCESS, 2022, 10 : 95451 - 95461
  • [27] Training Meta-Surrogate Model for Transferable Adversarial Attack
    Qin, Yunxiao
    Xiong, Yuanhao
    Yi, Jinfeng
    Hsieh, Cho-Jui
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9516 - 9524
  • [28] Transferable Interpolated Adversarial Attack with Random-Layer Mixup
    Ma, Size
    Han, Keji
    Long, Xianzhong
    Li, Yun
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2022, PT II, 2022, 13281 : 224 - 235
  • [29] Black-box Bayesian adversarial attack with transferable priors
    Shudong Zhang
    Haichang Gao
    Chao Shu
    Xiwen Cao
    Yunyi Zhou
    Jianping He
    Machine Learning, 2024, 113 : 1511 - 1528
  • [30] Improving transferable adversarial attack via feature-momentum
    He, Xianglong
    Li, Yuezun
    Qu, Haipeng
    Dong, Junyu
    COMPUTERS & SECURITY, 2023, 128