Improving adversarial robustness of deep neural networks by using semantic information

被引:11
|
作者
Wang, Lina [1 ]
Chen, Xingshu [1 ,2 ]
Tang, Rui [1 ]
Yue, Yawei [1 ]
Zhu, Yi [2 ]
Zeng, Xuemei [2 ]
Wang, Wei [2 ]
机构
[1] Sichuan Univ, Sch Cyber Sci & Engn, Chengdu 610065, Peoples R China
[2] Sichuan Univ, Cyber Sci Res Inst, Chengdu 610065, Peoples R China
基金
中国国家自然科学基金;
关键词
Adversarial robustness; Semantic information; Region adversarial training; Targeted universal perturbations; CLASSIFICATION;
D O I
10.1016/j.knosys.2021.107141
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The vulnerability of deep neural networks (DNNs) to adversarial attack, which is an attack that can mislead state-of-the-art classifiers into making an incorrect classification with high confidence by deliberately perturbing the original inputs, raises concerns about the robustness of DNNs to such attacks. Adversarial training, which is the main heuristic method for improving adversarial robustness and the first line of defense against adversarial attacks, requires many sample-by-sample calculations to increase training size and is usually insufficiently strong for an entire network. This paper provides a new perspective on the issue of adversarial robustness, one that shifts the focus from the network as a whole to the critical part of the region close to the decision boundary corresponding to a given class. From this perspective, we propose a method to generate a single but image-agnostic adversarial perturbation that carries the semantic information implying the directions to the fragile parts on the decision boundary and causes inputs to be misclassified as a specified target. We call the adversarial training based on such perturbations "region adversarial training" (RAT), which resembles classical adversarial training but is distinguished in that it reinforces the semantic information missing in the relevant regions. Experimental results on the MNIST and CIFAR-10 datasets show that this approach greatly improves adversarial robustness even when a very small dataset from the training data is used; moreover, it can defend against fast gradient sign method, universal perturbation, projected gradient descent, and Carlini and Wagner adversarial attacks, which have a completely different pattern from those encountered by the model during retraining. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] A SIMPLE STOCHASTIC NEURAL NETWORK FOR IMPROVING ADVERSARIAL ROBUSTNESS
    Yang, Hao
    Wang, Min
    Yu, Zhengfei
    Zhou, Yun
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 2297 - 2302
  • [42] Improving the Adversarial Robustness of NLP Models by Information Bottleneck
    Zhang, Cenyuan
    Zhou, Xiang
    Wan, Yixin
    Zheng, Xiaoqing
    Chang, Kai-Wei
    Hsieh, Cho-Jui
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3588 - 3598
  • [43] Improving Adversarial Robustness via Information Bottleneck Distillation
    Kuang, Huafeng
    Liu, Hong
    Wu, YongJian
    Satoh, Shin'ichi
    Ji, Rongrong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [44] Improving Adversarial Robustness via Mutual Information Estimation
    Zhou, Dawei
    Wang, Nannan
    Gao, Xinbo
    Han, Bo
    Wang, Xiaoyu
    Zhan, Yibing
    Liu, Tongliang
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [45] Improving adversarial robustness using knowledge distillation guided by attention information bottleneck
    Gong, Yuxin
    Wang, Shen
    Yu, Tingyue
    Jiang, Xunzhi
    Sun, Fanghui
    INFORMATION SCIENCES, 2024, 665
  • [46] A regularization perspective based theoretical analysis for adversarial robustness of deep spiking neural networks
    Zhang, Hui
    Cheng, Jian
    Zhang, Jun
    Liu, Hongyi
    Wei, Zhihui
    NEURAL NETWORKS, 2023, 165 : 164 - 174
  • [47] Adversarial robustness in deep neural networks based on variable attributes of the stochastic ensemble model
    Qin, Ruoxi
    Wang, Linyuan
    Du, Xuehui
    Xie, Pengfei
    Chen, Xingyuan
    Yan, Bin
    FRONTIERS IN NEUROROBOTICS, 2023, 17
  • [48] Using deep neural networks to disentangle visual and semantic information in human perception and memory
    Shoham, Adva
    Grosbard, Idan Daniel
    Patashnik, Or
    Cohen-Or, Daniel
    Yovel, Galit
    NATURE HUMAN BEHAVIOUR, 2024, 8 (4) : 702 - 717
  • [49] Using deep neural networks to disentangle visual and semantic information in human perception and memory
    Adva Shoham
    Idan Daniel Grosbard
    Or Patashnik
    Daniel Cohen-Or
    Galit Yovel
    Nature Human Behaviour, 2024, 8 : 702 - 717
  • [50] A concealed poisoning attack to reduce deep neural networks' robustness against adversarial samples
    Zheng, Junhao
    Chan, Patrick P. K.
    Chi, Huiyang
    He, Zhimin
    INFORMATION SCIENCES, 2022, 615 : 758 - 773