The Adversarial Attack and Detection under the Fisher Information Metric

被引:0
|
作者
Zhao, Chenxiao [1 ]
Fletcher, P. Thomas [2 ,3 ]
Yu, Mixue [1 ]
Peng, Yaxin [4 ,5 ]
Zhang, Guixu [1 ]
Shen, Chaomin [1 ,5 ]
机构
[1] East China Normal Univ, Dept Comp Sci, Shanghai, Peoples R China
[2] Univ Virginia, Dept Elect & Comp Sci, Charlottesville, VA 22903 USA
[3] Univ Virginia, Dept Comp Sci, Charlottesville, VA 22903 USA
[4] Shanghai Univ, Dept Math, Shanghai, Peoples R China
[5] Westlake Inst Brain Like Sci & Technol, Hangzhou, Zhejiang, Peoples R China
来源
THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE | 2019年
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many deep learning models are vulnerable to the adversarial attack, i.e., imperceptible but intentionally-designed perturbations to the input can cause incorrect output of the networks. In this paper, using information geometry, we provide a reasonable explanation for the vulnerability of deep learning models. By considering the data space as a non-linear space with the Fisher information metric induced from a neural network, we first propose an adversarial attack algorithm termed one-step spectral attack (OSSA). The method is described by a constrained quadratic form of the Fisher information matrix, where the optimal adversarial perturbation is given by the first eigenvector, and the vulnerability is reflected by the eigenvalues. The larger an eigenvalue is, the more vulnerable the model is to be attacked by the corresponding eigenvector. Taking advantage of the property, we also propose an adversarial detection method with the eigenvalues serving as characteristics. Both our attack and detection algorithms are numerically optimized to work efficiently on large datasets. Our evaluations show superior performance compared with other methods, implying that the Fisher information is a promising approach to investigate the adversarial attacks and defenses.
引用
收藏
页码:5869 / 5876
页数:8
相关论文
共 50 条
  • [41] Detection of Iterative Adversarial Attacks via Counter Attack
    Rottmann, Matthias
    Maag, Kira
    Peyron, Mathis
    Gottschalk, Hanno
    Krejic, Natasa
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2023, 198 (03) : 892 - 929
  • [42] EnsembleDet: ensembling against adversarial attack on deepfake detection
    Dutta, Himanshu
    Pandey, Aditya
    Bilgaiyan, Saurabh
    JOURNAL OF ELECTRONIC IMAGING, 2021, 30 (06)
  • [43] Adversarial attack for object detectors under complex conditions
    Qin, Yingxin
    Zhang, Kejia
    Pan, Haiwei
    COMPUTERS & SECURITY, 2023, 134
  • [44] Improved Generative Adversarial Network for Phishing Attack Detection
    Shammi, L.
    Shyni, Dr C. Emilin
    2024 4TH INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND SOCIAL NETWORKING, ICPCSN 2024, 2024, : 11 - 17
  • [45] MAP: MULTISPECTRAL ADVERSARIAL PATCH TO ATTACK PERSON DETECTION
    Kim, Taeheon
    Lee, Hong Joo
    Ro, Yong Man
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4853 - 4857
  • [46] Malicious Attack Detection in IoT by Generative Adversarial Networks
    Srikanth Bethu
    SN Computer Science, 6 (4)
  • [47] An Enhanced Transferable Adversarial Attack Against Object Detection
    Shi, Guoqiang
    Lin, Zhi
    Peng, Anjie
    Zeng, Hui
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [48] Detection of Iterative Adversarial Attacks via Counter Attack
    Matthias Rottmann
    Kira Maag
    Mathis Peyron
    Hanno Gottschalk
    Nataša Krejić
    Journal of Optimization Theory and Applications, 2023, 198 : 892 - 929
  • [49] DIPA: Adversarial Attack on DNNs by Dropping Information and Pixel-Level Attack on Attention
    Liu, Jing
    Liu, Huailin
    Wang, Pengju
    Wu, Yang
    Li, Keqin
    INFORMATION, 2024, 15 (07)
  • [50] Decreasing adversarial transferability using gradient information of attack paths
    Xu, Mengjun
    Liu, Lei
    Xia, Pengfei
    Li, Ziqiang
    Li, Bin
    APPLIED SOFT COMPUTING, 2025, 170