Interpreting Adversarial Examples in Deep Learning: A Review

被引:29
|
作者
Han, Sicong [1 ]
Lin, Chenhao [1 ]
Shen, Chao [1 ]
Wang, Qian [2 ]
Guan, Xiaohong [1 ]
机构
[1] Xi An Jiao Tong Univ, 28 Xianning West Rd, Xian 710049, Shaanxi, Peoples R China
[2] Wuhan Univ, 299 Bayi Rd, Wuhan 430072, Hubei, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Deep learning; adversarial example; interpretability; adversarial robustness;
D O I
10.1145/3594869
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep learning technology is increasingly being applied in safety-critical scenarios but has recently been found to be susceptible to imperceptible adversarial perturbations. This raises a serious concern regarding the adversarial robustness of deep neural network (DNN)-based applications. Accordingly, various adversarial attacks and defense approaches have been proposed. However, current studies implement different types of attacks and defenses with certain assumptions. There is still a lack of full theoretical understanding and interpretation of adversarial examples. Instead of reviewing technical progress in adversarial attacks and defenses, this article presents a framework consisting of three perspectives to discuss recent works focusing on theoretically explaining adversarial examples comprehensively. In each perspective, various hypotheses are further categorized and summarized into several subcategories and introduced systematically. To the best of our knowledge, this study is the first to concentrate on surveying existing research on adversarial examples and adversarial robustness from the interpretability perspective. By drawing on the reviewed literature, this survey characterizes current problems and challenges that need to be addressed and highlights potential future research directions to further investigate adversarial examples.
引用
收藏
页数:38
相关论文
共 50 条
  • [31] Natural Black-Box Adversarial Examples against Deep Reinforcement Learning
    Yu, Mengran
    Sun, Shiliang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8936 - 8944
  • [32] Adversarial Examples Against the Deep Learning Based Network Intrusion Detection Systems
    Yang, Kaichen
    Liu, Jianqing
    Zhang, Chi
    Fang, Yuguang
    2018 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM 2018), 2018, : 559 - 564
  • [33] Exploiting Adversarial Examples to Drain Computational Resources on Mobile Deep Learning Systems
    Gao, Han
    Tian, Yulong
    Yao, Rongchun
    Xu, Fengyuan
    Fu, Xinyi
    Zhong, Sheng
    2020 IEEE/ACM SYMPOSIUM ON EDGE COMPUTING (SEC 2020), 2020, : 334 - 339
  • [34] Delving deep into adversarial perturbations initialization on adversarial examples generation
    Hu, Cong
    Wan, Peng
    Wu, Xiao-Jun
    Yin, He-Feng
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [35] Adversarial Training Methods for Deep Learning: A Systematic Review
    Zhao, Weimin
    Alwidian, Sanaa
    Mahmoud, Qusay H.
    ALGORITHMS, 2022, 15 (08)
  • [36] Learning to Generate Textual Adversarial Examples
    Guo, Xiangzhe
    Tu, Shikui
    Xu, Lei
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT I, 2022, 13529 : 195 - 206
  • [37] "Adversarial Examples" for Proof-of-Learning
    Zhang, Rui
    Liu, Jian
    Ding, Yuan
    Wang, Zhibo
    Wu, Qingbiao
    Ren, Kui
    43RD IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2022), 2022, : 1408 - 1422
  • [38] Learning Indistinguishable and Transferable Adversarial Examples
    Zhang, Wu
    Zou, Junhua
    Duan, Yexin
    Zhou, Xingyu
    Pan, Zhisong
    PATTERN RECOGNITION AND COMPUTER VISION, PT IV, 2021, 13022 : 152 - 164
  • [39] Deep neural rejection against adversarial examples
    Angelo Sotgiu
    Ambra Demontis
    Marco Melis
    Battista Biggio
    Giorgio Fumera
    Xiaoyi Feng
    Fabio Roli
    EURASIP Journal on Information Security, 2020
  • [40] Robustness of deep neural networks in adversarial examples
    Song, Xiao (songxiao@buaa.edu.cn), 1600, University of Cincinnati (24):