A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability?

被引:230
|
作者
Huang, Xiaowei [1 ]
Kroening, Daniel [2 ]
Ruan, Wenjie [3 ]
Sharp, James [4 ]
Sun, Youcheng [5 ]
Thamo, Emese [1 ]
Wu, Min [2 ]
Yi, Xinping [1 ]
机构
[1] Univ Liverpool, Liverpool, Merseyside, England
[2] Univ Oxford, Oxford, England
[3] Univ Lancaster, Lancaster, England
[4] Def Sci & Technol Lab Dstl, Porton Down Salisbury, England
[5] Queens Univ Belfast, Belfast, Antrim, North Ireland
基金
英国工程与自然科学研究理事会;
关键词
ABSTRACTION-REFINEMENT; ROBUSTNESS; EXTRACTION;
D O I
10.1016/j.cosrev.2020.100270
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the past few years, significant progress has been made on deep neural networks (DNNs) in achieving human-level performance on several long-standing tasks. With the broader deployment of DNNs on various applications, the concerns over their safety and trustworthiness have been raised in public, especially after the widely reported fatal incidents involving self-driving cars. Research to address these concerns is particularly active, with a significant number of papers released in the past few years. This survey paper conducts a review of the current research effort into making DNNs safe and trustworthy, by focusing on four aspects: verification, testing, adversarial attack and defence, and interpretability. In total, we survey 202 papers, most of which were published after 2017. (c) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页数:35
相关论文
共 50 条
  • [31] Dynamic Programming-Based White Box Adversarial Attack for Deep Neural Networks
    Aggarwal, Swati
    Mittal, Anshul
    Aggarwal, Sanchit
    Singh, Anshul Kumar
    AI, 2024, 5 (03) : 1216 - 1234
  • [32] Analyze textual data: deep neural network for adversarial inversion attack in wireless networks
    Al Ghamdi, Mohammed A.
    SN APPLIED SCIENCES, 2023, 5 (12):
  • [33] Testing and Verification of the Deep Neural Networks Against Sparse Pixel Defects
    Szczepankiewicz, Michal
    Radlak, Krystian
    Szczepankiewicz, Karolina
    Popowicz, Adam
    Zawistowski, Pawel
    COMPUTER SAFETY, RELIABILITY, AND SECURITY, SAFECOMP 2022 WORKSHOPS, 2022, 13415 : 71 - 82
  • [34] Optimizing for interpretability in deep neural networks with tree regularization
    Wu M.
    Parbhoo S.
    Hughes M.C.
    Roth V.
    Doshi-Velez F.
    Journal of Artificial Intelligence Research, 2021, 72
  • [35] IMPROVING THE INTERPRETABILITY OF DEEP NEURAL NETWORKS WITH STIMULATED LEARNING
    Tan, Shawn
    Sim, Khe Chai
    Gales, Mark
    2015 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2015, : 617 - 623
  • [36] Improving Interpretability of Deep Neural Networks with Semantic Information
    Dong, Yinpeng
    Su, Hang
    Zhu, Jun
    Zhang, Bo
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 975 - 983
  • [37] Regional Tree Regularization for Interpretability in Deep Neural Networks
    Wu, Mike
    Parbhoo, Sonali
    Hughes, Michael C.
    Kindle, Ryan
    Celi, Leo
    Zazzi, Maurizio
    Roth, Volker
    Doshi-Velez, Finale
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 6413 - 6421
  • [38] Optimizing for Interpretability in Deep Neural Networks with Tree Regularization
    Wu, Mike
    Parbhoo, Sonali
    Hughes, Michael C.
    Roth, Volker
    Doshi-Velez, Finale
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2021, 72 : 1 - 37
  • [39] Improving the Interpretability of Deep Neural Networks with Knowledge Distillation
    Liu, Xuan
    Wang, Xiaoguang
    Matwin, Stan
    2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2018, : 905 - 912
  • [40] Generative Adversarial Networks: A Survey on Attack and Defense Perspective
    Zhang, Chenhan
    Yu, Shui
    Tian, Zhiyi
    Yu, James J. Q.
    ACM COMPUTING SURVEYS, 2024, 56 (04)