A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability?

被引:230
|
作者
Huang, Xiaowei [1 ]
Kroening, Daniel [2 ]
Ruan, Wenjie [3 ]
Sharp, James [4 ]
Sun, Youcheng [5 ]
Thamo, Emese [1 ]
Wu, Min [2 ]
Yi, Xinping [1 ]
机构
[1] Univ Liverpool, Liverpool, Merseyside, England
[2] Univ Oxford, Oxford, England
[3] Univ Lancaster, Lancaster, England
[4] Def Sci & Technol Lab Dstl, Porton Down Salisbury, England
[5] Queens Univ Belfast, Belfast, Antrim, North Ireland
基金
英国工程与自然科学研究理事会;
关键词
ABSTRACTION-REFINEMENT; ROBUSTNESS; EXTRACTION;
D O I
10.1016/j.cosrev.2020.100270
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the past few years, significant progress has been made on deep neural networks (DNNs) in achieving human-level performance on several long-standing tasks. With the broader deployment of DNNs on various applications, the concerns over their safety and trustworthiness have been raised in public, especially after the widely reported fatal incidents involving self-driving cars. Research to address these concerns is particularly active, with a significant number of papers released in the past few years. This survey paper conducts a review of the current research effort into making DNNs safe and trustworthy, by focusing on four aspects: verification, testing, adversarial attack and defence, and interpretability. In total, we survey 202 papers, most of which were published after 2017. (c) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页数:35
相关论文
共 50 条
  • [21] Invisible Adversarial Attack against Deep Neural Networks: An Adaptive Penalization Approach
    Wang, Zhibo
    Song, Mengkai
    Zheng, Siyan
    Zhang, Zhifei
    Song, Yang
    Wang, Qian
    IEEE Transactions on Dependable and Secure Computing, 2021, 18 (03): : 1474 - 1488
  • [22] AdvGuard: Fortifying Deep Neural Networks Against Optimized Adversarial Example Attack
    Kwon, Hyun
    Lee, Jun
    IEEE ACCESS, 2024, 12 : 5345 - 5356
  • [23] New Perspective of Interpretability of Deep Neural Networks
    Kimura, Masanari
    Tanaka, Masayuki
    2020 3RD INTERNATIONAL CONFERENCE ON INFORMATION AND COMPUTER TECHNOLOGIES (ICICT 2020), 2020, : 78 - 85
  • [24] A Benchmark for Interpretability Methods in Deep Neural Networks
    Hooker, Sara
    Erhan, Dumitru
    Kindermans, Pieter-Jan
    Kim, Been
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [25] Generation of Adversarial Mechanisms in Deep Neural Networks: A Survey of the State of the Art
    Pavate, Aruna Animish
    Bansode, Rajesh
    International Journal of Ambient Computing and Intelligence, 2022, 13 (01)
  • [26] Analyze textual data: deep neural network for adversarial inversion attack in wireless networks
    Mohammed A. Al Ghamdi
    SN Applied Sciences, 2023, 5
  • [27] Blind Data Adversarial Bit-flip Attack against Deep Neural Networks
    Ghavami, Behnam
    Sadati, Mani
    Shahidzadeh, Mohammad
    Fang, Zhenman
    Shannon, Lesley
    2022 25TH EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD), 2022, : 899 - 904
  • [28] ADVERSPARSE: AN ADVERSARIAL ATTACK FRAMEWORK FOR DEEP SPATIAL-TEMPORAL GRAPH NEURAL NETWORKS
    Li, Jiayu
    Zhang, Tianyun
    Jin, Shengmin
    Fardad, Makan
    Zafarani, Reza
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 5857 - 5861
  • [29] Black-box Adversarial Attack against Visual Interpreters for Deep Neural Networks
    Hirose, Yudai
    Ono, Satoshi
    2023 18TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, MVA, 2023,
  • [30] A concealed poisoning attack to reduce deep neural networks' robustness against adversarial samples
    Zheng, Junhao
    Chan, Patrick P. K.
    Chi, Huiyang
    He, Zhimin
    INFORMATION SCIENCES, 2022, 615 : 758 - 773