A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability?

被引:230
|
作者
Huang, Xiaowei [1 ]
Kroening, Daniel [2 ]
Ruan, Wenjie [3 ]
Sharp, James [4 ]
Sun, Youcheng [5 ]
Thamo, Emese [1 ]
Wu, Min [2 ]
Yi, Xinping [1 ]
机构
[1] Univ Liverpool, Liverpool, Merseyside, England
[2] Univ Oxford, Oxford, England
[3] Univ Lancaster, Lancaster, England
[4] Def Sci & Technol Lab Dstl, Porton Down Salisbury, England
[5] Queens Univ Belfast, Belfast, Antrim, North Ireland
基金
英国工程与自然科学研究理事会;
关键词
ABSTRACTION-REFINEMENT; ROBUSTNESS; EXTRACTION;
D O I
10.1016/j.cosrev.2020.100270
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the past few years, significant progress has been made on deep neural networks (DNNs) in achieving human-level performance on several long-standing tasks. With the broader deployment of DNNs on various applications, the concerns over their safety and trustworthiness have been raised in public, especially after the widely reported fatal incidents involving self-driving cars. Research to address these concerns is particularly active, with a significant number of papers released in the past few years. This survey paper conducts a review of the current research effort into making DNNs safe and trustworthy, by focusing on four aspects: verification, testing, adversarial attack and defence, and interpretability. In total, we survey 202 papers, most of which were published after 2017. (c) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页数:35
相关论文
共 50 条
  • [41] Conformalized Adversarial Attack Detection for Graph Neural Networks
    Ennadir, Sofiane
    Alkhatib, Amr
    Bostrom, Henrik
    Vazirgiannis, Michalis
    CONFORMAL AND PROBABILISTIC PREDICTION WITH APPLICATIONS, VOL 204, 2023, 204 : 311 - 323
  • [42] Targeted Universal Adversarial Attack on Deep Hash Networks
    Meng, Fanlei
    Chen, Xiangru
    Cao, Yuan
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 165 - 174
  • [43] Trust Region Based Adversarial Attack on Neural Networks
    Yao, Zhewei
    Gholami, Amir
    Xu, Peng
    Keutzer, Kurt
    Mahoney, Michael W.
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 11342 - 11351
  • [44] Imperceptible Adversarial Attack via Invertible Neural Networks
    Chen, Zihan
    Wang, Ziyue
    Huang, Jun-Jie
    Zhao, Wentao
    Liu, Xiao
    Guan, Dejian
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 414 - 424
  • [45] DTFA: Adversarial attack with discrete cosine transform noise and target features on deep neural networks
    Yang, Dong
    Chen, Wei
    Wei, Songjie
    IET IMAGE PROCESSING, 2023, 17 (05) : 1464 - 1477
  • [46] A CMA-ES-Based Adversarial Attack on Black-Box Deep Neural Networks
    Kuang, Xiaohui
    Liu, Hongyi
    Wang, Ye
    Zhang, Qikun
    Zhang, Quanxin
    Zheng, Jun
    IEEE ACCESS, 2019, 7 : 172938 - 172947
  • [47] An adversarial attack detection method in deep neural networks based on re-attacking approach
    Ahmadi, Morteza Ali
    Dianat, Rouhollah
    Amirkhani, Hossein
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (07) : 10985 - 11014
  • [48] An adversarial attack detection method in deep neural networks based on re-attacking approach
    Morteza Ali Ahmadi
    Rouhollah Dianat
    Hossein Amirkhani
    Multimedia Tools and Applications, 2021, 80 : 10985 - 11014
  • [49] Frequency constraint-based adversarial attack on deep neural networks for medical image classification
    Chen, Fang
    Wang, Jian
    Liu, Han
    Kong, Wentao
    Zhao, Zhe
    Ma, Longfei
    Liao, Hongen
    Zhang, Daoqiang
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 164
  • [50] Attack on Deep Steganalysis Neural Networks
    Li, Shiyu
    Ye, Dengpan
    Jiang, Shunzhi
    Liu, Changrui
    Niu, Xiaoguang
    Luo, Xiangyang
    CLOUD COMPUTING AND SECURITY, PT IV, 2018, 11066 : 265 - 276