Interpretability of Deep Learning Models: A Survey of Results

被引:0
|
作者
Chakraborty, Supriyo [1 ]
Tomsett, Richard [4 ]
Raghavendra, Ramya [1 ]
Harborne, Daniel [2 ]
Alzantot, Moustafa [3 ]
Cerutti, Federico [2 ]
Srivastava, Mani [3 ]
Preece, Alun [2 ]
Julier, Simon [8 ]
Rao, Raghuveer M. [5 ]
Kelley, Troy D. [5 ]
Braines, Dave [4 ]
Sensoy, Murat [6 ]
Willis, Christopher J. [7 ]
Gurram, Prudhvi [5 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
[2] Cardiff Univ, Crime & Secur Res Inst, Cardiff, S Glam, Wales
[3] UCLA, Los Angeles, CA 90024 USA
[4] IBM United Kingdom Ltd, Portsmouth, Hants, England
[5] Army Res Lab, Adelphi, MD USA
[6] Ozyegin Univ, Istanbul, Turkey
[7] BAE Syst AI Labs, Great Baddow, England
[8] UCL, London, England
关键词
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep neural networks have achieved near-human accuracy levels in various types of classification and prediction tasks including images, text, speech, and video data. However, the networks continue to be treated mostly as black-box function approximators, mapping a given input to a classification output. The next step in this human-machine evolutionary process incorporating these networks into mission critical processes such as medical diagnosis, planning and control - requires a level of trust association with the machine output. Typically, statistical metrics are used to quantify the uncertainty of an output. However, the notion of trust also depends on the visibility that a human has into the working of the machine. In other words, the neural network should provide human-understandable justifications for its output leading to insights about the inner workings. We call such models as interpretable deep networks. Interpretability is not a monolithic notion. In fact, the subjectivity of an interpretation, due to different levels of human understanding, implies that there must be a multitude of dimensions that together constitute interpretability. In addition, the interpretation itself can be provided either in terms of the low-level network parameters, or in terms of input features used by the model. In this paper, we outline some of the dimensions that are useful for model interpretability, and categorize prior work along those dimensions. In the process, we perform a gap analysis of what needs to be done to improve model interpretability.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Visual interpretability for deep learning:a survey
    Quan-shi ZHANG
    Song-chun ZHU
    Frontiers of Information Technology & Electronic Engineering, 2018, 19 (01) : 27 - 39
  • [2] Visual interpretability for deep learning: a survey
    Zhang, Quan-shi
    Zhu, Song-chun
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2018, 19 (01) : 27 - 39
  • [3] Visual interpretability for deep learning: a survey
    Quan-shi Zhang
    Song-chun Zhu
    Frontiers of Information Technology & Electronic Engineering, 2018, 19 : 27 - 39
  • [4] Survey on Interpretability of Deep Models for Image Classification
    Yang P.-B.
    Sang J.-T.
    Zhang B.
    Feng Y.-G.
    Yu J.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (01): : 230 - 254
  • [5] A survey on the interpretability of deep learning in medical diagnosis
    Teng, Qiaoying
    Liu, Zhe
    Song, Yuqing
    Han, Kai
    Lu, Yang
    MULTIMEDIA SYSTEMS, 2022, 28 (06) : 2335 - 2355
  • [6] A survey on the interpretability of deep learning in medical diagnosis
    Qiaoying Teng
    Zhe Liu
    Yuqing Song
    Kai Han
    Yang Lu
    Multimedia Systems, 2022, 28 : 2335 - 2355
  • [7] Interpretability research of deep learning: A literature survey
    Xu, Biao
    Yang, Guanci
    INFORMATION FUSION, 2025, 115
  • [8] Visual interpretability of bioimaging deep learning models
    Rotem, Oded
    Zaritsky, Assaf
    NATURE METHODS, 2024, 21 (08) : 1394 - 1397
  • [9] Applications of interpretability in deep learning models for ophthalmology
    Hanif, Adam M.
    Beqiri, Sara
    Keane, Pearse A.
    Campbell, J. Peter
    CURRENT OPINION IN OPHTHALMOLOGY, 2021, 32 (05) : 452 - 458
  • [10] Interpretability of deep learning models for crop yield forecasting
    Paudel, Dilli
    de Wit, Allard
    Boogaard, Hendrik
    Marcos, Diego
    Osinga, Sjoukje
    Athanasiadis, Ioannis N.
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2023, 206