Interpretability of Deep Learning Models: A Survey of Results

被引:0
|
作者
Chakraborty, Supriyo [1 ]
Tomsett, Richard [4 ]
Raghavendra, Ramya [1 ]
Harborne, Daniel [2 ]
Alzantot, Moustafa [3 ]
Cerutti, Federico [2 ]
Srivastava, Mani [3 ]
Preece, Alun [2 ]
Julier, Simon [8 ]
Rao, Raghuveer M. [5 ]
Kelley, Troy D. [5 ]
Braines, Dave [4 ]
Sensoy, Murat [6 ]
Willis, Christopher J. [7 ]
Gurram, Prudhvi [5 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
[2] Cardiff Univ, Crime & Secur Res Inst, Cardiff, S Glam, Wales
[3] UCLA, Los Angeles, CA 90024 USA
[4] IBM United Kingdom Ltd, Portsmouth, Hants, England
[5] Army Res Lab, Adelphi, MD USA
[6] Ozyegin Univ, Istanbul, Turkey
[7] BAE Syst AI Labs, Great Baddow, England
[8] UCL, London, England
关键词
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep neural networks have achieved near-human accuracy levels in various types of classification and prediction tasks including images, text, speech, and video data. However, the networks continue to be treated mostly as black-box function approximators, mapping a given input to a classification output. The next step in this human-machine evolutionary process incorporating these networks into mission critical processes such as medical diagnosis, planning and control - requires a level of trust association with the machine output. Typically, statistical metrics are used to quantify the uncertainty of an output. However, the notion of trust also depends on the visibility that a human has into the working of the machine. In other words, the neural network should provide human-understandable justifications for its output leading to insights about the inner workings. We call such models as interpretable deep networks. Interpretability is not a monolithic notion. In fact, the subjectivity of an interpretation, due to different levels of human understanding, implies that there must be a multitude of dimensions that together constitute interpretability. In addition, the interpretation itself can be provided either in terms of the low-level network parameters, or in terms of input features used by the model. In this paper, we outline some of the dimensions that are useful for model interpretability, and categorize prior work along those dimensions. In the process, we perform a gap analysis of what needs to be done to improve model interpretability.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Multicriteria interpretability driven deep learning
    Repetto, Marco
    ANNALS OF OPERATIONS RESEARCH, 2022, 346 (2) : 1621 - 1635
  • [22] Reliability and Interpretability in Science and Deep Learning
    Scorzato, Luigi
    MINDS AND MACHINES, 2024, 34 (03)
  • [23] The survey: Text generation models in deep learning
    Iqbal, Touseef
    Qureshi, Shaima
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (06) : 2515 - 2528
  • [24] Compression of Deep Learning Models for Text: A Survey
    Gupta, Manish
    Agrawal, Puneet
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2022, 16 (04)
  • [25] Robustness of deep learning models on graphs: A survey
    Xu, Jiarong
    Chen, Junru
    You, Siqi
    Xiao, Zhiqing
    Yang, Yang
    Lu, Jiangang
    AI OPEN, 2021, 2 : 69 - 78
  • [26] A Survey of Deep Active Learning for Foundation Models
    Wan, Tianjiao
    Xu, Kele
    Yu, Ting
    Wang, Xu
    Feng, Dawei
    Ding, Bo
    Wang, Huaimin
    Intelligent Computing, 2023, 2
  • [27] Deep learning in citation recommendation models survey
    Ali, Zafar
    Kefalas, Pavlos
    Muhammad, Khan
    Ali, Bahadar
    Imran, Muhammad
    EXPERT SYSTEMS WITH APPLICATIONS, 2020, 162
  • [28] A Survey of Interpretability Research Methods for Reinforcement Learning
    Cao, Hong-Ye
    Liu, Xiao
    Dong, Shao-Kang
    Yang, Shang-Dong
    Huo, Jing
    Li, Wen-Bin
    Gao, Yang
    Jisuanji Xuebao/Chinese Journal of Computers, 2024, 47 (08): : 1853 - 1882
  • [29] Machine Learning Interpretability: A Survey on Methods and Metrics
    Carvalho, Diogo, V
    Pereira, Eduardo M.
    Cardoso, Jaime S.
    ELECTRONICS, 2019, 8 (08)
  • [30] Efficient Deep Learning: A Survey on Making Deep Learning Models Smaller, Faster, and Better
    Menghani, Gaurav
    ACM COMPUTING SURVEYS, 2023, 55 (12)