Interpretability of Deep Learning Models: A Survey of Results

被引:0
|
作者
Chakraborty, Supriyo [1 ]
Tomsett, Richard [4 ]
Raghavendra, Ramya [1 ]
Harborne, Daniel [2 ]
Alzantot, Moustafa [3 ]
Cerutti, Federico [2 ]
Srivastava, Mani [3 ]
Preece, Alun [2 ]
Julier, Simon [8 ]
Rao, Raghuveer M. [5 ]
Kelley, Troy D. [5 ]
Braines, Dave [4 ]
Sensoy, Murat [6 ]
Willis, Christopher J. [7 ]
Gurram, Prudhvi [5 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
[2] Cardiff Univ, Crime & Secur Res Inst, Cardiff, S Glam, Wales
[3] UCLA, Los Angeles, CA 90024 USA
[4] IBM United Kingdom Ltd, Portsmouth, Hants, England
[5] Army Res Lab, Adelphi, MD USA
[6] Ozyegin Univ, Istanbul, Turkey
[7] BAE Syst AI Labs, Great Baddow, England
[8] UCL, London, England
关键词
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep neural networks have achieved near-human accuracy levels in various types of classification and prediction tasks including images, text, speech, and video data. However, the networks continue to be treated mostly as black-box function approximators, mapping a given input to a classification output. The next step in this human-machine evolutionary process incorporating these networks into mission critical processes such as medical diagnosis, planning and control - requires a level of trust association with the machine output. Typically, statistical metrics are used to quantify the uncertainty of an output. However, the notion of trust also depends on the visibility that a human has into the working of the machine. In other words, the neural network should provide human-understandable justifications for its output leading to insights about the inner workings. We call such models as interpretable deep networks. Interpretability is not a monolithic notion. In fact, the subjectivity of an interpretation, due to different levels of human understanding, implies that there must be a multitude of dimensions that together constitute interpretability. In addition, the interpretation itself can be provided either in terms of the low-level network parameters, or in terms of input features used by the model. In this paper, we outline some of the dimensions that are useful for model interpretability, and categorize prior work along those dimensions. In the process, we perform a gap analysis of what needs to be done to improve model interpretability.
引用
收藏
页数:6
相关论文
共 50 条
  • [11] Interpretability of deep learning models in analysis of Spanish financial text
    César Vaca
    Manuel Astorgano
    Alfonso J. López-Rivero
    Fernando Tejerina
    Benjamín Sahelices
    Neural Computing and Applications, 2024, 36 : 7509 - 7527
  • [12] Interpretability of deep learning models in analysis of Spanish financial text
    Vaca, Cesar
    Astorgano, Manuel
    Lopez-Rivero, Alfonso J.
    Tejerina, Fernando
    Sahelices, Benjamin
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (13): : 7509 - 7527
  • [13] Improvement of Deep Learning Models by Excluding Inappropriate Data Based on Interpretability
    Yamaguchi, Saneyasu
    Hirabayashi, Fuma
    Tamekuri, Atsuki
    2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024, 2024, : 291 - 296
  • [14] Unveiling Interpretability: Analyzing Transfer Learning in Deep Learning Models for Traffic Sign Recognition
    Waziry S.
    Rasheed J.
    Ghabban F.M.
    Alsubai S.
    Elkiran H.
    Alqahtani A.
    SN Computer Science, 5 (6)
  • [15] Interpretability and fairness evaluation of deep learning models on MIMIC-IV dataset
    Meng, Chuizheng
    Trinh, Loc
    Xu, Nan
    Enouen, James
    Liu, Yan
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [16] Interpretability and fairness evaluation of deep learning models on MIMIC-IV dataset
    Chuizheng Meng
    Loc Trinh
    Nan Xu
    James Enouen
    Yan Liu
    Scientific Reports, 12
  • [17] INFORMER- Interpretability Founded Monitoring of Medical Image Deep Learning Models
    Shu, Shelley Zixin
    de Mortanges, Aurelie Pahud
    Poellinger, Alexander
    Mahapatra, Dwarikanath
    Reyes, Mauricio
    UNCERTAINTY FOR SAFE UTILIZATION OF MACHINE LEARNING IN MEDICAL IMAGING, UNSURE 2024, 2025, 15167 : 215 - 224
  • [18] TorchEsegeta: Framework for Interpretability and Explainability of Image-Based Deep Learning Models
    Chatterjee, Soumick
    Das, Arnab
    Mandal, Chirag
    Mukhopadhyay, Budhaditya
    Vipinraj, Manish
    Shukla, Aniruddh
    Rao, Rajatha Nagaraja
    Sarasaen, Chompunuch
    Speck, Oliver
    Nuernberger, Andreas
    APPLIED SCIENCES-BASEL, 2022, 12 (04):
  • [19] Research Advances in the Interpretability of Deep Learning
    Cheng K.
    Wang N.
    Shi W.
    Zhan Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2020, 57 (06): : 1208 - 1217
  • [20] Improving Interpretability and Regularization in Deep Learning
    Wu, Chunyang
    Gales, Mark J. F.
    Ragni, Anton
    Karanasou, Penny
    Sim, Khe Chai
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2018, 26 (02) : 256 - 265