Interpretability of Deep Learning Models: A Survey of Results

被引:0
|
作者
Chakraborty, Supriyo [1 ]
Tomsett, Richard [4 ]
Raghavendra, Ramya [1 ]
Harborne, Daniel [2 ]
Alzantot, Moustafa [3 ]
Cerutti, Federico [2 ]
Srivastava, Mani [3 ]
Preece, Alun [2 ]
Julier, Simon [8 ]
Rao, Raghuveer M. [5 ]
Kelley, Troy D. [5 ]
Braines, Dave [4 ]
Sensoy, Murat [6 ]
Willis, Christopher J. [7 ]
Gurram, Prudhvi [5 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
[2] Cardiff Univ, Crime & Secur Res Inst, Cardiff, S Glam, Wales
[3] UCLA, Los Angeles, CA 90024 USA
[4] IBM United Kingdom Ltd, Portsmouth, Hants, England
[5] Army Res Lab, Adelphi, MD USA
[6] Ozyegin Univ, Istanbul, Turkey
[7] BAE Syst AI Labs, Great Baddow, England
[8] UCL, London, England
关键词
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep neural networks have achieved near-human accuracy levels in various types of classification and prediction tasks including images, text, speech, and video data. However, the networks continue to be treated mostly as black-box function approximators, mapping a given input to a classification output. The next step in this human-machine evolutionary process incorporating these networks into mission critical processes such as medical diagnosis, planning and control - requires a level of trust association with the machine output. Typically, statistical metrics are used to quantify the uncertainty of an output. However, the notion of trust also depends on the visibility that a human has into the working of the machine. In other words, the neural network should provide human-understandable justifications for its output leading to insights about the inner workings. We call such models as interpretable deep networks. Interpretability is not a monolithic notion. In fact, the subjectivity of an interpretation, due to different levels of human understanding, implies that there must be a multitude of dimensions that together constitute interpretability. In addition, the interpretation itself can be provided either in terms of the low-level network parameters, or in terms of input features used by the model. In this paper, we outline some of the dimensions that are useful for model interpretability, and categorize prior work along those dimensions. In the process, we perform a gap analysis of what needs to be done to improve model interpretability.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Distinguishing between Crohn's disease and ulcerative colitis using deep learning models with interpretability
    Mauricio, Jose
    Domingues, Ines
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (01)
  • [32] Optimizing vitiligo diagnosis with ResNet and Swin transformer deep learning models: a study on performance and interpretability
    Zhong, Fan
    He, Kaiqiao
    Ji, Mengqi
    Chen, Jianru
    Gao, Tianwen
    Li, Shuli
    Zhang, Junpeng
    Li, Chunying
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [33] Explainability and uncertainty: Two sides of the same coin for enhancing the interpretability of deep learning models in healthcare☆
    Salvi, Massimo
    Seoni, Silvia
    Campagner, Andrea
    Gertych, Arkadiusz
    Acharya, U. Rajendra
    Molinari, Filippo
    Cabitza, Federico
    INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 2025, 197
  • [34] Distinguishing between Crohn’s disease and ulcerative colitis using deep learning models with interpretability
    José Maurício
    Inês Domingues
    Pattern Analysis and Applications, 2024, 27
  • [35] Robustness Certification Research on Deep Learning Models: A Survey
    Ji S.-L.
    Du T.-Y.
    Deng S.-G.
    Cheng P.
    Shi J.
    Yang M.
    Li B.
    Jisuanji Xuebao/Chinese Journal of Computers, 2022, 45 (01): : 190 - 206
  • [36] A comprehensive survey on optimizing deep learning models by metaheuristics
    Bahriye Akay
    Dervis Karaboga
    Rustu Akay
    Artificial Intelligence Review, 2022, 55 : 829 - 894
  • [37] Ensemble Multifeatured Deep Learning Models and Applications: A Survey
    Abimannan, Satheesh
    El-Alfy, El-Sayed M.
    Chang, Yue-Shan
    Hussain, Shahid
    Shukla, Saurabh
    Satheesh, Dhivyadharsini
    IEEE ACCESS, 2023, 11 : 107194 - 107217
  • [38] A survey of deep learning models in medical therapeutic areas
    Nogales, Alberto
    Garcia-Tejedor, Alvaro J.
    Monge, Diana
    Serrano Vara, Juan
    Anton, Cristina
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2021, 112
  • [39] A survey of deep learning models in medical therapeutic areas
    Nogales, Alberto
    García-Tejedor, Álvaro J.
    Monge, Diana
    Vara, Juan Serrano
    Antón, Cristina
    Nogales, Alberto (alberto.nogales@ceiec.es), 1600, Elsevier B.V. (112):
  • [40] Survey on Videos Data Augmentation for Deep Learning Models
    Cauli, Nino
    Recupero, Diego Reforgiato
    FUTURE INTERNET, 2022, 14 (03)