Interpretability of Deep Learning Models: A Survey of Results

被引:0
|
作者
Chakraborty, Supriyo [1 ]
Tomsett, Richard [4 ]
Raghavendra, Ramya [1 ]
Harborne, Daniel [2 ]
Alzantot, Moustafa [3 ]
Cerutti, Federico [2 ]
Srivastava, Mani [3 ]
Preece, Alun [2 ]
Julier, Simon [8 ]
Rao, Raghuveer M. [5 ]
Kelley, Troy D. [5 ]
Braines, Dave [4 ]
Sensoy, Murat [6 ]
Willis, Christopher J. [7 ]
Gurram, Prudhvi [5 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
[2] Cardiff Univ, Crime & Secur Res Inst, Cardiff, S Glam, Wales
[3] UCLA, Los Angeles, CA 90024 USA
[4] IBM United Kingdom Ltd, Portsmouth, Hants, England
[5] Army Res Lab, Adelphi, MD USA
[6] Ozyegin Univ, Istanbul, Turkey
[7] BAE Syst AI Labs, Great Baddow, England
[8] UCL, London, England
关键词
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep neural networks have achieved near-human accuracy levels in various types of classification and prediction tasks including images, text, speech, and video data. However, the networks continue to be treated mostly as black-box function approximators, mapping a given input to a classification output. The next step in this human-machine evolutionary process incorporating these networks into mission critical processes such as medical diagnosis, planning and control - requires a level of trust association with the machine output. Typically, statistical metrics are used to quantify the uncertainty of an output. However, the notion of trust also depends on the visibility that a human has into the working of the machine. In other words, the neural network should provide human-understandable justifications for its output leading to insights about the inner workings. We call such models as interpretable deep networks. Interpretability is not a monolithic notion. In fact, the subjectivity of an interpretation, due to different levels of human understanding, implies that there must be a multitude of dimensions that together constitute interpretability. In addition, the interpretation itself can be provided either in terms of the low-level network parameters, or in terms of input features used by the model. In this paper, we outline some of the dimensions that are useful for model interpretability, and categorize prior work along those dimensions. In the process, we perform a gap analysis of what needs to be done to improve model interpretability.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] A comprehensive survey on optimizing deep learning models by metaheuristics
    Akay, Bahriye
    Karaboga, Dervis
    Akay, Rustu
    ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (02) : 829 - 894
  • [42] Backdoor Attacks to Deep Learning Models and Countermeasures: A Survey
    Li, Yudong
    Zhang, Shigeng
    Wang, Weiping
    Song, Hong
    IEEE OPEN JOURNAL OF THE COMPUTER SOCIETY, 2023, 4 : 134 - 146
  • [43] A Survey of Deep Learning Models for Medical Image Analysis
    Umer, Mohammad
    Sharma, Shilpa
    Rattan, Punam
    2021 INTERNATIONAL CONFERENCE ON COMPUTING SCIENCES (ICCS 2021), 2021, : 65 - 69
  • [44] A survey of deep learning models in medical therapeutic areas
    Nogales, Alberto
    Garcia-Tejedor, Alvaro J.
    Monge, Diana
    Serrano Vara, Juan
    Anton, Cristina
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2021, 112
  • [45] Deep learning with the generative models for recommender systems: A survey
    Nahta, Ravi
    Chauhan, Ganpat Singh
    Meena, Yogesh Kumar
    Gopalani, Dinesh
    COMPUTER SCIENCE REVIEW, 2024, 53
  • [46] Beyond Sparsity: Tree Regularization of Deep Models for Interpretability
    Wu, Mike
    Hughes, Michael C.
    Parbhoo, Sonali
    Zazzi, Maurizio
    Roth, Volker
    Doshi-Velez, Finale
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 1670 - 1678
  • [47] Bandit Interpretability of Deep Models via Confidence Selection
    Duan, Xiaoyue
    Li, Hong
    Wang, Panpan
    Wang, Tiancheng
    Liu, Boyu
    Zhang, Baochang
    NEUROCOMPUTING, 2023, 544
  • [48] Survey on Deep Fuzzy Systems in Regression Applications: A View on Interpretability
    S. S. Junior, Jorge
    Mendes, Jerome
    Souza, Francisco
    Premebida, Cristiano
    INTERNATIONAL JOURNAL OF FUZZY SYSTEMS, 2023, 25 (07) : 2568 - 2589
  • [49] Deep learning interpretability analysis methods in image interpretation
    Gong J.
    Huan L.
    Zheng X.
    Cehui Xuebao/Acta Geodaetica et Cartographica Sinica, 2022, 51 (06): : 873 - 884
  • [50] Decoupling Deep Learning for Enhanced Image Recognition Interpretability
    Peng, Yitao
    Liu, Yihang
    Yang, Longzhen
    Shang, Shaohua
    He, Lianghua
    Hu, Die
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (10)