Explaining Neural Networks Using Attentive Knowledge Distillation

被引:4
|
作者
Lee, Hyeonseok [1 ]
Kim, Sungchan [1 ,2 ]
机构
[1] Jeonbuk Natl Univ, Div Comp Sci & Engn, Jeonju Si 54896, Jeollabuk Do, South Korea
[2] Jeonbuk Natl Univ, Res Ctr Artificial Intelligence Technol, Jeonju Si 54896, Jeollabuk Do, South Korea
基金
新加坡国家研究基金会;
关键词
deep neural networks; visual explanation; attention; knowledge distillation; fine-grained classification;
D O I
10.3390/s21041280
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Explaining the prediction of deep neural networks makes the networks more understandable and trusted, leading to their use in various mission critical tasks. Recent progress in the learning capability of networks has primarily been due to the enormous number of model parameters, so that it is usually hard to interpret their operations, as opposed to classical white-box models. For this purpose, generating saliency maps is a popular approach to identify the important input features used for the model prediction. Existing explanation methods typically only use the output of the last convolution layer of the model to generate a saliency map, lacking the information included in intermediate layers. Thus, the corresponding explanations are coarse and result in limited accuracy. Although the accuracy can be improved by iteratively developing a saliency map, this is too time-consuming and is thus impractical. To address these problems, we proposed a novel approach to explain the model prediction by developing an attentive surrogate network using the knowledge distillation. The surrogate network aims to generate a fine-grained saliency map corresponding to the model prediction using meaningful regional information presented over all network layers. Experiments demonstrated that the saliency maps are the result of spatially attentive features learned from the distillation. Thus, they are useful for fine-grained classification tasks. Moreover, the proposed method runs at the rate of 24.3 frames per second, which is much faster than the existing methods by orders of magnitude.
引用
收藏
页码:1 / 17
页数:17
相关论文
共 50 条
  • [41] Quality of Life Prediction on Walking Scenes Using Deep Neural Networks and Performance Improvement Using Knowledge Distillation
    Rithanasophon, Thanasit
    Thitisiriwech, Kitsaphon
    Kantavat, Pittipol
    Kijsirikul, Boonserm
    Iwahori, Yuji
    Fukui, Shinji
    Nakamura, Kazuki
    Hayashi, Yoshitsugu
    ELECTRONICS, 2023, 12 (13)
  • [42] Knowledge Reverse Distillation Based Confidence Calibration for Deep Neural Networks
    Jiang, Xianhui
    Deng, Xiaogang
    NEURAL PROCESSING LETTERS, 2023, 55 (01) : 345 - 360
  • [43] FreeKD: Free-direction Knowledge Distillation for Graph Neural Networks
    Feng, Kaituo
    Li, Changsheng
    Yuan, Ye
    Wang, Guoren
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 357 - 366
  • [44] Homogeneous teacher based buffer knowledge distillation for tiny neural networks
    Dai, Xinru
    Lu, Gang
    Shen, Jianhua
    Huang, Shuo
    Wei, Tongquan
    JOURNAL OF SYSTEMS ARCHITECTURE, 2024, 148
  • [45] Knowledge Distillation Improves Graph Structure Augmentation for Graph Neural Networks
    Wu, Lirong
    Lin, Haitao
    Huang, Yufei
    Li, Stan Z.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [46] Knowledge distillation with ensembles of convolutional neural networks for medical image segmentation
    Noothout, Julia M. H.
    Lessmann, Nikolas
    van Eede, Matthijs C.
    van Harten, Louis D.
    Sogancioglu, Ecem
    Heslinga, Friso G.
    Veta, Mitko
    van Ginneken, Bram
    Isgum, Ivana
    JOURNAL OF MEDICAL IMAGING, 2022, 9 (05)
  • [47] Knowledge Reverse Distillation Based Confidence Calibration for Deep Neural Networks
    Xianhui Jiang
    Xiaogang Deng
    Neural Processing Letters, 2023, 55 : 345 - 360
  • [48] Compressing Deep Graph Neural Networks via Adversarial Knowledge Distillation
    He, Huarui
    Wang, Jie
    Zhang, Zhanqiu
    Wu, Feng
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 534 - 544
  • [49] EGNN: Constructing explainable graph neural networks via knowledge distillation
    Li, Yuan
    Liu, Li
    Wang, Guoyin
    Du, Yong
    Chen, Penggang
    KNOWLEDGE-BASED SYSTEMS, 2022, 241
  • [50] Feature Distribution-based Knowledge Distillation for Deep Neural Networks
    Hong, Hyeonseok
    Kim, Hyun
    2022 19TH INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2022, : 75 - 76