Explaining Neural Networks Using Attentive Knowledge Distillation

被引:4
|
作者
Lee, Hyeonseok [1 ]
Kim, Sungchan [1 ,2 ]
机构
[1] Jeonbuk Natl Univ, Div Comp Sci & Engn, Jeonju Si 54896, Jeollabuk Do, South Korea
[2] Jeonbuk Natl Univ, Res Ctr Artificial Intelligence Technol, Jeonju Si 54896, Jeollabuk Do, South Korea
基金
新加坡国家研究基金会;
关键词
deep neural networks; visual explanation; attention; knowledge distillation; fine-grained classification;
D O I
10.3390/s21041280
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Explaining the prediction of deep neural networks makes the networks more understandable and trusted, leading to their use in various mission critical tasks. Recent progress in the learning capability of networks has primarily been due to the enormous number of model parameters, so that it is usually hard to interpret their operations, as opposed to classical white-box models. For this purpose, generating saliency maps is a popular approach to identify the important input features used for the model prediction. Existing explanation methods typically only use the output of the last convolution layer of the model to generate a saliency map, lacking the information included in intermediate layers. Thus, the corresponding explanations are coarse and result in limited accuracy. Although the accuracy can be improved by iteratively developing a saliency map, this is too time-consuming and is thus impractical. To address these problems, we proposed a novel approach to explain the model prediction by developing an attentive surrogate network using the knowledge distillation. The surrogate network aims to generate a fine-grained saliency map corresponding to the model prediction using meaningful regional information presented over all network layers. Experiments demonstrated that the saliency maps are the result of spatially attentive features learned from the distillation. Thus, they are useful for fine-grained classification tasks. Moreover, the proposed method runs at the rate of 24.3 frames per second, which is much faster than the existing methods by orders of magnitude.
引用
收藏
页码:1 / 17
页数:17
相关论文
共 50 条
  • [21] Automatic Modulation Classification with Neural Networks via Knowledge Distillation
    Wang, Shuai
    Liu, Chunwu
    ELECTRONICS, 2022, 11 (19)
  • [22] Soft Hybrid Knowledge Distillation against deep neural networks
    Zhang, Jian
    Tao, Ze
    Zhang, Shichao
    Qiao, Zike
    Guo, Kehua
    NEUROCOMPUTING, 2024, 570
  • [23] Accelerating Molecular Graph Neural Networks via Knowledge Distillation
    Kelvinius, Filip Ekstrom
    Georgiev, Dimitar
    Toshev, Artur Petrov
    Gasteiger, Johannes
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36, NEURIPS 2023, 2023,
  • [24] Data-free knowledge distillation in neural networks for regression
    Kang, Myeonginn
    Kang, Seokho
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 175
  • [25] Self-architectural knowledge distillation for spiking neural networks
    Qiu, Haonan
    Ning, Munan
    Song, Zeyin
    Fang, Wei
    Chen, Yanqi
    Sun, Tao
    Ma, Zhengyu
    Yuan, Li
    Tian, Yonghong
    NEURAL NETWORKS, 2024, 178
  • [26] KD-FIXMATCH: KNOWLEDGE DISTILLATION SIAMESE NEURAL NETWORKS
    Wang, Chien-Chih
    Xu, Shaoyuan
    Fu, Jinmiao
    Liu, Yang
    Wang, Bryan
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 341 - 345
  • [27] Emulating quantum dynamics with neural networks via knowledge distillation
    Yao, Yu
    Cao, Chao
    Haas, Stephan
    Agarwal, Mahak
    Khanna, Divyam
    Abram, Marcin
    FRONTIERS IN MATERIALS, 2023, 9
  • [28] Knowledge Distillation with Graph Neural Networks for Epileptic Seizure Detection
    Zheng, Qinyue
    Venkitaraman, Arun
    Petravic, Simona
    Frossard, Pascal
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: APPLIED DATA SCIENCE AND DEMO TRACK, ECML PKDD 2023, PT VI, 2023, 14174 : 547 - 563
  • [29] Knowledge distillation circumvents nonlinearity for optical convolutional neural networks
    Xiang, Jinlin
    Colburn, Shane
    Majumdar, Arka
    Shlizerman, Eli
    APPLIED OPTICS, 2022, 61 (09) : 2173 - 2183
  • [30] Graph-Free Knowledge Distillation for Graph Neural Networks
    Deng, Xiang
    Zhang, Zhongfei
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2321 - 2327