BR-NPA: A non-parametric high-resolution attention model to improve the interpretability of attention

被引:0
|
作者
Gomez T. [1 ,3 ]
Ling S. [1 ,3 ]
Fréour T. [2 ,4 ,5 ,6 ]
Mouchère H. [1 ,3 ]
机构
[1] Christian Pauc Street, Nantes
[2] 63 Magellan Quai, Nantes
[3] Nantes University, CNRS, LS2N, CNRS UMR 6004, Nantes
[4] Nantes University Hospital, Department of Reproductive Medicine and Biology, Nantes
[5] Nantes University, Nantes University Hospital, Inserm, CRTI, Inserm UMR 1064, Nantes
[6] Nantes University, Nantes University Hospital, Inserm, CNRS, SFR Santé, Inserm UMS 016, CNRS UMS, 3556, Nantes
关键词
Deep learning; Interpretability; Non-parametric; Resolution; Spatial attention;
D O I
10.1016/j.patcog.2022.108927
中图分类号
TB18 [人体工程学]; Q98 [人类学];
学科分类号
030303 ; 1201 ;
摘要
The prevalence of employing attention mechanisms has brought along concerns about the interpretability of attention distributions. Although it provides insights into how a model is operating, utilizing attention as the explanation of model predictions is still highly dubious. The community is still seeking more interpretable strategies for better identifying local active regions that contribute the most to the final decision. To improve the interpretability of existing attention models, we propose a novel Bilinear Representative Non-Parametric Attention (BR-NPA) strategy that captures the task-relevant human-interpretable information. The target model is first distilled to have higher-resolution intermediate feature maps. From which, representative features are then grouped based on local pairwise feature similarity, to produce finer-grained, more precise attention maps highlighting task-relevant parts of the input. The obtained attention maps are ranked according to the activity level of the compound feature, which provides information regarding the important level of the highlighted regions. The proposed model can be easily adapted in a wide variety of modern deep models, where classification is involved. Extensive quantitative and qualitative experiments showcase more comprehensive and accurate visual explanations compared to state-of-the-art attention models and visualization methods across multiple tasks including fine-grained image classification, few-shot classification, and person re-identification, without compromising the classification accuracy. The proposed visualization model sheds imperative light on how neural networks ‘pay their attention’ differently in different tasks. © 2022
引用
收藏
相关论文
共 50 条
  • [1] NPA: Improving Large-scale Graph Neural Networks with Non-parametric Attention
    Zhang, Wentao
    Yan, Guochen
    Shen, Yu
    Ling, Yang
    Tao, Yaoyu
    Cui, Bin
    Tang, Jian
    COMPANION OF THE 2024 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, SIGMOD-COMPANION 2024, 2024, : 414 - 427
  • [2] Non-Parametric High-Resolution SAR Imaging
    Glentis, George Othon
    Zhao, Kexin
    Jakobsson, Andreas
    Li, Jian
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2013, 61 (07) : 1614 - 1624
  • [3] High-Resolution Non-Parametric Spectral Estimation Using The Hirschman Uncertainty
    Liu, Guifeng
    DeBrunner, Victor
    2011 CONFERENCE RECORD OF THE FORTY-FIFTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS (ASILOMAR), 2011, : 336 - 340
  • [4] Personalized smile synthesis using attention-guided global parametric model and local non-parametric model
    Tu, Ching-Ting
    Hsieh, Sung-Hsien
    Chen, Kuan-Lin
    Lien, Jenn-Jier James
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (14) : 21585 - 21609
  • [5] Comparison of parametric and non-parametric methods for chlorophyll estimation based on high-resolution UAV imagery
    Singhal, Gaurav
    Bansod, Babankumar
    Mathew, Lini
    Goswami, Jonali
    Choudhury, B. U.
    Raju, P. L. N.
    CURRENT SCIENCE, 2019, 117 (11): : 1874 - 1879
  • [6] HIGH-RESOLUTION NON-PARAMETRIC SPECTRAL ESTIMATION USING THE HIRSCHMAN OPTIMAL TRANSFORM
    Liu, Guifeng
    DeBrunner, Victor
    2012 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2012, : 3721 - 3724
  • [7] Correction to: Personalized smile synthesis using attention-guided global parametric model and local non-parametric model
    Ching-Ting Tu
    Sung-Hsien Hsieh
    Kuan-Lin Chen
    Jenn-Jier James Lien
    Multimedia Tools and Applications, 2023, 82 : 21611 - 21611
  • [8] A non-parametric softmax for improving neural attention in time-series forecasting
    Totaro, Simone
    Hussain, Amir
    Scardapane, Simone
    NEUROCOMPUTING, 2020, 381 : 177 - 185
  • [9] HIGH-RESOLUTION ATTENTION NETWORK WITH ACOUSTIC SEGMENT MODEL FOR ACOUSTIC SCENE CLASSIFICATION
    Bai, Xue
    Du, Jun
    Pan, Jia
    Zhou, Heng-shun
    Tu, Yan-Hui
    Lee, Chin-Hui
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 656 - 660
  • [10] Semantic Layout Manipulation With High-Resolution Sparse Attention
    Zheng, Haitian
    Lin, Zhe
    Lu, Jingwan
    Cohen, Scott
    Zhang, Jianming
    Xu, Ning
    Luo, Jiebo
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3768 - 3782