IMF: Integrating Matched Features Using Attentive Logit in Knowledge Distillation

被引:0
|
作者
Kim, Jeongho [1 ]
Lee, Hanbeen [2 ]
Woo, Simon S. [3 ]
机构
[1] Korea Adv Inst Sci & Technol, Korea Adv Inst Sci & Technol, Daejeon, South Korea
[2] NAVER Z Corp, Seongnam, South Korea
[3] Sungkyunkwan Univ, Dept Artificial Intelligence, Seoul, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge distillation (KD) is an effective method for transferring the knowledge of a teacher model to a student model, that aims to improve the latter's performance efficiently. Although generic knowledge distillation methods such as softmax representation distillation and intermediate feature matching have demonstrated improvements with various tasks, only marginal improvements are shown in student networks due to their limited model capacity. In this work, to address the student model's limitation, we propose a novel flexible KD framework, Integrating Matched Features using Attentive Logit in Knowledge Distillation (IMF). Our approach introduces an intermediate feature distiller (IFD) to improve the overall performance of the student model by directly distilling the teacher's knowledge into branches of student models. The generated output of IFD, which is trained by the teacher model, is effectively combined by attentive logit. We use only a few blocks of the student and the trained IFD during inference, requiring an equal or less number of parameters. Through extensive experiments, we demonstrate that IMF consistently outperforms other state-of-the-art methods with a large margin over the various datasets in different tasks without extra computation.
引用
收藏
页码:974 / +
页数:10
相关论文
共 50 条
  • [21] Recommendation Model by Integrating Knowledge Graph and Image Features
    Chen Qiaosong
    Guo Aodong
    Du Yulu
    Zhang Yiwen
    Zhu Yue
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2022, 44 (05) : 1723 - 1733
  • [22] A Dynamic Knowledge Diagnosis Approach Integrating Cognitive Features
    Huang, Tao
    Yang, Huali
    Li, Zhi
    Xie, Hekun
    Geng, Jing
    Zhang, Hao
    IEEE ACCESS, 2021, 9 (09): : 116814 - 116829
  • [23] A hierarchical feature-logit-based knowledge distillation scheme for internal defect detection of magnetic tiles
    Xie, Luofeng
    Cen, Xuexiang
    Lu, Houhong
    Yin, Guofu
    Yin, Ming
    ADVANCED ENGINEERING INFORMATICS, 2024, 61
  • [24] Integrating Knowledge Distillation With Learning to Rank for Few-Shot Scene Classification
    Liu, Yishu
    Zhang, Liqiang
    Han, Zhengzhuo
    Chen, Conghui
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [25] Lightweight video object segmentation: Integrating online knowledge distillation for fast segmentation
    Hou, Zhiqiang
    Wang, Chenxu
    Ma, Sugang
    Dong, Jiale
    Wang, Yunchen
    Yu, Wangsheng
    Yang, Xiaobao
    KNOWLEDGE-BASED SYSTEMS, 2025, 308
  • [26] COOPERATING KNOWLEDGE INTEGRATING SYSTEMS FOR THE SYNTHESIS OF ENERGY-INTEGRATED DISTILLATION PROCESSES
    SCHEMBECKER, G
    SCHUTTENHELM, W
    SIMMROCK, KH
    COMPUTERS & CHEMICAL ENGINEERING, 1994, 18 : S131 - S135
  • [27] Heterogeneous Knowledge Distillation Using Conceptual Learning
    Yu, Yerin
    Kim, Namgyu
    IEEE ACCESS, 2024, 12 : 52803 - 52814
  • [28] Analysis of Model Compression Using Knowledge Distillation
    Hong, Yu-Wei
    Leu, Jenq-Shiou
    Faisal, Muhamad
    Prakosa, Setya Widyawan
    IEEE ACCESS, 2022, 10 : 85095 - 85105
  • [29] Knowledge Distillation in Histology Landscape by Multi-Layer Features Supervision
    Javed, Sajid
    Mahmood, Arif
    Qaiser, Talha
    Werghi, Naoufel
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2023, 27 (04) : 2037 - 2046
  • [30] The Chain of Self-Taught Knowledge Distillation Combining Output and Features
    Wang, Yunnan
    Chen, Hao
    Li, Jianxun
    PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 5115 - 5120