IMF: Integrating Matched Features Using Attentive Logit in Knowledge Distillation

被引:0
|
作者
Kim, Jeongho [1 ]
Lee, Hanbeen [2 ]
Woo, Simon S. [3 ]
机构
[1] Korea Adv Inst Sci & Technol, Korea Adv Inst Sci & Technol, Daejeon, South Korea
[2] NAVER Z Corp, Seongnam, South Korea
[3] Sungkyunkwan Univ, Dept Artificial Intelligence, Seoul, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge distillation (KD) is an effective method for transferring the knowledge of a teacher model to a student model, that aims to improve the latter's performance efficiently. Although generic knowledge distillation methods such as softmax representation distillation and intermediate feature matching have demonstrated improvements with various tasks, only marginal improvements are shown in student networks due to their limited model capacity. In this work, to address the student model's limitation, we propose a novel flexible KD framework, Integrating Matched Features using Attentive Logit in Knowledge Distillation (IMF). Our approach introduces an intermediate feature distiller (IFD) to improve the overall performance of the student model by directly distilling the teacher's knowledge into branches of student models. The generated output of IFD, which is trained by the teacher model, is effectively combined by attentive logit. We use only a few blocks of the student and the trained IFD during inference, requiring an equal or less number of parameters. Through extensive experiments, we demonstrate that IMF consistently outperforms other state-of-the-art methods with a large margin over the various datasets in different tasks without extra computation.
引用
收藏
页码:974 / +
页数:10
相关论文
共 50 条
  • [31] A Neural Attentive Model Using Human Semantic Knowledge for Clickbait Detection
    Wei, Feng
    Uyen Trang Nguyen
    2020 IEEE INTL SYMP ON PARALLEL & DISTRIBUTED PROCESSING WITH APPLICATIONS, INTL CONF ON BIG DATA & CLOUD COMPUTING, INTL SYMP SOCIAL COMPUTING & NETWORKING, INTL CONF ON SUSTAINABLE COMPUTING & COMMUNICATIONS (ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM 2020), 2020, : 770 - 776
  • [32] Embracing the Dark Knowledge: Domain Generalization Using Regularized Knowledge Distillation
    Wang, Yufei
    Li, Haoliang
    Chau, Lap-pui
    Kot, Alex C.
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2595 - 2604
  • [33] Knowledge-based operation optimization of a distillation unit integrating feedstock property considerations
    Li, Sihong
    Zheng, Yi
    Li, Shaoyuan
    Huang, Meng
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 107
  • [34] Integrating Knowledge Distillation and Transfer Learning for Enhanced QoT-Estimation in Optical Networks
    Usmani, Fehmida
    Khan, Ihtesham
    Mehran, Arsalan
    Ahmad, Arsalan
    Curri, Vittorio
    IEEE ACCESS, 2024, 12 : 156785 - 156802
  • [35] Micro-expression Action Unit Detection with Dual-view Attentive Similarity-Preserving Knowledge Distillation
    Li, Yante
    Peng, Wei
    Zhao, Guoying
    2021 16TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2021), 2021,
  • [36] Embedded mutual learning: A novel online distillation method integrating diverse knowledge sources
    Chuanxiu Li
    Guangli Li
    Hongbin Zhang
    Donghong Ji
    Applied Intelligence, 2023, 53 : 11524 - 11537
  • [37] Embedded mutual learning: A novel online distillation method integrating diverse knowledge sources
    Li, Chuanxiu
    Li, Guangli
    Zhang, Hongbin
    Ji, Donghong
    APPLIED INTELLIGENCE, 2023, 53 (10) : 11524 - 11537
  • [38] Image Super-Resolution Using Knowledge Distillation
    Gao, Qinquan
    Zhao, Yan
    Li, Gen
    Tong, Tong
    COMPUTER VISION - ACCV 2018, PT II, 2019, 11362 : 527 - 541
  • [39] A Fast Scene Text Detector Using Knowledge Distillation
    Yang, Peng
    Zhang, Fanlong
    Yang, Guowei
    IEEE ACCESS, 2019, 7 : 22588 - 22598
  • [40] TrustAL: Trustworthy Active Learning Using Knowledge Distillation
    Kwak, Beong-woo
    Kim, Youngwook
    Kim, Yu Jin
    Hwang, Seung-won
    Yeo, Jinyoung
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 7263 - 7271