Multi-level Logit Distillation

被引:38
|
作者
Jin, Ying [1 ]
Wang, Jiaqi [2 ]
Lin, Dahua [1 ,2 ]
机构
[1] Chinese Univ Hong Kong, CUHK Sense Time Joint Lab, Hong Kong, Peoples R China
[2] Shanghai AI Lab, Shanghai, Peoples R China
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2023年
关键词
D O I
10.1109/CVPR52729.2023.02325
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Knowledge Distillation (KD) aims at distilling the knowledge from the large teacher model to a lightweight student model. Mainstream KD methods can be divided into two categories, logit distillation, and feature distillation. The former is easy to implement, but inferior in performance, while the latter is not applicable to some practical circumstances due to concerns such as privacy and safety. Towards this dilemma, in this paper, we explore a stronger logit distillation method via making better utilization of logit outputs. Concretely, we propose a simple yet effective approach to logit distillation via multi-level prediction alignment. Through this framework, the prediction alignment is not only conducted at the instance level, but also at the batch and class level, through which the student model learns instance prediction, input correlation, and category correlation simultaneously. In addition, a prediction augmentation mechanism based on model calibration further boosts the performance. Extensive experiment results validate that our method enjoys consistently higher performance than previous logit distillation methods, and even reaches competitive performance with mainstream feature distillation methods. Code is available at https://github.com/Jin-Ying/Multi-Level-Logit-Distillation.
引用
收藏
页码:24276 / 24285
页数:10
相关论文
共 50 条
  • [31] MULTI-LEVEL INTERCONNECTIONS
    RUEHLEMA.HE
    ELECTRONIC ENGINEER, 1972, 31 (10): : 41 - &
  • [32] MULTI-LEVEL SYSTEMS
    LIN, Y
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 1989, 20 (10) : 1875 - 1889
  • [33] Multi-level governance
    James, O
    WEST EUROPEAN POLITICS, 2005, 28 (03) : 687 - 689
  • [34] Multi-Level Modeling with Openflexo/FML A Contribution to the Multi-Level Process Challenge
    Guerin, Sylvain
    Champeau, Joel
    Bach, Jean-Christophe
    Beugnard, Antoine
    Dagnat, Fabien
    Martinez, Salvador
    ENTERPRISE MODELLING AND INFORMATION SYSTEMS ARCHITECTURES-AN INTERNATIONAL JOURNAL, 2022, 17 : 1 - 21
  • [35] Multi-Level Symbolic Regression: Function Structure Learning for Multi-Level Data
    Sen Fong, Kei
    Motani, Mehul
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [36] Multi-level model predictive controller with satisfactory optimization for multi-level converters
    Tamim, Touati Mohamed
    Li, Shaoyuan
    Wu, Jing
    SIMULATION MODELLING PRACTICE AND THEORY, 2019, 92 : 1 - 16
  • [37] PKD-Net: Distillation of prior knowledge for image completion by multi-level semantic attention
    Lu, Qiong
    Lin, Huaizhong
    Xing, Wei
    Zhao, Lei
    Chen, Jingjing
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2024, 36 (10):
  • [38] A Multi-Level Adaptive Lightweight Net for Damaged Road Marking Detection Based on Knowledge Distillation
    Wang, Junwei
    Zeng, Xiangqiang
    Wang, Yong
    Ren, Xiang
    Wang, Dongliang
    Qu, Wenqiu
    Liao, Xiaohan
    Pan, Peifen
    REMOTE SENSING, 2024, 16 (14)
  • [39] Weather Classification for Autonomous Vehicles under Adverse Conditions Using Multi-Level Knowledge Distillation
    Manivannan, Parthasarathi
    Sathyaprakash, Palaniyappan
    Jayakumar, Vaithiyashankar
    Chandrasekaran, Jayakumar
    Ananthanarayanan, Bragadeesh Srinivasan
    Sayeed, Md Shohel
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 81 (03): : 4327 - 4347
  • [40] Multi-level knowledge distillation for low-resolution object detection and facial expression recognition
    Ma, Tingsong
    Tian, Wenhong
    Xie, Yuanlun
    KNOWLEDGE-BASED SYSTEMS, 2022, 240