Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition

被引:30
|
作者
Yang, Chuanguang [1 ,2 ]
An, Zhulin [1 ]
Zhou, Helong [3 ]
Zhuang, Fuzhen [4 ,5 ]
Xu, Yongjun [1 ]
Zhang, Qian
机构
[1] Chinese Acad Sci, Inst Comp Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[3] Horizon Robot, Beijing 100089, Peoples R China
[4] Beihang Univ, Inst Artificial Intelligence, Beijing 100191, Peoples R China
[5] Zhongguancun Lab, Beijing 100194, Peoples R China
关键词
Contrastive learning; mutual learning; online knowledge distillation; visual recognition;
D O I
10.1109/TPAMI.2023.3257878
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The teacher-free online Knowledge Distillation (KD) aims to train an ensemble of multiple student models collaboratively and distill knowledge from each other. Although existing online KD methods achieve desirable performance, they often focus on class probabilities as the core knowledge type, ignoring the valuable feature representational information. We present a Mutual Contrastive Learning (MCL) framework for online KD. The core idea ofMCLis to perform mutual interaction and transfer of contrastive distributions among a cohort of networks in an online manner. Our MCL can aggregate cross-network embedding information and maximize the lower bound to the mutual information between two networks. This enables each network to learn extra contrastive knowledge from others, leading to better feature representations, thus improving the performance of visual recognition tasks. Beyond the final layer, we extend MCL to intermediate layers and perform an adaptive layer-matching mechanism trained by meta-optimization. Experiments on image classification and transfer learning to visual recognition tasks show that layer-wise MCL can lead to consistent performance gains against state-of-the-art online KD approaches. The superiority demonstrates that layer-wise MCL can guide the network to generate better feature representations. Our code is publicly avaliable at https://github.com/winycg/L-MCL.
引用
收藏
页码:10212 / 10227
页数:16
相关论文
共 50 条
  • [1] Mutual mentor: Online contrastive distillation network for general continual learning
    Wang, Qiang
    Ji, Zhong
    Li, Jin
    Pang, Yanwei
    NEUROCOMPUTING, 2023, 537 : 37 - 48
  • [2] MULTI-VIEW CONTRASTIVE LEARNING FOR ONLINE KNOWLEDGE DISTILLATION
    Yang, Chuanguang
    An, Zhulin
    Xu, Yongjun
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3750 - 3754
  • [3] Improving Deep Mutual Learning via Knowledge Distillation
    Lukman, Achmad
    Yang, Chuan-Kai
    APPLIED SCIENCES-BASEL, 2022, 12 (15):
  • [4] DECENTRALIZED FEDERATED LEARNING VIA MUTUAL KNOWLEDGE DISTILLATION
    Huang, Yue
    Kong, Lanju
    Li, Qingzhong
    Zhang, Baochen
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 342 - 347
  • [5] Federated Split Learning via Mutual Knowledge Distillation
    Luo, Linjun
    Zhang, Xinglin
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2024, 11 (03): : 2729 - 2741
  • [6] Mutual Contrastive Learning for Visual Representation Learning
    Yang, Chuanguang
    An, Zhulin
    Cai, Linhang
    Xu, Yongjun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3045 - 3053
  • [7] Quantization via Distillation and Contrastive Learning
    Pei, Zehua
    Yao, Xufeng
    Zhao, Wenqian
    Yu, Bei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17164 - 17176
  • [8] Feature Fusion for Online Mutual Knowledge Distillation
    Kim, Jangho
    Hyun, Minsung
    Chung, Inseop
    Kwak, Nojun
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 4619 - 4625
  • [9] Boosting Contrastive Learning with Relation Knowledge Distillation
    Zheng, Kai
    Wang, Yuanjiang
    Yuan, Ye
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3508 - 3516
  • [10] Knowledge Distillation for Single Image Super-Resolution via Contrastive Learning
    Liu, Cencen
    Zhang, Dongyang
    Qin, Ke
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 1079 - 1083