Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition

被引:30
|
作者
Yang, Chuanguang [1 ,2 ]
An, Zhulin [1 ]
Zhou, Helong [3 ]
Zhuang, Fuzhen [4 ,5 ]
Xu, Yongjun [1 ]
Zhang, Qian
机构
[1] Chinese Acad Sci, Inst Comp Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[3] Horizon Robot, Beijing 100089, Peoples R China
[4] Beihang Univ, Inst Artificial Intelligence, Beijing 100191, Peoples R China
[5] Zhongguancun Lab, Beijing 100194, Peoples R China
关键词
Contrastive learning; mutual learning; online knowledge distillation; visual recognition;
D O I
10.1109/TPAMI.2023.3257878
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The teacher-free online Knowledge Distillation (KD) aims to train an ensemble of multiple student models collaboratively and distill knowledge from each other. Although existing online KD methods achieve desirable performance, they often focus on class probabilities as the core knowledge type, ignoring the valuable feature representational information. We present a Mutual Contrastive Learning (MCL) framework for online KD. The core idea ofMCLis to perform mutual interaction and transfer of contrastive distributions among a cohort of networks in an online manner. Our MCL can aggregate cross-network embedding information and maximize the lower bound to the mutual information between two networks. This enables each network to learn extra contrastive knowledge from others, leading to better feature representations, thus improving the performance of visual recognition tasks. Beyond the final layer, we extend MCL to intermediate layers and perform an adaptive layer-matching mechanism trained by meta-optimization. Experiments on image classification and transfer learning to visual recognition tasks show that layer-wise MCL can lead to consistent performance gains against state-of-the-art online KD approaches. The superiority demonstrates that layer-wise MCL can guide the network to generate better feature representations. Our code is publicly avaliable at https://github.com/winycg/L-MCL.
引用
收藏
页码:10212 / 10227
页数:16
相关论文
共 50 条
  • [21] Online Knowledge Distillation for Efficient Action Recognition
    Wang, Jiazheng
    Bian, Cunlin
    Zhou, Xian
    Lyu, Fan
    Niu, Zhibin
    Feng, Wei
    2022 IEEE 2ND INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND ARTIFICIAL INTELLIGENCE (CCAI 2022), 2022, : 177 - 181
  • [22] A Fast Knowledge Distillation Framework for Visual Recognition
    Shen, Zhiqiang
    Xing, Eric
    COMPUTER VISION, ECCV 2022, PT XXIV, 2022, 13684 : 673 - 690
  • [23] Correction to: Embedded mutual learning: a novel online distillation method integrating diverse knowledge sources
    Chuanxiu Li
    Guangli Li
    Hongbin Zhang
    Donghong Ji
    Applied Intelligence, 2023, 53 : 17240 - 17240
  • [24] Cuing Without Sharing: A Federated Cued Speech Recognition Framework via Mutual Knowledge Distillation
    Zhang, Yuxuan
    Liu, Lei
    Liu, Li
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 8781 - 8789
  • [25] ABUS tumor segmentation via decouple contrastive knowledge distillation
    Pan, Pan
    Li, Yanfeng
    Chen, Houjin
    Sun, Jia
    Li, Xiaoling
    Cheng, Lin
    PHYSICS IN MEDICINE AND BIOLOGY, 2024, 69 (01):
  • [26] Teacher or supervisor? Effective online knowledge distillation via guided collaborative learning
    Borza, Diana Laura
    Ileni, Tudor Alexandru
    Marinescu, Alexandru Ion
    Darabant, Sergiu Adrian
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 228
  • [27] Identity-Aware Contrastive Knowledge Distillation for Facial Attribute Recognition
    Chen, Si
    Zhu, Xueyan
    Yan, Yan
    Zhu, Shunzhi
    Li, Shao-Zi
    Wang, Da-Han
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (10) : 5692 - 5706
  • [28] Low-light image enhancement via Clustering Contrastive Learning for visual recognition
    Sheng, Guanglei
    Hu, Gang
    Wang, Xiaofeng
    Chen, Wei
    Jiang, Jinlin
    PATTERN RECOGNITION, 2025, 164
  • [29] Smarter peer learning for online knowledge distillation
    Lin, Yu-e
    Liang, Xingzhu
    Hu, Gan
    Fang, Xianjin
    MULTIMEDIA SYSTEMS, 2022, 28 (03) : 1059 - 1067
  • [30] Peer Collaborative Learning for Online Knowledge Distillation
    Wu, Guile
    Gong, Shaogang
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10302 - 10310