Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition

被引:30
|
作者
Yang, Chuanguang [1 ,2 ]
An, Zhulin [1 ]
Zhou, Helong [3 ]
Zhuang, Fuzhen [4 ,5 ]
Xu, Yongjun [1 ]
Zhang, Qian
机构
[1] Chinese Acad Sci, Inst Comp Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[3] Horizon Robot, Beijing 100089, Peoples R China
[4] Beihang Univ, Inst Artificial Intelligence, Beijing 100191, Peoples R China
[5] Zhongguancun Lab, Beijing 100194, Peoples R China
关键词
Contrastive learning; mutual learning; online knowledge distillation; visual recognition;
D O I
10.1109/TPAMI.2023.3257878
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The teacher-free online Knowledge Distillation (KD) aims to train an ensemble of multiple student models collaboratively and distill knowledge from each other. Although existing online KD methods achieve desirable performance, they often focus on class probabilities as the core knowledge type, ignoring the valuable feature representational information. We present a Mutual Contrastive Learning (MCL) framework for online KD. The core idea ofMCLis to perform mutual interaction and transfer of contrastive distributions among a cohort of networks in an online manner. Our MCL can aggregate cross-network embedding information and maximize the lower bound to the mutual information between two networks. This enables each network to learn extra contrastive knowledge from others, leading to better feature representations, thus improving the performance of visual recognition tasks. Beyond the final layer, we extend MCL to intermediate layers and perform an adaptive layer-matching mechanism trained by meta-optimization. Experiments on image classification and transfer learning to visual recognition tasks show that layer-wise MCL can lead to consistent performance gains against state-of-the-art online KD approaches. The superiority demonstrates that layer-wise MCL can guide the network to generate better feature representations. Our code is publicly avaliable at https://github.com/winycg/L-MCL.
引用
收藏
页码:10212 / 10227
页数:16
相关论文
共 50 条
  • [31] Smarter peer learning for online knowledge distillation
    Yu-e Lin
    Xingzhu Liang
    Gan Hu
    Xianjin Fang
    Multimedia Systems, 2022, 28 : 1059 - 1067
  • [32] Learning Rules in Knowledge Graphs via Contrastive Learning
    Feng, Xiaoyang
    Liu, Xueli
    Yang, Yajun
    Wang, Wenjun
    Wang, Jun
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2024, PT IV, 2024, 14853 : 408 - 424
  • [33] Asymmetrical Contrastive Learning Network via Knowledge Distillation for No-Service Rail Surface Defect Detection
    Zhou, Wujie
    Sun, Xinyu
    Qian, Xiaohong
    Fang, Meixin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [34] Leveraging Contrastive Learning and Knowledge Distillation for Incomplete Modality Rumor Detection
    Xul, Fan
    Fan, Pinyun
    Huang, Qi
    Zou, Bowei
    Awe, AiTi
    Wang, Mingwen
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13492 - 13503
  • [35] Balanced Knowledge Distillation with Contrastive Learning for Document Re-ranking
    Yang, Yingrui
    He, Shanxiu
    Qiao, Yifan
    Xie, Wentai
    Yang, Tao
    PROCEEDINGS OF THE 2023 ACM SIGIR INTERNATIONAL CONFERENCE ON THE THEORY OF INFORMATION RETRIEVAL, ICTIR 2023, 2023, : 247 - 255
  • [36] A Federated Domain Adaptation Algorithm Based on Knowledge Distillation and Contrastive Learning
    HUANG Fang
    FANG Zhijun
    SHI Zhicai
    ZHUANG Lehui
    LI Xingchen
    HUANG Bo
    WuhanUniversityJournalofNaturalSciences, 2022, 27 (06) : 499 - 507
  • [37] Improving Structural and Semantic Global Knowledge in Graph Contrastive Learning with Distillation
    Wen, Mi
    Wang, Hongwei
    Xue, Yunsheng
    Wu, Yi
    Wen, Hong
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PT II, PAKDD 2024, 2024, 14646 : 364 - 375
  • [38] A Medical Image Segmentation Method Combining Knowledge Distillation and Contrastive Learning
    Ma, Xiaoxuan
    Shan, Sihan
    Sui, Dong
    Journal of Computers (Taiwan), 2024, 35 (03) : 363 - 377
  • [39] Multi-Mode Online Knowledge Distillation for Self-Supervised Visual Representation Learning
    Song, Kaiyou
    Xie, Jin
    Zhang, Shan
    Luo, Zimeng
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11848 - 11857
  • [40] Domain Knowledge Distillation and Supervised Contrastive Learning for Industrial Process Monitoring
    Ai, Mingxi
    Xie, Yongfang
    Ding, Steven X. X.
    Tang, Zhaohui
    Gui, Weihua
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2023, 70 (09) : 9452 - 9462