Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition

被引:30
|
作者
Yang, Chuanguang [1 ,2 ]
An, Zhulin [1 ]
Zhou, Helong [3 ]
Zhuang, Fuzhen [4 ,5 ]
Xu, Yongjun [1 ]
Zhang, Qian
机构
[1] Chinese Acad Sci, Inst Comp Technol, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[3] Horizon Robot, Beijing 100089, Peoples R China
[4] Beihang Univ, Inst Artificial Intelligence, Beijing 100191, Peoples R China
[5] Zhongguancun Lab, Beijing 100194, Peoples R China
关键词
Contrastive learning; mutual learning; online knowledge distillation; visual recognition;
D O I
10.1109/TPAMI.2023.3257878
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The teacher-free online Knowledge Distillation (KD) aims to train an ensemble of multiple student models collaboratively and distill knowledge from each other. Although existing online KD methods achieve desirable performance, they often focus on class probabilities as the core knowledge type, ignoring the valuable feature representational information. We present a Mutual Contrastive Learning (MCL) framework for online KD. The core idea ofMCLis to perform mutual interaction and transfer of contrastive distributions among a cohort of networks in an online manner. Our MCL can aggregate cross-network embedding information and maximize the lower bound to the mutual information between two networks. This enables each network to learn extra contrastive knowledge from others, leading to better feature representations, thus improving the performance of visual recognition tasks. Beyond the final layer, we extend MCL to intermediate layers and perform an adaptive layer-matching mechanism trained by meta-optimization. Experiments on image classification and transfer learning to visual recognition tasks show that layer-wise MCL can lead to consistent performance gains against state-of-the-art online KD approaches. The superiority demonstrates that layer-wise MCL can guide the network to generate better feature representations. Our code is publicly avaliable at https://github.com/winycg/L-MCL.
引用
收藏
页码:10212 / 10227
页数:16
相关论文
共 50 条
  • [41] FedRCIL: Federated Knowledge Distillation for Representation based Contrastive Incremental Learning
    Psaltis, Athanasios
    Chatzikonstantinou, Christos
    Patrikakis, Charalampos Z.
    Daras, Petros
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 3455 - 3464
  • [42] JointContrast: Skeleton-Based Mutual Action Recognition with Contrastive Learning
    Jia, Xiangze
    Zhang, Ji
    Wang, Zhen
    Luo, Yonglong
    Chen, Fulong
    Xiao, Jing
    PRICAI 2022: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2022, 13631 : 478 - 489
  • [43] DCML: Deep contrastive mutual learning for COVID-19 recognition
    Zhang, Hongbin
    Liang, Weinan
    Li, Chuanxiu
    Xiong, Qipeng
    Shi, Haowei
    Hu, Lang
    Li, Guangli
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2022, 77
  • [44] FedGM: Heterogeneous Federated Learning via Generative Learning and Mutual Distillation
    Peng, Chao
    Guo, Yiming
    Chen, Yao
    Rui, Qilin
    Yang, Zhengfeng
    Xu, Chenyang
    EURO-PAR 2023: PARALLEL PROCESSING, 2023, 14100 : 339 - 351
  • [45] STABLE KNOWLEDGE TRANSFER FOR CONTRASTIVE DISTILLATION
    Tang, Qiankun
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 4995 - 4999
  • [46] Online Knowledge Distillation for Multi-task Learning
    Jacob, Geethu Miriam
    Agarwal, Vishal
    Stenger, Bjorn
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2358 - 2367
  • [47] TaCo: Textual Attribute Recognition via Contrastive Learning
    Nie, Chang
    Hu, Yiqing
    Qu, Yanqiu
    Liu, Hao
    Jiang, Deqiang
    Ren, Bo
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1949 - 1956
  • [48] Learning to Specialize with Knowledge Distillation for Visual Question Answering
    Mun, Jonghwan
    Lee, Kimin
    Shin, Jinwoo
    Han, Bohyung
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [49] Probabilistic Contrastive Learning for Long-Tailed Visual Recognition
    Du, Chaoqun
    Wang, Yulin
    Song, Shiji
    Huang, Gao
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 5890 - 5904
  • [50] Balanced Contrastive Learning for Long-Tailed Visual Recognition
    Zhu, Jianggang
    Wang, Zheng
    Chen, Jingjing
    Chen, Yi-Ping Phoebe
    Jiang, Yu-Gang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 6898 - 6907