One-stage self-distillation guided knowledge transfer for long-tailed visual recognition

被引:1
|
作者
Xia, Yuelong [1 ,2 ,3 ]
Zhang, Shu [1 ,2 ,3 ]
Wang, Jun [2 ,3 ]
Zou, Wei [1 ,2 ,3 ]
Zhou, Juxiang [2 ,3 ]
Wen, Bin [1 ,2 ,3 ]
机构
[1] Yunnan Normal Univ, Sch Informat Sci & Technol, Kunming, Peoples R China
[2] Yunnan Normal Univ, Minist Educ, Key Lab Educ Informatizat Nationalities, Kunming, Peoples R China
[3] Yunnan Normal Univ, Yunnan Key Lab Smart Educ, Kunming, Peoples R China
基金
中国国家自然科学基金;
关键词
knowledge transfer; long-tailed recognition; one-stage training; self-distillation;
D O I
10.1002/int.23068
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has achieved remarkable progress for visual recognition on balanced data sets but still performs poorly on real-world long-tailed data distribution. The existing methods mainly decouple the problem into the two-stage decoupling training, that is, representation learning and classifier training, or multistage training based on knowledge distillation, thus resulting in huge training steps and extra computation cost. In this paper, we propose a conceptually simple yet effective One-stage Long-tailed Self-Distillation framework, called OLSD, which simultaneously takes representation learning and classifier training into one-stage training. For representation learning, we take two different sampling distributions and mixup them to input them into two branches, where the collaborative consistency loss is introduced to train network consistency, and we theoretically show that the proposed mixup naturally generates a tail-majority distribution mixup. For classifier training, we introduce balanced self-distillation guided knowledge transfer to improve generalization performance, where we theoretically show that proposed knowledge transfer implicitly minimizes not only cross-entropy but also KL divergence between head-to-tail and tail-to-head. Extensive experiments on long-tailed CIFAR10/100, ImageNet-LT and multilabel long-tailed VOC-LT demonstrate the proposed method's effectiveness.
引用
收藏
页码:11893 / 11908
页数:16
相关论文
共 50 条
  • [31] Balanced knowledge distillation for one-stage object detector
    Lee, Sungwook
    Lee, Seunghyun
    Song, Byung Cheol
    NEUROCOMPUTING, 2022, 500 : 394 - 404
  • [32] Key Point Sensitive Loss for Long-Tailed Visual Recognition
    Li, Mengke
    Cheung, Yiu-Ming
    Hu, Zhikai
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (04) : 4812 - 4825
  • [33] Dynamic Learnable Logit Adjustment for Long-Tailed Visual Recognition
    Zhang, Enhao
    Geng, Chuanxing
    Li, Chaohua
    Chen, Songcan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (09) : 7986 - 7997
  • [34] Feature Re-Balancing for Long-Tailed Visual Recognition
    Zhao, Yan
    Chen, Weicong
    Huang, Kai
    Zhu, Jihong
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [35] FCC: Feature Clusters Compression for Long-Tailed Visual Recognition
    Li, Jian
    Meng, Ziyao
    Shi, Daqian
    Song, Rui
    Diao, Xiaolei
    Wang, Jingwen
    Xu, Hao
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 24080 - 24089
  • [36] Balanced clustering contrastive learning for long-tailed visual recognition
    Kim, Byeong-il
    Ko, Byoung Chul
    PATTERN ANALYSIS AND APPLICATIONS, 2025, 28 (01)
  • [37] Feature calibration and feature separation for long-tailed visual recognition
    Wang, Qianqian
    Zhou, Fangyu
    Zhao, Xiangge
    Lin, Yangtao
    Ye, Haibo
    NEUROCOMPUTING, 2025, 637
  • [38] Adaptive Logit Adjustment Loss for Long-Tailed Visual Recognition
    Zhao, Yan
    Chen, Weicong
    Tan, Xu
    Huang, Kai
    Zhu, Jihong
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 3472 - 3480
  • [39] Hierarchical block aggregation network for long-tailed visual recognition
    Pang, Shanmin
    Wang, Weiye
    Zhang, Renzhong
    Hao, Wenyu
    NEUROCOMPUTING, 2023, 549
  • [40] MetaSAug: Meta Semantic Augmentation for Long-Tailed Visual Recognition
    Li, Shuang
    Gong, Kaixiong
    Liu, Chi Harold
    Wang, Yulin
    Qiao, Feng
    Cheng, Xinjing
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 5208 - 5217