Self Supervision to Distillation for Long-Tailed Visual Recognition

被引:37
|
作者
Li, Tianhao [1 ]
Wang, Limin [1 ]
Wu, Gangshan [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
基金
中国国家自然科学基金;
关键词
SMOTE;
D O I
10.1109/ICCV48922.2021.00067
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has achieved remarkable progress for visual recognition on large-scale balanced datasets but still performs poorly on real-world long-tailed data. Previous methods often adopt class re-balanced training strategies to effectively alleviate the imbalance issue, but might be a risk of over-fitting tail classes. The recent decoupling method overcomes over-fitting issues by using a multi-stage training scheme, yet, it is still incapable of capturing tail class information in the feature learning stage. In this paper, we show that soft label can serve as a powerful solution to incorporate label correlation into a multi-stage training scheme for long-tailed recognition. The intrinsic relation between classes embodied by soft labels turns out to be helpful for long-tailed recognition by transferring knowledge from head to tail classes. Specifically, we propose a conceptually simple yet particularly effective multi-stage training scheme, termed as Self Supervised to Distillation (SSD). This scheme is composed of two parts. First, we introduce a self-distillation framework for long-tailed recognition, which can mine the label relation automatically. Second, we present a new distillation label generation module guided by self-supervision. The distilled labels integrate information from both label and data domains that can model long-tailed distribution effectively. We conduct extensive experiments and our method achieves the state-of-the-art results on three long-tailed recognition benchmarks: ImageNet-LT, CIFAR100-LT and iNaturalist 2018. Our SSD outperforms the strong LWS baseline by from 2.7% to 4.5% on various datasets.
引用
收藏
页码:610 / 619
页数:10
相关论文
共 50 条
  • [31] Joint weighted knowledge distillation and multi-scale feature distillation for long-tailed recognition
    He, Yiru
    Wang, Shiqian
    Yu, Junyang
    Liu, Chaoyang
    He, Xin
    Li, Han
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (04) : 1647 - 1661
  • [32] MEID: Mixture-of-Experts with Internal Distillation for Long-Tailed Video Recognition
    Li, Xinjie
    Xu, Huijuan
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 1451 - 1459
  • [33] Bilinear-experts network with self-adaptive sampler for long-tailed visual recognition
    Wang, Qin
    Kwong, Sam
    Wang, Xizhao
    NEUROCOMPUTING, 2025, 633
  • [34] Balanced knowledge distillation for long-tailed learning
    Zhang, Shaoyu
    Chen, Chen
    Hu, Xiyuan
    Peng, Silong
    NEUROCOMPUTING, 2023, 527 : 36 - 46
  • [35] Long-Tailed Recognition Based on Self-attention Mechanism
    Feng, Zekai
    Jia, Hong
    Li, Mengke
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14876 : 380 - 391
  • [36] Class-Difficulty Based Methods for Long-Tailed Visual Recognition
    Saptarshi Sinha
    Hiroki Ohashi
    Katsuyuki Nakamura
    International Journal of Computer Vision, 2022, 130 : 2517 - 2531
  • [37] CLASS-AWARE FEATURE PERTURBATION FOR LONG-TAILED VISUAL RECOGNITION
    Chen, Xicheng
    Yel, Haibo
    Zhou, Fangyu
    2024 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME 2024, 2024,
  • [38] Long-tailed Visual Recognition via Gaussian Clouded Logit Adjustment
    Li, Mengke
    Cheung, Yiu-Ming
    Lu, Yang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 6919 - 6928
  • [39] Prototype-based classifier learning for long-tailed visual recognition
    Xiu-Shen Wei
    Shu-Lin Xu
    Hao Chen
    Liang Xiao
    Yuxin Peng
    Science China Information Sciences, 2022, 65
  • [40] Rebalanced supervised contrastive learning with prototypes for long-tailed visual recognition
    Chang, Xuhui
    Zhai, Junhai
    Qiu, Shaoxin
    Sun, Zhengrong
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2025, 252