Few-Shot Class-Incremental Learning via Class-Aware Bilateral Distillation

被引:30
|
作者
Zhao, Linglan [1 ]
Lu, Jing [2 ]
Xu, Yunlu [2 ]
Cheng, Zhanzhan [2 ]
Guo, Dashan [1 ]
Niu, Yi [2 ]
Fang, Xiangzhong [1 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Elect Engn, Shanghai, Peoples R China
[2] Hikvis Res Inst, Hangzhou, Zhejiang, Peoples R China
关键词
D O I
10.1109/CVPR52729.2023.01139
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-Shot Class-Incremental Learning (FSCIL) aims to continually learn novel classes based on only few training samples, which poses a more challenging task than the well-studied Class-Incremental Learning (CIL) due to data scarcity. While knowledge distillation, a prevailing technique in CIL, can alleviate the catastrophic forgetting of older classes by regularizing outputs between current and previous model, it fails to consider the overfitting risk of novel classes in FSCIL. To adapt the powerful distillation technique for FSCIL, we propose a novel distillation structure, by taking the unique challenge of overfitting into account. Concretely, we draw knowledge from two complementary teachers. One is the model trained on abundant data from base classes that carries rich general knowledge, which can be leveraged for easing the overfitting of current novel classes. The other is the updated model from last incremental session that contains the adapted knowledge of previous novel classes, which is used for alleviating their forgetting. To combine the guidances, an adaptive strategy conditioned on the class-wise semantic similarities is introduced. Besides, for better preserving base class knowledge when accommodating novel concepts, we adopt a two-branch network with an attention-based aggregation module to dynamically merge predictions from two complementary branches. Extensive experiments on 3 popular FSCIL datasets: mini-ImageNet, CIFAR100 and CUB200 validate the effectiveness of our method by surpassing existing works by a significant margin. Code is available at https://github.com/LinglanZhao/BiDistFSCIL.
引用
收藏
页码:11838 / 11847
页数:10
相关论文
共 50 条
  • [1] Few-Shot Class-Incremental Learning via Relation Knowledge Distillation
    Dong, Songlin
    Hong, Xiaopeng
    Tao, Xiaoyu
    Chang, Xinyuan
    Wei, Xing
    Gong, Yihong
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 1255 - 1263
  • [2] Semantic-aware Knowledge Distillation for Few-Shot Class-Incremental Learning
    Cheraghian, Ali
    Rahman, Shafin
    Fang, Pengfei
    Roy, Soumava Kumar
    Petersson, Lars
    Harandi, Mehrtash
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 2534 - 2543
  • [3] A survey on few-shot class-incremental learning
    Tian, Songsong
    Li, Lusi
    Li, Weijun
    Ran, Hang
    Ning, Xin
    Tiwari, Prayag
    NEURAL NETWORKS, 2024, 169 : 307 - 324
  • [4] A survey on few-shot class-incremental learning
    Tian, Songsong
    Li, Lusi
    Li, Weijun
    Ran, Hang
    Ning, Xin
    Tiwari, Prayag
    Neural Networks, 2024, 169 : 307 - 324
  • [5] Graph Few-shot Class-incremental Learning
    Tan, Zhen
    Ding, Kaize
    Guo, Ruocheng
    Liu, Huan
    WSDM'22: PROCEEDINGS OF THE FIFTEENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2022, : 987 - 996
  • [6] Constrained Few-shot Class-incremental Learning
    Hersche, Michael
    Karunaratne, Geethan
    Cherubini, Giovanni
    Benini, Luca
    Sebastian, Abu
    Rahimi, Abbas
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9047 - 9057
  • [7] Uncertainty-Aware Distillation for Semi-Supervised Few-Shot Class-Incremental Learning
    Cui, Yawen
    Deng, Wanxia
    Chen, Haoyu
    Liu, Li
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (10) : 14259 - 14272
  • [8] ACTIVE CLASS SELECTION FOR FEW-SHOT CLASS-INCREMENTAL LEARNING
    McClurg, Christopher
    Ayub, Ali
    Tyagi, Harsh
    Rajtmajer, Sarah M.
    Wagner, Alan R.
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 811 - 827
  • [9] On the Approximation Risk of Few-Shot Class-Incremental Learning
    Wang, Xuan
    Ji, Zhong
    Liu, Xiyao
    Pang, Yanwei
    Han, Jungong
    COMPUTER VISION - ECCV 2024, PT LI, 2025, 15109 : 162 - 178
  • [10] Forward Compatible Few-Shot Class-Incremental Learning
    Zhou, Da-Wei
    Wang, Fu-Yun
    Ye, Han-Jia
    Ma, Liang
    Pu, Shiliang
    Zhan, De-Chuan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9036 - 9046