Continual Learning Based on Knowledge Distillation and Representation Learning

被引:0
|
作者
Chen, Xiu-Yan [1 ]
Liu, Jian-Wei [1 ]
Li, Wen-Tao [1 ]
机构
[1] China Univ Petr, Dept Automat, Coll Informat Sci & Engn, Beijing 102249, Peoples R China
关键词
Continual learning; Class incremental learning; Representation learning; Knowledge distillation; Catastrophic forgetting;
D O I
10.1007/978-3-031-15937-4_3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, continual learning that is more in line with real-world scenarios has received more attention. In order to solve the catastrophic forgetting problem in continual learning, researchers have put forward various solutions, which are simply summarized into three types: network structure-based methods, rehearsal-based methods and regularization-based methods. Inspired by pseudo-rehearsal and regularization methods, we propose a novel Continual Learning Based on Knowledge Distillation and Representation Learning (KRCL) model, which employs Beta-VAE as a representation learning module to extract a shared representation of learned tasks. In addition, Beta-VAE is also used as a generative model to generate pseudo samples of historical task, and KRCL trains the pseudo samples of the previous tasks together with the data of the current task, and then combines the knowledge distillation process to extract the dark knowledge from the old task model to alleviate the catastrophic forgetting. We compare KRCL with the Fine-tune, LWF, IRCL and KRCL real baseline methods on four benchmark datasets. The result shows that the KRCL model achieves state-of-the-art performance in standard continual learning tasks.
引用
收藏
页码:27 / 38
页数:12
相关论文
共 50 条
  • [1] Continual Learning With Knowledge Distillation: A Survey
    Li, Songze
    Su, Tonghua
    Zhang, Xuyao
    Wang, Zhongjie
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [2] Contrastive Supervised Distillation for Continual Representation Learning
    Barletti, Tommaso
    Biondi, Niccolo
    Pernici, Federico
    Bruni, Matteo
    Del Bimbo, Alberto
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2022, PT I, 2022, 13231 : 597 - 609
  • [3] Graph-based interactive knowledge distillation for social relation continual learning
    Tang, Wang
    Qing, Linbo
    Wang, Pingyu
    Li, Lindong
    Peng, Yonghong
    NEUROCOMPUTING, 2025, 634
  • [4] A Class-Added Continual Learning Method for Motor Fault Diagnosis Based on Knowledge Distillation of Representation Proximity Behavior
    Ding, Ao
    Qin, Yong
    Wang, Biao
    Jia, Limin
    2023 IEEE INTERNATIONAL CONFERENCE ON PROGNOSTICS AND HEALTH MANAGEMENT, ICPHM, 2023, : 224 - 231
  • [5] Variational Data-Free Knowledge Distillation for Continual Learning
    Li, Xiaorong
    Wang, Shipeng
    Sun, Jian
    Xu, Zongben
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 12618 - 12634
  • [6] Subspace distillation for continual learning
    Roy, Kaushik
    Simon, Christian
    Moghadam, Peyman
    Harandi, Mehrtash
    NEURAL NETWORKS, 2023, 167 : 65 - 79
  • [7] FedRCIL: Federated Knowledge Distillation for Representation based Contrastive Incremental Learning
    Psaltis, Athanasios
    Chatzikonstantinou, Christos
    Patrikakis, Charalampos Z.
    Daras, Petros
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 3455 - 3464
  • [8] Continual Unsupervised Representation Learning
    Rao, Dushyant
    Visin, Francesco
    Rusu, Andrei A.
    Teh, Yee Whye
    Pascanu, Razvan
    Hadsell, Raia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] Representation Learning and Knowledge Distillation for Lightweight Domain Adaptation
    Bin Shah, Sayed Rafay
    Putty, Shreyas Subhash
    Schwung, Andreas
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 1202 - 1207
  • [10] Disentangle-based Continual Graph Representation Learning
    Kou, Xiaoyu
    Lin, Yankai
    Liu, Shaobo
    Li, Peng
    Zhou, Jie
    Zhang, Yan
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 2961 - 2972