Beyond Not-Forgetting: Continual Learning with Backward Knowledge Transfer

被引:0
|
作者
Lin, Sen [1 ]
Yang, Li [1 ]
Fan, Deliang [1 ]
Zhang, Junshan [2 ]
机构
[1] Arizona State Univ, Sch ECEE, Tempe, AZ 85287 USA
[2] Univ Calif Davis, Dept ECE, Davis, CA 95616 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
By learning a sequence of tasks continually, an agent in continual learning (CL) can improve the learning performance of both a new task and 'old' tasks by leveraging the forward knowledge transfer and the backward knowledge transfer, respectively. However, most existing CL methods focus on addressing catastrophic forgetting in neural networks by minimizing the modification of the learnt model for old tasks. This inevitably limits the backward knowledge transfer from the new task to the old tasks, because judicious model updates could possibly improve the learning performance of the old tasks as well. To tackle this problem, we first theoretically analyze the conditions under which updating the learnt model of old tasks could be beneficial for CL and also lead to backward knowledge transfer, based on the gradient projection onto the input subspaces of old tasks. Building on the theoretical analysis, we next develop a ContinUal learning method with Backward knowlEdge tRansfer (CUBER), for a fixed capacity neural network without data replay. In particular, CUBER first characterizes the task correlation to identify the positively correlated old tasks in a layer-wise manner, and then selectively modifies the learnt model of the old tasks when learning the new task. Experimental studies show that CUBER can even achieve positive backward knowledge transfer on several existing CL benchmarks for the first time without data replay, where the related baselines still suffer from catastrophic forgetting (negative backward knowledge transfer). The superior performance of CUBER on the backward knowledge transfer also leads to higher accuracy accordingly.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Adversarial Targeted Forgetting in Regularization and Generative Based Continual Learning Models
    Umer, Muhammad
    Polikar, Robi
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [42] Continual Deep Reinforcement Learning to Prevent Catastrophic Forgetting in Jamming Mitigation
    Davaslioglu, Kemal
    Kompella, Sastry
    Erpek, Tugba
    Sagduyu, Yalin E.
    arXiv,
  • [43] Preventing Catastrophic Forgetting in Continual Learning of New Natural Language Tasks
    Kar, Sudipta
    Castellucci, Giuseppe
    Filice, Simone
    Malmasi, Shervin
    Rokhlenko, Oleg
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 3137 - 3145
  • [44] Overcoming Catastrophic Forgetting in Continual Learning by Exploring Eigenvalues of Hessian Matrix
    Kong, Yajing
    Liu, Liu
    Chen, Huanhuan
    Kacprzyk, Janusz
    Tao, Dacheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 16196 - 16210
  • [45] Continual Learning for Visual Search with Backward Consistent Feature Embedding
    Wan, Timmy S. T.
    Chen, Jun-Cheng
    Wu, Tzer-Yi
    Chen, Chu-Song
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16681 - 16690
  • [46] Preserving Linear Separability in Continual Learning by Backward Feature Projection
    Gu, Qiao
    Shim, Dongsub
    Shkurti, Florian
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 24286 - 24295
  • [47] Improving Replay Sample Selection and Storage for Less Forgetting in Continual Learning
    Brignac, Daniel
    Lobo, Niels
    Mahalanobis, Abhijit
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 3532 - 3541
  • [48] Selective Amnesia: A Continual Learning Approach to Forgetting in Deep Generative Models
    Heng, Alvin
    Soh, Harold
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [49] Formalizing the Generalization-Forgetting Trade-Off in Continual Learning
    Krishnan, R.
    Balaprakash, Prasanna
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [50] Dynamic memory to alleviate catastrophic forgetting in continual learning with medical imaging
    Perkonigg, Matthias
    Hofmanninger, Johannes
    Herold, Christian J.
    Brink, James A.
    Pianykh, Oleg
    Prosch, Helmut
    Langs, Georg
    NATURE COMMUNICATIONS, 2021, 12 (01)