Preempting Catastrophic Forgetting in Continual Learning Models by Anticipatory Regularization

被引:0
|
作者
El Khatib, Alaa [1 ]
Karray, Fakhri [1 ]
机构
[1] Univ Waterloo, Elect & Comp Engn, Waterloo, ON, Canada
关键词
D O I
10.1109/ijcnn.2019.8852426
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks trained on tasks sequentially tend to degrade in performance, on the average, the more tasks they see, as the representations learned for one task get progressively modified while learning subsequent tasks. This phenomenon-known as catastrophic forgetting-is a major obstacle on the road toward designing agents that can continually learn new concepts and tasks the way, say, humans do. A common approach to containing catastrophic forgetting is to use regularization to slow down learning on weights deemed important to previously learned tasks. We argue in this paper that, on their own, such post hoc measures to safeguard what has been learned can, even in their more sophisticated variants, paralyze the network and degrade its capacity to learn and counter forgetting as the number of tasks learned increases. We propose insteador possibly in conjunction-that, in anticipation of future tasks, regularization be applied to drive the optimization of network weights toward reusable solutions. We show that one way to achieve this is through an auxiliary unsupervised reconstruction loss that encourages the learned representations not only to be useful for solving, say, the current classification task, but also to reflect the content of the data being processed-content that is generally richer than it is discriminative for any one task. We compare our approach to the recent elastic weight consolidation regularization approach, and show that, although we do not explicitly try to preserve important weights or pass on any information about the data distribution of learned tasks, our model is comparable in performance, and in some cases better.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] Episodic memory based continual learning without catastrophic forgetting for environmental sound classification
    Karam S.
    Ruan S.-J.
    Haq Q.M.
    Li L.P.-H.
    Journal of Ambient Intelligence and Humanized Computing, 2023, 14 (04) : 4439 - 4449
  • [22] Overcoming Catastrophic Forgetting by Bayesian Generative Regularization
    Chen, Patrick H.
    Wei, Wei
    Hsieh, Cho-jui
    Dai, Bo
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [23] Overcoming Catastrophic Forgetting beyond Continual Learning: Balanced Training for Neural Machine Translation
    Shao, Chenze
    Feng, Yang
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2023 - 2036
  • [24] Assessment of catastrophic forgetting in continual credit card fraud detection
    Lebichot, B.
    Siblini, W.
    Paldino, G. M.
    Le Borgne, Y. -A.
    Oble, F.
    Bontempi, G.
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
  • [25] Selective Amnesia: A Continual Learning Approach to Forgetting in Deep Generative Models
    Heng, Alvin
    Soh, Harold
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [26] Example forgetting and rehearsal in continual learning
    Benko, Beatrix
    PATTERN RECOGNITION LETTERS, 2024, 179 : 65 - 72
  • [27] Vaccine Enhanced Continual Learning With TFE to Overcome Catastrophic Forgetting for Variable Speed-Bearing Fault Diagnosis
    Wang, Lu
    Liu, Shulin
    Xiao, Haihua
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (05) : 7112 - 7123
  • [28] Energy Minimum Regularization in Continual Learning
    Li, Xiaobin
    Shan, Lianlei
    Li, Minglong
    Wang, Weiqiang
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 6404 - 6409
  • [29] Latent spectral regularization for continual learning
    Frascaroli, Emanuele
    Benaglia, Riccardo
    Boschini, Matteo
    Moschella, Luca
    Fiorini, Cosimo
    Rodola, Emanuele
    Calderara, Simone
    PATTERN RECOGNITION LETTERS, 2024, 184 : 119 - 125
  • [30] Investigating Catastrophic Forgetting of Deep Learning Models Within Office 31 Dataset
    Hidayaturrahman
    Trisetyarso, Agung
    Kartowisastro, Iman Herwidiana
    Budiharto, Widodo
    IEEE ACCESS, 2024, 12 : 138501 - 138509