Continuous transfer of neural network representational similarity for incremental learning

被引:43
|
作者
Tian, Songsong [1 ,2 ]
Li, Weijun [1 ,3 ,4 ]
Ning, Xin [1 ,3 ,4 ,5 ]
Ran, Hang [1 ]
Qin, Hong [1 ,3 ,4 ]
Tiwari, Prayag [6 ]
机构
[1] Chinese Acad Sci, Inst Semicond, Beijing 100083, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100049, Peoples R China
[3] Univ Chinese Acad Sci, Ctr Mat Sci & Optoelect Engn, Beijing 100049, Peoples R China
[4] Univ Chinese Acad Sci, Sch Integrated Circuits, Beijing 100049, Peoples R China
[5] Zhongke Ruitu Technol Co Ltd, Beijing 100096, Peoples R China
[6] Halmstad Univ, Sch Informat Technol, S-30118 Halmstad, Sweden
关键词
Incremental learning; Pre-trained model; Knowledge distillation; Neural network representation;
D O I
10.1016/j.neucom.2023.126300
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The incremental learning paradigm in machine learning has consistently been a focus of academic research. It is similar to the way in which biological systems learn, and reduces energy consumption by avoiding excessive retraining. Existing studies utilize the powerful feature extraction capabilities of pre-trained models to address incremental learning, but there remains a problem of insufficient utiliza-tion of neural network feature knowledge. To address this issue, this paper proposes a novel method called Pre-trained Model Knowledge Distillation (PMKD) which combines knowledge distillation of neu-ral network representations and replay. This paper designs a loss function based on centered kernel align-ment to transfer neural network representations knowledge from the pre-trained model to the incremental model layer-by-layer. Additionally, the use of memory buffer for Dark Experience Replay helps the model retain past knowledge better. Experiments show that PMKD achieved superior perfor-mance on various datasets and different buffer sizes. Compared to other methods, our class incremental learning accuracy reached the best performance. The open-source code is published athttps://github.-com/TianSongS/PMKD-IL.(c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
引用
收藏
页数:11
相关论文
共 50 条
  • [31] A self-structurizing neural network for online incremental learning
    Hasegawa, O
    Shen, R
    SICE 2004 ANNUAL CONFERENCE, VOLS 1-3, 2004, : 2063 - 2069
  • [32] An incremental neural network for non-stationary unsupervised learning
    Furao, S
    Hasegawa, O
    NEURAL INFORMATION PROCESSING, 2004, 3316 : 641 - 646
  • [33] Incremental Neural-Network Learning for Big Fraud Data
    Anowar, Farzana
    Sadaoui, Samira
    2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 3551 - 3557
  • [34] An incremental learning preprocessor for feed-forward neural network
    Piyabute Fuangkhon
    Artificial Intelligence Review, 2014, 41 : 183 - 210
  • [35] Transferred Correlation Learning: An Incremental Scheme for Neural Network Ensembles
    Jiang, Lei
    Zhang, Jian
    Allen, Gabrielle
    2010 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS IJCNN 2010, 2010,
  • [36] A new ARTMAP-based neural network for incremental learning
    Su, Mu-Chun
    Lee, Jonathan
    Hsieh, Kuo-Lung
    NEUROCOMPUTING, 2006, 69 (16-18) : 2284 - 2300
  • [37] Development and research of a neural network alternate incremental learning algorithm
    Orlov, A. A.
    Abramova, E. S.
    COMPUTER OPTICS, 2023, 47 (03) : 491 - +
  • [38] A self-organizing incremental neural network for imbalance learning
    Shao, Yue
    Xu, Baile
    Shen, Furao
    Zhao, Jian
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (13): : 9789 - 9802
  • [39] An Incremental Probabilistic Neural Network for Regression and Reinforcement Learning Tasks
    Heinen, Milton Roberto
    Engel, Paulo Martins
    ARTIFICIAL NEURAL NETWORKS-ICANN 2010, PT II, 2010, 6353 : 170 - 179