Continuous transfer of neural network representational similarity for incremental learning

被引:43
|
作者
Tian, Songsong [1 ,2 ]
Li, Weijun [1 ,3 ,4 ]
Ning, Xin [1 ,3 ,4 ,5 ]
Ran, Hang [1 ]
Qin, Hong [1 ,3 ,4 ]
Tiwari, Prayag [6 ]
机构
[1] Chinese Acad Sci, Inst Semicond, Beijing 100083, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100049, Peoples R China
[3] Univ Chinese Acad Sci, Ctr Mat Sci & Optoelect Engn, Beijing 100049, Peoples R China
[4] Univ Chinese Acad Sci, Sch Integrated Circuits, Beijing 100049, Peoples R China
[5] Zhongke Ruitu Technol Co Ltd, Beijing 100096, Peoples R China
[6] Halmstad Univ, Sch Informat Technol, S-30118 Halmstad, Sweden
关键词
Incremental learning; Pre-trained model; Knowledge distillation; Neural network representation;
D O I
10.1016/j.neucom.2023.126300
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The incremental learning paradigm in machine learning has consistently been a focus of academic research. It is similar to the way in which biological systems learn, and reduces energy consumption by avoiding excessive retraining. Existing studies utilize the powerful feature extraction capabilities of pre-trained models to address incremental learning, but there remains a problem of insufficient utiliza-tion of neural network feature knowledge. To address this issue, this paper proposes a novel method called Pre-trained Model Knowledge Distillation (PMKD) which combines knowledge distillation of neu-ral network representations and replay. This paper designs a loss function based on centered kernel align-ment to transfer neural network representations knowledge from the pre-trained model to the incremental model layer-by-layer. Additionally, the use of memory buffer for Dark Experience Replay helps the model retain past knowledge better. Experiments show that PMKD achieved superior perfor-mance on various datasets and different buffer sizes. Compared to other methods, our class incremental learning accuracy reached the best performance. The open-source code is published athttps://github.-com/TianSongS/PMKD-IL.(c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Continuous neural network with windowed Hebbian learning
    Fotouhi, M.
    Heidari, M.
    Sharifitabar, M.
    BIOLOGICAL CYBERNETICS, 2015, 109 (03) : 321 - 332
  • [42] On Pattern Generating Methods for Incremental Learning by Chaotic Neural Network
    Deguchi, Toshinori
    Ishii, Naohiro
    2016 4TH INTL CONF ON APPLIED COMPUTING AND INFORMATION TECHNOLOGY/3RD INTL CONF ON COMPUTATIONAL SCIENCE/INTELLIGENCE AND APPLIED INFORMATICS/1ST INTL CONF ON BIG DATA, CLOUD COMPUTING, DATA SCIENCE & ENGINEERING (ACIT-CSII-BCD), 2016, : 271 - 276
  • [43] Neural Network Knowledge Transfer using Unsupervised Similarity Matching
    Passalis, Nikolaos
    Tefas, Anastasios
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 716 - 721
  • [44] Continuous Learning in a Hierarchical Multiscale Neural Network
    Wolf, Thomas
    Chaumond, Julien
    Delangue, Clement
    PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, 2018, : 1 - 7
  • [45] Continuous neural network with windowed Hebbian learning
    M. Fotouhi
    M. Heidari
    M. Sharifitabar
    Biological Cybernetics, 2015, 109 : 321 - 332
  • [46] Distinct contributions of functional and deep neural network features to representational similarity of scenes in human brain and behavior
    Groen, Iris I. A.
    Greene, Michelle R.
    Baldassano, Christopher
    Li Fei-Fei
    Beck, Diane M.
    Baker, Chris I.
    ELIFE, 2018, 7
  • [47] Predicting behavior change from persuasive messages using neural representational similarity and social network analyses
    Pegors, Teresa K.
    Tompson, Steven
    O'Donnell, Matthew Brook
    Falk, Emily B.
    NEUROIMAGE, 2017, 157 : 118 - 128
  • [48] Neural Coding of Cognitive Control: The Representational Similarity Analysis Approach
    Freund, Michael C.
    Etzel, Joset A.
    Braver, Todd S.
    TRENDS IN COGNITIVE SCIENCES, 2021, 25 (07) : 622 - 638
  • [49] A Bayesian method for reducing bias in neural representational similarity analysis
    Cai, Ming Bo
    Schuck, Nicolas W.
    Pillow, Jonathan W.
    Niv, Yael
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [50] Content and cluster analysis: assessing representational similarity in neural systems
    Laakso, A
    Cottrell, G
    PHILOSOPHICAL PSYCHOLOGY, 2000, 13 (01) : 47 - 76