Training Networks in Null Space of Feature Covariance With Self-Supervision for Incremental Learning

被引:0
|
作者
Wang, Shipeng [1 ,2 ]
Li, Xiaorong [2 ]
Sun, Jian [2 ]
Xu, Zongben [3 ,4 ]
机构
[1] Xran Jiaotong Univ, Key Lab Biomed Informat Engn, Minist Educ, Sch Life Sci & Technol, Xian 710049, Peoples R China
[2] Xi An Jiao Tong Univ, Sch Math & Stat, Xian 710049, Peoples R China
[3] Xi An Jiao Tong Univ, Sch Math & Stat, Xian 710049, Peoples R China
[4] Pazhou Lab Huangpu Guangzhou, Guangzhou 510555, Guangdong, Peoples R China
关键词
Incremental learning; Vectors; Null space; Covariance matrices; Knowledge engineering; Training; Approximation algorithms; Training data; Stability plasticity; Generative adversarial networks; Catastrophic forgetting; continual learning; null space; self-supervision; stability-plasticity dilemma;
D O I
10.1109/TPAMI.2024.3522258
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the context of incremental learning, a network is sequentially trained on a stream of tasks, where data from previous tasks are particularly assumed to be inaccessible. The major challenge is how to overcome the stability-plasticity dilemma, i.e., learning knowledge from new tasks without forgetting the knowledge of previous tasks. To this end, we propose two mathematical conditions for guaranteeing network stability and plasticity with theoretical analysis. The conditions demonstrate that we can restrict the parameter update in the null space of uncentered feature covariance at each linear layer to overcome the stability-plasticity dilemma, which can be realized by layerwise projecting gradient into the null space. Inspired by it, we develop two algorithms, dubbed Adam-NSCL and Adam-SFCL respectively, for incremental learning. Adam-NSCL and Adam-SFCL provide different ways to compute the projection matrix. The projection matrix in Adam-NSCL is constructed by singular vectors associated with the smallest singular values of the uncentered feature covariance matrix, while the projection matrix in Adam-SFCL is constructed by all singular vectors associated with adaptive scaling factors. Additionally, we explore adopting self-supervised techniques, including self-supervised label augmentation and a newly proposed contrastive loss, to improve the performance of incremental learning. These self-supervised techniques are orthogonal to Adam-NSCL and Adam-SFCL and can be incorporated with them seamlessly, leading to Adam-NSCL-SSL and Adam-SFCL-SSL respectively. The proposed algorithms are applied to task-incremental and class-incremental learning on various benchmark datasets with multiple backbones, and the results show that they outperform the compared incremental learning methods.
引用
收藏
页码:2563 / 2580
页数:18
相关论文
共 50 条
  • [1] Training Networks in Null Space of Feature Covariance for Continual Learning
    Wang, Shipeng
    Li, Xiaorong
    Sun, Jian
    Xu, Zongben
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 184 - 193
  • [2] Hybrid rotation self-supervision and feature space normalization for class incremental learning
    Feng, Wenyi
    Wang, Zhe
    Zhang, Qian
    Gong, Jiayi
    Xu, Xinlei
    Fu, Zhilin
    INFORMATION SCIENCES, 2025, 691
  • [3] Prototype Augmentation and Self-Supervision for Incremental Learning
    Zhu, Fei
    Zhang, Xu-Yao
    Wang, Chuang
    Yin, Fei
    Liu, Cheng-Lin
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 5867 - 5876
  • [4] Semantic alignment with self-supervision for class incremental learning
    Fu, Zhiling
    Wang, Zhe
    Xu, Xinlei
    Yang, Mengping
    Chi, Ziqiu
    Ding, Weichao
    KNOWLEDGE-BASED SYSTEMS, 2023, 282
  • [5] Rethinking Self-Supervision for Few-Shot Class-Incremental Learning
    Zhao, Linglan
    Lu, Jing
    Cheng, Zhanzhan
    Liu, Duo
    Fang, Xiangzhong
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 726 - 731
  • [6] Feature propagation as self-supervision signals on graphs
    Pina, Oscar
    Vilaplana, Veronica
    KNOWLEDGE-BASED SYSTEMS, 2024, 289
  • [7] Hyperspherically regularized networks for self-supervision
    Durrant, Aiden
    Leontidis, Georgios
    IMAGE AND VISION COMPUTING, 2022, 124
  • [8] Tailoring Self-Supervision for Supervised Learning
    Moon, WonJun
    Kim, Ji-Hwan
    Heo, Jae-Pil
    COMPUTER VISION, ECCV 2022, PT XXV, 2022, 13685 : 346 - 364
  • [9] Learning with self-supervision on EEG data
    Gramfort, Alexandre
    Banville, Hubert
    Chehab, Omar
    Hyvarinen, Aapo
    Engemann, Denis
    2021 9TH IEEE INTERNATIONAL WINTER CONFERENCE ON BRAIN-COMPUTER INTERFACE (BCI), 2021, : 28 - 29
  • [10] Hyperspherically regularized networks for self-supervision
    Durrant, Aiden
    Leontidis, Georgios
    Image and Vision Computing, 2022, 124