Dense Network Expansion for Class Incremental Learning

被引:14
|
作者
Hu, Zhiyuan [1 ]
Li, Yunsheng [2 ]
Lyu, Jiancheng [3 ]
Gao, Dashan [3 ]
Vasconcelos, Nuno [1 ]
机构
[1] Univ Calif San Diego, San Diego, CA 92093 USA
[2] Microsoft Cloud AI, Redmond, WA USA
[3] Qualcomm AI Res, San Diego, CA USA
关键词
D O I
10.1109/CVPR52729.2023.01141
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The problem of class incremental learning (CIL) is considered. State-of-the-art approaches use a dynamic architecture based on network expansion (NE), in which a task expert is added per task. While effective from a computational standpoint, these methods lead to models that grow quickly with the number of tasks. A new NE method, dense network expansion (DNE), is proposed to achieve a better trade-off between accuracy and model complexity. This is accomplished by the introduction of dense connections between the intermediate layers of the task expert networks, that enable the transfer of knowledge from old to new tasks via feature sharing and reusing. This sharing is implemented with a cross-task attention mechanism, based on a new task attention block (TAB), that fuses information across tasks. Unlike traditional attention mechanisms, TAB operates at the level of the feature mixing and is decoupled with spatial attentions. This is shown more effective than a joint spatial-and-task attention for CIL. The proposed DNE approach can strictly maintain the feature space of old classes while growing the network and feature scale at a much slower rate than previous methods. In result, it outperforms the previous SOTA methods by a margin of 4% in terms of accuracy, with similar or even smaller model scale.
引用
收藏
页码:11858 / 11867
页数:10
相关论文
共 50 条
  • [1] Rebalancing network with knowledge stability for class incremental learning
    Song, Jialun
    Chen, Jian
    Du, Lan
    PATTERN RECOGNITION, 2024, 153
  • [2] A Broad Neural Network Structure for Class Incremental Learning
    Liu, Wenzhang
    Yang, Haiqin
    Sun, Yuewen
    Sun, Changyin
    ADVANCES IN NEURAL NETWORKS - ISNN 2018, 2018, 10878 : 229 - 238
  • [3] Resolving Task Confusion in Dynamic Expansion Architectures for Class Incremental Learning
    Huang, Bingchen
    Chen, Zhineng
    Zhou, Peng
    Chen, Jiayin
    Wu, Zuxuan
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 908 - 916
  • [5] Model Attention Expansion for Few-Shot Class-Incremental Learning
    Wang, Xuan
    Ji, Zhong
    Yu, Yunlong
    Pang, Yanwei
    Han, Jungong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4419 - 4431
  • [6] Class Incremental Learning Based on Playback Images Generated by Classification Network
    Zhu, Qiuyu
    Zhang, Yunxiao
    Zhuo, Yunhang
    Chen, Junli
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2025,
  • [7] Dynamic Support Network for Few-Shot Class Incremental Learning
    Yang, Boyu
    Lin, Mingbao
    Zhang, Yunxiao
    Liu, Binghao
    Liang, Xiaodan
    Ji, Rongrong
    Ye, Qixiang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 2945 - 2951
  • [8] Essentials for Class Incremental Learning
    Mittal, Sudhanshu
    Galesso, Silvio
    Brox, Thomas
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3508 - 3517
  • [9] Class-Incremental Exemplar Compression for Class-Incremental Learning
    Luo, Zilin
    Liu, Yaoyao
    Schiele, Bernt
    Sun, Qianru
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11371 - 11380
  • [10] Generalized Class Incremental Learning
    Mi, Fei
    Kong, Lingjing
    Lin, Tao
    Yu, Kaicheng
    Faltings, Boi
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 970 - 974