Memory efficient data-free distillation for continual learning

被引:4
|
作者
Li, Xiaorong [1 ]
Wang, Shipeng [1 ]
Sun, Jian [1 ]
Xu, Zongben [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Math & Stat, Xian, Shaanxi, Peoples R China
基金
国家重点研发计划;
关键词
Continual learning; Catastrophic forgetting; Knowledge distillation;
D O I
10.1016/j.patcog.2023.109875
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks suffer from the catastrophic forgetting phenomenon when trained on sequential tasks in continual learning, especially when data from previous tasks are unavailable. To mitigate catastrophic forgetting, various methods either store data from previous tasks, which may raise privacy concerns, or require large memory storage. Particularly, the distillation-based methods mitigate catastrophic forgetting by using proxy datasets. However, proxy datasets may not match the distributions of the original datasets of previous tasks. To address these problems in a setting where the full training data of previous tasks are unavailable and memory resources are limited, we propose a novel data-free distillation method. Our method encodes knowledge of previous tasks into network parameter gradients by Taylor expansion, deducing a regularizer relying on gradients in network training loss. To improve memory efficiency, we design an approach to compressing the gradients in the regularizer. Moreover, we theoretically analyze the approximation error of our method. Experimental results on multiple datasets demonstrate that our proposed method outperforms the existing approaches in continual learning.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] ROBUSTNESS AND DIVERSITY SEEKING DATA-FREE KNOWLEDGE DISTILLATION
    Han, Pengchao
    Park, Jihong
    Wang, Shiqiang
    Liu, Yejun
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2740 - 2744
  • [22] Data-free knowledge distillation in neural networks for regression
    Kang, Myeonginn
    Kang, Seokho
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 175
  • [23] Empirical Study of Data-Free Iterative Knowledge Distillation
    Shah, Het
    Vaswani, Ashwin
    Dash, Tirtharaj
    Hebbalaguppe, Ramya
    Srinivasan, Ashwin
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT III, 2021, 12893 : 546 - 557
  • [24] Data-free Knowledge Distillation for Reusing Recommendation Models
    Wang, Cheng
    Sun, Jiacheng
    Dong, Zhenhua
    Zhu, Jieming
    Li, Zhenguo
    Li, Ruixuan
    Zhang, Rui
    PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023, 2023, : 386 - 395
  • [25] Contrastive Model Inversion for Data-Free Knowledge Distillation
    Fang, Gongfan
    Song, Jie
    Wang, Xinchao
    Shen, Chengchao
    Wang, Xingen
    Song, Mingli
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2374 - 2380
  • [26] FedTAD: Topology-aware Data-free Knowledge Distillation for Subgraph Federated Learning
    Zhu, Yinlin
    Lie, Xunkai
    Wu, Zhengyu
    Wu, Di
    Hu, Miao
    Li, Rong-Hua
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 5716 - 5724
  • [27] Data-Free Learning of Student Networks
    Chen, Hanting
    Wang, Yunhe
    Xu, Chang
    Yang, Zhaohui
    Liu, Chuanjian
    Shi, Boxin
    Xu, Chunjing
    Xu, Chao
    Tian, Qi
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 3513 - 3521
  • [28] Memory Efficient Continual Learning with Transformers
    Ermis, Beyza
    Zappella, Giovanni
    Wistuba, Martin
    Rawal, Aditya
    Archambeau, Cedric
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [29] Data-free knowledge distillation via generator-free data generation for Non-IID federated learning
    Zhao, Siran
    Liao, Tianchi
    Fu, Lele
    Chen, Chuan
    Bian, Jing
    Zheng, Zibin
    NEURAL NETWORKS, 2024, 179
  • [30] Data-Free Knowledge Distillation For Image Super-Resolution
    Zhang, Yiman
    Chen, Hanting
    Chen, Xinghao
    Deng, Yiping
    Xu, Chunjing
    Wang, Yunhe
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 7848 - 7857