Memory efficient data-free distillation for continual learning

被引:4
|
作者
Li, Xiaorong [1 ]
Wang, Shipeng [1 ]
Sun, Jian [1 ]
Xu, Zongben [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Math & Stat, Xian, Shaanxi, Peoples R China
基金
国家重点研发计划;
关键词
Continual learning; Catastrophic forgetting; Knowledge distillation;
D O I
10.1016/j.patcog.2023.109875
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks suffer from the catastrophic forgetting phenomenon when trained on sequential tasks in continual learning, especially when data from previous tasks are unavailable. To mitigate catastrophic forgetting, various methods either store data from previous tasks, which may raise privacy concerns, or require large memory storage. Particularly, the distillation-based methods mitigate catastrophic forgetting by using proxy datasets. However, proxy datasets may not match the distributions of the original datasets of previous tasks. To address these problems in a setting where the full training data of previous tasks are unavailable and memory resources are limited, we propose a novel data-free distillation method. Our method encodes knowledge of previous tasks into network parameter gradients by Taylor expansion, deducing a regularizer relying on gradients in network training loss. To improve memory efficiency, we design an approach to compressing the gradients in the regularizer. Moreover, we theoretically analyze the approximation error of our method. Experimental results on multiple datasets demonstrate that our proposed method outperforms the existing approaches in continual learning.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Up to 100 x Faster Data-Free Knowledge Distillation
    Fang, Gongfan
    Mo, Kanya
    Wang, Xinchao
    Song, Jie
    Bei, Shitao
    Zhang, Haofei
    Song, Mingli
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 6597 - 6604
  • [32] Dual discriminator adversarial distillation for data-free model compression
    Zhao, Haoran
    Sun, Xin
    Dong, Junyu
    Manic, Milos
    Zhou, Huiyu
    Yu, Hui
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2022, 13 (05) : 1213 - 1230
  • [33] Double-Generators Network for Data-Free Knowledge Distillation
    Zhang J.
    Ju J.
    Ren Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2023, 60 (07): : 1615 - 1627
  • [34] Effectiveness of Arbitrary Transfer Sets for Data-free Knowledge Distillation
    Nayak, Gaurav Kumar
    Mopuri, Konda Reddy
    Chakraborty, Anirban
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 1429 - 1437
  • [35] Dual discriminator adversarial distillation for data-free model compression
    Haoran Zhao
    Xin Sun
    Junyu Dong
    Milos Manic
    Huiyu Zhou
    Hui Yu
    International Journal of Machine Learning and Cybernetics, 2022, 13 : 1213 - 1230
  • [36] Unpacking the Gap Box Against Data-Free Knowledge Distillation
    Wang, Yang
    Qian, Biao
    Liu, Haipeng
    Rui, Yong
    Wang, Meng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 6280 - 6291
  • [37] Frequency Domain Distillation for Data-Free Quantization of Vision Transformer
    Nan, Gongrui
    Chao, Fei
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 : 205 - 216
  • [38] Model Conversion via Differentially Private Data-Free Distillation
    Liu, Bochao
    Wang, Pengju
    Li, Shikun
    Zeng, Dan
    Ge, Shiming
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 2187 - 2195
  • [39] Augmented Geometric Distillation for Data-Free Incremental Person ReID
    Lu, Yichen
    Wang, Mei
    Deng, Weihong
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7319 - 7328
  • [40] Data-free Knowledge Distillation based on GNN for Node Classification
    Zeng, Xinfeng
    Liu, Tao
    Zeng, Ming
    Wu, Qingqiang
    Wang, Meihong
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2024, PT 2, 2025, 14851 : 243 - 258