Overcoming Forgetting Catastrophe in Quantization-Aware Training

被引:0
|
作者
Chen, Ting-An [1 ,2 ]
Yang, De-Nian [2 ,3 ]
Chen, Ming-Syan [1 ,3 ]
机构
[1] Natl Taiwan Univ, Grad Inst Elect Engn, Taipei, Taiwan
[2] Acad Sinica, Inst Informat Sci, Taipei, Taiwan
[3] Acad Sinica, Res Ctr Informat Technol Innovat, Taipei, Taiwan
关键词
D O I
10.1109/ICCV51070.2023.01592
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Quantization is an effective approach for memory cost reduction by compressing networks to lower bits. However, existing quantization processes learned only from the current data tend to suffer from forgetting catastrophe on streaming data, i.e., significant performance decrement on old task data after being trained on new tasks. Therefore, we propose a lifelong quantization process, LifeQuant, to address the problem. We theoretically analyze the forgetting catastrophe from the shift of quantization search space with the change of data tasks. To overcome the forgetting catastrophe, we first minimize the space shift during quantization and propose Proximal Quantization Space Search (ProxQ), for regularizing the search space during quantization to be close to a pre-defined standard space. Afterward, we exploit replay data (a subset of old task data) for retraining in new tasks to alleviate the forgetting problem. However, the limited amount of replay data usually leads to biased quantization performance toward the new tasks. To address the imbalance issue, we design a Balanced Lifelong Learning (BaLL) Loss to reweight (to increase) the influence of replay data in new task learning, by leveraging the class distributions. Experimental results show that LifeQuant achieves outstanding accuracy performance with a low forgetting rate.
引用
收藏
页码:17312 / 17321
页数:10
相关论文
共 50 条
  • [1] Overcoming Oscillations in Quantization-Aware Training
    Nagel, Markus
    Fournarakis, Marios
    Bondarenko, Yelysei
    Blankevoort, Tijmen
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [2] Quantization-Aware Training With Dynamic and Static Pruning
    An, Sangho
    Shin, Jongyun
    Kim, Jangho
    IEEE ACCESS, 2025, 13 : 57476 - 57484
  • [3] Quantization-aware Training for Multi-Agent Reinforcement Learning
    Chandrinos, Nikolaos
    Amasialidis, Michalis
    Kirtas, Manos
    Tsampazis, Konstantinos
    Passalis, Nikolaos
    Tefas, Anastasios
    32ND EUROPEAN SIGNAL PROCESSING CONFERENCE, EUSIPCO 2024, 2024, : 1891 - 1895
  • [4] AdaQAT: Adaptive Bit-Width Quantization-Aware Training
    Gernigon, Cedric
    Filip, Silviu-Ioan
    Sentieys, Olivier
    Coggiola, Clement
    Bruno, Mickael
    2024 IEEE 6TH INTERNATIONAL CONFERENCE ON AI CIRCUITS AND SYSTEMS, AICAS 2024, 2024, : 442 - 446
  • [5] A Robust, Quantization-Aware Training Method for Photonic Neural Networks
    Oikonomou, A.
    Kirtas, M.
    Passalis, N.
    Mourgias-Alexandris, G.
    Moralis-Pegios, M.
    Pleros, N.
    Tefas, A.
    ENGINEERING APPLICATIONS OF NEURAL NETWORKS, EAAAI/EANN 2022, 2022, 1600 : 427 - 438
  • [6] Disentangled Loss for Low-Bit Quantization-Aware Training
    Allenet, Thibault
    Briand, David
    Bichler, Olivier
    Sentieys, Olivier
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 2787 - 2791
  • [7] Optimal Clipping and Magnitude-aware Differentiation for Improved Quantization-aware Training
    Sakr, Charbel
    Dai, Steve
    Venkatesan, Rangharajan
    Zimmer, Brian
    Dally, William J.
    Khailany, Brucek
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 19123 - 19138
  • [8] Approximation- and Quantization-Aware Training for Graph Neural Networks
    Novkin, Rodion
    Klemme, Florian
    Amrouch, Hussam
    IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (02) : 599 - 612
  • [9] Quantization-aware phase retrieval
    Mukherjee, Subhadip
    Seelamantula, Chandra Sekhar
    INTERNATIONAL JOURNAL OF WAVELETS MULTIRESOLUTION AND INFORMATION PROCESSING, 2022, 20 (03)
  • [10] Quantization-aware training for low precision photonic neural networks
    Kirtas, M.
    Oikonomou, A.
    Passalis, N.
    Mourgias-Alexandris, G.
    Moralis-Pegios, M.
    Pleros, N.
    Tefas, A.
    NEURAL NETWORKS, 2022, 155 : 561 - 573