Measuring Asymmetric Gradient Discrepancy in Parallel Continual Learning

被引:2
|
作者
Lyu, Fan [1 ]
Sun, Qing [1 ]
Shang, Fanhua [1 ]
Wan, Liang [1 ]
Feng, Wei [1 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.01048
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In Parallel Continual Learning (PCL), the parallel multiple tasks start and end training unpredictably, thus suffering from both training conflict and catastrophic forgetting issues. The two issues are raised because the gradients from parallel tasks differ in directions and magnitudes. Thus, in this paper, we formulate the PCL into a minimum distance optimization problem among gradients and propose an explicit Asymmetric Gradient Distance (AGD) to evaluate the gradient discrepancy in PCL. AGD considers both gradient magnitude ratios and directions, and has a tolerance when updating with a small gradient of inverse direction, which reduces the imbalanced influence of gradients on parallel task training. Moreover, we present a novel Maximum Discrepancy Optimization ( MaxDO) strategy to minimize the maximum discrepancy among multiple gradients. Solving by MaxDO with AGD, parallel training reduces the influence of the training conflict and suppresses the catastrophic forgetting of finished tasks. Extensive experiments validate the effectiveness of our approach on three image recognition datasets in task-incremental and class-incremental PCL. Our code is available at https://github.com/fanlyu/maxdo.
引用
收藏
页码:11377 / 11386
页数:10
相关论文
共 50 条
  • [21] Concept Accumulation and Gradient-Guided Adaption for continual learning in evolving streaming
    Xiong, Lin
    Chen, Shanxiong
    Zhou, Hao
    Xiong, Hailing
    NEUROCOMPUTING, 2024, 601
  • [22] UniGrad-FS: Unified Gradient Projection With Flatter Sharpness for Continual Learning
    Li, Wei
    Feng, Tao
    Yuan, Hangjie
    Bian, Ang
    Du, Guodong
    Liang, Sixin
    Gan, Jianhong
    Liu, Ziwei
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (12) : 13873 - 13882
  • [23] Continual learning
    King, Denise
    JOURNAL OF EMERGENCY NURSING, 2008, 34 (04) : 283 - 283
  • [24] CONTINUAL LEARNING
    BROWN, WE
    JOURNAL OF THE AMERICAN DENTAL ASSOCIATION, 1965, 71 (04): : 935 - &
  • [25] PARALLEL GRADIENT BLEND FOR CLASS INCREMENTAL LEARNING
    Zhao, Yunlong
    Deng, Xiaoheng
    Pei, Xinjun
    Chen, Xuechen
    Li, Deng
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1220 - 1224
  • [26] Efficient Data-Parallel Continual Learning with Asynchronous Distributed Rehearsal Buffers
    Bouvier, Thomas
    Nicolae, Bogdan
    Chaugier, Hugo
    Costan, Alexandru
    Foster, Ian
    Antoniu, Gabriel
    2024 IEEE 24TH INTERNATIONAL SYMPOSIUM ON CLUSTER, CLOUD AND INTERNET COMPUTING, CCGRID 2024, 2024, : 245 - 254
  • [27] Psycholinguistics meets Continual Learning: Measuring Catastrophic Forgetting in Visual Question Answering
    Greco, Claudio
    Plank, Barbara
    Fernandez, Raquel
    Bernardi, Raffaella
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3601 - 3605
  • [28] Continual compression model for online continual learning
    Ye, Fei
    Bors, Adrian G.
    APPLIED SOFT COMPUTING, 2024, 167
  • [29] Learning to Prompt for Continual Learning
    Wang, Zifeng
    Zhang, Zizhao
    Lee, Hen Yu
    Zhang, Han
    Sun, Ruoxi
    Ren, Xiaoqi
    Su, Guolong
    Perot, Vincent
    Dy, Jennifer
    Pfister, Tomas
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 139 - 149
  • [30] TIME AND TEMPORAL ABSTRACTION IN CONTINUAL LEARNING: TRADEOFFS, ANALOGIES AND REGRET IN AN ACTIVE MEASURING SETTING
    Letourneau, Vincent
    Bellinger, Colin
    Tamblyn, Isaac
    Fraser, Maia
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 470 - 480