FedGAMMA: Federated Learning With Global Sharpness-Aware Minimization

被引:8
|
作者
Dai, Rong [1 ]
Yang, Xun [1 ]
Sun, Yan [2 ]
Shen, Li [3 ]
Tian, Xinmei [1 ]
Wang, Meng [4 ]
Zhang, Yongdong [1 ]
机构
[1] Univ Sci & Technol China, Sch Informat Sci & Technol, Hefei 230026, Peoples R China
[2] Univ Sydney, Sch Comp Sci, Sydney, NSW 2008, Australia
[3] JD Explore Acad, Beijing 100000, Peoples R China
[4] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei 230009, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Servers; Federated learning; Data models; Minimization; Degradation; Convergence; Client-drift; deep learning; distributed learning; federated learning (FL);
D O I
10.1109/TNNLS.2023.3304453
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Federated learning (FL) is a promising framework for privacy-preserving and distributed training with decentralized clients. However, there exists a large divergence between the collected local updates and the expected global update, which is known as the client drift and mainly caused by heterogeneous data distribution among clients, multiple local training steps, and partial client participation training. Most existing works tackle this challenge based on the empirical risk minimization (ERM) rule, while less attention has been paid to the relationship between the global loss landscape and the generalization ability. In this work, we propose FedGAMMA, a novel FL algorithm with Global sharpness-Aware MiniMizAtion to seek a global flat landscape with high performance. Specifically, in contrast to FedSAM which only seeks the local flatness and still suffers from performance degradation when facing the client-drift issue, we adopt a local varieties control technique to better align each client's local updates to alleviate the client drift and make each client heading toward the global flatness together. Finally, extensive experiments demonstrate that FedGAMMA can substantially outperform several existing FL baselines on various datasets, and it can well address the client-drift issue and simultaneously seek a smoother and flatter global landscape.
引用
收藏
页码:17479 / 17492
页数:14
相关论文
共 50 条
  • [21] Binary Quantized Network Training With Sharpness-Aware Minimization
    Liu, Ren
    Bian, Fengmiao
    Zhang, Xiaoqun
    JOURNAL OF SCIENTIFIC COMPUTING, 2023, 94 (01)
  • [22] Detection method of wheat rust based on transfer learning and sharpness-aware minimization
    Xu, Zhengguo
    Pan, Hengtuo
    Ye, Wei
    Xu, Zhuangwei
    Wang, Hongkai
    PLANT PATHOLOGY, 2023, 72 (02) : 353 - 360
  • [23] Binary Quantized Network Training With Sharpness-Aware Minimization
    Ren Liu
    Fengmiao Bian
    Xiaoqun Zhang
    Journal of Scientific Computing, 2023, 94
  • [24] Adaptive Sharpness-Aware Minimization for Adversarial Domain Generalization
    Xie, Tianci
    Li, Tao
    Wu, Ruoxue
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [25] AdaSAM: Boosting sharpness-aware minimization with adaptive learning rate and momentum for neural networks
    Sun, Hao
    Shen, Li
    Zhong, Qihuang
    Ding, Liang
    Chen, Shixiang
    Sun, Jingwei
    Li, Jing
    Sun, Guangzhong
    Tao, Dacheng
    NEURAL NETWORKS, 2024, 169 : 506 - 519
  • [26] CR-SAM: Curvature Regularized Sharpness-Aware Minimization
    Wu, Tao
    Luo, Tie
    Wunsch, Donald C., II
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 6144 - 6152
  • [27] Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach
    Mi, Peng
    Shen, Li
    Ren, Tianhe
    Zhou, Yiyi
    Sun, Xiaoshuai
    Ji, Rongrong
    Tao, Dacheng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [28] Sharpness-Aware Minimization Leads to Low-Rank Features
    Andriushchenko, Maksym
    Bahri, Dara
    Mobahi, Hossein
    Flammarion, Nicolas
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [29] Sharpness-Aware Training for Free
    Du, Jiawei
    Zhou, Daquan
    Feng, Jiashi
    Tan, Vincent Y. F.
    Zhou, Joey Tianyi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [30] ASAM: Adaptive Sharpness-Aware Minimization for Scale-Invariant Learning of Deep Neural Networks
    Kwon, Jungmin
    Kim, Jeongseop
    Park, Hyunseo
    Choi, In Kwon
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139