FedGAMMA: Federated Learning With Global Sharpness-Aware Minimization

被引:8
|
作者
Dai, Rong [1 ]
Yang, Xun [1 ]
Sun, Yan [2 ]
Shen, Li [3 ]
Tian, Xinmei [1 ]
Wang, Meng [4 ]
Zhang, Yongdong [1 ]
机构
[1] Univ Sci & Technol China, Sch Informat Sci & Technol, Hefei 230026, Peoples R China
[2] Univ Sydney, Sch Comp Sci, Sydney, NSW 2008, Australia
[3] JD Explore Acad, Beijing 100000, Peoples R China
[4] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei 230009, Peoples R China
基金
中国国家自然科学基金;
关键词
Training; Servers; Federated learning; Data models; Minimization; Degradation; Convergence; Client-drift; deep learning; distributed learning; federated learning (FL);
D O I
10.1109/TNNLS.2023.3304453
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Federated learning (FL) is a promising framework for privacy-preserving and distributed training with decentralized clients. However, there exists a large divergence between the collected local updates and the expected global update, which is known as the client drift and mainly caused by heterogeneous data distribution among clients, multiple local training steps, and partial client participation training. Most existing works tackle this challenge based on the empirical risk minimization (ERM) rule, while less attention has been paid to the relationship between the global loss landscape and the generalization ability. In this work, we propose FedGAMMA, a novel FL algorithm with Global sharpness-Aware MiniMizAtion to seek a global flat landscape with high performance. Specifically, in contrast to FedSAM which only seeks the local flatness and still suffers from performance degradation when facing the client-drift issue, we adopt a local varieties control technique to better align each client's local updates to alleviate the client drift and make each client heading toward the global flatness together. Finally, extensive experiments demonstrate that FedGAMMA can substantially outperform several existing FL baselines on various datasets, and it can well address the client-drift issue and simultaneously seek a smoother and flatter global landscape.
引用
收藏
页码:17479 / 17492
页数:14
相关论文
共 50 条
  • [1] Random Sharpness-Aware Minimization
    Liu, Yong
    Mai, Siqi
    Cheng, Minhao
    Chen, Xiangning
    Hsieh, Cho-Jui
    You, Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [2] Friendly Sharpness-Aware Minimization
    Li, Tao
    Zhou, Pan
    He, Zhengbao
    Cheng, Xinwen
    Huang, Xiaolin
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 5631 - 5640
  • [3] Enhancing Sharpness-Aware Minimization by Learning Perturbation Radius
    Wang, Xuehao
    Jiang, Weisen
    Fu, Shuai
    Zhang, Yu
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT II, ECML PKDD 2024, 2024, 14942 : 375 - 391
  • [4] Convergence of Sharpness-Aware Minimization with Momentum
    Pham Duy Khanh
    Luong, Hoang-Chau
    Mordukhovich, Boris S.
    Dat Ba Tran
    Truc Vo
    INFORMATION TECHNOLOGIES AND THEIR APPLICATIONS, PT II, ITTA 2024, 2025, 2226 : 123 - 132
  • [5] Towards Understanding Sharpness-Aware Minimization
    Andriushchenko, Maksym
    Flammarion, Nicolas
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022, : 639 - 668
  • [6] Sharpness-Aware Minimization and the Edge of Stability
    Long, Philip M.
    Bartlett, Peter L.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 20
  • [7] Self-adaptive asynchronous federated optimizer with adversarial sharpness-aware minimization
    Zhang, Xiongtao
    Wang, Ji
    Bao, Weidong
    Xiao, Wenhua
    Zhang, Yaohong
    Liu, Lihua
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 161 : 638 - 654
  • [8] Federated Model-Agnostic Meta-Learning With Sharpness-Aware Minimization for Internet of Things Optimization
    Wu, Qingtao
    Zhang, Yong
    Liu, Muhua
    Zhu, Junlong
    Zheng, Ruijuan
    Zhang, Mingchuan
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (19): : 31317 - 31330
  • [9] Noise-resistant sharpness-aware minimization in deep learning
    Su, Dan
    Jin, Long
    Wang, Jun
    NEURAL NETWORKS, 2025, 181
  • [10] Towards Efficient and Scalable Sharpness-Aware Minimization
    Liu, Yong
    Mai, Siqi
    Chen, Xiangning
    Hsieh, Cho-Jui
    You, Yang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12350 - 12360