A Communication-Efficient Hierarchical Federated Learning Framework via Shaping Data Distribution at Edge

被引:7
|
作者
Deng, Yongheng [1 ]
Lyu, Feng [2 ]
Xia, Tengxi [1 ]
Zhou, Yuezhi [3 ]
Zhang, Yaoxue [1 ,3 ]
Ren, Ju [1 ,3 ]
Yang, Yuanyuan [4 ]
机构
[1] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol BNRist, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Cent South Univ, Sch Comp Sci & Engn, Changsha 410083, Peoples R China
[3] Zhongguancun Lab, Beijing 100084, Peoples R China
[4] SUNY Stony Brook, Dept Elect & Comp Engn, Stony Brook, NY 11794 USA
关键词
Costs; Data models; Servers; Computational modeling; Training data; Federated learning; Distributed databases; Hierarchical federated learning; communication efficiency; edge computing; distributed edge intelligence; RESOURCE-ALLOCATION;
D O I
10.1109/TNET.2024.3363916
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) enables collaborative model training over distributed computing nodes without sharing their privacy-sensitive raw data. However, in FL, iterative exchanges of model updates between distributed nodes and the cloud server can result in significant communication cost, especially when the data distributions at distributed nodes are imbalanced with requiring more rounds of iterations. In this paper, with our in-depth empirical studies, we disclose that extensive cloud aggregations can be avoided without compromising the learning accuracy if frequent aggregations can be enabled at edge network. To this end, we shed light on the hierarchical federated learning (HFL) framework, where a subset of distributed nodes can play as edge aggregators to support edge aggregations. Under the HFL framework, we formulate a communication cost minimization (CCM) problem to minimize the total communication cost required for model learning with a target accuracy by making decisions on edge aggragator selection and node-edge associations. Inspired by our data-driven insights that the potential of HFL lies in the data distribution at edge aggregators, we propose ShapeFL, i.e., SHaping dAta distRibution at Edge, to transform and solve the CCM problem. In ShapeFL, we divide the original problem into two sub-problems to minimize the per-round communication cost and maximize the data distribution diversity of edge aggregator data, respectively, and devise two light-weight algorithms to solve them accordingly. Extensive experiments are carried out based on several opened datasets and real-world network topologies, and the results demonstrate the efficacy of ShapeFL in terms of both learning accuracy and communication efficiency.
引用
收藏
页码:2600 / 2615
页数:16
相关论文
共 50 条
  • [1] SHARE: Shaping Data Distribution at Edge for Communication-Efficient Hierarchical Federated Learning
    Deng, Yongheng
    Lyu, Feng
    Ren, Ju
    Zhang, Yongmin
    Zhou, Yuezhi
    Zhang, Yaoxue
    Yang, Yuanyuan
    2021 IEEE 41ST INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2021), 2021, : 24 - 34
  • [2] Communication-efficient and Scalable Decentralized Federated Edge Learning
    Yapp, Austine Zong Han
    Koh, Hong Soo Nicholas
    Lai, Yan Ting
    Kang, Jiawen
    Li, Xuandi
    Ng, Jer Shyuan
    Jiang, Hongchao
    Lim, Wei Yang Bryan
    Xiong, Zehui
    Niyato, Dusit
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 5032 - 5035
  • [3] Communication-efficient hierarchical federated learning for IoT heterogeneous systems with imbalanced data
    Abdellatif, Alaa Awad
    Mhaisen, Naram
    Mohamed, Amr
    Erbad, Aiman
    Guizani, Mohsen
    Dawy, Zaher
    Nasreddine, Wassim
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2022, 128 : 406 - 419
  • [4] Communication-Efficient Federated Edge Learning via Optimal Probabilistic Device Scheduling
    Zhang, Maojun
    Zhu, Guangxu
    Wang, Shuai
    Jiang, Jiamo
    Liao, Qing
    Zhong, Caijun
    Cui, Shuguang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (10) : 8536 - 8551
  • [5] Communication-efficient federated learning method via redundant data elimination
    Li K.
    Xu Q.
    Wang H.
    Tongxin Xuebao/Journal on Communications, 2023, 44 (05): : 79 - 93
  • [6] FlocOff: Data Heterogeneity Resilient Federated Learning With Communication-Efficient Edge Offloading
    Ma, Mulei
    Gong, Chenyu
    Zeng, Liekang
    Yang, Yang
    Wu, Liantao
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2024, 42 (11) : 3262 - 3277
  • [7] Communication-efficient federated learning
    Chen, Mingzhe
    Shlezinger, Nir
    Poor, H. Vincent
    Eldar, Yonina C.
    Cui, Shuguang
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2021, 118 (17)
  • [8] Communication-efficient federated learning via knowledge distillation
    Wu, Chuhan
    Wu, Fangzhao
    Lyu, Lingjuan
    Huang, Yongfeng
    Xie, Xing
    NATURE COMMUNICATIONS, 2022, 13 (01)
  • [9] Communication-Efficient Federated Learning via Predictive Coding
    Yue, Kai
    Jin, Richeng
    Wong, Chau-Wai
    Dai, Huaiyu
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (03) : 369 - 380
  • [10] CBFL: A Communication-Efficient Federated Learning Framework From Data Redundancy Perspective
    Li, Kaiju
    Xiao, Chunhua
    IEEE SYSTEMS JOURNAL, 2022, 16 (04): : 5572 - 5583