Heterogeneity-Aware Memory Efficient Federated Learning via Progressive Layer Freezing

被引:0
|
作者
Wu, Yebo [1 ]
Li, Li [1 ]
Tian, Chunlin [1 ]
Chang, Tao [2 ]
Lin, Chi [3 ]
Wang, Cong [4 ]
Xu, Cheng-Zhong [1 ]
机构
[1] Univ Macau, State Key Lab IoTSC, Taipa, Macao, Peoples R China
[2] Natl Univ Def Technol, Changsha, Peoples R China
[3] Dalian Univ Technol, Dalian, Peoples R China
[4] Zhejiang Univ, Hangzhou, Peoples R China
关键词
Federated Learning; On-Device Training; Heterogeneous Memory;
D O I
10.1109/IWQoS61813.2024.10682916
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning (FL) emerges as a new learning paradigm that enables multiple devices to collaboratively train a shared model while preserving data privacy. However, intensive memory footprint during the training process severely bottlenecks the deployment of FL on resource-limited mobile devices in real-world cases. Thus, a framework that can effectively reduce the memory footprint while guaranteeing training efficiency and model accuracy is crucial for FL. In this paper, we propose SmartFreeze, a framework that effectively reduces the memory footprint by conducting the training in a progressive manner. Instead of updating the full model in each training round, SmartFreeze divides the shared model into blocks consisting of a specified number of layers. It first trains the front block with a well-designed output module, safely freezes it after convergence, and then triggers the training of the next one. This process iterates until the whole model has been successfully trained. In this way, the backward computation of the frozen blocks and the corresponding memory space for storing the intermediate outputs and gradients are effectively saved. Except for the progressive training framework, SmartFreeze consists of the following two core components: a pace controller and a participant selector. The pace controller is designed to effectively monitor the training progress of each block at runtime and safely freezes them after convergence while the participant selector selects the right devices to participate in the training for each block by jointly considering the memory capacity, the statistical and system heterogeneity. Extensive experiments are conducted to evaluate the effectiveness of SmartFreeze on both simulation and hardware testbeds. The results demonstrate that SmartFreeze effectively reduces average memory usage by up to 82%. Moreover, it simultaneously improves the model accuracy by up to 83.1% and accelerates the training process up to 2.02x.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] SCHEDTUNE: A Heterogeneity-Aware GPU Scheduler for Deep Learning
    Albahar, Hadeel
    Dongare, Shruti
    Du, Yanlin
    Zhao, Nannan
    Paul, Arnab K.
    Butt, Ali R.
    2022 22ND IEEE/ACM INTERNATIONAL SYMPOSIUM ON CLUSTER, CLOUD AND INTERNET COMPUTING (CCGRID 2022), 2022, : 695 - 705
  • [32] Personalized Heterogeneity-aware Federated Search Towards Better Accuracy and Energy Efficiency
    Yang, Zhao
    Sun, Qingshuang
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [33] Heterogeneity-aware Cross-school Electives Recommendation: a Hybrid Federated Approach
    Ju, Chengyi
    Cao, Jiannong
    Yang, Yu
    Yang, Zhen-Qun
    Lee, Ho Man
    2023 23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW 2023, 2023, : 1500 - 1508
  • [34] Heterogeneity-aware and communication-efficient distributed statistical inference
    Duan, Rui
    Ning, Yang
    Chen, Yong
    BIOMETRIKA, 2022, 109 (01) : 67 - 83
  • [35] Heterogeneity-Aware Cluster Scheduling Policies for Deep Learning Workloads
    Narayanan, Deepak
    Santhanam, Keshav
    Kazhamiaka, Fiodar
    Phanishayee, Amar
    Zaharia, Matei
    PROCEEDINGS OF THE 14TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION (OSDI '20), 2020, : 481 - 498
  • [36] Heterogeneity-aware Deep Learning Workload Deployments on the Computing Continuum
    Bouvier, Thomas
    2021 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2021, : 1027 - 1027
  • [37] Heterogeneity-aware Energy-efficient Clustering (HEC) Technique for WSNs
    Sharma, Sukhwinder
    Bansal, Rakesh Kumar
    Bansal, Savina
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2017, 11 (04): : 1866 - 1888
  • [38] A Memory Heterogeneity-Aware Runtime System for bandwidth-sensitive HPC applications
    Chandrasekar, Kavitha
    Ni, Xiang
    Kale, Laxmikant V.
    2017 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2017, : 1293 - 1300
  • [39] A Heterogeneity-Aware Semi-Decentralized Model for a Lightweight Intrusion Detection System for IoT Networks Based on Federated Learning and BiLSTM
    Alsaleh, Shuroog
    Menai, Mohamed El Bachir
    Al-Ahmadi, Saad
    SENSORS, 2025, 25 (04)
  • [40] HedgeRank: Heterogeneity-Aware, Energy-Efficient Partitioning of Personalized PageRank at the Edge
    Gong, Young-Ho
    MICROMACHINES, 2023, 14 (09)