FedPFT: Federated Proxy Fine-Tuning of Foundation Models

被引:0
|
作者
Peng, Zhaopeng [1 ]
Fan, Xiaoliang [1 ]
Chen, Yufan [1 ]
Wang, Zheng [1 ]
Pan, Shirui [2 ]
Wen, Chenglu [1 ]
Zhang, Ruisheng [3 ]
Wang, Cheng [1 ]
机构
[1] Xiamen Univ, Sch Informat, Fujian Key Lab Sensing & Comp Smart Cities, Xiamen, Fujian, Peoples R China
[2] Griffith Univ, Sch Informat & Commun Technol, Nathan, Qld, Australia
[3] Lanzhou Univ, Sch Informat Sci & Engn, Lanzhou, Gansu, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adapting Foundation Models (FMs) for downstream tasks through Federated Learning (FL) emerges a promising strategy for protecting data privacy and valuable FMs. Existing methods finetune FM by allocating sub-FM to clients in FL, however, leading to suboptimal performance due to insufficient tuning and inevitable error accumulations of gradients. In this paper, we propose Federated Proxy Fine-Tuning (FedPFT), a novel method enhancing FMs adaptation in downstream tasks through FL by two key modules. First, the sub-FM construction module employs a layer-wise compression approach, facilitating comprehensive FM fine-tuning across all layers by emphasizing those crucial neurons. Second, the sub-FM alignment module conducts a two-step distillations-layerlevel and neuron-level-before and during FL finetuning respectively, to reduce error of gradient by accurately aligning sub-FM with FM under theoretical guarantees. Experimental results on seven commonly used datasets (i.e., four text and three vision) demonstrate the superiority of FedPFT. Our code is available at https://github.com/pzpdzd/FedPFT.
引用
收藏
页码:4806 / 4814
页数:9
相关论文
共 50 条
  • [1] Scaling Federated Learning for Fine-Tuning of Large Language Models
    Hilmkil, Agrin
    Callh, Sebastian
    Barbieri, Matteo
    Sutfeld, Leon Rene
    Zec, Edvin Listo
    Mogren, Olof
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2021), 2021, 12801 : 15 - 23
  • [2] Foundation Models and Fine-Tuning: A Benchmark for Out of Distribution Detection
    Cappio Borlino, Francesco
    Lu, Lorenzo
    Tommasi, Tatiana
    IEEE ACCESS, 2024, 12 : 79401 - 79414
  • [3] Federated Fine-Tuning Performance on Edge Devices
    Orescanin, Marko
    Ergezer, Mehmet
    Singh, Gurminder
    Baxter, Matthew
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 1174 - 1181
  • [4] Fine-Tuning Foundation Models With Confidence Assessment for Enhanced Semantic Segmentation
    Dionelis, Nikolaos
    Longepe, Nicolas
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2025, 22
  • [5] Tomtit: Hierarchical Federated Fine-Tuning of Giant Models based on Autonomous Synchronization
    Qi, Tianyu
    Zhan, Yufeng
    Li, Peng
    Xia, Yuanqing
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, 2024, : 1910 - 1919
  • [6] Fine-tuning 3D foundation models for geometric object retrieval
    Van den Herrewegen, Jarne
    Tourwe, Tom
    Ovsjanikov, Maks
    Wyffels, Francis
    COMPUTERS & GRAPHICS-UK, 2024, 122
  • [7] Learning Task-Specific Initialization for Effective Federated Continual Fine-Tuning of Foundation Model Adapters
    Peng, Danni
    Wang, Yuan
    Fu, Huazhu
    Wee, Qingsong
    Liu, Yong
    Goh, Rick Siow Mong
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 811 - 816
  • [8] Fine-tuning constraints on supergravity models
    Bastero-Gil, M
    Kane, GL
    King, SF
    PHYSICS LETTERS B, 2000, 474 (1-2) : 103 - 112
  • [9] Full fine-tuning strategy for endoscopic foundation models with expanded learnable offset parameters
    Dong, Minghan
    Zheng, Xiangwei
    Zhang, Xia
    Zhang, Xingyu
    Zhang, Mingzhe
    BIOMEDICAL PHYSICS & ENGINEERING EXPRESS, 2025, 11 (02):
  • [10] Titan: A Scheduler for Foundation Model Fine-tuning Workloads
    Gao, Wei
    Sun, Peng
    Wen, Yonggang
    Zhangi, Tianwei
    PROCEEDINGS OF THE 13TH SYMPOSIUM ON CLOUD COMPUTING, SOCC 2022, 2022, : 348 - 354