MEFT: Memory-Efficient Fine-Tuning through Sparse Adapter

被引:0
|
作者
Hao, Jitai [1 ]
Sun, Weiwei [1 ,2 ]
Xin, Xin [1 ]
Meng, Qi [3 ]
Chen, Zhumin [1 ]
Ren, Pengjie [1 ]
Ren, Zhaochun [4 ]
机构
[1] Shandong Univ, Qingdao, Peoples R China
[2] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
[3] Acad Math & Syst Sci, Beijing, Peoples R China
[4] Leiden Univ, Leiden, Netherlands
来源
PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS | 2024年
基金
国家重点研发计划;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Parameter-Efficient Fine-tuning (PEFT) facilitates the fine-tuning of Large Language Models (LLMs) under limited resources. However, the fine-tuning performance with PEFT on complex, knowledge-intensive tasks is limited due to the constrained model capacity, which originates from the limited number of additional trainable parameters. To overcome this limitation, we introduce a novel mechanism that fine-tunes LLMs with adapters of larger size yet memory-efficient. This is achieved by leveraging the inherent activation sparsity in the Feed-Forward Networks (FFNs) of LLMs and utilizing the larger capacity of Central Processing Unit (CPU) memory compared to Graphics Processing Unit (GPU). We store and update the parameters of larger adapters on the CPU. Moreover, we employ a Mixture of Experts (MoE)-like architecture to mitigate unnecessary CPU computations and reduce the communication volume between the GPU and CPU. This is particularly beneficial over the limited bandwidth of PCI Express (PCIe). Our method can achieve fine-tuning results comparable to those obtained with larger memory capacities, even when operating under more limited resources such as a 24GB memory single GPU setup, with acceptable loss in training efficiency. Our codes are available at https://github.com/CURRENTF/MEFT.
引用
收藏
页码:2375 / 2388
页数:14
相关论文
共 50 条
  • [21] SpotTune: Transfer Learning through Adaptive Fine-tuning
    Guo, Yunhui
    Shi, Honghui
    Kumar, Abhishek
    Grauman, Kristen
    Rosing, Tajana
    Feris, Rogerio
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4800 - 4809
  • [22] ClassWise-SAM-Adapter: Parameter-Efficient Fine-Tuning Adapts Segment Anything to SAR Domain for Semantic Segmentation
    Pu, Xinyang
    Jia, Hecheng
    Zheng, Linghao
    Wang, Feng
    Xu, Feng
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2025, 18 : 4791 - 4804
  • [23] Parameter-Efficient and Memory-Efficient Tuning for Vision Transformer: A Disentangled Approach
    Zhang, Taolin
    Bai, Jiawang
    Lu, Zhihe
    Lian, Dongze
    Wang, Genping
    Wang, Xinchao
    Xia, Shu-Tao
    COMPUTER VISION - ECCV 2024, PT XLV, 2025, 15103 : 346 - 363
  • [24] Fine-tuning memory macros using variable internal delays
    Gray, K
    IEEE SPECTRUM, 1999, 36 (08) : 44 - 49
  • [25] Efficient Index Learning via Model Reuse and Fine-tuning
    Liu, Guanli
    Qi, Jianzhong
    Kulik, Lars
    Soga, Kazuya
    Borovica-Gajic, Renata
    Rubinstein, Benjamin I. P.
    2023 IEEE 39TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING WORKSHOPS, ICDEW, 2023, : 60 - 66
  • [26] Pluto and Charon: A Time and Memory Efficient Collaborative Edge AI Framework for Personal LLMs Fine-Tuning
    Bei, Ouyang
    Ye, Shengyuan
    Zeng, Liekang
    Qian, Tianyi
    Li, Jingyi
    Chen, Xu
    53RD INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2024, 2024, : 762 - 771
  • [27] Salt adaptation requires efficient fine-tuning of jasmonate signalling
    Ismail, Ahmed
    Seo, Mitsunori
    Takebayashi, Yumiko
    Kamiya, Yuji
    Eiche, Elisabeth
    Nick, Peter
    PROTOPLASMA, 2014, 251 (04) : 881 - 898
  • [28] Salt adaptation requires efficient fine-tuning of jasmonate signalling
    Ahmed Ismail
    Mitsunori Seo
    Yumiko Takebayashi
    Yuji Kamiya
    Elisabeth Eiche
    Peter Nick
    Protoplasma, 2014, 251 : 881 - 898
  • [29] Frozen Weights as Prior for Parameter-Efficient Fine-Tuning
    Ma, Xiaolong
    Liu, Peishun
    Gao, Haojie
    Yan, Zikang
    Ma, Ningning
    Liu, Wenqiang
    Wang, Xuefang
    Tang, Ruichun
    IEEE ACCESS, 2025, 13 : 24411 - 24425
  • [30] Efficient Large Models Fine-tuning on Commodity Servers via Memory-balanced Pipeline Parallelism
    Liu, Yujie
    Lai, Zhiquan
    Liu, Weijie
    Wang, Wei
    Li, Dongsheng
    Proceedings - 2023 IEEE International Conference on High Performance Computing and Communications, Data Science and Systems, Smart City and Dependability in Sensor, Cloud and Big Data Systems and Application, HPCC/DSS/SmartCity/DependSys 2023, 2023, : 726 - 727