MEFT: Memory-Efficient Fine-Tuning through Sparse Adapter

被引:0
|
作者
Hao, Jitai [1 ]
Sun, Weiwei [1 ,2 ]
Xin, Xin [1 ]
Meng, Qi [3 ]
Chen, Zhumin [1 ]
Ren, Pengjie [1 ]
Ren, Zhaochun [4 ]
机构
[1] Shandong Univ, Qingdao, Peoples R China
[2] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
[3] Acad Math & Syst Sci, Beijing, Peoples R China
[4] Leiden Univ, Leiden, Netherlands
基金
国家重点研发计划;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Parameter-Efficient Fine-tuning (PEFT) facilitates the fine-tuning of Large Language Models (LLMs) under limited resources. However, the fine-tuning performance with PEFT on complex, knowledge-intensive tasks is limited due to the constrained model capacity, which originates from the limited number of additional trainable parameters. To overcome this limitation, we introduce a novel mechanism that fine-tunes LLMs with adapters of larger size yet memory-efficient. This is achieved by leveraging the inherent activation sparsity in the Feed-Forward Networks (FFNs) of LLMs and utilizing the larger capacity of Central Processing Unit (CPU) memory compared to Graphics Processing Unit (GPU). We store and update the parameters of larger adapters on the CPU. Moreover, we employ a Mixture of Experts (MoE)-like architecture to mitigate unnecessary CPU computations and reduce the communication volume between the GPU and CPU. This is particularly beneficial over the limited bandwidth of PCI Express (PCIe). Our method can achieve fine-tuning results comparable to those obtained with larger memory capacities, even when operating under more limited resources such as a 24GB memory single GPU setup, with acceptable loss in training efficiency. Our codes are available at https://github.com/CURRENTF/MEFT.
引用
收藏
页码:2375 / 2388
页数:14
相关论文
共 50 条
  • [1] Memory-Efficient Fine-Tuning for Quantized Diffusion Model
    Ryu, Hyogon
    Lim, Seohyun
    Shim, Hyunjung
    COMPUTER VISION - ECCV 2024, PT XVI, 2025, 15074 : 356 - 372
  • [2] PockEngine: Sparse and Efficient Fine-tuning in a Pocket
    Zhu, Ligeng
    Hu, Lanxiang
    Lin, Ji
    Wang, Wei-Chen
    Chen, Wei-Ming
    Gan, Chuang
    Han, Song
    56TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, MICRO 2023, 2023, : 1381 - 1394
  • [3] Memory-Efficient Fine-Tuning of Compressed Large Language Models via sub-4-bit Integer Quantization
    Kim, Jeonghoon
    Lee, Jung Hyun
    Kim, Sungdong
    Park, Joonsuk
    Yoo, Kang Min
    Kwon, Se Jung
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [4] A survey of efficient fine-tuning methods for Vision-Language Models - Prompt and Adapter
    Xing, Jialu
    Liu, Jianping
    Wang, Jian
    Sun, Lulu
    Chen, Xi
    Gu, Xunxun
    Wang, Yingfei
    COMPUTERS & GRAPHICS-UK, 2024, 119
  • [5] One is Not Enough: Parameter-Efficient Fine-Tuning With Multiplicative Sparse Factorization
    Chen, Xuxi
    Chen, Tianlong
    Cheng, Yu
    Chen, Weizhu
    Awadallah, Ahmed Hassan
    Wang, Zhangyang
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2024, 18 (06) : 1059 - 1069
  • [6] Transferrable DP-Adapter Tuning: A Privacy-Preserving Multimodal Parameter-Efficient Fine-Tuning Framework
    Ji, Lixia
    Xiao, Shijie
    Xu, Bingzhi
    Zhang, Han
    2024 IEEE 24TH INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY AND SECURITY, QRS, 2024, : 471 - 482
  • [7] LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models
    Hu, Zhiqiang
    Wang, Lei
    Lan, Yihuai
    Xu, Wanyu
    Lim, Ee-Peng
    Bing, Lidong
    Xu, Xing
    Poria, Soujanya
    Lee, Roy Ka-Wei
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 5254 - 5276
  • [8] On the Effectiveness of Parameter-Efficient Fine-Tuning
    Fu, Zihao
    Yang, Haoran
    So, Anthony Man-Cho
    Lam, Wai
    Bing, Lidong
    Collier, Nigel
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 12799 - 12807
  • [9] Efficient Fine-Tuning of BERT Models on the Edge
    Vucetic, Danilo
    Tayaranian, Mohammadreza
    Ziaeefard, Maryam
    Clark, James J.
    Meyer, Brett H.
    Gross, Warren J.
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1838 - 1842
  • [10] Personalized Large Language Models through Parameter Efficient Fine-Tuning Techniques
    Braga, Marco
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 3076 - 3076