Parameter-efficient fine-tuning in large language models: a survey of methodologies

被引:0
|
作者
Luping Wang [1 ]
Sheng Chen [1 ]
Linnan Jiang [1 ]
Shu Pan [1 ]
Runze Cai [1 ]
Sen Yang [1 ]
Fei Yang [1 ]
机构
[1] Zhejiang Laboratory,
关键词
Fine-tuning; Parameter-efficient; Large language model; Deep learning; Artificial intelligence;
D O I
10.1007/s10462-025-11236-4
中图分类号
学科分类号
摘要
The large language models, as predicted by scaling law forecasts, have made groundbreaking progress in many fields, particularly in natural language generation tasks, where they have approached or even surpassed human levels. However, the unprecedented scale of their parameters brings significant computational and storage costs. These large language models require substantial computational resources and GPU memory to operate. When adapting large language models to specific downstream tasks, their massive parameter scale poses a significant challenge in fine-tuning on hardware platforms with limited computational power and GPU memory. To address this issue, parameter-efficient fine-tuning (PEFT) offers a practical solution by efficiently adjusting the parameters of large pre-trained models to suit various downstream tasks. Specifically, PEFT adjusts the parameters of pre-trained large language models to adapt to specific tasks or domains, minimizing the introduction of additional parameters and the computational resources required. This review mainly introduces the preliminary knowledge of PEFT, the core ideas and principles of various PEFT algorithms, the applications of PEFT, and potential future research directions. By reading this review, we believe that interested parties can quickly grasp the PEFT methodology, thereby accelerating its development and innovation.
引用
收藏
相关论文
共 50 条
  • [21] AutoPEFT : Automatic Configuration Search for Parameter-Efficient Fine-Tuning
    Zhou, Han
    Wan, Xingchen
    Vulic, Ivan
    Korhonen, Anna
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 525 - 542
  • [22] Sensitivity-Aware Visual Parameter-Efficient Fine-Tuning
    He, Haoyu
    Cai, Jianfei
    Zhang, Jing
    Tao, Dacheng
    Zhuang, Bohan
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11791 - 11801
  • [23] AdapterGNN: Parameter-Efficient Fine-Tuning Improves Generalization in GNNs
    Li, Shengrui
    Han, Xueting
    Bai, Jing
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 12, 2024, : 13600 - 13608
  • [24] Leveraging Parameter-Efficient Fine-Tuning for Multilingual Abstractive Summarization
    Shen, Jialun
    Wang, Yusong
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT III, NLPCC 2024, 2025, 15361 : 293 - 303
  • [25] Parameter-efficient fine-tuning for single image snow removal
    Dai, Xinwei
    Zhou, Yuanbo
    Qiu, Xintao
    Tang, Hui
    Tong, Tong
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
  • [26] Bayesian Parameter-Efficient Fine-Tuning for Overcoming Catastrophic Forgetting
    Chen, Haolin
    Garner, Philip N.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 4253 - 4262
  • [27] One is Not Enough: Parameter-Efficient Fine-Tuning With Multiplicative Sparse Factorization
    Chen, Xuxi
    Chen, Tianlong
    Cheng, Yu
    Chen, Weizhu
    Awadallah, Ahmed Hassan
    Wang, Zhangyang
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2024, 18 (06) : 1059 - 1069
  • [28] Strong Baselines for Parameter-Efficient Few-Shot Fine-Tuning
    Basu, Samyadeep
    Hu, Shell
    Massiceti, Daniela
    Feizi, Soheil
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11024 - 11031
  • [29] An Empirical Study of Parameter-Efficient Fine-Tuning Methods for Pre-trained Code Models
    Liu, Jiaxing
    Sha, Chaofeng
    Peng, Xin
    2023 38TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE, 2023, : 397 - 408
  • [30] SCT: A Simple Baseline for Parameter-Efficient Fine-Tuning via Salient Channels
    Zhao, Henry Hengyuan
    Wang, Pichao
    Zhao, Yuyang
    Luo, Hao
    Wang, Fan
    Shou, Mike Zheng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (03) : 731 - 749