Hydra: Multi-head low-rank adaptation for parameter efficient fine-tuning

被引:3
|
作者
Kim, Sanghyeon [1 ]
Yang, Hyunmo [2 ]
Kim, Yunghyun [2 ]
Hong, Youngjoon [3 ]
Park, Eunbyung [1 ,2 ]
机构
[1] Sungkyunkwan Univ, Dept Elect & Comp Engn, 2066 Seobu Ro, Suwon 16419, South Korea
[2] Sungkyunkwan Univ, Dept Artificial Intelligence, 2066 Seobu Ro, Suwon 16419, South Korea
[3] Korea Adv Inst Sci & Technol, Dept Math Sci, 291 Daehak Ro, Taejon 305701, South Korea
基金
新加坡国家研究基金会;
关键词
Parameter efficient fine-tuning; Adapter; Transformer; BENCHMARK;
D O I
10.1016/j.neunet.2024.106414
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recent surge in large-scale foundation models has spurred the development of efficient methods for adapting these models to various downstream tasks. Low-rank adaptation methods, such as LoRA, have gained significant attention due to their outstanding parameter efficiency and no additional inference latency. This paper investigates a more general form of adapter module based on the analysis that parallel and sequential adaptation branches learn novel and general features during fine-tuning, respectively. The proposed method, named Hydra, combines parallel and sequential branch to integrate capabilities, which is more expressive than existing single branch methods and enables the exploration of a broader range of optimal points in the finetuning process. In addition, the proposed method explicitly leverages the pre-trained weights by performing a linear combination of the pre-trained features. It allows the learned features to have better generalization performance across diverse downstream tasks. Furthermore, we perform a comprehensive analysis of the characteristics of each adaptation branch with empirical evidence. Through an extensive range of experiments, we substantiate the efficiency and demonstrate the superior performance of Hydra. This comprehensive evaluation underscores the potential impact and effectiveness of Hydra in a variety of applications. The source code of this work is publicly opened on https://github.com/extremebird/Hydra.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] MTLoRA: A Low-Rank Adaptation Approach for Efficient Multi-Task Learning
    Agiza, Ahmed
    Neseem, Marina
    Reda, Sherief
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 16196 - 16205
  • [22] Frozen Weights as Prior for Parameter-Efficient Fine-Tuning
    Ma, Xiaolong
    Liu, Peishun
    Gao, Haojie
    Yan, Zikang
    Ma, Ningning
    Liu, Wenqiang
    Wang, Xuefang
    Tang, Ruichun
    IEEE ACCESS, 2025, 13 : 24411 - 24425
  • [23] Salt adaptation requires efficient fine-tuning of jasmonate signalling
    Ismail, Ahmed
    Seo, Mitsunori
    Takebayashi, Yumiko
    Kamiya, Yuji
    Eiche, Elisabeth
    Nick, Peter
    PROTOPLASMA, 2014, 251 (04) : 881 - 898
  • [24] Salt adaptation requires efficient fine-tuning of jasmonate signalling
    Ahmed Ismail
    Mitsunori Seo
    Yumiko Takebayashi
    Yuji Kamiya
    Elisabeth Eiche
    Peter Nick
    Protoplasma, 2014, 251 : 881 - 898
  • [25] Tree Prompting: Efficient Task Adaptation without Fine-Tuning
    Morris, John X.
    Singh, Chandan
    Rush, Alexander M.
    Gao, Jianfeng
    Deng, Yuntian
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 6253 - 6267
  • [26] Efficient Low-rank Backpropagation for Vision Transformer Adaptation
    Yang, Yuedong
    Chiang, Hung-Yueh
    Li, Guihong
    Marculescu, Diana
    Marculescu, Radu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [27] Parameter-efficient Multi-task Fine-tuning for Transformers via Shared Hypernetworks
    Mahabadi, Rabeeh Karimi
    Ruder, Sebastian
    Dehghani, Mostafa
    Henderson, James
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 565 - 576
  • [28] Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model Fine-tuning
    Zhang, Zhen-Ru
    Tan, Chuanqi
    Xu, Haiyang
    Wang, Chengyu
    Huang, Jun
    Huang, Songfang
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1239 - 1248
  • [29] Parameter-Efficient Fine-Tuning without Introducing New Latency
    Liao, Baohao
    Meng, Yan
    Monz, Christof
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 4242 - 4260
  • [30] Democratizing protein language models with parameter-efficient fine-tuning
    Sledzieski, Samuel
    Kshirsagar, Meghana
    Baek, Minkyung
    Dodhia, Rahul
    Ferres, Juan Lavista
    Berger, Bonnie
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2024, 121 (26)