Neural Architecture Search for Parameter-Efficient Fine-tuning of Large Pre-trained Language Models

被引:0
|
作者
Lawton, Neal [1 ]
Kumar, Anoop [2 ]
Thattai, Govind [2 ]
Galstyan, Aram [2 ]
Ver Steeg, Greg [2 ]
机构
[1] Informat Sci Inst, Marina Del Rey, CA 90292 USA
[2] Amazon Alexa AI, Redmond, WA USA
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Parameter-efficient tuning (PET) methods fit pre-trained language models (PLMs) to downstream tasks by either computing a small compressed update for a subset of model parameters, or appending and fine-tuning a small number of new model parameters to the pre-trained network. Hand-designed PET architectures from the literature perform well in practice, but have the potential to be improved via automated neural architecture search (NAS). We propose an efficient NAS method for learning PET architectures via structured and unstructured pruning. We present experiments on GLUE demonstrating the effectiveness of our algorithm and discuss how PET architectural design choices affect performance in practice.
引用
收藏
页码:8506 / 8515
页数:10
相关论文
共 50 条
  • [41] Disfluencies and Fine-Tuning Pre-trained Language Models for Detection of Alzheimer's Disease
    Yuan, Jiahong
    Bian, Yuchen
    Cai, Xingyu
    Huang, Jiaji
    Ye, Zheng
    Church, Kenneth
    INTERSPEECH 2020, 2020, : 2162 - 2166
  • [42] Towards Adaptive Prefix Tuning for Parameter-Efficient Language Model Fine-tuning
    Zhang, Zhen-Ru
    Tan, Chuanqi
    Xu, Haiyang
    Wang, Chengyu
    Huang, Jun
    Huang, Songfang
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 1239 - 1248
  • [43] Towards Efficient Fine-Tuning of Pre-trained Code Models: An Experimental Study and Beyond
    Shi, Ensheng
    Wang, Yanlin
    Zhang, Hongyu
    Du, Lun
    Han, Shi
    Zhang, Dongmei
    Sun, Hongbin
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 39 - 51
  • [44] SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models through Principled Regularized Optimization
    Jiang, Haoming
    He, Pengcheng
    Chen, Weizhu
    Liu, Xiaodong
    Gao, Jianfeng
    Zhao, Tuo
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 2177 - 2190
  • [45] Confounder balancing in adversarial domain adaptation for pre-trained large models fine-tuning
    Jiang, Shuoran
    Chen, Qingcai
    Xiang, Yang
    Pan, Youcheng
    Wu, Xiangping
    Lin, Yukang
    NEURAL NETWORKS, 2024, 173
  • [46] Fine-Tuning Pre-Trained CodeBERT for Code Search in Smart Contract
    JIN Huan
    LI Qinying
    Wuhan University Journal of Natural Sciences, 2023, 28 (03) : 237 - 245
  • [47] Fine-tuning Pre-trained Models for Robustness under Noisy Labels
    Ahn, Sumyeong
    Kim, Sihyeon
    Ko, Jongwoo
    Yun, Se-Young
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 3643 - 3651
  • [48] Exploiting Syntactic Information to Boost the Fine-tuning of Pre-trained Models
    Liu, Chaoming
    Zhu, Wenhao
    Zhang, Xiaoyu
    Zhai, Qiuhong
    2022 IEEE 46TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE (COMPSAC 2022), 2022, : 575 - 582
  • [49] Parameter-efficient fine-tuning large language model approach for hospital discharge paper summarization
    Goswami, Joyeeta
    Prajapati, Kaushal Kumar
    Saha, Ashim
    Saha, Apu Kumar
    APPLIED SOFT COMPUTING, 2024, 157
  • [50] Parameter-Efficient Fine-Tuning Large Speech Model Based on LoRA
    Ou, Ling
    Feng, Gen
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 36 - 41