HyperPELT: Unified Parameter-Efficient Language Model Tuning for Both Language and Vision-and-Language Tasks

被引:0
|
作者
Zhang, Zhengkun [1 ]
Guo, Wenya [1 ]
Meng, Xiaojun [2 ]
Wang, Yasheng [2 ]
Wang, Yadao [2 ]
Jiang, Xin [2 ]
Liu, Qun [2 ]
Yang, Zhenglu [1 ]
机构
[1] Nankai Univ, CS, TKLNDST, Tianjin, Peoples R China
[2] Huawei Technol, Noahs Ark Lab, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
学科分类号
摘要
With the scale and capacity of pretrained models growing rapidly, parameter-efficient language model tuning has emerged as a popular paradigm for solving various NLP and Vision-and-Language (V&L) tasks. In this paper, we design a unified parameter-efficient multitask learning framework that works effectively on both NLP and V&L tasks. In particular, we use a shared hypernetwork that takes trainable hyper-embeddings and visual modality as input, and outputs weights for different modules in a pretrained language model, such as the parameters inserted into multi-head attention blocks (i.e., prefix-tuning) and feed-forward blocks (i.e., adapter-tuning.). Our proposed framework adds fewer trainable parameters in multi-task learning while achieving superior performances and transfer ability compared to state-of-the-art methods. Empirical results on the GLUE benchmark and multiple V&L tasks confirm the effectiveness of our framework.
引用
收藏
页码:11442 / 11453
页数:12
相关论文
共 50 条
  • [21] A Parameter-efficient Language Extension Framework for Multilingual ASR
    Liu, Wei
    Hou, Jingyong
    Yang, Dong
    Cao, Muyong
    Lee, Tan
    INTERSPEECH 2024, 2024, : 3929 - 3933
  • [22] Characterizing Communication in Distributed Parameter-Efficient Fine-Tuning for Large Language Models
    Alnaasan, Nawras
    Huang, Horng-Ruey
    Shafi, Aamir
    Subramoni, Hari
    Panda, Dhabaleswar K.
    2024 IEEE SYMPOSIUM ON HIGH-PERFORMANCE INTERCONNECTS, HOTI 2024, 2024, : 11 - 19
  • [23] Introducing Routing Functions to Vision-Language Parameter-Efficient Fine-Tuning with Low-Rank Bottlenecks
    Qui, Tingyu
    Tuytelaars, Tinne
    Moens, Marie-Francine
    COMPUTER VISION - ECCV 2024, PT LXXXVIII, 2025, 15146 : 291 - 308
  • [24] Parameter-Efficient Adaptation of Large Vision-Language Models for Video Memorability Prediction
    Martin-Fernandez, Ivan
    Esteban-Romero, Sergio
    Fernandez-Martinez, Fernando
    Gil-Martin, Manuel
    SENSORS, 2025, 25 (06)
  • [25] Multimodal high-order relational network for vision-and-language tasks
    Pan, Hao
    Huang, Jun
    NEUROCOMPUTING, 2022, 492 : 62 - 75
  • [26] Exploring Versatile Generative Language Model Via Parameter-Efficient Transfer Learning
    Lin, Zhaojiang
    Madotto, Andrea
    Fung, Pascale
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 441 - 459
  • [27] Parameter-efficient online knowledge distillation for pretrained language models
    Wang, Yukun
    Wang, Jin
    Zhang, Xuejie
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
  • [28] Recent Advances in Vision-and-language Navigation
    Sima S.-L.
    Huang Y.
    He K.-J.
    An D.
    Yuan H.
    Wang L.
    Zidonghua Xuebao/Acta Automatica Sinica, 2023, 49 (01): : 1 - 14
  • [29] On the Evaluation of Vision-and-Language Navigation Instructions
    Zhao, Ming
    Anderson, Peter
    Jain, Vihan
    Wang, Su
    Ku, Alexander
    Baldridge, Jason
    Ie, Eugene
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1302 - 1316
  • [30] Curriculum Learning for Vision-and-Language Navigation
    Zhang, Jiwen
    Wei, Zhongyu
    Fan, Jianqing
    Peng, Jiajie
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34