Bayesian Multi-Task Transfer Learning for Soft Prompt Tuning

被引:0
|
作者
Lee, Haeju [1 ]
Jeong, Minchan [1 ]
Yun, Se-Young [1 ]
Kim, Kee-Eung [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch AI, Seoul, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompt tuning, in which prompts are optimized to adapt large-scale pre-trained language models to downstream tasks instead of fine-tuning the full model parameters, has been shown to be particularly effective when the prompts are trained in the multi-task transfer learning setting. These methods generally involve individually training prompts for each source task and then aggregating them to provide the initialization of the prompt for the target task. However, this approach critically ignores the fact that some of the source tasks could be negatively or positively interfering with each other. We argue that when we extract knowledge from source tasks via training source prompts, we need to consider this correlation among source tasks for better transfer to target tasks. To this end, we propose a Bayesian approach where we work with the posterior distribution of prompts across source tasks. We obtain representative source prompts corresponding to the samples from the posterior utilizing Stein Variational Gradient Descent, which are then aggregated to constitute the initial target prompt. We show extensive experimental results on the standard benchmark NLP tasks, where our Bayesian multi-task transfer learning approach outperforms the state-of-the-art methods in many settings. Furthermore, our approach requires no auxiliary models other than the prompt itself, achieving high degree of parameter-efficiency.(1)
引用
收藏
页码:4942 / 4958
页数:17
相关论文
共 50 条
  • [1] Hierarchical Prompt Tuning for Few-Shot Multi-Task Learning
    Liu, Jingping
    Chen, Tao
    Liang, Zujie
    Jiang, Haiyun
    Xiao, Yanghua
    Wei, Feng
    Qian, Yuxi
    Hao, Zhenghong
    Han, Bing
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 1556 - 1565
  • [2] Multi-task Transfer Learning for Bayesian Network Structures
    Benikhlef, Sarah
    Leray, Philippe
    Raschia, Guillaume
    Ben Messaoud, Montassar
    Sakly, Fayrouz
    SYMBOLIC AND QUANTITATIVE APPROACHES TO REASONING WITH UNCERTAINTY, ECSQARU 2021, 2021, 12897 : 217 - 228
  • [3] Hierarchical Prompt Learning for Multi-Task Learning
    Liu, Yajing
    Lu, Yuning
    Liu, Hao
    An, Yaozu
    Xu, Zhuoran
    Yao, Zhuokun
    Zhang, Baofeng
    Xiong, Zhiwei
    Gui, Chenguang
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 10888 - 10898
  • [4] Multi-task prompt tuning with soft context sharing for vision-language models
    Ding, Kun
    Wang, Ying
    Liu, Pengzhang
    Yu, Qiang
    Zhang, Haojian
    Xiang, Shiming
    Pan, Chunhong
    NEUROCOMPUTING, 2024, 603
  • [5] MULTI-TASK DISTILLATION: TOWARDS MITIGATING THE NEGATIVE TRANSFER IN MULTI-TASK LEARNING
    Meng, Ze
    Yao, Xin
    Sun, Lifeng
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 389 - 393
  • [6] Multi-Task Reinforcement Learning with Soft Modularization
    Yang, Ruihan
    Xu, Huazhe
    Wu, Yi
    Wang, Xiaolong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [7] Few-shot Log Analysis with Prompt-based Multi-task Transfer Learning
    Zhou, Mingjie
    Yang, Weidong
    Ma, Lipeng
    Jiang, Sihang
    Xu, Bo
    Xiao, Yanghua
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2024, PT 2, 2025, 14851 : 466 - 475
  • [8] NONPARAMETRIC BAYESIAN FEATURE SELECTION FOR MULTI-TASK LEARNING
    Li, Hui
    Liao, Xuejun
    Carin, Lawrence
    2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 2236 - 2239
  • [9] MultiAICL: Multi-task Tuning for Augmented In-Context Learning in Text Style Transfer
    Zhu, Linan
    Zhou, Zehai
    Chen, Xiangfan
    Guo, Xiaolei
    Kong, Xiangjie
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT III, NLPCC 2024, 2025, 15361 : 55 - 66
  • [10] Bayesian Multi-task Relationship Learning with Link Structure
    Li, Yingming
    Yang, Ming
    Qi, Zhongang
    Zhang, Zhongfei
    2013 IEEE 13TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2013, : 1115 - 1120