Self-supervised Meta-Prompt Learning with Meta-Gradient Regularization for Few-shot Generalization

被引:0
|
作者
Pan, Kaihang [1 ,2 ,4 ]
Li, Juncheng [1 ]
Song, Hongye [2 ]
Lin, Jun [2 ]
Liu, Xiaozhong
Tang, Siliang [1 ,3 ]
机构
[1] Zhejiang Univ, Hangzhou, Peoples R China
[2] Alibaba Grp, DAMO Acad, Hangzhou, Peoples R China
[3] Worcester Polytech Inst, Worcester, MA USA
[4] Alibaba DAMO Acad, Hangzhou, Peoples R China
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompt tuning is a parameter-efficient method, which learns soft prompts and conditions frozen language models to perform specific downstream tasks. Though effective, prompt tuning under few-shot settings on the one hand heavily relies on a good initialization of soft prompts. On the other hand, it can easily overfit to few-shot training samples, thereby undermining generalizability. Existing works leverage pre-training or supervised meta-learning to initialize soft prompts but they fail to data-efficiently generalize to unseen downstream tasks. To address the above problems, this paper proposes a novel Self-sUpervised meta-Prompt learning framework with MEta-gradient Regularization for few-shot generalization (SUPMER). SUPMER leverages self-supervised meta-learning with a diverse set of well-designed meta-training tasks to learn a universal prompt initialization for efficient adaptation using only unlabeled data. Additionally, it jointly meta-learns a gradient regularization function to transform raw gradients into a domain-generalizable direction, thus alleviating the problem of overfitting. Extensive experiments show that SUPMER achieves better performance for different few-shot downstream tasks, and also exhibits a stronger domain generalization ability. The code for SUPMER will be available at https://github.com/beepkh/SUPMER.
引用
收藏
页码:1059 / 1077
页数:19
相关论文
共 50 条
  • [1] Meta Self-Supervised Learning for Distribution Shifted Few-Shot Scene Classification
    Gong, Tengfei
    Zheng, Xiangtao
    Lu, Xiaoqiang
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [2] MaskSplit: Self-supervised Meta-learning for Few-shot Semantic Segmentation
    Amac, Mustafa Sercan
    Sencan, Ahmet
    Baran, Orhun Bugra
    Ikizler-Cinbis, Nazli
    Cinbis, Ramazan Gokberk
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 428 - 438
  • [3] Knowledge-enhanced meta-prompt for few-shot relation extraction
    Cui, Jinman
    Xu, Fu
    Wang, Xinyang
    Li, Yakun
    Qu, Xiaolong
    Yao, Lei
    Li, Dongmei
    COMPUTER SPEECH AND LANGUAGE, 2025, 91
  • [4] Self-Supervised Meta-Learning for Few-Shot Natural Language Classification Tasks
    Bansal, Trapit
    Jha, Rishikesh
    Munkhdalai, Tsendsuren
    McCallum, Andrew
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 522 - 534
  • [5] Effectively learn how to learn: a novel few-shot learning with meta-gradient memory
    Hui, Lin
    Chen, Yi-Cheng
    INTERNATIONAL JOURNAL OF WEB AND GRID SERVICES, 2024, 20 (01) : 3 - 24
  • [6] Reinforced Self-Supervised Training for Few-Shot Learning
    Yan, Zhichao
    An, Yuexuan
    Xue, Hui
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 731 - 735
  • [7] Conditional Self-Supervised Learning for Few-Shot Classification
    An, Yuexuan
    Xue, Hui
    Zhao, Xingyu
    Zhang, Lu
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2140 - 2146
  • [8] Self-Supervised Few-Shot Learning on Point Clouds
    Sharma, Charu
    Kaul, Manohar
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [9] Pareto Self-Supervised Training for Few-Shot Learning
    Chen, Zhengyu
    Ge, Jixie
    Zhan, Heshen
    Huang, Siteng
    Wang, Donglin
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 13658 - 13667
  • [10] SELF-SUPERVISED LEARNING FOR FEW-SHOT IMAGE CLASSIFICATION
    Chen, Da
    Chen, Yuefeng
    Li, Yuhong
    Mao, Feng
    He, Yuan
    Xue, Hui
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 1745 - 1749