Few-Shot Text Classification with an Efficient Prompt Tuning Method in Meta-Learning Framework

被引:1
|
作者
Lv, Xiaobao [1 ,2 ]
机构
[1] Southeast Univ, Sch Comp Sci & Engn, 2 Southeast Univ Rd, Nanjing, Jiangsu, Peoples R China
[2] Zhongke Shuguang Nanjing Res Inst Co Ltd, 519 Chengxin Rd, Nanjing, Jiangsu, Peoples R China
关键词
Few-shot learning; meta-learning; prompt tuning; text classification; pre-trained language model;
D O I
10.1142/S0218001424510066
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Meta-learning stands as a prevalent framework utilized in few-shot learning methods. Nonetheless, its efficacy hinges on substantial data availability during meta-training. Recent work adeptly tackled this hurdle by synergizing prompt tuning with the meta-learning paradigm, consequently attaining unparalleled performance on four benchmarks (FewRel, HuffPost, Reuters and Amazon). Nonetheless, the implementation efficacy of the previous method leaves room for enhancement, which is especially crucial when tuning larger language models. To this end, we introduce another expedited prompt tuning approach nested within the meta-learning framework. The novel approach normalizes the label information and sample information and uses the regression method to obtain the closed-form solution of each few-shot task, which significantly enhances inference speed, achieving a twofold improvement, while concurrently elevating average accuracy by 1.7 similar to 3.0% on the same benchmarks. Moreover, it demonstrates enhanced stability when faced with limited meta-training data, which is more applicable in many real scenarios where parallel data is rare. The source code is available to reproduce the results (http://github.com/Dr-Lv/EMPT).
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Prompt-Based Graph Convolution Adversarial Meta-Learning for Few-Shot Text Classification
    Gong, Ruwei
    Qin, Xizhong
    Ran, Wensheng
    APPLIED SCIENCES-BASEL, 2023, 13 (16):
  • [2] Enhanced Prompt Learning for Few-shot Text Classification Method
    Li R.
    Wei Z.
    Fan Y.
    Ye S.
    Zhang G.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2024, 60 (01): : 1 - 12
  • [3] Meta-learning triplet contrast network for few-shot text classification
    Dong, Kaifang
    Jiang, Baoxing
    Li, Hongye
    Zhu, Zhenfang
    Liu, Peiyu
    KNOWLEDGE-BASED SYSTEMS, 2024, 303
  • [4] MEDA: Meta-Learning with Data Augmentation for Few-Shot Text Classification
    Sun, Pengfei
    Ouyang, Yawen
    Zhang, Wenming
    Dai, Xin-yu
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3929 - 3935
  • [5] Fair Meta-Learning For Few-Shot Classification
    Zhao, Chen
    Li, Changbin
    Li, Jincheng
    Chen, Feng
    11TH IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG 2020), 2020, : 275 - 282
  • [6] A META-LEARNING FRAMEWORK FOR FEW-SHOT CLASSIFICATION OF REMOTE SENSING SCENE
    Zhang, Pei
    Bai, Yunpeng
    Wang, Dong
    Bai, Bendu
    Li, Ying
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4590 - 4594
  • [7] MedOptNet: Meta-Learning Framework for Few-Shot Medical Image Classification
    Lu, Liangfu
    Cui, Xudong
    Tan, Zhiyuan
    Wu, Yulei
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2024, 21 (04) : 725 - 736
  • [8] Few-shot classification via efficient meta-learning with hybrid optimization
    Jia, Jinfang
    Feng, Xiang
    Yu, Huiqun
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 127
  • [9] Improving Meta-learning for Few-Shot Text Classification via Label Propagation
    Li, Haorui
    Shao, Jie
    Zeng, Xiangqiang
    Xu, Hui
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT V, ECML PKDD 2024, 2024, 14945 : 389 - 405
  • [10] Meta-Learning Adversarial Domain Adaptation Network for Few-Shot Text Classification
    Han, ChengCheng
    Fan, Zeqiu
    Zhang, Dongxiang
    Qiu, Minghui
    Gao, Ming
    Zhou, Aoying
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1664 - 1673