PTE: Prompt tuning with ensemble verbalizers

被引:0
|
作者
Liang, Liheng [1 ]
Wang, Guancheng [2 ]
Lin, Cong [2 ]
Feng, Zhuowen [3 ]
机构
[1] Guangdong Ocean Univ, Fac Math & Comp Sci, Zhanjiang 524088, Peoples R China
[2] Guangdong Ocean Univ, Coll Elect & Informat Engn, Zhanjiang 524088, Peoples R China
[3] Guangdong Ocean Univ, Coll Literature & News Commun, Zhanjiang 524088, Peoples R China
关键词
Prompt tuning; Few-shot learning; Text classification; Pre-trained language models;
D O I
10.1016/j.eswa.2024.125600
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompt tuning has achieved remarkable success in facilitating the performance of Pre-trained Language Models (PLMs) across various downstream NLP tasks, particularly in scenarios with limited downstream data. Reframing tasks as fill-in-the-blank questions represents an effective approach within prompt tuning. However, this approach necessitates the mapping of labels through a verbalizer consisting of one or more label tokens, constrained by manually crafted prompts. Furthermore, most existing automatic crafting methods either introduce external resources or rely solely on discrete or continuous optimization strategies. To address this issue, we have proposed a methodology for optimizing discrete verbalizers based on gradient descent, which we refer to this approach as PTE. This method integrates discrete tokens into verbalizers that can be continuously optimized, combining the distinct advantages of both discrete and continuous optimization strategies. In contrast to prior approaches, ours eschews reliance on prompts generated by other models or prior knowledge, merely augmenting a matrix. This approach boasts remarkable simplicity and flexibility, enabling prompt optimization while preserving the interpretability of output label tokens without constraints imposed by discrete vocabularies. Finally, employing this method in text classification tasks, we observe that PTE achieves results comparable to, if not surpassing, previous methods even under extreme conciseness. This furnishes a simple, intuitive, and efficient solution for automatically constructing verbalizers. Moreover, through quantitative analysis of optimized verbalizers, we uncover that language models likely rely not only on semantic information but also on other features for text classification. This revelation unveils new avenues for future research and model enhancements.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Prompt Tuning with Contradictory Intentions for Sarcasm Recognition
    Liu, Yiyi
    Zhang, Ruqing
    Fan, Yixing
    Guo, Jiafeng
    Cheng, Xueqi
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 328 - 339
  • [22] Prompt Tuning for Unified Multimodal Pretrained Models
    Yang, Hao
    Lin, Junyang
    Yang, An
    Wang, Peng
    Zhou, Chang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 402 - 416
  • [23] PTAU: Prompt Tuning for Attributing Unanswerable Questions
    Liao, Jinzhi
    Zhao, Xiang
    Zheng, Jianming
    Li, Xinyi
    Cai, Fei
    Tang, Jiuyang
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1219 - 1229
  • [24] On Transferability of Prompt Tuning for Natural Language Processing
    Su, Yusheng
    Wang, Xiaozhi
    Qin, Yujia
    Chan, Chi-Min
    Lin, Yankai
    Wang, Huadong
    Wen, Kaiyue
    Liu, Zhiyuan
    Li, Peng
    Li, Juanzi
    Hou, Lei
    Sun, Maosong
    Zhou, Jie
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3949 - 3969
  • [25] Hard Sample Aware Prompt-Tuning
    Xu, Yuanjian
    An, Qi
    Zhang, Jiahuan
    Li, Peng
    Nie, Zaiqing
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 12356 - 12369
  • [26] Visual Prompt Tuning for Generative Transfer Learning
    Sohn, Kihyuk
    Chang, Huiwen
    Lezama, Jose
    Polania, Luisa
    Zhang, Han
    Hao, Yuan
    Essa, Irfan
    Jiang, Lu
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 19840 - 19851
  • [27] Prompt Tuning in Code Intelligence: An Experimental Evaluation
    Wang, Chaozheng
    Yang, Yuanhang
    Gao, Cuiyun
    Peng, Yun
    Zhang, Hongyu
    Lyu, Michael R.
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2023, 49 (11) : 4869 - 4885
  • [28] SharPT: Shared Latent Space Prompt Tuning
    Pang, Bo
    Yavuz, Semih
    Xiong, Caiming
    Zhou, Yingbo
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1244 - 1250
  • [29] PTR: Prompt Tuning with Rules for Text Classification
    Han, Xu
    Zhao, Weilin
    Ding, Ning
    Liu, Zhiyuan
    Sun, Maosong
    AI OPEN, 2022, 3 : 182 - 192
  • [30] Fine-Grained Retrieval Prompt Tuning
    Wang, Shijie
    Chang, Jianlong
    Wang, Zhihui
    Li, Haojie
    Ouyang, Wanli
    Tian, Qi
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 2, 2023, : 2644 - 2652