PTE: Prompt tuning with ensemble verbalizers

被引:0
|
作者
Liang, Liheng [1 ]
Wang, Guancheng [2 ]
Lin, Cong [2 ]
Feng, Zhuowen [3 ]
机构
[1] Guangdong Ocean Univ, Fac Math & Comp Sci, Zhanjiang 524088, Peoples R China
[2] Guangdong Ocean Univ, Coll Elect & Informat Engn, Zhanjiang 524088, Peoples R China
[3] Guangdong Ocean Univ, Coll Literature & News Commun, Zhanjiang 524088, Peoples R China
关键词
Prompt tuning; Few-shot learning; Text classification; Pre-trained language models;
D O I
10.1016/j.eswa.2024.125600
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompt tuning has achieved remarkable success in facilitating the performance of Pre-trained Language Models (PLMs) across various downstream NLP tasks, particularly in scenarios with limited downstream data. Reframing tasks as fill-in-the-blank questions represents an effective approach within prompt tuning. However, this approach necessitates the mapping of labels through a verbalizer consisting of one or more label tokens, constrained by manually crafted prompts. Furthermore, most existing automatic crafting methods either introduce external resources or rely solely on discrete or continuous optimization strategies. To address this issue, we have proposed a methodology for optimizing discrete verbalizers based on gradient descent, which we refer to this approach as PTE. This method integrates discrete tokens into verbalizers that can be continuously optimized, combining the distinct advantages of both discrete and continuous optimization strategies. In contrast to prior approaches, ours eschews reliance on prompts generated by other models or prior knowledge, merely augmenting a matrix. This approach boasts remarkable simplicity and flexibility, enabling prompt optimization while preserving the interpretability of output label tokens without constraints imposed by discrete vocabularies. Finally, employing this method in text classification tasks, we observe that PTE achieves results comparable to, if not surpassing, previous methods even under extreme conciseness. This furnishes a simple, intuitive, and efficient solution for automatically constructing verbalizers. Moreover, through quantitative analysis of optimized verbalizers, we uncover that language models likely rely not only on semantic information but also on other features for text classification. This revelation unveils new avenues for future research and model enhancements.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] No More Fine-Tuning? An Experimental Evaluation of Prompt Tuning in Code Intelligence
    Wang, Chaozheng
    Yang, Yuanhang
    Gao, Cuiyun
    Peng, Yun
    Zhang, Hongyu
    Lyu, Michael R.
    PROCEEDINGS OF THE 30TH ACM JOINT MEETING EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, ESEC/FSE 2022, 2022, : 382 - 394
  • [32] P-Tuning: Prompt Tuning Can Be Comparable to Fine-tuning Across Scales and Tasks
    Liu, Xiao
    Ji, Kaixuan
    Fu, Yicheng
    Tam, Weng Lam
    Du, Zhengxiao
    Yang, Zhilin
    Tang, Jie
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 61 - 68
  • [33] The Power of Scale for Parameter-Efficient Prompt Tuning
    Lester, Brian
    Al-Rfou, Rami
    Constant, Noah
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3045 - 3059
  • [34] Modality-Consistent Prompt Tuning With Optimal Transport
    Ren, Hairui
    Tang, Fan
    Zheng, Huangjie
    Zhao, He
    Guo, Dandan
    Chang, Yi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 2499 - 2512
  • [35] Building a Personalized Dialogue System with Prompt-Tuning
    Kasahara, Tomohito
    Kawahara, Daisuke
    Tung, Nguyen
    Li, Shengzhe
    Shinzato, Kenta
    Sato, Toshinori
    NAACL 2022 - 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Student Research Workshop, 2022, : 96 - 105
  • [36] Black-Box Prompt Tuning With Subspace Learning
    Zheng, Yuanhang
    Tan, Zhixing
    Li, Peng
    Liu, Yang
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3002 - 3013
  • [37] PTSTEP: Prompt Tuning for Semantic Typing of Event Processes
    Zhu, Wenhao
    Xu, Yongxiu
    Xu, Hongbo
    Tang, Minghao
    Zhu, Dongwei
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT III, 2023, 14256 : 541 - 553
  • [38] Building a Personalized Dialogue System with Prompt-Tuning
    Kasahara, Tomohito
    Kawahara, Daisuke
    Nguyen Tung
    Li, Shengzhe
    Shinzato, Kenta
    Sato, Toshinori
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES: PROCEEDINGS OF THE STUDENT RESEARCH WORKSHOP, 2022, : 96 - 105
  • [39] LIPT: Improving Prompt Tuning with Late Inception Reparameterization
    He, Yawen
    Feng, Ao
    Gao, Zhengjie
    Song, Xinyu
    ELECTRONICS, 2024, 13 (23):
  • [40] ProTeCt: Prompt Tuning for Taxonomic Open Set Classification
    Wu, Tz-Ying
    Ho, Chih-Hui
    Vasconcelos, Nuno
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 16531 - 16540