REKP: Refined External Knowledge into Prompt-Tuning for Few-Shot Text Classification

被引:1
|
作者
Dang, Yuzhuo [1 ]
Chen, Weijie [1 ]
Zhang, Xin [1 ]
Chen, Honghui [1 ]
机构
[1] Natl Univ Def Technol, Sci & Technol Informat Syst Engn Lab, Changsha 410073, Peoples R China
基金
中国国家自然科学基金;
关键词
few-shot learning; text classification; prompt learning; pre-trained language model;
D O I
10.3390/math11234780
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Text classification is a machine learning technique employed to assign a given text to predefined categories, facilitating the automatic analysis and processing of textual data. However, an important problem is that the number of new text categories is growing faster than that of human annotation data, which makes many new categories of text data lack a lot of annotation data. As a result, the conventional deep neural network is forced to over-fit, which damages the application in the real world. As a solution to this problem, academics recommend addressing data scarcity through few-shot learning. One of the efficient methods is prompt-tuning, which transforms the input text into a mask prediction problem featuring [MASK]. By utilizing descriptors, the model maps output words to labels, enabling accurate prediction. Nevertheless, the previous prompt-based adaption approaches often relied on manually produced verbalizers or a single label to represent the entire label vocabulary, which makes the mapping granularity low, resulting in words not being accurately mapped to their label. To address these issues, we propose to enhance the verbalizer and construct the refined external knowledge into a prompt-tuning (REKP) model. We employ the external knowledge bases to increase the mapping space of tagged terms and design three refinement methods to remove noise data. We conduct comprehensive experiments on four benchmark datasets, namely AG's News, Yahoo, IMDB, and Amazon. The results demonstrate that REKP can outperform the state-of-the-art baselines in terms of Micro-F1 on knowledge-enhanced text classification. In addition, we conduct an ablation study to ascertain the functionality of each module in our model, revealing that the refinement module significantly contributes to enhancing classification accuracy.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] A Neural Few-Shot Text Classification Reality Check
    Dopierre, Thomas
    Gravier, Christophe
    Logerais, Wilfried
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 935 - 943
  • [42] Retrieval-Augmented Few-shot Text Classification
    Yu, Guoxin
    Liu, Lemao
    Jiang, Haiyun
    Shi, Shuming
    Ao, Xiang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 6721 - 6735
  • [43] Distinct Label Representations for Few-Shot Text Classification
    Ohashi, Sora
    Takayama, Junya
    Kajiwara, Tomoyuki
    Arase, Yuki
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 831 - 836
  • [44] Hierarchical Verbalizer for Few-Shot Hierarchical Text Classification
    Ji, Ke
    Lian, Yixin
    Gao, Jingsheng
    Wang, Baoyuan
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 2918 - 2933
  • [45] Investigating Prompt Learning for Chinese Few-Shot Text Classification with Pre-Trained Language Models
    Song, Chengyu
    Shao, Taihua
    Lin, Kejing
    Liu, Dengfeng
    Wang, Siyuan
    Chen, Honghui
    APPLIED SCIENCES-BASEL, 2022, 12 (21):
  • [46] Prompt-Based Graph Convolution Adversarial Meta-Learning for Few-Shot Text Classification
    Gong, Ruwei
    Qin, Xizhong
    Ran, Wensheng
    APPLIED SCIENCES-BASEL, 2023, 13 (16):
  • [47] BioKnowPrompt: Incorporating imprecise knowledge into prompt-tuning verbalizer with biomedical text for relation extraction
    Li, Qing
    Wang, Yichen
    You, Tao
    Lu, Yantao
    INFORMATION SCIENCES, 2022, 617 : 346 - 358
  • [48] Hierarchy-Aware Interactive Prompt Learning for Few-Shot Classification
    Yin, Xiaotian
    Wu, Jiamin
    Yang, Wenfei
    Zhou, Xu
    Zhang, Shifeng
    Zhang, Tianzhu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (12) : 12221 - 12232
  • [49] Dual selective knowledge transfer for few-shot classification
    He, Kai
    Pu, Nan
    Lao, Mingrui
    Bakker, Erwin M.
    Lew, Michael S.
    APPLIED INTELLIGENCE, 2023, 53 (22) : 27779 - 27789
  • [50] Dual selective knowledge transfer for few-shot classification
    Kai He
    Nan Pu
    Mingrui Lao
    Erwin M. Bakker
    Michael S. Lew
    Applied Intelligence, 2023, 53 : 27779 - 27789