Prompt tuning discriminative language models for hierarchical text classification

被引:0
|
作者
du Toit, Jaco [1 ,2 ]
Dunaiski, Marcel [1 ,2 ]
机构
[1] Stellenbosch Univ, Dept Math Sci, Comp Sci Div, Stellenbosch, South Africa
[2] Stellenbosch Univ, Sch Data Sci & Computat Thinking, Stellenbosch, South Africa
来源
NATURAL LANGUAGE PROCESSING | 2024年
关键词
Large language models; discriminative language models; hierarchical text classification; prompt tuning;
D O I
10.1017/nlp.2024.51
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Hierarchical text classification (HTC) is a natural language processing task which aims to categorise a text document into a set of classes from a hierarchical class structure. Recent approaches to solve HTC tasks focus on leveraging pre-trained language models (PLMs) and the hierarchical class structure by allowing these components to interact in various ways. Specifically, the Hierarchy-aware Prompt Tuning (HPT) method has proven to be effective in applying the prompt tuning paradigm to Bidirectional Encoder Representations from Transformers (BERT) models for HTC tasks. Prompt tuning aims to reduce the gap between the pre-training and fine-tuning phases by transforming the downstream task into the pre-training task of the PLM. Discriminative PLMs, which use a replaced token detection (RTD) pre-training task, have also shown to perform better on flat text classification tasks when using prompt tuning instead of vanilla fine-tuning. In this paper, we propose the Hierarchy-aware Prompt Tuning for Discriminative PLMs (HPTD) approach which injects the HTC task into the RTD task used to pre-train discriminative PLMs. Furthermore, we make several improvements to the prompt tuning approach of discriminative PLMs that enable HTC tasks to scale to much larger hierarchical class structures. Through comprehensive experiments, we show that our method is robust and outperforms current state-of-the-art approaches on two out of three HTC benchmark datasets.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Prompt Tuning for Discriminative Pre-trained Language Models
    Yao, Yuan
    Dong, Bowen
    Zhang, Ao
    Zhang, Zhengyan
    Xie, Ruobing
    Liu, Zhiyuan
    Lin, Leyu
    Sun, Maosong
    Wang, Jianyong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3468 - 3473
  • [2] COPHTC: CONTRASTIVE LEARNING WITH PROMPT TUNING FOR HIERARCHICAL TEXT CLASSIFICATION
    Cai, Fuhan
    Zhang, Zhongqiang
    Liu, Duo
    Fang, Xiangzhong
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 5400 - 5404
  • [3] Enhanced Discriminative Fine-Tuning of Large Language Models for Chinese Text Classification
    Song, Jinwang
    Zan, Hongying
    Zhang, Kunli
    2024 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING, IALP 2024, 2024, : 168 - 174
  • [4] Medical text classification based on the discriminative pre-training model and prompt-tuning
    Wang, Yu
    Wang, Yuan
    Peng, Zhenwan
    Zhang, Feifan
    Zhou, Luyao
    Yang, Fei
    DIGITAL HEALTH, 2023, 9
  • [5] PTR: Prompt Tuning with Rules for Text Classification
    Han, Xu
    Zhao, Weilin
    Ding, Ning
    Liu, Zhiyuan
    Sun, Maosong
    AI OPEN, 2022, 3 : 182 - 192
  • [6] Knowledgeable Prompt-tuning: Incorporating Knowledge into Prompt Verbalizer for Text Classification
    Hu, Shengding
    Ding, Ning
    Wang, Huadong
    Liu, Zhiyuan
    Wang, Jingang
    Li, Juanzi
    Wu, Wei
    Sun, Maosong
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 2225 - 2240
  • [7] Short text classification with Soft Knowledgeable Prompt-tuning
    Zhu, Yi
    Wang, Ye
    Mu, Jianyuan
    Li, Yun
    Qiang, Jipeng
    Yuan, Yunhao
    Wu, Xindong
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 246
  • [8] Adversarial Prompt Tuning for Vision-Language Models
    Zhang, Jiaming
    Ma, Xingjun
    Wang, Xin
    Qiu, Lingyu
    Wang, Jiaqi
    Jiang, Yu-Gang
    Sang, Jitao
    COMPUTER VISION - ECCV 2024, PT XLV, 2025, 15103 : 56 - 72
  • [9] Prompt Tuning on Graph-Augmented Low-Resource Text Classification
    Wen, Zhihao
    Fang, Yuan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 9080 - 9095
  • [10] Unifying Graph Retrieval and Prompt Tuning for Graph-Grounded Text Classification
    Dai, Le
    Yin, Yu
    Chen, Enhong
    Xiong, Hui
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2682 - 2686