Integrating Task Specific Information into Pretrained Language Models for Low Resource Fine Tuning

被引:0
|
作者
Wang, Rui [1 ]
Si, Shijing [1 ]
Wang, Guoyin [1 ,2 ]
Zhang, Lei [3 ]
Carin, Lawrence [1 ]
Henao, Ricardo [1 ]
机构
[1] Duke Univ, Durham, NC 27706 USA
[2] Amazon Alexa AI, Cambridge, MA USA
[3] Fidel Investments, Raleigh, NC USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained Language Models (PLMs) have improved the performance of natural language understanding in recent years. Such models are pretrained on large corpora, which encode the general prior knowledge of natural languages but are agnostic to information characteristic of downstream tasks. This often results in overfitting when fine-tuned with low resource datasets where task-specific information is limited. In this paper, we integrate label information as a task-specific prior into the self-attention component of pretrained BERT models. Experiments on several benchmarks and real-word datasets suggest that the proposed approach can largely improve the performance of pretrained models when finetuning with small datasets. The code repository is released in https://github.com/RayWangWR/BERT_label_embedding.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] ACTUNE: Uncertainty-Based Active Self-Training for Active Fine-Tuning of Pretrained Language Models
    Yu, Yue
    Kong, Lingkai
    Zhang, Jieyu
    Zhang, Rongzhi
    Zhang, Chao
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1422 - 1436
  • [22] Causal-Debias: Unifying Debiasing in Pretrained Language Models and Fine-tuning via Causal Invariant Learning
    Zhou, Fan
    Mao, Yuzhou
    Yu, Liu
    Yang, Yi
    Zhong, Ting
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 4227 - 4241
  • [23] Leveraging Pretrained Language Models for Enhanced Entity Matching: A Comprehensive Study of Fine-Tuning and Prompt Learning Paradigms
    Wang, Yu
    Zhou, Luyao
    Wang, Yuan
    Peng, Zhenwan
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2024, 2024
  • [24] Towards Low-Resource Automatic Program Repair with Meta-Learning and Pretrained Language Models
    Wang, Weishi
    Wang, Yue
    Hoi, Steven C. H.
    Joty, Shafiq
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 6954 - 6968
  • [25] AgglutiFiT: Efficient Low-Resource Agglutinative Language Model Fine-Tuning
    Li, Zhe
    Li, Xiuhong
    Sheng, Jiabao
    Slamu, Wushour
    IEEE ACCESS, 2020, 8 : 148489 - 148499
  • [26] Evaluating the Effectiveness of Fine-Tuning Large Language Model for Domain-Specific Task
    Dabhi, Saumya
    Martinez, Joseph
    Poursardar, Faryaneh
    2024 IEEE INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION FOR DATA SCIENCE, IRI 2024, 2024, : 176 - 177
  • [27] Health Care Language Models and Their Fine-Tuning for Information Extraction: Scoping Review
    Nunes, Miguel
    Bone, Joao
    Ferreira, Joao C.
    Elvas, Luis B.
    JMIR MEDICAL INFORMATICS, 2024, 12
  • [28] Fine-tuning Happens in Tiny Subspaces: Exploring Intrinsic Task-specific Subspaces of Pre-trained Language Models
    Zhang, Zhong
    Liu, Bang
    Shao, Junming
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 1701 - 1713
  • [29] Rebetiko Singer Identification: Fine-tuning and explaining deep pretrained transformer models
    Papakostas, Maximos Kaliakatsos
    Zacharakis, Asterios
    Velenis, Konstantinos
    Cambouropoulos, Emilios
    PROCEEDINGS OF THE 19TH INTERNATIONAL AUDIO MOSTLY CONFERENCE, AM 2024, 2024, : 285 - 291
  • [30] GO BEYOND PLAIN FINE-TUNING: IMPROVING PRETRAINED MODELS FOR SOCIAL COMMONSENSE
    Chang, Ting-Yun
    Liu, Yang
    Gopalakrishnan, Karthik
    Hedayatnia, Behnam
    Zhou, Pei
    Hakkani-Tur, Dilek
    2021 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP (SLT), 2021, : 1028 - 1035