Task-adaptive Pre-training and Self-training are Complementary for Natural Language Understanding

被引:0
|
作者
Li, Shiyang [1 ]
Yavuz, Semih [2 ]
Chen, Wenhu [1 ,3 ]
Yan, Xifeng [1 ]
机构
[1] UC Santa Barbara, Santa Barbara, CA 93106 USA
[2] Salesforce Res, Menlo Pk, CA USA
[3] Google Research, New York, NY USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Task-adaptive pre-training (TAPT) and Selftraining (ST) have emerged as the major semisupervised approaches to improve natural language understanding (NLU) tasks with massive amount of unlabeled data. However, it's unclear whether they learn similar representations or they can be effectively combined. In this paper, we show that TAPT and ST can be complementary with simple TFS protocol by following TAPT ! Finetuning ! Selftraining (TFS) process. Experimental results show that TFS protocol can effectively utilize unlabeled data to achieve strong combined gains consistently across six datasets covering sentiment classification, paraphrase identification, natural language inference, named entity recognition and dialogue slot classification. We investigate various semi-supervised settings and consistently show that gains from TAPT and ST can be strongly additive by following TFS procedure. We hope that TFS could serve as an important semi-supervised baseline for future NLP studies.
引用
收藏
页码:1006 / 1015
页数:10
相关论文
共 50 条
  • [1] Self-training Improves Pre-training for Natural Language Understanding
    Du, Jingfei
    Grave, Edouard
    Gunel, Beliz
    Chaudhary, Vishrav
    Celebi, Onur
    Auli, Michael
    Stoyanov, Veselin
    Conneau, Alexis
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5408 - 5418
  • [2] SELF-TRAINING AND PRE-TRAINING ARE COMPLEMENTARY FOR SPEECH RECOGNITION
    Xu, Qiantong
    Baevski, Alexei
    Likhomanenko, Tatiana
    Tomasello, Paden
    Conneau, Alexis
    Collobert, Ronan
    Synnaeve, Gabriel
    Auli, Michael
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3030 - 3034
  • [3] Task-adaptive Pre-training of Language Models withWord Embedding Regularization
    Nishida, Kosuke
    Nishida, Kyosuke
    Yoshida, Sen
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 4546 - 4553
  • [4] Rethinking Pre-training and Self-training
    Zoph, Barret
    Ghiasi, Golnaz
    Lin, Tsung-Yi
    Cui, Yin
    Liu, Hanxiao
    Cubuk, Ekin D.
    Le, Quoc V.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [5] Code Question Answering via Task-Adaptive Sequence-to-Sequence Pre-training
    Yu, Tingrui
    Gu, Xiaodong
    Shen, Beijun
    2022 29TH ASIA-PACIFIC SOFTWARE ENGINEERING CONFERENCE, APSEC, 2022, : 229 - 238
  • [6] Unified Language Model Pre-training for Natural Language Understanding and Generation
    Dong, Li
    Yang, Nan
    Wang, Wenhui
    Wei, Furu
    Liu, Xiaodong
    Wang, Yu
    Gao, Jianfeng
    Zhou, Ming
    Hon, Hsiao-Wuen
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [7] A pre-training and self-training approach for biomedical named entity recognition
    Gao, Shang
    Kotevska, Olivera
    Sorokine, Alexandre
    Christian, J. Blair
    PLOS ONE, 2021, 16 (02):
  • [8] Self-training Improves Pre-training for Few-shot Learning in Task-oriented Dialog Systems
    Mi, Fei
    Zhou, Wanhao
    Cai, Fengyu
    Kong, Lingjing
    Huang, Minlie
    Faltings, Boi
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1887 - 1898
  • [9] Unsupervised Video Domain Adaptation with Masked Pre-Training and Collaborative Self-Training
    Reddy, Arun
    Paul, William
    Rivera, Corban
    Shah, Ketul
    de Melo, Celso M.
    Chellappa, Rama
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 18919 - 18929
  • [10] MVP: Multi-task Supervised Pre-training for Natural Language Generation
    Tang, Tianyi
    Li, Junyi
    Zhao, Wayne Xin
    Wen, Ji-Rong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8758 - 8794