An End-to-End Contrastive Self-Supervised Learning Framework for Language Understanding

被引:6
|
作者
Fang, Hongchao [1 ]
Xie, Pengtao [1 ]
机构
[1] Univ Calif San Diego, La Jolla, CA 92093 USA
关键词
D O I
10.1162/tacl_a_00521
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised learning (SSL) methods such as Word2vec, BERT, and GPT have shown great effectiveness in language understanding. Contrastive learning, as a recent SSL approach, has attracted increasing attention in NLP. Contrastive learning learns data representations by predicting whether two augmented data instances are generated from the same original data example. Previous contrastive learning methods perform data augmentation and contrastive learning separately. As a result, the augmented data may not be optimal for contrastive learning. To address this problem, we propose a four-level optimization framework that performs data augmentation and contrastive learning end-to-end, to enable the augmented data to be tailored to the contrastive learning task. This framework consists of four learning stages, including training machine translation models for sentence augmentation, pretraining a text encoder using contrastive learning, finetuning a text classification model, and updating weights of translation data by minimizing the validation loss of the classification model, which are performed in a unified way. Experiments on datasets in the GLUE benchmark (Wang et al., 2018a) and on datasets used in Gururangan et al. (2020) demonstrate the effectiveness of our method.
引用
收藏
页码:1324 / 1340
页数:17
相关论文
共 50 条
  • [31] Learning end-to-end patient representations through self-supervised covariate balancing for causal treatment effect estimation
    Tesei, Gino
    Giampanis, Stefanos
    Shi, Jingpu
    Norgeot, Beau
    JOURNAL OF BIOMEDICAL INFORMATICS, 2023, 140
  • [32] FeaRLESS: Feature Refinement Loss for Ensembling Self-Supervised Learning Features in Robust End-to-end Speech Recognition
    Chen, Szu-Jui
    Xie, Jiamin
    Hansen, John H. L.
    INTERSPEECH 2022, 2022, : 3058 - 3062
  • [33] Self-supervised End-to-End ASR for Low Resource L2 Swedish
    Al-Ghezi, Ragheb
    Getman, Yaroslav
    Rouhe, Aku
    Hilden, Raili
    Kurimo, Mikko
    INTERSPEECH 2021, 2021, : 1429 - 1433
  • [34] SAR: Self-Supervised Anti-Distortion Representation for End-To-End Speech Model
    Wang, Jianzong
    Zhang, Xulong
    Tang, Haobin
    Sun, Aolan
    Cheng, Ning
    Xiao, Jing
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [35] Adversarial Self-Supervised Contrastive Learning
    Kim, Minseon
    Tack, Jihoon
    Hwang, Sung Ju
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS (NEURIPS 2020), 2020, 33
  • [36] A Survey on Contrastive Self-Supervised Learning
    Jaiswal, Ashish
    Babu, Ashwin Ramesh
    Zadeh, Mohammad Zaki
    Banerjee, Debapriya
    Makedon, Fillia
    TECHNOLOGIES, 2021, 9 (01)
  • [37] Self-Supervised Learning: Generative or Contrastive
    Liu, Xiao
    Zhang, Fanjin
    Hou, Zhenyu
    Mian, Li
    Wang, Zhaoyu
    Zhang, Jing
    Tang, Jie
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (01) : 857 - 876
  • [38] SCL-Net: An End-to-End Supervised Contrastive Learning Network for Hyperspectral Image Classification
    Lu, Ting
    Hu, Yaochen
    Fu, Wei
    Ding, Kexin
    Bai, Beifang
    Fang, Leyuan
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [39] END-TO-END JOINT LEARNING OF NATURAL LANGUAGE UNDERSTANDING AND DIALOGUE MANAGER
    Yang, Xuesong
    Chen, Yun-Nung
    Hakkani-Tur, Dilek
    Crook, Paul
    Li, Xiujun
    Gao, Jianfeng
    Deng, Li
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 5690 - 5694
  • [40] Self-supervised Visual Feature Learning and Classification Framework: Based on Contrastive Learning
    Wang, Zhibo
    Yan, Shen
    Zhang, Xiaoyu
    Lobo, Niels Da Vitoria
    16TH IEEE INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV 2020), 2020, : 719 - 725