Weakly-Supervised Contrastive Learning Framework for Few-Shot Sentiment Classification Tasks

被引:0
|
作者
Lu S. [1 ]
Chen L. [1 ]
Lu G. [1 ]
Guan Z. [2 ]
Xie F. [3 ]
机构
[1] School of Communications and Information Engineering, Xi'an University of Posts & Telecommunications, Xi'an
[2] School of Computer Science and Technology, Xidian University, Xi'an
[3] Academy of Advanced Interdisciplinary Research, Xidian University, Xi'an
基金
中国国家自然科学基金;
关键词
Few-shot learning; Sentiment classification; Supervised contrastive learning; Transfer learning; Weakly-supervised learning;
D O I
10.7544/issn1000-1239.20210699
中图分类号
学科分类号
摘要
Text sentiment classification is a challenge research topic in natural language processing. Lexicon-based methods and traditional machine learning-based methods rely on high-quality sentiment lexicon and robust feature engineering respectively, whereas most deep learning methods are heavily reliant on large human-annotated data sets. Fortunately, users on various social platforms generate massive amounts of tagged opinioned texts which can be deemed as weakly-labeled data for sentiment classification. However, noisy labeled instances in weakly-labeled data have a negative impact on the training phase. In this paper, we present a weakly-supervised contrastive learning framework for few-shot sentiment classification that learns the sentiment semantics from large user-tagged data with noisy labels while also exploiting inter-class contrastive patterns hidden in small labeled data. The framework consists of two steps: first, we design a weakly-supervised pre-training strategy to reduce the influence of the noisy labeled samples, and then the contrastive strategy is used in supervised fine-tuning to capture the contrast patterns in the small labeled data. The experimental results on Amazon review data set show that our approach outperforms the other baseline methods. When fine-tuned on only 0.5% (i.e. 32 samples) of the labels, we achieve comparable performance among the deep baselines, showing its robustness in the data sparsity scenario. © 2022, Science Press. All right reserved.
引用
收藏
页码:2003 / 2014
页数:11
相关论文
共 47 条
  • [1] Nasukawa T, Yi J., Sentiment analysis: Capturing favorability using natural language processing, Proc of the 2nd Int Conf on Knowledge Capture, pp. 70-77, (2003)
  • [2] Pang Bo, Lee L, Vaithyanathan S., Thumbs up: Sentiment classification using machine learning techniques, Proc of the Conf on Empirical Methods in Natural Language Processing, pp. 79-86, (2002)
  • [3] Dave K, Lawrence S, Pennock D M., Mining the peanut gallery: Opinion extraction and semantic classification of product reviews, Proc of the 12th Int World Wide Web Conf, pp. 519-528, (2003)
  • [4] Glorot X, Bordes A, Bengio Y., Domain adaptation for large-scale sentiment classification: A deep learning approach, Proc of the 28th Int Conf on Machine Learning, pp. 513-520, (2011)
  • [5] Kim Y., Convolutional neural networks for sentence classification, (2014)
  • [6] Mikolov T, Chen Kai, Corrado G, Et al., Efficient estimation of word representations in vector space, (2013)
  • [7] Pennington J, Socher R, Manning C D., Glove: Global vectors for word representation, Proc of the Conf on Empirical Methods in Natural Language Processing, pp. 1532-1543, (2014)
  • [8] Devlin J, Chang Mingwei, Lee K, Et al., BERT: Pre-training of deep bidirectional transformers for language understanding, (2018)
  • [9] Deriu J, Lucchi A, De Luca V, Et al., Leveraging large amounts of weakly supervised data for multi-language sentiment classification, Proc of the 26th Int Conf on World Wide Web, pp. 1045-1052, (2017)
  • [10] Chen Ting, Kornblith S, Norouzi M, Et al., A simple framework for contrastive learning of visual representations, (2020)