TwitterBERT: Framework for Twitter Sentiment Analysis Based on Pre-trained Language Model Representations

被引:24
|
作者
Azzouza, Noureddine [1 ]
Akli-Astouati, Karima [1 ]
Ibrahim, Roliana [2 ]
机构
[1] Univ Sci & Technol Houari Boumediene, FEI Dept Comp Sci, RIIMA Lab, Algiers, Algeria
[2] Univ Teknol Malaysia UTM, Fac Engn, Sch Comp, Johor Baharu 81310, Johor, Malaysia
关键词
Twitter Sentiment Analysis; Word embedding; CNN; LSTM; BERT;
D O I
10.1007/978-3-030-33582-3_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentiment analysis has been a topic of discussion in the exploration domain of language understanding. Yet, the neural networks deployed in it are deficient to some extent. Currently, the majority of the studies proceeds on identifying the sentiments by focusing on vocabulary and syntax. Moreover, the task is recognised in Natural Language Processing (NLP) and, for calculating the noteworthy and exceptional results, Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) have been employed. In this study, we propose a four-phase framework for Twitter Sentiment Analysis. This setup is based on the pre-trained Bidirectional Encoder Representations from Transformers (BERT) model as an encoder for generating sentence depictions. For more effective utilisation of this model, we deploy various classification models. Additionally, we concatenate pre-trained representations of word embeddings with BERT representation method to enhance sentiment classification. Experimental results show better implementation when it is evaluated against the baseline framework on all datasets. For example, our best model attains an F1-score of 71.82% on the SemEval 2017 dataset. A comparative analysis on experimental results offers some recommendations on choosing pretraining steps to obtain improved results. The outcomes of the experiment confirm the effectiveness of our system.
引用
收藏
页码:428 / 437
页数:10
相关论文
共 50 条
  • [31] Interpretability of Entity Matching Based on Pre-trained Language Model
    Liang Z.
    Wang H.-Z.
    Dai J.-J.
    Shao X.-Y.
    Ding X.-O.
    Mu T.-Y.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (03): : 1087 - 1108
  • [32] Explainable Pre-Trained Language Models for Sentiment Analysis in Low-Resourced Languages
    Mabokela, Koena Ronny
    Primus, Mpho
    Celik, Turgay
    BIG DATA AND COGNITIVE COMPUTING, 2024, 8 (11)
  • [33] LETS: A Label-Efficient Training Scheme for Aspect-Based Sentiment Analysis by Using a Pre-Trained Language Model
    Shim, Heereen
    Lowet, Dietwig
    Luca, Stijn
    Vanrumste, Bart
    IEEE ACCESS, 2021, 9 : 115563 - 115578
  • [34] Adder Encoder for Pre-trained Language Model
    Ding, Jianbang
    Zhang, Suiyun
    Li, Linlin
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2023, 2023, 14232 : 339 - 347
  • [35] Fine-Grained Sentiment-Controlled Text Generation Approach Based on Pre-Trained Language Model
    Zhu, Linan
    Xu, Yifei
    Zhu, Zhechao
    Bao, Yinwei
    Kong, Xiangjie
    APPLIED SCIENCES-BASEL, 2023, 13 (01):
  • [36] A Study of Vietnamese Sentiment Classification with Ensemble Pre-trained Language Models
    Thin, Dang Van
    Hao, Duong Ngoc
    Nguyen, Ngan Luu-Thuy
    VIETNAM JOURNAL OF COMPUTER SCIENCE, 2024, 11 (01) : 137 - 165
  • [37] A Light Bug Triage Framework for Applying Large Pre-trained Language Model
    Lee, Jaehyung
    Han, Kisun
    Yu, Hwanjo
    PROCEEDINGS OF THE 37TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE 2022, 2022,
  • [38] Robust Sentiment Classification of Metaverse Services Using a Pre-trained Language Model with Soft Voting
    Lee, Haein
    Jung, Hae Sun
    Lee, Seon Hong
    Kim, Jang Hyun
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2023, 17 (09): : 2334 - 2347
  • [39] The Biases of Pre-Trained Language Models: An Empirical Study on Prompt-Based Sentiment Analysis and Emotion Detection
    Mao, Rui
    Liu, Qian
    He, Kai
    Li, Wei
    Cambria, Erik
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 1743 - 1753
  • [40] Schema matching based on energy domain pre-trained language model
    Pan Z.
    Yang M.
    Monti A.
    Energy Informatics, 2023, 6 (Suppl 1)