TwitterBERT: Framework for Twitter Sentiment Analysis Based on Pre-trained Language Model Representations

被引:24
|
作者
Azzouza, Noureddine [1 ]
Akli-Astouati, Karima [1 ]
Ibrahim, Roliana [2 ]
机构
[1] Univ Sci & Technol Houari Boumediene, FEI Dept Comp Sci, RIIMA Lab, Algiers, Algeria
[2] Univ Teknol Malaysia UTM, Fac Engn, Sch Comp, Johor Baharu 81310, Johor, Malaysia
关键词
Twitter Sentiment Analysis; Word embedding; CNN; LSTM; BERT;
D O I
10.1007/978-3-030-33582-3_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentiment analysis has been a topic of discussion in the exploration domain of language understanding. Yet, the neural networks deployed in it are deficient to some extent. Currently, the majority of the studies proceeds on identifying the sentiments by focusing on vocabulary and syntax. Moreover, the task is recognised in Natural Language Processing (NLP) and, for calculating the noteworthy and exceptional results, Convolutional Neural Networks (CNN) and Recurrent Neural Networks (RNN) have been employed. In this study, we propose a four-phase framework for Twitter Sentiment Analysis. This setup is based on the pre-trained Bidirectional Encoder Representations from Transformers (BERT) model as an encoder for generating sentence depictions. For more effective utilisation of this model, we deploy various classification models. Additionally, we concatenate pre-trained representations of word embeddings with BERT representation method to enhance sentiment classification. Experimental results show better implementation when it is evaluated against the baseline framework on all datasets. For example, our best model attains an F1-score of 71.82% on the SemEval 2017 dataset. A comparative analysis on experimental results offers some recommendations on choosing pretraining steps to obtain improved results. The outcomes of the experiment confirm the effectiveness of our system.
引用
收藏
页码:428 / 437
页数:10
相关论文
共 50 条
  • [41] Surgicberta: a pre-trained language model for procedural surgical language
    Bombieri, Marco
    Rospocher, Marco
    Ponzetto, Simone Paolo
    Fiorini, Paolo
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2024, 18 (01) : 69 - 81
  • [42] A teacher action recognition model based on pre-trained language and video model
    Luo, Sen
    Zhou, Juxiang
    Wen, Xiaoyu
    Li, Hao
    PROCEEDINGS OF THE 15TH INTERNATIONAL CONFERENCE ON EDUCATION TECHNOLOGY AND COMPUTERS, ICETC 2023, 2023, : 335 - 340
  • [43] Data Augmentation Based on Pre-trained Language Model for Event Detection
    Zhang, Meng
    Xie, Zhiwen
    Liu, Jin
    CCKS 2021 - EVALUATION TRACK, 2022, 1553 : 59 - 68
  • [44] Detection of Chinese Deceptive Reviews Based on Pre-Trained Language Model
    Weng, Chia-Hsien
    Lin, Kuan-Cheng
    Ying, Jia-Ching
    APPLIED SCIENCES-BASEL, 2022, 12 (07):
  • [45] Neural Transfer Learning For Vietnamese Sentiment Analysis Using Pre-trained Contextual Language Models
    An Pha Le
    Tran Vu Pham
    Thanh-Van Le
    Huynh, Duy, V
    2021 IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLIED NETWORK TECHNOLOGIES (ICMLANT II), 2021, : 84 - 88
  • [46] Fusion Pre-trained Emoji Feature Enhancement for Sentiment Analysis
    Chen, Jie
    Yao, Zhiqiang
    Zhao, Shu
    Zhang, Yanping
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (04)
  • [47] HORNET: Enriching Pre-trained Language Representations with Heterogeneous Knowledge Sources
    Zhang, Taolin
    Cai, Zerui
    Wang, Chengyu
    Li, Peng
    Li, Yang
    Qiu, Minghui
    Tang, Chengguang
    He, Xiaofeng
    Huang, Jun
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2608 - 2617
  • [48] Diffused Redundancy in Pre-trained Representations
    Nanda, Vedant
    Speicher, Till
    Dickerson, John P.
    Gummadi, Krishna P.
    Feizi, Soheil
    Weller, Adrian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [49] Pre-trained Affective Word Representations
    Chawla, Kushal
    Khosla, Sopan
    Chhaya, Niyati
    Jaidka, Kokil
    2019 8TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII), 2019,
  • [50] Learning to Select Pre-trained Deep Representations with Bayesian Evidence Framework
    Kim, Yong-Deok
    Jang, Taewoong
    Han, Bohyung
    Choi, Seungjin
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 5318 - 5326