ArabBert-LSTM: improving Arabic sentiment analysis based on transformer model and Long Short-Term Memory

被引:0
|
作者
Alosaimi, Wael [1 ]
Saleh, Hager [2 ,3 ,4 ]
Hamzah, Ali A. [5 ]
El-Rashidy, Nora [6 ]
Alharb, Abdullah [1 ]
Elaraby, Ahmed [7 ]
Mostafa, Sherif [2 ]
机构
[1] Taif Univ, Coll Comp & Informat Technol, Dept Informat Technol, Taif, Saudi Arabia
[2] South Valley Univ, Fac Comp & Artificial Intelligence, Hurghada, Egypt
[3] Galway Univ, Data Sci Inst, Galway, Ireland
[4] Atlantic Technol Univ, Letterkenny, Ireland
[5] Ahram Canadian Univ, 6th Of October City, Egypt
[6] Kafrelsheiksh Univ, Fac Artificial Intelligence, ML & Informat Retrieval Dept, Kafrelsheiksh, Egypt
[7] South Valley Univ, Fac Comp & Informat, Dept Comp Sci, Qena, Egypt
来源
关键词
sentiment analysis; transformer models; deep learning; machine learning; Arabic sentiment analysis; Long Short-Term Memory;
D O I
10.3389/frai.2024.1408845
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentiment analysis also referred to as opinion mining, plays a significant role in automating the identification of negative, positive, or neutral sentiments expressed in textual data. The proliferation of social networks, review sites, and blogs has rendered these platforms valuable resources for mining opinions. Sentiment analysis finds applications in various domains and languages, including English and Arabic. However, Arabic presents unique challenges due to its complex morphology characterized by inflectional and derivation patterns. To effectively analyze sentiment in Arabic text, sentiment analysis techniques must account for this intricacy. This paper proposes a model designed using the transformer model and deep learning (DL) techniques. The word embedding is represented by Transformer-based Model for Arabic Language Understanding (ArabBert), and then passed to the AraBERT model. The output of AraBERT is subsequently fed into a Long Short-Term Memory (LSTM) model, followed by feedforward neural networks and an output layer. AraBERT is used to capture rich contextual information and LSTM to enhance sequence modeling and retain long-term dependencies within the text data. We compared the proposed model with machine learning (ML) algorithms and DL algorithms, as well as different vectorization techniques: term frequency-inverse document frequency (TF-IDF), ArabBert, Continuous Bag-of-Words (CBOW), and skipGrams using four Arabic benchmark datasets. Through extensive experimentation and evaluation of Arabic sentiment analysis datasets, we showcase the effectiveness of our approach. The results underscore significant improvements in sentiment analysis accuracy, highlighting the potential of leveraging transformer models for Arabic Sentiment Analysis. The outcomes of this research contribute to advancing Arabic sentiment analysis, enabling more accurate and reliable sentiment analysis in Arabic text. The findings reveal that the proposed framework exhibits exceptional performance in sentiment classification, achieving an impressive accuracy rate of over 97%.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] CNN-OLSTM: Convolutional Neural Network with Optimized Long Short-Term Memory Model for Twitter based Sentiment Analysis
    Vanam, Harika
    Raj, Jeberson Retna
    IETE JOURNAL OF RESEARCH, 2024, 70 (03) : 2558 - 2569
  • [42] MFOA-Bi-LSTM: An optimized bidirectional long short-term memory model for short-term traffic flow prediction
    Naheliya, Bharti
    Redhu, Poonam
    Kumar, Kranti
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2024, 634
  • [43] Comparison of Long Short-Term Memory Networks and Temporal Convolutional Networks for Sentiment Analysis
    Hekman, Samuel
    Brock, Meghan
    Khan, Md Abdullah Al Hafiz
    Zhang, Xinyue
    PROCEEDINGS OF THE 2023 ACM SOUTHEAST CONFERENCE, ACMSE 2023, 2023, : 2 - 9
  • [44] Variants of Long Short-Term Memory for Sentiment Analysis on Vietnamese Students' Feedback Corpus
    Vu Duc Nguyen
    Kiet Van Nguyen
    Ngan Luu-Thuy Nguyen
    PROCEEDINGS OF 2018 10TH INTERNATIONAL CONFERENCE ON KNOWLEDGE AND SYSTEMS ENGINEERING (KSE), 2018, : 306 - 311
  • [45] Multilayer Long Short-Term Memory (LSTM) Neural Networks in Time Series Analysis
    Malinovic, Nemanja S.
    Predic, Bratislav B.
    Roganovic, Milos
    2020 55TH INTERNATIONAL SCIENTIFIC CONFERENCE ON INFORMATION, COMMUNICATION AND ENERGY SYSTEMS AND TECHNOLOGIES (IEEE ICEST 2020), 2020, : 11 - 14
  • [46] Cascade architecture with rhetoric long short-term memory for complex sentence sentiment analysis
    Ji, Chaojie
    Wu, Hongyan
    NEUROCOMPUTING, 2020, 405 : 161 - 172
  • [47] The Efficiency of Long Short-Term Memory (LSTM) in Phenology-Based Crop Classification
    Rahimi, Ehsan
    Jung, Chuleui
    KOREAN JOURNAL OF REMOTE SENSING, 2024, 40 (01) : 57 - 69
  • [48] A Reversible-Logic based Architecture for Long Short-Term Memory (LSTM) Network
    Khalil, Kasem
    Dey, Bappaditya
    Kumar, Ashok
    Bayoumi, Magdy
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [49] Control of an active suspension system based on long short-term memory (LSTM) learning
    Dridi, Issam
    Hamza, Anis
    Ben Yahia, Noureddine
    ADVANCES IN MECHANICAL ENGINEERING, 2023, 15 (02)
  • [50] Well performance prediction based on Long Short-Term Memory (LSTM) neural network
    Huang, Ruijie
    Wei, Chenji
    Wang, Baohua
    Yang, Jian
    Xu, Xin
    Wu, Suwei
    Huang, Suqi
    JOURNAL OF PETROLEUM SCIENCE AND ENGINEERING, 2022, 208