Multi-Model Fusion Framework Using Deep Learning for Visual-Textual Sentiment Classification

被引:0
|
作者
Salman Al-Tameemi I.K. [1 ,2 ,3 ]
Feizi-Derakhshi M.-R. [1 ,2 ]
Pashazadeh S. [2 ]
Asadpour M. [2 ]
机构
[1] Computerized Intelligence Systems Laboratory, Department of Computer Engineering, Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz
[2] Department of Computer Engineering, Faculty of Electrical and Computer Engineering, University of Tabriz, Tabriz
[3] State Company for Engineering Rehabilitation and Testing, Iraqi Ministry of Industry and Minerals, Baghdad
来源
Computers, Materials and Continua | 2023年 / 76卷 / 02期
关键词
decision fusion; deep learning; interpretability; joint fusion; multimodal classification; Sentiment analysis;
D O I
10.32604/CMC.2023.040997
中图分类号
学科分类号
摘要
Multimodal Sentiment Analysis (SA) is gaining popularity due to its broad application potential. The existing studies have focused on the SA of single modalities, such as texts or photos, posing challenges in effectively handling social media data with multiple modalities. Moreover, most multimodal research has concentrated on merely combining the two modalities rather than exploring their complex correlations, leading to unsatisfactory sentiment classification results. Motivated by this, we propose a new visual-textual sentiment classification model named Multi-Model Fusion (MMF), which uses a mixed fusion framework for SA to effectively capture the essential information and the intrinsic relationship between the visual and textual content. The proposed model comprises three deep neural networks. Two different neural networks are proposed to extract the most emotionally relevant aspects of image and text data. Thus, more discriminative features are gathered for accurate sentiment classification. Then, a multichannel joint fusion model with a self-attention technique is proposed to exploit the intrinsic correlation between visual and textual characteristics and obtain emotionally rich information for joint sentiment classification. Finally, the results of the three classifiers are integrated using a decision fusion scheme to improve the robustness and generalizability of the proposed model. An interpretable visual-textual sentiment classification model is further developed using the Local Interpretable Model-agnostic Explanation model (LIME) to ensure the model’s explainability and resilience. The proposed MMF model has been tested on four real-world sentiment datasets, achieving (99.78%) accuracy on Binary_Getty (BG), (99.12%) on Binary_iStock (BIS), (95.70%) on Twitter, and (79.06%) on the Multi-View Sentiment Analysis (MVSA) dataset. These results demonstrate the superior performance of our MMF model compared to single-model approaches and current state-of-the-art techniques based on model evaluation criteria. © 2023 Tech Science Press. All rights reserved.
引用
收藏
页码:2145 / 2177
页数:32
相关论文
共 50 条
  • [1] Multi-Model Fusion Framework Using Deep Learning for Visual-Textual Sentiment Classification
    Al-Tameemi, Israa K. Salman
    Feizi-Derakhshi, Mohammad-Reza
    Pashazadeh, Saeed
    Asadpour, Mohammad
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 76 (02): : 2145 - 2177
  • [2] A Multi-model Fusion Framework based on Deep Learning for Sentiment Classification
    Yang, Fen
    Zhu, Jia
    Wang, Xuming
    Wu, Xingcheng
    Tang, Yong
    Luo, Long
    PROCEEDINGS OF THE 2018 IEEE 22ND INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN ((CSCWD)), 2018, : 433 - 437
  • [3] Sentiment Recognition for Short Annotated GIFs Using Visual-Textual Fusion
    Liu, Tianliang
    Wan, Junwei
    Dai, Xiubin
    Liu, Feng
    You, Quanzeng
    Luo, Jiebo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (04) : 1098 - 1110
  • [4] Hybrid Representation and Decision Fusion towards Visual-textual Sentiment
    Yin, Chunyong
    Zhang, Sun
    Zeng, Qingkui
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (03)
  • [5] Joint Visual-Textual Sentiment Analysis with Deep Neural Networks
    You, Quanzeng
    Luo, Jiebo
    Jin, Hailin
    Yang, Jianchao
    MM'15: PROCEEDINGS OF THE 2015 ACM MULTIMEDIA CONFERENCE, 2015, : 1071 - 1074
  • [6] A Novel Visual-Textual Sentiment Analysis Framework for Social Media Data
    Jindal, Kanika
    Aron, Rajni
    COGNITIVE COMPUTATION, 2021, 13 (06) : 1433 - 1450
  • [7] A multimodal fusion network with attention mechanisms for visual-textual sentiment analysis
    Gan, Chenquan
    Fu, Xiang
    Feng, Qingdong
    Zhu, Qingyi
    Cao, Yang
    Zhu, Ye
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 242
  • [8] Visual-textual sentiment classification with bi-directional multi-level attention networks
    Xu, Jie
    Huang, Feiran
    Zhang, Xiaoming
    Wang, Senzhang
    Li, Chaozhuo
    Li, Zhoujun
    He, Yueying
    KNOWLEDGE-BASED SYSTEMS, 2019, 178 : 61 - 73
  • [9] A Novel Visual-Textual Sentiment Analysis Framework for Social Media Data
    Kanika Jindal
    Rajni Aron
    Cognitive Computation, 2021, 13 : 1433 - 1450
  • [10] END-TO-END DEEP MEMORY NETWORK FOR VISUAL-TEXTUAL SENTIMENT ANALYSIS
    Miao, Hang
    Liu, Ruifang
    Gao, Sheng
    Zhou, Xin
    He, Xiaoxin
    PROCEEDINGS OF 2018 INTERNATIONAL CONFERENCE ON NETWORK INFRASTRUCTURE AND DIGITAL CONTENT (IEEE IC-NIDC), 2018, : 399 - 403