Attention-based multimodal contextual fusion for sentiment and emotion classification using bidirectional LSTM

被引:0
|
作者
Mahesh G. Huddar
Sanjeev S. Sannakki
Vijay S. Rajpurohit
机构
[1] Hirasugar Institute of Technology,Department of Computer Science and Engineering
[2] Gogte Institute of Technology,Department of Computer Science and Engineering
来源
关键词
Multimodal fusion; Contextual information; Attention model; Bidirectional LSTM;
D O I
暂无
中图分类号
学科分类号
摘要
Due to the availability of an enormous amount of multimodal content on the social web and its applications, automatic sentiment analysis, and emotion detection has become an important and widely researched topic. Improving the quality of multimodal fusion is an important issue in this field of research. In this paper, we present a novel attention-based multimodal contextual fusion strategy, which extract the contextual information among the utterances before fusion. Initially, we fuse two-two modalities at a time and finally, we fuse all three modalities. We use a bidirectional LSTM with an attention model for extracting important contextual information among the utterances. The proposed model was tested on IEMOCAP dataset for emotion classification and CMU-MOSI dataset for sentiment classification. By incorporating the contextual information among utterances in the same video, our proposed method outperforms the existing methods by over 3% in emotion classification and over 2% in sentiment classification.
引用
收藏
页码:13059 / 13076
页数:17
相关论文
共 50 条
  • [31] Attention-based bidirectional LSTM with embedding technique for classification of COVID-19 articles
    Dutta, Rakesh
    Majumder, Mukta
    INTELLIGENT DECISION TECHNOLOGIES-NETHERLANDS, 2022, 16 (01): : 205 - 215
  • [32] Multi-attention Fusion for Multimodal Sentiment Classification
    Li, Guangmin
    Zeng, Xin
    Chen, Chi
    Zhou, Long
    PROCEEDINGS OF 2024 ACM ICMR WORKSHOP ON MULTIMODAL VIDEO RETRIEVAL, ICMR-MVR 2024, 2024, : 1 - 7
  • [33] Attention-based bidirectional LSTM with embedding technique for classification of COVID-19 articles
    Dutta, Rakesh
    Majumder, Mukta
    Intelligent Decision Technologies, 2022, 16 (01) : 205 - 215
  • [34] Gated attention fusion network for multimodal sentiment classification
    Du, Yongping
    Liu, Yang
    Peng, Zhi
    Jin, Xingnan
    KNOWLEDGE-BASED SYSTEMS, 2022, 240
  • [35] AtCAF: Attention-based causality-aware fusion network for multimodal sentiment analysis
    Huang, Changqin
    Chen, Jili
    Huang, Qionghao
    Wang, Shijin
    Tu, Yaxin
    Huang, Xiaodi
    INFORMATION FUSION, 2025, 114
  • [36] Siamese Attention-Based LSTM for Speech Emotion Recognition
    Nizamidin, Tashpolat
    Zhao, Li
    Liang, Ruiyu
    Xie, Yue
    Hamdulla, Askar
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2020, E103A (07) : 937 - 941
  • [37] Attention-Based Dense LSTM for Speech Emotion Recognition
    Xie, Yue
    Liang, Ruiyu
    Liang, Zhenlin
    Zhao, Li
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2019, E102D (07): : 1426 - 1429
  • [38] Attention-Based Multimodal Fusion for Video Description
    Hori, Chiori
    Hori, Takaaki
    Lee, Teng-Yok
    Zhang, Ziming
    Harsham, Bret
    Hershey, John R.
    Marks, Tim K.
    Sumi, Kazuhiko
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 4203 - 4212
  • [39] Attention-Based Multimodal Fusion for Estimating Human Emotion in Real-World HRI
    Li, Yuanchao
    Zhao, Tianyu
    Shen, Xun
    HRI'20: COMPANION OF THE 2020 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2020, : 340 - 342
  • [40] Multimodal Emotion Detection via Attention-Based Fusion of Extracted Facial and Speech Features
    Mamieva, Dilnoza
    Abdusalomov, Akmalbek Bobomirzaevich
    Kutlimuratov, Alpamis
    Muminov, Bahodir
    Whangbo, Taeg Keun
    SENSORS, 2023, 23 (12)