Bimodal Fusion Network with Multi-Head Attention for Multimodal Sentiment Analysis

被引:3
|
作者
Zhang, Rui [1 ,2 ]
Xue, Chengrong [1 ,2 ]
Qi, Qingfu [3 ]
Lin, Liyuan [2 ]
Zhang, Jing [1 ,2 ]
Zhang, Lun [1 ,2 ]
机构
[1] Tianjin Sino German Univ Appl Sci, Sch Software & Commun, Tianjin 300222, Peoples R China
[2] Tianjin Univ Sci & Technol, Coll Elect Informat & Automation, Tianjin 300222, Peoples R China
[3] Gaussian Robot Pte Ltd, Tianjin 200100, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 03期
关键词
multimodal sentiment analysis; bimodal fusion; multi-head attention; EMOTION RECOGNITION; FEATURES;
D O I
10.3390/app13031915
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
The enrichment of social media expression makes multimodal sentiment analysis a research hotspot. However, modality heterogeneity brings great difficulties to effective cross-modal fusion, especially the modality alignment problem and the uncontrolled vector offset during fusion. In this paper, we propose a bimodal multi-head attention network (BMAN) based on text and audio, which adaptively captures the intramodal utterance features and complex intermodal alignment relationships. Specifically, we first set two independent unimodal encoders to extract the semantic features within each modality. Considering that different modalities deserve different weights, we further built a joint decoder to fuse the audio information into the text representation, based on learnable weights to avoid an unreasonable vector offset. The obtained cross-modal representation is used to improve the sentiment prediction performance. Experiments on both the aligned and unaligned CMU-MOSEI datasets show that our model achieves better performance than multiple baselines, and it has outstanding advantages in solving the problem of cross-modal alignment.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Multimodal sentiment analysis based on multi-head attention mechanism
    Xi, Chen
    Lu, Guanming
    Yan, Jingjie
    ICMLSC 2020: PROCEEDINGS OF THE 4TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND SOFT COMPUTING, 2020, : 34 - 39
  • [2] Targeted Aspect-Based Multimodal Sentiment Analysis: An Attention Capsule Extraction and Multi-Head Fusion Network
    Gu, Donghong
    Wang, Jiaqian
    Cai, Shaohua
    Yang, Chi
    Song, Zhengxin
    Zhao, Haoliang
    Xiao, Luwei
    Wang, Hua
    IEEE ACCESS, 2021, 9 : 157329 - 157336
  • [3] Video sentiment analysis with bimodal information-augmented multi-head attention
    Wu, Ting
    Peng, Junjie
    Zhang, Wenqiang
    Zhang, Huiran
    Tan, Shuhua
    Yi, Fen
    Ma, Chuanshuai
    Huang, Yansong
    KNOWLEDGE-BASED SYSTEMS, 2022, 235
  • [4] Attention fusion network for multimodal sentiment analysis
    Yuanyi Luo
    Rui Wu
    Jiafeng Liu
    Xianglong Tang
    Multimedia Tools and Applications, 2024, 83 : 8207 - 8217
  • [5] Attention fusion network for multimodal sentiment analysis
    Luo, Yuanyi
    Wu, Rui
    Liu, Jiafeng
    Tang, Xianglong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (03) : 8207 - 8217
  • [6] Deep Multi-Head Attention Network for Aspect-Based Sentiment Analysis
    Yan, Danfeng
    Chen, Jiyuan
    Cui, Jianfei
    Shan, Ao
    Shi, Wenting
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 695 - 700
  • [7] Multi-head attention model for aspect level sentiment analysis
    Zhang, Xinsheng
    Gao, Teng
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2020, 38 (01) : 89 - 96
  • [8] Memory network with hierarchical multi-head attention for aspect-based sentiment analysis
    Chen, Yuzhong
    Zhuang, Tianhao
    Guo, Kun
    APPLIED INTELLIGENCE, 2021, 51 (07) : 4287 - 4304
  • [9] Memory network with hierarchical multi-head attention for aspect-based sentiment analysis
    Yuzhong Chen
    Tianhao Zhuang
    Kun Guo
    Applied Intelligence, 2021, 51 : 4287 - 4304
  • [10] Multi-Task Multi-Head Attention Memory Network for Fine-Grained Sentiment Analysis
    Dai, Zehui
    Dai, Wei
    Liu, Zhenhua
    Rao, Fengyun
    Chen, Huajie
    Zhang, Guangpeng
    Ding, Yadong
    Liu, Jiyang
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 609 - 620