Attention fusion network for multimodal sentiment analysis

被引:0
|
作者
Yuanyi Luo
Rui Wu
Jiafeng Liu
Xianglong Tang
机构
[1] Harbin Institute of Technology,
来源
关键词
Multimodal sentiment analysis; Attention mechanism; Multimodal fusion;
D O I
暂无
中图分类号
学科分类号
摘要
The main research problem in multimodal sentiment analysis is to model inter-modality dynamics. However, most of the current work cannot consider enough in this aspect. In this study, we propose a multimodal fusion network MSA-AFN, which considers both multimodal relationships and differences in modal contributions to the task. Specifically, in the feature extraction process, we consider not only the relationship between audio and text, but also the contribution of temporal features to the task. In the process of multimodal fusion, based on the soft attention mechanism, the feature representation of each modality is weighted and connected according to their contribution to the task. We evaluate our proposed approach on the Chinese multimodal sentiment analysis dataset: CH-SIMS. Results show that our model achieves better results than comparison models. Moreover, the performance of some baselines has been improved by 0.28% to 9.5% after adding the component of our network.
引用
收藏
页码:8207 / 8217
页数:10
相关论文
共 50 条
  • [21] Global Local Fusion Neural Network for Multimodal Sentiment Analysis
    Hu, Xiaoran
    Yamamura, Masayuki
    APPLIED SCIENCES-BASEL, 2022, 12 (17):
  • [22] Multi-attention Fusion for Multimodal Sentiment Classification
    Li, Guangmin
    Zeng, Xin
    Chen, Chi
    Zhou, Long
    PROCEEDINGS OF 2024 ACM ICMR WORKSHOP ON MULTIMODAL VIDEO RETRIEVAL, ICMR-MVR 2024, 2024, : 1 - 7
  • [23] Various syncretic co-attention network for multimodal sentiment analysis
    Cao, Meng
    Zhu, Yonghua
    Gao, Wenjing
    Li, Mengyao
    Wang, Shaoxiu
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2020, 32 (24):
  • [24] Multi-Level Attention Map Network for Multimodal Sentiment Analysis
    Xue, Xiaojun
    Zhang, Chunxia
    Niu, Zhendong
    Wu, Xindong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (05) : 5105 - 5118
  • [25] TETFN: A text enhanced transformer fusion network for multimodal sentiment analysis
    Wang, Di
    Guo, Xutong
    Tian, Yumin
    Liu, Jinhui
    He, LiHuo
    Luo, Xuemei
    PATTERN RECOGNITION, 2023, 136
  • [26] SCANET: Improving multimodal representation and fusion with sparse- and cross-attention for multimodal sentiment analysis
    Wang, Hao
    Yang, Mingchuan
    Li, Zheng
    Liu, Zhenhua
    Hu, Jie
    Fu, Ziwang
    Liu, Feng
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2022, 33 (3-4)
  • [27] Prompt Link Multimodal Fusion in Multimodal Sentiment Analysis
    Zhu, Kang
    Fan, Cunhang
    Tao, Jianhua
    Lv, Zhao
    INTERSPEECH 2024, 2024, : 4668 - 4672
  • [28] MULTI-CHANNEL ATTENTIVE GRAPH CONVOLUTIONAL NETWORK WITH SENTIMENT FUSION FOR MULTIMODAL SENTIMENT ANALYSIS
    Xiao, Luwei
    Wu, Xingjiao
    Wu, Wen
    Yang, Jing
    He, Liang
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4578 - 4582
  • [29] Entity-Sensitive Attention and Fusion Network for Entity-Level Multimodal Sentiment Classification
    Yu, Jianfei
    Jiang, Jing
    Xia, Rui
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 429 - 439
  • [30] Targeted Aspect-Based Multimodal Sentiment Analysis: An Attention Capsule Extraction and Multi-Head Fusion Network
    Gu, Donghong
    Wang, Jiaqian
    Cai, Shaohua
    Yang, Chi
    Song, Zhengxin
    Zhao, Haoliang
    Xiao, Luwei
    Wang, Hua
    IEEE ACCESS, 2021, 9 : 157329 - 157336