Video multimodal sentiment analysis using cross-modal feature translation and dynamical propagation

被引:2
|
作者
Gan, Chenquan [1 ,2 ,3 ]
Tang, Yu [1 ]
Fu, Xiang [1 ]
Zhu, Qingyi [2 ]
Jain, Deepak Kumar [4 ,5 ]
Garcia, Salvador [6 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Commun & Informat Engn, Chongqing 400065, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Sch Cyber Secur & Informat Law, Chongqing 400065, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Key Lab Big Data Intelligent Comp, Chongqing 400065, Peoples R China
[4] Dalian Univ Technol, Key Lab Intelligent Control & Optimizat Ind Equipm, Minist Educ, Dalian 116024, Peoples R China
[5] Symbiosis Int Univ, Symbiosis Inst Technol, Pune 412115, India
[6] Univ Granada, Andalusian Res Inst Data Sci & Computat Intelligen, Dept Comp Sci & Artificial Intelligence, Granada 18071, Spain
关键词
Video multimodal sentiment analysis; Public emotion feature; Cross-modal feature translation; Dynamical propagation model;
D O I
10.1016/j.knosys.2024.111982
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal sentiment analysis on social platforms is crucial for comprehending public opinions and attitudes, thus garnering substantial interest in knowledge engineering. Existing methods like implicit interaction, explicit interaction, and cross -modal translation can effectively integrate sentiment information, but they encounter challenges in establishing efficient emotional correlations across modalities due to data heterogeneity and concealed emotional relationships. To tackle this issue, we propose a video multimodal sentiment analysis model called PEST, which leverages cross -modal feature translation and a dynamic propagation model. Specifically, cross -modal feature translation translates textual, visual, and acoustic features into a common feature space, eliminating heterogeneity and enabling initial modal interaction. Additionally, the dynamic propagation model facilitates in-depth interaction and aids in establishing stable and reliable emotional correlations across modalities. Extensive experiments on the three multimodal sentiment datasets, CMU-MOSI, CMU-MOSEI, and CH-SIMS, demonstrate that PEST exhibits superior performance in both word -aligned and unaligned settings.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Cross-Modal Modulating for Multimodal Sentiment Analysis
    Cheng, Zichen
    Li, Yan
    Ge, Jiangwei
    Jiu, Mengfei
    Zhang, Jingwei
    Computer Engineering and Applications, 2023, 59 (10) : 171 - 179
  • [2] Video-Based Cross-Modal Auxiliary Network for Multimodal Sentiment Analysis
    Chen, Rongfei
    Zhou, Wenju
    Li, Yang
    Zhou, Huiyu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (12) : 8703 - 8716
  • [3] Cross-Modal Enhancement Network for Multimodal Sentiment Analysis
    Wang, Di
    Liu, Shuai
    Wang, Quan
    Tian, Yumin
    He, Lihuo
    Gao, Xinbo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4909 - 4921
  • [4] Hybrid cross-modal interaction learning for multimodal sentiment analysis
    Fu, Yanping
    Zhang, Zhiyuan
    Yang, Ruidi
    Yao, Cuiyou
    NEUROCOMPUTING, 2024, 571
  • [5] CMJRT: Cross-Modal Joint Representation Transformer for Multimodal Sentiment Analysis
    Xu, Meng
    Liang, Feifei
    Su, Xiangyi
    Fang, Cheng
    IEEE ACCESS, 2022, 10 : 131671 - 131679
  • [6] Multimodal Sentiment Analysis Based on Cross-Modal Joint-Encoding
    Sun, Bin
    Jiang, Tao
    Jia, Li
    Cui, Yiming
    Computer Engineering and Applications, 2024, 60 (18) : 208 - 216
  • [7] Multimodal GRU with directed pairwise cross-modal attention for sentiment analysis
    Zhenkai Qin
    Qining Luo
    Zhidong Zang
    Hongpeng Fu
    Scientific Reports, 15 (1)
  • [8] Multimodal Sentiment Analysis Based on a Cross-Modal Multihead Attention Mechanism
    Deng, Lujuan
    Liu, Boyi
    Li, Zuhe
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 78 (01): : 1157 - 1170
  • [9] CFMISA: Cross-Modal Fusion of Modal Invariant and Specific Representations for Multimodal Sentiment Analysis
    Xia, Haiying
    Chen, Jingwen
    Tan, Yumei
    Tang, Xiaohu
    PATTERN RECOGNITION AND COMPUTER VISION, PT III, PRCV 2024, 2025, 15033 : 423 - 437
  • [10] Cross-modal contrastive learning for multimodal sentiment recognition
    Yang, Shanliang
    Cui, Lichao
    Wang, Lei
    Wang, Tao
    APPLIED INTELLIGENCE, 2024, 54 (05) : 4260 - 4276