Video multimodal sentiment analysis using cross-modal feature translation and dynamical propagation

被引:2
|
作者
Gan, Chenquan [1 ,2 ,3 ]
Tang, Yu [1 ]
Fu, Xiang [1 ]
Zhu, Qingyi [2 ]
Jain, Deepak Kumar [4 ,5 ]
Garcia, Salvador [6 ]
机构
[1] Chongqing Univ Posts & Telecommun, Sch Commun & Informat Engn, Chongqing 400065, Peoples R China
[2] Chongqing Univ Posts & Telecommun, Sch Cyber Secur & Informat Law, Chongqing 400065, Peoples R China
[3] Chongqing Univ Posts & Telecommun, Key Lab Big Data Intelligent Comp, Chongqing 400065, Peoples R China
[4] Dalian Univ Technol, Key Lab Intelligent Control & Optimizat Ind Equipm, Minist Educ, Dalian 116024, Peoples R China
[5] Symbiosis Int Univ, Symbiosis Inst Technol, Pune 412115, India
[6] Univ Granada, Andalusian Res Inst Data Sci & Computat Intelligen, Dept Comp Sci & Artificial Intelligence, Granada 18071, Spain
关键词
Video multimodal sentiment analysis; Public emotion feature; Cross-modal feature translation; Dynamical propagation model;
D O I
10.1016/j.knosys.2024.111982
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal sentiment analysis on social platforms is crucial for comprehending public opinions and attitudes, thus garnering substantial interest in knowledge engineering. Existing methods like implicit interaction, explicit interaction, and cross -modal translation can effectively integrate sentiment information, but they encounter challenges in establishing efficient emotional correlations across modalities due to data heterogeneity and concealed emotional relationships. To tackle this issue, we propose a video multimodal sentiment analysis model called PEST, which leverages cross -modal feature translation and a dynamic propagation model. Specifically, cross -modal feature translation translates textual, visual, and acoustic features into a common feature space, eliminating heterogeneity and enabling initial modal interaction. Additionally, the dynamic propagation model facilitates in-depth interaction and aids in establishing stable and reliable emotional correlations across modalities. Extensive experiments on the three multimodal sentiment datasets, CMU-MOSI, CMU-MOSEI, and CH-SIMS, demonstrate that PEST exhibits superior performance in both word -aligned and unaligned settings.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Multimodal Sentiment Analysis Network Based on Distributional Transformation and Gated Cross-Modal Fusion
    Zhang, Yuchen
    Thong, Hong
    Chen, Guilin
    Alhusaini, Naji
    Zhao, Shenghui
    Wu, Cheng
    2024 INTERNATIONAL CONFERENCE ON NETWORKING AND NETWORK APPLICATIONS, NANA 2024, 2024, : 496 - 503
  • [22] Multimodal Sentiment Analysis in Realistic Environments Based on Cross-Modal Hierarchical Fusion Network
    Huang, Ju
    Lu, Pengtao
    Sun, Shuifa
    Wang, Fangyi
    ELECTRONICS, 2023, 12 (16)
  • [23] Cross-modal dynamic sentiment annotation for speech sentiment analysis
    Chen, Jincai
    Sun, Chao
    Zhang, Sheng
    Zeng, Jiangfeng
    COMPUTERS & ELECTRICAL ENGINEERING, 2023, 106
  • [24] Cross-modal complementary network with hierarchical fusion for multimodal sentiment classification
    Peng, Cheng
    Zhang, Chunxia
    Xue, Xiaojun
    Gao, Jiameng
    Liang, Hongjian
    Niu, Zhengdong
    TSINGHUA SCIENCE AND TECHNOLOGY, 2022, 27 (04) : 664 - 679
  • [25] Cross-Modal Complementary Network with Hierarchical Fusion for Multimodal Sentiment Classification
    Cheng Peng
    Chunxia Zhang
    Xiaojun Xue
    Jiameng Gao
    Hongjian Liang
    Zhengdong Niu
    TsinghuaScienceandTechnology, 2022, 27 (04) : 664 - 679
  • [26] CiteNet: Cross-modal incongruity perception network for multimodal sentiment prediction
    Wang, Jie
    Yang, Yan
    Liu, Keyu
    Xie, Zhuyang
    Zhang, Fan
    Li, Tianrui
    KNOWLEDGE-BASED SYSTEMS, 2024, 295
  • [27] Multimodal Sentiment Analysis Based on Cross-Modal Attention and Gated Cyclic Hierarchical Fusion Networks
    Quan, Zhibang
    Sun, Tao
    Su, Mengli
    Wei, Jishu
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [28] Multichannel Cross-Modal Fusion Network for Multimodal Sentiment Analysis Considering Language Information Enhancement
    Hu, Ronglong
    Yi, Jizheng
    Chen, Aibin
    Chen, Lijiang
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (07) : 9814 - 9824
  • [29] Multimodal Sentiment Analysis Method Based on Cross-Modal Attention and Gated Unit Fusion Network
    Chen, Yansong
    Zhang, Le
    Zhang, Leihan
    Lü, Xueqiang
    Data Analysis and Knowledge Discovery, 2024, 8 (07) : 67 - 76
  • [30] Cross-Modal Transformer Combination Model for Sentiment Analysis
    Wang, Liang
    Wang, Yi
    Wang, Jun
    Computer Engineering and Applications, 2024, 60 (13) : 124 - 1350