Multimodal sentiment analysis with unimodal label generation and modality decomposition

被引:1
|
作者
Zhu, Linan [1 ]
Zhao, Hongyan [1 ]
Zhu, Zhechao [1 ]
Zhang, Chenwei [2 ]
Kong, Xiangjie [1 ]
机构
[1] Zhejiang Univ Technol, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[2] Univ Hong Kong, Sch Fac Educ, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Multimodal sentiment analysis; Unimodal label generation; Modality decomposition; FUSION;
D O I
10.1016/j.inffus.2024.102787
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal sentiment analysis aims to combine information from different modalities to enhance the understanding of emotions and achieve accurate prediction. However, existing methods face issues of information redundancy and modality heterogeneity during the fusion process, and common multimodal sentiment analysis datasets lack unimodal labels. To address these issues, this paper proposes a multimodal sentiment analysis approach based on unimodal label generation and modality decomposition (ULMD). This method employs a multi-task learning framework, dividing the multimodal sentiment analysis task into a multimodal task and three unimodal tasks. Additionally, a modality representation separator is introduced to decompose modality representations into modality-invariant representations and modality-specific representations. This approach explores the fusion between modalities and generates unimodal labels to enhance the performance of the multimodal sentiment analysis task. Extensive experiments on two public benchmark datasets demonstrate the effectiveness of this method.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Self-Supervised Unimodal Label Generation Strategy Using Recalibrated Modality Representations for Multimodal Sentiment Analysis
    Hwang, Yewon
    Kim, Jong-Hwan
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 35 - 46
  • [2] Multimodal Sentiment Analysis with Temporal Modality Modality
    Qian, Fan
    Han, Jiqing
    INTERSPEECH 2021, 2021, : 3385 - 3389
  • [3] Joint training strategy of unimodal and multimodal for multimodal sentiment analysis
    Li, Meng
    Zhu, Zhenfang
    Li, Kefeng
    Zhou, Lihua
    Zhao, Zhen
    Pei, Hongli
    IMAGE AND VISION COMPUTING, 2024, 149
  • [4] Multimodal sentiment analysis with unidirectional modality translation
    Yang, Bo
    Shao, Bo
    Wu, Lijun
    Lin, Xiaola
    NEUROCOMPUTING, 2022, 467 : 130 - 137
  • [5] Analyzing Modality Robustness in Multimodal Sentiment Analysis
    Hazarika, Devamanyu
    Li, Yingting
    Cheng, Bo
    Zhao, Shuai
    Zimmermann, Roger
    Pone, Soujanya
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 685 - 696
  • [6] Multimodal transformer with adaptive modality weighting for multimodal sentiment analysis
    Wang, Yifeng
    He, Jiahao
    Wang, Di
    Wang, Quan
    Wan, Bo
    Luo, Xuemei
    NEUROCOMPUTING, 2024, 572
  • [7] Arabic language investigation in the context of unimodal and multimodal sentiment analysis
    Youcef, Fatima Zohra
    Barigou, Fatiha
    2021 22ND INTERNATIONAL ARAB CONFERENCE ON INFORMATION TECHNOLOGY (ACIT), 2021, : 19 - 25
  • [8] Adaptive Modality Distillation for Separable Multimodal Sentiment Analysis
    Peng, Wei
    Zhao, Guoying
    Hong, Xiaopeng
    IEEE INTELLIGENT SYSTEMS, 2021, 36 (03) : 82 - 89
  • [9] Active Exploration of Modality Complementarity for Multimodal Sentiment Analysis
    Xu, Junjun
    Song, Peipei
    2024 2ND ASIA CONFERENCE ON COMPUTER VISION, IMAGE PROCESSING AND PATTERN RECOGNITION, CVIPPR 2024, 2024,
  • [10] A Unimodal Reinforced Transformer With Time Squeeze Fusion for Multimodal Sentiment Analysis
    He, Jiaxuan
    Mai, Sijie
    Hu, Haifeng
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 992 - 996