CubeMLP: A MLP-based Model for Multimodal Sentiment Analysis and Depression Estimation

被引:55
|
作者
Sun, Hao [1 ]
Wang, Hongyi [1 ]
Liu, Jiaqing [2 ]
Chen, Yen-Wei [2 ]
Lin, Lanfen [1 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[2] Ritsumeikan Univ, Coll Informat Sci & Engn, Kusatsu, Shiga, Japan
来源
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022 | 2022年
关键词
multimodal processing; multimodal fusion; multimodal interaction; multimedia; MLP; sentiment analysis; depression detection;
D O I
10.1145/3503161.3548025
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Multimodal sentiment analysis and depression estimation are two important research topics that aim to predict human mental states using multimodal data. Previous research has focused on developing effective fusion strategies for exchanging and integrating mind-related information from different modalities. Some MLP-based techniques have recently achieved considerable success in a variety of computer vision tasks. Inspired by this, we explore multimodal approaches with a feature-mixing perspective in this study. To this end, we introduce CubeMLP, a multimodal feature processing framework based entirely on MLP. CubeMLP consists of three independent MLP units, each of which has two affine transformations. CubeMLP accepts all relevant modality features as input and mixes them across three axes. After extracting the characteristics using CubeMLP, the mixed multimodal features are flattened for task predictions. Our experiments are conducted on sentiment analysis datasets: CMU-MOSI and CMU-MOSEI, and depression estimation dataset: AVEC2019. The results show that CubeMLP can achieve state-of-the-art performance with a much lower computing cost.
引用
收藏
页码:3722 / 3729
页数:8
相关论文
共 50 条
  • [31] Multimodal Sentiment Analysis Based on Composite Hierarchical Fusion
    Lei, Yu
    Qu, Keshuai
    Zhao, Yifan
    Han, Qing
    Wang, Xuguang
    COMPUTER JOURNAL, 2024, 67 (06): : 2230 - 2245
  • [32] A Survey on Multimodal Aspect-Based Sentiment Analysis
    Zhao, Hua
    Yang, Manyu
    Bai, Xueyang
    Liu, Han
    IEEE ACCESS, 2024, 12 : 12039 - 12052
  • [33] Survey of Sentiment Analysis Algorithms Based on Multimodal Fusion
    Guo, Xu
    Mairidan, Wushouer
    Gulanbaier, Tuerhong
    Computer Engineering and Applications, 2024, 60 (02) : 1 - 18
  • [34] Joint multimodal sentiment analysis based on information relevance
    Chen, Danlei
    Su, Wang
    Wu, Peng
    Hua, Bolin
    INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (02)
  • [35] Multimodal social sentiment analysis based on semantic correlation
    Hu H.
    Feng M.
    Cao M.
    Liu M.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2021, 47 (03): : 469 - 477
  • [36] Design and Implementation of Attention Depression Detection Model Based on Multimodal Analysis
    Park, Junhee
    Moon, Nammee
    SUSTAINABILITY, 2022, 14 (06)
  • [37] Hierarchical Interactive Multimodal Transformer for Aspect-Based Multimodal Sentiment Analysis
    Yu, Jianfei
    Chen, Kai
    Xia, Rui
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 1966 - 1978
  • [38] Joint Sentiment Part Topic Regression Model for Multimodal Analysis
    Li, Mengyao
    Zhu, Yonghua
    Gao, Wenjing
    Cao, Meng
    Wang, Shaoxiu
    INFORMATION, 2020, 11 (10) : 1 - 16
  • [39] ANALYSIS OF PHONE POSTERIOR FEATURE SPACE EXPLOITING CLASS-SPECIFIC SPARSITY AND MLP-BASED SIMILARITY MEASURE
    Asaei, Afsaneh
    Picart, Benjamin
    Bourlard, Herve
    2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, : 4886 - 4889
  • [40] Novel OGBEE-based feature selection and feature-level fusion with MLP neural network for social media multimodal sentiment analysis
    Bairavel, S.
    Krishnamurthy, M.
    SOFT COMPUTING, 2020, 24 (24) : 18431 - 18445