Cross-modal dynamic sentiment annotation for speech sentiment analysis

被引:0
|
作者
Chen, Jincai [1 ]
Sun, Chao [1 ]
Zhang, Sheng [1 ]
Zeng, Jiangfeng [2 ]
机构
[1] Huazhong Univ Sci & Technol, Wuhan Natl Lab Optoelect, Wuhan 430074, Peoples R China
[2] Cent China Normal Univ, Sch Informat Management, Wuhan 430079, Peoples R China
基金
中国国家自然科学基金;
关键词
Speech sentiment analysis; Multi-modal video; Sentiment profiles; Cross-modal annotation;
D O I
10.1016/j.compeleceng.2023.108598
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Traditionally, one single hard label determines the sentiment label of an entire utterance for speech sentiment analysis. It obviously ignores the inherent dynamic and ambiguity of speech sentiments. Moreover, there are few segment-level ground truth labels in the most existing sentiment corpora, due to the label ambiguity and annotation cost. In this work, to capture segment-level sentiment fluctuations across one utterance, we propose sentiment profiles (SPs) to express segment-level soft labels. Meanwhile, we introduce massive multi-modal wild video data to solve the data shortage problem, and facial expression knowledge is used to guide audio segments generate soft labels through the Cross-modal Sentiment Annotation Module. Then, we design a Speech Encoder Module to encode audio segments into SPs. We further exploit the sentiment profile purifier (SPP) to iteratively improve the accuracy of SPs. Numerous experiments show that our model achieves state-of-the-art performance on CH-SIMS and IEMOCAP datasets with unlabeled data respectively.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Cross-Modal Modulating for Multimodal Sentiment Analysis
    Cheng, Zichen
    Li, Yan
    Ge, Jiangwei
    Jiu, Mengfei
    Zhang, Jingwei
    Computer Engineering and Applications, 2023, 59 (10) : 171 - 179
  • [2] Cross-Modal Transformer Combination Model for Sentiment Analysis
    Wang, Liang
    Wang, Yi
    Wang, Jun
    Computer Engineering and Applications, 2024, 60 (13) : 124 - 1350
  • [3] Cross-Modal Enhancement Network for Multimodal Sentiment Analysis
    Wang, Di
    Liu, Shuai
    Wang, Quan
    Tian, Yumin
    He, Lihuo
    Gao, Xinbo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4909 - 4921
  • [4] Dynamic Weighted Gating for Enhanced Cross-Modal Interaction in Multimodal Sentiment Analysis
    Wang, Nan
    Wang, Qi
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 21 (01)
  • [5] Fine-grained sentiment Feature Extraction Method for Cross-modal Sentiment Analysis
    Sun, Ye
    Jin, Guozhe
    Zhao, Yahui
    Cui, Rongyi
    2024 16TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND COMPUTING, ICMLC 2024, 2024, : 602 - 608
  • [6] Hybrid cross-modal interaction learning for multimodal sentiment analysis
    Fu, Yanping
    Zhang, Zhiyuan
    Yang, Ruidi
    Yao, Cuiyou
    NEUROCOMPUTING, 2024, 571
  • [7] The Weighted Cross-Modal Attention Mechanism With Sentiment Prediction Auxiliary Task for Multimodal Sentiment Analysis
    Chen, Qiupu
    Huang, Guimin
    Wang, Yabing
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 2689 - 2695
  • [8] CMJRT: Cross-Modal Joint Representation Transformer for Multimodal Sentiment Analysis
    Xu, Meng
    Liang, Feifei
    Su, Xiangyi
    Fang, Cheng
    IEEE ACCESS, 2022, 10 : 131671 - 131679
  • [9] Multimodal Sentiment Analysis Based on a Cross-Modal Multihead Attention Mechanism
    Deng, Lujuan
    Liu, Boyi
    Li, Zuhe
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 78 (01): : 1157 - 1170
  • [10] Text-Dominant Interactive Attention for Cross-Modal Sentiment Analysis
    Zhang, Zebao
    Yang, Shuang
    Pan, Haiwei
    PATTERN RECOGNITION AND COMPUTER VISION, PT V, PRCV 2024, 2025, 15035 : 201 - 215