Cross-Corpus Speech Emotion Recognition Based on Causal Emotion Information Representation

被引:0
|
作者
Fu, Hongliang [1 ]
Li, Qianqian [1 ]
Tao, Huawei [1 ]
Zhu, Chunhua [1 ]
Xie, Yue [2 ]
Guo, Ruxue [3 ]
机构
[1] Henan Univ Technol, Key Lab Grain Informat Proc & Control, Minist Educ, Zhengzhou 450001, Peoples R China
[2] Nanjing Inst Technol, Sch Commun Engn, Nanjing 211167, Peoples R China
[3] IFLYTEK Res, Hefei 230088, Peoples R China
基金
中国国家自然科学基金;
关键词
cross-corpus speech emotion recognition; causal representation learning; domain adaptation;
D O I
10.1587/transinf.2023EDL8087
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Speech emotion recognition (SER) is a key research technology to realize the third generation of artificial intelligence, which is widely used in human-computer interaction, emotion diagnosis, interpersonal communication and other fields. However, the aliasing of language and semantic information in speech tends to distort the alignment of emotion features, which affects the performance of cross-corpus SER system. This paper proposes a cross-corpus SER model based on causal emotion information representation (CEIR). The model uses the reconstruction loss of the deep autoencoder network and the source domain label information to realize the preliminary separation of causal features. Then, the causal correlation matrix is constructed, and the local maximum mean difference (LMMD) feature alignment technology is combined to make the causal features of different dimensions jointly distributed independent. Finally, the supervised fine-tuning of labeled data is used to achieve effective extraction of causal emotion information. The experimental results show that the average unweighted average recall (UAR) of the proposed algorithm is increased by 3.4% to 7.01% compared with the latest partial algorithms in the field.
引用
收藏
页码:1097 / 1100
页数:4
相关论文
共 50 条
  • [11] Transfer Linear Subspace Learning for Cross-Corpus Speech Emotion Recognition
    Song, Peng
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2019, 10 (02) : 265 - 275
  • [12] Progressively Discriminative Transfer Network for Cross-Corpus Speech Emotion Recognition
    Lu, Cheng
    Tang, Chuangao
    Zhang, Jiacheng
    Zong, Yuan
    ENTROPY, 2022, 24 (08)
  • [13] Deep Cross-Corpus Speech Emotion Recognition: Recent Advances and Perspectives
    Zhang, Shiqing
    Liu, Ruixin
    Tao, Xin
    Zhao, Xiaoming
    FRONTIERS IN NEUROROBOTICS, 2021, 15
  • [14] Synthesized speech for model training in cross-corpus recognition of human emotion
    Björn Schuller
    Zixing Zhang
    Felix Weninger
    Felix Burkhardt
    International Journal of Speech Technology, 2012, 15 (3) : 313 - 323
  • [15] Adversarial Domain Generalized Transformer for Cross-Corpus Speech Emotion Recognition
    Gao, Yuan
    Wang, Longbiao
    Liu, Jiaxing
    Dang, Jianwu
    Okada, Shogo
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (02) : 697 - 708
  • [16] Emo-DNA: Emotion Decoupling and Alignment Learning for Cross-Corpus Speech Emotion Recognition
    Ye, Jiaxin
    Wei, Yujie
    Wen, Xin-Cheng
    Ma, Chenglong
    Huang, Zhizhong
    Liu, Kunhong
    Shan, Hongming
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5956 - 5965
  • [17] Transferable discriminant linear regression for cross-corpus speech emotion recognition
    Li, Shaokai
    Song, Peng
    Zhang, Wenjing
    APPLIED ACOUSTICS, 2022, 197
  • [18] Analysis of Deep Learning Architectures for Cross-corpus Speech Emotion Recognition
    Parry, Jack
    Palaz, Dimitri
    Clarke, Georgia
    Lecomte, Pauline
    Mead, Rebecca
    Berger, Michael
    Hofer, Gregor
    INTERSPEECH 2019, 2019, : 1656 - 1660
  • [19] Domain Generalization with Triplet Network for Cross-Corpus Speech Emotion Recognition
    Lee, Shi-Wook
    2021 IEEE Spoken Language Technology Workshop, SLT 2021 - Proceedings, 2021, : 389 - 396
  • [20] Transfer Subspace Learning for Unsupervised Cross-Corpus Speech Emotion Recognition
    Liu, Na
    Zhang, Baofeng
    Liu, Bin
    Shi, Jingang
    Yang, Lei
    Li, Zhiwei
    Zhu, Junchao
    IEEE ACCESS, 2021, 9 : 95925 - 95937