Context-Independent Multilingual Emotion Recognition from Speech Signals

被引:42
|
作者
Vladimir Hozjan
Zdravko Kačič
机构
[1] University of Maribor,
[2] Faculty of Electrical Engineering and Computer Science,undefined
关键词
emotions; speech; emotion recognition; cross language emotion recognition;
D O I
10.1023/A:1023426522496
中图分类号
学科分类号
摘要
This paper presents and discusses an analysis of multilingual emotion recognition from speech with database-specific emotional features. Recognition was performed on English, Slovenian, Spanish, and French InterFace emotional speech databases. The InterFace databases included several neutral speaking styles and six emotions: disgust, surprise, joy, fear, anger and sadness. Speech features for emotion recognition were determined in two steps. In the first step, low-level features were defined and in the second high-level features were calculated from low-level features. Low-level features are composed from pitch, derivative of pitch, energy, derivative of energy, and duration of speech segments. High-level features are statistical presentations of low-level features. Database-specific emotional features were selected from high-level features that contain the most information about emotions in speech. Speaker-dependent and monolingual emotion recognisers were defined, as well as multilingual recognisers. Emotion recognition was performed using artificial neural networks. The achieved recognition accuracy was highest for speaker-dependent emotion recognition, smaller for monolingual emotion recognition and smallest for multilingual recognition. The database-specific emotional features are most convenient for use in multilingual emotion recognition. Among speaker-dependent, monolingual, and multilingual emotion recognition, the difference between emotion recognition with all high-level features and emotion recognition with database-specific emotional features is smallest for multilingual emotion recognition—3.84%.
引用
收藏
页码:311 / 320
页数:9
相关论文
共 50 条
  • [41] Concept Narrowing: The Role of Context-independent Information
    Rubio-Fernandez, Paula
    JOURNAL OF SEMANTICS, 2008, 25 (04) : 381 - 409
  • [42] Context-Independent Claim Detection for Argument Mining
    Lippi, Marco
    Torroni, Paolo
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 185 - 191
  • [43] Multilingual phone models for vocabulary-independent speech recognition tasks
    Köhler, J
    SPEECH COMMUNICATION, 2001, 35 (1-2) : 21 - 30
  • [44] Multilingual Speech Emotion Recognition System based on a Three-layer Model
    Li, Xingfeng
    Akagi, Masato
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 3608 - 3612
  • [45] Context-Independent Facial Action Unit Recognition Using Shape and Gabor Phase Information
    Gonzalez, Isabel
    Sahli, Hichem
    Enescu, Valentin
    Verhelst, Werner
    AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION, PT I, 2011, 6974 : 548 - 557
  • [46] A Multilingual Framework Based on Pre-training Model for Speech Emotion Recognition
    Zhang, Zhaohang
    Zhang, Xiaohui
    Guo, Min
    Zhang, Wei-Qiang
    Li, Ke
    Huang, Yukai
    2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 750 - 755
  • [47] Multilingual, Cross-lingual, and Monolingual Speech Emotion Recognition on EmoFilm Dataset
    Atmaja, Bagus Tris
    Sasou, Akira
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 1019 - 1025
  • [48] Context-independent expression of spatial code in hippocampus
    Kapl, S.
    Tichanek, F.
    Zitricky, F.
    Jezek, K.
    SCIENTIFIC REPORTS, 2022, 12 (01)
  • [49] Acoustic Phonetic Decoding Oriented to Multilingual Speech Recognition in the Basque Context
    Barroso, N.
    Lopez de Ipina, K.
    Ezeiza, A.
    TRENDS IN PRACTICAL APPLICATIONS OF AGENTS AND MULTIAGENT SYSTEMS, 2010, 71 : 697 - +
  • [50] Bimodal Emotion Recognition Based on Speech Signals and Facial Expression
    Tu, Binbin
    Yu, Fengqin
    FOUNDATIONS OF INTELLIGENT SYSTEMS (ISKE 2011), 2011, 122 : 691 - 696