Knowledge-based framework for intelligent emotion recognition in spontaneous speech

被引:9
|
作者
Chakraborty, Rupayan [1 ]
Pandharipande, Meghna [1 ]
Kopparapu, Sunil Kumar [1 ]
机构
[1] TCS Innovat Labs Mumbai, Yantra Pk, Thane West 400601, India
关键词
Knowledge-based framework; emotion recognition; intelligent systems; spontaneous speech; non-acted emotion; CLASSIFICATION;
D O I
10.1016/j.procs.2016.08.239
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Automatic speech emotion recognition plays an important role in intelligent human computer interaction. Identifying emotion in natural, day to day, spontaneous conversational speech is difficult because most often the emotion expressed by the speaker are not necessarily as prominent as in acted speech. In this paper, we propose a novel spontaneous speech emotion recognition framework that makes use of the available knowledge. The framework is motivated by the observation that there is significant disagreement amongst human annotators when they annotate spontaneous speech; the disagreement largely reduces when they are provided with additional knowledge related to the conversation. The proposed framework makes use of the contexts (derived from linguistic contents) and the knowledge regarding the time lapse of the spoken utterances in the context of an audio call to reliably recognize the current emotion of the speaker in spontaneous audio conversations. Our experimental results demonstrate that there is a significant improvement in the performance of spontaneous speech emotion recognition using the proposed framework. (C) 2016 The Authors. Published by Elsevier B.V.
引用
收藏
页码:587 / 596
页数:10
相关论文
共 50 条
  • [41] A Multilingual Framework Based on Pre-training Model for Speech Emotion Recognition
    Zhang, Zhaohang
    Zhang, Xiaohui
    Guo, Min
    Zhang, Wei-Qiang
    Li, Ke
    Huang, Yukai
    2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 750 - 755
  • [42] Hierarchical sparse coding framework for speech emotion recognition
    Torres-Boza, Diana
    Oveneke, Meshia Cedric
    Wang, Fengna
    Jiang, Dongmei
    Verhelst, Werner
    Sahli, Hichem
    SPEECH COMMUNICATION, 2018, 99 : 80 - 89
  • [43] Probing Speech Emotion Recognition Transformers for Linguistic Knowledge
    Triantafyllopoulos, Andreas
    Wagner, Johannes
    Wierstorf, Hagen
    Schmitt, Maximilian
    Reichel, Uwe
    Eyben, Florian
    Burkhardt, Felix
    Schuller, Bjoern W.
    INTERSPEECH 2022, 2022, : 146 - 150
  • [44] Speech Emotion Recognition Exploiting ASR-based and Phonological Knowledge Representations
    Liang, Shuang
    Xie, Xiang
    Zhan, Qingran
    Cheng, Hao
    6TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE, ICIAI2022, 2022, : 216 - 220
  • [45] Subjective Evaluation of a Speech Emotion Recognition Interaction Framework
    Vryzas, Nikolaos
    Matsiola, Maria
    Kotsakis, Rigas
    Dimoulas, Charalampos
    Kalliris, George
    2018 CONFERENCE ON INTERACTION WITH SOUND (AUDIO MOSTLY): SOUND IN IMMERSION AND EMOTION (AM'18), 2018,
  • [46] Exploration of an Independent Training Framework for Speech Emotion Recognition
    Zhong, Shunming
    Yu, Baoxian
    Zhang, Han
    IEEE ACCESS, 2020, 8 : 222533 - 222543
  • [47] Lightweight Deep Learning Framework for Speech Emotion Recognition
    Akinpelu, Samson
    Viriri, Serestina
    Adegun, Adekanmi
    IEEE ACCESS, 2023, 11 : 77086 - 77098
  • [48] A Speech Emotion Recognition Framework for Better Discrimination of Confusions
    Liu, Jiawang
    Wang, Haoxiang
    INTERSPEECH 2021, 2021, : 4483 - 4487
  • [49] Emotion recognition of speech based on RNN
    Park, CH
    Lee, DW
    Sim, KB
    2002 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-4, PROCEEDINGS, 2002, : 2210 - 2213
  • [50] Spontaneous speech emotion recognition via multiple kernel learning
    Zha, Cheng
    Yang, Ping
    Zhang, Xinran
    Zhao, Li
    PROCEEDINGS 2016 EIGHTH INTERNATIONAL CONFERENCE ON MEASURING TECHNOLOGY AND MECHATRONICS AUTOMATION ICMTMA 2016, 2016, : 621 - 623