Utterance and Syllable Level Prosodic Features for Automatic Emotion Recognition

被引:0
|
作者
Ben Alex, Starlet [1 ]
Babu, Ben P. [2 ]
Mary, Leena [2 ]
机构
[1] Rajiv Gandhi Inst Technol, Dept Elect & Commun, Kottayam, Kerala, India
[2] Govt Coll Engn, Dept Elect & Commun, Idukki, Kerala, India
关键词
automatic emotion recognition; prosodic features; syllable level segmentation;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes an automatic emotion recognition (AER) system that combines prosodic features extracted at utterance level and syllable level to recognize its emotional content. The prosodic features are extracted after identifying speech/non-speech intervals, followed by syllable level segmentation. Prosodic features chosen include parameters for representing dynamics of pitch and energy, along with duration information. Two separate classifiers are built using Deep Neural Networks (DNN). The decision scores based on both levels are fused to identify the emotion of a test utterance from the German Emotion Database (Emo-DB) which contains seven emotions, namely anger, boredom, disgust, fear, happiness, sadness and neutral. The proposed system gives a Weighted Average Recall (WAR) of 58.88% for both utterance level and syllable level prosodic features. Fusion of scores by merely adding the scores gives an overall WAR of 61.68%.
引用
收藏
页码:31 / 35
页数:5
相关论文
共 50 条
  • [31] Fusion Techniques for Utterance-Level Emotion Recognition Combining Speech and Transcripts
    Sebastian, Jilt
    Pierucci, Piero
    INTERSPEECH 2019, 2019, : 51 - 55
  • [32] USE OF PROSODIC FEATURES TO DERIVE WORD AND SYLLABLE BOUNDARIES
    WILLEMS, YD
    ALLEN, J
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1973, 53 (01): : 355 - &
  • [33] Learning Utterance-level Representations with Label Smoothing for Speech Emotion Recognition
    Huang, Jian
    Tao, Jianhua
    Liu, Bin
    Lian, Zheng
    INTERSPEECH 2020, 2020, : 4079 - 4083
  • [34] Newborns discriminate utterance-level prosodic contours
    Martinez-Alvarez, Anna
    Benavides-Varela, Silvia
    Lapillonne, Alexandre
    Gervain, Judit
    DEVELOPMENTAL SCIENCE, 2023, 26 (02)
  • [35] Emotion recognition from speech using wavelet packet transform and prosodic features
    Gupta, Manish
    Bharti, Shambhu Shankar
    Agarwal, Suneeta
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2018, 35 (02) : 1541 - 1553
  • [36] Study of speech emotion recognition based on prosodic parameters and facial expression features
    Wang, Yu Tai
    Han, Jie
    Jiang, Xiao Qing
    Zou, Jing
    Zhao, Hui
    INDUSTRIAL INSTRUMENTATION AND CONTROL SYSTEMS, PTS 1-4, 2013, 241-244 : 1677 - 1681
  • [37] Performance Analysis of Emotion Recognition from Speech Using Combined Prosodic Features
    Palo, Hemanta K.
    Mohanty, Mihir N.
    ADVANCED SCIENCE LETTERS, 2016, 22 (02) : 288 - 293
  • [38] Emotion Recognition Using Prosodic and Spectral Features of Speech and Naive Bayes Classifier
    Khan, Atreyee
    Roy, Uttam Kumar
    2017 2ND IEEE INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS, SIGNAL PROCESSING AND NETWORKING (WISPNET), 2017, : 1017 - 1021
  • [40] Identifying Uncertain Words within an Utterance via Prosodic Features
    Pon-Barry, Heather
    Shieber, Stuart
    INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5, 2009, : 1543 - 1546