Utterance and Syllable Level Prosodic Features for Automatic Emotion Recognition

被引:0
|
作者
Ben Alex, Starlet [1 ]
Babu, Ben P. [2 ]
Mary, Leena [2 ]
机构
[1] Rajiv Gandhi Inst Technol, Dept Elect & Commun, Kottayam, Kerala, India
[2] Govt Coll Engn, Dept Elect & Commun, Idukki, Kerala, India
关键词
automatic emotion recognition; prosodic features; syllable level segmentation;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes an automatic emotion recognition (AER) system that combines prosodic features extracted at utterance level and syllable level to recognize its emotional content. The prosodic features are extracted after identifying speech/non-speech intervals, followed by syllable level segmentation. Prosodic features chosen include parameters for representing dynamics of pitch and energy, along with duration information. Two separate classifiers are built using Deep Neural Networks (DNN). The decision scores based on both levels are fused to identify the emotion of a test utterance from the German Emotion Database (Emo-DB) which contains seven emotions, namely anger, boredom, disgust, fear, happiness, sadness and neutral. The proposed system gives a Weighted Average Recall (WAR) of 58.88% for both utterance level and syllable level prosodic features. Fusion of scores by merely adding the scores gives an overall WAR of 61.68%.
引用
收藏
页码:31 / 35
页数:5
相关论文
共 50 条
  • [1] Attention and Feature Selection for Automatic Speech Emotion Recognition Using Utterance and Syllable-Level Prosodic Features
    Starlet Ben Alex
    Leena Mary
    Ben P. Babu
    Circuits, Systems, and Signal Processing, 2020, 39 : 5681 - 5709
  • [2] Attention and Feature Selection for Automatic Speech Emotion Recognition Using Utterance and Syllable-Level Prosodic Features
    Ben Alex, Starlet
    Mary, Leena
    Babu, Ben P.
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2020, 39 (11) : 5681 - 5709
  • [3] Automatic Emotion Recognition using Auditory and Prosodic Indicative Features
    Gharsellaoui, Soumaya
    Selouani, Sid-Ahmed
    Dahmane, Adel Omar
    2015 IEEE 28TH CANADIAN CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING (CCECE), 2015, : 1265 - 1270
  • [4] Speech emotion recognition based on prosodic segment level features
    Han, Wenjing
    Li, Haifeng
    Qinghua Daxue Xuebao/Journal of Tsinghua University, 2009, 49 (SUPPL. 1): : 1363 - 1368
  • [5] Improved Emotion Recognition with Novel Global Utterance-level Features
    Huang, Yongming
    Zhang, Guobao
    Li, Xiong
    Da, Feipeng
    APPLIED MATHEMATICS & INFORMATION SCIENCES, 2011, 5 (02): : 147 - 153
  • [6] INVESTIGATIONS INTO PROSODIC SYLLABLE CONTOUR FEATURES FOR SPEAKER RECOGNITION
    Kockmann, Marcel
    Burget, Lukas
    Cernocky, Jan Honza
    2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, : 4418 - 4421
  • [7] Automatic Utterance Selection Based on Prosodic Features of Children's Vocalizations
    Kubo, Yuki
    Oka, Natsuki
    Hanada, Subaru
    Tanaka, Kazuaki
    Takahashi, Tomomi
    ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 1423 : 45 - 56
  • [8] Experimental Study in Emotion Recognition using Prosodic Features
    Pavaloi, Ioan
    Musca, Elena
    2015 E-HEALTH AND BIOENGINEERING CONFERENCE (EHB), 2015,
  • [9] Emotion recognition method based on normalization of prosodic features
    Suzuki, Motoyuki
    Nakagawa, Shohei
    Kita, Kenji
    2013 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2013,
  • [10] Prosodic Features Based Text-dependent Speaker Recognition with Short Utterance
    Zhang, Jianwu
    He, Jianchao
    Wu, Zhendong
    Li, Ping
    COMPUTATIONAL INTELLIGENCE AND INTELLIGENT SYSTEMS, (ISICA 2015), 2016, 575 : 541 - 552