Children's Emotion Recognition from Spontaneous Speech Using a Reduced Set of Acoustic and Linguistic Features

被引:9
|
作者
Planet, Santiago [1 ]
Iriondo, Ignasi [1 ]
机构
[1] Univ Ramon Llull, Barcelona 08022, Spain
关键词
Emotion recognition; Spontaneous speech; Acoustic and linguistic features; Feature selection; Feature-level fusion; Speaker-independent;
D O I
10.1007/s12559-012-9174-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The aim of this article is to classify children's affective states in a real-life non-prototypical emotion recognition scenario. The framework is the same as that proposed in the Interspeech 2009 Emotion Challenge. We used a large set of acoustic features and five linguistic parameters based on the concept of emotional salience. Features were extracted from the spontaneous speech recordings of the FAU Aibo Corpus and their transcriptions. We used a wrapper method to reduce the acoustic set of features from 384 to 28 elements and feature-level fusion to merge them with the set of linguistic parameters. We studied three classification approaches: a Naive-Bayes classifier, a support vector machine and a logistic model tree. Results show that the linguistic features improve the performances of the classifiers that use only acoustic datasets. Additionally, merging the linguistic features with the reduced acoustic set is more effective than working with the full dataset. The best classifier performance is achieved with the logistic model tree and the reduced set of acoustic and linguistic features, which improves the performance obtained with the full dataset by 4.15 % absolute (10.14 % relative) and improves the performance of the Naive-Bayes classifier by 9.91 % absolute (28.18 % relative). For the same conditions proposed in the Emotion Challenge, this simple scheme slightly improves a much more complex structure involving seven classifiers and a larger number of features.
引用
收藏
页码:526 / 532
页数:7
相关论文
共 50 条
  • [41] A Novel Emotion Recognizer from Speech Using Both Prosodic and Linguistic Features
    Suzuki, Motoyuki
    Tsuchiya, Seiji
    Ren, Fuji
    KNOWLEDGE-BASED AND INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT I: 15TH INTERNATIONAL CONFERENCE, KES 2011, 2011, 6881 : 456 - 465
  • [42] Anger recognition in speech using acoustic and linguistic cues
    Polzehl, Tim
    Schmitt, Alexander
    Metze, Florian
    Wagner, Michael
    SPEECH COMMUNICATION, 2011, 53 (9-10) : 1198 - 1209
  • [43] Alzheimer's Dementia Recognition From Spontaneous Speech Using Disfluency and Interactional Features
    Nasreen, Shamila
    Rohanian, Morteza
    Hough, Julian
    Purver, Matthew
    FRONTIERS IN COMPUTER SCIENCE, 2021, 3
  • [44] Spontaneous Speech Emotion Recognition using Prior Knowledge
    Chakraborty, Rupayan
    Pandharipande, Meghna
    Kopparapu, Sunil Kumar
    2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 2866 - 2871
  • [45] Speech emotion recognition using a novel feature set
    Yang, J. (jsjyj0801@163.com), 1600, Binary Information Press, P.O. Box 162, Bethel, CT 06801-0162, United States (09):
  • [46] Emotion recognition from speech using global and local prosodic features
    Rao K.S.
    Koolagudi S.G.
    Vempada R.R.
    International Journal of Speech Technology, 2013, 16 (2) : 143 - 160
  • [47] Explainable Alzheimer's Disease Detection Using linguistic features From Automatic Speech Recognition
    Tang, Lijuan
    Zhang, Zhenglin
    Feng, Feifan
    Yang, Li-Zhuang
    Li, Hai
    DEMENTIA AND GERIATRIC COGNITIVE DISORDERS, 2023, 52 (04) : 240 - 248
  • [48] Emotion recognition from speech using source, system, and prosodic features
    Koolagudi, Shashidhar G.
    Rao, K. Sreenivasa
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2012, 15 (02) : 265 - 289
  • [49] Emotion recognition from speech signals using new harmony features
    Yang, B.
    Lugger, M.
    SIGNAL PROCESSING, 2010, 90 (05) : 1415 - 1423
  • [50] Learning deep multimodal affective features for spontaneous speech emotion recognition
    Zhang, Shiqing
    Tao, Xin
    Chuang, Yuelong
    Zhao, Xiaoming
    SPEECH COMMUNICATION, 2021, 127 : 73 - 81