Children's Emotion Recognition from Spontaneous Speech Using a Reduced Set of Acoustic and Linguistic Features

被引:9
|
作者
Planet, Santiago [1 ]
Iriondo, Ignasi [1 ]
机构
[1] Univ Ramon Llull, Barcelona 08022, Spain
关键词
Emotion recognition; Spontaneous speech; Acoustic and linguistic features; Feature selection; Feature-level fusion; Speaker-independent;
D O I
10.1007/s12559-012-9174-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The aim of this article is to classify children's affective states in a real-life non-prototypical emotion recognition scenario. The framework is the same as that proposed in the Interspeech 2009 Emotion Challenge. We used a large set of acoustic features and five linguistic parameters based on the concept of emotional salience. Features were extracted from the spontaneous speech recordings of the FAU Aibo Corpus and their transcriptions. We used a wrapper method to reduce the acoustic set of features from 384 to 28 elements and feature-level fusion to merge them with the set of linguistic parameters. We studied three classification approaches: a Naive-Bayes classifier, a support vector machine and a logistic model tree. Results show that the linguistic features improve the performances of the classifiers that use only acoustic datasets. Additionally, merging the linguistic features with the reduced acoustic set is more effective than working with the full dataset. The best classifier performance is achieved with the logistic model tree and the reduced set of acoustic and linguistic features, which improves the performance obtained with the full dataset by 4.15 % absolute (10.14 % relative) and improves the performance of the Naive-Bayes classifier by 9.91 % absolute (28.18 % relative). For the same conditions proposed in the Emotion Challenge, this simple scheme slightly improves a much more complex structure involving seven classifiers and a larger number of features.
引用
收藏
页码:526 / 532
页数:7
相关论文
共 50 条
  • [21] Automatic Recognition of Spontaneous Emotions in Speech Using Acoustic and Lexical Features
    Truong, Khict P.
    Raaijmakers, Stephan
    MACHINE LEARNING FOR MULTIMODAL INTERACTION, PROCEEDINGS, 2008, 5237 : 161 - +
  • [22] Phoneme Set Design Based on Integrated Acoustic and Linguistic Features for Second Language Speech Recognition
    Wang, Xiaoyun
    Kato, Tsuneo
    Yamamoto, Seiichi
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2017, E100D (04): : 857 - 864
  • [23] FUSION APPROACHES FOR EMOTION RECOGNITION FROM SPEECH USING ACOUSTIC AND TEXT-BASED FEATURES
    Pepino, Leonardo
    Riera, Pablo
    Ferrer, Luciana
    Gravano, Agustin
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6484 - 6488
  • [24] Deep fusion framework for speech command recognition using acoustic and linguistic features
    Sunakshi Mehra
    Seba Susan
    Multimedia Tools and Applications, 2023, 82 : 38667 - 38691
  • [25] Deep fusion framework for speech command recognition using acoustic and linguistic features
    Mehra, Sunakshi
    Susan, Seba
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (25) : 38667 - 38691
  • [26] Recognizing emotion from Turkish speech using acoustic features
    Oflazoglu, Caglar
    Yildirim, Serdar
    EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2013,
  • [27] Recognizing emotion from Turkish speech using acoustic features
    Caglar Oflazoglu
    Serdar Yildirim
    EURASIP Journal on Audio, Speech, and Music Processing, 2013
  • [28] Machine learning techniques for speech emotion recognition using paralinguistic acoustic features
    Jha T.
    Kavya R.
    Christopher J.
    Arunachalam V.
    International Journal of Speech Technology, 2022, 25 (03): : 707 - 725
  • [29] Emotion Recognition in Spontaneous Speech Using GMMs
    Neiberg, Daniel
    Elenius, Kjell
    Laskowski, Kornel
    INTERSPEECH 2006 AND 9TH INTERNATIONAL CONFERENCE ON SPOKEN LANGUAGE PROCESSING, VOLS 1-5, 2006, : 809 - +
  • [30] Speech Emotion Classification using Acoustic Features
    Chen, Shizhe
    Jin, Qin
    Li, Xirong
    Yang, Gang
    Xu, Jieping
    2014 9TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2014, : 579 - 583