Formant-based Feature Extraction for Emotion Classification from Speech

被引:0
|
作者
Kim, Jonathan C. [1 ]
Clements, Mark A. [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
关键词
affective computing; speech analysis; formant; Gaussian mixture model;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In a previous study, a robust formant-tracking algorithm was introduced to model formant and spectral properties of speech. The algorithm utilizes Gaussian mixtures to estimate spectral parameters, and refines the estimates by using a maximum a posteriori adaptation (MAP) algorithm. In this paper, the formant-tracking algorithm was used to extract the formant-based features for emotion classification. The classification results were compared to a linear predictive coding (LPC) based algorithm for evaluation. On average, the formant features extracted using the algorithm improved the unweighted accuracy by 2.1 percentage points when compared to a LPC-based algorithm. The combination of formant features and other acoustic features statistically significantly improved the unweighted accuracy by 2.7 percentage points, whereas the LPC-based features barely improved it by 1 percentage point. The results clearly indicate that an improved formant-tracking method improved emotion classification accuracy. The effect of formant-based features in emotion classification is also discussed.
引用
收藏
页码:477 / 481
页数:5
相关论文
共 50 条
  • [31] Paralinguistic and spectral feature extraction for speech emotion classification using machine learning techniques
    Tong Liu
    Xiaochen Yuan
    EURASIP Journal on Audio, Speech, and Music Processing, 2023
  • [32] Introducing New Feature Set based on Wavelets for Speech Emotion Classification
    Tanmoy, Roy
    Tshilidzi, Marwala
    Snehashish, Chakraverty
    Paul, Satyakama
    PROCEEDINGS OF 2018 IEEE APPLIED SIGNAL PROCESSING CONFERENCE (ASPCON), 2018, : 124 - 128
  • [33] The formant structure based feature parameter for speech recognition
    Zhao, JH
    Kuang, JM
    Xie, X
    PROCEEDINGS OF THE 2003 IEEE WORKSHOP ON STATISTICAL SIGNAL PROCESSING, 2003, : 605 - 608
  • [34] Classification of Speech Emotion State Based on Feature Map Fusion of TCN and Pretrained CNN Model From Korean Speech Emotion Data
    Jo, A-Hyeon
    Kwak, Keun-Chang
    IEEE ACCESS, 2025, 13 : 19947 - 19963
  • [35] Feature Vector Classification based Speech Emotion Recognition for Service Robots
    Park, Jeong-Sik
    Kim, Ji-Hwan
    Oh, Yung-Hwan
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2009, 55 (03) : 1590 - 1596
  • [36] Text-to-speech conversion system for Brazilian Portuguese using a formant-based synthesis technique
    Gomes, LDT
    Nagle, EJ
    Chiquito, JG
    ITS '98 PROCEEDINGS - SBT/IEEE INTERNATIONAL TELECOMMUNICATIONS SYMPOSIUM, VOLS 1 AND 2, 1998, : 219 - 224
  • [37] A Pattern Mining Approach in Feature Extraction for Emotion Recognition from Speech
    Avci, Umut
    Akkurt, Gamze
    Unay, Devrim
    SPEECH AND COMPUTER, SPECOM 2019, 2019, 11658 : 54 - 63
  • [38] FORMANT-BASED AUDIO SYNTHESIS USING NONLINEAR DISTORTION
    PUCKETTE, M
    JOURNAL OF THE AUDIO ENGINEERING SOCIETY, 1995, 43 (1-2): : 40 - 47
  • [39] Pitch- and Formant-Based Order Adaptation of the Fractional Fourier Transform and Its Application to Speech Recognition
    Yin, Hui
    Nadeu, Climent
    Hohmann, Volker
    EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2009,
  • [40] A Quest for Formant-Based Compact Nonuniform Trapezoidal Filter Banks for Speech Processing with VGG16
    Parlak, Cevahir
    Altun, Yusuf
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2024, 43 (11) : 7309 - 7338