Analysis of Speaker Recognition in Blended Emotional Environment Using Deep Learning Approaches

被引:1
|
作者
Tomar, Shalini [1 ]
Koolagudi, Shashidhar G. [1 ]
机构
[1] Natl Inst Technol, Dept Comp Sci & Engn, Mangalore, Karnataka, India
关键词
Blended emotion; Mel Frequency Cepstral Coefficients; Convolutional Neural Network; Speaker Recognition; Speaker Recognition in Blended Emotion Environment; Valence;
D O I
10.1007/978-3-031-45170-6_72
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generally, human conversation has some emotion, and natural emotions are often blended. Today's Speaker Recognition systems lack the component of emotion. This work proposes a Speaker Recognition approaches in Blended Emotion Environment (SRBEE) system to enhance Speaker Recognition (SR) in an emotional context. Speaker Recognition algorithms nearly always achieve perfect performance in the case of neutral speech, but it is not true from an emotional perspective. This work attempts the recognition of speakers in blended emotion with the Mel-Frequency Cepstral Coefficients (MFCC) feature extraction using the Conv2D classifier. In the blended emotional environment, calculating the accuracy of the Speaker Recognition task is complex. The blend of four basic natural emotions (happy, sad, angry, and fearful) utterances tested in the proposed system to reduce SR's complexity in a blended emotional environment. The proposed system achieves an average accuracy of 99.3% for blended emotion with neutral speech and 92.8% for four basic blended natural emotions (happy, sad, angry, and fearful). The dataset was prepared by blending two emotions in one utterance.
引用
收藏
页码:691 / 698
页数:8
相关论文
共 50 条
  • [21] Experimental Face Recognition System Using Deep Learning Approaches
    Imoh, Nsikak
    Vajjhala, Narasimha Rao
    Rakshit, Sandip
    PROCEEDINGS OF SECOND INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTER ENGINEERING AND COMMUNICATION SYSTEMS, ICACECS 2021, 2022, : 131 - 140
  • [22] DOMAIN ROBUST DEEP EMBEDDING LEARNING FOR SPEAKER RECOGNITION
    Hu, Hang-Rui
    Song, Yan
    Liu, Ying
    Dai, Li-Rong
    McLoughlin, Ian
    Liu, Lin
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7182 - 7186
  • [23] A Comparison of MFCC and LPCC with Deep Learning for Speaker Recognition
    Yang, Haiyan
    Deng, Yanrong
    Zhao, Hua-An
    ICBDC 2019: PROCEEDINGS OF 2019 4TH INTERNATIONAL CONFERENCE ON BIG DATA AND COMPUTING, 2019, : 160 - 164
  • [24] An Analysis of Emotional Speech Recognition for Tamil Language Using Deep Learning Gate Recurrent Unit
    Fernandes, Bennilo
    Mannepalli, Kasiprasad
    PERTANIKA JOURNAL OF SCIENCE AND TECHNOLOGY, 2021, 29 (03): : 1937 - 1961
  • [25] Improved Deep Speaker Feature Learning for Text-Dependent Speaker Recognition
    Li, Lantian
    Lin, Yiye
    Zhang, Zhiyong
    Wang, Dong
    2015 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2015, : 426 - 429
  • [26] Emotional Climate Recognition in Interactive Conversational Speech Using Deep Learning
    Alhussein, Ghada
    Alkhodari, Mohanad
    Khandokher, Ahsan
    Hadjileontiadis, Leontios J.
    2022 IEEE INTERNATIONAL CONFERENCE ON DIGITAL HEALTH (IEEE ICDH 2022), 2022, : 96 - 103
  • [27] Facial Emotional Expression Recognition Using Hybrid Deep Learning Algorithm
    Phattarasooksirot, Phasook
    Sento, Adna
    2022 7TH INTERNATIONAL CONFERENCE ON BUSINESS AND INDUSTRIAL RESEARCH (ICBIR2022), 2022, : 323 - 329
  • [28] Curriculum Learning Based Approaches for Noise Robust Speaker Recognition
    Ranjan, Shivesh
    Hansen, John H. L.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2018, 26 (01) : 197 - 210
  • [29] Applying Emotional Factor Analysis and I-Vector to Emotional Speaker Recognition
    Chen, Li
    Yang, Yingchun
    BIOMETRIC RECOGNITION: CCBR 2011, 2011, 7098 : 174 - 179
  • [30] Disentangled Representation Learning for Environment-agnostic Speaker Recognition
    Nam, Kihyun
    Heo, Hee-Soo
    June, Jee-weon
    Chung, Joon Son
    INTERSPEECH 2024, 2024, : 2130 - 2134