Toward Multi-modal Music Emotion Classification

被引:0
|
作者
Yang, Yi-Hsuan [1 ]
Lin, Yu-Ching [1 ]
Cheng, Heng-Tze [1 ]
Liao, I-Bin [2 ]
Ho, Yeh-Chin [2 ]
Chen, Homer H. [1 ]
机构
[1] Natl Taiwan Univ, Taipei, Taiwan
[2] Chunghwa Telecom, Telecommun Labs, Taipei, Taiwan
关键词
Music emotion recognition; multi-modal fusion; lyrics; natural language processing; probabilistic latent semantic analysis;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The performance of categorical music emotion classification that divides emotion into classes and uses audio features alone for emotion classification has reached a limit due to the presence of a semantic gap between the object feature level and the human cognitive level of emotion perception. Motivated by the fact that lyrics carry rich semantic information of a song, we propose a multi-modal approach to help improve categorical music emotion classification. By exploiting both the audio features and the lyrics of a song, the proposed approach improves the 4-class emotion classification accuracy from 46.6% to 57.1%. The results also show that the incorporation of lyrics significantly enhances the classification accuracy of valence.
引用
收藏
页码:70 / +
页数:3
相关论文
共 50 条
  • [31] ATTENTION DRIVEN FUSION FOR MULTI-MODAL EMOTION RECOGNITION
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3227 - 3231
  • [32] Multi-modal Correlated Network for emotion recognition in speech
    Ren, Minjie
    Nie, Weizhi
    Liu, Anan
    Su, Yuting
    VISUAL INFORMATICS, 2019, 3 (03) : 150 - 155
  • [33] A Multi-Modal Deep Learning Approach for Emotion Recognition
    Shahzad, H. M.
    Bhatti, Sohail Masood
    Jaffar, Arfan
    Rashid, Muhammad
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 36 (02): : 1561 - 1570
  • [34] Multi-modal Emotion Recognition for Determining Employee Satisfaction
    Zaman, Farhan Uz
    Zaman, Maisha Tasnia
    Alam, Md Ashraful
    Alam, Md Golam Rabiul
    2021 IEEE ASIA-PACIFIC CONFERENCE ON COMPUTER SCIENCE AND DATA ENGINEERING (CSDE), 2021,
  • [35] Emotion recognition with multi-modal peripheral physiological signals
    Gohumpu, Jennifer
    Xue, Mengru
    Bao, Yanchi
    FRONTIERS IN COMPUTER SCIENCE, 2023, 5
  • [36] Semantic Alignment Network for Multi-Modal Emotion Recognition
    Hou, Mixiao
    Zhang, Zheng
    Liu, Chang
    Lu, Guangming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (09) : 5318 - 5329
  • [37] Facial emotion recognition using multi-modal information
    De Silva, LC
    Miyasato, T
    Nakatsu, R
    ICICS - PROCEEDINGS OF 1997 INTERNATIONAL CONFERENCE ON INFORMATION, COMMUNICATIONS AND SIGNAL PROCESSING, VOLS 1-3: THEME: TRENDS IN INFORMATION SYSTEMS ENGINEERING AND WIRELESS MULTIMEDIA COMMUNICATIONS, 1997, : 397 - 401
  • [38] A multi-modal emotion fusion classification method combined expression and speech based on attention mechanism
    Liu, Dong
    Chen, Longxi
    Wang, Lifeng
    Wang, Zhiyong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (29) : 41677 - 41695
  • [39] Emotion Aided Dialogue Act Classification for Task-Independent Conversations in a Multi-modal Framework
    Saha, Tulika
    Gupta, Dhawal
    Saha, Sriparna
    Bhattacharyya, Pushpak
    COGNITIVE COMPUTATION, 2021, 13 (02) : 277 - 289
  • [40] Emotion Aided Dialogue Act Classification for Task-Independent Conversations in a Multi-modal Framework
    Tulika Saha
    Dhawal Gupta
    Sriparna Saha
    Pushpak Bhattacharyya
    Cognitive Computation, 2021, 13 : 277 - 289