Toward Multi-modal Music Emotion Classification

被引:0
|
作者
Yang, Yi-Hsuan [1 ]
Lin, Yu-Ching [1 ]
Cheng, Heng-Tze [1 ]
Liao, I-Bin [2 ]
Ho, Yeh-Chin [2 ]
Chen, Homer H. [1 ]
机构
[1] Natl Taiwan Univ, Taipei, Taiwan
[2] Chunghwa Telecom, Telecommun Labs, Taipei, Taiwan
关键词
Music emotion recognition; multi-modal fusion; lyrics; natural language processing; probabilistic latent semantic analysis;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The performance of categorical music emotion classification that divides emotion into classes and uses audio features alone for emotion classification has reached a limit due to the presence of a semantic gap between the object feature level and the human cognitive level of emotion perception. Motivated by the fact that lyrics carry rich semantic information of a song, we propose a multi-modal approach to help improve categorical music emotion classification. By exploiting both the audio features and the lyrics of a song, the proposed approach improves the 4-class emotion classification accuracy from 46.6% to 57.1%. The results also show that the incorporation of lyrics significantly enhances the classification accuracy of valence.
引用
收藏
页码:70 / +
页数:3
相关论文
共 50 条
  • [1] Research on Multi-modal Music Emotion Classification Based on Audio and Lyirc
    Liu, Gaojun
    Tan, Zhiyuan
    PROCEEDINGS OF 2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2020), 2020, : 2331 - 2335
  • [2] Multi-modal, Multi-task and Multi-label for Music Genre Classification and Emotion Regression
    Pandeya, Yagya Raj
    You, Jie
    Bhattarai, Bhuwan
    Lee, Joonwhoan
    12TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE (ICTC 2021): BEYOND THE PANDEMIC ERA WITH ICT CONVERGENCE INNOVATION, 2021, : 1042 - 1045
  • [3] Multi-modal Music Genre Classification Approach
    Zhen, Chao
    Xu, Jieping
    PROCEEDINGS OF 2010 3RD IEEE INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND INFORMATION TECHNOLOGY (ICCSIT 2010), VOL 8, 2010, : 398 - 402
  • [4] Research on Emotion Classification Based on Multi-modal Fusion
    Xiang, Zhihua
    Radzi, Nor Haizan Mohamed
    Hashim, Haslina
    BAGHDAD SCIENCE JOURNAL, 2024, 21 (02) : 548 - 560
  • [5] Utilizing Multi-modal Emotion Information in Dialogue Strategy Classification
    Jang, Jin Yea
    Kim, Jieun
    Jung, Minyoung
    Jung, Hyedong
    Shin, Saim
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 943 - 945
  • [6] EXTRACTING AND RECOGNISING MUSIC FEATURES THROUGH MULTI-MODAL EMOTION RECOGNITION
    Xu, Chi
    MECHATRONIC SYSTEMS AND CONTROL, 2024, 52 (03): : 140 - 146
  • [7] Towards Emotion-aided Multi-modal Dialogue Act Classification
    Saha, Tulika
    Patra, Aditya Prakash
    Saha, Sriparna
    Bhattacharyya, Pushpak
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 4361 - 4372
  • [8] Automatic music mood classification using multi-modal attention framework
    Sujeesha, A. S.
    Mala, J. B.
    Rajan, Rajeev
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 128
  • [9] Multi-modal Extreme Classification
    Mittal, Anshul
    Dahiya, Kunal
    Malani, Shreya
    Ramaswamy, Janani
    Kuruvilla, Seba
    Ajmera, Jitendra
    Chang, Keng-Hao
    Agarwal, Sumeet
    Kar, Purushottam
    Varma, Manik
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12383 - 12392
  • [10] Towards Sentiment and Emotion aided Multi-modal Speech Act Classification in Twitter
    Saha, Tulika
    Upadhyaya, Apoorva
    Saha, Sriparna
    Bhattacharyya, Pushpak
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 5727 - 5737