Toward Multi-modal Music Emotion Classification

被引:0
|
作者
Yang, Yi-Hsuan [1 ]
Lin, Yu-Ching [1 ]
Cheng, Heng-Tze [1 ]
Liao, I-Bin [2 ]
Ho, Yeh-Chin [2 ]
Chen, Homer H. [1 ]
机构
[1] Natl Taiwan Univ, Taipei, Taiwan
[2] Chunghwa Telecom, Telecommun Labs, Taipei, Taiwan
关键词
Music emotion recognition; multi-modal fusion; lyrics; natural language processing; probabilistic latent semantic analysis;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The performance of categorical music emotion classification that divides emotion into classes and uses audio features alone for emotion classification has reached a limit due to the presence of a semantic gap between the object feature level and the human cognitive level of emotion perception. Motivated by the fact that lyrics carry rich semantic information of a song, we propose a multi-modal approach to help improve categorical music emotion classification. By exploiting both the audio features and the lyrics of a song, the proposed approach improves the 4-class emotion classification accuracy from 46.6% to 57.1%. The results also show that the incorporation of lyrics significantly enhances the classification accuracy of valence.
引用
收藏
页码:70 / +
页数:3
相关论文
共 50 条
  • [21] Multi-modal long document classification based on Hierarchical Prompt and Multi-modal Transformer
    Liu, Tengfei
    Hu, Yongli
    Gao, Junbin
    Wang, Jiapu
    Sun, Yanfeng
    Yin, Baocai
    NEURAL NETWORKS, 2024, 176
  • [22] MM-Net: A Multi-Modal Approach Toward Automatic Modulation Classification
    Triaridis, Konstantinos
    Doumanidis, Constantine
    Chatzidiamantis, Nestor D.
    Karagiannidis, George K.
    IEEE COMMUNICATIONS LETTERS, 2024, 28 (02) : 328 - 331
  • [23] Multi-Modal Emotion Classification in Virtual Reality Using Reinforced Self-Training
    Liu, Yi
    Li, Jianzhang
    Cui, Dewen
    Sato-Shimokawara, Eri
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2023, 27 (05) : 967 - 975
  • [24] Emotion classification with multi-modal physiological signals using multi-attention-based neural network
    Zou, Chengsheng
    Deng, Zhen
    He, Bingwei
    Yan, Maosong
    Wu, Jie
    Zhu, Zhaoju
    COGNITIVE COMPUTATION AND SYSTEMS, 2024, 6 (1-3) : 1 - 11
  • [25] Lightweight multi-modal emotion recognition model based on modal generation
    Liu, Peisong
    Che, Manqiang
    Luo, Jiangchuan
    2022 9TH INTERNATIONAL FORUM ON ELECTRICAL ENGINEERING AND AUTOMATION, IFEEA, 2022, : 430 - 435
  • [26] Cross-modal dynamic convolution for multi-modal emotion recognition
    Wen, Huanglu
    You, Shaodi
    Fu, Ying
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2021, 78
  • [27] Transformer-based Automatic Music Mood Classification Using Multi-modal Framework
    Kumar, Sujeesha Ajithakumari Suresh
    Rajan, Rajeev
    JOURNAL OF COMPUTER SCIENCE & TECHNOLOGY, 2023, 23 (01): : 18 - 34
  • [28] Multi-modal Emotion Recognition Based on Speech and Image
    Li, Yongqiang
    He, Qi
    Zhao, Yongping
    Yao, Hongxun
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT I, 2018, 10735 : 844 - 853
  • [29] Multi-Modal Emotion Recognition Fusing Video and Audio
    Xu, Chao
    Du, Pufeng
    Feng, Zhiyong
    Meng, Zhaopeng
    Cao, Tianyi
    Dong, Caichao
    APPLIED MATHEMATICS & INFORMATION SCIENCES, 2013, 7 (02): : 455 - 462
  • [30] Hidden Emotion Detection using Multi-modal Signals
    Kim, Dae Ha
    Song, Byung Cheol
    EXTENDED ABSTRACTS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'21), 2021,