MF-Net: a multimodal fusion network for emotion recognition based on multiple physiological signals

被引:0
|
作者
Zhu, Lei [1 ]
Ding, Yu [1 ]
Huang, Aiai [1 ]
Tan, Xufei [2 ]
Zhang, Jianhai [3 ,4 ]
机构
[1] Hangzhou Dianzi Univ, Sch Automat, Hangzhou 310000, Peoples R China
[2] Hangzhou City Univ, Sch Med, Hangzhou 310015, Peoples R China
[3] Hangzhou Dianzi Univ, Sch Comp Sci, Hangzhou 310000, Peoples R China
[4] Hangzhou City Univ, Key Lab Brain Machine Collaborat Intelligence Zhej, Hangzhou 310015, Peoples R China
关键词
Deep learning; Physiological signal; Multimodal fusion; Emotion recognition; EEG;
D O I
10.1007/s11760-024-03632-0
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Currently, research on emotion recognition has shown that multi-modal data fusion has advantages in improving the accuracy and robustness of human emotion recognition, outperforming single-modal methods. Despite the promising results of existing methods, significant challenges remain in effectively fusing data from multiple modalities to achieve superior performance. Firstly, existing works tend to focus on generating a joint representation by fusing multi-modal data, with fewer methods considering the specific characteristics of each modality. Secondly, most methods fail to fully capture the intricate correlations among multiple modalities, often resorting to simplistic combinations of latent features. To address these challenges, we propose a novel fusion network for multi-modal emotion recognition. This network enhances the efficacy of multi-modal fusion while preserving the distinct characteristics of each modality. Specifically, a dual-stream multi-scale feature encoding (MFE) is designed to extract emotional information from both electroencephalogram (EEG) and peripheral physiological signals (PPS) temporal slices. Subsequently, a cross-modal global-local feature fusion module (CGFFM) is proposed to integrate global and local information from multi-modal data and then assign different importance to each modality, which makes the fusion data tend to the more important modalities. Meanwhile, the transformer module is employed to further learn the modality-specific information. Moreover, we introduce the adaptive collaboration block (ACB), which optimally leverages both modality-specific and cross-modality relations for enhanced integration and feature representation. Following extensive experiments on the DEAP and DREAMER multimodal datasets, our model achieves state-of-the-art performance.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Multimodal Physiological Signals Fusion for Online Emotion Recognition
    Pan, Tongjie
    Ye, Yalan
    Cai, Hecheng
    Huang, Shudong
    Yang, Yang
    Wang, Guoqing
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5879 - 5888
  • [2] MF-Net: A Multimodal Fusion Model for Fast Multi-Object Tracking
    Tian, Shirui
    Duan, Mingxing
    Deng, Jiayan
    Luo, Huizhang
    Hu, Yikun
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (08) : 10948 - 10962
  • [3] Emotion recognition based on multimodal physiological electrical signals
    Wang, Zhuozheng
    Wang, Yihan
    FRONTIERS IN NEUROSCIENCE, 2025, 19
  • [4] Feature-Level Fusion of Multimodal Physiological Signals for Emotion Recognition
    Chen, Jing
    Ru, Bin
    Xu, Lixin
    Moore, Philip
    Su, Yun
    PROCEEDINGS 2015 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE, 2015, : 395 - 399
  • [5] Emotion recognition based on multiple physiological signals
    Li, Qi
    Liu, Yunqing
    Yan, Fei
    Zhang, Qiong
    Liu, Cong
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 85
  • [6] Emotion Recognition based on the multiple physiological signals
    Gong, Ping
    Ma, Heather T.
    Wang, Yutong
    2016 IEEE INTERNATIONAL CONFERENCE ON REAL-TIME COMPUTING AND ROBOTICS (IEEE RCAR), 2016, : 140 - 143
  • [7] A Convolution Neural Network Based Emotion Recognition System using Multimodal Physiological Signals
    Yang, Cheng-Jie
    Fahier, Nicolas
    Li, Wei-Chih
    Fang, Wai-Chi
    2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [8] MF-Net: Meta Fusion Network for 3D object detection
    Meng, Zhaoxin
    Luo, Guiyang
    Yuan, Quan
    Li, Jinglin
    Yang, Fangchun
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [9] TAGformer: A Multimodal Physiological Signals Fusion Network for Pilot Stress Recognition
    Wang, Shaofan
    Li, Yuangan
    Zhang, Tao
    Li, Ke
    IEEE SENSORS JOURNAL, 2024, 24 (13) : 20842 - 20854
  • [10] Emotion recognition based on a limited number of multimodal physiological signals channels
    Wan, Chunting
    Xu, Chuanpei
    Chen, Dongyi
    Wei, Daohong
    Li, Xiang
    MEASUREMENT, 2025, 242