Visual-to-EEG cross-modal knowledge distillation for continuous emotion recognition

被引:27
|
作者
Zhang, Su [1 ]
Tang, Chuangao [2 ]
Guan, Cuntai [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Southeast Univ, Sch Biol Sci & Med Engn, Key Lab Child Dev & Learning Sci, Minist Educ, Nanjing 210096, Peoples R China
关键词
Continuous emotion recognition; Knowledge distillation; Cross-modality; BRAIN;
D O I
10.1016/j.patcog.2022.108833
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual modality is one of the most dominant modalities for current continuous emotion recognition methods. Compared to which the EEG modality is relatively less sound due to its intrinsic limitation such as subject bias and low spatial resolution. This work attempts to improve the continuous prediction of the EEG modality by using the dark knowledge from the visual modality. The teacher model is built by a cascade convolutional neural network - temporal convolutional network (CNN-TCN) architecture, and the student model is built by TCNs. They are fed by video frames and EEG average band power features, respectively. Two data partitioning schemes are employed, i.e., the trial-level random shuffling (TRS) and the leave-one-subject-out (LOSO). The standalone teacher and student can produce continuous prediction superior to the baseline method, and the employment of the visual-to-EEG cross-modal KD further improves the prediction with statistical significance, i.e., p-value < 0.01 for TRS and p-value < 0.05 for LOSO partitioning. The saliency maps of the trained student model show that the brain areas associated with the active valence state are not located in precise brain areas. Instead, it results from synchronized activity among various brain areas. And the fast beta and gamma waves, with the frequency of 18 - 30Hz and 30 - 45Hz, contribute the most to the human emotion process compared to other bands. The code is available at https://github.com/sucv/Visual_to_EEG_Cross_Modal_KD_for_CER. (C) 2022 The Authors. Published by Elsevier Ltd.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] A Cross-Modal Correlation Fusion Network for Emotion Recognition in Conversations
    Tang, Xiaolyu
    Cai, Guoyong
    Chen, Ming
    Yuan, Peicong
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT V, NLPCC 2024, 2025, 15363 : 55 - 68
  • [32] Cross-Modal Guiding Neural Network for Multimodal Emotion Recognition From EEG and Eye Movement Signals
    Fu, Baole
    Chu, Wenhao
    Gu, Chunrui
    Liu, Yinhua
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (10) : 5865 - 5876
  • [33] Combining cross-modal knowledge transfer and semi-supervised learning for speech emotion recognition
    Zhang, Sheng
    Chen, Min
    Chen, Jincai
    Li, Yuan-Fang
    Wu, Yiling
    Li, Minglei
    Zhu, Chuanbo
    KNOWLEDGE-BASED SYSTEMS, 2021, 229
  • [34] Emotion Recognition in Speech using Cross-Modal Transfer in the Wild
    Albanie, Samuel
    Nagrani, Arsha
    Vedaldi, Andrea
    Zisserman, Andrew
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 292 - 301
  • [35] Visual, haptic and cross-modal recognition of objects and scenes
    Woods, AT
    Newell, FN
    JOURNAL OF PHYSIOLOGY-PARIS, 2004, 98 (1-3) : 147 - 159
  • [36] Contextual and Cross-Modal Interaction for Multi-Modal Speech Emotion Recognition
    Yang, Dingkang
    Huang, Shuai
    Liu, Yang
    Zhang, Lihua
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 2093 - 2097
  • [37] Cross-modal learning with prior visual relation knowledge
    Yu, Jing
    Zhang, Weifeng
    Yang, Zhuoqian
    Qin, Zengchang
    Hu, Yue
    KNOWLEDGE-BASED SYSTEMS, 2020, 203 (203)
  • [38] Bridging Modality Gap for Visual Grounding with Effecitve Cross-Modal Distillation
    Wang, Jiaxi
    Hu, Wenhui
    Liu, Xueyang
    Wu, Beihu
    Qiu, Yuting
    Cai, YingYing
    PATTERN RECOGNITION AND COMPUTER VISION, PT V, PRCV 2024, 2025, 15035 : 347 - 363
  • [39] Learnable Cross-modal Knowledge Distillation for Multi-modal Learning with Missing Modality
    Wang, Hu
    Ma, Congbo
    Zhang, Jianpeng
    Zhang, Yuan
    Avery, Jodie
    Hull, Louise
    Carneiro, Gustavo
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT IV, 2023, 14223 : 216 - 226
  • [40] Cross-Modal Knowledge Distillation in Deep Networks for SAR Image Classification
    Jahan, Chowdhury Sadman
    Savakis, Andreas
    Blasch, Erik
    GEOSPATIAL INFORMATICS XII, 2022, 12099