Decoding EEG Brain Activity for Multi-Modal Natural Language Processing

被引:15
|
作者
Hollenstein, Nora [1 ]
Renggli, Cedric [2 ]
Glaus, Benjamin [2 ]
Barrett, Maria [3 ]
Troendle, Marius [4 ]
Langer, Nicolas [4 ]
Zhang, Ce [2 ]
机构
[1] Univ Copenhagen, Dept Nord Studies & Linguist, Copenhagen, Denmark
[2] Swiss Fed Inst Technol, Swiss Fed Inst Technol, Dept Comp Sci, Zurich, Switzerland
[3] IT Univ Copenhagen, Dept Comp Sci, Copenhagen, Denmark
[4] Univ Zurich, Dept Psychol, Zurich, Switzerland
来源
关键词
EEG; natural language processing; frequency bands; brain activity; machine learning; multi-modal learning; physiological data; neural network; REGRESSION-BASED ESTIMATION; COGNITIVE NEUROSCIENCE; EYE-MOVEMENTS; THETA; SPEECH; NEUROBIOLOGY; OSCILLATIONS; RESPONSES; MODELS;
D O I
10.3389/fnhum.2021.659410
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Until recently, human behavioral data from reading has mainly been of interest to researchers to understand human cognition. However, these human language processing signals can also be beneficial in machine learning-based natural language processing tasks. Using EEG brain activity for this purpose is largely unexplored as of yet. In this paper, we present the first large-scale study of systematically analyzing the potential of EEG brain activity data for improving natural language processing tasks, with a special focus on which features of the signal are most beneficial. We present a multi-modal machine learning architecture that learns jointly from textual input as well as from EEG features. We find that filtering the EEG signals into frequency bands is more beneficial than using the broadband signal. Moreover, for a range of word embedding types, EEG data improves binary and ternary sentiment classification and outperforms multiple baselines. For more complex tasks such as relation detection, only the contextualized BERT embeddings outperform the baselines in our experiments, which raises the need for further research. Finally, EEG data shows to be particularly promising when limited training data is available.
引用
收藏
页数:19
相关论文
共 50 条
  • [31] Multi-modal PixelNet for Brain Tumor Segmentation
    Islam, Mobarakol
    Ren, Hongliang
    BRAINLESION: GLIOMA, MULTIPLE SCLEROSIS, STROKE AND TRAUMATIC BRAIN INJURIES, BRAINLES 2017, 2018, 10670 : 298 - 308
  • [32] Multi-lingual and multi-modal speech processing and applications
    Ivanecky, J
    Fischer, J
    Mast, M
    Kunzmann, S
    Ross, T
    Fischer, V
    PATTERN RECOGNITION, PROCEEDINGS, 2005, 3663 : 149 - 159
  • [33] Decoding the ABCs of natural language processing in dermatology
    Nambudiri, Vinod E.
    JOURNAL OF THE EUROPEAN ACADEMY OF DERMATOLOGY AND VENEREOLOGY, 2024, 38 (12) : 2201 - 2202
  • [34] Multi-modal and Multi-spectral Registration for Natural Images
    Shen, Xiaoyong
    Xu, Li
    Zhang, Qi
    Jia, Jiaya
    COMPUTER VISION - ECCV 2014, PT IV, 2014, 8692 : 309 - 324
  • [35] Multi-modal emotion identification fusing facial expression and EEG
    Yongzhen Wu
    Jinhua Li
    Multimedia Tools and Applications, 2023, 82 : 10901 - 10919
  • [36] Multi-modal emotion identification fusing facial expression and EEG
    Wu, Yongzhen
    Li, Jinhua
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (07) : 10901 - 10919
  • [37] Multi-modal EEG Online Visualization and Neuro-Feedback
    Hong, Kan
    Zhang, Liqing
    Li, Jie
    Li, Junhua
    ADVANCES IN NEURAL NETWORKS - ISNN 2010, PT 2, PROCEEDINGS, 2010, 6064 : 360 - 367
  • [38] Multi-modal emotion recognition using EEG and speech signals
    Wang, Qian
    Wang, Mou
    Yang, Yan
    Zhang, Xiaolei
    COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 149
  • [39] Attention-based convolutional neural network with multi-modal temporal information fusion for motor imagery EEG decoding
    Ma X.
    Chen W.
    Pei Z.
    Zhang Y.
    Chen J.
    Computers in Biology and Medicine, 2024, 175
  • [40] Multi-modal Language Models for Lecture Video Retrieval
    Chen, Huizhong
    Cooper, Matthew
    Joshi, Dhiraj
    Girod, Bernd
    PROCEEDINGS OF THE 2014 ACM CONFERENCE ON MULTIMEDIA (MM'14), 2014, : 1081 - 1084