Decoding imagined speech from EEG signals using hybrid-scale spatial-temporal dilated convolution network

被引:16
|
作者
Li, Fu [1 ]
Chao, Weibing [1 ]
Li, Yang [1 ]
Fu, Boxun [1 ]
Ji, Youshuo [1 ]
Wu, Hao [1 ]
Shi, Guangming [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Key Lab Intelligent Percept & Image Understanding, Minist Educ, Xian, Peoples R China
基金
中国博士后科学基金;
关键词
brain-computer interface (BCI); EEG-based imagined speech recognition; hybrid-scale; spatial-temporal network; BRAIN-COMPUTER INTERFACES; NEURAL-NETWORKS; CLASSIFICATION; COMMUNICATION; IMAGERY;
D O I
10.1088/1741-2552/ac13c0
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective. Directly decoding imagined speech from electroencephalogram (EEG) signals has attracted much interest in brain-computer interface applications, because it provides a natural and intuitive communication method for locked-in patients. Several methods have been applied to imagined speech decoding, but how to construct spatial-temporal dependencies and capture long-range contextual cues in EEG signals to better decode imagined speech should be considered. Approach. In this study, we propose a novel model called hybrid-scale spatial-temporal dilated convolution network (HS-STDCN) for EEG-based imagined speech recognition. HS-STDCN integrates feature learning from temporal and spatial information into a unified end-to-end model. To characterize the temporal dependencies of the EEG sequences, we adopted a hybrid-scale temporal convolution layer to capture temporal information at multiple levels. A depthwise spatial convolution layer was then designed to construct intrinsic spatial relationships of EEG electrodes, which can produce a spatial-temporal representation of the input EEG data. Based on the spatial-temporal representation, dilated convolution layers were further employed to learn long-range discriminative features for the final classification. Main results. To evaluate the proposed method, we compared the HS-STDCN with other existing methods on our collected dataset. The HS-STDCN achieved an averaged classification accuracy of 54.31% for decoding eight imagined words, which is significantly better than other methods at a significance level of 0.05. Significance. The proposed HS-STDCN model provided an effective approach to make use of both the temporal and spatial dependencies of the input EEG signals for imagined speech recognition. We also visualized the word semantic differences to analyze the impact of word semantics on imagined speech recognition, investigated the important regions in the decoding process, and explored the use of fewer electrodes to achieve comparable performance.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Decoding of the speech envelope from EEG using the VLAAI deep neural network
    Bernd Accou
    Jonas Vanthornhout
    Hugo Van hamme
    Tom Francart
    Scientific Reports, 13
  • [22] Decoding of the speech envelope from EEG using the VLAAI deep neural network
    Accou, Bernd
    Vanthornhout, Jonas
    Van Hamme, Hugo
    Francart, Tom
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [23] An IoT based smart water quality assessment framework for aqua-ponds management using Dilated Spatial-temporal Convolution Neural Network (DSTCNN)
    Arepalli, Peda Gopi
    Naik, K. Jairam
    AQUACULTURAL ENGINEERING, 2024, 104
  • [24] Emotion recognition using spatial-temporal EEG features through convolutional graph attention network
    Li, Zhongjie
    Zhang, Gaoyan
    Wang, Longbiao
    Wei, Jianguo
    Dang, Jianwu
    JOURNAL OF NEURAL ENGINEERING, 2023, 20 (01)
  • [25] Recognition of grammatical class of imagined words from EEG signals using convolutional neural network
    Datta, Sahil
    Boulgouris, Nikolaos, V
    NEUROCOMPUTING, 2021, 465 : 301 - 309
  • [26] Identification of Students' Confusion in Classes from EEG Signals Using Convolution Neural Network
    Sahu R.
    Dash S.R.
    Baral A.
    Informatica (Slovenia), 2024, 48 (01): : 45 - 56
  • [27] A Convolution Network of Multi-Windows Spatial-Temporal Feature Analysis For Single-trial EEG Classification in RSVP Task
    Tan, Ying
    Zang, Boyu
    Lin, Yanfei
    Gao, Xiaorong
    2021 14TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS (CISP-BMEI 2021), 2021,
  • [28] Estimating residual bait density using hybrid dilated convolution and attention multi-scale network
    Zhang, Lizhen
    Li, Yantian
    Li, Zhijian
    Meng, Xiongdong
    Zhang, Yongqi
    Wu, Di
    Nongye Gongcheng Xuebao/Transactions of the Chinese Society of Agricultural Engineering, 2024, 40 (14): : 137 - 145
  • [29] MDSTGCN : Multi-Scale Dynamic Spatial-Temporal Graph Convolution Network With Edge Feature Embedding for Traffic Forecasting
    Liu, Sijia
    Xu, Hui
    Meng, Fanyu
    Ren, Qianqian
    2024 IEEE 24TH INTERNATIONAL SYMPOSIUM ON CLUSTER, CLOUD AND INTERNET COMPUTING, CCGRID 2024, 2024, : 284 - 290
  • [30] Skeleton Driven Action Recognition Using an Image-Based Spatial-Temporal Representation and Convolution Neural Network
    Silva, Vinicius
    Soares, Filomena
    Leao, Celina P.
    Esteves, Joao Sena
    Vercelli, Gianni
    SENSORS, 2021, 21 (13)