Decoding Envelope and Frequency-Following EEG Responses to Continuous Speech Using Deep Neural Networks

被引:2
|
作者
Thornton, Mike D. [1 ]
Mandic, Danilo P. [2 ]
Reichenbach, Tobias J. [3 ]
机构
[1] Imperial Coll London, Dept Comp, London SW7 2RH, England
[2] Imperial Coll London, Dept Elect & Elect Engn, London SW7 2RH, England
[3] Friedrich Alexander Univ Erlangen Nurnberg, Dept Artificial Intelligence Biomed Engn, D-91052 Erlangen, Germany
关键词
Auditory attention decoding; deep learning; EEG signal processing;
D O I
10.1109/OJSP.2024.3378593
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The electroencephalogram (EEG) offers a non-invasive means by which a listener's auditory system may be monitored during continuous speech perception. Reliable auditory-EEG decoders could facilitate the objective diagnosis of hearing disorders, or find applications in cognitively-steered hearing aids. Previously, we developed decoders for the ICASSP Auditory EEG Signal Processing Grand Challenge (SPGC). These decoders placed first in the match-mismatch task: given a short temporal segment of EEG recordings, and two candidate speech segments, the task is to identify which of the two speech segments is temporally aligned, or matched, with the EEG segment. The decoders made use of cortical responses to the speech envelope, as well as speech-related frequency-following responses, to relate the EEG recordings to the speech stimuli. Here we comprehensively document the methods by which the decoders were developed. We extend our previous analysis by exploring the association between speaker characteristics (pitch and sex) and classification accuracy, and provide a full statistical analysis of the final performance of the decoders as evaluated on a heldout portion of the dataset. Finally, the generalisation capabilities of the decoders are characterised, by evaluating them using an entirely different dataset which contains EEG recorded under a variety of speech-listening conditions. The results show that the match-mismatch decoders achieve accurate and robust classification accuracies, and they can even serve as auditory attention decoders without additional training.
引用
收藏
页码:700 / 716
页数:17
相关论文
共 50 条
  • [1] Decoding of the speech envelope from EEG using the VLAAI deep neural network
    Bernd Accou
    Jonas Vanthornhout
    Hugo Van hamme
    Tom Francart
    Scientific Reports, 13
  • [2] Decoding of the speech envelope from EEG using the VLAAI deep neural network
    Accou, Bernd
    Vanthornhout, Jonas
    Van Hamme, Hugo
    Francart, Tom
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [3] Robust decoding of the speech envelope from EEG recordings through deep neural networks
    Thornton, Mike
    Mandic, Danilo
    Reichenbach, Tobias
    JOURNAL OF NEURAL ENGINEERING, 2022, 19 (04)
  • [4] Relating EEG to continuous speech using deep neural networks: a review
    Puffay, Corentin
    Accou, Bernd
    Bollens, Lies
    Monesi, Mohammad Jalilpour
    Vanthornhout, Jonas
    Van Hamme, Hugo
    Francart, Tom
    JOURNAL OF NEURAL ENGINEERING, 2023, 20 (04)
  • [5] Envelope and spectral frequency-following responses to vowel sounds
    Aiken, Steven J.
    Picton, Terence W.
    HEARING RESEARCH, 2008, 245 (1-2) : 35 - 47
  • [6] Predicting EEG Responses to Attended Speech via Deep Neural Networks for Speech
    Alickovic, Emina
    Dorszewski, Tobias
    Christiansen, Thomas U.
    Eskelund, Kasper
    Gizzi, Leonardo
    Skoglund, Martin A.
    Wendt, Dorothea
    2023 45TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY, EMBC, 2023,
  • [7] COHERENCE ANALYSIS OF ENVELOPE-FOLLOWING RESPONSES (EFRS) AND FREQUENCY-FOLLOWING RESPONSES (FFRS) IN INFANTS AND ADULTS
    LEVI, EC
    FOLSOM, RC
    DOBIE, RA
    HEARING RESEARCH, 1995, 89 (1-2) : 21 - 27
  • [8] Decoding Imagined Speech using Wavelet Features and Deep Neural Networks
    Panachakel, Jerrin Thomas
    Ramakrishnan, A. G.
    Ananthapadmanabha, T., V
    2019 IEEE 16TH INDIA COUNCIL INTERNATIONAL CONFERENCE (IEEE INDICON 2019), 2019,
  • [9] Decoding Visual Responses based on Deep Neural Networks with Ear-EEG Signals
    Lee, Young-Eun
    Lee, Minji
    2020 8TH INTERNATIONAL WINTER CONFERENCE ON BRAIN-COMPUTER INTERFACE (BCI), 2020, : 153 - 158
  • [10] Attentional Modulation of the Cortical Contribution to the Frequency-Following Response Evoked by Continuous Speech
    Schueller, Alina
    Schilling, Achim
    Krauss, Patrick
    Rampp, Stefan
    Reichenbach, Tobias
    JOURNAL OF NEUROSCIENCE, 2023, 43 (44): : 7429 - 7440