Decoding Envelope and Frequency-Following EEG Responses to Continuous Speech Using Deep Neural Networks

被引:2
|
作者
Thornton, Mike D. [1 ]
Mandic, Danilo P. [2 ]
Reichenbach, Tobias J. [3 ]
机构
[1] Imperial Coll London, Dept Comp, London SW7 2RH, England
[2] Imperial Coll London, Dept Elect & Elect Engn, London SW7 2RH, England
[3] Friedrich Alexander Univ Erlangen Nurnberg, Dept Artificial Intelligence Biomed Engn, D-91052 Erlangen, Germany
关键词
Auditory attention decoding; deep learning; EEG signal processing;
D O I
10.1109/OJSP.2024.3378593
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The electroencephalogram (EEG) offers a non-invasive means by which a listener's auditory system may be monitored during continuous speech perception. Reliable auditory-EEG decoders could facilitate the objective diagnosis of hearing disorders, or find applications in cognitively-steered hearing aids. Previously, we developed decoders for the ICASSP Auditory EEG Signal Processing Grand Challenge (SPGC). These decoders placed first in the match-mismatch task: given a short temporal segment of EEG recordings, and two candidate speech segments, the task is to identify which of the two speech segments is temporally aligned, or matched, with the EEG segment. The decoders made use of cortical responses to the speech envelope, as well as speech-related frequency-following responses, to relate the EEG recordings to the speech stimuli. Here we comprehensively document the methods by which the decoders were developed. We extend our previous analysis by exploring the association between speaker characteristics (pitch and sex) and classification accuracy, and provide a full statistical analysis of the final performance of the decoders as evaluated on a heldout portion of the dataset. Finally, the generalisation capabilities of the decoders are characterised, by evaluating them using an entirely different dataset which contains EEG recorded under a variety of speech-listening conditions. The results show that the match-mismatch decoders achieve accurate and robust classification accuracies, and they can even serve as auditory attention decoders without additional training.
引用
收藏
页码:700 / 716
页数:17
相关论文
共 50 条
  • [21] Modelling envelope and temporal fine structure components of frequency-following responses in rat inferior colliculus
    Wang, Qian
    Li, Liang
    SCIENCE CHINA-TECHNOLOGICAL SCIENCES, 2017, 60 (07) : 966 - 973
  • [22] Modelling envelope and temporal fine structure components of frequency-following responses in rat inferior colliculus
    WANG Qian
    LI Liang
    Science China(Technological Sciences), 2017, (07) : 966 - 973
  • [23] Modelling envelope and temporal fine structure components of frequency-following responses in rat inferior colliculus
    Qian Wang
    Liang Li
    Science China Technological Sciences, 2017, 60 : 966 - 973
  • [24] The frequency-following response to assess the neural representation of spectral speech cues in older adults
    Chauvette, L.
    Fournier, P.
    Sharp, A.
    HEARING RESEARCH, 2022, 418
  • [25] Modelling envelope and temporal fine structure components of frequency-following responses in rat inferior colliculus
    WANG Qian
    LI Liang
    Science China(Technological Sciences), 2017, 60 (07) : 966 - 973
  • [26] DEEP LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR DECODING AND VISUALIZATION OF EEG PATHOLOGY
    Schirrmeister, R.
    Gemein, L.
    Eggensperger, K.
    Hutter, F.
    Ball, T.
    2017 IEEE SIGNAL PROCESSING IN MEDICINE AND BIOLOGY SYMPOSIUM (SPMB), 2017,
  • [27] Decoding Neural Responses to Motion-in-Depth Using EEG
    Himmelberg, Marc M.
    Segala, Federico G.
    Maloney, Ryan T.
    Harris, Julie M.
    Wade, Alex R.
    FRONTIERS IN NEUROSCIENCE, 2020, 14
  • [28] Speech Imagery Decoding Using EEG Signals and Deep Learning: A Survey
    Zhang, Liying
    Zhou, Yueying
    Gong, Peiliang
    Zhang, Daoqiang
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2025, 17 (01) : 22 - 39
  • [29] Neural encoding of voice pitch and formant structure at birth as revealed by frequency-following responses
    Sonia Arenillas-Alcón
    Jordi Costa-Faidella
    Teresa Ribas-Prats
    María Dolores Gómez-Roig
    Carles Escera
    Scientific Reports, 11
  • [30] Neural encoding of voice pitch and formant structure at birth as revealed by frequency-following responses
    Arenillas-Alcon, Sonia
    Costa-Faidella, Jordi
    Ribas-Prats, Teresa
    Dolores Gomez-Roig, Maria
    Escera, Carles
    SCIENTIFIC REPORTS, 2021, 11 (01)