Decoding imagined speech from EEG signals using hybrid-scale spatial-temporal dilated convolution network

被引:16
|
作者
Li, Fu [1 ]
Chao, Weibing [1 ]
Li, Yang [1 ]
Fu, Boxun [1 ]
Ji, Youshuo [1 ]
Wu, Hao [1 ]
Shi, Guangming [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Key Lab Intelligent Percept & Image Understanding, Minist Educ, Xian, Peoples R China
基金
中国博士后科学基金;
关键词
brain-computer interface (BCI); EEG-based imagined speech recognition; hybrid-scale; spatial-temporal network; BRAIN-COMPUTER INTERFACES; NEURAL-NETWORKS; CLASSIFICATION; COMMUNICATION; IMAGERY;
D O I
10.1088/1741-2552/ac13c0
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective. Directly decoding imagined speech from electroencephalogram (EEG) signals has attracted much interest in brain-computer interface applications, because it provides a natural and intuitive communication method for locked-in patients. Several methods have been applied to imagined speech decoding, but how to construct spatial-temporal dependencies and capture long-range contextual cues in EEG signals to better decode imagined speech should be considered. Approach. In this study, we propose a novel model called hybrid-scale spatial-temporal dilated convolution network (HS-STDCN) for EEG-based imagined speech recognition. HS-STDCN integrates feature learning from temporal and spatial information into a unified end-to-end model. To characterize the temporal dependencies of the EEG sequences, we adopted a hybrid-scale temporal convolution layer to capture temporal information at multiple levels. A depthwise spatial convolution layer was then designed to construct intrinsic spatial relationships of EEG electrodes, which can produce a spatial-temporal representation of the input EEG data. Based on the spatial-temporal representation, dilated convolution layers were further employed to learn long-range discriminative features for the final classification. Main results. To evaluate the proposed method, we compared the HS-STDCN with other existing methods on our collected dataset. The HS-STDCN achieved an averaged classification accuracy of 54.31% for decoding eight imagined words, which is significantly better than other methods at a significance level of 0.05. Significance. The proposed HS-STDCN model provided an effective approach to make use of both the temporal and spatial dependencies of the input EEG signals for imagined speech recognition. We also visualized the word semantic differences to analyze the impact of word semantics on imagined speech recognition, investigated the important regions in the decoding process, and explored the use of fewer electrodes to achieve comparable performance.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] A Spatial-Temporal Graph Attention Network for Automated Detection and Width Estimation of Cortical Spreading Depression Using Scalp EEG
    Wang, Han Yi
    Liu, Xujin
    Grover, Pulkit
    Chamanzar, Alireza
    2023 45TH ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY, EMBC, 2023,
  • [42] Learning Multiband-Temporal-Spatial EEG Representations of Emotions Using Lightweight Temporal Convolution and 3D Convolutional Neural Network
    Wu, Fengjie
    Liu, Jiarui
    Yang, Jisen
    Zhang, Lihan
    He, Yan
    Lin, Zhaolong
    IEEE ACCESS, 2024, 12 : 132016 - 132026
  • [43] MAST-GCN: Multi-Scale Adaptive Spatial-Temporal Graph Convolutional Network for EEG-Based Depression Recognition
    Lu, Haifeng
    You, Zhiyang
    Guo, Yi
    Hu, Xiping
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (04) : 1985 - 1996
  • [44] Status detection from spatial-temporal data in pipeline network using data transformation convolutional neural network
    Hu, Xuguang
    Zhang, Huaguang
    Ma, Dazhong
    Wang, Rui
    NEUROCOMPUTING, 2019, 358 : 401 - 413
  • [45] Two Feature Fusion Network Based on Efficient Deep Residual Shrinkage Spatial-Temporal Graph Convolution for Emotion Recognition from Gaits
    Liu, Xiaoqing
    Zhang, Sen
    Xiao, Wendong
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 5824 - 5829
  • [46] Evaluation of temporal, spatial and spectral filtering in CSP-based methods for decoding pedaling-based motor tasks using EEG signals
    Blanco-Diaz, Cristian Felipe
    Guerrero-Mendez, Cristian David
    Delisle-Rodriguez, Denis
    Jaramillo-Isaza, Sebastian
    Ruiz-Olaya, Andres Felipe
    Frizera-Neto, Anselmo
    de Souza, Alberto Ferreira
    Bastos-Filho, Teodiano
    BIOMEDICAL PHYSICS & ENGINEERING EXPRESS, 2024, 10 (03)
  • [47] A hybrid deep learning based intrusion detection system using spatial-temporal representation of in-vehicle network traffic
    Lo, Wei
    Alqahtani, Hamed
    Thakur, Kutub
    Almadhor, Ahmad
    Chander, Subhash
    Kumar, Gulshan
    VEHICULAR COMMUNICATIONS, 2022, 35
  • [48] UApredictor: Urban Anomaly Prediction from Spatial-Temporal Data using Graph Transformer Neural Network
    Bhumika
    Das, Debasis
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [49] Single upper limb functional movements decoding from motor imagery EEG signals using wavelet neural network
    Zhou, Xiaobo
    Zou, Renling
    Huang, Xiayang
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2021, 70
  • [50] Upper limb complex movements decoding from pre-movement EEG signals using wavelet common spatial patterns
    Mohseni, Mahdieh
    Shalchyan, Vahid
    Jochumsen, Mads
    Niazi, Imran Khan
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2020, 183 (183)