Decoding imagined speech from EEG signals using hybrid-scale spatial-temporal dilated convolution network

被引:16
|
作者
Li, Fu [1 ]
Chao, Weibing [1 ]
Li, Yang [1 ]
Fu, Boxun [1 ]
Ji, Youshuo [1 ]
Wu, Hao [1 ]
Shi, Guangming [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Key Lab Intelligent Percept & Image Understanding, Minist Educ, Xian, Peoples R China
基金
中国博士后科学基金;
关键词
brain-computer interface (BCI); EEG-based imagined speech recognition; hybrid-scale; spatial-temporal network; BRAIN-COMPUTER INTERFACES; NEURAL-NETWORKS; CLASSIFICATION; COMMUNICATION; IMAGERY;
D O I
10.1088/1741-2552/ac13c0
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective. Directly decoding imagined speech from electroencephalogram (EEG) signals has attracted much interest in brain-computer interface applications, because it provides a natural and intuitive communication method for locked-in patients. Several methods have been applied to imagined speech decoding, but how to construct spatial-temporal dependencies and capture long-range contextual cues in EEG signals to better decode imagined speech should be considered. Approach. In this study, we propose a novel model called hybrid-scale spatial-temporal dilated convolution network (HS-STDCN) for EEG-based imagined speech recognition. HS-STDCN integrates feature learning from temporal and spatial information into a unified end-to-end model. To characterize the temporal dependencies of the EEG sequences, we adopted a hybrid-scale temporal convolution layer to capture temporal information at multiple levels. A depthwise spatial convolution layer was then designed to construct intrinsic spatial relationships of EEG electrodes, which can produce a spatial-temporal representation of the input EEG data. Based on the spatial-temporal representation, dilated convolution layers were further employed to learn long-range discriminative features for the final classification. Main results. To evaluate the proposed method, we compared the HS-STDCN with other existing methods on our collected dataset. The HS-STDCN achieved an averaged classification accuracy of 54.31% for decoding eight imagined words, which is significantly better than other methods at a significance level of 0.05. Significance. The proposed HS-STDCN model provided an effective approach to make use of both the temporal and spatial dependencies of the input EEG signals for imagined speech recognition. We also visualized the word semantic differences to analyze the impact of word semantics on imagined speech recognition, investigated the important regions in the decoding process, and explored the use of fewer electrodes to achieve comparable performance.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Decoding Imagined Speech from EEG Data: A Hybrid Deep Learning Approach to Capturing Spatial and Temporal Features
    Alharbi, Yasser F.
    Alotaibi, Yousef A.
    LIFE-BASEL, 2024, 14 (11):
  • [2] Convolution spatial-temporal attention network for EEG emotion recognition
    Cao, Lei
    Yu, Binlong
    Dong, Yilin
    Liu, Tianyu
    Li, Jie
    PHYSIOLOGICAL MEASUREMENT, 2024, 45 (12)
  • [3] SPATIAL-TEMPORAL GRAPH CONVOLUTION NETWORK FOR MULTICHANNEL SPEECH ENHANCEMENT
    Hao, Minghui
    Yu, Jingjing
    Zhang, Luyao
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6512 - 6516
  • [4] STA-Net: Spatial-temporal alignment network for hybrid EEG-fNIRS decoding
    Liu, Mutian
    Yang, Banghua
    Meng, Lin
    Zhang, Yonghuai
    Gao, Shouwei
    Zan, Peng
    Xia, Xinxing
    INFORMATION FUSION, 2025, 119
  • [5] Decoding Imagined Speech From EEG Using Transfer Learning
    Panachakel, Jerrin Thomas
    Ganesan, Ramakrishnan Angarai
    IEEE ACCESS, 2021, 9 : 135371 - 135383
  • [6] A Hybrid-Convolution Spatial-Temporal Recurrent Network For Traffic Flow Prediction
    Zhang, Xu
    Wen, Shunjie
    Yan, Liang
    Feng, Jiangfan
    Xia, Ying
    COMPUTER JOURNAL, 2024, 67 (01): : 236 - 252
  • [7] Deep Spatial-Temporal Network Based on Residual Networks and Dilated Convolution for Traffic Flow Prediction
    Dong, Liangming
    Zhang, Xinxin
    Liu, Lijuan
    2022 IEEE 7TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION ENGINEERING, ICITE, 2022, : 284 - 289
  • [8] A Spatial Filter Temporal Graph Convolutional Network for decoding motor imagery EEG signals
    Tang, Xianlun
    Zhang, Jing
    Qi, Yidan
    Liu, Ke
    Li, Rui
    Wang, Huiming
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [9] Decoding Braking Intentions From EEG Signals: A Convolution-Transformer Network Approach
    Li, Jiaxiang
    Liang, Xinbin
    Liu, Yadong
    Hu, Dewen
    Sun, Jianxiang
    Zhou, Tianjian
    IEEE SENSORS JOURNAL, 2024, 24 (22) : 38329 - 38343
  • [10] Attention-Based Multiscale Spatial-Temporal Convolutional Network for Motor Imagery EEG Decoding
    Zhang, Yu
    Li, Penghai
    Cheng, Longlong
    Li, Mingji
    Li, Hongji
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (01) : 2423 - 2434