Decoding imagined speech from EEG signals using hybrid-scale spatial-temporal dilated convolution network

被引:16
|
作者
Li, Fu [1 ]
Chao, Weibing [1 ]
Li, Yang [1 ]
Fu, Boxun [1 ]
Ji, Youshuo [1 ]
Wu, Hao [1 ]
Shi, Guangming [1 ]
机构
[1] Xidian Univ, Sch Artificial Intelligence, Key Lab Intelligent Percept & Image Understanding, Minist Educ, Xian, Peoples R China
基金
中国博士后科学基金;
关键词
brain-computer interface (BCI); EEG-based imagined speech recognition; hybrid-scale; spatial-temporal network; BRAIN-COMPUTER INTERFACES; NEURAL-NETWORKS; CLASSIFICATION; COMMUNICATION; IMAGERY;
D O I
10.1088/1741-2552/ac13c0
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective. Directly decoding imagined speech from electroencephalogram (EEG) signals has attracted much interest in brain-computer interface applications, because it provides a natural and intuitive communication method for locked-in patients. Several methods have been applied to imagined speech decoding, but how to construct spatial-temporal dependencies and capture long-range contextual cues in EEG signals to better decode imagined speech should be considered. Approach. In this study, we propose a novel model called hybrid-scale spatial-temporal dilated convolution network (HS-STDCN) for EEG-based imagined speech recognition. HS-STDCN integrates feature learning from temporal and spatial information into a unified end-to-end model. To characterize the temporal dependencies of the EEG sequences, we adopted a hybrid-scale temporal convolution layer to capture temporal information at multiple levels. A depthwise spatial convolution layer was then designed to construct intrinsic spatial relationships of EEG electrodes, which can produce a spatial-temporal representation of the input EEG data. Based on the spatial-temporal representation, dilated convolution layers were further employed to learn long-range discriminative features for the final classification. Main results. To evaluate the proposed method, we compared the HS-STDCN with other existing methods on our collected dataset. The HS-STDCN achieved an averaged classification accuracy of 54.31% for decoding eight imagined words, which is significantly better than other methods at a significance level of 0.05. Significance. The proposed HS-STDCN model provided an effective approach to make use of both the temporal and spatial dependencies of the input EEG signals for imagined speech recognition. We also visualized the word semantic differences to analyze the impact of word semantics on imagined speech recognition, investigated the important regions in the decoding process, and explored the use of fewer electrodes to achieve comparable performance.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Temporal-spatial convolutional residual network for decoding attempted movement related EEG signals of subjects with spinal cord injury
    Mirzabagherian, Hamed
    Menhaj, Mohammad Bagher
    Suratgar, Amir Abolfazl
    Talebi, Nasibeh
    Sardari, Mohammad Reza Abbasi
    Sajedin, Atena
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 164
  • [32] A Neural Network for EEG Emotion Recognition that Combines CNN and Transformer for Multi-scale Spatial-temporal Feature Extraction
    Hu, Zhangfang
    Wu, Haoze
    He, Lingxiao
    IAENG International Journal of Computer Science, 2024, 51 (08) : 1094 - 1104
  • [33] Dynamic Origin-Destination Flow Prediction Using Spatial-Temporal Graph Convolution Network With Mobile Phone Data
    Liu, Zhichen
    Liu, Zhiyuan
    Fu, Xiao
    IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE, 2022, 14 (05) : 147 - 161
  • [34] CWSTR-Net: A Channel-Weighted Spatial-Temporal Residual Network based on nonsmooth nonnegative matrix factorization for fatigue detection using EEG signals
    Li, Xueping
    Tang, Jiahao
    Li, Xue
    Yang, Yuan
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 97
  • [35] From Regional to Global Brain: A Novel Hierarchical Spatial-Temporal Neural Network Model for EEG Emotion Recognition
    Li, Yang
    Zheng, Wenming
    Wang, Lei
    Zong, Yuan
    Cui, Zhen
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (02) : 568 - 578
  • [36] Pain Detection from Facial Landmarks Using Spatial-temporal Deep Neural Network
    Choo, Keng Wah
    Du, Tiehua
    THIRTEENTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2021), 2021, 11878
  • [37] Decoding Visual Imagery from EEG Signals using Visual Perception Guided Network Training Method
    Kwon, Byoung-Hee
    Lee, Byeong-Hoo
    Cho, Jeong-Hyun
    Jeong, Ji-Hoon
    10TH INTERNATIONAL WINTER CONFERENCE ON BRAIN-COMPUTER INTERFACE (BCI2022), 2022,
  • [38] Subject-independent emotion recognition of EEG signals using graph attention-based spatial-temporal pattern learning
    Zhu, Yiwen
    Guo, Yeshuang
    Zhu, Wenzhe
    Di, Lare
    Yin, Thong
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 7070 - 7075
  • [39] Hybrid Network Using Dynamic Graph Convolution and Temporal Self-Attention for EEG-Based Emotion Recognition
    Dalian University of Technology, Department of Computer Science and Technology, Dalian
    116024, China
    不详
    313000, China
    IEEE Trans. Neural Networks Learn. Sys., 2162, 12 (18565-18575):
  • [40] Hybrid Network Using Dynamic Graph Convolution and Temporal Self-Attention for EEG-Based Emotion Recognition
    Cheng, Cheng
    Yu, Zikang
    Zhang, Yong
    Feng, Lin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 11