STAnet: A Spatiotemporal Attention Network for Decoding Auditory Spatial Attention From EEG

被引:56
|
作者
Su, Enze [1 ]
Cai, Siqi [2 ]
Xie, Longhan [1 ]
Li, Haizhou [2 ,3 ,4 ]
Schultz, Tanja [5 ]
机构
[1] South China Univ Technol, Shien Ming Wu Sch Intelligent Engn, Guangzhou 510460, Guangdong, Peoples R China
[2] Natl Univ Singapore, Dept Elect & Comp Engn, Singapore, Singapore
[3] Chinese Univ Hong Kong Shenzhen, Sch Data Sci, Shenzhen, Peoples R China
[4] Univ Bremen, Machine Listening Lab, Bremen, Germany
[5] Univ Bremen, Cognit Syst Lab, Bremen, Germany
基金
中国国家自然科学基金;
关键词
Electroencephalography; Spatiotemporal phenomena; Feature extraction; Brain modeling; Decoding; Speech enhancement; Pipelines; Auditory attention; brain-computer interface; electroencephalography; spatial attention; temporal attention; CORTICAL REPRESENTATION; SELECTIVE ATTENTION; ATTENDED SPEECH; TRACKING; BRAIN; MODULATION; RESPONSES; DYNAMICS; CHAOS; HAND;
D O I
10.1109/TBME.2022.3140246
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective: Humans are able to localize the source of a sound. This enables them to direct attention to a particular speaker in a cocktail party. Psycho-acoustic studies show that the sensory cortices of the human brain respond to the location of sound sources differently, and the auditory attention itself is a dynamic and temporally based brain activity. In this work, we seek to build a computational model which uses both spatial and temporal information manifested in EEG signals for auditory spatial attention detection (ASAD). Methods: We propose an end-to-end spatiotemporal attention network, denoted as STAnet, to detect auditory spatial attention from EEG. The STAnet is designed to assign differentiated weights dynamically to EEG channels through a spatial attention mechanism, and to temporal patterns in EEG signals through a temporal attention mechanism. Results: We report the ASAD experiments on two publicly available datasets. The STAnet outperforms other competitive models by a large margin under various experimental conditions. Its attention decision for 1-second decision window outperforms that of the state-of-the-art techniques for 10-second decision window. Experimental results also demonstrate that the STAnet achieves competitive performance on EEG signals ranging from 64 to as few as 16 channels. Conclusion: This study provides evidence suggesting that efficient low-density EEG online decoding is within reach. Significance: This study also marks an important step towards the practical implementation of ASAD in real life applications.
引用
收藏
页码:2233 / 2242
页数:10
相关论文
共 50 条
  • [41] Decoding Visual Spatial Attention Control
    Meyyappan, Sreenivasan
    Rajan, Abhijit
    Yang, Qiang
    Mangun, George R.
    Ding, Mingzhou
    ENEURO, 2025, 12 (03)
  • [42] Brain Topology Modeling With EEG-Graphs for Auditory Spatial Attention Detection
    Cai, Siqi
    Schultz, Tanja
    Li, Haizhou
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2024, 71 (01) : 171 - 182
  • [43] Spatiotemporal dynamics of auditory attention synchronize with speech
    Woestmann, Malte
    Herrmann, Bjoern
    Maess, Burkhard
    Obleser, Jonas
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2016, 113 (14) : 3873 - 3878
  • [44] Privacy-preserving Network BMI Decoding of Covert Spatial Attention
    Nakachi, Takayuki
    Ishihara, Hiroyuki
    Kiya, Hitoshi
    2018 12TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATION SYSTEMS (ICSPCS), 2018,
  • [45] EEG-based auditory attention decoding with audiovisual speech for hearing-impaired listeners
    Wang, Bo
    Xu, Xiran
    Niu, Yadong
    Wu, Chao
    Wu, Xihong
    Chen, Jing
    CEREBRAL CORTEX, 2023, 33 (22) : 10972 - 10983
  • [46] Decoding auditory attention to instruments in polyphonic music using single-trial EEG classification
    Treder, M. S.
    Purwins, H.
    Miklody, D.
    Sturm, I.
    Blankertz, B.
    JOURNAL OF NEURAL ENGINEERING, 2014, 11 (02)
  • [47] CSDA: Cross-Session Domain Adaptation in Auditory Attention Decoding of EEG for a Single Subject
    Dong, Xingguang
    Fan, Cunhang
    Zhang, Hongyu
    Yang, Xiaoke
    Zhang, Sheng
    Zhou, Jian
    Lv, Zhao
    2024 IEEE 14TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING, ISCSLP 2024, 2024, : 451 - 455
  • [48] COGNITIVE-DRIVEN CONVOLUTIONAL BEAMFORMING USING EEG-BASED AUDITORY ATTENTION DECODING
    Aroudi, Ali
    Delcroix, Marc
    Nakatani, Tomohiro
    Kinoshita, Keisuke
    Araki, Shoko
    Doclo, Simon
    PROCEEDINGS OF THE 2020 IEEE 30TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2020,
  • [49] Decoding auditory and tactile attention for use in an EEG-based brain-computer interface
    An, Winko W.
    Si-Mohammed, Hakim
    Huang, Nicholas
    Gamper, Hannes
    Lee, Adrian K. C.
    Holz, Christian
    Johnston, David
    Jalobeanu, Mihai
    Emmanouilidou, Dimitra
    Cutrell, Edward
    Wilson, Andrew
    Tashev, Ivan
    2020 8TH INTERNATIONAL WINTER CONFERENCE ON BRAIN-COMPUTER INTERFACE (BCI), 2020, : 42 - 47
  • [50] Cognitive-Driven Binaural Beamforming Using EEG-Based Auditory Attention Decoding
    Aroudi, Ali
    Doclo, Simon
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 862 - 875