Transformer-based network with temporal depthwise convolutions for sEMG recognition

被引:7
|
作者
Wang, Zefeng [1 ]
Yao, Junfeng [1 ,2 ,3 ]
Xu, Meiyan [4 ]
Jiang, Min [1 ,3 ]
Su, Jinsong [1 ]
机构
[1] Xiamen Univ, Sch Informat, Xiamen 361001, Fujian, Peoples R China
[2] Xiamen Univ, Sch Film, Xiamen 361000, Fujian, Peoples R China
[3] Taiwan Minist Culture & Tourism, Key Lab Digital Protect & Intelligent Proc Intang, Taipei City, Taiwan
[4] Minnan Normal Univ, Sch Comp Sci, Zhangzhou 363000, Fujian, Peoples R China
关键词
Surface electromyography; Feature learning; Gesture recognition; Transformer; Self-attention; Temporal depthwise convolution;
D O I
10.1016/j.patcog.2023.109967
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Considerable progress has been made in pattern recognition of surface electromyography (sEMG) with deep learning, bringing improvements to sEMG-based gesture classification. Current deep learning techniques are mainly based on convolutional neural networks (CNNs), recurrent neural networks (RNNs), and their hybrids. However, CNNs focus on spatial and local information, while RNNs are unparallelizable, and they suffer from gradient vanishing/exploding. Their hybrids often face problems of model complexity and high computational cost. Because sEMG signals have a sequential nature, motivated by the sequence modeling network Transformer and its self-attention mechanism, we propose a Transformer-based network, temporal depthwise convolutional Transformer (TDCT), for sparse sEMG recognition. With this network, higher recognition accuracy is achieved with fewer convolution parameters and a lower computational cost. Specifically, this network has parallel capability and can capture long-range features inside sEMG signals. We improve the locality and channel correlation capture of multi-head self-attention (MSA) for sEMG modeling by replacing the linear transforma-tion with the proposed temporal depthwise convolution (TDC), which can reduce the convolution parameters and computations for feature learning performance. Four sEMG datasets, Ninapro DB1, DB2, DB5, and OYDB, are used for evaluations and comparisons. In the results, our model outperforms other methods, including Transformer-based networks, in most windows at recognizing the raw signals of sparse sEMG, thus achieving state-of-the-art classification accuracy.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] A transformer-based network for speech recognition
    Tang L.
    [J]. International Journal of Speech Technology, 2023, 26 (2) : 531 - 539
  • [2] A Transformer-Based Network for Dynamic Hand Gesture Recognition
    D'Eusanio, Andrea
    Simoni, Alessandro
    Pini, Stefano
    Borghi, Guido
    Vezzani, Roberto
    Cucchiara, Rita
    [J]. 2020 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2020), 2020, : 623 - 632
  • [3] Traffic sign recognition algorithm based on depthwise separable convolutions
    Yang Jin-sheng
    Yang Yan-nan
    Li Tian-jiao
    [J]. CHINESE JOURNAL OF LIQUID CRYSTALS AND DISPLAYS, 2019, 34 (12) : 1191 - 1201
  • [4] A Transformer-based Radical Analysis Network for Chinese Character Recognition
    Yang, Chen
    Wang, Qing
    Du, Jun
    Zhang, Jianshu
    Wu, Changjie
    Wang, Jiaming
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 3714 - 3719
  • [5] End to end transformer-based contextual speech recognition based on pointer network
    Lin, Binghuai
    Wang, Liyuan
    [J]. INTERSPEECH 2021, 2021, : 2087 - 2091
  • [6] A Transformer-Based Deep Learning Network for Underwater Acoustic Target Recognition
    Feng, Sheng
    Zhu, Xiaoqian
    [J]. IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [7] Learning continuous dynamic network representation with transformer-based temporal graph neural network
    Li, Yingji
    Wu, Yue
    Sun, Mingchen
    Yang, Bo
    Wang, Ying
    [J]. INFORMATION SCIENCES, 2023, 649
  • [8] Pedestrian Attribute Recognition Model based on Adaptive Weight and Depthwise Separable Convolutions
    Lin, Xiao
    [J]. PROCEEDINGS OF 2020 IEEE 5TH INFORMATION TECHNOLOGY AND MECHATRONICS ENGINEERING CONFERENCE (ITOEC 2020), 2020, : 830 - 833
  • [9] Carbon emissions forecasting based on temporal graph transformer-based attentional neural network
    Wu, Xingping
    Yuan, Qiheng
    Zhou, Chunlei
    Chen, Xiang
    Xuan, Donghai
    Song, Jinwei
    [J]. JOURNAL OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING, 2024, 24 (03) : 1405 - 1421
  • [10] Temporal fusion transformer-based prediction in aquaponics
    Metin, Ahmet
    Kasif, Ahmet
    Catal, Cagatay
    [J]. JOURNAL OF SUPERCOMPUTING, 2023, 79 (17): : 19934 - 19958