Cross-Modality Self-Attention and Fusion-Based Neural Network for Lower Limb Locomotion Mode Recognition

被引:1
|
作者
Zhao, Changchen [1 ]
Liu, Kai [2 ]
Zheng, Hao [3 ]
Song, Wenbo [4 ]
Pei, Zhongcai [3 ]
Chen, Weihai [5 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci, Hangzhou 310018, Peoples R China
[2] Zhejiang Univ Technol, Coll Informat Engn, Hangzhou 310023, Peoples R China
[3] Beihang Univ, Hangzhou Innovat Inst, Hangzhou 310051, Peoples R China
[4] Jilin Normal Univ, Coll Phys Educ, Siping 136000, Peoples R China
[5] Anhui Univ, Sch Elect Engn & Automat, Hefei 230601, Peoples R China
关键词
Cross-modality interaction; self-attention; locomotion mode recognition; lower limb; neural network; INTENT RECOGNITION; PREDICTION; STRATEGY; GAZE;
D O I
10.1109/TASE.2024.3421276
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Although there are many wearable sensors that make the acquisition of multi-modality data easier, effective feature extraction and fusion of the data is still challenging for lower limb locomotion mode recognition. In this article, a novel neural network is proposed for accurate prediction of five common lower limb locomotion modes including level walking, ramp ascent, ramp descent, stair ascent, and stair descent. First, the encoder-decoder structure is employed to enrich the channel diversity for the separation of the useful patterns from combined patterns. Second, a self-attention based cross-modality interaction module is proposed, which enables bilateral information flow between two encoding paths to fully exploit the interdependencies and to find complementary information between modalities. Third, a multi-modality fusion module is designed where the complementary features are fused by a channel-wise weighted summation whose coefficients are learned end-to-end. A benchmark dataset is collected from 10 health subjects containing EMG and IMU signals and five locomotion modes. Extensive experiments are conducted on one publicly available dataset ENABL3S and one self-collected dataset. The results show that the proposed method outperforms the compared methods with higher classification accuracy. The proposed method achieves a classification accuracy of 98.25 $\%$ on ENABL3S dataset and 95.51 $\%$ on the self-collected dataset. Note to Practitioners-This article aims to solve the real challenges encountered when intelligent recognition algorithms are applied in wearable robots: how to effectively and efficiently fuse the multi-modality data for better decision-making. First, most existing methods directly concatenate the multi-modality data, which increases the data dimensionality and brings computational burden. Second, existing recognition neural networks continuously compress the feature size such that the discriminative patterns are submerged in the noise and thus difficult to be identified. This research decomposes the mixed input signals on the channel dimension such that the useful patterns can be separated. Moreover, this research employs self-attention mechanism to associate correlations between two modalities and use this correlation as a new feature for subsequent representation learning, generating new, compact, and complementary features for classification. We demonstrate that the proposed network achieves 98.25 $\%$ accuracy and 3.5 ms prediction time. We anticipate that the proposed network could be a general scientific and practical methodology of multi-modality signal fusion and feature learning for intelligent systems.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Locomotion Mode Recognition for Walking on Three Terrains Based on sEMG of Lower Limb and Back Muscles
    Zhou, Hui
    Yang, Dandan
    Li, Zhengyi
    Zhou, Dao
    Gao, Junfeng
    Guan, Jinan
    SENSORS, 2021, 21 (09)
  • [42] Cross-modal hashing network based on self-attention similarity transfer
    Liang H.
    Wang H.
    Wang D.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2024, 50 (02): : 615 - 622
  • [43] ConViViT - A Deep Neural Network Combining Convolutions and Factorized Self-Attention for Human Activity Recognition
    Dokkar, Rachid Reda
    Chaieb, Faten
    Drira, Hassen
    Aberkane, Arezki
    2023 IEEE 25TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING, MMSP, 2023,
  • [44] Infrared Small Target Detection Based on Fusion Full Convolutional Network and Self-Attention
    Duan, Peipei
    Zhang, Yan
    SECOND IYSF ACADEMIC SYMPOSIUM ON ARTIFICIAL INTELLIGENCE AND COMPUTER ENGINEERING, 2021, 12079
  • [45] Electrocardiogram signal classification based on fusion method of residual network and self-attention mechanism
    Yuan C.
    Liu Z.
    Wang C.
    Yang F.
    Shengwu Yixue Gongchengxue Zazhi/Journal of Biomedical Engineering, 2023, 40 (03): : 474 - 481
  • [46] Recursive Self-Attention Modules-Based Network for Panchromatic and Multispectral Image Fusion
    Liu, Chuang
    Wei, Lu
    Zhang, Zhiqi
    Feng, Xiaoxiao
    Xiang, Shao
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2023, 16 : 10067 - 10083
  • [47] Self-attention Based Multimodule Fusion Graph Convolution Network for Traffic Flow Prediction
    Li, Lijie
    Shao, Hongyang
    Chen, Junhao
    Wang, Ye
    DATA SCIENCE (ICPCSEE 2022), PT I, 2022, 1628 : 3 - 16
  • [48] EEG-Based Emotion Recognition Using Convolutional Recurrent Neural Network with Multi-Head Self-Attention
    Hu, Zhangfang
    Chen, Libujie
    Luo, Yuan
    Zhou, Jingfan
    APPLIED SCIENCES-BASEL, 2022, 12 (21):
  • [49] Mandarin Recognition Based on Self-Attention Mechanism with Deep Convolutional Neural Network (DCNN)-Gated Recurrent Unit (GRU)
    Chen, Xun
    Wang, Chengqi
    Hu, Chao
    Wang, Qin
    Big Data and Cognitive Computing, 2024, 8 (12)
  • [50] Convolutional neural network with coarse-to-fine resolution fusion and residual learning structures for cross-modality image synthesis
    Wu, Guoqing
    Chen, Xi
    Shi, Zhifeng
    Zhang, Dachuan
    Hu, Zhaoyu
    Mao, Ying
    Wang, Yuanyuan
    Yu, Jinhua
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2022, 71