Cross-Modality Self-Attention and Fusion-Based Neural Network for Lower Limb Locomotion Mode Recognition

被引:1
|
作者
Zhao, Changchen [1 ]
Liu, Kai [2 ]
Zheng, Hao [3 ]
Song, Wenbo [4 ]
Pei, Zhongcai [3 ]
Chen, Weihai [5 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci, Hangzhou 310018, Peoples R China
[2] Zhejiang Univ Technol, Coll Informat Engn, Hangzhou 310023, Peoples R China
[3] Beihang Univ, Hangzhou Innovat Inst, Hangzhou 310051, Peoples R China
[4] Jilin Normal Univ, Coll Phys Educ, Siping 136000, Peoples R China
[5] Anhui Univ, Sch Elect Engn & Automat, Hefei 230601, Peoples R China
关键词
Cross-modality interaction; self-attention; locomotion mode recognition; lower limb; neural network; INTENT RECOGNITION; PREDICTION; STRATEGY; GAZE;
D O I
10.1109/TASE.2024.3421276
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Although there are many wearable sensors that make the acquisition of multi-modality data easier, effective feature extraction and fusion of the data is still challenging for lower limb locomotion mode recognition. In this article, a novel neural network is proposed for accurate prediction of five common lower limb locomotion modes including level walking, ramp ascent, ramp descent, stair ascent, and stair descent. First, the encoder-decoder structure is employed to enrich the channel diversity for the separation of the useful patterns from combined patterns. Second, a self-attention based cross-modality interaction module is proposed, which enables bilateral information flow between two encoding paths to fully exploit the interdependencies and to find complementary information between modalities. Third, a multi-modality fusion module is designed where the complementary features are fused by a channel-wise weighted summation whose coefficients are learned end-to-end. A benchmark dataset is collected from 10 health subjects containing EMG and IMU signals and five locomotion modes. Extensive experiments are conducted on one publicly available dataset ENABL3S and one self-collected dataset. The results show that the proposed method outperforms the compared methods with higher classification accuracy. The proposed method achieves a classification accuracy of 98.25 $\%$ on ENABL3S dataset and 95.51 $\%$ on the self-collected dataset. Note to Practitioners-This article aims to solve the real challenges encountered when intelligent recognition algorithms are applied in wearable robots: how to effectively and efficiently fuse the multi-modality data for better decision-making. First, most existing methods directly concatenate the multi-modality data, which increases the data dimensionality and brings computational burden. Second, existing recognition neural networks continuously compress the feature size such that the discriminative patterns are submerged in the noise and thus difficult to be identified. This research decomposes the mixed input signals on the channel dimension such that the useful patterns can be separated. Moreover, this research employs self-attention mechanism to associate correlations between two modalities and use this correlation as a new feature for subsequent representation learning, generating new, compact, and complementary features for classification. We demonstrate that the proposed network achieves 98.25 $\%$ accuracy and 3.5 ms prediction time. We anticipate that the proposed network could be a general scientific and practical methodology of multi-modality signal fusion and feature learning for intelligent systems.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Gabor Log-Euclidean Gaussian and its fusion with deep network based on self-attention for face recognition
    Li, Chaorong
    Huang, Wei
    Huang, Yuanyuan
    APPLIED SOFT COMPUTING, 2022, 116
  • [32] A self-attention based neural architecture for Chinese medical named entity recognition
    Wan, Qian
    Liu, Jie
    Wei, Luona
    Ji, Bin
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2020, 17 (04) : 3498 - 3511
  • [33] Dunhuang murals contour generation network based on convolution and self-attention fusion
    Liu, Baokai
    He, Fengjie
    Du, Shiqiang
    Zhang, Kaiwu
    Wang, Jianhua
    APPLIED INTELLIGENCE, 2023, 53 (19) : 22073 - 22085
  • [34] GDN-CMCF: A Gated Disentangled Network With Cross-Modality Consensus Fusion for Multimodal Named Entity Recognition
    Huang, Guoheng
    He, Qin
    Dai, Zihao
    Zhong, Guo
    Yuan, Xiaochen
    Pun, Chi-Man
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, 11 (03) : 3944 - 3954
  • [35] Dunhuang murals contour generation network based on convolution and self-attention fusion
    Baokai Liu
    Fengjie He
    Shiqiang Du
    Kaiwu Zhang
    Jianhua Wang
    Applied Intelligence, 2023, 53 : 22073 - 22085
  • [36] Multi-level feature fusion capsule network with self-attention for facial expression recognition
    Huang, Zhiji
    Yu, Songsen
    Liang, Jun
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (02)
  • [37] Masked face recognition based on knowledge distillation and convolutional self-attention network
    Wan, Weiguo
    Wen, Runlin
    Yao, Li
    Yang, Yong
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, : 2269 - 2284
  • [38] SAFSN: A Self-Attention Based Neural Network for Encrypted Mobile Traffic Classification
    Zhang, Chengyuan
    An, Changqing
    Wang, Jessie Hui
    Zhao, Ziyi
    Yu, Tao
    Wang, Jilong
    IEEE CONGRESS ON CYBERMATICS / 2021 IEEE INTERNATIONAL CONFERENCES ON INTERNET OF THINGS (ITHINGS) / IEEE GREEN COMPUTING AND COMMUNICATIONS (GREENCOM) / IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING (CPSCOM) / IEEE SMART DATA (SMARTDATA), 2021, : 330 - 337
  • [39] Self-Attention Based Neural Network for Behavioral Modeling and Predistortion of Power Amplifiers
    Xiong, Zeren
    Guo, Zhiheng
    Wang, Xijun
    Chen, Xiang
    2023 INTERNATIONAL CONFERENCE ON FUTURE COMMUNICATIONS AND NETWORKS, FCN, 2023,
  • [40] Rapid nuclide identification algorithm based on self-attention mechanism neural network
    Sun, Jiaqian
    Niu, Deqing
    Liang, Jie
    Hou, Xin
    Li, Linshan
    ANNALS OF NUCLEAR ENERGY, 2024, 207