A Transformer-Based Gesture Prediction Model via sEMG Sensor for Human-Robot Interaction

被引:2
|
作者
Liu, Yanhong [1 ]
Li, Xingyu [1 ]
Yang, Lei [1 ]
Yu, Hongnian [1 ,2 ]
机构
[1] Zhengzhou Univ, Sch Elect & Informat Engn, Zhengzhou 450001, Henan, Peoples R China
[2] Edinburgh Napier Univ, Built Environm, Edinburgh, Scotland
关键词
Feature fusion; hand gesture recognition; human-robot interaction; surface electromyography (sEMG) sensor; transformer; IDENTIFICATION;
D O I
10.1109/TIM.2024.3373045
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
As one of the most direct and pivotal modes of human-computer interaction (HCI), the application of surface electromyography (sEMG) signals in the domain of gesture prediction has emerged as a prominent area of research. To enhance the performance of gesture prediction system based on multichannel sEMG signals, a novel gesture prediction framework is proposed that: 1) conversion of original biological signals from multichannel sEMG into 2-D time-frequency maps is achieved through the incorporation of continuous wavelet transform (CWT) and 2) for 2-D time-frequency map inputs, a Transformer-based classification network that effectively learns local and global context information is proposed, named DIFT-Net, with the goal of implementing sEMG-based gesture prediction for robot interaction. Proposed DIFT-Net employs a dual-branch interactive fusion structure based on the Swin Transformer, enabling effective acquisition of global contextual information and local details. Additionally, an attention guidance module (AGM) and an attentional interaction module (AIM) are proposed to guide network feature extraction and fusion processes in proposed DIFT-Net. The AGM module takes intermediate features from the same stage of both branches as input and guides the network to extract more localized and detailed features through convolutional attention. Meanwhile, the AIM module integrates output features from both branches to enhance the aggregation of global context information across various scales. To substantiate the efficacy of DIFT-Net, a multichannel EMG bracelet is utilized to collect and construct an sEMG signal dataset. Experimental results demonstrate that the proposed DIFT-Net attains an accuracy of 98.36% in self-built dataset and 82.64% accuracy on the public Nanapro DB1 dataset.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Empowering human-robot interaction using sEMG sensor: Hybrid deep learning model for accurate hand gesture recognition
    Zafar, Muhammad Hamza
    Langas, Even Falkenberg
    Sanfilippo, Filippo
    [J]. RESULTS IN ENGINEERING, 2023, 20
  • [2] A Gesture Based Interface for Human-Robot Interaction
    Stefan Waldherr
    Roseli Romero
    Sebastian Thrun
    [J]. Autonomous Robots, 2000, 9 : 151 - 173
  • [3] A gesture based interface for human-robot interaction
    Waldherr, S
    Romero, R
    Thrun, S
    [J]. AUTONOMOUS ROBOTS, 2000, 9 (02) : 151 - 173
  • [4] Human-robot interaction based on gesture and movement recognition
    Li, Xing
    [J]. SIGNAL PROCESSING-IMAGE COMMUNICATION, 2020, 81
  • [5] A flexible system for gesture based human-robot interaction
    Tellaeche, Alberto
    Kildal, Johan
    Maurtua, Inaki
    [J]. 51ST CIRP CONFERENCE ON MANUFACTURING SYSTEMS, 2018, 72 : 57 - 62
  • [6] Gesture analysis for human-robot interaction
    Kim, KK
    Kwak, KC
    Chi, SY
    [J]. 8TH INTERNATIONAL CONFERENCE ON ADVANCED COMMUNICATION TECHNOLOGY, VOLS 1-3: TOWARD THE ERA OF UBIQUITOUS NETWORKS AND SOCIETIES, 2006, : U1824 - U1827
  • [7] Gesture-based human-robot interaction for human assistance in manufacturing
    Neto, Pedro
    Simao, Miguel
    Mendes, Nuno
    Safeea, Mohammad
    [J]. INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2019, 101 (1-4): : 119 - 135
  • [8] Research on multimodal human-robot interaction based on speech and gesture
    Deng Yongda
    Li Fang
    Xin Huang
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2018, 72 : 443 - 454
  • [9] Gesture-based human-robot interaction for human assistance in manufacturing
    Pedro Neto
    Miguel Simão
    Nuno Mendes
    Mohammad Safeea
    [J]. The International Journal of Advanced Manufacturing Technology, 2019, 101 : 119 - 135
  • [10] A Gesture-based Multimodal Interface for Human-Robot Interaction
    Uimonen, Mikael
    Kemppi, Paul
    Hakanen, Taru
    [J]. 2023 32ND IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, RO-MAN, 2023, : 165 - 170