A Dynamic Head Gesture Recognition Method for Real-time Intention Inference and Its Application to Visual Human-robot Interaction

被引:1
|
作者
Xie, Jialong [1 ]
Zhang, Botao [1 ]
Lu, Qiang [1 ]
Borisov, Oleg [2 ]
机构
[1] Hangzhou Dianzi Univ, Sch Automat, Hangzhou, Peoples R China
[2] ITMO Univ, Fac Control Syst & Robot, St Petersburg, Russia
基金
中国国家自然科学基金;
关键词
Computer vision; deep learning; head gesture; human-robot interaction; MOTION;
D O I
10.1007/s12555-022-0051-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Head gesture is a natural and non-verbal communication method for human-computer and human-robot interaction, conveying attitudes and intentions. However, the existing vision-based recognition methods cannot meet the precision and robustness of interaction requirements. Due to the limited computational resources, applying most high-accuracy methods to mobile and onboard devices is challenging. Moreover, the wearable device-based approach is inconvenient and expensive. To deal with these problems, an end-to-end two-stream fusion network named TSIR3D is proposed to identify head gestures from videos for analyzing human attitudes and intentions. Inspired by Inception and ResNet architecture, the width and depth of the network are increased to capture motion features sufficiently. Meanwhile, convolutional kernels are expanded from the spatial domain to the spatiotemporal domain for temporal feature extraction. The fusion position of the two-stream channel is explored under an accuracy/complexity trade-off to a certain extent. Furthermore, a dynamic head gesture dataset named DHG and a behavior tree are designed for human-robot interaction. Experimental results show that the proposed method has advantages in real-time performance on the remote server or the onboard computer. Furthermore, its accuracy on the DHG can surpass most state-of-the-art vision-based methods and is even better than most previous approaches based on head-mounted sensors. Finally, TSIR3D is applied on Pepper Robot equipped with Jetson TX2.
引用
收藏
页码:252 / 264
页数:13
相关论文
共 50 条
  • [21] The Influence of Speed and Position in Dynamic Gesture Recognition for Human-Robot Interaction
    Carlos Castillo, Jose
    Alonso-Martin, Fernando
    Caceres-Dominguez, David
    Malfaz, Maria
    Salichs, Miguel A.
    JOURNAL OF SENSORS, 2019, 2019
  • [22] An Approach to Dynamic Gesture Recognition for Real-Time Interaction
    Zhao, Jinli
    Chen, Tianding
    SIXTH INTERNATIONAL SYMPOSIUM ON NEURAL NETWORKS (ISNN 2009), 2009, 56 : 369 - 377
  • [23] Human-robot collaborative handling of curtain walls using dynamic motion primitives and real-time human intention recognition
    Li, Fengming
    Sun, Huayan
    Liu, Enguang
    Du, Fuxin
    BIOMIMETIC INTELLIGENCE AND ROBOTICS, 2024, 4 (04):
  • [24] Face and gesture recognition using subspace method for human-robot interaction
    Hasanuzzaman, M
    Zhang, T
    Ampornaramveth, V
    Bhuiyan, MA
    Shirai, Y
    Ueno, H
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2004, PT 1, PROCEEDINGS, 2004, 3331 : 369 - 376
  • [25] Real-time human-robot interaction underlying neurorobotic trust and intent recognition
    Bray, Laurence C. Jayet
    Anumandla, Sridhar R.
    Thibeault, Corey M.
    Hoang, Roger V.
    Goodman, Philip H.
    Dascalu, Sergiu M.
    Bryant, Bobby D.
    Harris, Frederick C., Jr.
    NEURAL NETWORKS, 2012, 32 : 130 - 137
  • [26] Attention-Oriented Action Recognition for Real-Time Human-Robot Interaction
    Song, Ziyang
    Yin, Ziyi
    Yuan, Zejian
    Zhang, Chong
    Chi, Wanchao
    Ling, Yonggen
    Zhang, Shenghao
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7087 - 7094
  • [27] Advancements in Visual Gesture Recognition for Underwater Human-Robot Interaction: A Comprehensive Review
    Hozyn, Stanislaw
    IEEE ACCESS, 2024, 12 : 163131 - 163142
  • [28] A Human-Robot Interaction for a Mecanum Wheeled Mobile Robot with Real-Time 3D Two-Hand Gesture Recognition
    Luo, Xueling
    Amighetti, Andrea
    Zhang, Dan
    2019 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, AUTOMATION AND CONTROL TECHNOLOGIES (AIACT 2019), 2019, 1267
  • [29] A real-time human-robot interaction framework with robust background invariant hand gesture detection
    Mazhar, Osama
    Navarro, Benjamin
    Ramdani, Sofiane
    Passama, Robin
    Cherubini, Andrea
    ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2019, 60 : 34 - 48
  • [30] A descriptive behavior intention inference framework using spatio-temporal semantic features for human-robot real-time interaction
    Wang, Liangliang
    Huo, Guanglei
    Li, Ruifeng
    Liang, Peidong
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 128