Multi-modal 3D Human Pose Estimation for Human-Robot Collaborative Applications

被引:1
|
作者
Peppas, Konstantinos [1 ]
Tsiolis, Konstantinos [1 ]
Mariolis, Ioannis [1 ]
Topalidou-Kyniazopoulou, Angeliki [1 ]
Tzovaras, Dimitrios [1 ]
机构
[1] Ctr Res & Technol Hellas CERTH, Inst Informat Technol, 6th Km Charilaou Thermi Rd, Thessaloniki, Greece
关键词
Multi-modal learning; 3D human pose estimation; Collaborative tasks; Deep learning; CNN;
D O I
10.1007/978-3-030-73973-7_34
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a multi-modal 3D human pose estimation approach which combines a 2D human pose estimation network utilizing RGB data with a 3D human pose estimation network utilizing the 2D pose estimation results and depth information, in order to predict 3D human poses. We improve upon the state-of-the-art by proposing the use of a more accurate 2D human pose estimation network, as well as by introducing squeeze-excite blocks into the architecture of the 3D pose estimation network. More importantly, we focused on the challenging application of 3D human pose estimation during collaborative tasks. In that direction, we selected appropriate sub-sets that address collaborative tasks from a large-scale multi-view RGB-D dataset and generated a novel one-view RGB-D dataset for training and testing respectively. We achieved above state-of-the-art performance among RGB-D approaches when tested on a novel benchmark RGB-D dataset on collaborative assembly that we have created and made publicly available.
引用
收藏
页码:355 / 364
页数:10
相关论文
共 50 条
  • [21] Multi-modal Proactive Approaching of Humans for Human-Robot Cooperative Tasks
    Naik, Lakshadeep
    Palinko, Oskar
    Bodenhagen, Leon
    Krueger, Norbert
    2021 30TH IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2021, : 323 - 329
  • [22] Bidirectional Multi-modal Signs of Checking Human-Robot Engagement and Interaction
    Umberto Maniscalco
    Pietro Storniolo
    Antonio Messina
    International Journal of Social Robotics, 2022, 14 : 1295 - 1309
  • [23] A Multi-modal Gesture Recognition System in a Human-Robot Interaction Scenario
    Li, Zhi
    Jarvis, Ray
    2009 IEEE INTERNATIONAL WORKSHOP ON ROBOTIC AND SENSORS ENVIRONMENTS (ROSE 2009), 2009, : 41 - 46
  • [24] Are You Sure? - Multi-Modal Human Decision Uncertainty Detection in Human-Robot Interaction
    Scherf, Lisa
    Gasche, Lisa Alina
    Chemangui, Eya
    Koert, Dorothea
    PROCEEDINGS OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024, 2024, : 621 - 629
  • [25] A Multi-modal Sensor Array for Safe Human-Robot Interaction and Mapping
    Abah, Colette
    Orekhov, Andrew L.
    Johnston, Garrison L. H.
    Yin, Peng
    Choset, Howie
    Simaan, Nabil
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 3768 - 3774
  • [26] Bidirectional Multi-modal Signs of Checking Human-Robot Engagement and Interaction
    Maniscalco, Umberto
    Storniolo, Pietro
    Messina, Antonio
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2022, 14 (05) : 1295 - 1309
  • [27] Intuitive Multi-modal Human-Robot Interaction via Posture and Voice
    Lai, Yuzhi
    Radke, Mario
    Nassar, Youssef
    Gopal, Atmaraaj
    Weber, Thomas
    Liu, ZhaoHua
    Zhang, Yihong
    Raetsch, Matthias
    ROBOTICS, COMPUTER VISION AND INTELLIGENT SYSTEMS, ROBOVIS 2024, 2024, 2077 : 441 - 456
  • [28] Dynamic Hand Gesture Recognition Based on 3D Hand Pose Estimation for Human-Robot Interaction
    Gao, Qing
    Chen, Yongquan
    Ju, Zhaojie
    Liang, Yi
    IEEE SENSORS JOURNAL, 2022, 22 (18) : 17421 - 17430
  • [29] A multi-modal and collaborative human-machine interface for a walking robot
    Estremera, J
    Garcia, E
    de Santos, PG
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2002, 35 (04) : 397 - 425
  • [30] Multi-modal human-robot interface for interaction with a remotely operating mobile service robot
    Fischer, C
    Schmidt, G
    ADVANCED ROBOTICS, 1998, 12 (04) : 397 - 409