Multi-modal 3D Human Pose Estimation for Human-Robot Collaborative Applications

被引:1
|
作者
Peppas, Konstantinos [1 ]
Tsiolis, Konstantinos [1 ]
Mariolis, Ioannis [1 ]
Topalidou-Kyniazopoulou, Angeliki [1 ]
Tzovaras, Dimitrios [1 ]
机构
[1] Ctr Res & Technol Hellas CERTH, Inst Informat Technol, 6th Km Charilaou Thermi Rd, Thessaloniki, Greece
关键词
Multi-modal learning; 3D human pose estimation; Collaborative tasks; Deep learning; CNN;
D O I
10.1007/978-3-030-73973-7_34
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a multi-modal 3D human pose estimation approach which combines a 2D human pose estimation network utilizing RGB data with a 3D human pose estimation network utilizing the 2D pose estimation results and depth information, in order to predict 3D human poses. We improve upon the state-of-the-art by proposing the use of a more accurate 2D human pose estimation network, as well as by introducing squeeze-excite blocks into the architecture of the 3D pose estimation network. More importantly, we focused on the challenging application of 3D human pose estimation during collaborative tasks. In that direction, we selected appropriate sub-sets that address collaborative tasks from a large-scale multi-view RGB-D dataset and generated a novel one-view RGB-D dataset for training and testing respectively. We achieved above state-of-the-art performance among RGB-D approaches when tested on a novel benchmark RGB-D dataset on collaborative assembly that we have created and made publicly available.
引用
下载
收藏
页码:355 / 364
页数:10
相关论文
共 50 条
  • [1] Multi-Modal Interfaces for Human-Robot Communication in Collaborative Assembly
    Horvath, Gergely
    Kardos, Csaba
    Kemeny, Zsolt
    Kovacs, Andras
    Pataki, Balazs E.
    Vancza, Jozsef
    ERCIM NEWS, 2018, (114): : 15 - 16
  • [2] Weakly Supervised Multi-Modal 3D Human Body Pose Estimation for Autonomous Driving
    Bauer, Peter
    Bouazizi, Arij
    Kressel, Ulrich
    Flohr, Fabian B.
    2023 IEEE INTELLIGENT VEHICLES SYMPOSIUM, IV, 2023,
  • [3] Multi-modal anchoring for human-robot interaction
    Fritsch, J
    Kleinehagenbrock, M
    Lang, S
    Plötz, T
    Fink, GA
    Sagerer, G
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2003, 43 (2-3) : 133 - 147
  • [4] Assessment of the Human-Robot Collaborative Polishing Task by Using EMG Sensors and 3D Pose Estimation
    Petrovic, Milos
    Vukicevic, Arso M.
    Lukic, Branko
    Jovanovic, Kosta
    ADVANCES IN SERVICE AND INDUSTRIAL ROBOTICS, RAAD 2022, 2022, 120 : 564 - 570
  • [5] Multi-modal 3D Human Pose Estimation with 2DWeak Supervision in Autonomous Driving
    Zheng, Jingxiao
    Shi, Xinwei
    Gorban, Alexander
    Mao, Junhua
    Song, Yang
    Qi, Charles R.
    Liu, Ting
    Chari, Visesh
    Cornman, Andre
    Zhou, Yin
    Li, Congcong
    Anguelov, Dragomir
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 4477 - 4486
  • [6] Multi-modal interfaces for natural Human-Robot Interaction
    Andronas, Dionisis
    Apostolopoulos, George
    Fourtakas, Nikos
    Makris, Sotiris
    10TH CIRP SPONSORED CONFERENCE ON DIGITAL ENTERPRISE TECHNOLOGIES (DET 2020) - DIGITAL TECHNOLOGIES AS ENABLERS OF INDUSTRIAL COMPETITIVENESS AND SUSTAINABILITY, 2021, 54 : 197 - 202
  • [7] Multi-modal Language Models for Human-Robot Interaction
    Janssens, Ruben
    COMPANION OF THE 2024 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, HRI 2024 COMPANION, 2024, : 109 - 111
  • [8] Human-Robot Interaction with Multi-Human Social Pattern Inference on a Multi-Modal Robot
    Tseng, Shih-Huan
    Wu, Tung-Yen
    Cheng, Ching-Ying
    Fu, Li-Chen
    2014 14TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2014), 2014, : 819 - 824
  • [9] Regression-Based 3D Hand Pose Estimation for Human-Robot Interaction
    Bandi, Chaitanya
    Thomas, Ulrike
    COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS, VISIGRAPP 2020, 2022, 1474 : 507 - 529
  • [10] Multi-View Human Pose Estimation in Human-Robot Interaction
    Xu, Chengjun
    Yu, Xinyi
    Wang, Zhengan
    Ou, Linlin
    IECON 2020: THE 46TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2020, : 4769 - 4775