Multi-modal 3D Human Pose Estimation for Human-Robot Collaborative Applications

被引:1
|
作者
Peppas, Konstantinos [1 ]
Tsiolis, Konstantinos [1 ]
Mariolis, Ioannis [1 ]
Topalidou-Kyniazopoulou, Angeliki [1 ]
Tzovaras, Dimitrios [1 ]
机构
[1] Ctr Res & Technol Hellas CERTH, Inst Informat Technol, 6th Km Charilaou Thermi Rd, Thessaloniki, Greece
关键词
Multi-modal learning; 3D human pose estimation; Collaborative tasks; Deep learning; CNN;
D O I
10.1007/978-3-030-73973-7_34
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a multi-modal 3D human pose estimation approach which combines a 2D human pose estimation network utilizing RGB data with a 3D human pose estimation network utilizing the 2D pose estimation results and depth information, in order to predict 3D human poses. We improve upon the state-of-the-art by proposing the use of a more accurate 2D human pose estimation network, as well as by introducing squeeze-excite blocks into the architecture of the 3D pose estimation network. More importantly, we focused on the challenging application of 3D human pose estimation during collaborative tasks. In that direction, we selected appropriate sub-sets that address collaborative tasks from a large-scale multi-view RGB-D dataset and generated a novel one-view RGB-D dataset for training and testing respectively. We achieved above state-of-the-art performance among RGB-D approaches when tested on a novel benchmark RGB-D dataset on collaborative assembly that we have created and made publicly available.
引用
收藏
页码:355 / 364
页数:10
相关论文
共 50 条
  • [31] Multi-modal referring expressions in human-human task descriptions and their implications for human-robot interaction
    Gross, Stephanie
    Krenn, Brigitte
    Scheutz, Matthias
    INTERACTION STUDIES, 2016, 17 (02) : 180 - 210
  • [32] Designing and Implementing a Platform for Collecting Multi-Modal Data of Human-Robot Interaction
    Vaughan, Brian
    Han, Jing Guang
    Gilmartin, Emer
    Campbell, Nick
    ACTA POLYTECHNICA HUNGARICA, 2012, 9 (01) : 7 - 17
  • [33] A Multi-party Multi-modal Dataset for Focus of Visual Attention in Human-human and Human-robot Interaction
    Stefanov, Kalin
    Beskow, Jonas
    LREC 2016 - TENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2016, : 4440 - 4444
  • [34] A Probabilistic Approach for Attention-Based Multi-Modal Human-Robot Interaction
    Begum, Momotaz
    Karray, Fakhri
    Mann, George K. I.
    Gosine, Raymond
    RO-MAN 2009: THE 18TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, VOLS 1 AND 2, 2009, : 909 - +
  • [35] Investigation of multi-modal interface features for adaptive automation of a human-robot system
    Kaber, DB
    Wright, MC
    Sheik-Nainar, MA
    INTERNATIONAL JOURNAL OF HUMAN-COMPUTER STUDIES, 2006, 64 (06) : 527 - 540
  • [36] Editorial: Integrated Multi-modal and Sensorimotor Coordination for Enhanced Human-Robot Interaction
    Fang, Bin
    Fang, Cheng
    Wen, Li
    Manoonpong, Poramate
    FRONTIERS IN NEUROROBOTICS, 2021, 15
  • [37] Editorial: Integrated Multi-modal and Sensorimotor Coordination for Enhanced Human-Robot Interaction
    Fang, Bin
    Fang, Cheng
    Wen, Li
    Manoonpong, Poramate
    Fang, Bin (fangbin@tsinghua.edu.cn), 1600, Frontiers Media S.A. (15):
  • [38] Navigating to Success in Multi-Modal Human-Robot Collaboration: Analysis and Corpus Release
    Lukin, Stephanie M.
    Pollard, Kimberly A.
    Bonial, Claire
    Hudson, Taylor
    Artstein, Ron
    Voss, Clare
    Traum, David
    2023 32ND IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, RO-MAN, 2023, : 1859 - 1865
  • [39] Towards Multi-Modal Intention Interfaces for Human-Robot Co-Manipulation
    Peternel, Luka
    Tsagarakis, Nikos
    Ajoudani, Arash
    2016 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2016), 2016, : 2663 - 2669
  • [40] Multi-modal AI Systems for Human and Animal Pose Estimation in Challenging Conditions
    Deng, Qianyi
    2023 IEEE INTERNATIONAL CONFERENCE ON SMART COMPUTING, SMARTCOMP, 2023, : 239 - 240