D-SPDH: Improving 3D Robot Pose Estimation in Sim2Real Scenario via Depth Data

被引:0
|
作者
Simoni, Alessandro [1 ]
Borghi, Guido [2 ]
Garattoni, Lorenzo [3 ]
Francesca, Gianpiero [3 ]
Vezzani, Roberto [1 ]
机构
[1] Univ Modena & Reggio Emilia, Dept Engn Enzo Ferrari, I-42122 Reggio Emilia, Italy
[2] Univ Modena & Reggio Emilia, Dipartimento Educ & Sci Umane, I-42122 Reggio Emilia, Italy
[3] Toyota Motor Europe, B-1130 Brussels, Belgium
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Robots; Three-dimensional displays; Robot kinematics; Cameras; Robot vision systems; Pose estimation; Synthetic data; Training; Rendering (computer graphics); Deep learning; Human-machine systems; Computer vision; Human-machine interaction; human-robot interaction; collaborative robots (Cobots); robot pose estimation; deep learning; computer vision; depth maps;
D O I
10.1109/ACCESS.2024.3492812
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, there has been a notable surge in the significance attributed to technologies facilitating secure and efficient cohabitation and collaboration between humans and machines, with a particular interest in robotic systems. A pivotal element in actualizing this novel and challenging collaborative paradigm involves different technical tasks, including the comprehension of 3D poses exhibited by both humans and robots through the utilization of non-intrusive systems, such as cameras. In this scenario, the availability of vision-based systems capable of detecting in real-time the robot's pose is needed as a first step towards a safe and effective interaction to, for instance, avoid collisions. Therefore, in this work, we propose a vision-based system, referred to as D-SPDH, able to estimate the 3D robot pose. The system is based on double-branch architecture and depth data as a single input; any additional information regarding the state of the internal encoders of the robot is not required. The working scenario is the Sim2Real, i.e., the system is trained only with synthetic data and then tested on real sequences, thus eliminating the time-consuming acquisition and annotation procedures of real data, common phases in deep learning algorithms. Moreover, we introduce SimBa++, a dataset featuring both synthetic and real sequences with new real-world double-arm movements, and that represents a challenging setting in which the proposed approach is tested. Experimental results show that our D-SPDH method achieves state-of-the-art and real-time performance, paving the way a possible future non-invasive systems to monitor human-robot interactions.
引用
收藏
页码:166660 / 166673
页数:14
相关论文
共 50 条
  • [1] Sim2real transfer learning for 3D human pose estimation: motion to the rescue
    Doersch, Carl
    Zisserman, Andrew
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Sim2Real Instance-Level Style Transfer for 6D Pose Estimation
    Ikeda, Takuya
    Tanishige, Suomi
    Amma, Ayako
    Sudano, Michael
    Audren, Herv Prime E.
    Nishiwaki, Koichi
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 3225 - 3232
  • [3] Adapted human pose: monocular 3D human pose estimation with zero real 3D pose data
    Liu, Shuangjun
    Sehgal, Naveen
    Ostadabbas, Sarah
    APPLIED INTELLIGENCE, 2022, 52 (12) : 14491 - 14506
  • [4] Adapted human pose: monocular 3D human pose estimation with zero real 3D pose data
    Shuangjun Liu
    Naveen Sehgal
    Sarah Ostadabbas
    Applied Intelligence, 2022, 52 : 14491 - 14506
  • [5] Improving 3D Human Pose Estimation via 3D Part Affinity Fields
    Liu, Ding
    Zhao, Zixu
    Wang, Xinchao
    Hu, Yuxiao
    Zhang, Lei
    Huang, Thomas S.
    2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, : 1004 - 1013
  • [6] 3D Human Pose Estimation Using Egocentric Depth Data
    Baek, Seongmin
    Gil, Youn-Hee
    Kim, Yejin
    30TH ACM SYMPOSIUM ON VIRTUAL REALITY SOFTWARE AND TECHNOLOGY, VRST 2024, 2024,
  • [7] Monocular 3D Pose Estimation via Pose Grammar and Data Augmentation
    Xu, Yuanlu
    Wang, Wenguan
    Liu, Tengyu
    Liu, Xiaobai
    Xie, Jianwen
    Zhu, Song-Chun
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (10) : 6327 - 6344
  • [8] Part-Guided 3D RL for Sim2Real Articulated Object Manipulation
    Xie, Pengwei
    Chen, Rui
    Chen, Siang
    Qin, Yuzhe
    Xiang, Fanbo
    Sun, Tianyu
    Xu, Jing
    Wang, Guijin
    Su, Hao
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (11) : 7178 - 7185
  • [9] TGF-Net: Sim2Real Transparent Object 6D Pose Estimation Based on Geometric Fusion
    Yu, Haixin
    Li, Shoujie
    Liu, Houde
    Xia, Chongkun
    Ding, Wenbo
    Liang, Bin
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (06): : 3867 - 3874
  • [10] 3D Human Pose Estimation via Explicit Compositional Depth Maps
    Wu, Haiping
    Xiao, Bin
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12378 - 12385