YOLO-6D-Pose: Enhancing YOLO for Single-Stage Monocular Multi-Object 6D Pose Estimation

被引:0
|
作者
Maji, Debapriya [1 ]
Nagori, Soyeb [1 ]
Mathew, Manu [1 ]
Poddar, Deepak [1 ]
机构
[1] Texas Instruments Inc, Bangalore, India
关键词
D O I
10.1109/3DV62453.2024.00160
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Directly regressing 6 degrees of freedom for all the objects from a single RGB image is not well explored. Even end-to-end pose estimation approaches for a single object are inferior compared to state-of-the-art methods in terms of accuracy. Most 6D pose estimation frameworks are multi-stage relying on off-the-shelf deep networks for object and keypoint detection to establish correspondences between 3D object keypoints and 2D image locations. This is followed by applying a variant of a RANSAC-based Perspective-n-Point (PnP) followed by complex refinement operation. In this work, we propose a multi-object 6D pose estimation framework by enhancing the popular YOLOX object detector. The network is end-to-end trainable and detects each object along with its pose from a single RGB image without any additional post-processing. We show that by properly parameterizing the 6D pose and carefully designing the loss function, we can achieve state-of-theart accuracy without further refinement or any intermediate representations. YOLO-6D-Pose achieves SOTA results on YCBV and LMO dataset, surpassing all existing monocular approaches. We systematically analyze various 6D augmentations to verify their correctness and propose a new translation augmentation for this task. The network does not rely on any correspondences and is independent of the CAD model during inference. Code is available at https:// github. com/ TexasInstruments/ edgeai-yolox.
引用
收藏
页码:1616 / 1625
页数:10
相关论文
共 50 条
  • [1] Single-Stage 6D Object Pose Estimation
    Hu, Yinlin
    Fua, Pascal
    Wang, Wei
    Salzmann, Mathieu
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 2927 - 2936
  • [2] PoET: Pose Estimation Transformer for Single-View, Multi-Object 6D Pose Estimation
    Jantos, Thomas
    Hamdad, Mohamed Amin
    Granig, Wolfgang
    Weiss, Stephan
    Steinbrener, Jan
    CONFERENCE ON ROBOT LEARNING, VOL 205, 2022, 205 : 1060 - 1070
  • [3] Coupled Iterative Refinement for 6D Multi-Object Pose Estimation
    Lipson, Lahav
    Teed, Zachary
    Goyal, Ankit
    Deng, Jia
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 6718 - 6727
  • [4] YOLO-6D+: Single Shot 6D Pose Estimation Using Privileged Silhouette Information
    Kang, Jia
    Liu, Wenjun
    Tu, Wenzhe
    Yang, Lu
    2020 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND ROBOTICS (ICIPROB 2020, 2020,
  • [5] Single Shot 6D Object Pose Estimation
    Kleeberger, Kilian
    Huber, Marco F.
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 6239 - 6245
  • [6] On Evaluation of 6D Object Pose Estimation
    Hodan, Tomas
    Matas, Jiri
    Obdrzalek, Stephan
    COMPUTER VISION - ECCV 2016 WORKSHOPS, PT III, 2016, 9915 : 606 - 619
  • [7] YOLO-Pose: Enhancing YOLO for Multi Person Pose Estimation Using Object Keypoint Similarity Loss
    Maji, Debapriya
    Nagori, Soyeb
    Mathew, Manu
    Poddar, Deepak
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 2636 - 2645
  • [8] Joint Optimization of the 3D Model and 6D Pose for Monocular Pose Estimation
    Guo, Liangchao
    Chen, Lin
    Wang, Qiufu
    Zhang, Zhuo
    Sun, Xiaoliang
    DRONES, 2024, 8 (11)
  • [9] Estimation method of category-level multi-object rigid body 6D pose
    Cheng, Shuo
    Jia, Di
    Yang, Liu
    He, Dekun
    CHINESE JOURNAL OF LIQUID CRYSTALS AND DISPLAYS, 2025, 40 (03) : 457 - 471
  • [10] Active 6D Multi-Object Pose Estimation in Cluttered Scenarios with Deep Reinforcement Learning
    Sock, Juil
    Garcia-Hernando, Guillermo
    Kim, Tae-Kyun
    2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 10564 - 10571