MANUS: Markerless Grasp Capture using Articulated 3D Gaussians

被引:0
|
作者
Pokhariya, Chandradeep [1 ,2 ]
Shah, Ishaan Nikhil [1 ]
Xing, Angela [2 ]
Li, Zekun [2 ]
Chen, Kefan [2 ]
Sharma, Avinash [1 ]
Sridhar, Srinath [2 ]
机构
[1] IIIT Hyderabad, Hyderabad, India
[2] Brown Univ, Providence, RI USA
关键词
D O I
10.1109/CVPR52733.2024.00214
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Understanding how we grasp objects with our hands has important applications in areas like robotics and mixed reality. However, this challenging problem requires accurate modeling of the contact between hands and objects. To capture grasps, existing methods use skeletons, meshes, or parametric models that does not represent hand shape accurately resulting in inaccurate contacts. We present MANUS, a method for Markerless Hand-Object Grasp Capture using Articulated 3D Gaussians. We build a novel articulated 3D Gaussians representation that extends 3D Gaussian splatting [29] for high-fidelity representation of articulating hands. Since our representation uses Gaussian primitives optimized from the multi-view pixel-aligned losses, it enables us to efficiently and accurately estimate contacts between the hand and the object. For the most accurate results, our method requires tens of camera views that current datasets do not provide. We therefore build MANUS-Grasps, a new dataset that contains hand-object grasps viewed from 50+ cameras across 30+ scenes, 3 subjects, and comprising over 7M frames. In addition to extensive qualitative results, we also show that our method outperforms others on a quantitative contact evaluation method that uses paint transfer from the object to the hand.
引用
收藏
页码:2197 / 2208
页数:12
相关论文
共 50 条
  • [31] Markerless motion capture can provide reliable 3D gait kinematics in the sagittal and frontal plane
    Sandau, Martin
    Koblauch, Henrik
    Moeslund, Thomas B.
    Aanaes, Henrik
    Alkjaer, Tine
    Simonsen, Erik B.
    MEDICAL ENGINEERING & PHYSICS, 2014, 36 (09) : 1168 - 1175
  • [32] Animatable 3D Gaussians for modeling dynamic humans
    Xu, Yukun
    Ye, Keyang
    Shao, Tianjia
    Weng, Yanlin
    FRONTIERS OF COMPUTER SCIENCE, 2025, 19 (09)
  • [33] Using DeepLabCut for 3D markerless pose estimation across species and behaviors
    Nath, Tanmay
    Mathis, Alexander
    Chen, An Chi
    Patel, Amir
    Bethge, Matthias
    Mathis, Mackenzie Weygandt
    NATURE PROTOCOLS, 2019, 14 (07) : 2152 - 2176
  • [34] Using DeepLabCut for 3D markerless pose estimation across species and behaviors
    Tanmay Nath
    Alexander Mathis
    An Chi Chen
    Amir Patel
    Matthias Bethge
    Mackenzie Weygandt Mathis
    Nature Protocols, 2019, 14 : 2152 - 2176
  • [35] Retrieving Articulated 3D Objects Using Normalized Distance Function
    Mohamed, Waleed
    Ben Hamza, A.
    ARTICULATED MOTION AND DEFORMABLE OBJECTS, PROCEEDINGS, 2010, 6169 : 21 - 30
  • [36] 3D Surgical Overlay with Markerless Image Registration Using a Single Camera
    Wang, Junchen
    Suenaga, Hideyuki
    Yang, Liangjing
    Liao, Hongen
    Ando, Takehiro
    Kobayashi, Etsuko
    Sakuma, Ichiro
    AUGMENTED ENVIRONMENTS FOR COMPUTER-ASSISTED INTERVENTIONS, AE-CAI 2015, 2015, 9365 : 124 - 133
  • [37] Using Geometry to Detect Grasp Poses in 3D Point Clouds
    ten Pas, Andreas
    Platt, Robert
    ROBOTICS RESEARCH, VOL 1, 2018, 2 : 307 - 324
  • [38] Markerless 3D kinematics and force estimation in cheetahs
    da Silva, Zico
    Shield, Stacey
    Hudson, Penny E.
    Wilson, Alan M.
    Nicolls, Fred
    Patel, Amir
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [39] Matching 2D and 3D articulated shapes using the eccentricity transform
    Ion, Adrian
    Artner, Nicole M.
    Peyre, Gabriel
    Kropatsch, Walter G.
    Cohen, Laurent D.
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2011, 115 (06) : 817 - 834
  • [40] Align Your Gaussians: Text-to-4D with Dynamic 3D Gaussians and Composed Diffusion Models
    Ling, Huan
    Kim, Seung Wook
    Torralba, Antonio
    Fidler, Sanja
    Kreis, Karsten
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 8576 - 8588