A Learned Stereo Depth System for Robotic Manipulation in Homes

被引:6
|
作者
Shankar, Krishna [1 ]
Tjersland, Mark [1 ]
Ma, Jeremy [1 ]
Stone, Kevin [1 ]
Bajracharya, Max [1 ]
机构
[1] Toyota Res Inst, Los Altos, CA 94022 USA
关键词
RGB-D perception; deep learning for visual perception; perception for grasping and manipulation;
D O I
10.1109/LRA.2022.3143895
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
We present a passive stereo depth system that produces dense and accurate point clouds optimized for human environments, including dark, textureless, thin, reflective and specular surfaces and objects, at 2560 x 2048 resolution, with 384 disparities, in 30 ms. The system consists of an algorithm combining learned stereo matching with engineered filtering, a training and data-mixing methodology, and a sensor hardware design. Our architecture is 15x faster than approaches that perform similarly on the Middlebury and Flying Things Stereo Benchmarks. To effectively supervise the training of this model, we combine real data labelled using off-the-shelf depth sensors, as well as a number of different rendered, simulated labeled datasets. We demonstrate the efficacy of our system by presenting a large number of qualitative results in the form of depth maps and point-clouds, experiments validating the metric accuracy of our system and comparisons to other sensors on challenging objects and scenes. We also show the competitiveness of our algorithm compared to state-of-the-art learned models using the Middlebury and FlyingThings datasets.
引用
收藏
页码:2305 / 2312
页数:8
相关论文
共 50 条
  • [1] Parareal with a learned coarse model for robotic manipulation
    Agboh, Wisdom
    Grainger, Oliver
    Ruprecht, Daniel
    Dogar, Mehmet
    [J]. COMPUTING AND VISUALIZATION IN SCIENCE, 2020, 23 (1-4)
  • [2] Hybrid image plane/stereo (HIPS) manipulation for robotic space applications
    Matthew L. Robinson
    Eric T. Baumgartner
    Kevin M. Nickels
    Todd E. Litwin
    [J]. Autonomous Robots, 2007, 23 : 83 - 96
  • [3] Hybrid image plane/stereo (HIPS) manipulation for robotic space applications
    Robinson, Matthew L.
    Baumgartner, Eric T.
    Nickels, Kevin M.
    Litwin, Todd E.
    [J]. AUTONOMOUS ROBOTS, 2007, 23 (02) : 83 - 96
  • [4] CodedStereo: Learned Phase Masks for Large Depth-of-field Stereo
    Tan, Shiyu
    Wu, Yicheng
    Yu, Shoou-, I
    Veeraraghavan, Ashok
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 7166 - 7175
  • [5] A novel computational model of stereo depth estimation for robotic vision systems
    Chen, K. -Y.
    Chen, Ch. -H.
    Chien, Ch. -Ch.
    [J]. SCIENTIA IRANICA, 2015, 22 (06) : 2188 - 2197
  • [6] A VISION SYSTEM FOR ROBOTIC INSPECTION AND MANIPULATION
    TRIVEDI, MM
    CHEN, CX
    MARAPANE, SB
    [J]. COMPUTER, 1989, 22 (06) : 91 - 97
  • [7] A TACTILE SENSING SYSTEM FOR ROBOTIC MANIPULATION
    SPEETER, TH
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 1990, 9 (06): : 25 - 36
  • [8] Parallel microassembly with a robotic manipulation system
    Chu, Henry K.
    Mills, James K.
    Cleghorn, William L.
    [J]. JOURNAL OF MICROMECHANICS AND MICROENGINEERING, 2010, 20 (12)
  • [9] Intelligent stereo vision system for robotic applications
    Tél, Ferenc
    Lantos, Béla
    [J]. Periodica Polytechnica Electrical Engineering, 1999, 43 (03): : 153 - 162
  • [10] A Practical Stereo Depth System for Smart Glasses
    Wang, Jialiang
    Scharstein, Daniel
    Bapat, Akash
    Blackburn-Matzen, Kevin
    Yu, Matthew
    Lehman, Jonathan
    Alsisan, Suhib
    Wang, Yanghan
    Tsai, Sam
    Frahm, Jan-Michael
    He, Zijian
    Vajda, Peter
    Cohen, Michael F.
    Uyttendaele, Matt
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 21498 - 21507