Deep Visual Constraints: Neural Implicit Models for Manipulation Planning From Visual Input

被引:10
|
作者
Ha, Jung-Su [1 ]
Driess, Danny [1 ,2 ]
Toussaint, Marc [1 ]
机构
[1] TU Berlin, Learning & Intelligent Syst Lab, Berlin, Germany
[2] TU Berlin, Sci Intelligence Excellence Cluster, Berlin, Germany
关键词
Integrated planning and learning; manipulation planning; representation learning;
D O I
10.1109/LRA.2022.3194955
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Manipulation planning is the problem of finding a sequence of robot configurations that involves interactions with objects in the scene, e.g., grasping and placing an object, or more general tool-use. To achieve such interactions, traditional approaches require hand-engineering of object representations and interaction constraints, which easily becomes tedious when complex objects/interactions are considered. Inspired by recent advances in 3D modeling, e.g. NeRF, we propose a method to represent objects as continuous functions upon which constraint features are defined and jointly trained. In particular, the proposed pixel-aligned representation is directly inferred from images with known camera geometry and naturally acts as a perception component in the whole manipulation pipeline, thereby enabling long-horizon planning only from visual input.
引用
收藏
页码:10857 / 10864
页数:8
相关论文
共 50 条
  • [41] Understanding and Visualizing Deep Visual Saliency Models
    He, Sen
    Tavakoli, Hamed R.
    Borji, Ali
    Mi, Yang
    Pugeault, Nicolas
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 10198 - 10207
  • [42] ANALOGOUS NEURAL MODELS FOR TACTUAL AND VISUAL LEARNING
    MISHKIN, M
    NEUROPSYCHOLOGIA, 1979, 17 (02) : 139 - 151
  • [43] Visual Metacognition: Measures, Models, and Neural Correlates
    Rahnev, Dobromir
    AMERICAN PSYCHOLOGIST, 2021, 76 (09) : 1445 - 1453
  • [44] From Visual Understanding to Complex Object Manipulation
    Butepage, Judith
    Cruciani, Silvia
    Kokic, Mia
    Welle, Michael
    Kragic, Danica
    ANNUAL REVIEW OF CONTROL, ROBOTICS, AND AUTONOMOUS SYSTEMS, VOL 2, 2019, 2 : 161 - 179
  • [45] Multimodal Learning of Keypoint Predictive Models for Visual Object Manipulation
    Bechtle, Sarah
    Das, Neha
    Meier, Franziska
    IEEE TRANSACTIONS ON ROBOTICS, 2023, 39 (02) : 1212 - 1224
  • [46] Visual Semantic Planning using Deep Successor Representations
    Zhu, Yuke
    Gordon, Daniel
    Kolve, Eric
    Fox, Dieter
    Li Fei-Fei
    Gupta, Abhinav
    Mottaghi, Roozbeh
    Farhadi, Ali
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 483 - 492
  • [47] A neural model for the visual navigation and planning of a mobile robot
    Quoy, M
    Gaussier, P
    Leprêtre, S
    Revel, A
    ADVANCES IN ARTIFICIAL LIFE, PROCEEDINGS, 1999, 1674 : 319 - 323
  • [48] Path-Planning and Manipulation of Nanotubes Using Visual and Haptic Guidance
    Gao, Zhan
    Lecuyer, Anatole
    2009 IEEE INTERNATIONAL CONFERENCE ON VIRTUAL ENVIRONMENTS, HUMAN-COMPUTER INTERFACES AND MEASUREMENT SYSTEMS, 2009, : 1 - 5
  • [49] Convolutional Neural Network with Structural Input for Visual Object Tracking
    Fiaz, Mustansar
    Mahmood, Arif
    Jung, Soon Ki
    SAC '19: PROCEEDINGS OF THE 34TH ACM/SIGAPP SYMPOSIUM ON APPLIED COMPUTING, 2019, : 1345 - 1352
  • [50] Modular neural-visual servoing with image compression input
    Stanley, K
    Wu, QMJ
    de Silva, CW
    Gruver, WA
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2001, 10 (01) : 1 - 11