Deep Visual Constraints: Neural Implicit Models for Manipulation Planning From Visual Input

被引:10
|
作者
Ha, Jung-Su [1 ]
Driess, Danny [1 ,2 ]
Toussaint, Marc [1 ]
机构
[1] TU Berlin, Learning & Intelligent Syst Lab, Berlin, Germany
[2] TU Berlin, Sci Intelligence Excellence Cluster, Berlin, Germany
关键词
Integrated planning and learning; manipulation planning; representation learning;
D O I
10.1109/LRA.2022.3194955
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Manipulation planning is the problem of finding a sequence of robot configurations that involves interactions with objects in the scene, e.g., grasping and placing an object, or more general tool-use. To achieve such interactions, traditional approaches require hand-engineering of object representations and interaction constraints, which easily becomes tedious when complex objects/interactions are considered. Inspired by recent advances in 3D modeling, e.g. NeRF, we propose a method to represent objects as continuous functions upon which constraint features are defined and jointly trained. In particular, the proposed pixel-aligned representation is directly inferred from images with known camera geometry and naturally acts as a perception component in the whole manipulation pipeline, thereby enabling long-horizon planning only from visual input.
引用
收藏
页码:10857 / 10864
页数:8
相关论文
共 50 条
  • [11] Learning Task Constraints in Visual-Action Planning from Demonstrations
    Esposito, Francesco
    Pek, Christian
    Welle, Michael C.
    Kragic, Danica
    2021 30TH IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2021, : 131 - 138
  • [12] Learning Robotic Manipulation through Visual Planning and Acting
    Wang, Angelina
    Kurutach, Thanard
    Liu, Kara
    Abbeel, Pieter
    Tamar, Aviv
    ROBOTICS: SCIENCE AND SYSTEMS XV, 2019,
  • [13] Visual Genealogy of Deep Neural Networks
    Wang, Qianwen
    Yuan, Jun
    Chen, Shuxin
    Su, Hang
    Qu, Huamin
    Liu, Shixia
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2020, 26 (11) : 3340 - 3352
  • [14] Visual Attention with Deep Neural Networks
    Canziani, Alfredo
    Culurciello, Eugenio
    2015 49TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2015,
  • [15] Visual perception of liquids: Insights from deep neural networks
    van Assen, Jan Jaap R.
    Nishida, Shin'ya
    Fleming, Roland W.
    PLOS COMPUTATIONAL BIOLOGY, 2020, 16 (08)
  • [16] Action-conditional implicit visual dynamics for deformable object manipulation
    Shen, Bokui
    Jiang, Zhenyu
    Choy, Christopher
    Savarese, Silvio
    Guibas, Leonidas J. J.
    Anandkumar, Anima
    Zhu, Yuke
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2024, 43 (04): : 437 - 455
  • [17] Output planning at the input stage in visual working memory
    Boettcher, Sage E. P.
    Gresch, Daniela
    Nobre, Anna C.
    van Ede, Freek
    SCIENCE ADVANCES, 2021, 7 (13):
  • [18] Object Tracking Using Deep Convolutional Neural Networks and Visual Appearance Models
    Mocanu, Bogdan
    Tapu, Ruxandra
    Zaharia, Titus
    ADVANCED CONCEPTS FOR INTELLIGENT VISION SYSTEMS (ACIVS 2017), 2017, 10617 : 114 - 125
  • [19] Factorized visual representations in the primate visual system and deep neural networks
    Lindsey, Jack W.
    Issa, Elias B.
    ELIFE, 2024, 13
  • [20] AcTiVis: Visual Exploration of Industry-Scale Deep Neural Network Models
    Kahng, Minsuk
    Andrews, Pierre Y.
    Kalro, Aditya
    Chau, Duen Horng
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2018, 24 (01) : 88 - 97