Deep Visual Constraints: Neural Implicit Models for Manipulation Planning From Visual Input

被引:10
|
作者
Ha, Jung-Su [1 ]
Driess, Danny [1 ,2 ]
Toussaint, Marc [1 ]
机构
[1] TU Berlin, Learning & Intelligent Syst Lab, Berlin, Germany
[2] TU Berlin, Sci Intelligence Excellence Cluster, Berlin, Germany
关键词
Integrated planning and learning; manipulation planning; representation learning;
D O I
10.1109/LRA.2022.3194955
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Manipulation planning is the problem of finding a sequence of robot configurations that involves interactions with objects in the scene, e.g., grasping and placing an object, or more general tool-use. To achieve such interactions, traditional approaches require hand-engineering of object representations and interaction constraints, which easily becomes tedious when complex objects/interactions are considered. Inspired by recent advances in 3D modeling, e.g. NeRF, we propose a method to represent objects as continuous functions upon which constraint features are defined and jointly trained. In particular, the proposed pixel-aligned representation is directly inferred from images with known camera geometry and naturally acts as a perception component in the whole manipulation pipeline, thereby enabling long-horizon planning only from visual input.
引用
收藏
页码:10857 / 10864
页数:8
相关论文
共 50 条
  • [31] Visual Udder Detection with Deep Neural Networks
    Katamreddy, Sukumar
    Doody, Pat
    Walsh, Joseph
    Riordan, Daniel
    2018 12TH INTERNATIONAL CONFERENCE ON SENSING TECHNOLOGY (ICST), 2018, : 166 - 171
  • [32] Visual Search With Deep Convolutional Neural Network
    Poder, Endel
    PERCEPTION, 2019, 48 : 134 - 134
  • [33] Training Deep Neural Networks for Visual Servoing
    Bateux, Quentin
    Marchand, Eric
    Leitner, Jurgen
    Chaumette, Francois
    Corke, Peter
    2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 3307 - 3314
  • [34] The Role of Architectural and Learning Constraints in Neural Network Models: A Case Study on Visual Space Coding
    Testolin, Alberto
    De Grazia, Michele De Filippo
    Zorzi, Marco
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2017, 11
  • [35] Visual servoing with deep learning and data augmentation for robotic manipulation
    Liu J.
    Li Y.
    Journal of Advanced Computational Intelligence and Intelligent Informatics, 2020, 24 (07): : 953 - 962
  • [36] CAN: Concept-Aligned Neurons for Visual Comparison of Deep Neural Network Models
    Li, M.
    Jeong, S.
    Liu, S.
    Berger, M.
    COMPUTER GRAPHICS FORUM, 2024, 43 (03)
  • [37] Visual Data Simulation for Deep Learning in Robot Manipulation Tasks
    Surak, Miroslav
    Kosnar, Karel
    Kulich, Miroslav
    Kozak, Viktor
    Peeucil, Libor
    MODELLING AND SIMULATION FOR AUTONOMOUS SYSTEMS (MESAS 2018), 2019, 11472 : 402 - 411
  • [38] Symmetric Networks with Geometric Constraints as Models of Visual Illusions
    Stewart, Ian
    Golubitsky, Martin
    SYMMETRY-BASEL, 2019, 11 (06):
  • [39] Towards explainable deep visual saliency models
    Malladi, Sai Phani Kumar
    Mukherjee, Jayanta
    Larabi, Mohamed-Chaker
    Chaudhury, Santanu
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 235
  • [40] Visual interpretability of bioimaging deep learning models
    Rotem, Oded
    Zaritsky, Assaf
    NATURE METHODS, 2024, 21 (08) : 1394 - 1397