Implicit Kinematic Policies: Unifying Joint and Cartesian Action Spaces in End-to-End Robot Learning

被引:4
|
作者
Ganapathi, Aditya [1 ,2 ]
Florence, Pete [1 ]
Varley, Jake [1 ]
Burns, Kaylee [1 ,3 ]
Goldberg, Ken [2 ]
Zeng, Andy [1 ]
机构
[1] Google, Robot, Mountain View, CA 94043 USA
[2] Univ Calif Berkeley, Berkeley, CA 90095 USA
[3] Stanford Univ, Stanford, CA USA
来源
2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022) | 2022年
关键词
D O I
10.1109/ICRA46639.2022.9812165
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Action representation is an important yet often overlooked aspect in end-to-end robot learning with deep networks. Choosing one action space over another (e.g. target joint positions, or Cartesian end-effector poses) can result in surprisingly stark performance differences between various downstream tasks - and as a result, considerable research has been devoted to finding the right action space for a given application. However, in this work, we instead investigate how our models can discover and learn for themselves which action space to use. Leveraging recent work on implicit behavioral cloning, which takes both observations and actions as input, we demonstrate that it is possible to present the same action in multiple different spaces to the same policy - allowing it to learn inductive patterns from each space. Specifically, we study the benefits of combining Cartesian and joint action spaces in the context of learning manipulation skills. To this end, we present Implicit Kinematic Policies (IKP), which incorporates the kinematic chain as a differentiable module within the deep network. Quantitative experiments across several simulated continuous control tasks-from scooping piles of small objects, to lifting boxes with elbows, to precise block insertion with miscalibrated robots-suggest IKP not only learns complex prehensile and non-prehensile manipulation from pixels better than baseline alternatives, but also can learn to compensate for small joint encoder offset errors. Finally, we also run qualitative experiments on a real UR5e to demonstrate the feasibility of our algorithm on a physical robotic system with real data. See https://tinyurl.com/4wz3nf86 for code and supplementary material.
引用
收藏
页码:2656 / 2662
页数:7
相关论文
共 50 条
  • [31] Learning to predict action potentials end-to-end from calcium imaging data
    Linsley, Drew
    Linsley, Jeremy W.
    Sharma, Tarun
    Meyers, Nathan
    Serre, Thomas
    2018 52ND ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2018,
  • [32] Beamer: An End-to-End Deep Learning Framework for Unifying Data Cleaning in DNN Model Training and Inference
    Bi, Nifei
    Chen, Xiansen
    Xu, Chen
    Zhou, Aoying
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 4685 - 4689
  • [33] End-to-End Model-Free Reinforcement Learning for Urban Driving using Implicit Affordances
    Toromanoff, Marin
    Wirbel, Emilie
    Moutarde, Fabien
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 7151 - 7160
  • [34] VGAI: END-TO-END LEARNING OF VISION-BASED DECENTRALIZED CONTROLLERS FOR ROBOT SWARMS
    Hu, Ting-Kuei
    Gama, Fernando
    Chen, Tianlong
    Wang, Zhangyang
    Ribeiro, Alejandro
    Sadler, Brian M.
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4900 - 4904
  • [35] Joint Design of Radar Waveform and Detector via End-to-End Learning With Waveform Constraints
    Jiang, Wei
    Haimovich, Alexander M.
    Simeone, Osvaldo
    IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2022, 58 (01) : 552 - 567
  • [36] Coded Vision: Towards Joint Image Understanding and Compression via End-to-End Learning
    Shen, Qiu
    Cai, Juanjuan
    Liu, Linfeng
    Liu, Haojie
    Chen, Tong
    Ye, Long
    Ma, Zhan
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT I, 2018, 11164 : 3 - 14
  • [37] Audio-Driven Facial Animation by Joint End-to-End Learning of Pose and Emotion
    Karras, Tero
    Aila, Timo
    Laine, Samuli
    Herva, Antti
    Lehtinen, Jaakko
    ACM TRANSACTIONS ON GRAPHICS, 2017, 36 (04):
  • [38] Developing End-to-End Control Policies for Robotic Swarms Using Deep Q-learning
    Wei, Yufei
    Nie, Xiaotong
    Hiraga, Motoaki
    Ohkura, Kazuhiro
    Car, Zlatan
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2019, 23 (05) : 920 - 927
  • [39] End-to-end grasping policies for human-in-the-loop robots via deep reinforcement learning
    Sharif, Mohammadreza
    Erdogmus, Deniz
    Amato, Christopher
    Padir, Taskin
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 2768 - 2774
  • [40] Developing End-to-End Control Policies for Robotic Swarms Using Deep Q-learning
    Wei Y.
    Nie X.
    Hiraga M.
    Ohkura K.
    Car Z.
    Journal of Advanced Computational Intelligence and Intelligent Informatics, 2019, 23 (05): : 920 - 927