Optimistic Reinforcement Learning-Based Skill Insertions for Task and Motion Planning

被引:1
|
作者
Liu, Gaoyuan [1 ,2 ]
de Winter, Joris [1 ]
Durodie, Yuri [1 ,2 ]
Steckelmacher, Denis [3 ]
Nowe, Ann [3 ]
Vanderborght, Bram [1 ,2 ]
机构
[1] Vrije Univ Brussel, Brubot, B-1050 Brussels, Belgium
[2] IMEC, B-3001 Leuven, Belgium
[3] Vrije Univ Brussel, Artificial Intelligence AI Lab, B-1050 Brussels, Belgium
来源
关键词
Manipulation planning; reinforcement learning; task and motion planning; SAMPLING-BASED METHODS;
D O I
10.1109/LRA.2024.3398402
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Task and motion planning (TAMP) for robotics manipulation necessitates long-horizon reasoning involving versatile actions and skills. While deterministic actions can be crafted by sampling or optimizing with certain constraints, planning actions with uncertainty, i.e., probabilistic actions, remains a challenge for TAMP. On the contrary, Reinforcement Learning (RL) excels in acquiring versatile, yet short-horizon, manipulation skills that are robust with uncertainties. In this letter, we design a method that integrates RL skills into TAMP pipelines. Besides the policy, a RL skill is defined with data-driven logical components that enable the skill to be deployed by symbolic planning. A plan refinement sub-routine is designed to further tackle the inevitable effect uncertainties. In the experiments, we compare our method with baseline hierarchical planning from both TAMP and RL fields and illustrate the strength of the method. The results show that by embedding RL skills, we extend the capability of TAMP to domains with probabilistic skills, and improve the planning efficiency compared to the previous methods.
引用
收藏
页码:5974 / 5981
页数:8
相关论文
共 50 条
  • [21] Robotic Arm Motion Planning Based on Residual Reinforcement Learning
    Zhou, Dongxu
    Jia, Ruiqing
    Yao, Haifeng
    Xie, Mingzuo
    2021 THE 13TH INTERNATIONAL CONFERENCE ON COMPUTER AND AUTOMATION ENGINEERING (ICCAE 2021), 2021, : 89 - 94
  • [22] Humanoid motion planning of robotic arm based on reinforcement learning
    Yang A.
    Chen Y.
    Xu Y.
    Yi Qi Yi Biao Xue Bao/Chinese Journal of Scientific Instrument, 2021, 42 (12): : 136 - 145
  • [23] Bootstrapping Motor Skill Learning with Motion Planning
    Abbatematteo, Ben
    Rosen, Eric
    Tellex, Stefanie
    Konidaris, George
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 4926 - 4933
  • [24] A Motion Planning Framework for Skill Coordination and Learning
    Kallmann, Marcelo
    Jiang, Xiaoxi
    MOTION PLANNING FOR HUMANOID ROBOTS, 2010, : 277 - 306
  • [25] Cross-Observability Optimistic-Pessimistic Safe Reinforcement Learning for Interactive Motion Planning With Visual Occlusion
    Hou, Xiaohui
    Gan, Minggang
    Wu, Wei
    Ji, Yuan
    Zhao, Shiyue
    Chen, Jie
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (11) : 17602 - 17613
  • [26] Parting with Misconceptions about Learning-based Vehicle Motion Planning
    Dauner, Daniel
    Hallgarten, Marcel
    Geiger, Andreas
    Chitta, Kashyap
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [27] Structure Learning-Based Task Decomposition for Reinforcement Learning in Non-stationary Environments
    Woo, Honguk
    Yoo, Gwangpyo
    Yoo, Minjong
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8657 - 8665
  • [28] Reinforcement learning-based motion control for snake robots in complex environments
    Zhang, Dong
    Ju, Renjie
    Cao, Zhengcai
    ROBOTICA, 2024, 42 (04) : 947 - 961
  • [29] Deep reinforcement learning-based path planning of underactuated surface vessels
    Xu H.
    Wang N.
    Zhao H.
    Zheng Z.
    Cyber-Physical Systems, 2019, 5 (01): : 1 - 17
  • [30] Perception-based learning for motion in contact in task planning
    Cervera, E
    DelPobil, AP
    Marta, E
    Serna, MA
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 1996, 17 (03) : 283 - 308