Leveraging the efficiency of multi-task robot manipulation via task-evoked planner and reinforcement learning

被引:1
|
作者
Qi, Haofu [1 ,2 ]
Zheng, Haoyang [1 ,2 ]
Shao, Jun [1 ,2 ]
Zhang, Jiatao [1 ,2 ]
Gu, Jason [2 ]
Song, Wei [1 ,2 ]
Zhu, Shiqiang [1 ]
机构
[1] Zhejiang Univ, Hangzhou 310030, Peoples R China
[2] Zhejiang Lab, Res Inst Interdisciplinary Innovat, Res Ctr Intelligent Robot, Hangzhou 311100, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICRA57147.2024.10611076
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-task learning has expanded the boundaries of robotic manipulation, enabling the execution of increasingly complex tasks. However, policies learned through reinforcement learning exhibit limited generalization and narrow distributions, which restrict their effectiveness in multi-task training. Addressing the challenge of obtaining policies with generalization and stability represents a non-trivial problem. To tackle this issue, we propose a planning-guided reinforcement learning method. It leverages a task-evoked planner(TEP) and a reinforcement learning approach with planner's guidance. TEP utilizes reusable samples as the source, with the aim of learning reachability information across different task scenarios. Then in reinforcement learning, TEP assesses and guides the Actor towards better outputs and smoothly enhances the performance in multi-task benchmarks. We evaluate this approach within the Meta-World framework and compare it with prior works in terms of learning efficiency and effectiveness. Depending on experimental results, our method has more efficiency, higher success rates, and demonstrates more realistic behavior.
引用
收藏
页码:9220 / 9226
页数:7
相关论文
共 50 条
  • [21] Multi-Task Multi-Agent Reinforcement Learning With Interaction and Task Representations
    Li, Chao
    Dong, Shaokang
    Yang, Shangdong
    Hu, Yujing
    Ding, Tianyu
    Li, Wenbin
    Gao, Yang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [22] Adversarial Online Multi-Task Reinforcement Learning
    Nguyen, Quan
    Mehta, Nishant A.
    INTERNATIONAL CONFERENCE ON ALGORITHMIC LEARNING THEORY, VOL 201, 2023, 201 : 1124 - 1165
  • [23] Adversarial Online Multi-Task Reinforcement Learning
    Nguyen, Quan
    Mehta, Nishant A.
    Proceedings of Machine Learning Research, 2023, 201 : 1124 - 1165
  • [24] Learning Multi-Task Transferable Rewards via Variational Inverse Reinforcement Learning
    Yoo, Se-Wook
    Seo, Seung-Woo
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022,
  • [25] Multi-task Batch Reinforcement Learning with Metric Learning
    Li, Jiachen
    Quan Vuong
    Liu, Shuang
    Liu, Minghua
    Ciosek, Kamil
    Christensen, Henrik
    Su, Hao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [26] Skills Regularized Task Decomposition for Multi-task Offline Reinforcement Learning
    Yoo, Minjong
    Cho, Sangwoo
    Woo, Honguk
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [27] Multi-task Deep Reinforcement Learning for Scalable Parallel Task Scheduling
    Zhang, Lingxin
    Qi, Qi
    Wang, Jingyu
    Sun, Haifeng
    Liao, Jianxin
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 2992 - 3001
  • [28] Reset-Free Reinforcement Learning via Multi-Task Learning: Learning Dexterous Manipulation Behaviors without Human Intervention
    Gupta, Abhishek
    Yu, Justin
    Zhao, Tony Z.
    Kumar, Vikash
    Rovinsky, Aaron
    Xu, Kelvin
    Devlin, Thomas
    Levine, Sergey
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 6664 - 6671
  • [29] HIERARCHICAL MULTI-TASK LEARNING VIA TASK AFFINITY GROUPINGS
    Srivastava, Siddharth
    Bhugra, Swati
    Kaushik, Vinay
    Lall, Brejesh
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3289 - 3293
  • [30] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    Memetic Computing, 2020, 12 : 355 - 369