A Hierarchical Robot Learning Framework for Manipulator Reactive Motion Generation via Multi-Agent Reinforcement Learning and Riemannian Motion Policies

被引:1
|
作者
Wang, Yuliu [1 ,2 ]
Sagawa, Ryusuke [1 ,2 ]
Yoshiyasu, Yusuke [2 ]
机构
[1] Univ Tsukuba, Intelligent & Mech Interact Syst Program, Tsukuba, Ibaraki 3058577, Japan
[2] Natl Inst Adv Ind Sci & Technol, Artificial Intelligence Res Ctr, Comp Vis Res Team, Tsukuba, Ibaraki 3058560, Japan
基金
日本学术振兴会;
关键词
Riemannian motion policies; motion generation; motion planning; robot learning; multi-agent reinforcement learning; hierarchical reinforcement learning;
D O I
10.1109/ACCESS.2023.3324039
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Manipulators motion planning faces new challenges as robots are increasingly used in dense, cluttered and dynamic environments. The recently proposed technique called Riemannian motion policies(RMPs) provides an elegant solution with clear mathematical interpretations to such challenging scenarios. It is based on differential geometry policies that generate reactive motions in dynamic environments with real-time performance. However, designing and combining RMPs is still a difficult task involving extensive parameter tuning, and typically seven or more RMPs need to be combined by using RMPflow to realize motions of a robot manipulator with more than 6 degrees-of-freedoms, where the RMPs parameters have to be empirically set each time. In this paper, we take a policy to decompose such complex policies into multiple learning modules based on reinforcement learning. Specifically, we propose a three-layer robot learning framework that consists of the basic-level, middle-level and top-level layers. At the basic layer, only two base RMPs i.e. target and collision avoidance are used to output reactive actions. At the middle-level layer, a hierarchical reinforcement learning approach is used to train an agent that automatically selects those RMPs and their parameters based on environmental changes and will be deployed at each joint. At the top-level layer, a multi-agent reinforcement learning approach trains all the joints with high-level collaborative policies to accomplish actions such as tracking a target and avoiding obstacles. With simulation experiments, we compare the proposed method with the baseline method and find that our method effectively produces superior actions and is better at avoiding obstacles, handling self-collisions, and avoiding singularities in dynamic environments. In addition, the proposed framework possesses higher training efficiency while leveraging the generalization ability of reinforcement learning to dynamic environments and improving safety and interpretability.
引用
收藏
页码:126979 / 126994
页数:16
相关论文
共 50 条
  • [1] Learning Advanced Locomotion for Quadrupedal Robots: A Distributed Multi-Agent Reinforcement Learning Framework with Riemannian Motion Policies
    Wang, Yuliu
    Sagawa, Ryusuke
    Yoshiyasu, Yusuke
    ROBOTICS, 2024, 13 (06)
  • [2] Manipulator Motion Planning via Centralized Training and Decentralized Execution Multi-Agent Reinforcement Learning
    Wang, Yuliu
    Sagawa, Ryusuke
    2022 INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2022), 2022, : 812 - 817
  • [3] Safe multi-agent motion planning via filtered reinforcement learning
    Vinod, Abraham P.
    Safaoui, Sleiman
    Chakrabarty, Ankush
    Quirynen, Rien
    Yoshikawa, Nobuyuki
    Di Cairano, Stefano
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2022, 2022, : 7270 - 7276
  • [4] Hierarchical DDPG Based Reinforcement Learning Framework for Multi-Agent Collective Motion With Short Communication Ranges
    Li, Jiaxin
    Yi, Peng
    Duan, Tong
    Zhang, Zhen
    Hu, Tao
    IEEE Transactions on Machine Learning in Communications and Networking, 2024, 2 : 633 - 644
  • [5] Composable Energy Policies for Reactive Motion Generation and Reinforcement Learning
    IAS, TU Darmstadt
    不详
    Robot. Sci. Syst., 1600,
  • [6] Composable energy policies for reactive motion generation and reinforcement learning
    Urain, Julen
    Li, Anqi
    Liu, Puze
    D'Eramo, Carlo
    Peters, Jan
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2023, 42 (10): : 827 - 858
  • [7] Composable Energy Policies for Reactive Motion Generation and Reinforcement Learning
    Urain, Julen
    Li, Anqi
    Liu, Puze
    D'Eramo, Carlo
    Peters, Jan
    ROBOTICS: SCIENCE AND SYSTEM XVII, 2021,
  • [8] Modeling collective motion for fish schooling via multi-agent reinforcement learning
    Wang, Xin
    Liu, Shuo
    Yu, Yifan
    Yue, Shengzhi
    Liu, Ying
    Zhang, Fumin
    Lin, Yuanshan
    ECOLOGICAL MODELLING, 2023, 477
  • [9] Hierarchical multi-agent reinforcement learning
    Mohammad Ghavamzadeh
    Sridhar Mahadevan
    Rajbala Makar
    Autonomous Agents and Multi-Agent Systems, 2006, 13 : 197 - 229
  • [10] Hierarchical multi-agent reinforcement learning
    Ghavamzadeh, Mohammad
    Mahadevan, Sridhar
    Makar, Rajbala
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2006, 13 (02) : 197 - 229