A Hierarchical Robot Learning Framework for Manipulator Reactive Motion Generation via Multi-Agent Reinforcement Learning and Riemannian Motion Policies

被引:1
|
作者
Wang, Yuliu [1 ,2 ]
Sagawa, Ryusuke [1 ,2 ]
Yoshiyasu, Yusuke [2 ]
机构
[1] Univ Tsukuba, Intelligent & Mech Interact Syst Program, Tsukuba, Ibaraki 3058577, Japan
[2] Natl Inst Adv Ind Sci & Technol, Artificial Intelligence Res Ctr, Comp Vis Res Team, Tsukuba, Ibaraki 3058560, Japan
基金
日本学术振兴会;
关键词
Riemannian motion policies; motion generation; motion planning; robot learning; multi-agent reinforcement learning; hierarchical reinforcement learning;
D O I
10.1109/ACCESS.2023.3324039
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Manipulators motion planning faces new challenges as robots are increasingly used in dense, cluttered and dynamic environments. The recently proposed technique called Riemannian motion policies(RMPs) provides an elegant solution with clear mathematical interpretations to such challenging scenarios. It is based on differential geometry policies that generate reactive motions in dynamic environments with real-time performance. However, designing and combining RMPs is still a difficult task involving extensive parameter tuning, and typically seven or more RMPs need to be combined by using RMPflow to realize motions of a robot manipulator with more than 6 degrees-of-freedoms, where the RMPs parameters have to be empirically set each time. In this paper, we take a policy to decompose such complex policies into multiple learning modules based on reinforcement learning. Specifically, we propose a three-layer robot learning framework that consists of the basic-level, middle-level and top-level layers. At the basic layer, only two base RMPs i.e. target and collision avoidance are used to output reactive actions. At the middle-level layer, a hierarchical reinforcement learning approach is used to train an agent that automatically selects those RMPs and their parameters based on environmental changes and will be deployed at each joint. At the top-level layer, a multi-agent reinforcement learning approach trains all the joints with high-level collaborative policies to accomplish actions such as tracking a target and avoiding obstacles. With simulation experiments, we compare the proposed method with the baseline method and find that our method effectively produces superior actions and is better at avoiding obstacles, handling self-collisions, and avoiding singularities in dynamic environments. In addition, the proposed framework possesses higher training efficiency while leveraging the generalization ability of reinforcement learning to dynamic environments and improving safety and interpretability.
引用
收藏
页码:126979 / 126994
页数:16
相关论文
共 50 条
  • [31] Hierarchical reinforcement learning via dynamic subspace search for multi-agent planning
    Ma, Aaron
    Ouimet, Michael
    Cortes, Jorge
    AUTONOMOUS ROBOTS, 2020, 44 (3-4) : 485 - 503
  • [32] Cooperative Learning of Multi-Agent Systems Via Reinforcement Learning
    Wang, Xin
    Zhao, Chen
    Huang, Tingwen
    Chakrabarti, Prasun
    Kurths, Juergen
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2023, 9 : 13 - 23
  • [33] Hierarchical reinforcement learning via dynamic subspace search for multi-agent planning
    Aaron Ma
    Michael Ouimet
    Jorge Cortés
    Autonomous Robots, 2020, 44 : 485 - 503
  • [34] Emergent Social Learning via Multi-agent Reinforcement Learning
    Ndousse, Kamal
    Eck, Douglas
    Levine, Sergey
    Jaques, Natasha
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [35] A deep learning framework for realistic robot motion generation
    Dong, Ran
    Chang, Qiong
    Ikuno, Soichiro
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (32): : 23343 - 23356
  • [36] A deep learning framework for realistic robot motion generation
    Ran Dong
    Qiong Chang
    Soichiro Ikuno
    Neural Computing and Applications, 2023, 35 : 23343 - 23356
  • [37] MAVIPER: Learning Decision Tree Policies for Interpretable Multi-agent Reinforcement Learning
    Milani, Stephanie
    Zhang, Zhicheng
    Topin, Nicholay
    Shi, Zheyuan Ryan
    Kamhoua, Charles
    Papalexakis, Evangelos E.
    Fang, Fei
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2022, PT IV, 2023, 13716 : 251 - 266
  • [38] Reinforcement Learning with Quantitative Verification for Assured Multi-Agent Policies
    Riley, Joshua
    Calinescu, Radu
    Paterson, Colin
    Kudenko, Daniel
    Banks, Alec
    ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 2, 2021, : 237 - 245
  • [39] Towards Interpretable Policies in Multi-agent Reinforcement Learning Tasks
    Crespi, Marco
    Custode, Leonardo Lucio
    Iacca, Giovanni
    BIOINSPIRED OPTIMIZATION METHODS AND THEIR APPLICATIONS, 2022, 13627 : 262 - 276
  • [40] Multi-agent reinforcement learning based on policies of global objective
    张化祥
    黄上腾
    Journal of Systems Engineering and Electronics, 2005, (03) : 676 - 681