Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning

被引:131
|
作者
Morimoto, J
Doya, K
机构
[1] JST, ERATO, Kawato Dynam Brain Project, Kyoto 6190288, Japan
[2] Nara Inst Sci & Technol, Grad Sch Informat Sci, Nara 6300101, Japan
[3] JST, CREST, ATR Int, Kyoto 6190288, Japan
关键词
reinforcement learning; hierarchical; real robot; stand-up; motor control;
D O I
10.1016/S0921-8890(01)00113-0
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose a hierarchical reinforcement learning architecture that realizes practical learning speed in real hardware control tasks. In order to enable learning in a practical number of trials, we introduce a low-dimensional representation of the state of the robot for higher-level planning. The upper level learns a discrete sequence of sub-goals in a low-dimensional state space for achieving the main goal of the task. The lower-level modules learn local trajectories in the original high-dimensional state space to achieve the sub-goal specified by the upper level. We applied the hierarchical architecture to a three-link, two-joint robot for the task of learning to stand up by trial and error. The upper-level learning was implemented by Q-learning, while the lower-level learning was implemented by a continuous actor-critic method. The robot successfully learned to stand up within 750 trials in simulation and then in an additional 170 trials using real hardware. The effects of the setting of the search steps in the upper level and the use of a supplementary reward for achieving sub-goals are also tested in simulation. (C) 2001 Elsevier Science B.V. All rights reserved.
引用
下载
收藏
页码:37 / 51
页数:15
相关论文
共 50 条
  • [41] Mobile robot navigation control based on hierarchical reinforcement learning
    Department of Automation, University of Science and Technology of China, Hefei 230027, China
    Nanjing Hangkong Hangtian Daxue Xuebao, 2006, 1 (70-75):
  • [42] Fast Robot Hierarchical Exploration Based on Deep Reinforcement Learning
    Zuo, Shun
    Niu, Jianwei
    Ren, Lu
    Ouyang, Zhenchao
    2023 INTERNATIONAL WIRELESS COMMUNICATIONS AND MOBILE COMPUTING, IWCMC, 2023, : 138 - 143
  • [43] Cooperative behavior acquisition mechanism for a multi-robot system based on reinforcement learning in continuous space
    Yasuda, T
    Ohkura, K
    Taura, T
    2003 IEEE INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE IN ROBOTICS AND AUTOMATION, VOLS I-III, PROCEEDINGS, 2003, : 1539 - 1544
  • [44] Neural reinforcement learning controllers for a real robot application
    Hafner, Roland
    Riedmiller, Martin
    PROCEEDINGS OF THE 2007 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-10, 2007, : 2098 - +
  • [45] Cooperative behavior acquisition mechanism for a multi-robot system based on reinforcement learning in continuous space
    IEEE Robotics and Automation Society (Institute of Electrical and Electronics Engineers Inc., United States):
  • [46] Modular reinforcement learning:: An application to a real robot task
    Kalmár, Z
    Szepesvári, C
    Lörincz, A
    LEARNING ROBOTS, PROCEEDINGS, 1998, 1545 : 29 - 45
  • [47] Speed Up the Process of Reinforcement Learning for Robot Grasping
    Ke, Fengkai
    Zhao, Daxing
    Sun, Guodong
    Feng, Wei
    FUZZY SYSTEMS AND DATA MINING V (FSDM 2019), 2019, 320 : 255 - 260
  • [48] Real-time Motion Generation for Imaginary Creatures Using Hierarchical Reinforcement Learning
    Ogaki, Keisuke
    Nakamura, Masayoshi
    SIGGRAPH'18: ACM SIGGRAPH 2018 STUDIO, 2018,
  • [49] Learning Socially Appropriate Robot Approaching Behavior Toward Groups using Deep Reinforcement Learning
    Gao, Yuan
    Yang, Fangkai
    Frisk, Martin
    Hernandez, Daniel
    Peters, Christopher
    Castellano, Ginevra
    2019 28TH IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2019,
  • [50] Hierarchical Reinforcement Learning Explains Task Interleaving Behavior
    Gebhardt C.
    Oulasvirta A.
    Hilliges O.
    Computational Brain & Behavior, 2021, 4 (3) : 284 - 304