Reinforcement learning of dynamic motor sequence: Learning to stand up

被引:0
|
作者
Morimoto, J [1 ]
Doya, K [1 ]
机构
[1] Nara Inst Sci & Technol, Grad Sch Informat Sci, ATR, Human Informat Proc Lab, Nara 6300101, Japan
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose a learning method for implementing human-like sequential movements in robots. As an example of dynamic sequential movement, we consider the "stand-up" task for a two-joint, three-link: robot. In contrast to the case of steady walking or standing, the desired trajectory for such a transient behavior is very difficult to derive. The goal of the task is to find a path that links a lying state to an upright state under the constraints of the system dynamics. The geometry of the robot is such that there is no static solution; the robot has to stand up dynamically utilizing the momentum of its body. We use reinforcement learning, in particular, a continuous time and state temporal difference (TD) learning method. For successful results, we use 1) an efficient method of value function approximation in a high-dimensional state space, and 2) a hierarchical architecture which divides a large state space into a few smaller pieces.
引用
收藏
页码:1721 / 1726
页数:6
相关论文
共 50 条
  • [1] Hierarchical reinforcement learning for motion learning: learning 'stand-up' trajectories
    Morimoto, J
    Doya, K
    [J]. ADVANCED ROBOTICS, 1999, 13 (03) : 267 - 268
  • [2] Hierarchical reinforcement learning for motion learning: Learning `stand-up' trajectories
    Morimoto, Jun
    Doya, Kenji
    [J]. Advanced Robotics, 1998, 13 (03): : 267 - 268
  • [3] Reinforcement learning of a continuous motor sequence with hidden states
    Arie, Hiroaki
    Ogata, Tetsuya
    Tani, Jun
    Sugano, Shigeki
    [J]. ADVANCED ROBOTICS, 2007, 21 (10) : 1215 - 1229
  • [4] Dynamic assembly sequence selection using reinforcement learning
    Lowe, G
    Shirinzadeh, B
    [J]. 2004 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1- 5, PROCEEDINGS, 2004, : 2633 - 2638
  • [6] Guided Reinforcement Learning via Sequence Learning
    Ramamurthy, Rajkumar
    Sifa, Rafet
    Luebbering, Max
    Bauckhage, Christian
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 335 - 345
  • [7] Machining sequence learning via inverse reinforcement learning
    Sugisawa, Yasutomo
    Takasugi, Keigo
    Asakawa, Naoki
    [J]. PRECISION ENGINEERING-JOURNAL OF THE INTERNATIONAL SOCIETIES FOR PRECISION ENGINEERING AND NANOTECHNOLOGY, 2022, 73 : 477 - 487
  • [8] The error-related negativity as a reinforcement learning signal in motor sequence acquisition
    Mathewson, Kyle
    Krigolson, Olav
    Holroyd, Clay
    [J]. CANADIAN JOURNAL OF EXPERIMENTAL PSYCHOLOGY-REVUE CANADIENNE DE PSYCHOLOGIE EXPERIMENTALE, 2007, 61 (04): : 369 - 370
  • [9] Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning
    Morimoto, J
    Doya, K
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2001, 36 (01) : 37 - 51
  • [10] Dynamic pricing and reinforcement learning
    Carvalho, AX
    Puterman, ML
    [J]. PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS 2003, VOLS 1-4, 2003, : 2916 - 2921