Trial and Error Using Previous Experiences as Simulation Models in Humanoid Motor Learning

被引:4
|
作者
Sugimoto, Norikazu [1 ]
Tangkaratt, Voot [2 ]
Wensveen, Thijs [3 ]
Zhao, Tingting [4 ]
Sugiyama, Masashi [2 ]
Morimoto, Jun [5 ]
机构
[1] Natl Inst Informat & Commun Technol, Osaka, Japan
[2] Univ Tokyo, Tokyo 1138654, Japan
[3] Delft Univ Technol, NL-2600 AA Delft, Netherlands
[4] Tianjin Univ Sci & Technol, Tianjin, Peoples R China
[5] ATR Computat Neurosci Labs, Kyoto, Japan
关键词
POLICY GRADIENTS; SAMPLE REUSE; ROBOTICS;
D O I
10.1109/MRA.2015.2511681
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Since biological systems have the ability to efficiently reuse previous experiences to change their behavioral strategies to avoid enemies or find food, the number of required samples from real environments to improve behavioral policy is greatly reduced. Even for real robotic systems, it is desirable to use only a limited number of samples from real environments due to the limited durability of real systems to reduce the required time to improve control performance. In this article, we used previous experiences as environmental local models so that the movement policy of a humanoid robot can be efficiently improved with a limited number of samples from its real environment. We applied our proposed learning method to a real humanoid robot and successfully achieve two challenging control tasks. We applied our proposed learning approach to acquire a policy for a cart-pole swing-up task in a real-virtual hybrid task environment, where the robot waves a PlayStation (PS) Move motion controller to move a cart-pole in a virtual simulator. Furthermore, we applied our proposed method to a challenging basketball-shooting task in a real environment. © 1994-2011 IEEE.
引用
收藏
页码:96 / 105
页数:10
相关论文
共 50 条
  • [1] Efficient Reuse of Previous Experiences in Humanoid Motor Learning
    Sugimoto, Norikazu
    Tangkaratt, Voot
    Wensveen, Thijs
    Zhao, Tingting
    Sugiyama, Masashi
    Morimoto, Jun
    2014 14TH IEEE-RAS INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS), 2014, : 554 - 559
  • [2] A matter of trial and error for motor learning
    Pritchett, Dominique L.
    Carey, Megan R.
    TRENDS IN NEUROSCIENCES, 2014, 37 (09) : 465 - 466
  • [3] Effect of Trial Order and Error Magnitude on Motor Learning by Observing
    Brown, Liana E.
    Wilson, Elizabeth T.
    Obhi, Sukhvinder S.
    Gribble, Paul L.
    JOURNAL OF NEUROPHYSIOLOGY, 2010, 104 (03) : 1409 - 1416
  • [4] TRIAL-AND-ERROR LEARNING AND ECONOMIC-MODELS
    ARCHIBALD, RB
    ELLIOTT, CS
    KYKLOS, 1989, 42 (01) : 39 - 59
  • [5] Anatomy of motor learning .2. Subcortical structures and learning by trial and error
    Jueptner, M
    Frith, CD
    Brooks, DJ
    Frackowiak, RSJ
    Passingham, RE
    JOURNAL OF NEUROPHYSIOLOGY, 1997, 77 (03) : 1325 - 1337
  • [6] Joint Angle Error Reduction for Humanoid Robots Using Dynamics Learning Tree
    Hirai, Ryo
    Gouko, Manabu
    Kim, Chyon Hae
    RECENT TRENDS AND FUTURE TECHNOLOGY IN APPLIED INTELLIGENCE, IEA/AIE 2018, 2018, 10868 : 221 - 232
  • [7] Trial and error in urban restructuring processes: learning from Dutch experiences
    Kokx, Anita
    Spit, Tejo
    van Beckhoven, Ellen
    URBAN RESEARCH & PRACTICE, 2009, 2 (02) : 169 - 188
  • [8] Using Previous Models to Bias Structural Learning in the Hierarchical BOA
    Hauschild, M. W.
    Pelikan, M.
    Sastry, K.
    Goldberg, D. E.
    EVOLUTIONARY COMPUTATION, 2012, 20 (01) : 135 - 160
  • [9] Trial-and-Error Learning of Repulsors for Humanoid QP-based Whole-Body Control
    Spitz, Jonathan
    Bouyarmane, Karim
    Ivaldi, Serena
    Mouret, Jean-Baptiste
    2017 IEEE-RAS 17TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTICS (HUMANOIDS), 2017, : 468 - 475
  • [10] Dynamic Neural Correlates of Motor Error Monitoring and Adaptation during Trial-to-Trial Learning
    Tan, Huiling
    Jenkinson, Ned
    Brown, Peter
    JOURNAL OF NEUROSCIENCE, 2014, 34 (16): : 5678 - 5688