Generalized Model Learning for Reinforcement Learning on a Humanoid Robot

被引:42
|
作者
Hester, Todd [1 ]
Quinlan, Michael [1 ]
Stone, Peter [1 ]
机构
[1] Univ Texas Austin, Dept Comp Sci, Austin, TX 78712 USA
基金
美国国家科学基金会;
关键词
D O I
10.1109/ROBOT.2010.5509181
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) algorithms have long been promising methods for enabling an autonomous robot to improve its behavior on sequential decision-making tasks. The obvious enticement is that the robot should be able to improve its own behavior without the need for detailed step-by-step programming. However, for RL to reach its full potential, the algorithms must be sample efficient: they must learn competent behavior from very few real-world trials. From this perspective, model-based methods, which use experiential data more efficiently than model-free approaches, are appealing. But they often require exhaustive exploration to learn an accurate model of the domain. In this paper, we present an algorithm, Reinforcement Learning with Decision Trees (RLDT), that uses decision trees to learn the model by generalizing the relative effect of actions across states. The agent explores the environment until it believes it has a reasonable policy. The combination of the learning approach with the targeted exploration policy enables fast learning of the model. We compare RL-DT against standard model-free and model-based learning methods, and demonstrate its effectiveness on an Aldebaran Nao humanoid robot scoring goals in a penalty kick scenario.
引用
收藏
页码:2369 / 2374
页数:6
相关论文
共 50 条
  • [21] Benchmarking the Full-Order Model Optimization Based Imitation in the Humanoid Robot Reinforcement Learning Walk
    Chaikovskaya, Ekaterina
    Minashina, Inna
    Litvinenko, Vladimir
    Davydenko, Egor
    Makarov, Dmitry
    Danik, Yulia
    Gorbachev, Roman
    [J]. 2023 21ST INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS, ICAR, 2023, : 206 - 211
  • [22] Learning low level skills from scratch for humanoid robot soccer using deep reinforcement learning
    Abreu, Miguel
    Lau, Nuno
    Sousa, Armando
    Reis, Luis Paulo
    [J]. 2019 19TH IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS (ICARSC 2019), 2019, : 256 - 263
  • [23] Locomotion Control Method for Humanoid Robot Based on United Hierarchical Reinforcement Learning
    Liu, Boying
    Ma, Lu
    Liu, Chenju
    Xu, BinChen
    [J]. 2020 IEEE 16TH INTERNATIONAL CONFERENCE ON CONTROL & AUTOMATION (ICCA), 2020, : 1161 - 1166
  • [24] Reinforcement Learning for Bipedal Gait with MAX-E2 Humanoid Robot
    Yanguas-Rojas, David
    Mojica-Nava, Eduardo
    Cardenas, Alben
    [J]. INTERNATIONAL JOURNAL OF HUMANOID ROBOTICS, 2022, 19 (05)
  • [25] Acceleration of reinforcement learning by a mobile robot using generalized rules
    Inoue, K
    Ota, J
    Katayama, T
    Arai, T
    [J]. 2000 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2000), VOLS 1-3, PROCEEDINGS, 2000, : 885 - 890
  • [26] Generalized reinforcement learning model of dopamine activity
    Kaveri, Sivaramakrishnan
    Nakahara, Hiroyuki
    [J]. NEUROSCIENCE RESEARCH, 2008, 61 : S195 - S195
  • [27] A Neurocomputational Model Implemented on Humanoid Robot for Learning Action Selection
    Ercelik, Emec
    Sengor, Neslihan Serap
    [J]. 2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [28] GENETIC ALGORITHM FOR A LEARNING HUMANOID ROBOT
    Toskova, Asya
    Toskov, Borislav
    Stoyanov, Stanimir
    Popchev, Ivan
    [J]. COMPTES RENDUS DE L ACADEMIE BULGARE DES SCIENCES, 2019, 72 (08): : 1102 - 1110
  • [29] Kinesthetic Learning of Behaviors in a Humanoid Robot
    Cho, Sumin
    Jo, Sungho
    [J]. 2011 11TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2011, : 1108 - 1112
  • [30] Learning the Odometry on a Small Humanoid Robot
    Rouxel, Quentin
    Passault, Gregoire
    Hofer, Ludovic
    N'Guyen, Steve
    Ly, Olivier
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2016, : 1810 - 1816