Relational reinforcement learning with guided demonstrations

被引:22
|
作者
Martinez, David [1 ]
Alenya, Guillem [1 ]
Torras, Carme [1 ]
机构
[1] UPC, CSIC, Inst Robot & Informat Ind, Llorens & Artigas 4-6, Barcelona 08028, Spain
关键词
Active learning; Learning guidance; Planning excuse; Reinforcement learning; Robot learning; Teacher demonstration; Teacher guidance; POLYNOMIAL-TIME;
D O I
10.1016/j.artint.2015.02.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Model-based reinforcement learning is a powerful paradigm for learning tasks in robotics. However, in-depth exploration is usually required and the actions have to be known in advance. Thus, we propose a novel algorithm that integrates the option of requesting teacher demonstrations to learn new domains with fewer action executions and no previous knowledge. Demonstrations allow new actions to be learned and they greatly reduce the amount of exploration required, but they are only requested when they are expected to yield a significant improvement because the teacher's time is considered to be more valuable than the robot's time. Moreover, selecting the appropriate action to demonstrate is not an easy task, and thus some guidance is provided to the teacher. The rule-based model is analyzed to determine the parts of the state that may be incomplete, and to provide the teacher with a set of possible problems for which a demonstration is needed. Rule analysis is also used to find better alternative models and to complete subgoals before requesting help, thereby minimizing the number of requested demonstrations. These improvements were demonstrated in a set of experiments, which included domains from the international planning competition and a robotic task. Adding teacher demonstrations and rule analysis reduced the amount of exploration required by up to 60% in some domains, and improved the success ratio by 35% in other domains. (C) 2015 Elsevier B.V. All rights reserved.
引用
收藏
页码:295 / 312
页数:18
相关论文
共 50 条
  • [1] Learning from Ambiguous Demonstrations with Self-Explanation Guided Reinforcement Learning
    Zha, Yantian
    Guan, Lin
    Kambhampati, Subbarao
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 9, 2024, : 10395 - 10403
  • [2] Relational reinforcement learning
    Driessens, K
    [J]. AI COMMUNICATIONS, 2005, 18 (01) : 71 - 73
  • [3] Relational reinforcement learning
    Driessens, K
    [J]. MULTI-AGENT SYSTEMS AND APPLICATIONS, 2001, 2086 : 271 - 280
  • [4] Relational reinforcement learning
    Dzeroski, S
    De Raedt, L
    Driessens, K
    [J]. MACHINE LEARNING, 2001, 43 (1-2) : 7 - 52
  • [5] Relational Reinforcement Learning
    Sašo Džeroski
    Luc De Raedt
    Kurt Driessens
    [J]. Machine Learning, 2001, 43 : 7 - 52
  • [6] Overcoming Exploration in Reinforcement Learning with Demonstrations
    Nair, Ashvin
    McGrew, Bob
    Andrychowicz, Marcin
    Zaremba, Wojciech
    Abbeel, Pieter
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2018, : 6292 - 6299
  • [7] Self-Guided Actor-Critic: Reinforcement Learning from Adaptive Expert Demonstrations
    Zhang, Haoran
    Yin, Chenkun
    Zhang, Yanxin
    Jin, Shangtai
    [J]. 2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 572 - 577
  • [8] Learning relational options for inductive transfer in relational reinforcement learning
    Croonenborghs, Tom
    Driessens, Kurt
    Bruynooghe, Maurice
    [J]. INDUCTIVE LOGIC PROGRAMMING, 2008, 4894 : 88 - 97
  • [9] An Enhancement of Relational Reinforcement Learning
    da Silva, Renato R.
    Policastro, Claudio A.
    Romero, Roseli A. F.
    [J]. 2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 2055 - 2060
  • [10] Analysis of Inverse Reinforcement Learning with Perturbed Demonstrations
    Melo, Francisco S.
    Lopes, Manuel
    Ferreira, Ricardo
    [J]. ECAI 2010 - 19TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2010, 215 : 349 - 354