Integrating guidance into relational reinforcement learning

被引:53
|
作者
Driessens, K
Dzeroski, S
机构
[1] Katholieke Univ Leuven, Dept Comp Sci, B-3001 Heverlee, Belgium
[2] Jozef Stefan Inst, Dept Intelligent Syst, SI-1000 Ljubljana, Slovenia
关键词
reinforcement learning; relational learning; guided exploration;
D O I
10.1023/B:MACH.0000039779.47329.3a
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning, and Q-learning in particular, encounter two major problems when dealing with large state spaces. First, learning the Q-function in tabular form may be infeasible because of the excessive amount of memory needed to store the table, and because the Q-function only converges after each state has been visited multiple times. Second, rewards in the state space may be so sparse that with random exploration they will only be discovered extremely slowly. The first problem is often solved by learning a generalization of the encountered examples ( e. g., using a neural net or decision tree). Relational reinforcement learning (RRL) is such an approach; it makes Q-learning feasible in structural domains by incorporating a relational learner into Q-learning. The problem of sparse rewards has not been addressed for RRL. This paper presents a solution based on the use of "reasonable policies" to provide guidance. Different types of policies and different strategies to supply guidance through these policies are discussed and evaluated experimentally in several relational domains to show the merits of the approach.
引用
收藏
页码:271 / 304
页数:34
相关论文
共 50 条
  • [1] Integrating Guidance into Relational Reinforcement Learning
    Kurt Driessens
    Sašo Džeroski
    [J]. Machine Learning, 2004, 57 : 271 - 304
  • [2] Relational reinforcement learning
    Driessens, K
    [J]. AI COMMUNICATIONS, 2005, 18 (01) : 71 - 73
  • [3] Relational reinforcement learning
    Driessens, K
    [J]. MULTI-AGENT SYSTEMS AND APPLICATIONS, 2001, 2086 : 271 - 280
  • [4] Relational reinforcement learning
    Dzeroski, S
    De Raedt, L
    Driessens, K
    [J]. MACHINE LEARNING, 2001, 43 (1-2) : 7 - 52
  • [5] Relational Reinforcement Learning
    Sašo Džeroski
    Luc De Raedt
    Kurt Driessens
    [J]. Machine Learning, 2001, 43 : 7 - 52
  • [6] Hybrid Deep RePReL: Integrating Relational Planning and Reinforcement Learning for Information Fusion
    Kokel, Harsha
    Prabhakar, Nikhilesh
    Ravindran, Balaraman
    Blasch, Erik
    Tadepalli, Prasad
    Natarajan, Sriraam
    [J]. 2022 25TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION 2022), 2022,
  • [7] Learning relational options for inductive transfer in relational reinforcement learning
    Croonenborghs, Tom
    Driessens, Kurt
    Bruynooghe, Maurice
    [J]. INDUCTIVE LOGIC PROGRAMMING, 2008, 4894 : 88 - 97
  • [8] An Enhancement of Relational Reinforcement Learning
    da Silva, Renato R.
    Policastro, Claudio A.
    Romero, Roseli A. F.
    [J]. 2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 2055 - 2060
  • [9] Relational reinforcement learning with guided demonstrations
    Martinez, David
    Alenya, Guillem
    Torras, Carme
    [J]. ARTIFICIAL INTELLIGENCE, 2017, 247 : 295 - 312
  • [10] Utile Distinctions for Relational Reinforcement Learning
    Dabney, William
    McGovern, Amy
    [J]. 20TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2007, : 738 - 743