Integrating guidance into relational reinforcement learning

被引:55
|
作者
Driessens, K
Dzeroski, S
机构
[1] Katholieke Univ Leuven, Dept Comp Sci, B-3001 Heverlee, Belgium
[2] Jozef Stefan Inst, Dept Intelligent Syst, SI-1000 Ljubljana, Slovenia
关键词
reinforcement learning; relational learning; guided exploration;
D O I
10.1023/B:MACH.0000039779.47329.3a
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning, and Q-learning in particular, encounter two major problems when dealing with large state spaces. First, learning the Q-function in tabular form may be infeasible because of the excessive amount of memory needed to store the table, and because the Q-function only converges after each state has been visited multiple times. Second, rewards in the state space may be so sparse that with random exploration they will only be discovered extremely slowly. The first problem is often solved by learning a generalization of the encountered examples ( e. g., using a neural net or decision tree). Relational reinforcement learning (RRL) is such an approach; it makes Q-learning feasible in structural domains by incorporating a relational learner into Q-learning. The problem of sparse rewards has not been addressed for RRL. This paper presents a solution based on the use of "reasonable policies" to provide guidance. Different types of policies and different strategies to supply guidance through these policies are discussed and evaluated experimentally in several relational domains to show the merits of the approach.
引用
收藏
页码:271 / 304
页数:34
相关论文
共 50 条
  • [41] Modelling Shared Attention Through Relational Reinforcement Learning
    Renato Ramos da Silva
    Roseli Aparecida Francelin Romero
    Journal of Intelligent & Robotic Systems, 2012, 66 : 167 - 182
  • [42] Graph kernels and Gaussian processes for relational reinforcement learning
    Gärtner, T
    Driessens, K
    Ramon, J
    INDUCTIVE LOGIC PROGRAMMING, PROCEEDINGS, 2003, 2835 : 146 - 163
  • [43] Graph kernels and Gaussian processes for relational reinforcement learning
    Driessens, Kurt
    Ramon, Jan
    Gaertner, Thomas
    MACHINE LEARNING, 2006, 64 (1-3) : 91 - 119
  • [44] Modelling Shared Attention Through Relational Reinforcement Learning
    da Silva, Renato Ramos
    Francelin Romero, Roseli Aparecida
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2012, 66 (1-2) : 167 - 182
  • [45] Graph kernels and Gaussian processes for relational reinforcement learning
    Kurt Driessens
    Jan Ramon
    Thomas Gärtner
    Machine Learning, 2006, 64 : 91 - 119
  • [46] Relational Deep Reinforcement Learning for Routing in Wireless Networks
    Manfredi, Victoria
    Wolfe, Alicia P.
    Wang, Bing
    Zhang, Xiaolan
    2021 IEEE 22ND INTERNATIONAL SYMPOSIUM ON A WORLD OF WIRELESS, MOBILE AND MULTIMEDIA NETWORKS (WOWMOM 2021), 2021, : 159 - 168
  • [47] Integrating Classical Control into Reinforcement Learning Policy
    Huang, Ye
    Gu, Chaochen
    Guan, Xinping
    NEURAL PROCESSING LETTERS, 2021, 53 (03) : 1709 - 1722
  • [48] Integrating Classical Control into Reinforcement Learning Policy
    Ye Huang
    Chaochen Gu
    Xinping Guan
    Neural Processing Letters, 2021, 53 : 1709 - 1722
  • [49] Integrating reinforcement learning, bidding and genetic algorithms
    Qi, DH
    Sun, R
    IEEE/WIC INTERNATIONAL CONFERENCE ON INTELLIGENT AGENT TECHNOLOGY, PROCEEDINGS, 2003, : 53 - 59
  • [50] Integrating Expert Knowledge into Fuzzy Reinforcement Learning
    Tompa, Tomas
    Kovacs, Szilveszter
    18TH INTERNATIONAL SYMPOSIUM ON APPLIED COMPUTATIONAL INTELLIGENCE AND INFORMATICS, SACI 2024, 2024, : 333 - 337