Tree based discretization for continuous state space reinforcement learning

被引:0
|
作者
Uther, WTB [1 ]
Veloso, MM [1 ]
机构
[1] Carnegie Mellon Univ, Dept Comp Sci, Pittsburgh, PA 15213 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is an effective technique for learning action policies in discrete stochastic environments, but its efficiency can decay exponentially with the size of the state space. In many situations significant portions of a large state space may be irrelevant to a specific goal and can be aggregated into a few, relevant, states. The U Tree algorithm generates a tree based state discretization that efficiently finds the relevant state chunks of large propositional domains. In this paper, we extend the U Tree algorithm to challenging domains with a continuous state space for which there is no initial discretization. This Continuous U Tree algorithm transfers traditional regression tree techniques to reinforcement learning. We have performed experiments in a variety of domains that show that Continuous U Tree effectively handles large continuous state spaces. In this paper, we report on results in two domains, one gives a clear visualization of the algorithm and another empirically demonstrates an effective state discretization in a simple multi-agent environment.
引用
收藏
页码:769 / 774
页数:6
相关论文
共 50 条
  • [21] Instance-based reinforcement learning for robot path finding in continuous space
    Nakamura, J
    Ohnishi, S
    Ohkura, K
    Ueda, K
    [J]. SMC '97 CONFERENCE PROCEEDINGS - 1997 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS, VOLS 1-5: CONFERENCE THEME: COMPUTATIONAL CYBERNETICS AND SIMULATION, 1997, : 1229 - 1234
  • [22] THE DISCRETIZATION OF THE CONTINUOUS LINEAR CASCADE BY MEANS OF STATE-SPACE ANALYSIS
    SZOLLOSINAGY, A
    [J]. JOURNAL OF HYDROLOGY, 1982, 58 (3-4) : 223 - 236
  • [23] Adaptive state space formation in reinforcement learning
    Samejima, K
    Omori, T
    [J]. ICONIP'98: THE FIFTH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING JOINTLY WITH JNNS'98: THE 1998 ANNUAL CONFERENCE OF THE JAPANESE NEURAL NETWORK SOCIETY - PROCEEDINGS, VOLS 1-3, 1998, : 251 - 255
  • [24] Adaptive state space partitioning for reinforcement learning
    Lee, ISK
    Lau, HYK
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2004, 17 (06) : 577 - 588
  • [25] A reinforcement learning accelerated by state space reduction
    Senda, K
    Mano, S
    Fujii, S
    [J]. SICE 2003 ANNUAL CONFERENCE, VOLS 1-3, 2003, : 1992 - 1997
  • [26] Barycentric interpolators for continuous space & time reinforcement learning
    Munos, R
    Moore, A
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 11, 1999, 11 : 1024 - 1030
  • [27] Linear inverse reinforcement learning in continuous time and space
    Kamalapurkar, Rushikesh
    [J]. 2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, : 1683 - 1688
  • [28] Reinforcement learning algorithm with CTRNN in continuous action space
    Arie, Hiroaki
    Namikawa, Jun
    Ogata, Tetsuya
    Tani, Jun
    Sugano, Shigeki
    [J]. NEURAL INFORMATION PROCESSING, PT 1, PROCEEDINGS, 2006, 4232 : 387 - 396
  • [29] A reinforcement learning with switching controllers for a continuous action space
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    [J]. ARTIFICIAL LIFE AND ROBOTICS, 2010, 15 (01) : 97 - 100
  • [30] Application of Reinforcement Learning with Continuous State Space to Ramp Metering in Real-world Conditions
    Rezaee, Kasra
    Abdulhai, Baher
    Abdelgawad, Hossam
    [J]. 2012 15TH INTERNATIONAL IEEE CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2012, : 1590 - 1595