Tree based discretization for continuous state space reinforcement learning

被引:0
|
作者
Uther, WTB [1 ]
Veloso, MM [1 ]
机构
[1] Carnegie Mellon Univ, Dept Comp Sci, Pittsburgh, PA 15213 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is an effective technique for learning action policies in discrete stochastic environments, but its efficiency can decay exponentially with the size of the state space. In many situations significant portions of a large state space may be irrelevant to a specific goal and can be aggregated into a few, relevant, states. The U Tree algorithm generates a tree based state discretization that efficiently finds the relevant state chunks of large propositional domains. In this paper, we extend the U Tree algorithm to challenging domains with a continuous state space for which there is no initial discretization. This Continuous U Tree algorithm transfers traditional regression tree techniques to reinforcement learning. We have performed experiments in a variety of domains that show that Continuous U Tree effectively handles large continuous state spaces. In this paper, we report on results in two domains, one gives a clear visualization of the algorithm and another empirically demonstrates an effective state discretization in a simple multi-agent environment.
引用
收藏
页码:769 / 774
页数:6
相关论文
共 50 条
  • [1] Budgeted Reinforcement Learning in Continuous State Space
    Carrara, Nicolas
    Leurent, Edouard
    Laroche, Romain
    Urvoy, Tanguy
    Maillard, Odalric-Ambrym
    Pietquin, Olivier
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] A state space filter for reinforcement learning in POMDPs - Application to a continuous state space -
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    [J]. 2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 3098 - +
  • [3] Reinforcement Learning Method for Continuous State Space Based on Dynamic Neural Network
    Sun, Wei
    Wang, Xuesong
    Cheng, Yuhu
    [J]. 2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 750 - 754
  • [4] On the Convergence of Reinforcement Learning in Nonlinear Continuous State Space Problems
    Goyal, Raman
    Chakravorty, Suman
    Wang, Ran
    Mohamed, Mohamed Naveed Gul
    [J]. 2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 2969 - 2975
  • [5] Inverse Reinforcement Learning in a Continuous State Space with Formal Guarantees
    Dexter, Gregory
    Bello, Kevin
    Honorio, Jean
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [6] Hierarchical Reinforcement Learning Based on Continuous Subgoal Space
    Wang, Chen
    Zeng, Fanyu
    Ge, Shuzhi Sam
    Jiang, Xin
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON REAL-TIME COMPUTING AND ROBOTICS (IEEE-RCAR 2020), 2020, : 74 - 80
  • [7] Multi-Robot Cooperation Based on Continuous Reinforcement Learning with Two State Space Representations
    Yasuda, Toshiyuki
    Ohkura, Kazuhiro
    Yamada, Kazuaki
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC 2013), 2013, : 4470 - 4475
  • [8] Reinforcement learning in continuous time and space
    Doya, K
    [J]. NEURAL COMPUTATION, 2000, 12 (01) : 219 - 245
  • [9] Self-organizing decision tree based on reinforcement learning and its application on state space partition
    Kao-Shing Hwang
    Tsung-Wen Yang
    Chia-Ju Lin
    [J]. 2006 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS, VOLS 1-6, PROCEEDINGS, 2006, : 5088 - 5093
  • [10] Energy management of hybrid electric bus based on deep reinforcement learning in continuous state and action space
    Tan, Huachun
    Zhang, Hailong
    Peng, Jiankun
    Jiang, Zhuxi
    Wu, Yuankai
    [J]. ENERGY CONVERSION AND MANAGEMENT, 2019, 195 : 548 - 560