Adaptive state space formation in reinforcement learning

被引:0
|
作者
Samejima, K [1 ]
Omori, T [1 ]
机构
[1] Tokyo Univ Agr & Technol, Fac Engn, Koganei, Tokyo 184, Japan
关键词
reinforcement learning; locally weighted learning; function approximation; collision avoidance problem; basis division;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Difficulties are encountered during the application of the reinforcement learning method to real-world problems. One of them is the formation of a discrete state space from a continuous input signal. In the absence of a priori knowledge about input space, a straightforward approach to this problem is to separate the input space into a grid, and to use lookup tables. But this method suffers from the dimensionality problem. Some studies use continuous function approximaters such as neural networks instead of lookup tables. However, when a global basis function such as sigmoid function is used, convergence can't be guaranteed. For this problem we propose a method in which local basis functions are assigned depending on the task requirement. In the initial state of the learning, only one basis function is presented over the entire space. The basis function is divided by the statistical property of locally weighted temporal difference error. We applied the method to an autonomous robot collision avoidance problem, and evaluated the validity of the algorithm.
引用
收藏
页码:251 / 255
页数:5
相关论文
共 50 条
  • [1] Adaptive state space partitioning for reinforcement learning
    Lee, ISK
    Lau, HYK
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2004, 17 (06) : 577 - 588
  • [2] A proposition of adaptive state space partition in reinforcement learning with Voronoi tessellation
    Fuchida, Takayasu
    Aung, Kathy Thi
    [J]. ARTIFICIAL LIFE AND ROBOTICS, 2013, 18 (3-4) : 172 - 177
  • [3] A proposition of adaptive state space partition in reinforcement learning with Voronoi Tessellation
    Aung, Kathy Thi
    Fuchida, Takayasu
    [J]. PROCEEDINGS OF THE SEVENTEENTH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL LIFE AND ROBOTICS (AROB 17TH '12), 2012, : 638 - 641
  • [4] A reinforcement learning with adaptive state space construction for mobile robot navigation
    Li, Guizhi
    Pang, Jie
    [J]. PROCEEDINGS OF THE 2006 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING AND CONTROL, 2006, : 84 - 88
  • [5] Adaptive State Aggregation for Reinforcement Learning
    Hwang, Kao-Shing
    Chen, Yu-Jen
    Jiang, Wei-Cheng
    [J]. PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 2452 - 2456
  • [6] A reinforcement learning with adaptive state space recruitment strategy for real autonomous mobile robots
    Kondo, T
    Ito, K
    [J]. 2002 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-3, PROCEEDINGS, 2002, : 897 - 902
  • [7] Budgeted Reinforcement Learning in Continuous State Space
    Carrara, Nicolas
    Leurent, Edouard
    Laroche, Romain
    Urvoy, Tanguy
    Maillard, Odalric-Ambrym
    Pietquin, Olivier
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [8] A reinforcement learning accelerated by state space reduction
    Senda, K
    Mano, S
    Fujii, S
    [J]. SICE 2003 ANNUAL CONFERENCE, VOLS 1-3, 2003, : 1992 - 1997
  • [9] A state space filter for reinforcement learning in POMDPs - Application to a continuous state space -
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    [J]. 2006 SICE-ICASE INTERNATIONAL JOINT CONFERENCE, VOLS 1-13, 2006, : 3098 - +
  • [10] Adaptive internal state space construction method for Reinforcement learning of a real-world agent
    Samejima, K
    Omori, T
    [J]. NEURAL NETWORKS, 1999, 12 (7-8) : 1143 - 1155