A state space filter for reinforcement learning in POMDPs - Application to a continuous state space -

被引:0
|
作者
Nagayoshi, Masato [1 ,2 ]
Murao, Hajime [3 ]
Tamaki, Hisashi [1 ]
机构
[1] Kobe Univ, Grad Sch Sci & Technol, Nada Ku, Kobe, Hyogo 6578501, Japan
[2] Hyogo Assistive Technol Res & Design Inst, Kobe 6512181, Japan
[3] Kobe Univ, Fac Cross Cultural Studies, Kobe 6578501, Japan
关键词
reinforcement learning; state space design; POMDPs; state space filtering; continuous state space; entropy;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a technique to deal with both discrete and continuous state space systems in POMDPs for reinforcement learning while keeping the state space of an agent compact. First our computational model for MDP environments, where a concept of "state space filtering" has been introduced and constructed to make properly the state space of an agent smaller by referring to "entropy" calculated based on the state-action mapping, is extended to be applicable in POMDP environments by introducing the mechanism of utilizing effectively of history information. Then, it is possible to deal with a continuous state space as well as a discrete state space. Here, the mechanism of adjusting the amount of history information is also introduced so that the state space of an agent should be compact. Moreover, some computational experiments with a robot navigation problem with a continuous state space have been carried out. The potential and the effectiveness of the extended approach have been confirmed through these experiments.
引用
收藏
页码:3098 / +
页数:2
相关论文
共 50 条
  • [41] Reinforcement Learning in a Birth and Death Process: Breaking the Dependence on the State Space
    Anselmi, Jonatha
    Gaujal, Bruno
    Rebuffi, Louis Sebastien
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [42] Efficient state-space representation by neural maps for reinforcement learning
    Herrmann, M
    Der, R
    [J]. CLASSIFICATION IN THE INFORMATION AGE, 1999, : 302 - 309
  • [43] Hierarchical reinforcement learning algorithm based on structural state-space
    Meng, Jiang-Hua
    Zhu, Ji-Hong
    Sun, Zeng-Qi
    [J]. Kongzhi yu Juece/Control and Decision, 2007, 22 (02): : 233 - 237
  • [44] Multiagent reinforcement learning with the partly high-dimensional state space
    Department of Electrical and Computer Engineering, Nagoya Institute of Technology, Nagoya, 466-8555, Japan
    [J]. Syst Comput Jpn, 2006, 9 (22-31):
  • [45] A proposition of adaptive state space partition in reinforcement learning with Voronoi tessellation
    Fuchida, Takayasu
    Aung, Kathy Thi
    [J]. ARTIFICIAL LIFE AND ROBOTICS, 2013, 18 (3-4) : 172 - 177
  • [46] Anomaly detection using state-space models and reinforcement learning
    Khazaeli, Shervin
    Nguyen, Luong Ha
    Goulet, James A.
    [J]. STRUCTURAL CONTROL & HEALTH MONITORING, 2021, 28 (06):
  • [47] A proposition of adaptive state space partition in reinforcement learning with Voronoi Tessellation
    Aung, Kathy Thi
    Fuchida, Takayasu
    [J]. PROCEEDINGS OF THE SEVENTEENTH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL LIFE AND ROBOTICS (AROB 17TH '12), 2012, : 638 - 641
  • [48] A reinforcement learning with adaptive state space construction for mobile robot navigation
    Li, Guizhi
    Pang, Jie
    [J]. PROCEEDINGS OF THE 2006 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING AND CONTROL, 2006, : 84 - 88
  • [49] State space maximum correntropy filter
    Liu, Xi
    Qu, Hua
    Zhao, Jihong
    Chen, Badong
    [J]. SIGNAL PROCESSING, 2017, 130 : 152 - 158
  • [50] Bayesian Reinforcement Learning in Continuous POMDPs with Gaussian Processes
    Dallaire, Patrick
    Besse, Camille
    Ross, Stephane
    Chaib-draa, Brahim
    [J]. 2009 IEEE-RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, 2009, : 2604 - 2609