Reinforcement learning in nonstationary environment navigation tasks

被引:0
|
作者
Lane, Terran [1 ]
Ridens, Martin [1 ]
Stevens, Scott [1 ]
机构
[1] Univ New Mexico, Dept Comp Sci, Albuquerque, NM 87131 USA
来源
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The field of reinforcement learning (RL) has achieved great strides in learning control knowledge from closed-loop interaction with environments. "Classical" RL, based on atomic state space representations, suffers from an inability to adapt to nonstationarities in the target Markov decision process (i.e., environment). Relational RL is widely seen as being a potential solution to this shortcoming. In this paper, we demonstrate a class of "pseudo-relational" learning methods for nonstationary navigational RL domains - domains in which the location of the goal, or even the structure of the environment, can change over time. Our approach is closely related to deictic representations, which have previously been found to be troublesome for RL. The key insight of this paper is that navigational problems are a highly constrained class of MDP, possessing a strong native topology that relaxes some of the partial observability difficulties arising from deixis. Agents can employ local information that is relevant to their near-term action choices to act effectively. We demonstrate that, unlike an atomic representation, our agents can learn to fluidly adapt to changing goal locations and environment structure.
引用
收藏
页码:429 / +
页数:3
相关论文
共 50 条
  • [1] An environment model for nonstationary reinforcement learning
    Choi, SPM
    Yeung, DY
    Zhang, NL
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 12, 2000, 12 : 987 - 993
  • [2] Quantum Deep Reinforcement Learning for Robot Navigation Tasks
    Hohenfeld, Hans
    Heimann, Dirk
    Wiebe, Felix
    Kirchner, Frank
    [J]. IEEE ACCESS, 2024, 12 : 87217 - 87236
  • [3] Reinforcement learning with function approximation for cooperative navigation tasks
    Melo, Francisco S.
    Ribeiro, M. Isabel
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-9, 2008, : 3321 - +
  • [4] Proxemic behavior in navigation tasks using reinforcement learning
    Millan-Arias, Cristian
    Fernandes, Bruno
    Cruz, Francisco
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 35 (23): : 16723 - 16738
  • [5] Proxemic behavior in navigation tasks using reinforcement learning
    Cristian Millán-Arias
    Bruno Fernandes
    Francisco Cruz
    [J]. Neural Computing and Applications, 2023, 35 : 16723 - 16738
  • [6] Environment Exploration for Mapless Navigation based on Deep Reinforcement Learning
    Toan, Nguyen Duc
    Gon-Woo, Kim
    [J]. 2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 17 - 20
  • [7] A Model of External Memory for Navigation in Partially Observable Visual Reinforcement Learning Tasks
    Smith, Robert J.
    Heywood, Malcolm I.
    [J]. GENETIC PROGRAMMING, EUROGP 2019, 2019, 11451 : 162 - 177
  • [8] Asynchronous Deep Reinforcement Learning for the Mobile Robot Navigation with Supervised Auxiliary Tasks
    Tongloy, T.
    Chuwongin, S.
    Jaksukam, K.
    Chousangsuntorn, C.
    Boonsang, S.
    [J]. 2017 2ND INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION ENGINEERING (ICRAE), 2017, : 68 - 72
  • [9] Task-oriented reinforcement learning for continuous tasks in dynamic environment
    Kamal, MAS
    Murata, J
    Hirasawa, K
    [J]. SICE 2002: PROCEEDINGS OF THE 41ST SICE ANNUAL CONFERENCE, VOLS 1-5, 2002, : 829 - 832
  • [10] A Deep Reinforcement Learning Environment for Particle Robot Navigation and Object Manipulation
    Shen, Jeremy
    Xiao, Erdong
    Liu, Yuchen
    Feng, Chen
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2022, 2022, : 6232 - 6239