Reinforcement learning in nonstationary environment navigation tasks

被引:0
|
作者
Lane, Terran [1 ]
Ridens, Martin [1 ]
Stevens, Scott [1 ]
机构
[1] Univ New Mexico, Dept Comp Sci, Albuquerque, NM 87131 USA
来源
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The field of reinforcement learning (RL) has achieved great strides in learning control knowledge from closed-loop interaction with environments. "Classical" RL, based on atomic state space representations, suffers from an inability to adapt to nonstationarities in the target Markov decision process (i.e., environment). Relational RL is widely seen as being a potential solution to this shortcoming. In this paper, we demonstrate a class of "pseudo-relational" learning methods for nonstationary navigational RL domains - domains in which the location of the goal, or even the structure of the environment, can change over time. Our approach is closely related to deictic representations, which have previously been found to be troublesome for RL. The key insight of this paper is that navigational problems are a highly constrained class of MDP, possessing a strong native topology that relaxes some of the partial observability difficulties arising from deixis. Agents can employ local information that is relevant to their near-term action choices to act effectively. We demonstrate that, unlike an atomic representation, our agents can learn to fluidly adapt to changing goal locations and environment structure.
引用
收藏
页码:429 / +
页数:3
相关论文
共 50 条
  • [21] Combining Motion Planner and Deep Reinforcement Learning for UAV Navigation in Unknown Environment
    Xue, Yuntao
    Chen, Weisheng
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (01) : 635 - 642
  • [22] Deep Reinforcement Learning-Based Automatic Exploration for Navigation in Unknown Environment
    Li, Haoran
    Zhang, Qichao
    Zhao, Dongbin
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (06) : 2064 - 2076
  • [23] Safe Reinforcement Learning of Dynamic High-Dimensional Robotic Tasks: Navigation, Manipulation, Interaction
    Liu, Puze
    Zhang, Kuo
    Tateo, Davide
    Jauhri, Snehal
    Hu, Zhiyuan
    Peters, Jan
    Chalvatzaki, Georgia
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 9449 - 9456
  • [24] Reinforcement learning with nonstationary reward depending on the episode
    Shibuya, Takeshi
    Yasunobu, Seiji
    [J]. 2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 2145 - 2150
  • [25] Nonstationary Reinforcement Learning: The Blessing of (More) Optimism
    Cheung, Wang Chi
    Simchi-Levi, David
    Zhu, Ruihao
    [J]. MANAGEMENT SCIENCE, 2023, 69 (10) : 5722 - 5739
  • [26] Automated Transfer for Reinforcement Learning Tasks
    Ammar, Haitham Bou
    Chen, Siqi
    Tuyls, Karl
    Weiss, Gerhard
    [J]. KUNSTLICHE INTELLIGENZ, 2014, 28 (01): : 7 - 14
  • [27] Strategic Tasks for Explainable Reinforcement Learning
    Pocius, Rey
    Neal, Lawrence
    Fern, Alan
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 10007 - 10008
  • [28] CoverNav: Cover Following Navigation Planning in Unstructured Outdoor Environment with Deep Reinforcement Learning
    Hossain, Jumman
    Faridee, Abu-Zaher
    Roy, Nirmalya
    Basak, Anjan
    Asher, Derrik E.
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON AUTONOMIC COMPUTING AND SELF-ORGANIZING SYSTEMS, ACSOS, 2023, : 127 - 132
  • [29] Multi-Agent Deep Reinforcement Learning for UAVs Navigation in Unknown Complex Environment
    Xue, Yuntao
    Chen, Weisheng
    [J]. IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (01): : 2290 - 2303
  • [30] LEARNING AND IMPRINTING IN STATIONARY AND NONSTATIONARY ENVIRONMENT
    PFAFFELHUBER, E
    DAMLE, PS
    [J]. KYBERNETIK, 1973, 13 (04): : 229 - 237