Memory-based reinforcement learning algorithm for autonomous exploration in unknown environment

被引:6
|
作者
Dooraki, Amir Ramezani [1 ]
Lee, Deok Jin [1 ]
机构
[1] Kunsan Natl Univ, Smart Autonomous Syst Lab, Gunsan 54150, South Korea
来源
基金
新加坡国家研究基金会;
关键词
Reinforcement learning; autonomous exploration; adaptive agent; depth map; artificial neural network; sensor fusion; memory-based; obstacle avoidance; ROBOT;
D O I
10.1177/1729881418775849
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In the near future, robots would be seen in almost every area of our life, in different shapes and with different objectives such as entertainment, surveillance, rescue, and navigation. In any shape and with any objective, it is necessary for them to be capable of successful exploration. They should be able to explore efficiently and be able to adapt themselves with changes in their environment. For successful navigation, it is necessary to recognize the difference between similar places of an environment. In order to achieve this goal without increasing the capability of sensors, having a memory is crucial. In this article, an algorithm for autonomous exploration and obstacle avoidance in an unknown environment is proposed. In order to make our self-learner algorithm, a memory-based reinforcement learning method using multilayer neural network is used with the aim of creating an agent having an efficient exploration and obstacle avoidance policy. Furthermore, this agent can automatically adapt itself to the changes of its environment. Finally, in order to test the capability of our algorithm, we have implemented it in a robot similar to a real model, simulated in the robust physics engine simulator of Gazebo.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Dynamic Memory-Based Curiosity: A Bootstrap Approach for Exploration in Reinforcement Learning
    Gao, Zijian
    Li, Yiying
    Xu, Kele
    Zhai, Yuanzhao
    Ding, Bo
    Feng, Dawei
    Mao, Xinjun
    Wang, Huaimin
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (02): : 1181 - 1193
  • [2] Memory-Based Explainable Reinforcement Learning
    Cruz, Francisco
    Dazeley, Richard
    Vamplew, Peter
    AI 2019: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, 11919 : 66 - 77
  • [3] Hierarchical memory-based reinforcement learning
    Hernandez-Gardiol, N
    Mahadevan, S
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 13, 2001, 13 : 1047 - 1053
  • [4] Multi-AGV's Temporal Memory-Based RRT Exploration in Unknown Environment
    Lau, Billy Pik Lik
    Ong, Brandon Jin Yang
    Loh, Leonard Kin Yung
    Liu, Ran
    Yuen, Chau
    Soh, Gim Song
    Tan, U-Xuan
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04) : 9256 - 9263
  • [5] Memory-based Deep Reinforcement Learning for POMDPs
    Meng, Lingheng
    Gorbet, Rob
    Kulic, Dana
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 5619 - 5626
  • [6] A Memory-based Reinforcement Learning Algorithm for Partially Observable Markovian Decision Processes
    Zheng, Lei
    Cho, Siu-Yeung
    Quek, Chai
    2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 800 - 805
  • [7] Deep Reinforcement Learning-based ROS-Controlled RC Car for Autonomous Path Exploration in the Unknown Environment
    Hossain, Sabir
    Doukhi, Oualid
    Jo, Yeonho
    Lee, Deok-Jin
    2020 20TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2020, : 1231 - 1236
  • [8] Deep Reinforcement Learning-Based Automatic Exploration for Navigation in Unknown Environment
    Li, Haoran
    Zhang, Qichao
    Zhao, Dongbin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (06) : 2064 - 2076
  • [9] Memory-Based Deep Reinforcement Learning for Obstacle Avoidance in UAV With Limited Environment Knowledge
    Singla, Abhik
    Padakandla, Sindhu
    Bhatnagar, Shalabh
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (01) : 107 - 118
  • [10] Perspective view of autonomous control in unknown environment: Dual control for exploitation and exploration vs reinforcement learning
    Chen, Wen-Hua
    NEUROCOMPUTING, 2022, 497 : 50 - 63