Deep Reinforcement Learning-Based Large-Scale Robot Exploration

被引:0
|
作者
Cao, Yuhong [1 ]
Zhao, Rui [1 ]
Wang, Yizhuo [1 ]
Xiang, Bairan [1 ]
Sartoretti, Guillaume [1 ]
机构
[1] Natl Univ Singapore, Coll Design & Engn, Dept Mech Engn, Singapore 117482, Singapore
关键词
View Planning for SLAM; reinforcement learning; motion and path planning; AUTONOMOUS EXPLORATION; EFFICIENT;
D O I
10.1109/LRA.2024.3379804
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In this work, we propose a deep reinforcement learning (DRL) based reactive planner to solve large-scale Lidar-based autonomous robot exploration problems in 2D action space. Our DRL-based planner allows the agent to reactively plan its exploration path by making implicit predictions about unknown areas, based on a learned estimation of the underlying transition model of the environment. To this end, our approach relies on learned attention mechanisms for their powerful ability to capture long-term dependencies at different spatial scales to reason about the robot's entire belief over known areas. Our approach relies on ground truth information (i.e., privileged learning) to guide the environment estimation during training, as well as on a graph rarefaction algorithm, which allows models trained in small-scale environments to scale to large-scale ones. Simulation results show that our model exhibits better exploration efficiency (12% in path length, 6% in makespan) and lower planning time (60%) than the state-of-the-art planners in a 130 m x 100 m benchmark scenario. We also validate our learned model on hardware.
引用
下载
收藏
页码:4631 / 4638
页数:8
相关论文
共 50 条
  • [21] A LARGE-SCALE PATH PLANNING ALGORITHM FOR UNDERWATER ROBOTS BASED ON DEEP REINFORCEMENT LEARNING
    Wang, Wenhui
    Li, Leqing
    Ye, Fumeng
    Peng, Yumin
    Ma, Yiming
    INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2024, 39 (03): : 204 - 210
  • [22] An Attention Reinforcement Learning-Based Strategy for Large-Scale Adaptive Traffic Signal Control System
    Han, Gengyue
    Liu, Xiaohan
    Wang, Hao
    Dong, Changyin
    Han, Yu
    JOURNAL OF TRANSPORTATION ENGINEERING PART A-SYSTEMS, 2024, 150 (03)
  • [23] Deep Learning-Based Symbol-Level Precoding for Large-Scale Antenna System
    Xie, Changxu
    Du, Huiqin
    Liu, Xialing
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2021, 2021
  • [24] Deep Reinforcement Learning for Stabilization of Large-Scale Probabilistic Boolean Networks
    Moschoyiannis, Sotiris
    Chatzaroulas, Evangelos
    Sliogeris, Vytenis
    Wu, Yuhu
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2023, 10 (03): : 1412 - 1423
  • [25] Planning Large-scale Object Rearrangement Using Deep Reinforcement Learning
    Ghosh, Sourav
    Das, Dipanjan
    Chakraborty, Abhishek
    Agarwal, Marichi
    Bhowmick, Brojeshwar
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [26] Deep reinforcement learning for scheduling in large-scale networked control systems
    Redder, Adrian
    Ramaswamy, Arunselvan
    Quevedo, Daniel E.
    IFAC PAPERSONLINE, 2019, 52 (20): : 333 - 338
  • [27] Deep Reinforcement Learning for Network Service Recovery in Large-scale Failures
    Akashi, Kazuaki
    Fukuda, Nobukazu
    Kanai, Shunsuke
    Tayama, Kenichi
    2023 19TH INTERNATIONAL CONFERENCE ON NETWORK AND SERVICE MANAGEMENT, CNSM, 2023,
  • [28] Large-Scale and Adaptive Service Composition Using Deep Reinforcement Learning
    Wang, Hongbing
    Gu, Mingzhu
    Yu, Qi
    Fei, Huanhuan
    Li, Jiajie
    Tao, Yong
    SERVICE-ORIENTED COMPUTING, ICSOC 2017, 2017, 10601 : 383 - 391
  • [29] An efficient evolutionary algorithm based on deep reinforcement learning for large-scale sparse multiobjective optimization
    Gao, Mengqi
    Feng, Xiang
    Yu, Huiqun
    Li, Xiuquan
    APPLIED INTELLIGENCE, 2023, 53 (18) : 21116 - 21139
  • [30] Coordinated Charging Strategy Applicable to Large-scale Charging Stations Based on Deep Reinforcement Learning
    Chen G.
    Wang X.
    Yuan S.
    Shuai X.
    Zhou Q.
    Dianli Xitong Zidonghua/Automation of Electric Power Systems, 2023, 47 (02): : 88 - 95