Design and Experimental Validation of Deep Reinforcement Learning-Based Fast Trajectory Planning and Control for Mobile Robot in Unknown Environment

被引:75
|
作者
Chai, Runqi [1 ,2 ]
Niu, Hanlin [1 ]
Carrasco, Joaquin [1 ]
Arvin, Farshad [3 ]
Yin, Hujun [1 ]
Lennox, Barry [1 ]
机构
[1] Univ Manchester, Dept Elect & Elect Engn, Manchester M13 9PL, Lancs, England
[2] Beijing Inst Technol, Sch Automat, Beijing 100081, Peoples R China
[3] Univ Durham, Dept Comp Sci, Durham DH1 3LE, England
基金
英国工程与自然科学研究理事会;
关键词
Mobile robots; Trajectory; Planning; Collision avoidance; Training; Robot sensing systems; Noise measurement; Deep reinforcement learning (DRL); mobile robot; motion control; noisy prioritized experience replay (PER); optimal motion planning; recurrent neural network; unexpected obstacles; ROBUST; IMPLEMENTATION; VEHICLES; ASTERISK;
D O I
10.1109/TNNLS.2022.3209154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article is concerned with the problem of planning optimal maneuver trajectories and guiding the mobile robot toward target positions in uncertain environments for exploration purposes. A hierarchical deep learning-based control framework is proposed which consists of an upper level motion planning layer and a lower level waypoint tracking layer. In the motion planning phase, a recurrent deep neural network (RDNN)-based algorithm is adopted to predict the optimal maneuver profiles for the mobile robot. This approach is built upon a recently proposed idea of using deep neural networks (DNNs) to approximate the optimal motion trajectories, which has been validated that a fast approximation performance can be achieved. To further enhance the network prediction performance, a recurrent network model capable of fully exploiting the inherent relationship between preoptimized system state and control pairs is advocated. In the lower level, a deep reinforcement learning (DRL)-based collision-free control algorithm is established to achieve the waypoint tracking task in an uncertain environment (e.g., the existence of unexpected obstacles). Since this approach allows the control policy to directly learn from human demonstration data, the time required by the training process can be significantly reduced. Moreover, a noisy prioritized experience replay (PER) algorithm is proposed to improve the exploring rate of control policy. The effectiveness of applying the proposed deep learning-based control is validated by executing a number of simulation and experimental case studies. The simulation result shows that the proposed DRL method outperforms the vanilla PER algorithm in terms of training speed. Experimental videos are also uploaded, and the corresponding results confirm that the proposed strategy is able to fulfill the autonomous exploration mission with improved motion planning performance, enhanced collision avoidance ability, and less training time.
引用
收藏
页码:5778 / 5792
页数:15
相关论文
共 50 条
  • [1] Immune deep reinforcement learning-based path planning for mobile robot in unknown environment
    Yan, Chengliang
    Chen, Guangzhu
    Li, Yang
    Sun, Fuchun
    Wu, Yuanyuan
    [J]. APPLIED SOFT COMPUTING, 2023, 145
  • [2] Deep Reinforcement Learning-Based Robot Exploration for Constructing Map of Unknown Environment
    Shih-Yeh Chen
    Qi-Fong He
    Chin-Feng Lai
    [J]. Information Systems Frontiers, 2024, 26 : 63 - 74
  • [3] Deep Reinforcement Learning-Based Robot Exploration for Constructing Map of Unknown Environment
    Chen, Shih-Yeh
    He, Qi-Fong
    Lai, Chin-Feng
    [J]. INFORMATION SYSTEMS FRONTIERS, 2021, 26 (1) : 63 - 74
  • [4] Path Planning of Autonomous Mobile Robot in Comprehensive Unknown Environment Using Deep Reinforcement Learning
    Bai, Zekun
    Pang, Hui
    He, Zhaonian
    Zhao, Bin
    Wang, Tong
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (12): : 22153 - 22166
  • [5] Deep reinforcement learning-based rehabilitation robot trajectory planning with optimized reward functions
    Wang, Xusheng
    Xie, Jiexin
    Guo, Shijie
    Li, Yue
    Sun, Pengfei
    Gan, Zhongxue
    [J]. ADVANCES IN MECHANICAL ENGINEERING, 2021, 13 (12)
  • [6] Correction to: Deep Reinforcement Learning-Based Robot Exploration for Constructing Map of Unknown Environment
    Shih-Yeh Chen
    Qi-Fong He
    Chin-Feng Lai
    [J]. Information Systems Frontiers, 2023, 25 : 2115 - 2115
  • [7] Q -learning-based Collision -free Path Planning for Mobile Robot in Unknown Environment
    Wang, Yuxiang
    Wang, Shuting
    Xie, Yuanlong
    Hu, Yiming
    Li, Hu
    [J]. 2022 IEEE 17TH CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA), 2022, : 1104 - 1109
  • [8] Deep Learning-based Algorithm for Mobile Robot Control in Textureless Environment
    Petrovic, Milica
    Mystkowski, Arkadiusz
    Jokic, Aleksandar
    Dokic, Lazar
    Miljkovic, Zoran
    [J]. 15TH INTERNATIONAL CONFERENCE MECHATRONIC SYSTEMS AND MATERIALS, MSM'20, 2020, : 143 - 146
  • [9] TRAJECTORY PLANNING OF MOBILE ROBOT MOVEMENT IN UNKNOWN ENVIRONMENT
    Nemeiksis, Andrius
    Osadcuks, Vitalijs
    [J]. 16TH INTERNATIONAL SCIENTIFIC CONFERENCE: ENGINEERING FOR RURAL DEVELOPMENT, 2017, : 1157 - 1166
  • [10] Path Planning for Mobile Robot Based on Deep Reinforcement Learning and Fuzzy Control
    Liu, Chunling
    Xu, Jun
    Guo, Kaiwen
    [J]. 2022 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, COMPUTER VISION AND MACHINE LEARNING (ICICML), 2022, : 533 - 537