Design and Experimental Validation of Deep Reinforcement Learning-Based Fast Trajectory Planning and Control for Mobile Robot in Unknown Environment

被引:75
|
作者
Chai, Runqi [1 ,2 ]
Niu, Hanlin [1 ]
Carrasco, Joaquin [1 ]
Arvin, Farshad [3 ]
Yin, Hujun [1 ]
Lennox, Barry [1 ]
机构
[1] Univ Manchester, Dept Elect & Elect Engn, Manchester M13 9PL, Lancs, England
[2] Beijing Inst Technol, Sch Automat, Beijing 100081, Peoples R China
[3] Univ Durham, Dept Comp Sci, Durham DH1 3LE, England
基金
英国工程与自然科学研究理事会;
关键词
Mobile robots; Trajectory; Planning; Collision avoidance; Training; Robot sensing systems; Noise measurement; Deep reinforcement learning (DRL); mobile robot; motion control; noisy prioritized experience replay (PER); optimal motion planning; recurrent neural network; unexpected obstacles; ROBUST; IMPLEMENTATION; VEHICLES; ASTERISK;
D O I
10.1109/TNNLS.2022.3209154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article is concerned with the problem of planning optimal maneuver trajectories and guiding the mobile robot toward target positions in uncertain environments for exploration purposes. A hierarchical deep learning-based control framework is proposed which consists of an upper level motion planning layer and a lower level waypoint tracking layer. In the motion planning phase, a recurrent deep neural network (RDNN)-based algorithm is adopted to predict the optimal maneuver profiles for the mobile robot. This approach is built upon a recently proposed idea of using deep neural networks (DNNs) to approximate the optimal motion trajectories, which has been validated that a fast approximation performance can be achieved. To further enhance the network prediction performance, a recurrent network model capable of fully exploiting the inherent relationship between preoptimized system state and control pairs is advocated. In the lower level, a deep reinforcement learning (DRL)-based collision-free control algorithm is established to achieve the waypoint tracking task in an uncertain environment (e.g., the existence of unexpected obstacles). Since this approach allows the control policy to directly learn from human demonstration data, the time required by the training process can be significantly reduced. Moreover, a noisy prioritized experience replay (PER) algorithm is proposed to improve the exploring rate of control policy. The effectiveness of applying the proposed deep learning-based control is validated by executing a number of simulation and experimental case studies. The simulation result shows that the proposed DRL method outperforms the vanilla PER algorithm in terms of training speed. Experimental videos are also uploaded, and the corresponding results confirm that the proposed strategy is able to fulfill the autonomous exploration mission with improved motion planning performance, enhanced collision avoidance ability, and less training time.
引用
收藏
页码:5778 / 5792
页数:15
相关论文
共 50 条
  • [41] Multi-Agent Deep Reinforcement Learning-Based Trajectory Planning for Multi-UAV Assisted Mobile Edge Computing
    Wang, Liang
    Wang, Kezhi
    Pan, Cunhua
    Xu, Wei
    Aslam, Nauman
    Hanzo, Lajos
    [J]. IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2021, 7 (01) : 73 - 84
  • [42] Mobile Robot Navigation based on Deep Reinforcement Learning
    Ruan, Xiaogang
    Ren, Dingqi
    Zhu, Xiaoqing
    Huang, Jing
    [J]. PROCEEDINGS OF THE 2019 31ST CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2019), 2019, : 6174 - 6178
  • [43] Deep Reinforcement Learning-Based Distributed 3D UAV Trajectory Design
    He, Huasen
    Yuan, Wenke
    Chen, Shuangwu
    Jiang, Xiaofeng
    Yang, Feng
    Yang, Jian
    [J]. IEEE TRANSACTIONS ON COMMUNICATIONS, 2024, 72 (06) : 3736 - 3751
  • [44] An Efficiently Convergent Deep Reinforcement Learning-Based Trajectory Planning Method for Manipulators in Dynamic Environments
    Li Zheng
    YaHao Wang
    Run Yang
    Shaolei Wu
    Rui Guo
    Erbao Dong
    [J]. Journal of Intelligent & Robotic Systems, 2023, 107
  • [45] Deep Reinforcement Learning-Based 3D Trajectory Planning for Cellular Connected UAV
    Liu, Xiang
    Zhong, Weizhi
    Wang, Xin
    Duan, Hongtao
    Fan, Zhenxiong
    Jin, Haowen
    Huang, Yang
    Lin, Zhipeng
    [J]. DRONES, 2024, 8 (05)
  • [46] Robot path planning based on deep reinforcement learning
    Long, Yinxin
    He, Huajin
    [J]. 2020 IEEE CONFERENCE ON TELECOMMUNICATIONS, OPTICS AND COMPUTER SCIENCE (TOCS), 2020, : 151 - 154
  • [47] An Efficiently Convergent Deep Reinforcement Learning-Based Trajectory Planning Method for Manipulators in Dynamic Environments
    Zheng, Li
    Wang, YaHao
    Yang, Run
    Wu, Shaolei
    Guo, Rui
    Dong, Erbao
    [J]. JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2023, 107 (04)
  • [48] MOBILE ROBOT MOVEMENT PLANNING AND CONTROL INTELLIGENT SYSTEM IN UNKNOWN ENVIRONMENT
    Nemeiksis, Andrius
    Osadcuks, Vitalijs
    [J]. 17TH INTERNATIONAL SCIENTIFIC CONFERENCE: ENGINEERING FOR RURAL DEVELOPMENT, 2018, : 948 - 957
  • [49] Causal deconfounding deep reinforcement learning for mobile robot motion planning
    Tang, Wenbing
    Wu, Fenghua
    Lin, Shang-wei
    Ding, Zuohua
    Liu, Jing
    Liu, Yang
    He, Jifeng
    [J]. KNOWLEDGE-BASED SYSTEMS, 2024, 303
  • [50] A Review of Deep Reinforcement Learning Algorithms for Mobile Robot Path Planning
    Singh, Ramanjeet
    Ren, Jing
    Lin, Xianke
    [J]. VEHICLES, 2023, 5 (04): : 1423 - 1451