Trajectory Planning for Autonomous Vehicles Using Hierarchical Reinforcement Learning

被引:24
|
作者
Ben Naveed, Kaleb [1 ]
Qiao, Zhiqian [2 ]
Dolan, John M. [3 ]
机构
[1] Hong Kong Polytech Univ, Student Elect & Informat Engn, Hong Kong, Peoples R China
[2] Carnegie Mellon Univ, Elect & Comp Engn, Pittsburgh, PA 15213 USA
[3] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
关键词
Trajectory Planning; Hierarchical Deep Reinforcement Learning; Double Deep Q-Learning; PID controller;
D O I
10.1109/ITSC48978.2021.9564634
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Planning safe trajectories under uncertain and dynamic conditions makes the autonomous driving problem significantly complex. Current heuristic-based algorithms such as the slot-based method rely heavily on hand-engineered parameters and are restricted to specific scenarios. Supervised learning methods such as Imitation Learning lack generalization and safety guarantees. To address these problems and to ensure a robust framework, we propose a Robust-Hierarchical Reinforcement Learning (HRL) framework for learning autonomous driving policies. We adapt a state-of-the-art algorithm, Hierarchical Double Deep Q-learning (h-DDQN), and make the framework robust by (1) constituting the decision of selecting driving maneuver as a high-level option; (2) for the lower-level controller, outputting waypoint trajectories to track with a Proportional-Integral-Derivative (PID) controller instead of direct acceleration/steering actions; and (3) using a Long-Short-Term-Memory (LSTM) layer in the network to alleviate the effects of observation noise and dynamic driving behaviors. Moreover, to improve the sample efficiency, we use Hybrid Reward Mechanism and Reward-Driven Exploration. Results from the high-fidelity CARLA simulator while simulating different interactive lane change scenarios indicate that the proposed framework reduces convergence time, generates smoother trajectories, and can better handle dynamic surroundings and noisy observations as compared to other traditional RL approaches.
引用
收藏
页码:601 / 606
页数:6
相关论文
共 50 条
  • [1] Predictive trajectory planning for autonomous vehicles at intersections using reinforcement learning
    Zhang, Ethan
    Zhang, Ruixuan
    Masoud, Neda
    [J]. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2023, 149
  • [2] Hierarchical Reinforcement Learning for Autonomous Decision Making and Motion Planning of Intelligent Vehicles
    Lu, Yang
    Xu, Xin
    Zhang, Xinglong
    Qian, Lilin
    Zhou, Xing
    [J]. IEEE ACCESS, 2020, 8 : 209776 - 209789
  • [3] Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning
    You, Changxi
    Lu, Jianbo
    Filev, Dimitar
    Tsiotras, Panagiotis
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 114 : 1 - 18
  • [4] Trajectory tracking algorithm for autonomous vehicles using adaptive reinforcement learning
    De Paula, Mariano
    Acosta, Gerardo G.
    [J]. OCEANS 2015 - MTS/IEEE WASHINGTON, 2015,
  • [5] Trajectory Planning for Hypersonic Vehicles with Reinforcement Learning
    Chi, Haihong
    Thou, Mingxin
    [J]. 2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 3721 - 3726
  • [6] Obstacle avoidance planning of autonomous vehicles using deep reinforcement learning
    Qian, Yubin
    Feng, Song
    Hu, Wenhao
    Wang, Wanqiu
    [J]. ADVANCES IN MECHANICAL ENGINEERING, 2022, 14 (12)
  • [7] Motion Planning for Autonomous Vehicles in the Presence of Uncertainty Using Reinforcement Learning
    Rezaee, Kasra
    Yadmellat, Peyman
    Chamorro, Simon
    [J]. 2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 3506 - 3511
  • [8] Autonomous Vehicles Roundup Strategy by Reinforcement Learning with Prediction Trajectory
    Ni, Jiayang
    Ma, Rubing
    Zhong, Hua
    Wang, Bo
    [J]. 2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 3370 - 3375
  • [9] Hierarchical dynamic trajectory planning for autonomous underwater vehicles: Algorithms and experiments
    Liu, Guoshun
    Zheng, Huarong
    Liu, Shuo
    Shen, Binjian
    Xu, Wen
    [J]. OCEAN ENGINEERING, 2024, 307
  • [10] Hierarchical Reinforcement Learning using Gaussian Random Trajectory Generation in Autonomous Furniture Assembly
    Yun, Won Joon
    Mohaisen, David
    Jung, Soyi
    Kim, Jong-Kook
    Kim, Joongheon
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3624 - 3633