Trajectory Planning for Autonomous Vehicles Using Hierarchical Reinforcement Learning

被引:24
|
作者
Ben Naveed, Kaleb [1 ]
Qiao, Zhiqian [2 ]
Dolan, John M. [3 ]
机构
[1] Hong Kong Polytech Univ, Student Elect & Informat Engn, Hong Kong, Peoples R China
[2] Carnegie Mellon Univ, Elect & Comp Engn, Pittsburgh, PA 15213 USA
[3] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
关键词
Trajectory Planning; Hierarchical Deep Reinforcement Learning; Double Deep Q-Learning; PID controller;
D O I
10.1109/ITSC48978.2021.9564634
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Planning safe trajectories under uncertain and dynamic conditions makes the autonomous driving problem significantly complex. Current heuristic-based algorithms such as the slot-based method rely heavily on hand-engineered parameters and are restricted to specific scenarios. Supervised learning methods such as Imitation Learning lack generalization and safety guarantees. To address these problems and to ensure a robust framework, we propose a Robust-Hierarchical Reinforcement Learning (HRL) framework for learning autonomous driving policies. We adapt a state-of-the-art algorithm, Hierarchical Double Deep Q-learning (h-DDQN), and make the framework robust by (1) constituting the decision of selecting driving maneuver as a high-level option; (2) for the lower-level controller, outputting waypoint trajectories to track with a Proportional-Integral-Derivative (PID) controller instead of direct acceleration/steering actions; and (3) using a Long-Short-Term-Memory (LSTM) layer in the network to alleviate the effects of observation noise and dynamic driving behaviors. Moreover, to improve the sample efficiency, we use Hybrid Reward Mechanism and Reward-Driven Exploration. Results from the high-fidelity CARLA simulator while simulating different interactive lane change scenarios indicate that the proposed framework reduces convergence time, generates smoother trajectories, and can better handle dynamic surroundings and noisy observations as compared to other traditional RL approaches.
引用
收藏
页码:601 / 606
页数:6
相关论文
共 50 条
  • [31] Path planning of autonomous UAVs using reinforcement learning
    Chronis, Christos
    Anagnostopoulos, Georgios
    Politi, Elena
    Garyfallou, Antonios
    Varlamis, Iraklis
    Dimitrakopoulos, George
    [J]. 12TH EASN INTERNATIONAL CONFERENCE ON "INNOVATION IN AVIATION & SPACE FOR OPENING NEW HORIZONS", 2023, 2526
  • [32] Planning for Negotiations in Autonomous Driving using Reinforcement Learning
    Reshef, Roi
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 10595 - 10602
  • [33] Continuous advantage learning for minimum-time trajectory planning of autonomous vehicles
    Li, Zhuo
    Wu, Weiran
    Wang, Jialin
    Wang, Gang
    Sun, Jian
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (07)
  • [34] A Hierarchical Trajectory Planning Framework for Autonomous Driving
    Li, Jiangnan
    Gong, Jianwei
    Kong, Guojie
    Zhao, Yaogang
    Zhang, Xi
    [J]. PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), 2020, : 428 - 434
  • [35] Continuous advantage learning for minimum-time trajectory planning of autonomous vehicles
    Zhuo LI
    Weiran WU
    Jialin WANG
    Gang WANG
    Jian SUN
    [J]. Science China(Information Sciences), 2024, (07) - 294
  • [36] Trajectory Planning for Autonomous Vehicles combining Nonlinear Optimal Control and Supervised Learning
    Markolf, Lukas
    Eilbrecht, Jan
    Stursberg, Olaf
    [J]. IFAC PAPERSONLINE, 2020, 53 (02): : 15608 - 15614
  • [37] Path Planning for Autonomous Vehicles by Trajectory Smoothing Using Motion Primitives
    Bottasso, Carlo L.
    Leonello, Domenico
    Savini, Barbara
    [J]. IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2008, 16 (06) : 1152 - 1168
  • [38] Continuous advantage learning for minimum-time trajectory planning of autonomous vehicles
    Zhuo LI
    Weiran WU
    Jialin WANG
    Gang WANG
    Jian SUN
    [J]. Science China(Information Sciences)., 2024, 67 (07) - 294
  • [39] Vision-Based Trajectory Planning via Imitation Learning for Autonomous Vehicles
    Cai, Peide
    Sun, Yuxiang
    Chen, Yuying
    Liu, Ming
    [J]. 2019 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2019, : 2736 - 2742
  • [40] Analysis of Reinforcement Learning in Autonomous Vehicles
    Jebessa, Estephanos
    Olana, Kidus
    Getachew, Kidus
    Isteefanos, Stuart
    Mohd, Tauheed Khan
    [J]. 2022 IEEE 12TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE (CCWC), 2022, : 87 - 91