Path planning via reinforcement learning with closed-loop motion control and field tests

被引:0
|
作者
Feher, Arpad [1 ]
Domina, Adam [2 ]
Bardos, Adam [2 ]
Aradi, Szilard [1 ]
Becsi, Tamas [1 ]
机构
[1] Budapest Univ Technol & Econ, Fac Transportat Engn & Vehicle Engn, Dept Control Transportat & Vehicle Syst, Muegyet Rkp 3, H-1111 Budapest, Hungary
[2] Budapest Univ Technol & Econ, Dept Automot Technol, Fac Transportat Engn & Vehicle Engn, Muegyetem Rkp 3, H-1111 Budapest, Hungary
关键词
Vehicle dynamics; Advanced driver assistance systems; Machine learning; Reinforcement learning; Model predictive control; ACTIVE STEERING CONTROL; MODEL; SIMULATION; VEHICLES;
D O I
10.1016/j.engappai.2024.109870
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Performing evasive maneuvers with highly automated vehicles is a challenging task. The algorithm must fulfill safety constraints and complete the task while keeping the car in a controllable state. Furthermore, considering all aspects of vehicle dynamics, the path generation problem is numerically complex. Hence its classical solutions can hardly meet real-time requirements. On the other hand, single reinforcement learning based approaches only could handle this problem as a simple driving task and would not provide feasibility information on the whole task's horizon. Therefore, this paper presents a hierarchical method for obstacle avoidance of an automated vehicle to overcome this issue, where the geometric path generation is provided by a single-step continuous Reinforcement Learning agent, while a model-predictive controller deals with lateral control to perform a double lane change maneuver. As the agent plays the optimization role in this architecture, it is trained in various scenarios to provide the necessary parameters fora geometric path generator in a onestep neural network output. During the training, the controller that follows the track evaluates the feasibility of the generated path whose performance metrics provide feedback to the agent so it can further improve its performance. The framework can train an agent fora given problem with various parameters. Asa use case, it is presented as a static obstacle avoidance maneuver. the proposed framework was tested on an automotive proving ground with the geometric constraints of the ISO-3888-2 test. The results proved its real-time capability and performance compared to human drivers' abilities.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Learning of Closed-Loop Motion Control
    Farshidian, Farbod
    Neunert, Michael
    Buchli, Jonas
    2014 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2014), 2014, : 1441 - 1446
  • [2] Closed-Loop Control of Direct Ink Writing via Reinforcement Learning
    Piovarci, Michal
    Foshey, Michael
    Xu, Jie
    Erps, Timmothy
    Babaei, Vahid
    Didyk, Piotr
    Rusinkiewicz, Szymon
    Matusik, Wojciech
    Bickel, Bernd
    ACM TRANSACTIONS ON GRAPHICS, 2022, 41 (04):
  • [3] Closed-loop control of a noisy qubit with reinforcement learning
    Ding, Yongcheng
    Chen, Xi
    Magdalena-Benedito, Rafael
    Martin-Guerrero, Jose D.
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2023, 4 (02):
  • [4] Deep Reinforcement Learning for Closed-Loop Blood Glucose Control
    Fox, Ian
    Lee, Joyce
    Pop-Busui, Rodica
    Wiens, Jenna
    MACHINE LEARNING FOR HEALTHCARE CONFERENCE, VOL 126, 2020, 126 : 508 - 535
  • [5] Closed-Loop Control of Fluid Resuscitation Using Reinforcement Learning
    Estiri, Elham
    Mirinejad, Hossein
    IEEE ACCESS, 2023, 11 : 140569 - 140581
  • [6] Reinforcement learning fbr mixed open-loop and closed-loop control
    Hansen, EA
    Barto, AG
    Zilberstein, S
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 9: PROCEEDINGS OF THE 1996 CONFERENCE, 1997, 9 : 1026 - 1032
  • [7] A MICROCONTROLLER FOR CLOSED-LOOP MOTION CONTROL
    DALAY, BS
    PARKIN, RM
    MICROPROCESSORS AND MICROSYSTEMS, 1991, 15 (09) : 473 - 480
  • [8] Closed-loop control of bevel-tip needles based on path planning
    Huo, Benyan
    Zhao, Xingang
    Han, Jianda
    Xu, Weiliang
    ROBOTICA, 2018, 36 (12) : 1857 - 1873
  • [9] Closed-Loop Modeling of a Quintic Trajectory Planning and PID Motion Control System
    Awathe, Arpit
    Varunjikar, Tejas
    2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 1713 - 1720
  • [10] Application of Reinforcement Learning to Electrical Power System Closed-Loop Emergency Control
    Druet, C.
    Ernest, D.
    Wehenkel, L.
    LECTURE NOTES IN COMPUTER SCIENCE <D>, 2000, 1910 : 86 - 95