Safe reinforcement learning for high-speed autonomous racing

被引:0
|
作者
Evans B.D. [1 ]
Jordaan H.W. [1 ]
Engelbrecht H.A. [1 ]
机构
[1] Stellenbosch University, Electrical and Electronic Engineering, Stellenbosch, Banghoek Road
来源
Cognitive Robotics | 2023年 / 3卷
关键词
Autonomous racing; Reinforcement learning; Safe autonomous systems; Safe learning;
D O I
10.1016/j.cogr.2023.04.002
中图分类号
学科分类号
摘要
The conventional application of deep reinforcement learning (DRL) to autonomous racing requires the agent to crash during training, thus limiting training to simulation environments. Further, many DRL approaches still exhibit high crash rates after training, making them infeasible for real-world use. This paper addresses the problem of safely training DRL agents for autonomous racing. Firstly, we present a Viability Theory-based supervisor that ensures the vehicle does not crash and remains within the friction limit while maintaining recursive feasibility. Secondly, we use the supervisor to ensure the vehicle does not crash during the training of DRL agents for high-speed racing. The evaluation in the open-source F1Tenth simulator demonstrates that our safety system can ensure the safety of a worst-case scenario planner on four test maps up to speeds of 6 m/s. Training agents to race with the supervisor significantly improves sample efficiency, requiring only 10,000 steps. Our learning formulation leads to learning more conservative, safer policies with slower lap times and a higher success rate, resulting in our method being feasible for physical vehicle racing. Enabling DRL agents to learn to race without ever crashing is a step towards using DRL on physical vehicles. © 2023 The Authors
引用
收藏
页码:107 / 126
页数:19
相关论文
共 50 条
  • [31] High-speed quadrupedal locomotion by imitation-relaxation reinforcement learning
    Yongbin Jin
    Xianwei Liu
    Yecheng Shao
    Hongtao Wang
    Wei Yang
    Nature Machine Intelligence, 2022, 4 : 1198 - 1208
  • [32] Safe odometry for high-speed trains
    Murillas, Diego Herrero
    Poncet, Luc
    2016 IEEE INTERNATIONAL CONFERENCE ON INTELLIGENT RAIL TRANSPORTATION (ICIRT), 2016, : 244 - 248
  • [33] Reinforcement of high-speed steel ingots
    Baranov, A.A.
    Lyubchenko, A.G.
    Pashinskij, V.V.
    Antonov, V.V.
    Liteinoe Proizvodstvo, 1991, (05): : 34 - 35
  • [34] Reinforcement Learning-based Optimization of Speed Profile for High-speed Train with Temporary Speed Restriction
    Zhou M.
    Dong H.
    Zhou X.
    Xu W.
    Ning L.
    Tiedao Xuebao/Journal of the China Railway Society, 2023, 45 (02): : 84 - 92
  • [35] Autonomous multi-drone racing method based on deep reinforcement learning
    Kang, Yu
    Di, Jian
    Li, Ming
    Zhao, Yunbo
    Wang, Yuhui
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (08)
  • [36] Autonomous multi-drone racing method based on deep reinforcement learning
    Yu KANG
    Jian DI
    Ming LI
    Yunbo ZHAO
    Yuhui WANG
    Science China(Information Sciences), 2024, 67 (08) : 35 - 48
  • [37] Zero-Shot Policy Transfer in Autonomous Racing: Reinforcement Learning vs Imitation Learning
    Hamilton, Nathaniel
    Musau, Patrick
    Lopez, Diego Manzanas
    Johnson, Taylor T.
    2022 IEEE INTERNATIONAL CONFERENCE ON ASSURED AUTONOMY (ICAA 2022), 2022, : 11 - 20
  • [38] Safe Reinforcement Learning in Autonomous Driving With Epistemic Uncertainty Estimation
    Zhang, Zheng
    Liu, Qi
    Li, Yanjie
    Lin, Ke
    Li, Linyu
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (10) : 13653 - 13666
  • [39] High-speed exercise history and catastrophic racing fracture in Thoroughbreds
    Estberg, L
    Stover, SM
    Gardner, IA
    Drake, CM
    Johnson, B
    Ardans, A
    AMERICAN JOURNAL OF VETERINARY RESEARCH, 1996, 57 (11) : 1549 - 1555
  • [40] Robust High-Speed Running for Quadruped Robots via Deep Reinforcement Learning
    Bellegarda, Guillaume
    Chen, Yiyu
    Liu, Zhuochen
    Quan Nguyen
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 10364 - 10370