Safe reinforcement learning for high-speed autonomous racing

被引:0
|
作者
Evans B.D. [1 ]
Jordaan H.W. [1 ]
Engelbrecht H.A. [1 ]
机构
[1] Stellenbosch University, Electrical and Electronic Engineering, Stellenbosch, Banghoek Road
来源
Cognitive Robotics | 2023年 / 3卷
关键词
Autonomous racing; Reinforcement learning; Safe autonomous systems; Safe learning;
D O I
10.1016/j.cogr.2023.04.002
中图分类号
学科分类号
摘要
The conventional application of deep reinforcement learning (DRL) to autonomous racing requires the agent to crash during training, thus limiting training to simulation environments. Further, many DRL approaches still exhibit high crash rates after training, making them infeasible for real-world use. This paper addresses the problem of safely training DRL agents for autonomous racing. Firstly, we present a Viability Theory-based supervisor that ensures the vehicle does not crash and remains within the friction limit while maintaining recursive feasibility. Secondly, we use the supervisor to ensure the vehicle does not crash during the training of DRL agents for high-speed racing. The evaluation in the open-source F1Tenth simulator demonstrates that our safety system can ensure the safety of a worst-case scenario planner on four test maps up to speeds of 6 m/s. Training agents to race with the supervisor significantly improves sample efficiency, requiring only 10,000 steps. Our learning formulation leads to learning more conservative, safer policies with slower lap times and a higher success rate, resulting in our method being feasible for physical vehicle racing. Enabling DRL agents to learn to race without ever crashing is a step towards using DRL on physical vehicles. © 2023 The Authors
引用
收藏
页码:107 / 126
页数:19
相关论文
共 50 条
  • [21] Application of Reinforcement Learning on High-Speed Rail Cognitive Radio
    Wu, Qing-ting
    Wu, Cheng
    Wang, Yi-ming
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE: TECHNIQUES AND APPLICATIONS, AITA 2016, 2016, : 332 - 336
  • [22] Bayesian Learning for Safe High-Speed Navigation in Unknown Environments
    Richter, Charles
    Vega-Brown, William
    Roy, Nicholas
    ROBOTICS RESEARCH, VOL 2, 2018, 3 : 325 - 341
  • [23] Balanced Reward-inspired Reinforcement Learning for Autonomous Vehicle Racing
    Tian, Zhen
    Zhao, Dezong
    Lin, Zhihao
    Flynn, David
    Zhao, Wenjing
    Tian, Daxin
    6TH ANNUAL LEARNING FOR DYNAMICS & CONTROL CONFERENCE, 2024, 242 : 628 - 640
  • [24] Reaching the limit in autonomous racing: Optimal control versus reinforcement learning
    Song, Yunlong
    Romero, Angel
    Muller, Matthias
    Koltun, Vladlen
    Scaramuzza, Davide
    SCIENCE ROBOTICS, 2023, 8 (82)
  • [25] Autonomous Car Racing in Simulation Environment Using Deep Reinforcement Learning
    Guckiran, Kivanc
    Bolat, Bulent
    2019 INNOVATIONS IN INTELLIGENT SYSTEMS AND APPLICATIONS CONFERENCE (ASYU), 2019, : 329 - 334
  • [26] High-speed quadrupedal locomotion by imitation-relaxation reinforcement learning
    Jin, Yongbin
    Liu, Xianwei
    Shao, Yecheng
    Wang, Hongtao
    Yang, Wei
    NATURE MACHINE INTELLIGENCE, 2022, 4 (12) : 1198 - 1208
  • [27] A reinforcement learning approach to congestion control of high-speed multimedia networks
    Shaio, MC
    Tan, SW
    Hwang, KS
    Wu, CS
    CYBERNETICS AND SYSTEMS, 2005, 36 (02) : 181 - 202
  • [28] A Deep Reinforcement Learning Approach for the Traffic Management of High-Speed Railways
    Wu, Wei
    Yin, Jiateng
    Pu, Fan
    Su, Shuai
    Tang, Tao
    2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2021, : 2368 - 2373
  • [29] Association of high-speed exercise with racing injury in Thoroughbreds
    Cohen, ND
    Berry, SM
    Peloso, JG
    Mundy, GD
    Howard, IC
    JOURNAL OF THE AMERICAN VETERINARY MEDICAL ASSOCIATION, 2000, 216 (08): : 1273 - 1278
  • [30] An Intermittent Learning Algorithm for High-Speed Autonomous Driving in Unknown Environments
    Gundu, Pavan K.
    Vamvoudakis, Kyriakos G.
    Gerdes, Ryan M.
    2019 AMERICAN CONTROL CONFERENCE (ACC), 2019, : 4286 - 4292