Reducing Safety Interventions in Provably Safe Reinforcement Learning

被引:0
|
作者
Thumm, Jakob [1 ]
Pelat, Guillaume [1 ]
Althoff, Matthias [1 ]
机构
[1] Tech Univ Munich, Sch Informat, D-85748 Garching, Germany
基金
欧盟地平线“2020”;
关键词
OPTIMIZATION;
D O I
10.1109/IROS55552.2023.10342464
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Reinforcement Learning (RL) has shown promise in addressing complex robotic challenges. In real-world applications, RL is often accompanied by failsafe controllers as a last resort to avoid catastrophic events. While necessary for safety, these interventions can result in undesirable behaviors, such as abrupt braking or aggressive steering. This paper proposes two safety intervention reduction methods: proactive replacement and proactive projection, which change the action of the agent if it leads to a potential failsafe intervention. These approaches are compared to state-of-the-art constrained RL on the OpenAI safety gym benchmark and a human-robot collaboration task. Our study demonstrates that the combination of our method with provably safe RL leads to high-performing policies with zero safety violations and a low number of failsafe interventions. Our versatile method can be applied to a wide range of real-world robotic tasks, while effectively improving safety without sacrificing task performance.
引用
收藏
页码:7515 / 7522
页数:8
相关论文
共 50 条
  • [1] Provably Safe Deep Reinforcement Learning for Robotic Manipulation in Human Environments
    Thumm, Jakob
    Althoff, Matthias
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2022, 2022, : 6344 - 6350
  • [2] Provably Safe Reinforcement Learning with Step-wise Violation Constraints
    Xiong, Nuoya
    Du, Yihan
    Huang, Longbo
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Provably Efficient Generalized Lagrangian Policy Optimization for Safe Multi-Agent Reinforcement Learning
    Ding, Dongsheng
    Wei, Xiaohan
    Yang, Zhuoran
    Wang, Zhuoran
    Jovanovic, Mihailo R.
    [J]. LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [4] Provably Safe Model-Based Meta Reinforcement Learning: An Abstraction-Based Approach
    Sun, Xiaowu
    Fatnassi, Wael
    Santa Cruz, Ulices
    Shoukry, Yasser
    [J]. 2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 2963 - 2968
  • [5] Provably Safe Reinforcement Learning via Action Projection Using Reachability Analysis and Polynomial Zonotopes
    Kochdumper, Niklas
    Krasowski, Hanna
    Wang, Xiao
    Bak, Stanley
    Althoff, Matthias
    [J]. IEEE Open Journal of Control Systems, 2023, 2 : 79 - 92
  • [6] Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark
    Ji, Jiaming
    Zhang, Borong
    Zhou, Jiayi
    Pan, Xuehai
    Huang, Weidong
    Sun, Ruiyang
    Geng, Yiran
    Zhong, Yifan
    Dai, Juntao
    Yang, Yaodong
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [7] Provably Robust Blackbox Optimization for Reinforcement Learning
    Choromanski, Krzysztof
    Pacchiano, Aldo
    Parker-Holder, Jack
    Tang, Yunhao
    Jain, Deepali
    Yang, Yuxiang
    Iscen, Atil
    Hsu, Jasmine
    Sindhwani, Vikas
    [J]. CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [8] Provably adaptive reinforcement learning in metric spaces
    Cao, Tongyi
    Krishnamurthy, Akshay
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [9] Provably Efficient Exploration for Reinforcement Learning Using Unsupervised Learning
    Feng, Fei
    Wang, Ruosong
    Yin, Wotao
    Du, Simon S.
    Yang, Lin F.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [10] On Normative Reinforcement Learning via Safe Reinforcement Learning
    Neufeld, Emery A.
    Bartocci, Ezio
    Ciabattoni, Agata
    [J]. PRIMA 2022: PRINCIPLES AND PRACTICE OF MULTI-AGENT SYSTEMS, 2023, 13753 : 72 - 89