Safety Margins for Reinforcement Learning

被引:2
|
作者
Grushin, Alexander [1 ]
Woods, Walt [1 ]
Velasquez, Alvaro [2 ]
Khan, Simon [3 ]
机构
[1] Galois Inc, Portland, OR 97204 USA
[2] Univ Colorado Boulder, Boulder, CO USA
[3] Air Force Res Lab, Wright Patterson AFB, OH USA
关键词
D O I
10.1109/CAI54212.2023.00026
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Any autonomous controller will be unsafe in some situations. The ability to quantitatively identify when these unsafe situations are about to occur is crucial for drawing timely human oversight in, e.g., freight transportation applications. In this work, we demonstrate that the true criticality of an agent's situation can be robustly defined as the mean reduction in reward given some number of random actions. Proxy criticality metrics that are computable in real-time (i.e., without actually simulating the effects of random actions) can be compared to the true criticality, and we show how to leverage these proxy metrics to generate safety margins, which directly tie the consequences of potentially incorrect actions to an anticipated loss in overall performance. We evaluate our approach on learned policies from APE-X and A3C within an Atari environment, and demonstrate how safety margins decrease as agents approach failure states. The integration of safety margins into programs for monitoring deployed agents allows for the real-time identification of potentially catastrophic situations.
引用
收藏
页码:42 / 43
页数:2
相关论文
共 50 条
  • [1] Reconnaissance for Reinforcement Learning with Safety Constraints
    Maeda, Shin-ichi
    Watahiki, Hayato
    Ouyang, Yi
    Okada, Shintarou
    Koyama, Masanori
    Nagarajan, Prabhat
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT II, 2021, 12976 : 567 - 582
  • [2] Safety reinforcement learning control via transfer learning
    Zhang, Quanqi
    Wu, Chengwei
    Tian, Haoyu
    Gao, Yabin
    Yao, Weiran
    Wu, Ligang
    AUTOMATICA, 2024, 166
  • [3] Safety-constrained reinforcement learning with a distributional safety critic
    Qisong Yang
    Thiago D. Simão
    Simon H. Tindemans
    Matthijs T. J. Spaan
    Machine Learning, 2023, 112 : 859 - 887
  • [4] Safety-constrained reinforcement learning with a distributional safety critic
    Yang, Qisong
    Simao, Thiago D.
    Tindemans, Simon H.
    Spaan, Matthijs T. J.
    MACHINE LEARNING, 2023, 112 (03) : 859 - 887
  • [5] Task Independent Safety Assessment for Reinforcement Learning
    Jocas, Mark
    Zoghlami, Firas
    Kurrek, Philip
    Gianni, Mario
    Salehi, Vahid
    TOWARDS AUTONOMOUS ROBOTIC SYSTEMS, TAROS 2022, 2022, 13546 : 190 - 204
  • [6] Task-Agnostic Safety for Reinforcement Learning
    Rahman, Md Asifur
    Alqahtani, Sarra
    PROCEEDINGS OF THE 16TH ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2023, 2023, : 139 - 148
  • [7] Runtime Safety Assurance Using Reinforcement Learning
    Lazarus, Christopher
    Lopez, James G.
    Kochenderfer, Mykel J.
    2020 AIAA/IEEE 39TH DIGITAL AVIONICS SYSTEMS CONFERENCE (DASC) PROCEEDINGS, 2020,
  • [8] Solving Safety Problems with Ensemble Reinforcement Learning
    Ferreira, Leonardo A.
    dos Santos, Thiago F.
    Bianchi, Reinaldo A. C.
    Santos, Paulo E.
    AI 2019: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, 11919 : 203 - 214
  • [9] Safety-Constrained Reinforcement Learning for MDPs
    Junges, Sebastian
    Jansen, Nils
    Dehnert, Christian
    Topcu, Ufuk
    Katoen, Joost-Pieter
    TOOLS AND ALGORITHMS FOR THE CONSTRUCTION AND ANALYSIS OF SYSTEMS (TACAS 2016), 2016, 9636 : 130 - 146
  • [10] SAFETY MARGINS
    RINEHART, RW
    AVIATION WEEK & SPACE TECHNOLOGY, 1983, 119 (24): : 94 - 94