Probabilistic Safeguard for Reinforcement Learning Using Safety Index Guided Gaussian Process Models

被引:0
|
作者
Zhao, Weiye [1 ]
He, Tairan [1 ]
Liu, Changliu [1 ]
机构
[1] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
基金
美国国家科学基金会;
关键词
Safe control; Gaussian process; Dynamics learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Safety is one of the biggest concerns to applying reinforcement learning (RL) to the physical world. In its core part, it is challenging to ensure RL agents persistently satisfy a hard state constraint without white-box or black-box dynamics models. This paper presents an integrated model learning and safe control framework to safeguard any RL agent, where the environment dynamics are learned as Gaussian processes. The proposed theory provides (i) a novel method to construct an offline dataset for model learning that best achieves safety requirements; (ii) a design rule to construct the safety index to ensure the existence of safe control under control limits; (iii) a probablistic safety guarantee (i.e. probabilistic forward invariance) when the model is learned using the aforementioned dataset. Simulation results show that our framework achieves almost zero safety violation on various continuous control tasks.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Increasing the Safety of Adaptive Cruise Control Using Physics-Guided Reinforcement Learning
    Jurj, Sorin Liviu
    Grundt, Dominik
    Werner, Tino
    Borchers, Philipp
    Rothemann, Karina
    Mohlmann, Eike
    ENERGIES, 2021, 14 (22)
  • [22] Probabilistic Curve Learning: Coulomb Repulsion and the Electrostatic Gaussian Process
    Wang, Ye
    Dunson, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [23] Monopolistic Models for Resource Allocation: A Probabilistic Reinforcement Learning Approach
    Zhang, Yue
    Song, Bin
    Gao, Su
    Du, Xiaojiang
    Guizani, Mohsen
    IEEE ACCESS, 2018, 6 : 49721 - 49731
  • [24] Gaussian Process Reinforcement Learning for Fast Opportunistic Spectrum Access
    Yan, Zun
    Cheng, Peng
    Chen, Zhuo
    Li, Yonghui
    Vucetic, Branka
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 : 2613 - 2628
  • [25] Nonlinear Inverse Reinforcement Learning with Mutual Information and Gaussian Process
    Li, De C.
    He, Yu Q.
    Fu, Feng
    2014 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS IEEE-ROBIO 2014, 2014, : 1445 - 1450
  • [26] PROBABILISTIC UNCERTAINTY PROPAGATION USING GAUSSIAN PROCESS SURROGATES
    Manfredi, P.
    INTERNATIONAL JOURNAL FOR UNCERTAINTY QUANTIFICATION, 2024, 14 (06) : 71 - 104
  • [27] Gaussian Process Reinforcement Learning for Fast Opportunistic Spectrum Access
    Yan, Zun
    Cheng, Peng
    Chen, Zhuo
    Li, Yonghui
    Vucetic, Branka
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [28] Reinforcement learning for continuous spaces based on Gaussian process classifier
    Wang, Xue-Song
    Zhang, Yi-Yang
    Cheng, Yu-Hu
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2009, 37 (06): : 1153 - 1158
  • [29] Integrating Gaussian Process with Reinforcement Learning for Adaptive Service Composition
    Wang, Hongbing
    Wu, Qin
    Chen, Xin
    Yu, Qi
    SERVICE-ORIENTED COMPUTING, (ICSOC 2015), 2015, 9435 : 203 - 217
  • [30] Associative reinforcement learning using linear probabilistic concepts
    Abe, N
    Long, PM
    MACHINE LEARNING, PROCEEDINGS, 1999, : 3 - 11