Probabilistic Safeguard for Reinforcement Learning Using Safety Index Guided Gaussian Process Models

被引:0
|
作者
Zhao, Weiye [1 ]
He, Tairan [1 ]
Liu, Changliu [1 ]
机构
[1] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
基金
美国国家科学基金会;
关键词
Safe control; Gaussian process; Dynamics learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Safety is one of the biggest concerns to applying reinforcement learning (RL) to the physical world. In its core part, it is challenging to ensure RL agents persistently satisfy a hard state constraint without white-box or black-box dynamics models. This paper presents an integrated model learning and safe control framework to safeguard any RL agent, where the environment dynamics are learned as Gaussian processes. The proposed theory provides (i) a novel method to construct an offline dataset for model learning that best achieves safety requirements; (ii) a design rule to construct the safety index to ensure the existence of safe control under control limits; (iii) a probablistic safety guarantee (i.e. probabilistic forward invariance) when the model is learned using the aforementioned dataset. Simulation results show that our framework achieves almost zero safety violation on various continuous control tasks.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Reinforcement learning with guided policy search using Gaussian processes
    Jakab, Hunor S.
    Csato, Lehel
    2012 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2012,
  • [2] Inverse Reinforcement Learning with Gaussian Process
    Qiao, Qifeng
    Beling, Peter A.
    2011 AMERICAN CONTROL CONFERENCE, 2011, : 113 - 118
  • [3] Probabilistic Riemannian submanifold learning with wrapped Gaussian process latent variable models
    Mallasto, Anton
    Hauberg, Soren
    Feragen, Aasa
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [4] Probabilistic clear-sky index forecasts using Gaussian process ensembles
    van der Meer, Dennis
    Munkhammar, Joakim
    Widen, Joakim
    2018 IEEE 7TH WORLD CONFERENCE ON PHOTOVOLTAIC ENERGY CONVERSION (WCPEC) (A JOINT CONFERENCE OF 45TH IEEE PVSC, 28TH PVSEC & 34TH EU PVSEC), 2018, : 2724 - 2729
  • [5] Learning safety in model-based Reinforcement Learning using MPC and Gaussian Processes
    Airaldi, Filippo
    De Schutter, Bart
    Dabiri, Azita
    IFAC PAPERSONLINE, 2023, 56 (02): : 5759 - 5764
  • [6] Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models
    Chua, Kurtland
    Calandra, Roberto
    McAllister, Rowan
    Levine, Sergey
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [7] Reinforcement learning with Gaussian process regression using variational free energy
    Kameda, Kiseki
    Tanaka, Fuyuhiko
    JOURNAL OF INTELLIGENT SYSTEMS, 2023, 32 (01)
  • [8] Dialogue manager domain adaptation using Gaussian process reinforcement learning
    Gasic, Milica
    Mrksic, Nikola
    Rojas-Barahona, Lina M.
    Su, Pei-Hao
    Ultes, Stefan
    Vandyke, David
    Wen, Tsung-Hsien
    Young, Steve
    COMPUTER SPEECH AND LANGUAGE, 2017, 45 : 552 - 569
  • [9] Gaussian process model based reinforcement learning
    Yoo J.H.
    Journal of Institute of Control, Robotics and Systems, 2019, 25 (08) : 746 - 751
  • [10] Probabilistic Constraint for Safety-Critical Reinforcement Learning
    Chen, Weiqin
    Subramanian, Dharmashankar
    Paternain, Santiago
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (10) : 6789 - 6804