Probabilistic Safeguard for Reinforcement Learning Using Safety Index Guided Gaussian Process Models

被引:0
|
作者
Zhao, Weiye [1 ]
He, Tairan [1 ]
Liu, Changliu [1 ]
机构
[1] Carnegie Mellon Univ, Robot Inst, Pittsburgh, PA 15213 USA
基金
美国国家科学基金会;
关键词
Safe control; Gaussian process; Dynamics learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Safety is one of the biggest concerns to applying reinforcement learning (RL) to the physical world. In its core part, it is challenging to ensure RL agents persistently satisfy a hard state constraint without white-box or black-box dynamics models. This paper presents an integrated model learning and safe control framework to safeguard any RL agent, where the environment dynamics are learned as Gaussian processes. The proposed theory provides (i) a novel method to construct an offline dataset for model learning that best achieves safety requirements; (ii) a design rule to construct the safety index to ensure the existence of safe control under control limits; (iii) a probablistic safety guarantee (i.e. probabilistic forward invariance) when the model is learned using the aforementioned dataset. Simulation results show that our framework achieves almost zero safety violation on various continuous control tasks.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] An Index Advisor Using Deep Reinforcement Learning
    Lan, Hai
    Bao, Zhifeng
    Peng, Yuwei
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 2105 - 2108
  • [42] ACDRL: A constraint-guided dynamic reinforcement learning for index selection
    Yan, Yu
    Geng, Jian
    Wang, Hongzhi
    Zhang, Kaixin
    Wang, Yuzhuo
    INFORMATION SCIENCES, 2025, 689
  • [43] LEARNING-PROCESS PROBABILISTIC MODELS IN IDENTIFICATION PROBLEMS
    DRYNKOV, AV
    PSIKHOLOGICHESKII ZHURNAL, 1983, 4 (03) : 102 - 107
  • [44] Probabilistic Initial Orbit Determination Using Gaussian Mixture Models
    DeMars, Kyle J.
    Jah, Moriba K.
    JOURNAL OF GUIDANCE CONTROL AND DYNAMICS, 2013, 36 (05) : 1324 - 1335
  • [45] Learning Gaussian Process Models from Uncertain Data
    Dallaire, Patrick
    Besse, Camille
    Chaib-draa, Brahim
    NEURAL INFORMATION PROCESSING, PT 1, PROCEEDINGS, 2009, 5863 : 433 - 440
  • [46] Learning Stable Gaussian Process State Space Models
    Umlauft, Jonas
    Lederer, Armin
    Hirche, Sandra
    2017 AMERICAN CONTROL CONFERENCE (ACC), 2017, : 1499 - 1504
  • [47] ASSUMED DENSITY FILTERING FOR LEARNING GAUSSIAN PROCESS MODELS
    Ramakrishnan, Naveen
    Ertin, Emre
    Moses, Randolph L.
    2011 IEEE STATISTICAL SIGNAL PROCESSING WORKSHOP (SSP), 2011, : 257 - 260
  • [48] TAG: Teacher-Advice Mechanism With Gaussian Process for Reinforcement Learning
    Lin, Ke
    Li, Duantengchuan
    Li, Yanjie
    Chen, Shiyu
    Liu, Qi
    Gao, Jianqi
    Jin, Yanrui
    Gong, Liang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (09) : 12419 - 12433
  • [49] Flexible unmanned surface vehicles control using probabilistic model-based reinforcement learning with hierarchical Gaussian distribution
    Cui, Yunduan
    Xu, Kun
    Zheng, Chunhua
    Liu, Jia
    Peng, Lei
    Li, Huiyun
    OCEAN ENGINEERING, 2023, 285
  • [50] Probabilistic tracking control for non-Gaussian stochastic process using novel iterative learning algorithms
    Yi, Yang
    Sun, ChangYin
    Guo, Lei
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2013, 44 (07) : 1325 - 1332