Safe control of nonlinear systems in LPV framework using model-based reinforcement learning

被引:7
|
作者
Bao, Yajie [1 ]
Velni, Javad Mohammadpour [1 ]
机构
[1] Univ Georgia, Sch Elect & Comp Engn, Athens, GA 30602 USA
基金
美国国家科学基金会;
关键词
Safe nonlinear control; model-based reinforcement learning; LPV framework; PREDICTIVE CONTROL; IDENTIFICATION;
D O I
10.1080/00207179.2022.2029945
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a safe model-based reinforcement learning (MBRL) approach to control nonlinear systems described by linear parameter-varying (LPV) models. A variational Bayesian inference Neural Network (BNN) approach is first employed to learn a state-space model with uncertainty quantification from input-output data collected from the system; the model is then utilised for training MBRL to learn control actions for the system with safety guarantees. Specifically, MBRL employs the BNN model to generate simulation environments for training, which avoids safety violations in the exploration stage. To adapt to dynamically varying environments, knowledge on the evolution of LPV model scheduling variables is incorporated in simulation to reduce the discrepancy between the transition distributions of simulation and real environments. Experiments on a parameter-varying double integrator system and a control moment gyroscope (CMG) simulation model demonstrate that the proposed approach can safely achieve desired control performance.
引用
收藏
页码:1078 / 1089
页数:12
相关论文
共 50 条
  • [21] A Safe and Data-Efficient Model-Based Reinforcement Learning System for HVAC Control
    Ding, Xianzhong
    An, Zhiyu
    Rathee, Arya
    Du, Wan
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (07): : 8014 - 8032
  • [22] Model-Based Reinforcement Learning For Robot Control
    Li, Xiang
    Shang, Weiwei
    Cong, Shuang
    2020 5TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2020), 2020, : 300 - 305
  • [23] Conservative and Adaptive Penalty for Model-Based Safe Reinforcement Learning
    Ma, Yecheng Jason
    Shen, Andrew
    Bastani, Osbert
    Jayaraman, Dinesh
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 5404 - 5412
  • [24] Control of Nonlinear and LPV Systems: Interval Observer-Based Framework
    Efimov, Denis
    Raissi, Tarek
    Zolghadri, Ali
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2013, 58 (03) : 773 - 778
  • [25] Introducing a Model-based Learning Control Software for Nonlinear Systems: ROFALT
    Steinhauser, Armin
    Swevers, Jan
    2018 EUROPEAN CONTROL CONFERENCE (ECC), 2018, : 3149 - 3154
  • [26] Model-based reinforcement learning for nonlinear optimal control with practical asymptotic stability guarantees
    Kim, Yeonsoo
    Lee, Jong Min
    AICHE JOURNAL, 2020, 66 (10)
  • [27] Control Approach Combining Reinforcement Learning and Model-Based Control
    Okawa, Yoshihiro
    Sasaki, Tomotake
    Iwane, Hidenao
    2019 12TH ASIAN CONTROL CONFERENCE (ASCC), 2019, : 1419 - 1424
  • [28] Learning to Attack Federated Learning: A Model-based Reinforcement Learning Attack Framework
    Li, Henger
    Sun, Xiaolin
    Zheng, Zizhan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [29] Control of Nonlinear Physiological Systems via LPV Framework
    Eigner, Gyorgy
    Drexler, Daniel Andras
    Kovacs, Levente
    PROCEEDINGS OF THE 2017 12TH IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA), 2017, : 2088 - 2093
  • [30] Model-based inverse reinforcement learning for deterministic systems
    Self, Ryan
    Abudia, Moad
    Mahmud, S. M. Nahid
    Kamalapurkar, Rushikesh
    AUTOMATICA, 2022, 140