A reinforcement learning method with closed-loop stability guarantee

被引:5
|
作者
Osinenko, Pavel [1 ]
Beckenbach, Lukas [1 ]
Goehrt, Thomas [1 ]
Streif, Stefan [1 ]
机构
[1] Tech Univ Chemnitz, Automat Control & Syst Dynam Lab, Chemnitz, Germany
来源
IFAC PAPERSONLINE | 2020年 / 53卷 / 02期
关键词
Reinforcement learning control; Stability of nonlinear systems; Lyapunov methods; ITERATION;
D O I
10.1016/j.ifacol.2020.12.2237
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning (RL) in the context of control systems offers wide possibilities of controller adaptation. Given an infinite-horizon cost function, the so-called critic of RL approximates it with a neural net and sends this information to the controller (called "actor"). However, the issue of closed-loop stability under an RL-method is still not fully addressed. Since the critic delivers merely an approximation to the value function of the corresponding infinitehorizon problem, no guarantee can be given in general as to whether the actor's actions stabilize the system. Different approaches to this issue exist. The current work offers a particular one, which, starting with a (not necessarily smooth) control Lyapunov function (CLF), derives an online RL-scheme in such a way that practical semi-global stability property of the closed-loop can be established. The approach logically continues the work of the authors on parameterized controllers and Lyapunov-like constraints for RL, whereas the CLF now appears merely in one of the constraints of the control scheme. The analysis of the closed-loop behavior is done in a sample-and-hold (SH) manner thus offering a certain insight into the digital realization. The case study with a non-holonomic integrator shows the capabilities of the derived method to optimize the given cost function compared to a nominal stabilizing controller. Copyright (C) 2020 The Authors.
引用
收藏
页码:8043 / 8048
页数:6
相关论文
共 50 条
  • [31] Conceptual closed-loop design of automotive cooling systems leveraging Reinforcement Learning
    Vanhuyse, Johan
    Bertheaume, Clement
    Gumussoy, Suat
    Nicolai, Mike
    FORSCHUNG IM INGENIEURWESEN-ENGINEERING RESEARCH, 2025, 89 (01):
  • [32] A closed-loop algorithm to detect human face using color and reinforcement learning
    吴东晖
    叶秀清
    顾伟康
    "JournalofZhejiangUniversityScienceJ", 2002, (01) : 73 - 77
  • [33] A closed-loop algorithm to detect human face using color and reinforcement learning
    Wu Dong-hui
    Ye Xiu-qing
    Gu Wei-kang
    Journal of Zhejiang University-SCIENCE A, 2002, 3 (1): : 72 - 76
  • [34] Closed-Loop Dynamic Control of a Soft Manipulator Using Deep Reinforcement Learning
    Centurelli, Andrea
    Arleo, Luca
    Rizzo, Alessandro
    Tolu, Silvia
    Laschi, Cecilia
    Falotico, Egidio
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (02) : 4741 - 4748
  • [35] Deep reinforcement learning for closed-loop blood glucose control: two approaches
    Di Felice, Francesco
    Borri, Alessandro
    Di Benedetto, Maria Domenica
    IFAC PAPERSONLINE, 2022, 55 (40): : 115 - 120
  • [36] A CLOSED-LOOP CONTROLLER FOR VOLTAGE STABILITY
    RAJU, VV
    KUPPURAJULU, A
    INTERNATIONAL JOURNAL OF ELECTRICAL POWER & ENERGY SYSTEMS, 1993, 15 (05) : 283 - 292
  • [37] Designing closed-loop circuits for stability
    Burton, RV
    HYDRAULICS & PNEUMATICS, 1996, 49 (09) : 53 - &
  • [38] Dissipativity-constrained learning of MPC with guaranteeing closed-loop stability?
    Hara, Keita
    Inoue, Masaki
    Sebe, Noboru
    AUTOMATICA, 2023, 157
  • [39] Closed-loop Flow Control Method Based on Deep Reinforcement Learning using a Co-flow Jet
    Zhao, Y. R.
    Xu, H. Y.
    Xie, Z. Y.
    JOURNAL OF APPLIED FLUID MECHANICS, 2024, 17 (04) : 816 - 827
  • [40] CLOSED-LOOP THEORY OF MOTOR LEARNING
    ADAMS, JA
    JOURNAL OF MOTOR BEHAVIOR, 1971, 3 (02) : 111 - 150