Actor-Critic Physics-Informed Neural Lyapunov Control

被引:0
|
作者
Wang, Jiarui [1 ]
Fazlyab, Mahyar [2 ]
机构
[1] Johns Hopkins Univ, Comp Sci Dept, Baltimore, MD 21218 USA
[2] Johns Hopkins Univ, Elect & Comp Engn Dept, Baltimore, MD 21218 USA
来源
关键词
Lyapunov methods; stability of nonlinear systems; neural networks;
D O I
10.1109/LCSYS.2024.3416235
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Designing control policies for stabilization tasks with provable guarantees is a long-standing problem in nonlinear control. A crucial performance metric is the size of the resulting region of attraction, which essentially serves as a robustness "margin" of the closed-loop system against uncertainties. In this letter, we propose a new method to train a stabilizing neural network controller along with its corresponding Lyapunov certificate, aiming to maximize the resulting region of attraction while respecting the actuation constraints. Crucial to our approach is the use of Zubov's Partial Differential Equation (PDE), which precisely characterizes the true region of attraction of a given control policy. Our framework follows an actor-critic pattern where we alternate between improving the control policy (actor) and learning a Zubov function (critic). Finally, we compute the largest certifiable region of attraction by invoking an SMT solver after the training procedure. Our numerical experiments on several design problems show consistent and significant improvements in the size of the resulting region of attraction.
引用
收藏
页码:1751 / 1756
页数:6
相关论文
共 50 条
  • [21] Characterizing Motor Control of Mastication With Soft Actor-Critic
    Abdi, Amir H.
    Sagl, Benedikt
    Srungarapu, Venkata P.
    Stavness, Ian
    Prisman, Eitan
    Abolmaesumi, Purang
    Fels, Sidney
    FRONTIERS IN HUMAN NEUROSCIENCE, 2020, 14
  • [22] Actor-Critic Reinforcement Learning for Control With Stability Guarantee
    Han, Minghao
    Zhang, Lixian
    Wang, Jun
    Pan, Wei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 6217 - 6224
  • [23] Variational actor-critic algorithms*,**
    Zhu, Yuhua
    Ying, Lexing
    ESAIM-CONTROL OPTIMISATION AND CALCULUS OF VARIATIONS, 2023, 29
  • [24] Error controlled actor-critic
    Gao, Xingen
    Chao, Fei
    Zhou, Changle
    Ge, Zhen
    Yang, Longzhi
    Chang, Xiang
    Shang, Changjing
    Shen, Qiang
    INFORMATION SCIENCES, 2022, 612 : 62 - 74
  • [25] A Hessian Actor-Critic Algorithm
    Wang, Jing
    Paschalidis, Ioannis Ch
    2014 IEEE 53RD ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2014, : 1131 - 1136
  • [26] Learning State Representation for Deep Actor-Critic Control
    Munk, Jelle
    Kober, Jens
    Babuska, Robert
    2016 IEEE 55TH CONFERENCE ON DECISION AND CONTROL (CDC), 2016, : 4667 - 4673
  • [27] A Predictive Control Method Based on Neural Predictor and Soft Actor-Critic for Power Converters
    Liu, Chenghao
    Ma, Jien
    Liu, Xing
    Qiu, Lin
    Wu, Wenjie
    Fang, Youtong
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024,
  • [28] Actor-Critic Learning Algorithms for Mean-Field Control with Moment Neural Networks
    Pham, Huyen
    Warin, Xavier
    METHODOLOGY AND COMPUTING IN APPLIED PROBABILITY, 2025, 27 (01)
  • [29] Physics-guided actor-critic reinforcement learning for swimming in turbulence
    Koh, Christopher
    Pagnier, Laurent
    Chertkov, Michael
    PHYSICAL REVIEW RESEARCH, 2025, 7 (01):
  • [30] Physics-Informed Extreme Learning Machine Lyapunov Functions
    Zhou, Ruikun
    Fitzsimmons, Maxwell
    Meng, Yiming
    Liu, Jun
    IEEE CONTROL SYSTEMS LETTERS, 2024, 8 : 1763 - 1768