Actor-Critic Physics-Informed Neural Lyapunov Control

被引:0
|
作者
Wang, Jiarui [1 ]
Fazlyab, Mahyar [2 ]
机构
[1] Johns Hopkins Univ, Comp Sci Dept, Baltimore, MD 21218 USA
[2] Johns Hopkins Univ, Elect & Comp Engn Dept, Baltimore, MD 21218 USA
来源
关键词
Lyapunov methods; stability of nonlinear systems; neural networks;
D O I
10.1109/LCSYS.2024.3416235
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Designing control policies for stabilization tasks with provable guarantees is a long-standing problem in nonlinear control. A crucial performance metric is the size of the resulting region of attraction, which essentially serves as a robustness "margin" of the closed-loop system against uncertainties. In this letter, we propose a new method to train a stabilizing neural network controller along with its corresponding Lyapunov certificate, aiming to maximize the resulting region of attraction while respecting the actuation constraints. Crucial to our approach is the use of Zubov's Partial Differential Equation (PDE), which precisely characterizes the true region of attraction of a given control policy. Our framework follows an actor-critic pattern where we alternate between improving the control policy (actor) and learning a Zubov function (critic). Finally, we compute the largest certifiable region of attraction by invoking an SMT solver after the training procedure. Our numerical experiments on several design problems show consistent and significant improvements in the size of the resulting region of attraction.
引用
收藏
页码:1751 / 1756
页数:6
相关论文
共 50 条
  • [41] Separable Physics-Informed Neural Networks
    Cho, Junwoo
    Nam, Seungtae
    Yang, Hyunmo
    Yun, Seok-Bae
    Hong, Youngjoon
    Park, Eunbyung
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [42] Actor-Critic or Critic-Actor? A Tale of Two Time Scales
    Bhatnagar, Shalabh
    Borkar, Vivek S.
    Guin, Soumyajit
    IEEE CONTROL SYSTEMS LETTERS, 2023, 7 : 2671 - 2676
  • [43] Noisy Importance Sampling Actor-Critic: An Off-Policy Actor-Critic With Experience Replay
    Tasfi, Norman
    Capretz, Miriam
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [44] Actor-Critic Neural Network Based Finite-time Control for Uncertain Robotic Systems
    Lei, Changyi
    5TH INTERNATIONAL CONFERENCE ON INFORMATION SYSTEM AND DATA MINING (ICISDM 2021), 2021, : 34 - 40
  • [45] Numerical analysis of physics-informed neural networks and related models in physics-informed machine learning
    De Ryck, Tim
    Mishra, Siddhartha
    ACTA NUMERICA, 2024, 33 : 633 - 713
  • [46] On the Role of Models in Learning Control: Actor-Critic Iterative Learning Control
    Poot, Maurice
    Portegies, Jim
    Oomen, Tom
    IFAC PAPERSONLINE, 2020, 53 (02): : 1450 - 1455
  • [47] Development and Validation of Active Roll Control based on Actor-critic Neural Network Reinforcement Learning
    Bahr, Matthias
    Reicherts, Sebastian
    Sieberg, Philipp
    Morss, Luca
    Schramm, Dieter
    SIMULTECH: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON SIMULATION AND MODELING METHODOLOGIES, TECHNOLOGIES AND APPLICATIONS, 2019, 2019, : 36 - 46
  • [48] Temporal Logic Motion Control using Actor-Critic Methods
    Ding, Xu Chu
    Wang, Jing
    Lahijanian, Morteza
    Paschalidis, Ioannis Ch
    Belta, Calin A.
    2012 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2012, : 4687 - 4692
  • [49] Importance sampling actor-critic algorithms
    Williams, Jason L.
    Fisher, John W., III
    Willsky, Alan S.
    2006 AMERICAN CONTROL CONFERENCE, VOLS 1-12, 2006, 1-12 : 1625 - +
  • [50] A two-layer networked learning control system using actor-critic neural network
    Du, Dajun
    Fei, Minrui
    APPLIED MATHEMATICS AND COMPUTATION, 2008, 205 (01) : 26 - 36