Safety-Critical Optimal Control of Discrete-Time Non-Linear Systems via Policy Iteration-Based Q-Learning

被引:0
|
作者
Long, Lijun [1 ,2 ]
Liu, Xiaomei [1 ,2 ]
Huang, Xiaomin [1 ,2 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang, Peoples R China
[2] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang, Peoples R China
关键词
control barrier functions; discrete-time systems; neural networks; Q-learning; safety-critical control;
D O I
10.1002/rnc.7809
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper investigates the problem of safety-critical optimal control for discrete-time non-linear systems. A safety-critical control algorithm is developed based on Q-learning and an iterative adaptive dynamic programming, that is, policy iteration. Discrete-time control barrier functions (CBFs) are introduced into the utility function for guaranteeing safety, in which a novel definition of the safe set and its boundary with multiple discrete-time CBFs are given. Also, for discrete-time systems, by using multiple discrete-time CBFs, the safety-critical optimal control problem of multiple safety objectives is addressed. Meanwhile, safety, convergence, and stability of the developed algorithm are rigorously demonstrated. An effective method to obtain an initial safety-admissible control law is established. Also, the developed algorithm is implemented by building an actor-critic structure with neural networks. Finally, the effectiveness of the proposed algorithm is illustrated by three simulation examples.
引用
收藏
页数:19
相关论文
共 50 条
  • [41] Reinforcement Q-Learning and Non-Zero-Sum Games Optimal Tracking Control for Discrete-Time Linear Multi-Input Systems
    Zhao, Jin-Gang
    2023 IEEE 12TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE, DDCLS, 2023, : 277 - 282
  • [42] Policy Iteration for Optimal Control of Discrete-Time Time-Varying Nonlinear Systems
    Zhu, Guangyu
    Li, Xiaolu
    Sun, Ranran
    Yang, Yiyuan
    Zhang, Peng
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2023, 10 (03) : 781 - 791
  • [43] Policy Iteration for Optimal Control of Discrete-Time Time-Varying Nonlinear Systems
    Guangyu Zhu
    Xiaolu Li
    Ranran Sun
    Yiyuan Yang
    Peng Zhang
    IEEE/CAAJournalofAutomaticaSinica, 2023, 10 (03) : 781 - 791
  • [44] Optimal control for unknown mean-field discrete-time system based on Q-Learning
    Ge, Yingying
    Liu, Xikui
    Li, Yan
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2021, 52 (15) : 3335 - 3349
  • [45] Learning Optimal Control Policy for Unknown Discrete-Time Systems
    Lai, Jing
    Xiong, Junlin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (11) : 4191 - 4195
  • [46] The Adaptive Optimal Output Feedback Tracking Control of Unknown Discrete-Time Linear Systems Using a Multistep Q-Learning Approach
    Dong, Xunde
    Lin, Yuxin
    Suo, Xudong
    Wang, Xihao
    Sun, Weijie
    MATHEMATICS, 2024, 12 (04)
  • [47] H∞ Control for Discrete-time Linear Systems by Integrating Off-policy Q-learning and Zero-sum Game
    Li, Jinna
    Ding, Zhengtao
    Yang, Chunyu
    Niu, Hong
    2018 IEEE 14TH INTERNATIONAL CONFERENCE ON CONTROL AND AUTOMATION (ICCA), 2018, : 817 - 822
  • [48] Output Feedback Reinforcement Q-learning for Optimal Quadratic Tracking Control of Unknown Discrete-Time Linear Systems and Its Application
    Zhao, Guangyue
    Sun, Weijie
    Cai, He
    Peng, Yunjian
    2018 15TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2018, : 750 - 755
  • [49] Optimal Learning Control for Discrete-Time Nonlinear Systems Using Generalized Policy Iteration Based Adaptive Dynamic Programming
    Wei, Qinglai
    Liu, Derong
    2014 11TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2014, : 1781 - 1786
  • [50] H∞ Tracking Control of Unknown Discrete-Time Linear Systems via Output-Data-Driven Off-policy Q-learning Algorithm
    Zhang, Kun
    Liu, Xuantong
    Zhang, Lei
    Chen, Qian
    Peng, Yunjian
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 2350 - 2356