Robust Learning-Based Predictive Control for Discrete-Time Nonlinear Systems With Unknown Dynamics and State Constraints

被引:24
|
作者
Zhang, Xinglong [1 ]
Liu, Jiahang [1 ]
Xu, Xin [1 ]
Yu, Shuyou [2 ,3 ]
Chen, Hong [4 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Changsha 410073, Peoples R China
[2] Jilin Univ, State Key Lab Automot Simulat & Control, Changchun 130025, Peoples R China
[3] Jilin Univ, Dept Control Sci & Engn, Changchun 130025, Peoples R China
[4] Tongji Univ, Coll Elect & Informat Engn, Shanghai 201804, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Robustness; Predictive control; Aerospace electronics; Computational modeling; Predictive models; Optimal control; Heuristic algorithms; Model predictive control (MPC); nonlinear systems; reinforcement learning (RL); robustness; state constraints; TRACKING CONTROL; LINEAR-SYSTEMS; STABILITY;
D O I
10.1109/TSMC.2022.3146284
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robust model predictive control (MPC) is a well-known control technique for model-based control with constraints and uncertainties. In classic robust tube-based MPC approaches, an open-loop control sequence is computed via periodically solving an online nominal MPC problem, which requires prior model information and frequent access to onboard computational resources. In this article, we propose an efficient robust MPC solution based on receding horizon reinforcement learning, called r-LPC, for unknown nonlinear systems with state constraints and disturbances. The proposed r-LPC utilizes a Koopman operator-based prediction model obtained offline from precollected input-output datasets. Unlike classic tube-based MPC, in each prediction time interval of r-LPC, we use an actor-critic structure to learn a near-optimal feedback control policy rather than a control sequence. The resulting closed-loop control policy can be learned offline and deployed online or learned online in an asynchronous way. In the latter case, online learning can be activated whenever necessary; for instance, the safety constraint is violated with the deployed policy. The closed-loop recursive feasibility, robustness, and asymptotic stability are proven under function approximation errors of the actor-critic networks. Simulation and experimental results on two nonlinear systems with unknown dynamics and disturbances have demonstrated that our approach has better or comparable performance when compared with tube-based MPC and linear quadratic regulator, and outperforms a recently developed actor-critic learning approach.
引用
收藏
页码:7314 / 7327
页数:14
相关论文
共 50 条
  • [41] Robust model predictive control for discrete-time nonlinear systems with delay-range-dependent
    Zhao, Jie-Mei
    Hu, Zhong-Hui
    Zhang, Li-Jun
    Kongzhi yu Juece/Control and Decision, 2015, 30 (01): : 59 - 64
  • [42] Discrete-time online learning control for a class of unknown nonaffine nonlinear systems using reinforcement learning
    Yang, Xiong
    Liu, Derong
    Wang, Ding
    Wei, Qinglai
    NEURAL NETWORKS, 2014, 55 : 30 - 41
  • [43] Output Feedback Adaptive Iterative Learning Control for Nonlinear Discrete-Time Systems with Unknown Control Directions
    Yu, Miao
    Wang, Jiasen
    Xin, Huanhai
    Qi, Donglian
    2012 IEEE 51ST ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2012, : 4660 - 4665
  • [44] Event-based online learning control design with eligibility trace for discrete-time unknown nonlinear systems
    Wang, Ding
    Wang, Jiangyu
    Hu, Lingzhi
    Zhao, Mingming
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 123
  • [45] A Direct Adaptive Iterative Learning Control for Nonaffine Nonlinear Discrete-Time Systems with Unknown Control Directions
    Wang, Ying-Chung
    Chien, Chiang-Ju
    Chi, Ronghu
    Shen, Dong
    2016 INTERNATIONAL CONFERENCE ON FUZZY THEORY AND ITS APPLICATIONS (IFUZZY), 2016,
  • [46] Discrete-Time Positive Periodic Systems With State and Control Constraints
    Rami, Mustapha Ait
    Napp, Diego
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2016, 61 (01) : 234 - 239
  • [47] Broad Learning System Approximation-Based Adaptive Optimal Control for Unknown Discrete-Time Nonlinear Systems
    Yuan, Liang'en
    Li, Tieshan
    Tong, Shaocheng
    Xiao, Yang
    Shan, Qihe
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (08): : 5028 - 5038
  • [48] Robust memory state feedback model predictive control for discrete-time uncertain state delayed systems
    Ji, D. H.
    Park, Ju H.
    Yoo, W. J.
    Won, S. C.
    APPLIED MATHEMATICS AND COMPUTATION, 2009, 215 (06) : 2035 - 2044
  • [49] Adaptive robust control of a class of nonlinear strict-feedback discrete-time systems with unknown control directions
    Ge, Shuzhi Sam
    Yang, Chenguang
    Lee, Tong Heng
    SYSTEMS & CONTROL LETTERS, 2008, 57 (11) : 888 - 895
  • [50] Online adaptive policy learning algorithm for H∞ state feedback control of unknown affine nonlinear discrete-time systems
    College of Information Science and Engineering, Northeastern University, Shenyang
    110004, China
    不详
    110004, China
    不详
    475004, China
    不详
    210016, China
    IEEE Trans. Cybern., 12 (2706-2718):