Learning-Based Predictive Control for Discrete-Time Nonlinear Systems With Stochastic Disturbances

被引:55
|
作者
Xu, Xin [1 ]
Chen, Hong [2 ,3 ]
Lian, Chuanqiang [4 ]
Li, Dazi [5 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci, Changsha 410073, Hunan, Peoples R China
[2] Jilin Univ NanLing, State Key Lab Automot Simulat & Control, Changchun 130025, Jilin, Peoples R China
[3] Jilin Univ NanLing, Dept Control Sci & Engn, Changchun 130025, Jilin, Peoples R China
[4] Naval Univ Engn, Natl Key Lab Sci & Technol Vessel Integrated Powe, Wuhan 430032, Hubei, Peoples R China
[5] Beijing Univ Chem Technol, Dept Automat, Beijing 100029, Peoples R China
基金
中国国家自然科学基金;
关键词
Adaptive dynamic programming (ADP); function approximation; model predictive control (MPC); optimal control; receding horizon; reinforcement learning (RL); H-INFINITY CONTROL; POLICY ITERATION; CONTROL SCHEME;
D O I
10.1109/TNNLS.2018.2820019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a learning-based predictive control (LPC) scheme is proposed for adaptive optimal control of discrete-time nonlinear systems under stochastic disturbances. The proposed LPC scheme is different from conventional model predictive control (MPC), which uses open-loop optimization or simplified closed-loop optimal control techniques in each horizon. In LPC, the control task in each horizon is formulated as a closed-loop nonlinear optimal control problem and a finite-horizon iterative reinforcement learning (RL) algorithm is developed to obtain the closed-loop optimal/suboptimal solutions. Therefore, in LPC, RL and adaptive dynamic programming ( ADP) are used as a new class of closed-loop learning-based optimization techniques for nonlinear predictive control with stochastic disturbances. Moreover, LPC also decomposes the infinite-horizon optimal control problem in previous RL and ADP methods into a series of finite horizon problems, so that the computational costs are reduced and the learning efficiency can be improved. Convergence of the finite-horizon iterative RL algorithm in each prediction horizon and the Lyapunov stability of the closed-loop control system are proved. Moreover, by using successive policy updates between adjoint time horizons, LPC also has lower computational costs than conventional MPC which has independent optimization procedures between two different prediction horizons. Simulation results illustrate that compared with conventional nonlinear MPC as well as ADP, the proposed LPC scheme can obtain a better performance both in terms of policy optimality and computational efficiency.
引用
收藏
页码:6202 / 6213
页数:12
相关论文
共 50 条
  • [1] Reinforcement Learning-Based Model Predictive Control for Discrete-Time Systems
    Lin, Min
    Sun, Zhongqi
    Xia, Yuanqing
    Zhang, Jinhui
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3312 - 3324
  • [2] Robust Learning-Based Predictive Control for Discrete-Time Nonlinear Systems With Unknown Dynamics and State Constraints
    Zhang, Xinglong
    Liu, Jiahang
    Xu, Xin
    Yu, Shuyou
    Chen, Hong
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (12): : 7314 - 7327
  • [3] Online Learning-Based Predictive Control of Switched Nonlinear Systems With Disturbances
    Hu, Cheng
    Wu, Zhe
    2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 92 - 99
  • [4] Reinforcement Learning-Based Control for Nonlinear Discrete-Time Systems with Unknown Control Directions and Control Constraints
    Huang, Miao
    Liu, Cong
    He, Xiaoqi
    Ma, Longhua
    Lu, Zheming
    Su, Hongye
    NEUROCOMPUTING, 2020, 402 : 50 - 65
  • [5] Filtering for nonlinear discrete-time networked control systems with incomplete measurements and Stochastic disturbances
    Wei, Guoliang
    Wang, Zidong
    Shu, Huisheng
    He, Xiao
    PROCEEDINGS OF 2008 IEEE INTERNATIONAL CONFERENCE ON NETWORKING, SENSING AND CONTROL, VOLS 1 AND 2, 2008, : 330 - +
  • [6] A note on control of a class of discrete-time stochastic systems with distributed delays and nonlinear disturbances
    Wang, Zidong
    Liu, Yurong
    Wei, Guoliang
    Liu, Xiaohui
    AUTOMATICA, 2010, 46 (03) : 543 - 548
  • [7] Control of Equilibria for Nonlinear Stochastic Discrete-Time Systems
    Bashkirtseva, Irina
    Ryashko, Lev
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2011, 56 (09) : 2162 - 2166
  • [8] H∞ control for discrete-time nonlinear Stochastic systems
    Berman, N
    Shaked, U
    2004 43RD IEEE CONFERENCE ON DECISION AND CONTROL (CDC), VOLS 1-5, 2004, : 2578 - 2583
  • [9] H∞ control for discrete-time nonlinear stochastic systems
    Berman, Nadav
    Shaked, Uri
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2006, 51 (06) : 1041 - 1046
  • [10] Probabilistic Constrained Model Predictive Control for Linear Discrete-time Systems with Additive Stochastic Disturbances
    Hashimoto, Tomoaki
    2013 IEEE 52ND ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2013, : 6434 - 6439