H∞ Tracking Control for Linear Discrete-Time Systems: Model-Free Q-Learning Designs

被引:35
|
作者
Yang, Yunjie [1 ]
Wan, Yan [2 ]
Zhu, Jihong [1 ]
Lewis, Frank L. [3 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Univ Texas Arlington, Dept Elect Engn, Arlington, TX 76019 USA
[3] Univ Texas Arlington, UTA Res Inst, Ft Worth, TX 75052 USA
来源
IEEE CONTROL SYSTEMS LETTERS | 2021年 / 5卷 / 01期
基金
中国国家自然科学基金;
关键词
Linear discrete-time systems; H-infinity tracking control; Q-learning; ZERO-SUM GAMES;
D O I
10.1109/LCSYS.2020.3001241
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this letter, a novel model-free Q-learning based approach is developed to solve the H-infinity tracking problem for linear discrete-time systems. A new exponential discounted value function is introduced that includes the cost of the whole control input and tracking error. The tracking Bellman equation and the game algebraic Riccati equation (GARE) are derived. The solution to the GARE leads to the feedback and feedforward parts of the control input. A Q-learning algorithm is then developed to learn the solution of the GARE online without requiring any knowledge of the system dynamics. Convergence of the algorithm is analyzed, and it is also proved that probing noises in maintaining the persistence of excitation (PE) condition do not result in any bias. An example of the F-16 aircraft short period dynamics is developed to validate the proposed algorithm.
引用
收藏
页码:175 / 180
页数:6
相关论文
共 50 条
  • [41] Adaptive optimal output feedback tracking control for unknown discrete-time linear systems using a combined reinforcement Q-learning and internal model method
    Sun, Weijie
    Zhao, Guangyue
    Peng, Yunjian
    IET CONTROL THEORY AND APPLICATIONS, 2019, 13 (18): : 3075 - 3086
  • [42] Model-Free Optimal Tracking Design With Evolving Control Strategies via Q-Learning
    Wang, Ding
    Huang, Haiming
    Zhao, Mingming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (07) : 3373 - 3377
  • [43] Model-free control of nonlinear stochastic systems with discrete-time measurements
    Spall, JC
    Cristion, JA
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1998, 43 (09) : 1198 - 1210
  • [44] Model-Free Optimal Tracking Control via Critic-Only Q-Learning
    Luo, Biao
    Liu, Derong
    Huang, Tingwen
    Wang, Ding
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (10) : 2134 - 2144
  • [45] Model-free adaptive PID control for nonlinear discrete-time systems
    Zhang, Shuhua
    Chi, Ronghu
    TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2020, 42 (10) : 1797 - 1807
  • [46] H∞ Control for Discrete-time Linear Systems by Integrating Off-policy Q-learning and Zero-sum Game
    Li, Jinna
    Ding, Zhengtao
    Yang, Chunyu
    Niu, Hong
    2018 IEEE 14TH INTERNATIONAL CONFERENCE ON CONTROL AND AUTOMATION (ICCA), 2018, : 817 - 822
  • [47] ''Model-free'' stability analysis for discrete-time fuzzy control systems
    Cao, SG
    Rees, NW
    Feng, G
    PROCEEDINGS OF THE 35TH IEEE CONFERENCE ON DECISION AND CONTROL, VOLS 1-4, 1996, : 2725 - 2726
  • [48] Experience replay-based output feedback Q-learning scheme for optimal output tracking control of discrete-time linear systems
    Rizvi, Syed Ali Asad
    Lin, Zongli
    INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2019, 33 (12) : 1825 - 1842
  • [49] Output Feedback Q-Learning Control for the Discrete-Time Linear Quadratic Regulator Problem
    Rizvi, Syed Ali Asad
    Lin, Zongli
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (05) : 1523 - 1536
  • [50] Asynchronous iterative Q-learning based tracking control for nonlinear discrete-time multi-agent systems
    Shen, Ziwen
    Dong, Tao
    Huang, Tingwen
    NEURAL NETWORKS, 2024, 180