H∞ Tracking Control for Linear Discrete-Time Systems: Model-Free Q-Learning Designs

被引:35
|
作者
Yang, Yunjie [1 ]
Wan, Yan [2 ]
Zhu, Jihong [1 ]
Lewis, Frank L. [3 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Univ Texas Arlington, Dept Elect Engn, Arlington, TX 76019 USA
[3] Univ Texas Arlington, UTA Res Inst, Ft Worth, TX 75052 USA
来源
IEEE CONTROL SYSTEMS LETTERS | 2021年 / 5卷 / 01期
基金
中国国家自然科学基金;
关键词
Linear discrete-time systems; H-infinity tracking control; Q-learning; ZERO-SUM GAMES;
D O I
10.1109/LCSYS.2020.3001241
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this letter, a novel model-free Q-learning based approach is developed to solve the H-infinity tracking problem for linear discrete-time systems. A new exponential discounted value function is introduced that includes the cost of the whole control input and tracking error. The tracking Bellman equation and the game algebraic Riccati equation (GARE) are derived. The solution to the GARE leads to the feedback and feedforward parts of the control input. A Q-learning algorithm is then developed to learn the solution of the GARE online without requiring any knowledge of the system dynamics. Convergence of the algorithm is analyzed, and it is also proved that probing noises in maintaining the persistence of excitation (PE) condition do not result in any bias. An example of the F-16 aircraft short period dynamics is developed to validate the proposed algorithm.
引用
收藏
页码:175 / 180
页数:6
相关论文
共 50 条
  • [1] Model-Free Q-Learning for the Tracking Problem of Linear Discrete-Time Systems
    Li, Chun
    Ding, Jinliang
    Lewis, Frank L.
    Chai, Tianyou
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3191 - 3201
  • [2] Model-free H∞ control design for unknown linear discrete-time systems via Q-learning with LMI
    Kim, J. -H.
    Lewis, F. L.
    AUTOMATICA, 2010, 46 (08) : 1320 - 1326
  • [3] Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control
    Al-Tamimi, Asma
    Lewis, Frank L.
    Abu-Khalaf, Murad
    AUTOMATICA, 2007, 43 (03) : 473 - 481
  • [4] Optimal tracking control for discrete-time systems by model-free off-policy Q-learning approach
    Li, Jinna
    Yuan, Decheng
    Ding, Zhengtao
    2017 11TH ASIAN CONTROL CONFERENCE (ASCC), 2017, : 7 - 12
  • [5] Stochastic linear quadratic optimal control for model-free discrete-time systems based on Q-learning algorithm
    Wang, Tao
    Zhang, Huaguang
    Luo, Yanhong
    NEUROCOMPUTING, 2018, 312 : 1 - 8
  • [6] Reinforcement Q-Learning Algorithm for H∞ Tracking Control of Unknown Discrete-Time Linear Systems
    Peng, Yunjian
    Chen, Qian
    Sun, Weijie
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2020, 50 (11): : 4109 - 4122
  • [7] Model-free optimal tracking control for discrete-time system with delays using reinforcement Q-learning
    Liu, Yang
    Yu, Rui
    ELECTRONICS LETTERS, 2018, 54 (12) : 750 - 751
  • [8] Minimax Q-learning design for H∞ control of linear discrete-time systems
    Li, Xinxing
    Xi, Lele
    Zha, Wenzhong
    Peng, Zhihong
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2022, 23 (03) : 438 - 451
  • [9] Model-Free Algorithms for Containment Control of Saturated Discrete-Time Multiagent Systems via Q-Learning Method
    Long, Mingkang
    Su, Housheng
    Zeng, Zhigang
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (02): : 1308 - 1316
  • [10] Robust H8 tracking of linear discrete-time systems using Q-learning
    Valadbeigi, Amir Parviz
    Shu, Zhan
    Khaki Sedigh, Ali
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (10) : 5604 - 5623