Optimal tracking control for completely unknown nonlinear discrete-time Markov jump systems using data-based reinforcement learning method

被引:31
|
作者
Jiang, He [1 ]
Zhang, Huaguang [1 ]
Luo, Yanhong [1 ]
Wang, Junyi [1 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Box 134, Shenyang 110819, Peoples R China
基金
中国国家自然科学基金;
关键词
Optimal tracking control; Markov jump systems; Data-based; Reinforcement learning; Adaptive dynamic programming; Neural networks; SYNCHRONIZATION CONTROL; GRAPHICAL GAMES; CONTROL SCHEME; STABILITY; ALGORITHM;
D O I
10.1016/j.neucom.2016.02.029
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we develop a novel optimal tracking control scheme for a class of nonlinear discrete-time Markov jump systems (MJSs) by utilizing a data-based reinforcement learning method. It is not practical to obtain accurate system models of the real-world MJSs due to the existence of abrupt variations in their system structures. Consequently, most traditional model-based methods for MJSs are invalid for the practical engineering applications. In order to overcome the difficulties without any identification scheme which would cause estimation errors, a model-free adaptive dynamic programming (ADP) algorithm will be designed by using system data rather than accurate system functions. Firstly, we combine the tracking error dynamics and reference system dynamics to form an augmented system. Then, based on the augmented system, a new performance index function with discount factor is formulated for the optimal tracking control problem via Markov chain and weighted sum technique. Neural networks are employed to implement the on-line ADP learning algorithm. Finally, a simulation example is given to demonstrate the effectiveness of our proposed approach. (C) 2016 Elsevier B.V. All rights reserved.
引用
收藏
页码:176 / 182
页数:7
相关论文
共 50 条
  • [1] Off-policy reinforcement learning for tracking control of discrete-time Markov jump linear systems with completely unknown dynamics
    Huang Z.
    Tu Y.
    Fang H.
    Wang H.
    Zhang L.
    Shi K.
    He S.
    Journal of the Franklin Institute, 2023, 360 (03) : 2361 - 2378
  • [2] Data-Based Optimal Tracking Control of Nonaffine Nonlinear Discrete-Time Systems
    Luo, Biao
    Liu, Derong
    Huang, Tingwen
    Li, Chao
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT IV, 2016, 9950 : 573 - 581
  • [3] Optimal Control for Unknown Discrete-Time Nonlinear Markov Jump Systems Using Adaptive Dynamic Programming
    Zhong, Xiangnan
    He, Haibo
    Zhang, Huaguang
    Wang, Zhanshan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (12) : 2141 - 2155
  • [4] Reinforcement learning-based optimal control for Markov jump systems with completely unknown dynamics
    Shi, Xiongtao
    Li, Yanjie
    Du, Chenglong
    Chen, Chaoyang
    Zong, Guangdeng
    Gui, Weihua
    AUTOMATICA, 2025, 171
  • [5] Optimal Tracking Control for Linear Discrete-time Systems Using Reinforcement Learning
    Kiumarsi-Khomartash, Bahare
    Lewis, Frank L.
    Naghibi-Sistani, Mohammad-Bagher
    Karimpour, Ali
    2013 IEEE 52ND ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2013, : 3845 - 3850
  • [6] Reinforcement Q-learning algorithm for H∞ tracking control of discrete-time Markov jump systems
    Shi, Jiahui
    He, Dakuo
    Zhang, Qiang
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2024,
  • [7] Data-driven finite-horizon optimal tracking control scheme for completely unknown discrete-time nonlinear systems
    Song, Ruizhuo
    Xie, Yulong
    Zhang, Zenglian
    NEUROCOMPUTING, 2019, 356 : 206 - 216
  • [8] H∞ Tracking learning control for discrete-time Markov jump systems: A parallel off-policy reinforcement learning
    Zhang, Xuewen
    Xia, Jianwei
    Wang, Jing
    Chen, Xiangyong
    Shen, Hao
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2023, 360 (18): : 14878 - 14890
  • [9] Fuzzy-Based Adaptive Optimization of Unknown Discrete-Time Nonlinear Markov Jump Systems With Off-Policy Reinforcement Learning
    Fang, Haiyang
    Tu, Yidong
    Wang, Hai
    He, Shuping
    Liu, Fei
    Ding, Zhengtao
    Cheng, Shing Shin
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2022, 30 (12) : 5276 - 5290
  • [10] Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics
    Kiumarsi, Bahare
    Lewis, Frank L.
    Modares, Hamidreza
    Karimpour, Ali
    Naghibi-Sistani, Mohammad-Bagher
    AUTOMATICA, 2014, 50 (04) : 1167 - 1175