Research Progress on Learning-based Robust Adaptive Critic Control

被引:0
|
作者
Wang D. [1 ,2 ]
机构
[1] Faculty of Information Technology, Beijing University of Technology, Beijing
[2] Beijing Key Laboratory of Computational Intelligence and Intelligent System, Beijing
来源
基金
中国国家自然科学基金;
关键词
Adaptive critic control; Intelligent learning; Neural networks; Robust control; Uncertain systems;
D O I
10.16383/j.aas.c170701
中图分类号
学科分类号
摘要
In the machine learning fleld, the core technique of artiflcial intelligence, reinforcement learning is a class of strategies focusing on learning during the interaction process between machine and environment. As an important branch of reinforcement learning, the adaptive critic technique is closely related to dynamic programming and optimization design. In order to efiectively solve optimal control problems of complex dynamical systems, the adaptive dynamic programming approach was proposed by combining adaptive critic, dynamic programming with artiflcial neural networks and has been attracted extensive attention. Particularly, great progress has been obtained on robust adaptive critic control design with uncertainties and disturbances. Now, it has been regarded as a necessary outlet to construct intelligent learning systems and achieve true brain-like intelligence. This paper presents a comprehensive survey on the learning-based robust adaptive critic control theory and methods, including self-learning robust stabilization, adaptive trajectory tracking, event-driven robust control, and adaptive H∞ control design. Therein, it covers a general analysis for adaptive critic systems in terms of stability, convergence, optimality, and robustness. In addition, considering novel techniques such as artiflcial intelligence, big data, deep learning, and knowledge automation, it also discusses future prospects of robust adaptive critic control. Copyright © 2019 Acta Automatica Sinica. All rights reserved.
引用
收藏
页码:1031 / 1043
页数:12
相关论文
共 110 条
  • [31] Bian T., Jiang Z.P., Value iteration and adaptive dynamic programming for data-driven adaptive optimal control design, Automatica, 71, pp. 348-360, (2016)
  • [32] Lee J.Y., Park J.B., Choi Y.H., Integral reinforcement learning for continuous-time input-a-ne nonlinear systems with simultaneous invariant explorations, IEEE Transactions on Neural Networks and Learning Systems, 26, 5, pp. 916-932, (2015)
  • [33] Ha M.M., Wang D., Liu D.R., Event-triggered adaptive critic control design for discrete-time constrained nonlinear systems, IEEE Transactions on Systems, Man and Cybernetics: Systems, (2019)
  • [34] Wang F.Y., Zhang H.G., Liu D.R., Adaptive dynamic programming: an introduction, IEEE Computational Intelligence Magazine, 4, 2, pp. 39-47, (2009)
  • [35] Lewis F.L., Liu D.R., Reinforcement Learning and Approximate Dynamic Programming for Feedback Control, (2012)
  • [36] Zhang H.G., Liu D.R., Luo Y.H., Wang D., Adaptive Dynamic Programming for Control: Algorithms and Stability, (2013)
  • [37] Zhang H.-G., Zhang X., Luo Y.-H., Yang J., An overview of research on adaptive dynamic programming, Acta Automatica Sinica, 39, 4, pp. 303-311, (2013)
  • [38] Liu D.-R., Li H.-L., Wang D., Data-based selflearning optimal control: research progress and prospects, Acta Automatica Sinica, 39, 11, pp. 1858-1870, (2013)
  • [39] Wang D., He H.B., Liu D.R., Adaptive critic nonlinear robust control: a survey, IEEE Transactions on Cybernetics, 47, 10, pp. 3429-3451, (2017)
  • [40] Wang D., Mu C.X., Adaptive Critic Control with Robust Stabilization for Uncertain Nonlinear Systems, (2019)