Research Progress on Learning-based Robust Adaptive Critic Control

被引:0
|
作者
Wang D. [1 ,2 ]
机构
[1] Faculty of Information Technology, Beijing University of Technology, Beijing
[2] Beijing Key Laboratory of Computational Intelligence and Intelligent System, Beijing
来源
基金
中国国家自然科学基金;
关键词
Adaptive critic control; Intelligent learning; Neural networks; Robust control; Uncertain systems;
D O I
10.16383/j.aas.c170701
中图分类号
学科分类号
摘要
In the machine learning fleld, the core technique of artiflcial intelligence, reinforcement learning is a class of strategies focusing on learning during the interaction process between machine and environment. As an important branch of reinforcement learning, the adaptive critic technique is closely related to dynamic programming and optimization design. In order to efiectively solve optimal control problems of complex dynamical systems, the adaptive dynamic programming approach was proposed by combining adaptive critic, dynamic programming with artiflcial neural networks and has been attracted extensive attention. Particularly, great progress has been obtained on robust adaptive critic control design with uncertainties and disturbances. Now, it has been regarded as a necessary outlet to construct intelligent learning systems and achieve true brain-like intelligence. This paper presents a comprehensive survey on the learning-based robust adaptive critic control theory and methods, including self-learning robust stabilization, adaptive trajectory tracking, event-driven robust control, and adaptive H∞ control design. Therein, it covers a general analysis for adaptive critic systems in terms of stability, convergence, optimality, and robustness. In addition, considering novel techniques such as artiflcial intelligence, big data, deep learning, and knowledge automation, it also discusses future prospects of robust adaptive critic control. Copyright © 2019 Acta Automatica Sinica. All rights reserved.
引用
收藏
页码:1031 / 1043
页数:12
相关论文
共 110 条
  • [1] Silver D., Huang A., Maddison C.J., Guez A., Sifre L., Van Den Driessche G., Et al., Mastering the game of Go with deep neural networks and tree search, Nature, 529, 7587, pp. 484-489, (2016)
  • [2] LeCun Y., Bengio Y., Hinton G., Deep learning, Nature, 521, 7553, pp. 436-444, (2015)
  • [3] Schmidhuber J., Deep learning in neural networks: an overview, Neural Networks, 61, pp. 85-117, (2015)
  • [4] Haykin S., Neural Networks: A Comprehensive Foundation, (1999)
  • [5] Sutton R.S., Barto A.G., Reinforcement Learning: An Introduction, (1998)
  • [6] Silver D., Schrittwieser J., Simonyan K., Antonoglou I., Huang A., Guez A., Et al., Mastering the game of Go without human knowledge, Nature, 550, pp. 354-359, (2017)
  • [7] Bellman R.E., Dynamic Programming, (1957)
  • [8] Lewis F.L., Vrabie D., Syrmos V.L., Optimal Control (Third edition), (2012)
  • [9] Werbos P.J., Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences, (1974)
  • [10] Werbos P.J., Advanced forecasting methods for global crisis warning and models of intelligence, General Systems Yearbook, 22, 6, pp. 25-38, (1977)