Approximate Dynamic Programming for Self-Learning Control

被引:6
|
作者
Derong Liu (Department of Electrical and Computer Engineering
机构
基金
美国国家科学基金会;
关键词
Approximate dynamic programming; learning control; neural networks;
D O I
10.16383/j.aas.2005.01.002
中图分类号
TP273.22 [];
学科分类号
080201 ; 0835 ;
摘要
This paper introduces a self-learning control approach based on approximate dynamic programming. Dynamic programming was introduced by Bellman in the 1950’s for solving optimal control problems of nonlinear dynamical systems. Due to its high computational complexity, the applications of dynamic programming have been limited to simple and small problems. The key step in finding approximate solutions to dynamic programming is to estimate the performance index in dynamic programming. The optimal control signal can then be determined by minimizing (or maximizing) the performance index. Artificial neural networks are very efficient tools in representing the performance index in dynamic programming. This paper assumes the use of neural networks for estimating the performance index in dynamic programming and for generating optimal control signals, thus to achieve optimal control through self-learning.
引用
收藏
页码:13 / 18
页数:6
相关论文
共 3 条
  • [1] Technical Note: Q-Learning[J] . Christopher J.C.H. Watkins,Peter Dayan.Machine Learning . 1992 (3)
  • [2] Learning to predict by the methods of temporal differences[J] . Richard S. Sutton.Machine Learning . 1988 (1)
  • [3] Intelligent engine control using an adaptive critic .2 Kulkarni N V,KrishnaKumar K. IEEE Transactions on Control Systems Technology . 2003