Dynamic programming equations for discounted constrained stochastic control

被引:29
|
作者
Chen, RC [1 ]
Blankenship, GL
机构
[1] USN, Res Lab, Div Radar, Washington, DC 20375 USA
[2] Univ Maryland, Dept Elect Engn, College Pk, MD 20742 USA
关键词
constrained Markov decision processes; dynamic programming; pure policies; stochastic control;
D O I
10.1109/TAC.2004.826725
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, the application of the dynamic programming approach to constrained stochastic control problems with expected value constraints is demonstrated. Specifically, two such problems are analyzed using this approach. The problems analyzed are the problem of minimizing a discounted cost infinite horizon expectation objective subject to an identically structured constraint, and the problem of minimizing a discounted cost infinite horizon minimax objective subject to a discounted expectation constraint. Using the dynamic programming approach, optimality equations, which are the chief contribution of this paper, are obtained for these problems. In particular, the dynamic programming operators for problems with expectation constraints differ significantly from those of standard dynamic programming and problems with worst-case constraints. For the discounted cost infinite horizon cases, existence and uniqueness of solutions to the dynamic programming equations are explicitly shown by using the Banach fixed point theorem to show that the corresponding dynamic programming operators are contractions. The theory developed is illustrated by numerically solving the constrained stochastic control dynamic programming equations derived for simple example problems. The example problems are based on a two-state Markov model that represents an error prone system that is to be maintained.
引用
收藏
页码:699 / 709
页数:11
相关论文
共 50 条