ON THE FIRST PASSAGE g-MEAN-VARIANCE OPTIMALITY FOR DISCOUNTED CONTINUOUS-TIME MARKOV DECISION PROCESSES

被引:8
|
作者
Guo, Xianping [1 ]
Huang, Xiangxiang [1 ]
Zhang, Yi [2 ]
机构
[1] Sun Yat Sen Univ, Sch Math & Computat Sci, Guangzhou 510275, Guangdong, Peoples R China
[2] Univ Liverpool, Dept Math Sci, Liverpool L69 7ZL, Merseyside, England
关键词
continuous-time Markov decision processes; state-action-dependent discount factors; first passage mean-optimality; first passage g-mean-based variance minimization; PORTFOLIO SELECTION; STATE; CRITERION;
D O I
10.1137/140968872
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper considers the discounted continuous-time Markov decision processes (MDPs) in Borel spaces and with unbounded transition rates. The discount factors are allowed to depend on states and actions. Main attention is concentrated on the set F-g of stationary policies attaining a given mean performance g up to the first passage of the continuous-time MDP to an arbitrarily fixed target set. Under suitable conditions, we prove the existence of a g-mean-variance optimal policy that minimizes the first passage variance over the set F-g using a transformation technique, and also give the value iteration and policy iteration algorithms for computing the g-variance value function and a g-mean-variance optimal policy, respectively. Two examples are analytically solved to demonstrate the application of our results.
引用
收藏
页码:1406 / 1424
页数:19
相关论文
共 50 条