ON GRADUAL-IMPULSE CONTROL OF CONTINUOUS-TIME MARKOV DECISION PROCESSES WITH EXPONENTIAL UTILITY

被引:2
|
作者
Guo, Xin [1 ]
Kurushima, Aiko [2 ]
Piunovskiy, Alexey [3 ]
Zhang, Yi [3 ]
机构
[1] Tsinghua Univ, Sch Econ & Management, Beijing 100084, Peoples R China
[2] Sophia Univ, Dept Econ, Chiyoda Ku, 7-1 Kioi Cho, Tokyo 1028554, Japan
[3] Univ Liverpool, Dept Math Sci, Liverpool L69 72L, Merseyside, England
基金
英国工程与自然科学研究理事会;
关键词
Continuous-time Markov decision processes; dynamic programming; gradual-impulse control; optimality equation; RISK-SENSITIVE CONTROL; COST;
D O I
10.1017/apr.2020.64
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We consider a gradual-impulse control problem of continuous-time Markov decision processes, where the system performance is measured by the expectation of the exponential utility of the total cost. We show, under natural conditions on the system primitives, the existence of a deterministic stationary optimal policy out of a more general class of policies that allow multiple simultaneous impulses, randomized selection of impulses with random effects, and accumulation of jumps. After characterizing the value function using the optimality equation, we reduce the gradual-impulse control problem to an equivalent simple discrete-time Markov decision process, whose action space is the union of the sets of gradual and impulsive actions.
引用
收藏
页码:301 / 334
页数:34
相关论文
共 50 条