Discounted continuous-time Markov decision processes with unbounded rates and randomized history-dependent policies: the dynamic programming approach

From MaRDI portal
Publication:457293

DOI10.1007/S10288-013-0236-1zbMATH Open1297.90174arXiv1103.0134OpenAlexW2016588739MaRDI QIDQ457293FDOQ457293

A. B. Piunovskiy, Yi Zhang

Publication date: 26 September 2014

Published in: 4OR (Search for Journal in Brave)

Abstract: This paper deals with unconstrained discounted continuous-time Markov decision processes in Borel state and action spaces. Under some conditions imposed on the primitives, allowing unbounded transition rates and unbounded (from both above and below) cost rates, we show the regularity of the controlled process, which ensures the underlying models to be well defined. Then we develop the dynamic programming approach by showing that the Bellman equation is satisfied (by the optimal value). Finally, under some compactness-continuity conditions, we obtain the existence of a deterministic stationary optimal policy out of the class of randomized history-dependent policies.


Full work available at URL: https://arxiv.org/abs/1103.0134




Recommendations




Cites Work


Cited In (17)





This page was built for publication: Discounted continuous-time Markov decision processes with unbounded rates and randomized history-dependent policies: the dynamic programming approach

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q457293)