Near optimality of quantized policies in stochastic control under weak continuity conditions

From MaRDI portal
Publication:892326

DOI10.1016/J.JMAA.2015.10.008zbMATH Open1323.93078arXiv1410.6985OpenAlexW2963272571MaRDI QIDQ892326FDOQ892326


Authors: Naci Saldi, Serdar Yüksel, Tamás Linder Edit this on Wikidata


Publication date: 18 November 2015

Published in: Journal of Mathematical Analysis and Applications (Search for Journal in Brave)

Abstract: This paper studies the approximation of optimal control policies by quantized (discretized) policies for a very general class of Markov decision processes (MDPs). The problem is motivated by applications in networked control systems, computational methods for MDPs, and learning algorithms for MDPs. We consider the finite-action approximation of stationary policies for a discrete-time Markov decision process with discounted and average costs under a weak continuity assumption on the transition probability, which is a significant relaxation of conditions required in earlier literature. The discretization is constructive, and quantized policies are shown to approximate optimal deterministic stationary policies with arbitrary precision. The results are applied to the fully observed reduction of a partially observed Markov decision process, where weak continuity is a much more reasonable assumption than more stringent conditions such as strong continuity or continuity in total variation.


Full work available at URL: https://arxiv.org/abs/1410.6985




Recommendations




Cites Work


Cited In (11)





This page was built for publication: Near optimality of quantized policies in stochastic control under weak continuity conditions

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q892326)