Near-Optimal Time and Sample Complexities for Solving Discounted Markov Decision Process with a Generative Model
From MaRDI portal
Publication:6302549
Abstract: In this paper we consider the problem of computing an -optimal policy of a discounted Markov Decision Process (DMDP) provided we can only access its transition function through a generative sampling model that given any state-action pair samples from the transition function in time. Given such a DMDP with states , actions , discount factor , and rewards in range we provide an algorithm which computes an -optimal policy with probability where emph{both} the time spent and number of sample taken are upper bounded by [ Oleft[frac{|S||A|}{(1-gamma)^3 epsilon^2} log left(frac{|S||A|}{(1-gamma)delta epsilon}
ight) logleft(frac{1}{(1-gamma)epsilon}
ight)
ight] ~. ] For fixed values of , this improves upon the previous best known bounds by a factor of and matches the sample complexity lower bounds proved in Azar et al. (2013) up to logarithmic factors. We also extend our method to computing -optimal policies for finite-horizon MDP with a generative model and provide a nearly matching sample complexity lower bound.
This page was built for publication: Near-Optimal Time and Sample Complexities for Solving Discounted Markov Decision Process with a Generative Model
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6302549)