Simulation‐based Uniform Value Function Estimates of Markov Decision Processes
DOI10.1137/040619508zbMATH Open1119.90070OpenAlexW2084794933MaRDI QIDQ3593009FDOQ3593009
Publication date: 24 September 2007
Published in: SIAM Journal on Control and Optimization (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1137/040619508
Recommendations
- The value functions of Markov decision processes
- Simulation-based algorithms for Markov decision processes
- scientific article; zbMATH DE number 1509479
- Simulation-based optimization of Markov reward processes
- Uniform convergence of value iteration policies for discounted Markov decision processes
- Computing optimal policies for Markovian decision processes using simulation
- scientific article; zbMATH DE number 3916050
- Simulation-based optimization of Markov decision processes: an empirical process theory approach
- A class of procedures to compute the optimal value f unction in a Markovian decision problem
Markov decision processesMarkov gamesPAC learningempirical process theoryuniform rate of convergencevalue function estimation
Markov processes: estimation; hidden Markov models (62M05) Empirical decision procedures; empirical Bayes procedures (62C12) Markov and semi-Markov decision processes (90C40)
Cited In (6)
- A survey of some simulation-based algorithms for Markov decision processes
- CONVERGENCE OF SIMULATION-BASED POLICY ITERATION
- Title not available (Why is that?)
- Simulation-based optimization of Markov decision processes: an empirical process theory approach
- Near optimality of quantized policies in stochastic control under weak continuity conditions
- Empirical dynamic programming
This page was built for publication: Simulation‐based Uniform Value Function Estimates of Markov Decision Processes
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q3593009)