Approximation of average cost Markov decision processes using empirical distributions and concentration inequalities
From MaRDI portal
Publication:5265786
DOI10.1080/17442508.2014.939979zbMATH Open1317.90317OpenAlexW2011519669MaRDI QIDQ5265786FDOQ5265786
TomΓ‘s Prieto-Rumeau, F. Dufour
Publication date: 29 July 2015
Published in: Stochastics (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1080/17442508.2014.939979
Wasserstein distanceconcentration inequalitiesMarkov decision processeslong-run average costapproximation of the optimal value and an optimal policy
Cites Work
- Title not available (Why is that?)
- A time aggregation approach to Markov decision processes
- Approximate Dynamic Programming
- OnActor-Critic Algorithms
- Simulation-based optimization of Markov reward processes
- Discrete-Time Controlled Markov Processes with Average Cost Criterion: A Survey
- Finite linear programming approximations of constrained discounted Markov decision processes
- Simple bounds for the convergence of empirical and occupation measures in 1-Wasserstein distance
- Learning algorithms for Markov decision processes with average cost
- Approximation of Markov decision processes with general state space
- CONVERGENCE OF SIMULATION-BASED POLICY ITERATION
- Approximate gradient methods in policy-space optimization of Markov reward processes
- Approximate receding horizon approach for Markov decision processes: average reward case
- Average optimality for Markov decision processes in borel spaces: a new condition and approach
- Policy iteration for average cost Markov control processes on Borel spaces
- Convergence Results for Some Temporal Difference Methods Based on Least Squares
- Average cost Markov control processes: Stability with respect to the Kantorovich metric
- A policy improvement method for constrained average Markov decision processes
- Universal Reinforcement Learning
Cited In (10)
- The average cost of Markov chains subject to total variation distance uncertainty
- Approximation of discounted minimax Markov control problems and zero-sum Markov games using Hausdorff and Wasserstein distances
- Robustness to incorrect models and data-driven learning in average-cost optimal stochastic control
- Optimal deterministic controller synthesis from steady-state distributions
- Computable approximations for continuous-time Markov decision processes on Borel spaces based on empirical measures
- From Infinite to Finite Programs: Explicit Error Bounds with Applications to Approximate Dynamic Programming
- A stability result for linear Markovian stochastic optimization problems
- A perturbation approach to approximate value iteration for average cost Markov decision processes with Borel spaces and bounded costs
- A convex optimization approach to dynamic programming in continuous state and action spaces
- Computable approximations for average Markov decision processes in continuous time
Recommendations
- Approximation of average cost optimal policies for general Markov decision processes with unbounded costs π π
- Exact finite approximations of average-cost countable Markov decision processes π π
- Average cost Markov decision processes with weakly continuous transition probabilities π π
- Average cost Markov decision processes under the hypothesis of Doeblin π π
- Finite-State Approximations to Discounted and Average Cost Constrained Markov Decision Processes π π
- Average Cost Markov Decision Processes with Semi-Uniform Feller Transition Probabilities π π
- Average cost Markov decision processes: Optimality conditions π π
- The convergence of value iteration in average cost Markov decision chains π π
This page was built for publication: Approximation of average cost Markov decision processes using empirical distributions and concentration inequalities
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5265786)