A New Value Iteration method for the Average Cost Dynamic Programming Problem
From MaRDI portal
Publication:4388932
Recommendations
- Policy iteration and Newton-Raphson methods for Markov decision processes under average cost criterion
- Value iteration for average cost Markov decision processes in Borel spaces
- Value iteration in countable state average cost Markov decision processes with unbounded costs
- The value iteration method for countable state Markov decision processes
- Value Iteration in a Class of Communicating Markov Decision Chains with the Average Cost Criterion
Cited in
(10)- Policy iteration and Newton-Raphson methods for Markov decision processes under average cost criterion
- Convex relaxations for permutation problems
- An empirical study of policy convergence in Markov decision process value iteration
- Inertial Newton algorithms avoiding strict saddle points
- Valuing programs with deterministic and stochastic cycles
- Model-based average reward reinforcement learning
- Policy iteration type algorithms for recurrent state Markov decision processes
- Analyzing anonymity attacks through noisy channels
- A unified approach to time-aggregated Markov decision processes
- Solving average cost Markov decision processes by means of a two-phase time aggregation algorithm
This page was built for publication: A New Value Iteration method for the Average Cost Dynamic Programming Problem
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4388932)