A New Value Iteration method for the Average Cost Dynamic Programming Problem
From MaRDI portal
Publication:4388932
Recommendations
- Policy iteration and Newton-Raphson methods for Markov decision processes under average cost criterion
- Value iteration for average cost Markov decision processes in Borel spaces
- Value iteration in countable state average cost Markov decision processes with unbounded costs
- The value iteration method for countable state Markov decision processes
- Value Iteration in a Class of Communicating Markov Decision Chains with the Average Cost Criterion
Cited in
(10)- A unified approach to time-aggregated Markov decision processes
- Inertial Newton algorithms avoiding strict saddle points
- Convex relaxations for permutation problems
- Policy iteration type algorithms for recurrent state Markov decision processes
- Solving average cost Markov decision processes by means of a two-phase time aggregation algorithm
- Model-based average reward reinforcement learning
- An empirical study of policy convergence in Markov decision process value iteration
- Valuing programs with deterministic and stochastic cycles
- Analyzing anonymity attacks through noisy channels
- Policy iteration and Newton-Raphson methods for Markov decision processes under average cost criterion
This page was built for publication: A New Value Iteration method for the Average Cost Dynamic Programming Problem
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4388932)