The value iteration method for countable state Markov decision processes
From MaRDI portal
Publication:1306452
DOI10.1016/S0167-6377(99)00015-2zbMath0954.90060MaRDI QIDQ1306452
Publication date: 9 February 2001
Published in: Operations Research Letters (Search for Journal in Brave)
Markov decision processes; unbounded costs; optimal policy; countable state space; value-iteration method
90C40: Markov and semi-Markov decision processes
Related Items
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Value iteration in countable state average cost Markov decision processes with unbounded costs
- Stochastic optimal control. The discrete time case
- On strong average optimality of Markov decision processes with unbounded costs
- Comparing recent assumptions for the existence of average optimal stationary policies
- The convergence of value iteration in average cost Markov decision chains
- Denumerable Undiscounted Semi-Markov Decision Processes with Unbounded Rewards
- An Inventory Model with Limited Production Capacity and Uncertain Demands I. The Average-Cost Criterion
- Average Cost Optimal Stationary Policies in Infinite State Markov Decision Processes with Unbounded Costs
- On the Second Optimality Equation for Semi-Markov Decision Models
- The asymptotic behaviour of the minimal total expected cost for the denumerable state Markov decision model
- Computing the optimal policy for capacitated inventory models
- X-Y Band and Modified (s, S) Policy
- Bounds and Asymptotics for Planning Critical Safety Stocks
- Sensitivity Analysis for Base-Stock Levels in Multiechelon Production-Inventory Systems
- Stochastic Inventory Models with Limited Production Capacity and Periodically Varying Parameters
- Semi-Markov Decision Processes with Unbounded Rewards