A \(K\)-step look-ahead analysis of value iteration algorithms for Markov decision processes
From MaRDI portal
Publication:1266643
DOI10.1016/0377-2217(94)00208-8zbMath0913.90261OpenAlexW1996660883MaRDI QIDQ1266643
Publication date: 31 May 1999
Published in: European Journal of Operational Research (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/0377-2217(94)00208-8
value iterationmodified policy iterationadaptive relaxation factordiscounted and undiscounted Markov decision processeslook-ahead analysis
Related Items
Cites Work
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Truncated policy iteration methods
- A simple technique in Markovian control with applications to resource allocation to resource allocation in communication networks
- Criteria for selecting the relaxation factor of the value iteration algorithm for undiscounted Markov and semi-Markov decision processes
- The convergence of value iteration in discounted Markov decision processes
- Computational comparison of value iteration algorithms for discounted Markov decision processes
- Iterative solution of the functional equations of undiscounted Markov renewal programming
- Action Elimination Procedures for Modified Policy Iteration Algorithms
- Bounds and Transformations for Discounted Finite Markov Decision Chains
- Branch-and-Bound Strategies for Dynamic Programming
- A set of successive approximation methods for discounted Markovian decision problems
- Technical Note—Accelerated Computation of the Expected Discounted Return in a Markov Chain
- Modified Policy Iteration Algorithms for Discounted Markov Decision Problems
- Discrete versions of an algorithm due to Varaiya
- Accelerating Procedures of the Value Iteration Algorithm for Discounted Markov Decision Processes, Based on a One-Step Lookahead Analysis
- Computing Optimal Policies for Controlled Tandem Queueing Systems
- Letter to the Editor—A Test for Suboptimal Actions in Markovian Decision Problems
- Technical Note—Undiscounted Markov Renewal Programming Via Modified Successive Approximations
- Some Bounds for Discounted Sequential Decision Processes