The Simplex and Policy-Iteration Methods Are Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate

From MaRDI portal
Publication:2884291

DOI10.1287/moor.1110.0516zbMath1245.90140OpenAlexW1984901446MaRDI QIDQ2884291

Yinyu Ye

Publication date: 24 May 2012

Published in: Mathematics of Operations Research (Search for Journal in Brave)

Full work available at URL: https://doi.org/10.1287/moor.1110.0516




Related Items (40)

A complexity analysis of policy iteration through combinatorial matrices arising from unique sink orientationsThe greedy strategy for optimizing the Perron eigenvalueMore bounds on the diameters of convex polytopesOn Augmentation Algorithms for Linear and Integer-Linear Programming: From Edmonds--Karp to Bland and BeyondImproved and Generalized Upper Bounds on the Complexity of Policy IterationContinue, quit, restart probability modelThe Simplex Method is Strongly Polynomial for Deterministic Markov Decision ProcessesThe value iteration algorithm is not strongly polynomial for discounted dynamic programmingComplexity bounds for approximately solving discounted MDPs by value iterationsReformulation of the linear program for completely ergodic MDPs with average cost criteriaThe stochastic shortest path problem: a polyhedral combinatorics perspectiveOn the Number of Solutions Generated by the Simplex Method for LPRandom search for constrained Markov decision processes with multi-policy improvementUniform Turnpike Theorems for Finite Markov Decision ProcessesDual Ascent and Primal-Dual Algorithms for Infinite-Horizon Nonstationary Markov Decision ProcessesUnnamed ItemJoint chance-constrained Markov decision processesStrong polynomiality of the Gass-Saaty shadow-vertex pivoting rule for controlled random walksRandomized Linear Programming Solves the Markov Decision Problem in Nearly Linear (Sometimes Sublinear) TimeOn the number of solutions generated by the dual simplex methodA Friendly Smoothed Analysis of the Simplex MethodA double-pivot simplex algorithm and its upper bounds of the iteration numbersTowards solving 2-TBSG efficientlyStrong polynomiality of policy iterations for average-cost MDPs modeling replacement and maintenance problemsOptimal schedulers vs optimal bases: an approach for efficient exact solving of Markov decision processesWhat Tropical Geometry Tells Us about the Complexity of Linear ProgrammingThe double pivot simplex methodPolicy iteration for robust nonstationary Markov decision processesModified policy iteration algorithms are not strongly polynomial for discounted dynamic programmingA primal-simplex based Tardos' algorithmImproved bound on the worst case complexity of policy iterationComputing Kitahara-Mizuno's bound on the number of basic feasible solutions generated with the simplex algorithmComments on: Recent progress on the combinatorial diameter of polytopes and simplicial complexesVariance Reduced Value Iteration and Faster Algorithms for Solving Markov Decision ProcessesEfficient computation of a canonical form for a matrix with the generalized P-propertyThe simplex method using Tardos' basic algorithm is strongly polynomial for totally unimodular LP under nondegeneracy assumptionThe operator approach to entropy gamesOn the reduction of total‐cost and average‐cost MDPs to discounted MDPsMultiply Accelerated Value Iteration for NonSymmetric Affine Fixed Point Problems and Application to Markov Decision ProcessesA scaling-invariant algorithm for linear programming whose running time depends only on the constraint matrix




This page was built for publication: The Simplex and Policy-Iteration Methods Are Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate