The Complexity of Markov Decision Processes

From MaRDI portal
Publication:3780028

DOI10.1287/moor.12.3.441zbMath0638.90099OpenAlexW2032100464WikidataQ29038896 ScholiaQ29038896MaRDI QIDQ3780028

John N. Tsitsiklis, Christos H. Papadimitriou

Publication date: 1987

Published in: Mathematics of Operations Research (Search for Journal in Brave)

Full work available at URL: http://hdl.handle.net/1721.1/2893




Related Items (84)

What is decidable about partially observable Markov decision processes with \(\omega\)-regular objectivesMeeting a deadline: shortest paths on stochastic directed acyclic graphs with information gatheringProbabilistic Timed Automata with One Clock and Initialised Clock-Dependent ProbabilitiesThe Limitations of Optimization from SamplesA theory of strict P-completenessRuntime monitors for Markov decision processesPARTIALLY OBSERVABLE MARKOV DECISION PROCESSES AND PERIODIC POLICIES WITH APPLICATIONSWhen is a pair of matrices mortal?Approximation algorithms for stochastic combinatorial optimization problemsAn Incremental Fast Policy Search Using a Single Sample PathReachability analysis of quantum Markov decision processesProbabilistic planning with clear preferences on missing informationMyopic Bounds for Optimal Policy of POMDPs: An Extension of Lovejoy’s Structural ResultsGraph Games and Reactive SynthesisOn the complexity of partially observed Markov decision processesSolving H-horizon, stationary Markov decision problems in time proportional to log (H)Concavely-Priced Probabilistic Timed AutomataQuantitative verification and strategy synthesis for stochastic gamesA simulated annealing algorithm for the restricted stochastic traveling salesman problem with exponentially distributed arc lengthsVerifying Pufferfish privacy in hidden Markov modelsThe Simplex Method is Strongly Polynomial for Deterministic Markov Decision ProcessesModel Checking Linear-Time Properties of Probabilistic SystemsThe Lyapunov exponent and joint spectral radius of pairs of matrices are hard - when not impossible - to compute and to approximateCost-sensitive feature acquisition and classificationRemote state estimation with usage-dependent Markovian packet lossesOn the control of discrete-event dynamical systemsRobotic manipulation of multiple objects as a POMDPExact decomposition approaches for Markov decision processes: a surveyOptimal eviction policies for stochastic address tracesA mean-variance optimization problem for discounted Markov decision processesHybrid answer set programmingGraph planning with expected finite horizonMarkov Decision Processes with Incomplete Information and Semiuniform Feller Transition ProbabilitiesUnnamed Itemk-Certainty Exploration Method: an action selector to identify the environment in reinforcement learningApproximability and efficient algorithms for constrained fixed-horizon POMDPs with durative actionsA discrete-time optimal execution problem with market prices subject to random environmentsRisk-aware shielding of partially observable Monte Carlo planning policiesThe partially observable Markov decision processes in healthcare: an application to patients with ischemic heart disease (IHD)Future memories are not needed for large classes of POMDPsA survey of stochastic \(\omega \)-regular gamesOptimal Switching Sequence for Switched Linear SystemsFrom Reinforcement Learning to Deep Reinforcement Learning: An OverviewSeparation of learning and control for cyber-physical systemsOn the complexity of computational problems associated with simple stochastic gamesPartially observable Markov decision model for the treatment of early prostate cancerOn the computability of Solomonoff induction and AIXIUsing mathematical programming to solve factored Markov decision processes with imprecise probabilitiesExploiting symmetries for single- and multi-agent partially observable stochastic domainsDecentralized MDPs with sparse interactionsPageRank optimization by edge selectionUnnamed ItemA novel scheduling index rule proposal for QoE maximization in wireless networksLinear programming formulation for non-stationary, finite-horizon Markov decision process modelsNew complexity results about Nash equilibriaComputation of weighted sums of rewards for concurrent MDPsStrong planning under partial observabilityOn players with a bounded number of statesAlgorithms and conditional lower bounds for planning problemsRandomization for robot tasks: using dynamic programming in the space of knowledge statesA survey of computational complexity results in systems and controlA fast approximation method for partially observable Markov decision processesPartial-Observation Stochastic GamesProbabilistic Acceptors for Languages over Infinite WordsUnnamed ItemModel-based learning of interaction strategies in multi-agent systemsUnnamed ItemEmpirical Dynamic ProgrammingThe complexity of dynamic programmingReasoning about uncertain parameters and agent behaviors through encoded experiences and belief planningGame theory on attack graph for cyber deceptionOn the Complexity of Value IterationPartially observable Markov decision processes with imprecise parametersRobust Control of Partially Observable Failing SystemsA theory of strict P-completenessUnnamed ItemBayesian Decision Making in Groups is HardConstrained Multiagent Markov Decision Processes: a Taxonomy of Problems and AlgorithmsOn probabilistic timed automata.POMDPs under probabilistic semanticsOptimal decisions in stochastic graphs with uncorrelated and correlated edge weightsOptimal cost almost-sure reachability in POMDPsOn the undecidability of probabilistic planning and related stochastic optimization problemsSolving factored MDPs using non-homogeneous partitions




This page was built for publication: The Complexity of Markov Decision Processes