Approximate receding horizon approach for Markov decision processes: average reward case (Q1414220): Difference between revisions

From MaRDI portal
Created claim: MaRDI profile type (P1460): MaRDI publication profile (Q5976449), #quickstatements; #temporary_batch_1710670515113
ReferenceBot (talk | contribs)
Changed an Item
 
Property / cites work
 
Property / cites work: Discrete-Time Controlled Markov Processes with Average Cost Criterion: A Survey / rank
 
Normal rank
Property / cites work
 
Property / cites work: Enlarging the region of convergence of Newton's method for constrained optimization / rank
 
Normal rank
Property / cites work
 
Property / cites work: Rollout algorithms for stochastic scheduling problems / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4257216 / rank
 
Normal rank
Property / cites work
 
Property / cites work: An on-line procedure in discounted infinite-horizon stochastic optimal control / rank
 
Normal rank
Property / cites work
 
Property / cites work: On the structure of value functions for threshold policies in queueing models / rank
 
Normal rank
Property / cites work
 
Property / cites work: Moving horizon control in dynamic games / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q3237805 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q5688016 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Finite horizon approximations of infinite horizon linear programs / rank
 
Normal rank
Property / cites work
 
Property / cites work: Error bounds for rolling horizon policies in discrete-time Markov control processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: A forecast horizon and a stopping rule for general Markov decision processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Adaptive Markov control processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4002098 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Optimal infinite-horizon feedback laws for a general class of constrained discrete-time systems: Stability and moving-horizon approximations / rank
 
Normal rank
Property / cites work
 
Property / cites work: On the value function of a priority queue with an application to a controlled polling model / rank
 
Normal rank
Property / cites work
 
Property / cites work: A probabilistic analysis of bias optimality in unichain Markov decision processes / rank
 
Normal rank
Property / cites work
 
Property / cites work: Receding horizon control of nonlinear systems / rank
 
Normal rank
Property / cites work
 
Property / cites work: The policy iteration algorithm for average reward Markov decision processes with general state space / rank
 
Normal rank
Property / cites work
 
Property / cites work: Complexity of finite-horizon Markov decision process problems / rank
 
Normal rank
Property / cites work
 
Property / cites work: Separable routing: A scheme for state-dependent routing of circuit switched telephone traffic / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4386528 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Optimal Infinite-Horizon Undiscounted Control of Finite Probabilistic Systems / rank
 
Normal rank
Property / cites work
 
Property / cites work: Q4315289 / rank
 
Normal rank
Property / cites work
 
Property / cites work: Comparing neuro-dynamic programming algorithms for the vehicle routing problem with stochastic demands / rank
 
Normal rank
Property / cites work
 
Property / cites work: The Determination of Approximately Optimal Policies in Markov Decision Processes by the Use of Bounds / rank
 
Normal rank

Latest revision as of 12:45, 6 June 2024

scientific article
Language Label Description Also known as
English
Approximate receding horizon approach for Markov decision processes: average reward case
scientific article

    Statements

    Approximate receding horizon approach for Markov decision processes: average reward case (English)
    0 references
    0 references
    0 references
    20 November 2003
    0 references
    The authors consider an approximation scheme for solving Markov decision processes (MDPs) with countable state space, finite action space, and bounded rewards that uses an approximate solution of a fixed finite-horizon sub-MDP of a given infinite-horizon MDP to create a stationary policy, which they call ''approximate receding horizon control''. They analyze the performance of the approximate receding horizon control in some conditions, study two examples, also provide a simple proof on the policy improvement for countable state space, and discuss practical implementations of these schemes via simulation.
    0 references
    0 references
    Markov decision process
    0 references
    receding horizon control
    0 references
    Infinite-horizon average reward
    0 references
    policy improvement
    0 references
    ergodicity
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references
    0 references