Value iteration and rolling plans for Markov control processes with unbounded rewards (Q1260895)

From MaRDI portal
scientific article
Language Label Description Also known as
English
Value iteration and rolling plans for Markov control processes with unbounded rewards
scientific article

    Statements

    Value iteration and rolling plans for Markov control processes with unbounded rewards (English)
    0 references
    5 September 1993
    0 references
    The purpose is to extend known results for discounted Markov decision processes to the convergence of the value-iteration and the existence of error bounds for rolling horizon procedures to the case of a general state space and unbounded rewards. Now the error bounds are pointwise [w.r.t. the initial states] in contrast to the known uniform bounds. Uniformness is then obtained by using weighted norms. Further, under a strong ergodicity condition the bounds can be improved. The condition assumes a positive measure as lower bound for the distributions of states.
    0 references
    0 references
    discounted Markov decision processes
    0 references
    convergence of the value-iteration
    0 references
    strong ergodicity condition
    0 references
    0 references