Value iteration and rolling plans for Markov control processes with unbounded rewards (Q1260895)
From MaRDI portal
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Value iteration and rolling plans for Markov control processes with unbounded rewards |
scientific article |
Statements
Value iteration and rolling plans for Markov control processes with unbounded rewards (English)
0 references
5 September 1993
0 references
The purpose is to extend known results for discounted Markov decision processes to the convergence of the value-iteration and the existence of error bounds for rolling horizon procedures to the case of a general state space and unbounded rewards. Now the error bounds are pointwise [w.r.t. the initial states] in contrast to the known uniform bounds. Uniformness is then obtained by using weighted norms. Further, under a strong ergodicity condition the bounds can be improved. The condition assumes a positive measure as lower bound for the distributions of states.
0 references
discounted Markov decision processes
0 references
convergence of the value-iteration
0 references
strong ergodicity condition
0 references