Approximating Markov decision processes using expected state transitions (Q1075954): Difference between revisions

From MaRDI portal
RedirectionBot (talk | contribs)
Removed claim: author (P16): Item:Q181183
RedirectionBot (talk | contribs)
Changed an Item
Property / author
 
Property / author: Douglas J. White / rank
 
Normal rank

Revision as of 08:16, 10 February 2024

scientific article
Language Label Description Also known as
English
Approximating Markov decision processes using expected state transitions
scientific article

    Statements

    Approximating Markov decision processes using expected state transitions (English)
    0 references
    1985
    0 references
    For discounted Markov decision processes approximative solutions are obtained when in the optimality equation the expectation is replaced by the value at the expected new state. The approximate value and the true value for the approximative policy are compared to the unknown optimal value by establishing various types of bounds using e.g. Lipschitz conditions, Taylor expansions, convexity, monotonicity, and normal approximation. These bounds are demonstrated by a random walk model and by an inventory problem.
    0 references
    expected state approximation
    0 references
    discounted Markov decision processes approximative solutions
    0 references
    bounds
    0 references
    random walk model
    0 references
    0 references

    Identifiers