A counterexample on sample-path optimality in stable Markov decision chains with the average reward criterion
From MaRDI portal
Publication:481787
DOI10.1007/s10957-013-0474-6zbMath1302.90241OpenAlexW1999785104WikidataQ124799376 ScholiaQ124799376MaRDI QIDQ481787
Karel Sladký, Rolando Cavazos-Cadena, Raúl Montes-De-oca
Publication date: 15 December 2014
Published in: Journal of Optimization Theory and Applications (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1007/s10957-013-0474-6
expected average reward criterionLyapunov function conditionstationary policystrong sample-path optimality
Related Items
Sample-Path Optimal Stationary Policies in Stable Markov Decision Chains with the Average Reward Criterion ⋮ Discrete-time zero-sum games for Markov chains with risk-sensitive average cost criterion ⋮ Economic design of memory-type control charts: the fallacy of the formula proposed by Lorenzen and Vance (1986)
Cites Work
- Sample path optimality for a Markov optimization problem
- Sample-path average optimality for Markov control processes
- Sample-Path Optimality in Average Markov Decision Chains Under a Double Lyapunov Function Condition
- Denumerable controlled Markov chains with average reward criterion: Sample path optimality
- Unnamed Item
- Unnamed Item
- Unnamed Item
This page was built for publication: A counterexample on sample-path optimality in stable Markov decision chains with the average reward criterion