The Optimal Reward Operator in Negative Dynamic Programming
From MaRDI portal
Publication:4027786
Recommendations
- scientific article; zbMATH DE number 3843524
- Nearly optimal stationary policies in negative dynamic programming
- The Bellman's principle of optimality in the discounted dynamic programming
- A Note on Positive Dynamic Programming
- scientific article; zbMATH DE number 1786117
- A note on negative dynamic programming for risk-sensitive control
- Finite state dynamic programming with the total reward criterion
- scientific article; zbMATH DE number 2189772
- On iterative optimization ol structured Markov decision processes with discounted rewards
Cited in
(7)- A mixed value and policy iteration method for stochastic control with universally measurable policies
- Finitely additive dynamic programming
- On convergence of value iteration for a class of total cost Markov decision processes
- scientific article; zbMATH DE number 3843524 (Why is no real title available?)
- A note on negative dynamic programming for risk-sensitive control
- Average cost optimality inequality for Markov decision processes with Borel spaces and universally measurable policies
- Ashok Prasad Maitra (1938-2008)
This page was built for publication: The Optimal Reward Operator in Negative Dynamic Programming
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4027786)