Exponential Convergence and Stability of Howard's Policy Improvement Algorithm for Controlled Diffusions
DOI10.1137/19M1236758zbMATH Open1441.93343arXiv1812.07846WikidataQ114978697 ScholiaQ114978697MaRDI QIDQ5111071FDOQ5111071
Lukasz Szpruch, B. Kerimkulov, David Šiška
Publication date: 26 May 2020
Published in: SIAM Journal on Control and Optimization (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1812.07846
Stochastic ordinary differential equations (aspects of stochastic analysis) (60H10) Optimal stochastic control (93E20) Exponential stability (93D23)
Cites Work
- Title not available (Why is that?)
- Title not available (Why is that?)
- Continuous-time stochastic control and optimization with financial applications
- Controlled Markov processes and viscosity solutions
- Some Convergence Results for Howard's Algorithm
- Markovian quadratic and superquadratic BSDEs with an unbounded terminal condition
- Infinite horizon backward stochastic differential equations and elliptic equations in Hilbert spaces.
- FUNCTIONAL EQUATIONS IN THE THEORY OF DYNAMIC PROGRAMMING. V. POSITIVITY AND QUASI-LINEARITY
- Average Optimality in Markov Control Processes via Discounted-Cost Problems and Linear Programming
- SOME NEW RESULTS IN THE THEORY OF CONTROLLED DIFFUSION PROCESSES
- On the Convergence of Policy Iteration in Stationary Dynamic Programming
- Control improvement for jump-diffusion processes with applications to finance
- The rate of convergence of finite-difference approximations for parabolic bellman equations with Lipschitz coefficients in cylindrical domains
- On the convergence of policy iteration for controlled diffusions
- Convergence Properties of Policy Iteration
- On finite-difference approximations for normalized Bellman equations
- On the policy improvement algorithm in continuous time
Cited In (13)
- Rates of convergence for the policy iteration method for mean field games systems
- Gradient flows for regularized stochastic control problems
- Exploratory LQG mean field games with entropy regularization
- A modified MSA for stochastic control problems
- Linear Convergence of a Policy Gradient Method for Some Finite Horizon Continuous Time Control Problems
- A Modified Method of Successive Approximations for Stochastic Recursive Optimal Control Problems
- A neural network-based policy iteration algorithm with global \(H^2\)-superlinear convergence for stochastic games on domains
- The modified MSA, a gradient flow and convergence
- Market based mechanisms for incentivising exchange liquidity provision
- A policy iteration method for mean field games
- Reinforcement Learning for Linear-Convex Models with Jumps via Stability Analysis of Feedback Controls
- Improved order 1/4 convergence for piecewise constant policy approximation of stochastic control problems
- Policy iteration method for time-dependent mean field games systems with non-separable Hamiltonians
This page was built for publication: Exponential Convergence and Stability of Howard's Policy Improvement Algorithm for Controlled Diffusions
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5111071)