Convergence of policy gradient methods for finite-horizon exploratory linear-quadratic control problems
From MaRDI portal
Publication:6490237
Cites work
- scientific article; zbMATH DE number 1325009 (Why is no real title available?)
- scientific article; zbMATH DE number 7307478 (Why is no real title available?)
- A theoretical and empirical comparison of gradient approximations in derivative-free optimization
- Algorithmic trading, stochastic control, and mutually exciting processes
- Continuous-time mean-variance portfolio selection: a stochastic LQ framework
- Continuous‐time mean–variance portfolio selection: A reinforcement learning framework
- Exploratory LQG mean field games with entropy regularization
- Learning Optimal Controllers for Linear Systems With Multiplicative Noise via Policy Gradient
- Lectures on BSDEs, stochastic control, and stochastic differential games with financial applications
- Linear quadratic mean field type control and mean field games with common noise, with application to production of an exhaustible resource
- Open-loop and closed-loop solvabilities for stochastic linear quadratic optimal control problems
- Policy Gradient Methods for the Noisy Linear Quadratic Regulator over a Finite Horizon
- Policy gradient in continuous time
- Policy optimization for \(\mathcal{H}_2\) linear control with \(\mathcal{H}_\infty\) robustness guarantee: implicit regularization and global convergence
Cited in
(2)
This page was built for publication: Convergence of policy gradient methods for finite-horizon exploratory linear-quadratic control problems
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6490237)