Convergence of policy gradient methods for finite-horizon exploratory linear-quadratic control problems
From MaRDI portal
Publication:6490237
DOI10.1137/22M1533517MaRDI QIDQ6490237FDOQ6490237
Authors: Michael Giegrich, C. Reisinger
Publication date: 23 April 2024
Published in: SIAM Journal on Control and Optimization (Search for Journal in Brave)
global linear convergencemesh-independent convergencepolicy optimizationcontinuous-time linear-quadratic controlgeometry-aware gradientrelative Entropy
Cites Work
- Continuous-time mean-variance portfolio selection: a stochastic LQ framework
- Title not available (Why is that?)
- Lectures on BSDEs, stochastic control, and stochastic differential games with financial applications
- Policy gradient in continuous time
- Linear quadratic mean field type control and mean field games with common noise, with application to production of an exhaustible resource
- Open-loop and closed-loop solvabilities for stochastic linear quadratic optimal control problems
- Learning Optimal Controllers for Linear Systems With Multiplicative Noise via Policy Gradient
- A theoretical and empirical comparison of gradient approximations in derivative-free optimization
- Title not available (Why is that?)
- Continuous‐time mean–variance portfolio selection: A reinforcement learning framework
- Exploratory LQG mean field games with entropy regularization
- Algorithmic trading, stochastic control, and mutually exciting processes
- Policy Gradient Methods for the Noisy Linear Quadratic Regulator over a Finite Horizon
- Policy optimization for \(\mathcal{H}_2\) linear control with \(\mathcal{H}_\infty\) robustness guarantee: implicit regularization and global convergence
Cited In (2)
This page was built for publication: Convergence of policy gradient methods for finite-horizon exploratory linear-quadratic control problems
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6490237)