Accelerating Primal-Dual Methods for Regularized Markov Decision Processes
DOI10.1137/21M1468851arXiv2202.10506MaRDI QIDQ6202767FDOQ6202767
Authors: Haoya Li, Hsiang-Fu Yu, Lexing Ying, Inderjit S. Dhillon
Publication date: 27 February 2024
Published in: SIAM Journal on Optimization (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/2202.10506
Recommendations
- Approximate Newton Policy Gradient Algorithms
- A note on optimization formulations of Markov decision processes
- Fast global convergence of natural policy gradient methods with entropy regularization
- Entropy Regularization for Mean Field Games with Learning
- A First-Order Approach to Accelerated Value Iteration
Numerical optimization and variational techniques (65K10) Learning and adaptive systems in artificial intelligence (68T05) Minimax problems in mathematical programming (90C47) Markov and semi-Markov decision processes (90C40) Lyapunov and storage functions (93D30) Acceleration of convergence in numerical analysis (65B99)
Cites Work
- Title not available (Why is that?)
- Simple statistical gradient-following algorithms for connectionist reinforcement learning
- Reinforcement learning. An introduction
- 10.1162/jmlr.2003.3.4-5.803
- A note on optimization formulations of Markov decision processes
- Policy mirror descent for reinforcement learning: linear convergence, new sampling complexity, and generalized problem classes
- Approximate Newton Policy Gradient Algorithms
- Randomized linear programming solves the Markov decision problem in nearly linear (sometimes sublinear) time
Cited In (1)
This page was built for publication: Accelerating Primal-Dual Methods for Regularized Markov Decision Processes
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6202767)