Accelerating Primal-Dual Methods for Regularized Markov Decision Processes

From MaRDI portal
Publication:6202767

DOI10.1137/21M1468851arXiv2202.10506MaRDI QIDQ6202767FDOQ6202767

Lexing Ying, Inderjit S. Dhillon, Haoya Li, Hsiang-Fu Yu

Publication date: 27 February 2024

Published in: SIAM Journal on Optimization (Search for Journal in Brave)

Abstract: Entropy regularized Markov decision processes have been widely used in reinforcement learning. This paper is concerned with the primal-dual formulation of the entropy regularized problems. Standard first-order methods suffer from slow convergence due to the lack of strict convexity and concavity. To address this issue, we first introduce a new quadratically convexified primal-dual formulation. The natural gradient ascent descent of the new formulation enjoys global convergence guarantee and exponential convergence rate. We also propose a new interpolating metric that further accelerates the convergence significantly. Numerical results are provided to demonstrate the performance of the proposed methods under multiple settings.


Full work available at URL: https://arxiv.org/abs/2202.10506







Cites Work






This page was built for publication: Accelerating Primal-Dual Methods for Regularized Markov Decision Processes

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6202767)