Follow the flow: Proximal flow inspired multi-step methods
From MaRDI portal
Publication:6416628
arXiv2211.04653MaRDI QIDQ6416628FDOQ6416628
Publication date: 8 November 2022
Abstract: The gradient flow (GF) is an ODE for which its explicit Euler's discretization is the gradient descent method. In this work, we investigate a family of methods derived from emph{approximate implicit discretizations} of (GF), drawing the connection between larger stability regions and less sensitive hyperparameter tuning. We focus on the implicit -step backwards differentiation formulas (BDFs), approximated in an inner loop with a few iterations of vanilla gradient descent, and give their convergence rate when the objective function is convex, strongly convex, or nonconvex. Numerical experiments show the wide range of effects of these different methods on extremely poorly conditioned problems, especially those brought about in training deep neural networks.
This page was built for publication: Follow the flow: Proximal flow inspired multi-step methods
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6416628)