Extending the Step-Size Restriction for Gradient Descent to Avoid Strict Saddle Points
DOI10.1137/19M129509XzbMATH Open1484.90088arXiv1908.01753OpenAlexW3111569753MaRDI QIDQ5027014FDOQ5027014
Authors: Hayden Schaeffer, S. G. McCalla
Publication date: 3 February 2022
Published in: SIAM Journal on Mathematics of Data Science (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1908.01753
Recommendations
- A generalization of \(s\)-step variants of gradient methods
- On the steplength selection in gradient methods for unconstrained optimization
- New stepsizes for the gradient method
- On the steplength selection in stochastic gradient methods
- Revisiting Normalized Gradient Descent: Fast Evasion of Saddle Points
- A new gradient method with an optimal stepsize property
- Gradient methods with adaptive step-sizes
- An efficient gradient method with approximate optimal stepsize for large-scale unconstrained optimization
- Two novel gradient methods with optimal step sizes
- On the behavior of the gradient norm in the steepest descent method
Nonconvex programming, global optimization (90C26) Normal forms, center manifold theory, bifurcation theory for infinite-dimensional dissipative dynamical systems (37L10)
Cites Work
- Introductory lectures on convex optimization. A basic course.
- Trust Region Methods
- Convergence of the Iterates of Descent Methods for Analytic Cost Functions
- Nonconvergence to unstable points in urn models and stochastic approximations
- Cubic regularization of Newton method and its global performance
- Complete Dictionary Recovery Over the Sphere I: Overview and the Geometric Picture
- Complete Dictionary Recovery Over the Sphere II: Recovery by Riemannian Trust-Region Method
- On the use of directions of negative curvature in a modified newton method
- Complexity analysis of second-order line-search algorithms for smooth nonconvex optimization
- Accelerated methods for nonconvex optimization
- The zero set of a real analytic function
- A geometric analysis of phase retrieval
- Theoretical Insights Into the Optimization Landscape of Over-Parameterized Shallow Neural Networks
- Gradient descent only converges to minimizers: non-isolated critical points and invariant regions
- First-order methods almost always avoid strict saddle points
Cited In (5)
- Analysis of asymptotic escape of strict saddle sets in manifold optimization
- First-order methods almost always avoid strict saddle points
- A deterministic gradient-based approach to avoid saddle points
- Gradient descent only converges to minimizers: non-isolated critical points and invariant regions
- Behavior of accelerated gradient methods near critical points of nonconvex functions
This page was built for publication: Extending the Step-Size Restriction for Gradient Descent to Avoid Strict Saddle Points
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5027014)