Critical point-finding methods reveal gradient-flat regions of deep network losses
From MaRDI portal
Publication:5004367
Recommendations
- Shaping the learning landscape in neural networks around wide flat minima
- Landscape analysis for shallow neural networks: complete classification of critical points for affine target functions
- Optimization Landscape of Neural Networks
- The global optimization geometry of shallow linear neural networks
- Symmetry \& critical points for a model shallow neural network
Cites work
- scientific article; zbMATH DE number 3436540 (Why is no real title available?)
- scientific article; zbMATH DE number 2095946 (Why is no real title available?)
- scientific article; zbMATH DE number 2107836 (Why is no real title available?)
- scientific article; zbMATH DE number 5060482 (Why is no real title available?)
- Adaptive subgradient methods for online learning and stochastic optimization
- Analysis of Newton’s Method at Irregular Singularities
- Implementable tensor methods in unconstrained convex optimization
- MINRES-QLP: a Krylov subspace method for indefinite or singular symmetric systems
- Newton-Type Methods for Optimization and Variational Problems
- Numerically solving polynomial systems with Bertini
- On the convergence of Newton iterations to non-stationary points
- Residual and backward error bounds in minimum residual Krylov subspace methods
- Scikit-learn: machine learning in Python
- The Fundamental Theorem of Linear Algebra
Cited in
(3)
This page was built for publication: Critical point-finding methods reveal gradient-flat regions of deep network losses
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q5004367)