Stochastic subgradient method converges on tame functions

From MaRDI portal
Publication:2291732

DOI10.1007/s10208-018-09409-5zbMath1433.65141arXiv1804.07795OpenAlexW2964210434WikidataQ128619589 ScholiaQ128619589MaRDI QIDQ2291732

Damek Davis, Dmitriy Drusvyatskiy, Jason D. Lee, Sham M. Kakade

Publication date: 31 January 2020

Published in: Foundations of Computational Mathematics (Search for Journal in Brave)

Full work available at URL: https://arxiv.org/abs/1804.07795




Related Items (41)

Convergence of constant step stochastic gradient descent for non-smooth non-convex functionsA framework for differential calculus on persistence barcodesGraphical Convergence of Subgradients in Nonconvex Optimization and LearningA forward-backward dynamical approach for nonsmooth problems with block structure coupled by a smooth functionConservative and semismooth derivatives are equivalent for semialgebraic mapsA gradient sampling algorithm for stratified maps with applications to topological data analysisGlobal convergence of the gradient method for functions definable in o-minimal structuresLyapunov stability of the subgradient method with constant step sizeStochastic proximal subgradient descent oscillates in the vicinity of its accumulation setSubgradient Sampling for Nonsmooth Nonconvex MinimizationAn Improved Unconstrained Approach for Bilevel OptimizationMomentum-based variance-reduced proximal stochastic gradient method for composite nonconvex stochastic optimizationStochastic approximation with discontinuous dynamics, differential inclusions, and applicationsCertifying the Absence of Spurious Local Minima at InfinityPathological Subgradient DynamicsConservative parametric optimality and the ridge method for tame min-max problemsFirst-order methods for convex optimizationA Decomposition Algorithm for Two-Stage Stochastic Programs with Nonconvex Recourse FunctionsMultiComposite Nonconvex Optimization for Training Deep Neural NetworksGeneralized gradients in dynamic optimization, optimal control, and machine learning problemsDiscussion of: ``Nonparametric regression using deep neural networks with ReLU activation functionConvergence of a stochastic subgradient method with averaging for nonsmooth nonconvex constrained optimizationConvergence and Dynamical Behavior of the ADAM Algorithm for Nonconvex Stochastic OptimizationStochastic Approximation for Optimization in Shape SpacesDetermination of convex functions via subgradients of minimal normManifold Sampling for Optimizing Nonsmooth Nonconvex CompositionsStochastic proximal gradient methods for nonconvex problems in Hilbert spacesConservative set valued fields, automatic differentiation, stochastic gradient methods and deep learningIncremental without replacement sampling in nonconvex optimizationEvery Local Minimum Value Is the Global Minimum Value of Induced Model in Nonconvex Machine LearningStochastic generalized gradient methods for training nonconvex nonsmooth neural networksA stochastic approximation method for approximating the efficient frontier of chance-constrained nonlinear programsAn Inertial Newton Algorithm for Deep LearningA Stochastic Subgradient Method for Nonsmooth Nonconvex Multilevel Composition OptimizationUnnamed ItemGhost Penalties in Nonconvex Constrained Optimization: Diminishing Stepsizes and Iteration ComplexityPerturbed iterate SGD for Lipschitz continuous loss functionsThe Structure of Conservative Gradient FieldsAsymptotic Properties of Stationary Solutions of Coupled Nonconvex Nonsmooth Empirical Risk MinimizationExamples of Pathological Dynamics of the Subgradient Method for Lipschitz Path-Differentiable FunctionsFour algorithms to solve symmetric multi-type non-negative matrix tri-factorization problem


Uses Software


Cites Work


This page was built for publication: Stochastic subgradient method converges on tame functions