Stochastic heavy ball
From MaRDI portal
Abstract: This paper deals with a natural stochastic optimization procedure derived from the so-called Heavy-ball method differential equation, which was introduced by Polyak in the 1960s with his seminal contribution [Pol64]. The Heavy-ball method is a second-order dynamics that was investigated to minimize convex functions f . The family of second-order methods recently received a large amount of attention, until the famous contribution of Nesterov [Nes83], leading to the explosion of large-scale optimization problems. This work provides an in-depth description of the stochastic heavy-ball method, which is an adaptation of the deterministic one when only unbiased evalutions of the gradient are available and used throughout the iterations of the algorithm. We first describe some almost sure convergence results in the case of general non-convex coercive functions f . We then examine the situation of convex and strongly convex potentials and derive some non-asymptotic results about the stochastic heavy-ball method. We end our study with limit theorems on several rescaled algorithms.
Recommendations
- Stochastic heavy-ball method for constrained stochastic optimization problems
- Convergence rates of the heavy ball method for quasi-strongly convex optimization
- scientific article; zbMATH DE number 165426
- Non-monotone Behavior of the Heavy Ball Method
- Convergence rates for the heavy-ball continuous dynamics for non-convex optimization, under Polyak-Łojasiewicz condition
Cites work
- scientific article; zbMATH DE number 6475388 (Why is no real title available?)
- scientific article; zbMATH DE number 3850830 (Why is no real title available?)
- scientific article; zbMATH DE number 3826915 (Why is no real title available?)
- scientific article; zbMATH DE number 3951715 (Why is no real title available?)
- scientific article; zbMATH DE number 3747561 (Why is no real title available?)
- scientific article; zbMATH DE number 3790208 (Why is no real title available?)
- scientific article; zbMATH DE number 48893 (Why is no real title available?)
- scientific article; zbMATH DE number 578461 (Why is no real title available?)
- scientific article; zbMATH DE number 976356 (Why is no real title available?)
- scientific article; zbMATH DE number 1972910 (Why is no real title available?)
- scientific article; zbMATH DE number 2107836 (Why is no real title available?)
- scientific article; zbMATH DE number 1405930 (Why is no real title available?)
- A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems
- A Stochastic Approximation Method
- A differential equation for modeling Nesterov's accelerated gradient method: theory and insights
- A stochastic algorithm for feature selection in pattern recognition
- Accelerated gradient methods for nonconvex nonlinear and stochastic programming
- Acceleration of Stochastic Approximation by Averaging
- Adaptivity of averaged stochastic gradient descent to local strong convexity for logistic regression
- An adaptive scheme for the approximation of dissipative systems
- An optimal method for stochastic composite optimization
- Asymptotic pseudotrajectories and chain recurrent flows, with applications
- Concentration inequalities. A nonasymptotic theory of independence
- Do stochastic algorithms avoid traps?
- Ergodicity for SDEs and approximations: locally Lipschitz vector fields and degenerate noise.
- Hypocoercivity
- Introductory lectures on convex optimization. A basic course.
- Large-scale machine learning with stochastic gradient descent
- Local minima and convergence in low-rank semidefinite programming
- Long time behaviour and stationary regime of memory gradient diffusions
- Multidimensional diffusion processes.
- Nonconvergence to unstable points in urn models and stochastic approximations
- On the Long Time Behavior of Second Order Differential Equations with Asymptotically Small Dissipation
- On the convergence of gradient-like flows with noisy gradient input
- On the long time behavior of second order differential equations with asymptotically small dissipation
- Optimal non-asymptotic analysis of the Ruppert-Polyak averaging stochastic algorithm
- Parallelizing stochastic gradient descent for least squares regression: mini-batching, averaging, and model misspecification
- Self-interacting diffusions.
- Some methods of speeding up the convergence of iteration methods
- Stability of Markovian processes III: Foster–Lyapunov criteria for continuous-time processes
- Stochastic Estimation of the Maximum of a Regression Function
- Stochastic First- and Zeroth-Order Methods for Nonconvex Stochastic Programming
- Stochastic approximation with two time scales
- Stochastic approximation. A dynamical systems viewpoint.
Cited in
(25)- scientific article; zbMATH DE number 1421106 (Why is no real title available?)
- Non asymptotic controls on a recursive superquantile approximation
- Stochastic optimization with momentum: convergence, fluctuations, and traps avoidance
- Stochastic approximation algorithms for superquantiles estimation
- Stochastic heavy-ball method for constrained stochastic optimization problems
- On the rates of convergence of parallelized averaged stochastic gradient algorithms
- SRKCD: a stabilized Runge-Kutta method for stochastic optimization
- scientific article; zbMATH DE number 7370590 (Why is no real title available?)
- Stochastic differential equations for modeling first order optimization methods
- scientific article; zbMATH DE number 7370534 (Why is no real title available?)
- Nonsmooth nonconvex stochastic heavy ball
- Stochastic relaxed inertial forward-backward-forward splitting for monotone inclusions in Hilbert spaces
- Momentum and stochastic momentum for stochastic gradient, Newton, proximal point and subspace descent methods
- Convergence and dynamical behavior of the ADAM algorithm for nonconvex stochastic optimization
- Convergence rates of the heavy ball method for quasi-strongly convex optimization
- Differentially Private Accelerated Optimization Algorithms
- A general system of differential equations to model first-order adaptive algorithms
- Subgradient Sampling for Nonsmooth Nonconvex Minimization
- A robust control approach to asymptotic optimality of the heavy ball method for optimization of quadratic functions
- scientific article; zbMATH DE number 165426 (Why is no real title available?)
- An adaptive Polyak heavy-ball method
- Convergence of gradient algorithms for nonconvex \(C^{1+ \alpha}\) cost functions
- On the convergence analysis of aggregated heavy-ball method
- Optimal non-asymptotic analysis of the Ruppert-Polyak averaging stochastic algorithm
- Several kinds of acceleration techniques for unconstrained optimization first-order algorithms
This page was built for publication: Stochastic heavy ball
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q1697485)