Finite-sum smooth optimization with SARAH
From MaRDI portal
Abstract: The total complexity (measured as the total number of gradient computations) of a stochastic first-order optimization algorithm that finds a first-order stationary point of a finite-sum smooth nonconvex objective function has been proven to be at least for where denotes the attained accuracy for the outputted approximation (Fang et al., 2018). In this paper, we provide a convergence analysis for a slightly modified version of the SARAH algorithm (Nguyen et al., 2017a;b) and achieve total complexity that matches the lower-bound worst case complexity in (Fang et al., 2018) up to a constant factor when for nonconvex problems. For convex optimization, we propose SARAH++ with sublinear convergence for general convex and linear convergence for strongly convex problems; and we provide a practical version for which numerical experiments on various datasets show an improved performance.
Recommendations
- Inexact SARAH algorithm for stochastic optimization
- Smooth Optimization with Approximate Gradient
- Smooth Optimization Methods for Minimax Problems
- A smooth method for the finite minimax problem
- Solving semi-infinite programs by smoothing projected gradient method
- A smoothing iterative method for the finite minimax problem
- Smoothing accelerated algorithm for constrained nonsmooth convex optimization problems
- A Stochastic Smoothing Algorithm for Semidefinite Programming
- Smooth over-parameterized solvers for non-smooth structured optimization
- Smoothing augmented Lagrangian method for nonsmooth constrained optimization problems
Cites work
- scientific article; zbMATH DE number 7255141 (Why is no real title available?)
- A Stochastic Approximation Method
- Adaptive subgradient methods for online learning and stochastic optimization
- Inexact SARAH algorithm for stochastic optimization
- Introductory lectures on convex optimization. A basic course.
- Minimizing finite sums with the stochastic average gradient
- New Convergence Aspects of Stochastic Gradient Algorithms
- Optimization methods for large-scale machine learning
- Stochastic dual coordinate ascent methods for regularized loss minimization
Cited in
(9)- Adaptivity of stochastic gradient methods for nonconvex optimization
- DESTRESS: Computation-Optimal and Communication-Efficient Decentralized Nonconvex Finite-Sum Optimization
- A Convergence Study of SGD-Type Methods for Stochastic Optimization
- Accelerating mini-batch SARAH by step size rules
- Stochastic variance-reduced majorization-minimization algorithms
- Fast decentralized nonconvex finite-sum optimization with recursive variance reduction
- A stochastic extra-step quasi-Newton method for nonsmooth nonconvex optimization
- Inexact SARAH algorithm for stochastic optimization
- Stochastic gradient methods with preconditioned updates
This page was built for publication: Finite-sum smooth optimization with SARAH
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2149950)