Finite-sum smooth optimization with SARAH

From MaRDI portal




Abstract: The total complexity (measured as the total number of gradient computations) of a stochastic first-order optimization algorithm that finds a first-order stationary point of a finite-sum smooth nonconvex objective function F(w)=frac1nsumi=1nfi(w) has been proven to be at least Omega(sqrtn/epsilon) for nleqmathcalO(epsilon2) where epsilon denotes the attained accuracy mathbbE[|ablaF(ildew)|2]leqepsilon for the outputted approximation ildew (Fang et al., 2018). In this paper, we provide a convergence analysis for a slightly modified version of the SARAH algorithm (Nguyen et al., 2017a;b) and achieve total complexity that matches the lower-bound worst case complexity in (Fang et al., 2018) up to a constant factor when nleqmathcalO(epsilon2) for nonconvex problems. For convex optimization, we propose SARAH++ with sublinear convergence for general convex and linear convergence for strongly convex problems; and we provide a practical version for which numerical experiments on various datasets show an improved performance.





Describes a project that uses

Uses Software





This page was built for publication: Finite-sum smooth optimization with SARAH

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2149950)