Lower bounds for non-convex stochastic optimization

From MaRDI portal
Publication:6038643

DOI10.1007/S10107-022-01822-7zbMATH Open1517.90087arXiv1912.02365OpenAlexW2993258424MaRDI QIDQ6038643FDOQ6038643

Nathan Srebro, John C. Duchi, Yossi Arjevani, Blake Woodworth, Yair Carmon, Dylan J. Foster

Publication date: 2 May 2023

Published in: Mathematical Programming. Series A. Series B (Search for Journal in Brave)

Abstract: We lower bound the complexity of finding epsilon-stationary points (with gradient norm at most epsilon) using stochastic first-order methods. In a well-studied model where algorithms access smooth, potentially non-convex functions through queries to an unbiased stochastic gradient oracle with bounded variance, we prove that (in the worst case) any algorithm requires at least epsilon4 queries to find an epsilon stationary point. The lower bound is tight, and establishes that stochastic gradient descent is minimax optimal in this model. In a more restrictive model where the noisy gradient estimates satisfy a mean-squared smoothness property, we prove a lower bound of epsilon3 queries, establishing the optimality of recently proposed variance reduction techniques.


Full work available at URL: https://arxiv.org/abs/1912.02365




Recommendations



Cites Work


Cited In (8)





This page was built for publication: Lower bounds for non-convex stochastic optimization

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6038643)