Lower bounds for non-convex stochastic optimization

From MaRDI portal
Publication:6038643




Abstract: We lower bound the complexity of finding epsilon-stationary points (with gradient norm at most epsilon) using stochastic first-order methods. In a well-studied model where algorithms access smooth, potentially non-convex functions through queries to an unbiased stochastic gradient oracle with bounded variance, we prove that (in the worst case) any algorithm requires at least epsilon4 queries to find an epsilon stationary point. The lower bound is tight, and establishes that stochastic gradient descent is minimax optimal in this model. In a more restrictive model where the noisy gradient estimates satisfy a mean-squared smoothness property, we prove a lower bound of epsilon3 queries, establishing the optimality of recently proposed variance reduction techniques.



Cites work







This page was built for publication: Lower bounds for non-convex stochastic optimization

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6038643)