Why random reshuffling beats stochastic gradient descent
From MaRDI portal
Publication:2227529
DOI10.1007/s10107-019-01440-wzbMath1459.90199arXiv1510.08560OpenAlexW1866529865WikidataQ126862166 ScholiaQ126862166MaRDI QIDQ2227529
Pablo A. Parrilo, Mert Gürbüzbalaban, Asuman Ozdaglar
Publication date: 15 February 2021
Published in: Mathematical Programming. Series A. Series B (Search for Journal in Brave)
Full work available at URL: https://arxiv.org/abs/1510.08560
Convex programming (90C25) Large-scale problems in mathematical programming (90C06) Nonlinear programming (90C30) Stochastic programming (90C15)
Related Items
On the Efficiency of Random Permutation for ADMM and Coordinate Descent, Spherical graph drawing by multi-dimensional scaling, Simple and fast algorithm for binary integer and online linear programming, Convergence of Random Reshuffling under the Kurdyka–Łojasiewicz Inequality, Global Convergence Rate of Proximal Incremental Aggregated Gradient Methods, Unnamed Item, Two Symmetrized Coordinate Descent Methods Can Be $O(n^2)$ Times Slower Than the Randomized Version, Incremental without replacement sampling in nonconvex optimization, Convergence Rate of Incremental Gradient and Incremental Newton Methods, On the Convergence Rate of Incremental Aggregated Gradient Algorithms
Uses Software
Cites Work
- Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers
- An arithmetic-geometric mean inequality for products of three matrices
- Parallel stochastic gradient algorithms for large-scale matrix completion
- Large-Scale Machine Learning with Stochastic Gradient Descent
- Robust Stochastic Approximation Approach to Stochastic Programming
- Acceleration of Stochastic Approximation by Averaging
- A New Class of Incremental Gradient Methods for Least Squares Problems
- Incremental Least Squares Methods and the Extended Kalman Filter
- Distributed Subgradient Methods for Multi-Agent Optimization
- Convergence Rate of Incremental Gradient and Incremental Newton Methods
- Information-Theoretic Lower Bounds on the Oracle Complexity of Stochastic Convex Optimization
- On‐line learning for very large data sets
- Convergence of weighted averages of random variables revisited
- Stochastic Approximation of Minima with Improved Asymptotic Speed
- On Asymptotic Normality in Stochastic Approximation
- A Stochastic Approximation Method
- On a Stochastic Approximation Method
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item
- Unnamed Item