Batched Stochastic Gradient Descent with Weighted Sampling

From MaRDI portal
Publication:4609808

DOI10.1007/978-3-319-59912-0_14zbMATH Open1385.65041arXiv1608.07641OpenAlexW2963244042MaRDI QIDQ4609808FDOQ4609808


Authors: D. Needell, Rachel Ward Edit this on Wikidata


Publication date: 26 March 2018

Published in: Springer Proceedings in Mathematics & Statistics (Search for Journal in Brave)

Abstract: We analyze a batched variant of Stochastic Gradient Descent (SGD) with weighted sampling distribution for smooth and non-smooth objective functions. We show that by distributing the batches computationally, a significant speedup in the convergence rate is provably possible compared to either batched sampling or weighted sampling alone. We propose several computationally efficient schemes to approximate the optimal weights, and compute proposed sampling distributions explicitly for the least squares and hinge loss problems. We show both analytically and experimentally that substantial gains can be obtained.


Full work available at URL: https://arxiv.org/abs/1608.07641




Recommendations




Cites Work


Cited In (11)

Uses Software





This page was built for publication: Batched Stochastic Gradient Descent with Weighted Sampling

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q4609808)