Communication-Efficient Distributed Learning via Lazily Aggregated Quantized Gradients

From MaRDI portal
Publication:6325444

arXiv1909.07588MaRDI QIDQ6325444FDOQ6325444


Authors: Jun Sun, Tianyi Chen, Georgios B. Giannakis, Zaiyue Yang Edit this on Wikidata


Publication date: 17 September 2019

Abstract: The present paper develops a novel aggregated gradient approach for distributed machine learning that adaptively compresses the gradient communication. The key idea is to first quantize the computed gradients, and then skip less informative quantized gradient communications by reusing outdated gradients. Quantizing and skipping result in `lazy' worker-server communications, which justifies the term Lazily Aggregated Quantized gradient that is henceforth abbreviated as LAQ. Our LAQ can provably attain the same linear convergence rate as the gradient descent in the strongly convex case, while effecting major savings in the communication overhead both in transmitted bits as well as in communication rounds. Empirically, experiments with real data corroborate a significant communication reduction compared to existing gradient- and stochastic gradient-based algorithms.




Has companion code repository: https://github.com/sunjunaimer/LAQ









This page was built for publication: Communication-Efficient Distributed Learning via Lazily Aggregated Quantized Gradients

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6325444)