Low-Rank Approximation and Regression in Input Sparsity Time

From MaRDI portal
Publication:3177880

DOI10.1145/3019134zbMATH Open1426.65057arXiv1207.6365OpenAlexW2580753685MaRDI QIDQ3177880FDOQ3177880

Kenneth L. Clarkson, David P. Woodruff

Publication date: 2 August 2018

Published in: Journal of the ACM, Proceedings of the forty-eighth annual ACM symposium on Theory of Computing (Search for Journal in Brave)

Abstract: We design a new distribution over poly(reps1)imesn matrices S so that for any fixed nimesd matrix A of rank r, with probability at least 9/10, ormSAx2=(1pmeps)ormAx2 simultaneously for all xinmathbbRd. Such a matrix S is called a emph{subspace embedding}. Furthermore, SA can be computed in nz(A)+poly(deps1) time, where nz(A) is the number of non-zero entries of A. This improves over all previous subspace embeddings, which required at least Omega(ndlogd) time to achieve this property. We call our matrices S emph{sparse embedding matrices}. Using our sparse embedding matrices, we obtain the fastest known algorithms for (1+eps)-approximation for overconstrained least-squares regression, low-rank approximation, approximating all leverage scores, and ellp-regression. The leading order term in the time complexity of our algorithms is O(nz(A)) or O(nz(A)logn). We optimize the low-order poly(d/eps) terms in our running times (or for rank-k approximation, the n*poly(k/eps) term), and show various tradeoffs. For instance, we also use our methods to design new preconditioners that improve the dependence on eps in least squares regression to log1/eps. Finally, we provide preliminary experimental results which suggest that our algorithms are competitive in practice.


Full work available at URL: https://arxiv.org/abs/1207.6365




Recommendations





Cited In (99)





This page was built for publication: Low-Rank Approximation and Regression in Input Sparsity Time

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q3177880)