The Fast Cauchy Transform: with Applications to Basis Construction, Regression, and Subspace Approximation in L1
- OOD

We give fast algorithms for regression and related problems: for an input matrix and vector , in time we reduce the problem to the same problem with input matrix of dimension and corresponding of dimension ; and are a \emph{coreset} for the problem, consisting of sampled and rescaled rows of and . Here is independent of , and polynomial in . Our results improve on the best previous algorithms when , for all except , in particular the running time of Sohler and Woodruff (STOC, 2011) for , that uses asymptotically fast matrix multiplication, and the time of Dasgupta \emph{et al.} (SODA, 2008) for general . We also give a detailed empirical evaluation of implementations of our algorithms for , comparing them with several related algorithms. Among other things, our results clearly show that the practice follows the theory closely, in the asymptotic regime. In addition, we show near-optimal results for regression problems that are too large for any prior solution methods. Our algorithms use our faster constructions of well-conditioned bases for spaces, and for , a fast subspace embedding: a matrix , found obliviously to , that approximately preserves the norms of all vectors in ; that is, , for all , with distortion . Moreover, can be computed in time. Our techniques include fast Johnson-Lindenstrauss transforms, low coherence matrices, and rescaling by Cauchy random variables.
View on arXiv