LIVEJoin the current RTAI Connect sessionJoin now

37
6

Sketching Algorithms and Lower Bounds for Ridge Regression

Abstract

We give a sketching-based iterative algorithm that computes a 1+ε1+\varepsilon approximate solution for the ridge regression problem minxAxb22+λx22\min_x \|Ax-b\|_2^2 +\lambda\|x\|_2^2 where ARn×dA \in R^{n \times d} with dnd \ge n. Our algorithm, for a constant number of iterations (requiring a constant number of passes over the input), improves upon earlier work (Chowdhury et al.) by requiring that the sketching matrix only has a weaker Approximate Matrix Multiplication (AMM) guarantee that depends on ε\varepsilon, along with a constant subspace embedding guarantee. The earlier work instead requires that the sketching matrix has a subspace embedding guarantee that depends on ε\varepsilon. For example, to produce a 1+ε1+\varepsilon approximate solution in 11 iteration, which requires 22 passes over the input, our algorithm requires the OSNAP embedding to have m=O(nσ2/λε)m= O(n\sigma^2/\lambda\varepsilon) rows with a sparsity parameter s=O(log(n))s = O(\log(n)), whereas the earlier algorithm of Chowdhury et al. with the same number of rows of OSNAP requires a sparsity s=O(σ2/λεlog(n))s = O(\sqrt{\sigma^2/\lambda\varepsilon} \cdot \log(n)), where σ=\opnormA\sigma = \opnorm{A} is the spectral norm of the matrix AA. We also show that this algorithm can be used to give faster algorithms for kernel ridge regression. Finally, we show that the sketch size required for our algorithm is essentially optimal for a natural framework of algorithms for ridge regression by proving lower bounds on oblivious sketching matrices for AMM. The sketch size lower bounds for AMM may be of independent interest.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.