ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1912.04177
46
20
v1v2v3v4v5 (latest)

Robust and Sample Optimal Algorithms for PSD Low-Rank Approximation

9 December 2019
Ainesh Bakshi
Nadiia Chepurko
David P. Woodruff
ArXiv (abs)PDFHTML
Abstract

Recently, Musco and Woodruff (FOCS, 2017) showed that given an n×nn \times nn×n positive semidefinite (PSD) matrix AAA, it is possible to compute a (1+ϵ)(1+\epsilon)(1+ϵ)-approximate relative-error low-rank approximation to AAA by querying O(nk/ϵ2.5)O(nk/\epsilon^{2.5})O(nk/ϵ2.5) entries of AAA in time O(nk/ϵ2.5+nkω−1/ϵ2(ω−1))O(nk/\epsilon^{2.5} +n k^{\omega-1}/\epsilon^{2(\omega-1)})O(nk/ϵ2.5+nkω−1/ϵ2(ω−1)). They also showed that any relative-error low-rank approximation algorithm must query Ω(nk/ϵ)\Omega(nk/\epsilon)Ω(nk/ϵ) entries of AAA, this gap has since remained open. Our main result is to resolve this question by obtaining an optimal algorithm that queries O(nk/ϵ)O(nk/\epsilon)O(nk/ϵ) entries of AAA and outputs a relative-error low-rank approximation in O(n(k/ϵ)ω−1)O(n(k/\epsilon)^{\omega-1})O(n(k/ϵ)ω−1) time. Note, our running time improves that of Musco and Woodruff, and matches the information-theoretic lower bound if the matrix-multiplication exponent ω\omegaω is 222. We then extend our techniques to negative-type distance matrices. Bakshi and Woodruff (NeurIPS, 2018) showed a bi-criteria, relative-error low-rank approximation which queries O(nk/ϵ2.5)O(nk/\epsilon^{2.5})O(nk/ϵ2.5) entries and outputs a rank-(k+4)(k+4)(k+4) matrix. We show that the bi-criteria guarantee is not necessary and obtain an O(nk/ϵ)O(nk/\epsilon)O(nk/ϵ) query algorithm, which is optimal. Our algorithm applies to all distance matrices that arise from metrics satisfying negative-type inequalities, including ℓ1,ℓ2,\ell_1, \ell_2,ℓ1​,ℓ2​, spherical metrics and hypermetrics. Next, we introduce a new robust low-rank approximation model which captures PSD matrices that have been corrupted with noise. While a sample complexity lower bound precludes sublinear algorithms for arbitrary PSD matrices, we provide the first sublinear time and query algorithms when the corruption on the diagonal entries is bounded. As a special case, we show sample-optimal sublinear time algorithms for low-rank approximation of correlation matrices corrupted by noise.

View on arXiv
Comments on this paper