ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1811.01442
21
32

Towards a Zero-One Law for Column Subset Selection

4 November 2018
Zhao Song
David P. Woodruff
Peilin Zhong
ArXivPDFHTML
Abstract

There are a number of approximation algorithms for NP-hard versions of low rank approximation, such as finding a rank-kkk matrix BBB minimizing the sum of absolute values of differences to a given nnn-by-nnn matrix AAA, min⁡rank-k B∥A−B∥1\min_{\textrm{rank-}k~B}\|A-B\|_1minrank-k B​∥A−B∥1​, or more generally finding a rank-kkk matrix BBB which minimizes the sum of ppp-th powers of absolute values of differences, min⁡rank-k B∥A−B∥pp\min_{\textrm{rank-}k~B}\|A-B\|_p^pminrank-k B​∥A−B∥pp​. Many of these algorithms are linear time columns subset selection algorithms, returning a subset of poly(klog⁡n)\mathrm{poly}(k \log n)poly(klogn) columns whose cost is no more than a poly(k)\mathrm{poly}(k)poly(k) factor larger than the cost of the best rank-kkk matrix. The above error measures are special cases of the following general entrywise low rank approximation problem: given an arbitrary function g:R→R≥0g:\mathbb{R} \rightarrow \mathbb{R}_{\geq 0}g:R→R≥0​, find a rank-kkk matrix BBB which minimizes ∥A−B∥g=∑i,jg(Ai,j−Bi,j)\|A-B\|_g = \sum_{i,j}g(A_{i,j}-B_{i,j})∥A−B∥g​=∑i,j​g(Ai,j​−Bi,j​). A natural question is which functions ggg admit efficient approximation algorithms? Indeed, this is a central question of recent work studying generalized low rank models. In this work we give approximation algorithms for every\textit{every}every function ggg which is approximately monotone and satisfies an approximate triangle inequality, and we show both of these conditions are necessary. Further, our algorithm is efficient if the function ggg admits an efficient approximate regression algorithm. Our approximation algorithms handle functions which are not even scale-invariant, such as the Huber loss function, which we show have very different structural properties than ℓp\ell_pℓp​-norms, e.g., one can show the lack of scale-invariance causes any column subset selection algorithm to provably require a log⁡n\sqrt{\log n}logn​ factor larger number of columns than ℓp\ell_pℓp​-norms; nevertheless we design the first efficient column subset selection algorithms for such error measures.

View on arXiv
Comments on this paper