ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2306.16516
12
0

For Kernel Range Spaces a Constant Number of Queries Are Sufficient

28 June 2023
J. Phillips
H. Pourmahmood-Aghababa
ArXivPDFHTML
Abstract

We introduce the notion of an ε\varepsilonε-cover for a kernel range space. A kernel range space concerns a set of points X⊂RdX \subset \mathbb{R}^dX⊂Rd and the space of all queries by a fixed kernel (e.g., a Gaussian kernel K(p,⋅)=exp⁡(−∥p−⋅∥2)K(p,\cdot) = \exp(-\|p-\cdot\|^2)K(p,⋅)=exp(−∥p−⋅∥2)). For a point set XXX of size nnn, a query returns a vector of values Rp∈RnR_p \in \mathbb{R}^nRp​∈Rn, where the iiith coordinate (Rp)i=K(p,xi)(R_p)_i = K(p,x_i)(Rp​)i​=K(p,xi​) for xi∈Xx_i \in Xxi​∈X. An ε\varepsilonε-cover is a subset of points Q⊂RdQ \subset \mathbb{R}^dQ⊂Rd so for any p∈Rdp \in \mathbb{R}^dp∈Rd that 1n∥Rp−Rq∥1≤ε\frac{1}{n} \|R_p - R_q\|_1\leq \varepsilonn1​∥Rp​−Rq​∥1​≤ε for some q∈Qq \in Qq∈Q. This is a smooth analog of Haussler's notion of ε\varepsilonε-covers for combinatorial range spaces (e.g., defined by subsets of points within a ball query) where the resulting vectors RpR_pRp​ are in {0,1}n\{0,1\}^n{0,1}n instead of [0,1]n[0,1]^n[0,1]n. The kernel versions of these range spaces show up in data analysis tasks where the coordinates may be uncertain or imprecise, and hence one wishes to add some flexibility in the notion of inside and outside of a query range. Our main result is that, unlike combinatorial range spaces, the size of kernel ε\varepsilonε-covers is independent of the input size nnn and dimension ddd. We obtain a bound of (1/ε)O~(1/ε2)(1/\varepsilon)^{\tilde O(1/\varepsilon^2)}(1/ε)O~(1/ε2), where O~(f(1/ε))\tilde{O}(f(1/\varepsilon))O~(f(1/ε)) hides log factors in (1/ε)(1/\varepsilon)(1/ε) that can depend on the kernel. This implies that by relaxing the notion of boundaries in range queries, eventually the curse of dimensionality disappears, and may help explain the success of machine learning in very high-dimensions. We also complement this result with a lower bound of almost (1/ε)Ω(1/ε)(1/\varepsilon)^{\Omega(1/\varepsilon)}(1/ε)Ω(1/ε), showing the exponential dependence on 1/ε1/\varepsilon1/ε is necessary.

View on arXiv
Comments on this paper