ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2006.05441
26
6

Faster PAC Learning and Smaller Coresets via Smoothed Analysis

9 June 2020
Alaa Maalouf
Ibrahim Jubran
M. Tukan
Dan Feldman
ArXivPDFHTML
Abstract

PAC-learning usually aims to compute a small subset (ε\varepsilonε-sample/net) from nnn items, that provably approximates a given loss function for every query (model, classifier, hypothesis) from a given set of queries, up to an additive error ε∈(0,1)\varepsilon\in(0,1)ε∈(0,1). Coresets generalize this idea to support multiplicative error 1±ε1\pm\varepsilon1±ε. Inspired by smoothed analysis, we suggest a natural generalization: approximate the \emph{average} (instead of the worst-case) error over the queries, in the hope of getting smaller subsets. The dependency between errors of different queries implies that we may no longer apply the Chernoff-Hoeffding inequality for a fixed query, and then use the VC-dimension or union bound. This paper provides deterministic and randomized algorithms for computing such coresets and ε\varepsilonε-samples of size independent of nnn, for any finite set of queries and loss function. Example applications include new and improved coreset constructions for e.g. streaming vector summarization [ICML'17] and kkk-PCA [NIPS'16]. Experimental results with open source code are provided.

View on arXiv
Comments on this paper