ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1312.1054
122
91

Faster and Sample Near-Optimal Algorithms for Proper Learning Mixtures of Gaussians

4 December 2013
C. Daskalakis
Gautam Kamath
ArXivPDFHTML
Abstract

We provide an algorithm for properly learning mixtures of two single-dimensional Gaussians without any separability assumptions. Given O~(1/ε2)\tilde{O}(1/\varepsilon^2)O~(1/ε2) samples from an unknown mixture, our algorithm outputs a mixture that is ε\varepsilonε-close in total variation distance, in time O~(1/ε5)\tilde{O}(1/\varepsilon^5)O~(1/ε5). Our sample complexity is optimal up to logarithmic factors, and significantly improves upon both Kalai et al., whose algorithm has a prohibitive dependence on 1/ε1/\varepsilon1/ε, and Feldman et al., whose algorithm requires bounds on the mixture parameters and depends pseudo-polynomially in these parameters. One of our main contributions is an improved and generalized algorithm for selecting a good candidate distribution from among competing hypotheses. Namely, given a collection of NNN hypotheses containing at least one candidate that is ε\varepsilonε-close to an unknown distribution, our algorithm outputs a candidate which is O(ε)O(\varepsilon)O(ε)-close to the distribution. The algorithm requires O(log⁡N/ε2){O}(\log{N}/\varepsilon^2)O(logN/ε2) samples from the unknown distribution and O(Nlog⁡N/ε2){O}(N \log N/\varepsilon^2)O(NlogN/ε2) time, which improves previous such results (such as the Scheff\é estimator) from a quadratic dependence of the running time on NNN to quasilinear. Given the wide use of such results for the purpose of hypothesis selection, our improved algorithm implies immediate improvements to any such use.

View on arXiv
Comments on this paper