ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1706.01596
13
25

Sample-Efficient Learning of Mixtures

6 June 2017
H. Ashtiani
Shai Ben-David
Abbas Mehrabian
ArXivPDFHTML
Abstract

We consider PAC learning of probability distributions (a.k.a. density estimation), where we are given an i.i.d. sample generated from an unknown target distribution, and want to output a distribution that is close to the target in total variation distance. Let F\mathcal FF be an arbitrary class of probability distributions, and let Fk\mathcal{F}^kFk denote the class of kkk-mixtures of elements of F\mathcal FF. Assuming the existence of a method for learning F\mathcal FF with sample complexity mF(ϵ)m_{\mathcal{F}}(\epsilon)mF​(ϵ), we provide a method for learning Fk\mathcal F^kFk with sample complexity O(klog⁡k⋅mF(ϵ)/ϵ2)O({k\log k \cdot m_{\mathcal F}(\epsilon) }/{\epsilon^{2}})O(klogk⋅mF​(ϵ)/ϵ2). Our mixture learning algorithm has the property that, if the F\mathcal FF-learner is proper/agnostic, then the Fk\mathcal F^kFk-learner would be proper/agnostic as well. This general result enables us to improve the best known sample complexity upper bounds for a variety of important mixture classes. First, we show that the class of mixtures of kkk axis-aligned Gaussians in Rd\mathbb{R}^dRd is PAC-learnable in the agnostic setting with O~(kd/ϵ4)\widetilde{O}({kd}/{\epsilon ^ 4})O(kd/ϵ4) samples, which is tight in kkk and ddd up to logarithmic factors. Second, we show that the class of mixtures of kkk Gaussians in Rd\mathbb{R}^dRd is PAC-learnable in the agnostic setting with sample complexity O~(kd2/ϵ4)\widetilde{O}({kd^2}/{\epsilon ^ 4})O(kd2/ϵ4), which improves the previous known bounds of O~(k3d2/ϵ4)\widetilde{O}({k^3d^2}/{\epsilon ^ 4})O(k3d2/ϵ4) and O~(k4d4/ϵ2)\widetilde{O}(k^4d^4/\epsilon ^ 2)O(k4d4/ϵ2) in its dependence on kkk and ddd. Finally, we show that the class of mixtures of kkk log-concave distributions over Rd\mathbb{R}^dRd is PAC-learnable using O~(d(d+5)/2ϵ−(d+9)/2k)\widetilde{O}(d^{(d+5)/2}\epsilon^{-(d+9)/2}k)O(d(d+5)/2ϵ−(d+9)/2k) samples.

View on arXiv
Comments on this paper