ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.15150
17
3

Tight Bounds on the Hardness of Learning Simple Nonparametric Mixtures

28 March 2022
Bryon Aragam
W. Tai
ArXivPDFHTML
Abstract

We study the problem of learning nonparametric distributions in a finite mixture, and establish tight bounds on the sample complexity for learning the component distributions in such models. Namely, we are given i.i.d. samples from a pdf fff where f=w_1f_1+w_2f_2, \quad w_1+w_2=1, \quad w_1,w_2>0 and we are interested in learning each component fif_ifi​. Without any assumptions on fif_ifi​, this problem is ill-posed. In order to identify the components fif_ifi​, we assume that each fif_ifi​ can be written as a convolution of a Gaussian and a compactly supported density νi\nu_iνi​ with supp(ν1)∩supp(ν2)=∅\text{supp}(\nu_1)\cap \text{supp}(\nu_2)=\emptysetsupp(ν1​)∩supp(ν2​)=∅. Our main result shows that (1ε)Ω(log⁡log⁡1ε)(\frac{1}{\varepsilon})^{\Omega(\log\log \frac{1}{\varepsilon})}(ε1​)Ω(loglogε1​) samples are required for estimating each fif_ifi​. The proof relies on a quantitative Tauberian theorem that yields a fast rate of approximation with Gaussians, which may be of independent interest. To show this is tight, we also propose an algorithm that uses (1ε)O(log⁡log⁡1ε)(\frac{1}{\varepsilon})^{O(\log\log \frac{1}{\varepsilon})}(ε1​)O(loglogε1​) samples to estimate each fif_ifi​. Unlike existing approaches to learning latent variable models based on moment-matching and tensor methods, our proof instead involves a delicate analysis of an ill-conditioned linear system via orthogonal functions. Combining these bounds, we conclude that the optimal sample complexity of this problem properly lies in between polynomial and exponential, which is not common in learning theory.

View on arXiv
Comments on this paper