ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2210.02415
20
3

A Fourier Approach to Mixture Learning

5 October 2022
Mingda Qiao
Guru Guruganesh
A. S. Rawat
Kumar Avinava Dubey
Manzil Zaheer
ArXivPDFHTML
Abstract

We revisit the problem of learning mixtures of spherical Gaussians. Given samples from mixture 1k∑j=1kN(μj,Id)\frac{1}{k}\sum_{j=1}^{k}\mathcal{N}(\mu_j, I_d)k1​∑j=1k​N(μj​,Id​), the goal is to estimate the means μ1,μ2,…,μk∈Rd\mu_1, \mu_2, \ldots, \mu_k \in \mathbb{R}^dμ1​,μ2​,…,μk​∈Rd up to a small error. The hardness of this learning problem can be measured by the separation Δ\DeltaΔ defined as the minimum distance between all pairs of means. Regev and Vijayaraghavan (2017) showed that with Δ=Ω(log⁡k)\Delta = \Omega(\sqrt{\log k})Δ=Ω(logk​) separation, the means can be learned using poly(k,d)\mathrm{poly}(k, d)poly(k,d) samples, whereas super-polynomially many samples are required if Δ=o(log⁡k)\Delta = o(\sqrt{\log k})Δ=o(logk​) and d=Ω(log⁡k)d = \Omega(\log k)d=Ω(logk). This leaves open the low-dimensional regime where d=o(log⁡k)d = o(\log k)d=o(logk). In this work, we give an algorithm that efficiently learns the means in d=O(log⁡k/log⁡log⁡k)d = O(\log k/\log\log k)d=O(logk/loglogk) dimensions under separation d/log⁡kd/\sqrt{\log k}d/logk​ (modulo doubly logarithmic factors). This separation is strictly smaller than log⁡k\sqrt{\log k}logk​, and is also shown to be necessary. Along with the results of Regev and Vijayaraghavan (2017), our work almost pins down the critical separation threshold at which efficient parameter learning becomes possible for spherical Gaussian mixtures. More generally, our algorithm runs in time poly(k)⋅f(d,Δ,ϵ)\mathrm{poly}(k)\cdot f(d, \Delta, \epsilon)poly(k)⋅f(d,Δ,ϵ), and is thus fixed-parameter tractable in parameters ddd, Δ\DeltaΔ and ϵ\epsilonϵ. Our approach is based on estimating the Fourier transform of the mixture at carefully chosen frequencies, and both the algorithm and its analysis are simple and elementary. Our positive results can be easily extended to learning mixtures of non-Gaussian distributions, under a mild condition on the Fourier spectrum of the distribution.

View on arXiv
Comments on this paper