ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1908.10935
243
42

Randomly initialized EM algorithm for two-component Gaussian mixture achieves near optimality in O(n)O(\sqrt{n})O(n​) iterations

28 August 2019
Yihong Wu
Harrison H. Zhou
ArXivPDFHTML
Abstract

We analyze the classical EM algorithm for parameter estimation in the symmetric two-component Gaussian mixtures in ddd dimensions. We show that, even in the absence of any separation between components, provided that the sample size satisfies n=Ω(dlog⁡3d)n=\Omega(d \log^3 d)n=Ω(dlog3d), the randomly initialized EM algorithm converges to an estimate in at most O(n)O(\sqrt{n})O(n​) iterations with high probability, which is at most O((dlog⁡3nn)1/4)O((\frac{d \log^3 n}{n})^{1/4})O((ndlog3n​)1/4) in Euclidean distance from the true parameter and within logarithmic factors of the minimax rate of (dn)1/4(\frac{d}{n})^{1/4}(nd​)1/4. Both the nonparametric statistical rate and the sublinear convergence rate are direct consequences of the zero Fisher information in the worst case. Refined pointwise guarantees beyond worst-case analysis and convergence to the MLE are also shown under mild conditions. This improves the previous result of Balakrishnan et al \cite{BWY17} which requires strong conditions on both the separation of the components and the quality of the initialization, and that of Daskalakis et al \cite{DTZ17} which requires sample splitting and restarting the EM iteration.

View on arXiv
Comments on this paper