Randomly initialized EM algorithm for two-component Gaussian mixture achieves near optimality in O(√(n)) iterations

08/28/2019
by   Yihong Wu, et al.
0

We analyze the classical EM algorithm for parameter estimation in the symmetric two-component Gaussian mixtures in d dimensions. We show that, even in the absence of any separation between components, provided that the sample size satisfies n=Ω(d log^3 d), the randomly initialized EM algorithm converges to an estimate in at most O(√(n)) iterations with high probability, which is at most O((d log^3 n/n)^1/4) in Euclidean distance from the true parameter and within logarithmic factors of the minimax rate of (d/n)^1/4. Both the nonparametric statistical rate and the sublinear convergence rate are direct consequences of the zero Fisher information in the worst case. Refined pointwise guarantees beyond worst-case analysis and convergence to the MLE are also shown under mild conditions. This improves the previous result of Balakrishnan et al <cit.> which requires strong conditions on both the separation of the components and the quality of the initialization, and that of Daskalakis et al <cit.> which requires sample splitting and restarting the EM iteration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/02/2020

EM Algorithm is Sample-Optimal for Learning Mixtures of Well-Separated Gaussians

We consider the problem of spherical Gaussian Mixture models with k ≥ 3 ...
research
03/29/2021

The EM Algorithm is Adaptively-Optimal for Unbalanced Symmetric Gaussian Mixtures

This paper studies the problem of estimating the means ±θ_*∈ℝ^d of a sym...
research
10/01/2018

Singularity, Misspecification, and the Convergence Rate of EM

A line of recent work has characterized the behavior of the EM algorithm...
research
06/04/2020

On the Minimax Optimality of the EM Algorithm for Learning Two-Component Mixed Linear Regression

We study the convergence rates of the EM algorithm for learning two-comp...
research
01/03/2021

Improved Convergence Guarantees for Learning Gaussian Mixture Models by EM and Gradient EM

We consider the problem of estimating the parameters a Gaussian Mixture ...
research
08/07/2016

Statistical Guarantees for Estimating the Centers of a Two-component Gaussian Mixture by EM

Recently, a general method for analyzing the statistical accuracy of the...
research
05/02/2013

Learning Mixtures of Bernoulli Templates by Two-Round EM with Performance Guarantee

Dasgupta and Shulman showed that a two-round variant of the EM algorithm...

Please sign up or login with your details

Forgot password? Click here to reset