Understanding and Accelerating EM Algorithm's Convergence by Fair Competition Principle and Rate-Verisimilitude Function

04/21/2021
by   Chenguang Lu, et al.
0

Why can the Expectation-Maximization (EM) algorithm for mixture models converge? Why can different initial parameters cause various convergence difficulties? The Q-L synchronization theory explains that the observed data log-likelihood L and the complete data log-likelihood Q are positively correlated; we can achieve maximum L by maximizing Q. According to this theory, the Deterministic Annealing EM (DAEM) algorithm's authors make great efforts to eliminate locally maximal Q for avoiding L's local convergence. However, this paper proves that in some cases, Q may and should decrease for L to increase; slow or local convergence exists only because of small samples and unfair competition. This paper uses marriage competition to explain different convergence difficulties and proposes the Fair Competition Principle (FCP) with an initialization map for improving initializations. It uses the rate-verisimilitude function, extended from the rate-distortion function, to explain the convergence of the EM and improved EM algorithms. This convergence proof adopts variational and iterative methods that Shannon et al. used for analyzing rate-distortion functions. The initialization map can vastly save both algorithms' running times for binary Gaussian mixtures. The FCP and the initialization map are useful for complicated mixtures but not sufficient; we need further studies for specific methods.

READ FULL TEXT

page 10

page 19

research
07/25/2020

Fair Marriage Principle and Initialization Map for the EM Algorithm

The popular convergence theory of the EM algorithm explains that the obs...
research
10/26/2018

From the EM Algorithm to the CM-EM Algorithm for Global Convergence of Mixture Models

The Expectation-Maximization (EM) algorithm for mixture models often res...
research
02/10/2023

Efficient and Accurate Learning of Mixtures of Plackett-Luce Models

Mixture models of Plackett-Luce (PL) – one of the most fundamental ranki...
research
11/21/2022

EM's Convergence in Gaussian Latent Tree Models

We study the optimization landscape of the log-likelihood function and t...
research
10/26/2018

Benefits of over-parameterization with EM

Expectation Maximization (EM) is among the most popular algorithms for m...
research
02/19/2019

On the Convergence of EM for truncated mixtures of two Gaussians

Motivated by a recent result of Daskalakis et al. DGTZ18, we analyze the...
research
06/21/2023

Discovering Intrinsic Spatial-Temporal Logic Rules to Explain Human Actions

We propose a logic-informed knowledge-driven modeling framework for huma...

Please sign up or login with your details

Forgot password? Click here to reset