Accelerated Training of Large-Scale Gaussian Mixtures by a Merger of Sublinear Approaches

10/01/2018
by   Florian Hirschberger, et al.
0

We combine two recent lines of research on sublinear clustering to significantly increase the efficiency of training Gaussian mixture models (GMMs) on large scale problems. First, we use a novel truncated variational EM approach for GMMs with isotropic Gaussians in order to increase clustering efficiency for large C (many clusters). Second, we use recent coreset approaches to increase clustering efficiency for large N (many data points). In order to derive a novel accelerated algorithm, we first show analytically how variational EM and coreset objectives can be merged to give rise to a new, combined clustering objective. Each iteration of the novel algorithm derived from this merged objective is then shown to have a run-time cost of O(N' G^2 D) per iteration, where N'<N is the coreset size and G^2<C is a constant related to the extent of local cluster neighborhoods. While enabling clustering with a strongly reduced number of distance evaluations per iteration, the combined approach is observed to still very effectively increase the clustering objective. In a series of numerical experiments on standard benchmarks, we use efficient seeding for initialization and evaluate the net computational demand of the merged approach in comparison to (already highly efficient) recent approaches. As result, depending on the dataset and number of clusters, the merged algorithm shows several times (and up to an order of magnitude) faster execution times to reach the same quantization errors as algorithms based on coresets or on variational EM alone.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/09/2017

Can clustering scale sublinearly with its clusters? A variational EM acceleration of GMMs and k-means

One iteration of k-means or EM for Gaussian mixture models (GMMs) scales...
research
04/16/2017

k-Means is a Variational EM Approximation of Gaussian Mixture Models

We show that k-means (Lloyd's algorithm) is equivalent to a variational ...
research
12/27/2017

The information bottleneck and geometric clustering

The information bottleneck (IB) approach to clustering takes a joint dis...
research
09/26/2020

An Adaptive EM Accelerator for Unsupervised Learning of Gaussian Mixture Models

We propose an Anderson Acceleration (AA) scheme for the adaptive Expecta...
research
02/07/2017

Truncated Variational EM for Semi-Supervised Neural Simpletrons

Inference and learning for probabilistic generative networks is often ve...
research
06/09/2015

Stagewise Learning for Sparse Clustering of Discretely-Valued Data

The performance of EM in learning mixtures of product distributions ofte...
research
06/23/2022

Quantifying Distances Between Clusters with Elliptical or Non-Elliptical Shapes

Finite mixture models that allow for a broad range of potentially non-el...

Please sign up or login with your details

Forgot password? Click here to reset