Fast approximations of the Jeffreys divergence between univariate Gaussian mixture models via exponential polynomial densities

07/13/2021
by   Frank Nielsen, et al.
0

The Jeffreys divergence is a renown symmetrization of the statistical Kullback-Leibler divergence which is often used in statistics, machine learning, signal processing, and information sciences in general. Since the Jeffreys divergence between the ubiquitous Gaussian Mixture Models are not available in closed-form, many techniques with various pros and cons have been proposed in the literature to either (i) estimate, (ii) approximate, or (iii) lower and/or upper bound this divergence. In this work, we propose a simple yet fast heuristic to approximate the Jeffreys divergence between two univariate GMMs of arbitrary number of components. The heuristic relies on converting GMMs into pairs of dually parameterized probability densities belonging to exponential families. In particular, we consider Exponential-Polynomial Densities, and design a goodness-of-fit criterion to measure the dissimilarity between a GMM and a EPD which is a generalization of the Hyvärinen divergence. This criterion allows one to select the orders of the EPDs to approximate the GMMs. We demonstrate experimentally that the computational time of our heuristic improves over the stochastic Monte Carlo estimation baseline by several orders of magnitude while approximating reasonably well the Jeffreys divergence, specially when the univariate mixtures have a small number of modes.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/19/2016

Guaranteed bounds on the Kullback-Leibler divergence of univariate mixtures using piecewise log-sum-exp inequalities

Information-theoretic measures such as the entropy, cross-entropy and th...
research
01/09/2019

The statistical Minkowski distances: Closed-form formula for Gaussian Mixture Models

The traditional Minkowski distances are induced by the corresponding Min...
research
06/29/2018

Guaranteed Deterministic Bounds on the Total Variation Distance between Univariate Mixtures

The total variation distance is a core statistical distance between prob...
research
04/26/2021

Consistency issues in Gaussian Mixture Models reduction algorithms

In many contexts Gaussian Mixtures (GM) are used to approximate probabil...
research
01/03/2020

Pearson chi^2-divergence Approach to Gaussian Mixture Reduction and its Application to Gaussian-sum Filter and Smoother

The Gaussian mixture distribution is important in various statistical pr...
research
12/18/2020

Nonparametric Estimation of Repeated Densities with Heterogeneous Sample Sizes

We consider the estimation of densities in multiple subpopulations, wher...
research
11/03/2022

Using Signal Processing in Tandem With Adapted Mixture Models for Classifying Genomic Signals

Genomic signal processing has been used successfully in bioinformatics t...

Please sign up or login with your details

Forgot password? Click here to reset