Non-Asymptotic Performance Guarantees for Neural Estimation of 𝖿-Divergences

03/11/2021
by   Sreejith Sreekumar, et al.
4

Statistical distances (SDs), which quantify the dissimilarity between probability distributions, are central to machine learning and statistics. A modern method for estimating such distances from data relies on parametrizing a variational form by a neural network (NN) and optimizing it. These estimators are abundantly used in practice, but corresponding performance guarantees are partial and call for further exploration. In particular, there seems to be a fundamental tradeoff between the two sources of error involved: approximation and estimation. While the former needs the NN class to be rich and expressive, the latter relies on controlling complexity. This paper explores this tradeoff by means of non-asymptotic error bounds, focusing on three popular choices of SDs – Kullback-Leibler divergence, chi-squared divergence, and squared Hellinger distance. Our analysis relies on non-asymptotic function approximation theorems and tools from empirical process theory. Numerical results validating the theory are also provided.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/07/2021

Neural Estimation of Statistical Divergences

Statistical divergences (SDs), which quantify the dissimilarity between ...
research
11/24/2017

Central limit theorems for Sinkhorn divergence between probability distributions on finite spaces and statistical applications

The notion of Sinkhorn divergence has recently gained popularity in mach...
research
11/21/2022

Limit distribution theory for f-Divergences

f-divergences, which quantify discrepancy between probability distributi...
research
04/08/2023

Non-asymptotic approximations of Gaussian neural networks via second-order Poincaré inequalities

There is a growing interest on large-width asymptotic properties of Gaus...
research
01/02/2018

A Concentration Result of Estimating Phi-Divergence using Data Dependent Partition

Estimation of the ϕ-divergence between two unknown probability distribut...
research
02/03/2020

Limit Distribution for Smooth Total Variation and χ^2-Divergence in High Dimensions

Statistical divergences are ubiquitous in machine learning as tools for ...
research
06/15/2020

Optimizing variational representations of divergences and accelerating their statistical estimation

Variational representations of distances and divergences between high-di...

Please sign up or login with your details

Forgot password? Click here to reset