A Universal Approximation Theorem of Deep Neural Networks for Expressing Distributions

04/19/2020
by   Yulong Lu, et al.
0

This paper studies the universal approximation property of deep neural networks for representing probability distributions. Given a target distribution π and a source distribution p_z both defined on R^d, we prove under some assumptions that there exists a deep neural network g:R^dR with ReLU activation such that the push-forward measure (∇ g)_# p_z of p_z under the map ∇ g is arbitrarily close to the target measure π. The closeness are measured by three classes of integral probability metrics between probability distributions: 1-Wasserstein distance, maximum mean distance (MMD) and kernelized Stein discrepancy (KSD). We prove upper bounds for the size (width and depth) of the deep neural network in terms of the dimension d and the approximation error ε with respect to the three discrepancies. In particular, the size of neural network can grow exponentially in d when 1-Wasserstein distance is used as the discrepancy, whereas for both MMD and KSD the size of neural network only depends on d at most polynomially. Our proof relies on convergence estimates of empirical measures under aforementioned discrepancies and semi-discrete optimal transport.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/29/2021

On the capacity of deep generative networks for approximating distributions

We study the efficacy and efficiency of deep generative networks for app...
research
06/30/2020

Constructive Universal High-Dimensional Distribution Generation through Deep ReLU Networks

We present an explicit deep neural network construction that transforms ...
research
07/26/2021

High-Dimensional Distribution Generation Through Deep Neural Networks

We show that every d-dimensional probability distribution of bounded sup...
research
04/24/2021

A Class of Dimensionality-free Metrics for the Convergence of Empirical Measures

This paper concerns the convergence of empirical measures in high dimens...
research
11/22/2021

Cycle Consistent Probability Divergences Across Different Spaces

Discrepancy measures between probability distributions are at the core o...
research
03/23/2021

Depth-based pseudo-metrics between probability distributions

Data depth is a non parametric statistical tool that measures centrality...
research
07/26/2023

Fixed Integral Neural Networks

It is often useful to perform integration over learned functions represe...

Please sign up or login with your details

Forgot password? Click here to reset