Computing f-Divergences and Distances of High-Dimensional Probability Density Functions – Low-Rank Tensor Approximations

11/13/2021
by   Alexander Litvinenko, et al.
0

Very often, in the course of uncertainty quantification tasks or data analysis, one has to deal with high-dimensional random variables (RVs). A high-dimensional RV can be described by its probability density (pdf) and/or by the corresponding probability characteristic functions (pcf), or by a polynomial chaos (PCE) or similar expansion. Here the interest is mainly to compute characterisations like the entropy, or relations between two distributions, like their Kullback-Leibler divergence. These are all computed from the pdf, which is often not available directly, and it is a computational challenge to even represent it in a numerically feasible fashion in case the dimension d is even moderately large. In this regard, we propose to represent the density by a high order tensor product, and approximate this in a low-rank format. We show how to go from the pcf or functional representation to the pdf. This allows us to reduce the computational complexity and storage cost from an exponential to a linear. The characterisations such as entropy or the f-divergences need the possibility to compute point-wise functions of the pdf. This normally rather trivial task becomes more difficult when the pdf is approximated in a low-rank tensor format, as the point values are not directly accessible any more. The data is considered as an element of a high order tensor space. The considered algorithms are independent of the representation of the data as a tensor. All that we require is that the data can be considered as an element of an associative, commutative algebra with an inner product. Such an algebra is isomorphic to a commutative sub-algebra of the usual matrix algebra, allowing the use of matrix algorithms to accomplish the mentioned tasks.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 11

06/13/2019

Post-Processing of High-Dimensional Data

Scientific computations or measurements may result in huge volumes of da...
01/22/2020

Rank Bounds for Approximating Gaussian Densities in the Tensor-Train Format

Low rank tensor approximations have been employed successfully, for exam...
08/27/2020

Nonparametric Multivariate Density Estimation: A Low-Rank Characteristic Function Approach

Effective non-parametric density estimation is a key challenge in high-d...
09/04/2020

High-Dimensional Uncertainty Quantification via Active and Rank-Adaptive Tensor Regression

Uncertainty quantification based on stochastic spectral methods suffers ...
06/23/2021

Committor functions via tensor networks

We propose a novel approach for computing committor functions, which des...
08/10/2020

Low-rank tensor reconstruction of concentrated densities with application to Bayesian inversion

Transport maps have become a popular mechanic to express complicated pro...
07/14/2020

Deep Composition of Tensor Trains using Squared Inverse Rosenblatt Transports

Characterising intractable high-dimensional random variables is one of t...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.