Probabilistic Learning on Manifolds (PLoM) with Partition

02/22/2021
by   Christian Soize, et al.
0

The probabilistic learning on manifolds (PLoM) introduced in 2016 has solved difficult supervised problems for the “small data” limit where the number N of points in the training set is small. Many extensions have since been proposed, making it possible to deal with increasingly complex cases. However, the performance limit has been observed and explained for applications for which N is very small (50 for example) and for which the dimension of the diffusion-map basis is close to N. For these cases, we propose a novel extension based on the introduction of a partition in independent random vectors. We take advantage of this novel development to present improvements of the PLoM such as a simplified algorithm for constructing the diffusion-map basis and a new mathematical result for quantifying the concentration of the probability measure in terms of a probability upper bound. The analysis of the efficiency of this novel extension is presented through two applications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/28/2020

Probabilistic Learning on Manifolds

This paper presents mathematical results in support of the methodology o...
research
05/06/2022

Probabilistic learning constrained by realizations using a weak formulation of Fourier transform of probability measures

This paper deals with the taking into account a given set of realization...
research
03/21/2018

Entropy-based closure for probabilistic learning on manifolds

In a recent paper, the authors proposed a general methodology for probab...
research
06/05/2022

An information upper bound for probability sensitivity

Uncertain input of a mathematical model induces uncertainties in the out...
research
04/22/2013

Towards an Extension of the 2-tuple Linguistic Model to Deal With Unbalanced Linguistic Term sets

In the domain of Computing with words (CW), fuzzy linguistic approaches ...

Please sign up or login with your details

Forgot password? Click here to reset