Auto-adaptative Laplacian Pyramids for High-dimensional Data Analysis

11/26/2013
by   Ángela Fernández, et al.
0

Non-linear dimensionality reduction techniques such as manifold learning algorithms have become a common way for processing and analyzing high-dimensional patterns that often have attached a target that corresponds to the value of an unknown function. Their application to new points consists in two steps: first, embedding the new data point into the low dimensional space and then, estimating the function value on the test point from its neighbors in the embedded space. However, finding the low dimension representation of a test point, while easy for simple but often not powerful enough procedures such as PCA, can be much more complicated for methods that rely on some kind of eigenanalysis, such as Spectral Clustering (SC) or Diffusion Maps (DM). Similarly, when a target function is to be evaluated, averaging methods like nearest neighbors may give unstable results if the function is noisy. Thus, the smoothing of the target function with respect to the intrinsic, low-dimensional representation that describes the geometric structure of the examined data is a challenging task. In this paper we propose Auto-adaptive Laplacian Pyramids (ALP), an extension of the standard Laplacian Pyramids model that incorporates a modified LOOCV procedure that avoids the large cost of the standard one and offers the following advantages: (i) it selects automatically the optimal function resolution (stopping time) adapted to the data and its noise, (ii) it is easy to apply as it does not require parameterization, (iii) it does not overfit the training set and (iv) it adds no extra cost compared to other classical interpolation methods. We illustrate numerically ALP's behavior on a synthetic problem and apply it to the computation of the DM projection of new patterns and to the extension to them of target function values on a radiation forecasting problem over very high dimensional patterns.

READ FULL TEXT
research
02/01/2011

Vector Diffusion Maps and the Connection Laplacian

We introduce vector diffusion maps (VDM), a new mathematical framework ...
research
07/25/2022

Laplacian-based Cluster-Contractive t-SNE for High Dimensional Data Visualization

Dimensionality reduction techniques aim at representing high-dimensional...
research
06/25/2015

Diffusion Nets

Non-linear manifold learning enables high-dimensional data analysis, but...
research
07/01/2015

Bigeometric Organization of Deep Nets

In this paper, we build an organization of high-dimensional datasets tha...
research
09/09/2017

Simultaneously Learning Neighborship and Projection Matrix for Supervised Dimensionality Reduction

Explicitly or implicitly, most of dimensionality reduction methods need ...
research
03/22/2013

Sparse Projections of Medical Images onto Manifolds

Manifold learning has been successfully applied to a variety of medical ...
research
05/16/2021

Theoretical Foundations of t-SNE for Visualizing High-Dimensional Clustered Data

This study investigates the theoretical foundations of t-distributed sto...

Please sign up or login with your details

Forgot password? Click here to reset