Anisotropic k-Nearest Neighbor Search Using Covariance Quadtree

08/31/2011
by   Eraldo Pereira Marinho, et al.
0

We present a variant of the hyper-quadtree that divides a multidimensional space according to the hyperplanes associated to the principal components of the data in each hyperquadrant. Each of the 2^λ hyper-quadrants is a data partition in a λ-dimension subspace, whose intrinsic dimensionality λ≤ d is reduced from the root dimensionality d by the principal components analysis, which discards the irrelevant eigenvalues of the local covariance matrix. In the present method a component is irrelevant if its length is smaller than, or comparable to, the local inter-data spacing. Thus, the covariance hyper-quadtree is fully adaptive to the local dimensionality. The proposed data-structure is used to compute the anisotropic K nearest neighbors (kNN), supported by the Mahalanobis metric. As an application, we used the present k nearest neighbors method to perform density estimation over a noisy data distribution. Such estimation method can be further incorporated to the smoothed particle hydrodynamics, allowing computer simulations of anisotropic fluid flows.

READ FULL TEXT

page 4

page 8

page 9

page 10

page 11

page 13

page 18

research
06/29/2022

LIDL: Local Intrinsic Dimension Estimation Using Approximate Likelihood

Most of the existing methods for estimating the local intrinsic dimensio...
research
06/28/2016

A Local Density-Based Approach for Local Outlier Detection

This paper presents a simple but effective density-based outlier detecti...
research
07/06/2021

DEANN: Speeding up Kernel-Density Estimation using Approximate Nearest Neighbor Search

Kernel Density Estimation (KDE) is a nonparametric method for estimating...
research
06/16/2021

Covariance-based smoothed particle hydrodynamics. A machine-learning application to simulating disc fragmentation

A PCA-based, machine learning version of the SPH method is proposed. In ...
research
11/29/2022

Variable selection and covariance structure identification using loadings

We provide sparse principal loading analysis which is a new concept that...
research
11/13/2017

Randomized Near Neighbor Graphs, Giant Components, and Applications in Data Science

If we pick n random points uniformly in [0,1]^d and connect each point t...
research
08/24/2021

Quantification of intrinsic quality of a principal dimension in correspondence analysis and taxicab correspondence analysis

Collins(2002, 2011) raised a number of issues with regards to correspond...

Please sign up or login with your details

Forgot password? Click here to reset