Is the k-NN classifier in high dimensions affected by the curse of dimensionality?

10/19/2011
by   Vladimir Pestov, et al.
0

There is an increasing body of evidence suggesting that exact nearest neighbour search in high-dimensional spaces is affected by the curse of dimensionality at a fundamental level. Does it necessarily mean that the same is true for k nearest neighbours based learning algorithms such as the k-NN classifier? We analyse this question at a number of levels and show that the answer is different at each of them. As our first main observation, we show the consistency of a k approximate nearest neighbour classifier. However, the performance of the classifier in very high dimensions is provably unstable. As our second main observation, we point out that the existing model for statistical learning is oblivious of dimension of the domain and so every learning problem admits a universally consistent deterministic reduction to the one-dimensional case by means of a Borel isomorphism.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/10/2020

Universal consistency of Wasserstein k-NN classifier

The Wasserstein distance provides a notion of dissimilarities between pr...
research
11/29/2015

k-Nearest Neighbour Classification of Datasets with a Family of Distances

The k-nearest neighbour (k-NN) classifier is one of the oldest and most ...
research
07/01/2014

A Bayes consistent 1-NN classifier

We show that a simple modification of the 1-nearest neighbor classifier ...
research
10/31/2020

Optimal 1-NN Prototypes for Pathological Geometries

Using prototype methods to reduce the size of training datasets can dras...
research
02/28/2020

Universal consistency of the k-NN rule in metric spaces and Nagata dimension

The k nearest neighbour learning rule (under the uniform distance tie br...
research
04/08/2021

BR-NS: an Archive-less Approach to Novelty Search

As open-ended learning based on divergent search algorithms such as Nove...
research
03/02/2019

High-Dimensional Learning under Approximate Sparsity: A Unifying Framework for Nonsmooth Learning and Regularized Neural Networks

High-dimensional statistical learning (HDSL) has been widely applied in ...

Please sign up or login with your details

Forgot password? Click here to reset