Asymptotic optimality and minimal complexity of classification by random projection

08/11/2021
by   Mireille Boutin, et al.
0

The generalization error of a classifier is related to the complexity of the set of functions among which the classifier is chosen. Roughly speaking, the more complex the family, the greater the potential disparity between the training error and the population error of the classifier. This principle is embodied in layman's terms by Occam's razor principle, which suggests favoring low-complexity hypotheses over complex ones. We study a family of low-complexity classifiers consisting of thresholding the one-dimensional feature obtained by projecting the data on a random line after embedding it into a higher dimensional space parametrized by monomials of order up to k. More specifically, the extended data is projected n-times and the best classifier among those n (based on its performance on training data) is chosen. We obtain a bound on the generalization error of these low-complexity classifiers. The bound is less than that of any classifier with a non-trivial VC dimension, and thus less than that of a linear classifier. We also show that, given full knowledge of the class conditional densities, the error of the classifiers would converge to the optimal (Bayes) error as k and n go to infinity; if only a training dataset is given, we show that the classifiers will perfectly classify all the training points as k and n go to infinity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/14/2016

Generalization Error of Invariant Classifiers

This paper studies the generalization error of invariant classifiers. In...
research
10/16/2020

For self-supervised learning, Rationality implies generalization, provably

We prove a new upper bound on the generalization gap of classifiers that...
research
06/01/2020

Random Hyperboxes

This paper proposes a simple yet powerful ensemble classifier, called Ra...
research
08/22/2022

Low Complexity Classification Approach for Faster-than-Nyquist (FTN) Signalling Detection

Faster-than-Nyquist (FTN) signaling can improve the spectral efficiency ...
research
01/20/2017

Stability Enhanced Large-Margin Classifier Selection

Stability is an important aspect of a classification procedure because u...
research
11/07/2022

Highly over-parameterized classifiers generalize since bad solutions are rare

We study the generalization of over-parameterized classifiers where Empi...
research
10/09/2015

Some Theory For Practical Classifier Validation

We compare and contrast two approaches to validating a trained classifie...

Please sign up or login with your details

Forgot password? Click here to reset