Generalization Error of Invariant Classifiers

10/14/2016
by   Jure Sokolic, et al.
0

This paper studies the generalization error of invariant classifiers. In particular, we consider the common scenario where the classification task is invariant to certain transformations of the input, and that the classifier is constructed (or learned) to be invariant to these transformations. Our approach relies on factoring the input space into a product of a base space and a set of transformations. We show that whereas the generalization error of a non-invariant classifier is proportional to the complexity of the input space, the generalization error of an invariant classifier is proportional to the complexity of the base space. We also derive a set of sufficient conditions on the geometry of the base space and the set of transformations that ensure that the complexity of the base space is much smaller than the complexity of the input space. Our analysis applies to general classifiers such as convolutional neural networks. We demonstrate the implications of the developed theory for such classifiers with experiments on the MNIST and CIFAR-10 datasets.

READ FULL TEXT
research
11/10/2021

Understanding the Generalization Benefit of Model Invariance from a Data Perspective

Machine learning models that are developed to be invariant under certain...
research
08/11/2021

Asymptotic optimality and minimal complexity of classification by random projection

The generalization error of a classifier is related to the complexity of...
research
11/15/2017

Efficient Estimation of Generalization Error and Bias-Variance Components of Ensembles

For many applications, an ensemble of base classifiers is an effective s...
research
04/02/2018

Confidence from Invariance to Image Transformations

We develop a technique for automatically detecting the classification er...
research
07/14/2019

Feature space transformations and model selection to improve the performance of classifiers

Improving the performance of classifiers is the realm of prototype selec...
research
12/05/2012

On the Convergence Properties of Optimal AdaBoost

AdaBoost is one of the most popular machine-learning algorithms. It is s...
research
10/07/2013

Discriminative Features via Generalized Eigenvectors

Representing examples in a way that is compatible with the underlying cl...

Please sign up or login with your details

Forgot password? Click here to reset