Understanding the Generalization Benefit of Model Invariance from a Data Perspective

11/10/2021
by   Sicheng Zhu, et al.
16

Machine learning models that are developed to be invariant under certain types of data transformations have shown improved generalization in practice. However, a principled understanding of why invariance benefits generalization is limited. Given a dataset, there is often no principled way to select "suitable" data transformations under which model invariance guarantees better generalization. This paper studies the generalization benefit of model invariance by introducing the sample cover induced by transformations, i.e., a representative subset of a dataset that can approximately recover the whole dataset using transformations. For any data transformations, we provide refined generalization bounds for invariant models based on the sample cover. We also characterize the "suitability" of a set of data transformations by the sample covering number induced by transformations, i.e., the smallest size of its induced sample covers. We show that we may tighten the generalization bounds for "suitable" transformations that have a small sample covering number. In addition, our proposed sample covering number can be empirically evaluated and thus provides a guide for selecting transformations to develop model invariance for better generalization. In experiments on multiple datasets, we evaluate sample covering numbers for some commonly used transformations and show that the smaller sample covering number for a set of transformations (e.g., the 3D-view transformation) indicates a smaller gap between the test and training error for invariant models, which verifies our propositions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/14/2022

On the Strong Correlation Between Model Invariance and Generalization

Generalization and invariance are two essential properties of any machin...
research
02/09/2021

More Is More – Narrowing the Generalization Gap by Adding Classification Heads

Overfit is a fundamental problem in machine learning in general, and in ...
research
10/14/2016

Generalization Error of Invariant Classifiers

This paper studies the generalization error of invariant classifiers. In...
research
11/07/2022

Using Set Covering to Generate Databases for Holistic Steganalysis

Within an operational framework, covers used by a steganographer are lik...
research
11/08/2019

Discovering Invariances in Healthcare Neural Networks

We study the invariance characteristics of pre-trained predictive models...
research
05/30/2022

PAC Generalization via Invariant Representations

One method for obtaining generalizable solutions to machine learning tas...
research
12/02/2021

Invariant Priors for Bayesian Quadrature

Bayesian quadrature (BQ) is a model-based numerical integration method t...

Please sign up or login with your details

Forgot password? Click here to reset