Regularising for invariance to data augmentation improves supervised learning

03/07/2022
by   Aleksander Botev, et al.
0

Data augmentation is used in machine learning to make the classifier invariant to label-preserving transformations. Usually this invariance is only encouraged implicitly by including a single augmented input during training. However, several works have recently shown that using multiple augmentations per input can improve generalisation or can be used to incorporate invariances more explicitly. In this work, we first empirically compare these recently proposed objectives that differ in whether they rely on explicit or implicit regularisation and at what level of the predictor they encode the invariances. We show that the predictions of the best performing method are also the most similar when compared on different augmentations of the same input. Inspired by this observation, we propose an explicit regulariser that encourages this invariance on the level of individual model predictions. Through extensive experiments on CIFAR-100 and ImageNet we show that this explicit regulariser (i) improves generalisation and (ii) equalises performance differences between all considered objectives. Our results suggest that objectives that encourage invariance on the level of the neural network itself generalise better than those that achieve invariance by averaging predictions of non-invariant models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/01/2020

On the Benefits of Invariance in Neural Networks

Many real world data analysis problems exhibit invariant structure, and ...
research
06/11/2019

Learning robust visual representations using data augmentation invariance

Deep convolutional neural networks trained for image object categorizati...
research
02/26/2022

Automated Data Augmentations for Graph Classification

Data augmentations are effective in improving the invariance of learning...
research
06/09/2021

Grounding inductive biases in natural images:invariance stems from variations in data

To perform well on unseen and potentially out-of-distribution samples, i...
research
09/24/2022

A Simple Strategy to Provable Invariance via Orbit Mapping

Many applications require robustness, or ideally invariance, of neural n...
research
02/28/2017

Learning Discrete Representations via Information Maximizing Self-Augmented Training

Learning discrete representations of data is a central machine learning ...
research
07/14/2022

On the Strong Correlation Between Model Invariance and Generalization

Generalization and invariance are two essential properties of any machin...

Please sign up or login with your details

Forgot password? Click here to reset