Dropout as data augmentation

06/29/2015
by   Xavier Bouthillier, et al.
0

Dropout is typically interpreted as bagging a large number of models sharing parameters. We show that using dropout in a network can also be interpreted as a kind of data augmentation in the input space without domain knowledge. We present an approach to projecting the dropout noise within a network back into the input space, thereby generating augmented versions of the training data, and we show that training a deterministic network on the augmented samples yields similar results. Finally, we propose a new dropout noise scheme based on our observations and show that it improves dropout results without adding significant computational cost.

READ FULL TEXT

page 5

page 7

research
10/17/2021

Network Augmentation for Tiny Deep Learning

We introduce Network Augmentation (NetAug), a new training method for im...
research
06/08/2020

The Penalty Imposed by Ablated Data Augmentation

There is a set of data augmentation techniques that ablate parts of the ...
research
05/30/2023

ShuffleMix: Improving Representations via Channel-Wise Shuffle of Interpolated Hidden States

Mixup style data augmentation algorithms have been widely adopted in var...
research
09/13/2020

Machine Learning's Dropout Training is Distributionally Robust Optimal

This paper shows that dropout training in Generalized Linear Models is t...
research
08/06/2020

On the Accuracy of CRNNs for Line-Based OCR: A Multi-Parameter Evaluation

We investigate how to train a high quality optical character recognition...
research
02/26/2022

Dropout can Simulate Exponential Number of Models for Sample Selection Techniques

Following Coteaching, generally in the literature, two models are used i...
research
12/20/2014

Neural Network Regularization via Robust Weight Factorization

Regularization is essential when training large neural networks. As deep...

Please sign up or login with your details

Forgot password? Click here to reset