Data augmentation instead of explicit regularization

06/11/2018
by   Alex Hernández-García, et al.
4

Modern deep artificial neural networks have achieved impressive results through models with very large capacity---compared to the number of training examples---that control overfitting with the help of different forms of regularization. Regularization can be implicit, as is the case of stochastic gradient descent and parameter sharing in convolutional layers, or explicit. Most common explicit regularization techniques, such as weight decay and dropout, reduce the effective capacity of the model and typically require the use of deeper and wider architectures to compensate for the reduced capacity. Although these techniques have been proven successful in terms of improved generalization, they seem to waste capacity. In contrast, data augmentation techniques do not reduce the effective capacity and improve generalization by increasing the number of training examples. In this paper we systematically analyze the effect of data augmentation on some popular architectures and conclude that data augmentation alone---without any other explicit regularization techniques---can achieve the same performance or higher as regularized models, especially when training with fewer examples, and exhibits much higher adaptability to changes in the architecture.

READ FULL TEXT

page 6

page 7

page 8

research
02/20/2018

Do deep nets really need weight decay and dropout?

The impressive success of modern deep neural networks on computer vision...
research
06/26/2019

Further advantages of data augmentation on convolutional neural networks

Data augmentation is a popular technique largely used to enhance the tra...
research
05/29/2018

Improved Mixed-Example Data Augmentation

In order to reduce overfitting, neural networks are typically trained wi...
research
11/11/2019

Evaluating Combinatorial Generalization in Variational Autoencoders

We evaluate the ability of variational autoencoders to generalize to uns...
research
06/11/2020

Mixup Training as the Complexity Reduction

Machine learning has achieved remarkable results in recent years due to ...
research
01/24/2020

Stochastic Optimization of Plain Convolutional Neural Networks with Simple methods

Convolutional neural networks have been achieving the best possible accu...
research
04/30/2020

When does data augmentation help generalization in NLP?

Neural models often exploit superficial ("weak") features to achieve goo...

Please sign up or login with your details

Forgot password? Click here to reset