Multi-Sample Dropout for Accelerated Training and Better Generalization

05/23/2019
by   Hiroshi Inoue, et al.
0

Dropout is a simple but efficient regularization technique for achieving better generalization of deep neural networks (DNNs); hence it is widely used in tasks based on DNNs. During training, dropout randomly discards a portion of the neurons to avoid overfitting. This paper presents an enhanced dropout technique, which we call multi-sample dropout, for both accelerating training and improving generalization over the original dropout. The original dropout creates a randomly selected subset (called a dropout sample) from the input in each training iteration while the multi-sample dropout creates multiple dropout samples. The loss is calculated for each sample, and then the sample losses are averaged to obtain the final loss. This technique can be easily implemented without implementing a new operator by duplicating a part of the network after the dropout layer while sharing the weights among the duplicated fully connected layers. Experimental results showed that multi-sample dropout significantly accelerates training by reducing the number of iterations until convergence for image classification tasks using the ImageNet, CIFAR-10, CIFAR-100, and SVHN datasets. Multi-sample dropout does not significantly increase computation cost per iteration because most of the computation time is consumed in the convolution layers before the dropout layer, which are not duplicated. Experiments also showed that networks trained using multi-sample dropout achieved lower error rates and losses for both the training set and validation set.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2015

Efficient batchwise dropout training using submatrices

Dropout is a popular technique for regularizing artificial neural networ...
research
07/31/2015

Flip-Rotate-Pooling Convolution and Split Dropout on Convolution Neural Networks for Image Classification

This paper presents a new version of Dropout called Split Dropout (sDrop...
research
10/23/2020

On Convergence and Generalization of Dropout Training

We study dropout in two-layer neural networks with rectified linear unit...
research
10/11/2020

Advanced Dropout: A Model-free Methodology for Bayesian Dropout Optimization

Due to lack of data, overfitting ubiquitously exists in real-world appli...
research
09/04/2023

Dropout Attacks

Dropout is a common operator in deep learning, aiming to prevent overfit...
research
07/18/2023

Can Neural Network Memorization Be Localized?

Recent efforts at explaining the interplay of memorization and generaliz...
research
02/23/2020

Beyond Dropout: Feature Map Distortion to Regularize Deep Neural Networks

Deep neural networks often consist of a great number of trainable parame...

Please sign up or login with your details

Forgot password? Click here to reset