Dropout with Tabu Strategy for Regularizing Deep Neural Networks

08/29/2018
by   Zongjie Ma, et al.
4

Dropout has proven to be an effective technique for regularization and preventing the co-adaptation of neurons in deep neural networks (DNN). It randomly drops units with a probability p during the training stage of DNN. Dropout also provides a way of approximately combining exponentially many different neural network architectures efficiently. In this work, we add a diversification strategy into dropout, which aims at generating more different neural network architectures in a proper times of iterations. The dropped units in last forward propagation will be marked. Then the selected units for dropping in the current FP will be kept if they have been marked in the last forward propagation. We only mark the units from the last forward propagation. We call this new technique Tabu Dropout. Tabu Dropout has no extra parameters compared with the standard Dropout and also it is computationally cheap. The experiments conducted on MNIST, Fashion-MNIST datasets show that Tabu Dropout improves the performance of the standard dropout.

READ FULL TEXT

page 4

page 5

research
02/07/2019

Ising-Dropout: A Regularization Method for Training and Compression of Deep Neural Networks

Overfitting is a major problem in training machine learning models, spec...
research
04/13/2019

Shakeout: A New Approach to Regularized Deep Neural Network Training

Recent years have witnessed the success of deep neural networks in deali...
research
03/18/2017

Curriculum Dropout

Dropout is a very effective way of regularizing neural networks. Stochas...
research
11/18/2016

Compacting Neural Network Classifiers via Dropout Training

We introduce dropout compaction, a novel method for training feed-forwar...
research
11/28/2019

Continuous Dropout

Dropout has been proven to be an effective algorithm for training robust...
research
12/31/2018

Batch Size Influence on Performance of Graphic and Tensor Processing Units during Training and Inference Phases

The impact of the maximally possible batch size (for the better runtime)...
research
05/23/2018

Approximate Random Dropout

The training phases of Deep neural network (DNN) consume enormous proces...

Please sign up or login with your details

Forgot password? Click here to reset