Towards Dropout Training for Convolutional Neural Networks

12/01/2015
by   Haibing Wu, et al.
0

Recently, dropout has seen increasing use in deep learning. For deep convolutional neural networks, dropout is known to work well in fully-connected layers. However, its effect in convolutional and pooling layers is still not clear. This paper demonstrates that max-pooling dropout is equivalent to randomly picking activation based on a multinomial distribution at training time. In light of this insight, we advocate employing our proposed probabilistic weighted pooling, instead of commonly used max-pooling, to act as model averaging at test time. Empirical evidence validates the superiority of probabilistic weighted pooling. We also empirically show that the effect of convolutional dropout is not trivial, despite the dramatically reduced possibility of over-fitting due to the convolutional architecture. Elaborately designing dropout training simultaneously in max-pooling and fully-connected layers, we achieve state-of-the-art performance on MNIST, and very competitive results on CIFAR-10 and CIFAR-100, relative to other approaches without data augmentation. Finally, we compare max-pooling dropout and stochastic pooling, both of which introduce stochasticity based on multinomial distributions at pooling stage.

READ FULL TEXT
research
12/04/2015

Max-Pooling Dropout for Regularization of Convolutional Neural Networks

Recently, dropout has seen increasing use in deep learning. For deep con...
research
01/16/2013

Stochastic Pooling for Regularization of Deep Convolutional Neural Networks

We introduce a simple and effective method for regularizing large convol...
research
02/18/2013

Maxout Networks

We consider the problem of designing models to leverage a recently intro...
research
07/30/2020

Generalization Comparison of Deep Neural Networks via Output Sensitivity

Although recent works have brought some insights into the performance im...
research
04/21/2016

TI-POOLING: transformation-invariant pooling for feature learning in Convolutional Neural Networks

In this paper we present a deep neural network topology that incorporate...
research
02/09/2017

Effective face landmark localization via single deep network

In this paper, we propose a novel face alignment method using single dee...
research
02/20/2023

Kernel function impact on convolutional neural networks

This paper investigates the usage of kernel functions at the different l...

Please sign up or login with your details

Forgot password? Click here to reset