Effective and Efficient Dropout for Deep Convolutional Neural Networks

04/06/2019
by   Shaofeng Cai, et al.
0

Machine-learning-based data-driven applications have become ubiquitous, e.g., health-care analysis and database system optimization. Big training data and large (deep) models are crucial for good performance. Dropout has been widely used as an efficient regularization technique to prevent large models from overfitting. However, many recent works show that dropout does not bring much performance improvement for deep convolutional neural networks (CNNs), a popular deep learning model for data-driven applications. In this paper, we formulate existing dropout methods for CNNs under the same analysis framework to investigate the failures. We attribute the failure to the conflicts between the dropout and the batch normalization operation after it. Consequently, we propose to change the order of the operations, which results in new building blocks of CNNs.Extensive experiments on benchmark datasets CIFAR, SVHN and ImageNet have been conducted to compare the existing building blocks and our new building blocks with different dropout methods. The results confirm the superiority of our proposed building blocks due to the regularization and implicit model ensemble effect of dropout. In particular, we improve over state-of-the-art CNNs with significantly better performance of 3.17 1.44 respectively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/25/2018

Effective Building Block Design for Deep Convolutional Neural Networks using Search

Deep learning has shown promising results on many machine learning tasks...
research
01/28/2020

CSNNs: Unsupervised, Backpropagation-free Convolutional Neural Networks for Representation Learning

This work combines Convolutional Neural Networks (CNNs), clustering via ...
research
03/20/2019

Deep Octonion Networks

Deep learning is a research hot topic in the field of machine learning. ...
research
12/15/2014

MatConvNet - Convolutional Neural Networks for MATLAB

MatConvNet is an implementation of Convolutional Neural Networks (CNNs) ...
research
03/28/2023

Do Neural Topic Models Really Need Dropout? Analysis of the Effect of Dropout in Topic Modeling

Dropout is a widely used regularization trick to resolve the overfitting...
research
10/30/2014

DeepSentiBank: Visual Sentiment Concept Classification with Deep Convolutional Neural Networks

This paper introduces a visual sentiment concept classification method b...
research
11/07/2016

Regularizing CNNs with Locally Constrained Decorrelations

Regularization is key for deep learning since it allows training more co...

Please sign up or login with your details

Forgot password? Click here to reset