Improving neural networks by preventing co-adaptation of feature detectors

07/03/2012
by   Geoffrey E. Hinton, et al.
0

When a large feedforward neural network is trained on a small training set, it typically performs poorly on held-out test data. This "overfitting" is greatly reduced by randomly omitting half of the feature detectors on each training case. This prevents complex co-adaptations in which a feature detector is only helpful in the context of several other specific feature detectors. Instead, each neuron learns to detect a feature that is generally helpful for producing the correct answer given the combinatorially large variety of internal contexts in which it must operate. Random "dropout" gives big improvements on many benchmark tasks and sets new records for speech and object recognition.

READ FULL TEXT

page 5

page 10

research
11/28/2019

Continuous Dropout

Dropout has been proven to be an effective algorithm for training robust...
research
11/28/2019

Transform-Invariant Convolutional Neural Networks for Image Classification and Search

Convolutional neural networks (CNNs) have achieved state-of-the-art resu...
research
08/28/2015

Partitioning Large Scale Deep Belief Networks Using Dropout

Deep learning methods have shown great promise in many practical applica...
research
02/27/2019

Customizing Object Detectors for Indoor Robots

Object detection models based on convolutional neural networks (CNNs) de...
research
03/28/2023

Do Neural Topic Models Really Need Dropout? Analysis of the Effect of Dropout in Topic Modeling

Dropout is a widely used regularization trick to resolve the overfitting...
research
05/01/2018

Internal node bagging: an explicit ensemble learning method in neural network training

We introduce a novel view to understand how dropout works as an inexplic...

Please sign up or login with your details

Forgot password? Click here to reset