Contextual Dropout: An Efficient Sample-Dependent Dropout Module

03/06/2021
by   Xinjie Fan, et al.
28

Dropout has been demonstrated as a simple and effective module to not only regularize the training process of deep neural networks, but also provide the uncertainty estimation for prediction. However, the quality of uncertainty estimation is highly dependent on the dropout probabilities. Most current models use the same dropout distributions across all data samples due to its simplicity. Despite the potential gains in the flexibility of modeling uncertainty, sample-dependent dropout, on the other hand, is less explored as it often encounters scalability issues or involves non-trivial model changes. In this paper, we propose contextual dropout with an efficient structural design as a simple and scalable sample-dependent dropout module, which can be applied to a wide range of models at the expense of only slightly increased memory and computational cost. We learn the dropout probabilities with a variational objective, compatible with both Bernoulli dropout and Gaussian dropout. We apply the contextual dropout module to various models with applications to image classification and visual question answering and demonstrate the scalability of the method with large-scale datasets, such as ImageNet and VQA 2.0. Our experimental results show that the proposed method outperforms baseline methods in terms of both accuracy and quality of uncertainty estimation.

READ FULL TEXT

page 9

page 22

page 23

page 24

page 25

page 26

page 27

research
11/01/2018

Variational Dropout via Empirical Bayes

We study the Automatic Relevance Determination procedure applied to deep...
research
03/08/2017

Dropout Inference in Bayesian Neural Networks with Alpha-divergences

To obtain uncertainty estimates with real-world Bayesian deep learning m...
research
06/15/2021

CODA: Constructivism Learning for Instance-Dependent Dropout Architecture Construction

Dropout is attracting intensive research interest in deep learning as an...
research
03/06/2020

Dropout Strikes Back: Improved Uncertainty Estimation via Diversity Sampled Implicit Ensembles

Modern machine learning models usually do not extrapolate well, i.e., th...
research
10/20/2020

Bayesian Attention Modules

Attention modules, as simple and effective tools, have not only enabled ...
research
05/28/2018

Adaptive Network Sparsification via Dependent Variational Beta-Bernoulli Dropout

While variational dropout approaches have been shown to be effective for...
research
02/12/2020

Learnable Bernoulli Dropout for Bayesian Deep Learning

In this work, we propose learnable Bernoulli dropout (LBD), a new model-...

Please sign up or login with your details

Forgot password? Click here to reset