Machine Learning's Dropout Training is Distributionally Robust Optimal

09/13/2020
by   Jose Blanchet, et al.
0

This paper shows that dropout training in Generalized Linear Models is the minimax solution of a two-player, zero-sum game where an adversarial nature corrupts a statistician's covariates using a multiplicative nonparametric errors-in-variables model. In this game—known as a Distributionally Robust Optimization problem—nature's least favorable distribution is dropout noise, where nature independently deletes entries of the covariate vector with some fixed probability δ. Our decision-theoretic analysis shows that dropout training—the statistician's minimax strategy in the game—indeed provides out-of-sample expected loss guarantees for distributions that arise from multiplicative perturbations of in-sample data. This paper also provides a novel, parallelizable, Unbiased Multi-Level Monte Carlo algorithm to speed-up the implementation of dropout training. Our algorithm has a much smaller computational cost compared to the naive implementation of dropout, provided the number of data points is much smaller than the dimension of the covariate vector.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/29/2015

Dropout as data augmentation

Dropout is typically interpreted as bagging a large number of models sha...
research
02/26/2022

Dropout can Simulate Exponential Number of Models for Sample Selection Techniques

Following Coteaching, generally in the literature, two models are used i...
research
10/09/2018

Unifying the Dropout Family Through Structured Shrinkage Priors

Dropout regularization of deep neural networks has been a mysterious yet...
research
08/14/2021

Investigating the Relationship Between Dropout Regularization and Model Complexity in Neural Networks

Dropout Regularization, serving to reduce variance, is nearly ubiquitous...
research
09/20/2018

Playing the Game of Universal Adversarial Perturbations

We study the problem of learning classifiers robust to universal adversa...
research
03/18/2021

Decision Theoretic Bootstrapping

The design and testing of supervised machine learning models combine two...
research
11/01/2019

Kinetic foundation of the zero-inflated negative binomial model for single-cell RNA sequencing data

Single-cell RNA sequencing data have complex features such as dropout ev...

Please sign up or login with your details

Forgot password? Click here to reset