On approximating dropout noise injection

05/27/2019
by   Natalie Schluter, et al.
0

This paper examines the assumptions of the derived equivalence between dropout noise injection and L_2 regularisation for logistic regression with negative log loss. We show that the approximation method is based on a divergent Taylor expansion, making, subsequent work using this approximation to compare the dropout trained logistic regression model with standard regularisers, remains unfortunately ill-founded to date. Moreover, the approximation approach is shown to be invalid using any robust constraints. We show how this finding extends to general neural network topologies that use a cross-entropy prediction layer.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/12/2015

Bayesian Dropout

Dropout has recently emerged as a powerful and simple method for trainin...
research
03/02/2020

Score Engineered Logistic Regression

In several FICO studies logistic regression has been shown to be a very ...
research
03/31/2023

Almost Linear Constant-Factor Sketching for ℓ_1 and Logistic Regression

We improve upon previous oblivious sketching and turnstile streaming res...
research
02/21/2017

Delving Deeper into MOOC Student Dropout Prediction

In order to obtain reliable accuracy estimates for automatic MOOC dropou...
research
07/14/2021

Oblivious sketching for logistic regression

What guarantees are possible for solving logistic regression in one pass...
research
02/06/2023

Dropout Injection at Test Time for Post Hoc Uncertainty Quantification in Neural Networks

Among Bayesian methods, Monte-Carlo dropout provides principled tools fo...
research
06/27/2012

A Permutation Approach to Testing Interactions in Many Dimensions

To date, testing interactions in high dimensions has been a challenging ...

Please sign up or login with your details

Forgot password? Click here to reset