Regularizing Neural Networks by Stochastically Training Layer Ensembles

11/21/2019
by   Alex Labach, et al.
0

Dropout and similar stochastic neural network regularization methods are often interpreted as implicitly averaging over a large ensemble of models. We propose STE (stochastically trained ensemble) layers, which enhance the averaging properties of such methods by training an ensemble of weight matrices with stochastic regularization while explicitly averaging outputs. This provides stronger regularization with no additional computational cost at test time. We show consistent improvement on various image classification tasks using standard network topologies.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/30/2022

Improving Ensemble Distillation With Weight Averaging and Diversifying Perturbation

Ensembles of deep neural networks have demonstrated superior performance...
research
12/20/2014

Neural Network Regularization via Robust Weight Factorization

Regularization is essential when training large neural networks. As deep...
research
05/20/2016

Swapout: Learning an ensemble of deep architectures

We describe Swapout, a new stochastic training method, that outperforms ...
research
10/12/2022

Deep Combinatorial Aggregation

Neural networks are known to produce poor uncertainty estimations, and a...
research
02/20/2021

Learning Neural Network Subspaces

Recent observations have advanced our understanding of the neural networ...
research
12/21/2013

An empirical analysis of dropout in piecewise linear networks

The recently introduced dropout training criterion for neural networks h...
research
10/28/2019

Distributed Estimation via Network Regularization

We propose a new method for distributed estimation of a linear model by ...

Please sign up or login with your details

Forgot password? Click here to reset