Shake-Shake regularization

05/21/2017
by   Xavier Gastaldi, et al.
0

The method introduced in this paper aims at helping deep learning practitioners faced with an overfit problem. The idea is to replace, in a multi-branch network, the standard summation of parallel branches with a stochastic affine combination. Applied to 3-branch residual networks, shake-shake regularization improves on the best single shot published results on CIFAR-10 and CIFAR-100 by reaching test errors of 2.86 Experiments on architectures without skip connections or Batch Normalization show encouraging results and open the door to a large set of applications. Code is available at https://github.com/xgastaldi/shake-shake

READ FULL TEXT
research
06/30/2020

Deep Isometric Learning for Visual Recognition

Initialization, normalization, and skip connections are believed to be t...
research
12/23/2021

A Robust Initialization of Residual Blocks for Effective ResNet Training without Batch Normalization

Batch Normalization is an essential component of all state-of-the-art ne...
research
04/14/2016

Deep Residual Networks with Exponential Linear Unit

Very deep convolutional neural networks introduced new problems like van...
research
07/27/2016

Instance Normalization: The Missing Ingredient for Fast Stylization

It this paper we revisit the fast stylization method introduced in Ulyan...
research
08/18/2021

Generalizing MLPs With Dropouts, Batch Normalization, and Skip Connections

A multilayer perceptron (MLP) is typically made of multiple fully connec...
research
11/23/2016

Deep Convolutional Neural Networks with Merge-and-Run Mappings

A deep residual network, built by stacking a sequence of residual blocks...
research
09/29/2017

Deep Competitive Pathway Networks

In the design of deep neural architectures, recent studies have demonstr...

Please sign up or login with your details

Forgot password? Click here to reset