Regularizing Neural Networks via Stochastic Branch Layers

10/03/2019
by   Wonpyo Park, et al.
0

We introduce a novel stochastic regularization technique for deep neural networks, which decomposes a layer into multiple branches with different parameters and merges stochastically sampled combinations of the outputs from the branches during training. Since the factorized branches can collapse into a single branch through a linear operation, inference requires no additional complexity compared to the ordinary layers. The proposed regularization method, referred to as StochasticBranch, is applicable to any linear layers such as fully-connected or convolution layers. The proposed regularizer allows the model to explore diverse regions of the model parameter space via multiple combinations of branches to find better local minima. An extensive set of experiments shows that our method effectively regularizes networks and further improves the generalization performance when used together with other existing regularization techniques.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/03/2018

Exploring Multi-Branch and High-Level Semantic Networks for Improving Pedestrian Detection

To better detect pedestrians of various scales, deep multi-scale methods...
research
09/18/2017

Coupled Ensembles of Neural Networks

We investigate in this paper the architecture of deep convolutional netw...
research
06/12/2022

Analysis of Branch Specialization and its Application in Image Decomposition

Branched neural networks have been used extensively for a variety of tas...
research
05/28/2018

Non-bifurcating phylogenetic tree inference via the adaptive LASSO

Phylogenetic tree inference using deep DNA sequencing is reshaping our u...
research
04/28/2022

A Closer Look at Branch Classifiers of Multi-exit Architectures

Multi-exit architectures consist of a backbone and branch classifiers th...
research
11/11/2022

Multilevel-in-Layer Training for Deep Neural Network Regression

A common challenge in regression is that for many problems, the degrees ...
research
05/03/2015

Making Sense of Hidden Layer Information in Deep Networks by Learning Hierarchical Targets

This paper proposes an architecture for deep neural networks with hidden...

Please sign up or login with your details

Forgot password? Click here to reset