Coupled Ensembles of Neural Networks

09/18/2017
by   Anuvabh Dutt, et al.
0

We investigate in this paper the architecture of deep convolutional networks. Building on existing state of the art models, we propose a reconfiguration of the model parameters into several parallel branches at the global network level, with each branch being a standalone CNN. We show that this arrangement is an efficient way to significantly reduce the number of parameters without losing performance or to significantly improve the performance with the same level of performance. The use of branches brings an additional form of regularization. In addition to the split into parallel branches, we propose a tighter coupling of these branches by placing the "fuse (averaging) layer" before the Log-Likelihood and SoftMax layers during training. This gives another significant performance improvement, the tighter coupling favouring the learning of better representations, even at the level of the individual branches. We refer to this branched architecture as "coupled ensembles". The approach is very generic and can be applied with almost any DCNN architecture. With coupled ensembles of DenseNet-BC and parameter budget of 25M, we obtain error rates of 2.92 SVHN tasks. For the same budget, DenseNet-BC has error rate of 3.46 and 1.8 networks, with 50M total parameters, we obtain error rates of 2.72 1.42

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/03/2019

Regularizing Neural Networks via Stochastic Branch Layers

We introduce a novel stochastic regularization technique for deep neural...
research
10/19/2021

Rep Works in Speaker Verification

Multi-branch convolutional neural network architecture has raised lots o...
research
07/16/2020

On Power Laws in Deep Ensembles

Ensembles of deep neural networks are known to achieve state-of-the-art ...
research
05/14/2020

Deep Ensembles on a Fixed Memory Budget: One Wide Network or Several Thinner Ones?

One of the generally accepted views of modern deep learning is that incr...
research
10/15/2021

Receptive Field Broadening and Boosting for Salient Object Detection

Salient object detection requires a comprehensive and scalable receptive...
research
05/11/2015

Training Deeper Convolutional Networks with Deep Supervision

One of the most promising ways of improving the performance of deep conv...
research
02/07/2018

ShakeDrop regularization

This paper proposes a powerful regularization method named ShakeDrop reg...

Please sign up or login with your details

Forgot password? Click here to reset