The Shallow End: Empowering Shallower Deep-Convolutional Networks through Auxiliary Outputs

11/06/2016
by   Yong Guo, et al.
0

The depth is one of the key factors behind the great success of convolutional neural networks (CNNs), with the gradient vanishing issue having been largely addressed by various nets, e.g. ResNet. However, when the depth goes very deep, the supervision information from the loss function will vanish due to the long backpropagation path, especially for those shallow layers. This means that intermediate layers receive less supervision information and will lead to redundancy in models. As a result, the model becomes very redundant and the over-fitting issue may happen. To address this, we propose a model, called AuxNet, by introducing auxiliary outputs at intermediate layers. Different from existing approaches, we propose a Multi-path training method to propagate not only gradients but also sufficient supervision informationfrommultipleauxiliaryoutputs. TheproposedAuxNetwithmulti-pathtrainingmethodgivesrisetomorecompact networks which outperform their very deep equivalent (i.e. ResNet). For example, AuxNet with 44 layers performs better than the ResNet equivalent with 110 layers on several benchmark data sets, i.e. CIFAR-10, CIFAR-100 and SVHN.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/12/2022

Contrastive Deep Supervision

The success of deep learning is usually accompanied by the growth in neu...
research
05/11/2015

Training Deeper Convolutional Networks with Deep Supervision

One of the most promising ways of improving the performance of deep conv...
research
06/16/2016

Convolutional Residual Memory Networks

Very deep convolutional neural networks (CNNs) yield state of the art re...
research
08/07/2016

Residual CNDS

Convolutional Neural networks nowadays are of tremendous importance for ...
research
09/18/2014

Deeply-Supervised Nets

Our proposed deeply-supervised nets (DSN) method simultaneously minimize...
research
05/23/2017

Input Fast-Forwarding for Better Deep Learning

This paper introduces a new architectural framework, known as input fast...
research
11/28/2021

Implicit Equivariance in Convolutional Networks

Convolutional Neural Networks(CNN) are inherently equivariant under tran...

Please sign up or login with your details

Forgot password? Click here to reset