DiracNets: Training Very Deep Neural Networks Without Skip-Connections

06/01/2017
by   Sergey Zagoruyko, et al.
0

Deep neural networks with skip-connections, such as ResNet, show excellent performance in various image classification benchmarks. It is though observed that the initial motivation behind them - training deeper networks - does not actually hold true, and the benefits come from increased capacity, rather than from depth. Motivated by this, and inspired from ResNet, we propose a simple Dirac weight parameterization, which allows us to train very deep plain networks without skip-connections, and achieve nearly the same performance. This parameterization has a minor computational cost at training time and no cost at all at inference. We're able to achieve 95.5 34-layer deep plain network, surpassing 1001-layer deep ResNet, and approaching Wide ResNet. Our parameterization also mostly eliminates the need of careful initialization in residual and non-residual networks. The code and models for our experiments are available at https://github.com/szagoruyko/diracnets

READ FULL TEXT
research
06/30/2020

Deep Isometric Learning for Visual Recognition

Initialization, normalization, and skip connections are believed to be t...
research
11/01/2021

RMNet: Equivalently Removing Residual Connection from Networks

Although residual connection enables training very deep neural networks,...
research
03/24/2022

DyRep: Bootstrapping Training with Dynamic Re-parameterization

Structural re-parameterization (Rep) methods achieve noticeable improvem...
research
05/18/2018

Norm-Preservation: Why Residual Networks Can Become Extremely Deep?

Augmenting deep neural networks with skip connections, as introduced in ...
research
03/07/2017

Sharing Residual Units Through Collective Tensor Factorization in Deep Neural Networks

Residual units are wildly used for alleviating optimization difficulties...
research
04/26/2018

IamNN: Iterative and Adaptive Mobile Neural Network for Efficient Image Classification

Deep residual networks (ResNets) made a recent breakthrough in deep lear...

Please sign up or login with your details

Forgot password? Click here to reset