Layer Folding: Neural Network Depth Reduction using Activation Linearization

06/17/2021
by   Amir Ben Dror, et al.
0

Despite the increasing prevalence of deep neural networks, their applicability in resource-constrained devices is limited due to their computational load. While modern devices exhibit a high level of parallelism, real-time latency is still highly dependent on networks' depth. Although recent works show that below a certain depth, the width of shallower networks must grow exponentially, we presume that neural networks typically exceed this minimal depth to accelerate convergence and incrementally increase accuracy. This motivates us to transform pre-trained deep networks that already exploit such advantages into shallower forms. We propose a method that learns whether non-linear activations can be removed, allowing to fold consecutive linear layers into one. We apply our method to networks pre-trained on CIFAR-10 and CIFAR-100 and find that they can all be transformed into shallower forms that share a similar depth. Finally, we use our method to provide more efficient alternatives to MobileNetV2 and EfficientNet-Lite architectures on the ImageNet classification task.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/14/2021

Non-deep Networks

Depth is the hallmark of deep neural networks. But more depth means more...
research
04/02/2020

Learning Sparse Ternary Neural Networks with Entropy-Constrained Trained Ternarization (EC2T)

Deep neural networks (DNN) have shown remarkable success in a variety of...
research
02/04/2020

A Deep Conditioning Treatment of Neural Networks

We study the role of depth in training randomly initialized overparamete...
research
04/15/2018

SparseNet: A Sparse DenseNet for Image Classification

Deep neural networks have made remarkable progresses on various computer...
research
04/04/2019

Transfer Learning with Sparse Associative Memories

In this paper, we introduce a novel layer designed to be used as the out...
research
10/06/2021

ParaDiS: Parallelly Distributable Slimmable Neural Networks

When several limited power devices are available, one of the most effici...
research
11/24/2016

Survey of Expressivity in Deep Neural Networks

We survey results on neural network expressivity described in "On the Ex...

Please sign up or login with your details

Forgot password? Click here to reset