Single-bit-per-weight deep convolutional neural networks without batch-normalization layers for embedded systems

07/16/2019
by   Mark D. McDonnell, et al.
0

Batch-normalization (BN) layers are thought to be an integrally important layer type in today's state-of-the-art deep convolutional neural networks for computer vision tasks such as classification and detection. However, BN layers introduce complexity and computational overheads that are highly undesirable for training and/or inference on low-power custom hardware implementations of real-time embedded vision systems such as UAVs, robots and Internet of Things (IoT) devices. They are also problematic when batch sizes need to be very small during training, and innovations such as residual connections introduced more recently than BN layers could potentially have lessened their impact. In this paper we aim to quantify the benefits BN layers offer in image classification networks, in comparison with alternative choices. In particular, we study networks that use shifted-ReLU layers instead of BN layers. We found, following experiments with wide residual networks applied to the ImageNet, CIFAR 10 and CIFAR 100 image classification datasets, that BN layers do not consistently offer a significant advantage. We found that the accuracy margin offered by BN layers depends on the data set, the network size, and the bit-depth of weights. We conclude that in situations where BN layers are undesirable due to speed, memory or complexity costs, that using shifted-ReLU layers instead should be considered; we found they can offer advantages in all these areas, and often do not impose a significant accuracy cost.

READ FULL TEXT
research
04/14/2016

Deep Residual Networks with Exponential Linear Unit

Very deep convolutional neural networks introduced new problems like van...
research
10/04/2019

Farkas layers: don't shift the data, fix the geometry

Successfully training deep neural networks often requires either batch n...
research
02/23/2018

Training wide residual networks for deployment using a single bit for each weight

For fast and energy-efficient deployment of trained deep neural networks...
research
11/28/2021

EffCNet: An Efficient CondenseNet for Image Classification on NXP BlueBox

Intelligent edge devices with built-in processors vary widely in terms o...
research
02/18/2019

LocalNorm: Robust Image Classification through Dynamically Regularized Normalization

While modern convolutional neural networks achieve outstanding accuracy ...
research
10/26/2020

MarbleNet: Deep 1D Time-Channel Separable Convolutional Neural Network for Voice Activity Detection

We present MarbleNet, an end-to-end neural network for Voice Activity De...
research
12/31/2018

Deep Residual Learning in the JPEG Transform Domain

We introduce a general method of performing Residual Network inference a...

Please sign up or login with your details

Forgot password? Click here to reset