Low-Precision Batch-Normalized Activations

02/27/2017
by   Benjamin Graham, et al.
0

Artificial neural networks can be trained with relatively low-precision floating-point and fixed-point arithmetic, using between one and 16 bits. Previous works have focused on relatively wide-but-shallow, feed-forward networks. We introduce a quantization scheme that is compatible with training very deep neural networks. Quantizing the network activations in the middle of each batch-normalization module can greatly reduce the amount of memory and computational power needed, with little loss in accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/22/2014

Training deep neural networks with low precision multiplications

Multipliers are the most space and power-hungry arithmetic operators of ...
research
02/09/2015

Deep Learning with Limited Numerical Precision

Training of large-scale deep neural networks is often constrained by the...
research
01/25/2018

Investigating the Effects of Dynamic Precision Scaling on Neural Network Training

Training neural networks is a time- and compute-intensive operation. Thi...
research
02/13/2018

Training and Inference with Integers in Deep Neural Networks

Researches on deep neural networks with discrete parameters and their de...
research
08/30/2020

Optimal Quantization for Batch Normalization in Neural Network Deployments and Beyond

Quantized Neural Networks (QNNs) use low bit-width fixed-point numbers f...
research
11/21/2018

Regularizing by the Variance of the Activations' Sample-Variances

Normalization techniques play an important role in supporting efficient ...
research
07/23/2020

Efficient Residue Number System Based Winograd Convolution

Prior research has shown that Winograd algorithm can reduce the computat...

Please sign up or login with your details

Forgot password? Click here to reset