8-bit Numerical Formats for Deep Neural Networks

06/06/2022
by   Badreddine Noune, et al.
56

Given the current trend of increasing size and complexity of machine learning architectures, it has become of critical importance to identify new approaches to improve the computational efficiency of model training. In this context, we address the advantages of floating-point over fixed-point representation, and present an in-depth study on the use of 8-bit floating-point number formats for activations, weights, and gradients for both training and inference. We explore the effect of different bit-widths for exponents and significands and different exponent biases. The experimental results demonstrate that a suitable choice of these low-precision formats enables faster training and reduced power consumption without any degradation in accuracy for a range of deep learning models for image classification and language processing.

READ FULL TEXT

page 6

page 8

page 26

research
12/08/2022

Customizing Number Representation and Precision

There is a growing interest in the use of reduced-precision arithmetic, ...
research
06/20/2018

Binary Ensemble Neural Network: More Bits per Network or More Networks per Bit?

Binary neural networks (BNN) have been studied extensively since they ru...
research
07/10/2020

Efficient Unpaired Image Dehazing with Cyclic Perceptual-Depth Supervision

Image dehazing without paired haze-free images is of immense importance,...
research
03/20/2023

Unit Scaling: Out-of-the-Box Low-Precision Training

We present unit scaling, a paradigm for designing deep learning models t...
research
06/15/2020

Neural gradients are lognormally distributed: understanding sparse and quantized training

Neural gradient compression remains a main bottleneck in improving train...
research
06/16/2020

Multi-Precision Policy Enforced Training (MuPPET): A precision-switching strategy for quantised fixed-point training of CNNs

Large-scale convolutional neural networks (CNNs) suffer from very long t...
research
07/30/2023

An Efficient Approach to Mitigate Numerical Instability in Backpropagation for 16-bit Neural Network Training

In this research, we delve into the intricacies of the numerical instabi...

Please sign up or login with your details

Forgot password? Click here to reset