Fixflow: A Framework to Evaluate Fixed-point Arithmetic in Light-Weight CNN Inference

02/19/2023
by   Farhad Taheri, et al.
0

Convolutional neural networks (CNN) are widely used in resource-constrained devices in IoT applications. In order to reduce the computational complexity and memory footprint, the resource-constrained devices use fixed-point representation. This representation consumes less area and energy in hardware with similar classification accuracy compared to the floating-point ones. However, to employ the low-precision fixed-point representation, various considerations to gain high accuracy are required. Although many quantization and re-training techniques are proposed to improve the inference accuracy, these approaches are time-consuming and require access to the entire dataset. This paper investigates the effect of different fixed-point hardware units on CNN inference accuracy. To this end, we provide a framework called Fixflow to evaluate the effect of fixed-point computations performed at hardware level on CNN classification accuracy. We can employ different fixed-point considerations at the hardware accelerators.This includes rounding methods and adjusting the precision of the fixed-point operation's result. Fixflow can determine the impact of employing different arithmetic units (such as truncated multipliers) on CNN classification accuracy. Moreover, we evaluate the energy and area consumption of these units in hardware accelerators. We perform experiments on two common MNIST and CIFAR-10 datasets. Our results show that employing different methods at the hardware level specially with low-precision, can significantly change the classification accuracy.

READ FULL TEXT
research
02/09/2015

Deep Learning with Limited Numerical Precision

Training of large-scale deep neural networks is often constrained by the...
research
03/22/2016

Energy-Efficient ConvNets Through Approximate Computing

Recently ConvNets or convolutional neural networks (CNN) have come up as...
research
07/24/2018

Method for Hybrid Precision Convolutional Neural Network Representation

This invention addresses fixed-point representations of convolutional ne...
research
10/22/2019

Neural Network Training with Approximate Logarithmic Computations

The high computational complexity associated with training deep neural n...
research
01/08/2020

Training Progressively Binarizing Deep Networks Using FPGAs

While hardware implementations of inference routines for Binarized Neura...
research
02/19/2019

Towards Hardware Implementation of Neural Network-based Communication Algorithms

There is a recent interest in neural network (NN)-based communication al...
research
11/07/2016

Fixed-point Factorized Networks

In recent years, Deep Neural Networks (DNN) based methods have achieved ...

Please sign up or login with your details

Forgot password? Click here to reset