TENT: Efficient Quantization of Neural Networks on the tiny Edge with Tapered FixEd PoiNT

04/06/2021
by   Hamed F. Langroudi, et al.
0

In this research, we propose a new low-precision framework, TENT, to leverage the benefits of a tapered fixed-point numerical format in TinyML models. We introduce a tapered fixed-point quantization algorithm that matches the numerical format's dynamic range and distribution to that of the deep neural network model's parameter distribution at each layer. An accelerator architecture for the tapered fixed-point with TENT framework is proposed. Results show that the accuracy on classification tasks improves up to  31 with an energy overhead of  17-30 ResNet-18 models.

READ FULL TEXT

page 3

page 5

research
02/10/2022

F8Net: Fixed-Point 8-bit Only Multiplication for Network Quantization

Neural network quantization is a promising compression technique to redu...
research
02/09/2015

Deep Learning with Limited Numerical Precision

Training of large-scale deep neural networks is often constrained by the...
research
12/04/2015

Fixed-Point Performance Analysis of Recurrent Neural Networks

Recurrent neural networks have shown excellent performance in many appli...
research
11/19/2019

IFQ-Net: Integrated Fixed-point Quantization Networks for Embedded Vision

Deploying deep models on embedded devices has been a challenging problem...
research
05/10/2016

CORDIC-based Architecture for Powering Computation in Fixed-Point Arithmetic

We present a fixed point architecture (source VHDL code is provided) for...
research
10/07/2019

An Algorithmic Inference Approach to Learn Copulas

We introduce a new method for estimating the parameter of the bivariate ...
research
11/01/2019

Adaptive Precision Training: Quantify Back Propagation in Neural Networks with Fixed-point Numbers

Recent emerged quantization technique (i.e., using low bit-width fixed-p...

Please sign up or login with your details

Forgot password? Click here to reset