Quantization in Layer's Input is Matter

02/10/2022
by   Daning Cheng, et al.
0

In this paper, we will show that the quantization in layer's input is more important than parameters' quantization for loss function. And the algorithm which is based on the layer's input quantization error is better than hessian-based mixed precision layout algorithm.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2022

Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss

Based on the model's resilience to computational noise, model quantizati...
research
12/05/2016

Towards the Limit of Network Quantization

Network quantization is one of network compression techniques to reduce ...
research
01/30/2015

Vector Quantization by Minimizing Kullback-Leibler Divergence

This paper proposes a new method for vector quantization by minimizing t...
research
07/15/2020

Image De-Quantization Using Generative Models as Priors

Image quantization is used in several applications aiming in reducing th...
research
12/24/2021

Stochastic Learning Equation using Monotone Increasing Resolution of Quantization

In this paper, we propose a quantized learning equation with a monotone ...
research
08/19/2020

Channel-wise Hessian Aware trace-Weighted Quantization of Neural Networks

Second-order information has proven to be very effective in determining ...
research
06/05/2016

Pairwise Quantization

We consider the task of lossy compression of high-dimensional vectors th...

Please sign up or login with your details

Forgot password? Click here to reset