A Generalized Zero-Shot Quantization of Deep Convolutional Neural Networks via Learned Weights Statistics

12/06/2021
by   Prasen Kumar Sharma, et al.
7

Quantizing the floating-point weights and activations of deep convolutional neural networks to fixed-point representation yields reduced memory footprints and inference time. Recently, efforts have been afoot towards zero-shot quantization that does not require original unlabelled training samples of a given task. These best-published works heavily rely on the learned batch normalization (BN) parameters to infer the range of the activations for quantization. In particular, these methods are built upon either empirical estimation framework or the data distillation approach, for computing the range of the activations. However, the performance of such schemes severely degrades when presented with a network that does not accommodate BN layers. In this line of thought, we propose a generalized zero-shot quantization (GZSQ) framework that neither requires original data nor relies on BN layer statistics. We have utilized the data distillation approach and leveraged only the pre-trained weights of the model to estimate enriched data for range calibration of the activations. To the best of our knowledge, this is the first work that utilizes the distribution of the pretrained weights to assist the process of zero-shot quantization. The proposed scheme has significantly outperformed the existing zero-shot works, e.g., an improvement of   33 MobileNetV2 and several other models that are w w/o BN layers, for a variety of tasks. We have also demonstrated the efficacy of the proposed work across multiple open-source quantization frameworks. Importantly, our work is the first attempt towards the post-training zero-shot quantization of futuristic unnormalized deep neural networks.

READ FULL TEXT

page 1

page 6

page 9

page 10

page 14

page 15

research
12/09/2022

Genie: Show Me the Data for Quantization

Zero-shot quantization is a promising approach for developing lightweigh...
research
01/01/2020

ZeroQ: A Novel Zero Shot Quantization Framework

Quantization is a promising approach for reducing the inference time and...
research
01/21/2021

Generative Zero-shot Network Quantization

Convolutional neural networks are able to learn realistic image priors f...
research
11/13/2022

Long-Range Zero-Shot Generative Deep Network Quantization

Quantization approximates a deep network model with floating-point numbe...
research
07/13/2021

Deep Neural Networks are Surprisingly Reversible: A Baseline for Zero-Shot Inversion

Understanding the behavior and vulnerability of pre-trained deep neural ...
research
03/31/2022

It's All In the Teacher: Zero-Shot Quantization Brought Closer to the Teacher

Model quantization is considered as a promising method to greatly reduce...
research
03/28/2022

SPIQ: Data-Free Per-Channel Static Input Quantization

Computationally expensive neural networks are ubiquitous in computer vis...

Please sign up or login with your details

Forgot password? Click here to reset