Empirical Analysis of Knowledge Distillation Technique for Optimization of Quantized Deep Neural Networks

09/04/2019
by   Sungho Shin, et al.
0

Knowledge distillation (KD) is a very popular method for model size reduction. Recently, the technique is exploited for quantized deep neural networks (QDNNs) training as a way to restore the performance sacrificed by word-length reduction. KD, however, employs additional hyper-parameters, such as temperature, coefficient, and the size of teacher network for QDNN training. We analyze the effect of these hyper-parameters for QDNN optimization with KD. We find that these hyper-parameters are inter-related, and also introduce a simple and effective technique that reduces coefficient during training. With KD employing the proposed hyper-parameters, we achieve the test accuracy of 92.7 and CIFAR-100 data sets, respectively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/26/2021

PURSUhInT: In Search of Informative Hint Points Based on Layer Clustering for Knowledge Distillation

We propose a novel knowledge distillation methodology for compressing de...
research
09/30/2020

Stochastic Precision Ensemble: Self-Knowledge Distillation for Quantized Deep Neural Networks

The quantization of deep neural networks (QDNNs) has been actively studi...
research
03/25/2020

Circumventing Outliers of AutoAugment with Knowledge Distillation

AutoAugment has been a powerful algorithm that improves the accuracy of ...
research
02/17/2022

General Cyclical Training of Neural Networks

This paper describes the principle of "General Cyclical Training" in mac...
research
02/08/2018

Imitation networks: Few-shot learning of neural networks from scratch

In this paper, we propose imitation networks, a simple but effective met...
research
09/14/2021

Exploring the Connection between Knowledge Distillation and Logits Matching

Knowledge distillation is a generalized logits matching technique for mo...
research
09/29/2022

Hyper-Representations as Generative Models: Sampling Unseen Neural Network Weights

Learning representations of neural network weights given a model zoo is ...

Please sign up or login with your details

Forgot password? Click here to reset