Widening and Squeezing: Towards Accurate and Efficient QNNs

02/03/2020
by   Chuanjian Liu, et al.
0

Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters. Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques. However, we find the representation capability of quantization features is far weaker than full-precision features by experiments. We address this problem by projecting features in original full-precision networks to high-dimensional quantization features. Simultaneously, redundant quantization features will be eliminated in order to avoid unrestricted growth of dimensions for some datasets. Then, a compact quantization neural network but with sufficient representation ability will be established. Experimental results on benchmark datasets demonstrate that the proposed method is able to establish QNNs with much less parameters and calculations but almost the same performance as that of full-precision baseline models, e.g. 29.9% top-1 error of binary ResNet-18 on the ImageNet ILSVRC 2012 dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/24/2018

Precision Highway for Ultra Low-Precision Quantization

Neural network quantization has an inherent problem called accumulated q...
research
01/09/2020

Least squares binary quantization of neural networks

Quantizing weights and activations of deep neural networks results in si...
research
10/10/2020

Training Binary Neural Networks through Learning with Noisy Supervision

This paper formalizes the binarization operations over neural networks f...
research
09/16/2019

Searching for Accurate Binary Neural Architectures

Binary neural networks have attracted tremendous attention due to the ef...
research
11/23/2022

Join the High Accuracy Club on ImageNet with A Binary Neural Network Ticket

Binary neural networks are the extreme case of network quantization, whi...
research
10/24/2022

Precision Machine Learning

We explore unique considerations involved in fitting ML models to data w...
research
08/12/2020

FATNN: Fast and Accurate Ternary Neural Networks

Ternary Neural Networks (TNNs) have received much attention due to being...

Please sign up or login with your details

Forgot password? Click here to reset