Intra-layer Nonuniform Quantization for Deep Convolutional Neural Network

07/10/2016
by   Fangxuan Sun, et al.
0

Deep convolutional neural network (DCNN) has achieved remarkable performance on object detection and speech recognition in recent years. However, the excellent performance of a DCNN incurs high computational complexity and large memory requirement. In this paper, an equal distance nonuniform quantization (ENQ) scheme and a K-means clustering nonuniform quantization (KNQ) scheme are proposed to reduce the required memory storage when low complexity hardware or software implementations are considered. For the VGG-16 and the AlexNet, the proposed nonuniform quantization schemes reduce the number of required memory storage by approximately 50% while achieving almost the same or even better classification accuracy compared to the state-of-the-art quantization method. Compared to the ENQ scheme, the proposed KNQ scheme provides a better tradeoff when higher accuracy is required.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/18/2014

Compressing Deep Convolutional Networks using Vector Quantization

Deep convolutional neural networks (CNN) has become the most promising m...
research
07/23/2022

Low-complexity CNNs for Acoustic Scene Classification

This paper presents a low-complexity framework for acoustic scene classi...
research
08/26/2020

Low Complexity Trellis-Coded Quantization in Versatile Video Coding

The forthcoming Versatile Video Coding (VVC) standard adopts the trellis...
research
02/27/2022

Arrhythmia Classifier Using Convolutional Neural Network with Adaptive Loss-aware Multi-bit Networks Quantization

Cardiovascular disease (CVDs) is one of the universal deadly diseases, a...
research
08/23/2021

Dynamic Network Quantization for Efficient Video Inference

Deep convolutional networks have recently achieved great success in vide...
research
03/04/2021

Neural Network-based Quantization for Network Automation

Deep Learning methods have been adopted in mobile networks, especially f...
research
01/23/2018

Stacked Filters Stationary Flow For Hardware-Oriented Acceleration Of Deep Convolutional Neural Networks

To address memory and computation resource limitations for hardware-orie...

Please sign up or login with your details

Forgot password? Click here to reset