Multi-Precision Quantized Neural Networks via Encoding Decomposition of -1 and +1

05/31/2019
by   Qigong Sun, et al.
0

The training of deep neural networks (DNNs) requires intensive resources both for computation and for storage performance. Thus, DNNs cannot be efficiently applied to mobile phones and embedded devices, which seriously limits their applicability in industry applications. To address this issue, we propose a novel encoding scheme of using -1,+1 to decompose quantized neural networks (QNNs) into multi-branch binary networks, which can be efficiently implemented by bitwise operations (xnor and bitcount) to achieve model compression, computational acceleration and resource saving. Based on our method, users can easily achieve different encoding precisions arbitrarily according to their requirements and hardware resources. The proposed mechanism is very suitable for the use of FPGA and ASIC in terms of data storage and computation, which provides a feasible idea for smart chips. We validate the effectiveness of our method on both large-scale image classification tasks (e.g., ImageNet) and object detection tasks. In particular, our method with low-bit encoding can still achieve almost the same performance as its full-precision counterparts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/18/2021

Quantized Neural Networks via -1, +1 Encoding Decomposition and Acceleration

The training of deep neural networks (DNNs) always requires intensive re...
research
12/01/2022

Exploiting Kernel Compression on BNNs

Binary Neural Networks (BNNs) are showing tremendous success on realisti...
research
04/02/2020

Learning Sparse Ternary Neural Networks with Entropy-Constrained Trained Ternarization (EC2T)

Deep neural networks (DNN) have shown remarkable success in a variety of...
research
07/19/2021

A New Clustering-Based Technique for the Acceleration of Deep Convolutional Networks

Deep learning and especially the use of Deep Neural Networks (DNNs) prov...
research
07/31/2017

Streaming Architecture for Large-Scale Quantized Neural Networks on an FPGA-Based Dataflow Platform

Deep neural networks (DNNs) are used by different applications that are ...
research
04/05/2023

Adopting Two Supervisors for Efficient Use of Large-Scale Remote Deep Neural Networks

Recent decades have seen the rise of large-scale Deep Neural Networks (D...
research
08/26/2016

Scalable Compression of Deep Neural Networks

Deep neural networks generally involve some layers with mil- lions of pa...

Please sign up or login with your details

Forgot password? Click here to reset