NESTA: Hamming Weight Compression-Based Neural Proc. Engine

10/01/2019
by   Ali Mirzaeian, et al.
0

In this paper, we present NESTA, a specialized Neural engine that significantly accelerates the computation of convolution layers in a deep convolutional neural network, while reducing the computational energy. NESTA reformats Convolutions into 3 × 3 batches and uses a hierarchy of Hamming Weight Compressors to process each batch. Besides, when processing the convolution across multiple channels, NESTA, rather than computing the precise result of a convolution per channel, quickly computes an approximation of its partial sum, and a residual value such that if added to the approximate partial sum, generates the accurate output. Then, instead of immediately adding the residual, it uses (consumes) the residual when processing the next batch in the hamming weight compressors with available capacity. This mechanism shortens the critical path by avoiding the need to propagate carry signals during each round of computation and speeds up the convolution of each channel. In the last stage of computation, when the partial sum of the last channel is computed, NESTA terminates by adding the residual bits to the approximate output to generate a correct result.

READ FULL TEXT
research
12/17/2017

clcNet: Improving the Efficiency of Convolutional Neural Network using Channel Local Convolutions

Depthwise convolution and grouped convolution has been successfully appl...
research
02/25/2020

Generalized Hamming weight of Projective Toric Code over Hypersimplices

The d-th hypersimplex of R^s is the convex hull in R^s of all integral p...
research
10/30/2017

CrescendoNet: A Simple Deep Convolutional Neural Network with Ensemble Behavior

We introduce a new deep convolutional neural network, CrescendoNet, by s...
research
07/20/2018

Optimize Deep Convolutional Neural Network with Ternarized Weights and High Accuracy

Deep convolution neural network has achieved great success in many artif...
research
05/20/2018

Low-Cost Parameterizations of Deep Convolution Neural Networks

The main computational cost in the training of and prediction with Convo...
research
02/09/2022

Exploring Structural Sparsity in Neural Image Compression

Neural image compression have reached or out-performed traditional metho...
research
11/15/2017

Deep Epitome for Unravelling Generalized Hamming Network: A Fuzzy Logic Interpretation of Deep Learning

This paper gives a rigorous analysis of trained Generalized Hamming Netw...

Please sign up or login with your details

Forgot password? Click here to reset