Structured Sparse Ternary Weight Coding of Deep Neural Networks for Efficient Hardware Implementations

07/01/2017
by   Yoonho Boo, et al.
0

Deep neural networks (DNNs) usually demand a large amount of operations for real-time inference. Especially, fully-connected layers contain a large number of weights, thus they usually need many off-chip memory accesses for inference. We propose a weight compression method for deep neural networks, which allows values of +1 or -1 only at predetermined positions of the weights so that decoding using a table can be conducted easily. For example, the structured sparse (8,2) coding allows at most two non-zero values among eight weights. This method not only enables multiplication-free DNN implementations but also compresses the weight storage by up to x32 compared to floating-point networks. Weight distribution normalization and gradual pruning techniques are applied to mitigate the performance degradation. The experiments are conducted with fully-connected deep neural networks and convolutional neural networks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/19/2016

Quantized neural network design under weight capacity constraint

The complexity of deep neural network algorithms for hardware implementa...
research
07/02/2019

MimosaNet: An Unrobust Neural Network Preventing Model Stealing

Deep Neural Networks are robust to minor perturbations of the learned ne...
research
11/25/2020

Low Latency CMOS Hardware Acceleration for Fully Connected Layers in Deep Neural Networks

We present a novel low latency CMOS hardware accelerator for fully conne...
research
06/09/2023

Weight Freezing: A Regularization Approach for Fully Connected Layers with an Application in EEG Classification

In the realm of EEG decoding, enhancing the performance of artificial ne...
research
07/16/2019

An Inter-Layer Weight Prediction and Quantization for Deep Neural Networks based on a Smoothly Varying Weight Hypothesis

Network compression for deep neural networks has become an important par...
research
02/05/2018

Re-Weighted Learning for Sparsifying Deep Neural Networks

This paper addresses the topic of sparsifying deep neural networks (DNN'...
research
10/12/2020

TUTOR: Training Neural Networks Using Decision Rules as Model Priors

The human brain has the ability to carry out new tasks with limited expe...

Please sign up or login with your details

Forgot password? Click here to reset