GASL: Guided Attention for Sparsity Learning in Deep Neural Networks

01/07/2019
by   Amirsina Torfi, et al.
0

The main goal of network pruning is imposing sparsity on the neural network by increasing the number of parameters with zero value in order to reduce the architecture size and the computational speedup. In most of the previous research works, sparsity is imposed stochastically without considering any prior knowledge of the weights distribution or other internal network characteristics. Enforcing too much sparsity may induce accuracy drop due to the fact that a lot of important elements might have been eliminated. In this paper, we propose Guided Attention for Sparsity Learning (GASL) to achieve (1) model compression by having less number of elements and speed-up; (2) prevent the accuracy drop by supervising the sparsity operation via a guided attention mechanism and (3) introduce a generic mechanism that can be adapted for any type of architecture; Our work is aimed at providing a framework based on interpretable attention mechanisms for imposing structured and non-structured sparsity in deep neural networks. For Cifar-100 experiments, we achieved the state-of-the-art sparsity level and 2.91x speedup with competitive accuracy compared to the best method. For MNIST and LeNet architecture we also achieved the highest sparsity and speedup level.

READ FULL TEXT
research
05/19/2018

Sparse Architectures for Text-Independent Speaker Verification Using Deep Neural Networks

Network pruning is of great importance due to the elimination of the uni...
research
06/23/2020

Embedding Differentiable Sparsity into Deep Neural Network

In this paper, we propose embedding sparsity into the structure of deep ...
research
05/03/2019

Compressibility Loss for Neural Network Weights

In this paper we apply a compressibility loss that enables learning high...
research
10/28/2018

Learning Sparse Neural Networks via Sensitivity-Driven Regularization

The ever-increasing number of parameters in deep neural networks poses c...
research
09/12/2023

Accelerating Deep Neural Networks via Semi-Structured Activation Sparsity

The demand for efficient processing of deep neural networks (DNNs) on em...
research
09/22/2019

Blocking and sparsity for optimization of convolution calculation algorithm on GPUs

Convolution neural network (CNN) plays a paramount role in machine learn...
research
10/11/2019

Learning Cluster Structured Sparsity by Reweighting

Recently, the paradigm of unfolding iterative algorithms into finite-len...

Please sign up or login with your details

Forgot password? Click here to reset