SASL: Saliency-Adaptive Sparsity Learning for Neural Network Acceleration

03/12/2020
by   Jun Shi, et al.
0

Accelerating the inference speed of CNNs is critical to their deployment in real-world applications. Among all the pruning approaches, those implementing a sparsity learning framework have shown to be effective as they learn and prune the models in an end-to-end data-driven manner. However, these works impose the same sparsity regularization on all filters indiscriminately, which can hardly result in an optimal structure-sparse network. In this paper, we propose a Saliency-Adaptive Sparsity Learning (SASL) approach for further optimization. A novel and effective estimation of each filter, i.e., saliency, is designed, which is measured from two aspects: the importance for the prediction performance and the consumed computational resources. During sparsity learning, the regularization strength is adjusted according to the saliency, so our optimized format can better preserve the prediction performance while zeroing out more computation-heavy filters. The calculation for saliency introduces minimum overhead to the training process, which means our SASL is very efficient. During the pruning phase, in order to optimize the proposed data-dependent criterion, a hard sample mining strategy is utilized, which shows higher effectiveness and efficiency. Extensive experiments demonstrate the superior performance of our method. Notably, on ILSVRC-2012 dataset, our approach can reduce 49.7 and 0.05

READ FULL TEXT
research
06/04/2020

Weight Pruning via Adaptive Sparsity Loss

Pruning neural networks has regained interest in recent years as a means...
research
10/20/2019

Self-Adaptive Network Pruning

Deep convolutional neural networks have been proved successful on a wide...
research
07/03/2023

Structured Network Pruning by Measuring Filter-wise Interactions

Structured network pruning is a practical approach to reduce computation...
research
05/06/2020

Dependency Aware Filter Pruning

Convolutional neural networks (CNNs) are typically over-parameterized, b...
research
01/23/2019

Towards Compact ConvNets via Structure-Sparsity Regularized Filter Pruning

The success of convolutional neural networks (CNNs) in computer vision a...
research
04/03/2020

Composition of Saliency Metrics for Channel Pruning with a Myopic Oracle

The computation and memory needed for Convolutional Neural Network (CNN)...
research
03/21/2023

Performance-aware Approximation of Global Channel Pruning for Multitask CNNs

Global channel pruning (GCP) aims to remove a subset of channels (filter...

Please sign up or login with your details

Forgot password? Click here to reset