CAP: instance complexity-aware network pruning

09/08/2022
by   Jiapeng Wang, et al.
0

Existing differentiable channel pruning methods often attach scaling factors or masks behind channels to prune filters with less importance, and assume uniform contribution of input samples to filter importance. Specifically, the effects of instance complexity on pruning performance are not yet fully investigated. In this paper, we propose a simple yet effective differentiable network pruning method CAP based on instance complexity-aware filter importance scores. We define instance complexity related weight for each sample by giving higher weights to hard samples, and measure the weighted sum of sample-specific soft masks to model non-uniform contribution of different inputs, which encourages hard samples to dominate the pruning process and the model performance to be well preserved. In addition, we introduce a new regularizer to encourage polarization of the masks, such that a sweet spot can be easily found to identify the filters to be pruned. Performance evaluations on various network architectures and datasets demonstrate CAP has advantages over the state-of-the-arts in pruning large networks. For instance, CAP improves the accuracy of ResNet56 on CIFAR-10 dataset by 0.33 FLOPs, and prunes 87.75 Top-1 accuracy loss.

READ FULL TEXT
research
02/15/2022

Pruning Networks with Cross-Layer Ranking k-Reciprocal Nearest Filters

This paper focuses on filter-level network pruning. A novel pruning meth...
research
07/08/2020

Operation-Aware Soft Channel Pruning using Differentiable Masks

We propose a simple but effective data-driven channel pruning algorithm,...
research
10/20/2022

Pruning by Active Attention Manipulation

Filter pruning of a CNN is typically achieved by applying discrete masks...
research
04/15/2022

End-to-End Sensitivity-Based Filter Pruning

In this paper, we present a novel sensitivity-based filter pruning algor...
research
06/16/2022

Asymptotic Soft Cluster Pruning for Deep Neural Networks

Filter pruning method introduces structural sparsity by removing selecte...
research
09/30/2020

Pruning Filter in Filter

Pruning has become a very powerful and effective technique to compress a...
research
10/15/2021

Fire Together Wire Together: A Dynamic Pruning Approach with Self-Supervised Mask Prediction

Dynamic model pruning is a recent direction that allows for the inferenc...

Please sign up or login with your details

Forgot password? Click here to reset