Fire Together Wire Together: A Dynamic Pruning Approach with Self-Supervised Mask Prediction

10/15/2021
by   Sara Elkerdawy, et al.
0

Dynamic model pruning is a recent direction that allows for the inference of a different sub-network for each input sample during deployment. However, current dynamic methods rely on learning a continuous channel gating through regularization by inducing sparsity loss. This formulation introduces complexity in balancing different losses (e.g task loss, regularization loss). In addition, regularization-based methods lack transparent tradeoff hyperparameter selection to realize computational budget. Our contribution is twofold: 1) decoupled task and pruning training. 2) Simple hyperparameter selection that enables FLOPs reduction estimation before training. We propose to predict a mask to process k filters in a layer based on the activation of its previous layer. We pose the problem as a self-supervised binary classification problem. Each mask predictor module is trained to predict if the log-likelihood of each filter in the current layer belongs to the top-k activated filters. The value k is dynamically estimated for each input based on a novel criterion using the mass of heatmaps. We show experiments on several neural architectures, such as VGG, ResNet, and MobileNet on CIFAR and ImageNet datasets. On CIFAR, we reach similar accuracy to SOTA methods with 15 higher FLOPs reduction. Similarly in ImageNet, we achieve a lower drop in accuracy with up to 13

READ FULL TEXT

page 2

page 6

research
10/28/2020

Data Agnostic Filter Gating for Efficient Deep Networks

To deploy a well-trained CNN model on low-end computation edge devices, ...
research
01/18/2022

Pruning-aware Sparse Regularization for Network Pruning

Structural neural network pruning aims to remove the redundant channels ...
research
10/20/2022

Pruning by Active Attention Manipulation

Filter pruning of a CNN is typically achieved by applying discrete masks...
research
07/14/2020

REPrune: Filter Pruning via Representative Election

Even though norm-based filter pruning methods are widely accepted, it is...
research
03/22/2019

Towards Optimal Structured CNN Pruning via Generative Adversarial Learning

Structured pruning of filters or neurons has received increased focus fo...
research
09/08/2022

CAP: instance complexity-aware network pruning

Existing differentiable channel pruning methods often attach scaling fac...
research
03/08/2023

InfoBatch: Lossless Training Speed Up by Unbiased Dynamic Data Pruning

Data pruning aims to obtain lossless performances as training on the ori...

Please sign up or login with your details

Forgot password? Click here to reset