PDP: Parameter-free Differentiable Pruning is All You Need

05/18/2023
by   Minsik Cho, et al.
0

DNN pruning is a popular way to reduce the size of a model, improve the inference latency, and minimize the power consumption on DNN accelerators. However, existing approaches might be too complex, expensive or ineffective to apply to a variety of vision/language tasks, DNN architectures and to honor structured pruning constraints. In this paper, we propose an efficient yet effective train-time pruning scheme, Parameter-free Differentiable Pruning (PDP), which offers state-of-the-art qualities in model size, accuracy, and training cost. PDP uses a dynamic function of weights during training to generate soft pruning masks for the weights in a parameter-free manner for a given pruning target. While differentiable, the simplicity and efficiency of PDP make it universal enough to deliver state-of-the-art random/structured/channel pruning results on various vision and natural language tasks. For example, for MobileNet-v1, PDP can achieve 68.2 ImageNet1k accuracy at 86.6 from the state-of-the-art algorithms. Also, PDP yields over 83.1 Multi-Genre Natural Language Inference with 90 next best from the existing techniques shows 81.5 can be applied to structured pruning, such as N:M pruning and channel pruning. For 1:4 structured pruning of ResNet18, PDP improved the top-1 ImageNet1k accuracy by over 3.6 ResNet50, PDP reduced the top-1 ImageNet1k accuracy by 0.6 state-of-the-art.

READ FULL TEXT

page 4

page 14

page 16

research
05/30/2021

MLPruning: A Multilevel Structured Pruning Framework for Transformer-based Models

Pruning is an effective method to reduce the memory footprint and comput...
research
03/26/2023

Task-oriented Memory-efficient Pruning-Adapter

The Outstanding performance and growing size of Large Language Models ha...
research
05/30/2022

Parameter Efficient Diff Pruning for Bias Mitigation

In recent years language models have achieved state of the art performan...
research
04/27/2022

Channel Pruned YOLOv5-based Deep Learning Approach for Rapid and Accurate Outdoor Obstacles Detection

One-stage algorithm have been widely used in target detection systems th...
research
03/07/2023

Gradient-Free Structured Pruning with Unlabeled Data

Large Language Models (LLMs) have achieved great success in solving diff...
research
07/08/2020

Operation-Aware Soft Channel Pruning using Differentiable Masks

We propose a simple but effective data-driven channel pruning algorithm,...
research
04/05/2023

To Asymmetry and Beyond: Structured Pruning of Sequence to Sequence Models for Improved Inference Efficiency

Sequence-to-sequence language models can be used to produce abstractive ...

Please sign up or login with your details

Forgot password? Click here to reset