Channel Pruning via Optimal Thresholding

03/10/2020
by   Yun Ye, et al.
17

Structured pruning, especially channel pruning is widely used for the reduced computational cost and the compatibility with off-the-shelf hardware devices. Among existing works, weights are typically removed using a predefined global threshold, or a threshold computed from a predefined metric. The predefined global threshold based designs ignore the variation among different layers and weights distribution, therefore, they may often result in sub-optimal performance caused by over-pruning or under-pruning. In this paper, we present a simple yet effective method, termed Optimal Thresholding (OT), to prune channels with layer dependent thresholds that optimally separate important from negligible channels. By using OT, most negligible or unimportant channels are pruned to achieve high sparsity while minimizing performance degradation. Since most important weights are preserved, the pruned model can be further fine-tuned and quickly converge with very few iterations. Our method demonstrates superior performance, especially when compared to the state-of-the-art designs at high levels of sparsity. On CIFAR-100, a pruned and fine-tuned DenseNet-121 by using OT achieves 75.99 FLOPs and 0.71M parameters. code is available at: https://github.com/yeyun11/netslim.

READ FULL TEXT
research
03/02/2021

Network Pruning via Resource Reallocation

Channel pruning is broadly recognized as an effective approach to obtain...
research
01/23/2020

Channel Pruning via Automatic Structure Search

Channel pruning is among the predominant approaches to compress deep neu...
research
09/29/2022

Is Complexity Required for Neural Network Pruning? A Case Study on Global Magnitude Pruning

Pruning neural networks has become popular in the last decade when it wa...
research
10/28/2022

Determining Ratio of Prunable Channels in MobileNet by Sparsity for Acoustic Scene Classification

MobileNet is widely used for Acoustic Scene Classification (ASC) in embe...
research
08/13/2023

Influence Function Based Second-Order Channel Pruning-Evaluating True Loss Changes For Pruning Is Possible Without Retraining

A challenge of channel pruning is designing efficient and effective crit...
research
06/20/2023

A Simple and Effective Pruning Approach for Large Language Models

As their size increases, Large Languages Models (LLMs) are natural candi...
research
01/12/2023

Why is the State of Neural Network Pruning so Confusing? On the Fairness, Comparison Setup, and Trainability in Network Pruning

The state of neural network pruning has been noticed to be unclear and e...

Please sign up or login with your details

Forgot password? Click here to reset