Differentiable Pruning Method for Neural Networks

04/24/2019
by   Jaedeok Kim, et al.
0

Architecture optimization is a promising technique to find an efficient neural network to meet certain requirements, which is usually a problem of selections. This paper introduces a concept of a trainable gate function and proposes a channel pruning method which finds automatically the optimal combination of channels using a simple gradient descent training procedure. The trainable gate function, which confers a differentiable property to discrete-valued variables, allows us to directly optimize loss functions that include discrete values such as the number of parameters or FLOPs that are generally non-differentiable. Channel pruning can be applied simply by appending trainable gate functions to each intermediate output tensor followed by fine-tuning the overall model, using any gradient-based training methods. Our experiments show that the proposed method can achieve better compression results on various models. For instance, our proposed method compresses ResNet-56 on CIFAR-10 dataset by half in terms of the number of FLOPs without accuracy drop.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2021

A Framework For Pruning Deep Neural Networks Using Energy-Based Models

A typical deep neural network (DNN) has a large number of trainable para...
research
05/07/2020

DMCP: Differentiable Markov Channel Pruning for Neural Networks

Recent works imply that the channel pruning can be regarded as searching...
research
02/28/2020

Learned Threshold Pruning

This paper presents a novel differentiable method for unstructured weigh...
research
02/23/2020

Gradual Channel Pruning while Training using Feature Relevance Scores for Convolutional Neural Networks

The enormous inference cost of deep neural networks can be scaled down b...
research
07/08/2020

Operation-Aware Soft Channel Pruning using Differentiable Masks

We propose a simple but effective data-driven channel pruning algorithm,...
research
10/28/2020

Differentiable Channel Pruning Search

In this paper, we propose the differentiable channel pruning search (DCP...
research
05/31/2023

Lottery Tickets in Evolutionary Optimization: On Sparse Backpropagation-Free Trainability

Is the lottery ticket phenomenon an idiosyncrasy of gradient-based train...

Please sign up or login with your details

Forgot password? Click here to reset