C2S2: Cost-aware Channel Sparse Selection for Progressive Network Pruning

04/06/2019
by   Chih-Yao Chiu, et al.
0

This paper describes a channel-selection approach for simplifying deep neural networks. Specifically, we propose a new type of generic network layer, called pruning layer, to seamlessly augment a given pre-trained model for compression. Each pruning layer, comprising 1 × 1 depth-wise kernels, is represented with a dual format: one is real-valued and the other is binary. The former enables a two-phase optimization process of network pruning to operate with an end-to-end differentiable network, and the latter yields the mask information for channel selection. Our method progressively performs the pruning task layer-wise, and achieves channel selection according to a sparsity criterion to favor pruning more channels. We also develop a cost-aware mechanism to prevent the compression from sacrificing the expected network performance. Our results for compressing several benchmark deep networks on image classification and semantic segmentation are comparable to those by state-of-the-art.

READ FULL TEXT
research
02/27/2019

Multi-loss-aware Channel Pruning of Deep Networks

Channel pruning, which seeks to reduce the model size by removing redund...
research
06/14/2018

PCAS: Pruning Channels with Attention Statistics

To implement deep neural networks on small embedded devices, conventiona...
research
04/30/2020

Out-of-the-box channel pruned networks

In the last decade convolutional neural networks have become gargantuan....
research
10/21/2021

CATRO: Channel Pruning via Class-Aware Trace Ratio Optimization

Deep convolutional neural networks are shown to be overkill with high pa...
research
11/01/2021

Back to Basics: Efficient Network Compression via IMP

Network pruning is a widely used technique for effectively compressing D...
research
07/30/2020

Growing Efficient Deep Networks by Structured Continuous Sparsification

We develop an approach to training deep networks while dynamically adjus...
research
01/04/2020

Discrimination-aware Network Pruning for Deep Model Compression

We study network pruning which aims to remove redundant channels/kernels...

Please sign up or login with your details

Forgot password? Click here to reset