Dynamic Runtime Feature Map Pruning

12/24/2018
by   Tailin Liang, et al.
0

High bandwidth requirements are an obstacle for accelerating the training and inference of deep neural networks. Most previous research focuses on reducing the size of kernel maps for inference. We analyze parameter sparsity of six popular convolutional neural networks - AlexNet, MobileNet, ResNet-50, SqueezeNet, TinyNet, and VGG16. Of the networks considered, those using ReLU (AlexNet, SqueezeNet, VGG16) contain a high percentage of 0-valued parameters and can be statically pruned. Networks with Non-ReLU activation functions in some cases may not contain any 0-valued parameters (ResNet-50, TinyNet). We also investigate runtime feature map usage and find that input feature maps comprise the majority of bandwidth requirements when depth-wise convolution and point-wise convolutions used. We introduce dynamic runtime pruning of feature maps and show that 10 loss of accuracy. We then extend dynamic pruning to allow for values within an epsilon of zero and show a further 5 1

READ FULL TEXT
research
02/24/2020

HRank: Filter Pruning using High-Rank Feature Map

Neural network pruning offers a promising prospect to facilitate deployi...
research
10/30/2016

Compact Deep Convolutional Neural Networks With Coarse Pruning

The learning capability of a neural network improves with increasing dep...
research
07/27/2019

Learning Instance-wise Sparsity for Accelerating Deep Models

Exploring deep convolutional neural networks of high efficiency and low ...
research
11/22/2019

SparseTrain:Leveraging Dynamic Sparsity in Training DNNs on General-Purpose SIMD Processors

Our community has greatly improved the efficiency of deep learning appli...
research
03/25/2017

More is Less: A More Complicated Network with Less Inference Complexity

In this paper, we present a novel and general network structure towards ...
research
04/13/2023

Boosting Convolutional Neural Networks with Middle Spectrum Grouped Convolution

This paper proposes a novel module called middle spectrum grouped convol...
research
05/02/2022

Zebra: Memory Bandwidth Reduction for CNN Accelerators With Zero Block Regularization of Activation Maps

The large amount of memory bandwidth between local buffer and external D...

Please sign up or login with your details

Forgot password? Click here to reset