Principal Filter Analysis for Guided Network Compression

07/20/2018
by   Xavier Suau, et al.
0

Principal Filter Analysis (PFA), is an elegant, easy to implement, yet effective methodology for neural network compression. PFA exploits the intrinsic correlation between filter responses within network layers to recommend a smaller network footprint. We propose two compression algorithms: the first allows a user to specify the proportion of the original spectral energy that should be preserved in each layer after compression, while the second is a parameter-free approach that automatically selects the compression used at each layer. Both algorithms are evaluated against several architectures and datasets, and we show considerable compression rates without compromising accuracy, e.g., for VGG-16 on CIFAR-10 and CIFAR-100 PFA achieves a compression rate of 8x and 3x with an accuracy gain of 0.4 respectively. In our tests we also demonstrate that networks compressed with PFA achieve an accuracy that is very close to the empirical upper bound for a given compression ratio.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/29/2020

Compressing Deep Neural Networks via Layer Fusion

This paper proposes layer fusion - a model compression technique that di...
research
02/08/2019

Architecture Compression

In this paper we propose a novel approach to model compression termed Ar...
research
07/25/2018

Coreset-Based Neural Network Compression

We propose a novel Convolutional Neural Network (CNN) compression algori...
research
06/07/2022

Neural Network Compression via Effective Filter Analysis and Hierarchical Pruning

Network compression is crucial to making the deep networks to be more ef...
research
01/26/2019

DeepSZ: A Novel Framework to Compress Deep Neural Networks by Using Error-Bounded Lossy Compression

DNNs have been quickly and broadly exploited to improve the data analysi...
research
11/20/2018

Multi-layer Pruning Framework for Compressing Single Shot MultiBox Detector

We propose a framework for compressing state-of-the-art Single Shot Mult...
research
09/13/2017

Flexible Network Binarization with Layer-wise Priority

How to effectively approximate real-valued parameters with binary codes ...

Please sign up or login with your details

Forgot password? Click here to reset