Compression of Deep Neural Networks on the Fly

09/29/2015
by   Guillaume Soulié, et al.
0

Thanks to their state-of-the-art performance, deep neural networks are increasingly used for object recognition. To achieve these results, they use millions of parameters to be trained. However, when targeting embedded applications the size of these models becomes problematic. As a consequence, their usage on smartphones or other resource limited devices is prohibited. In this paper we introduce a novel compression method for deep neural networks that is performed during the learning phase. It consists in adding an extra regularization term to the cost function of fully-connected layers. We combine this method with Product Quantization (PQ) of the trained weights for higher savings in storage consumption. We evaluate our method on two data sets (MNIST and CIFAR10), on which we achieve significantly larger compression rates than state-of-the-art methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2018

Wide Compression: Tensor Ring Nets

Deep neural networks have demonstrated state-of-the-art performance in a...
research
12/18/2014

Compressing Deep Convolutional Networks using Vector Quantization

Deep convolutional neural networks (CNN) has become the most promising m...
research
12/07/2017

AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training

Highly distributed training of Deep Neural Networks (DNNs) on future com...
research
02/13/2019

Structured Bayesian Compression for Deep models in mobile enabled devices for connected healthcare

Deep Models, typically Deep neural networks, have millions of parameters...
research
02/21/2023

FrankenSplit: Saliency Guided Neural Feature Compression with Shallow Variational Bottleneck Injection

The rise of mobile AI accelerators allows latency-sensitive applications...
research
05/25/2018

Tensorized Spectrum Preserving Compression for Neural Networks

Modern neural networks can have tens of millions of parameters, and are ...
research
07/15/2020

Compression strategies and space-conscious representations for deep neural networks

Recent advances in deep learning have made available large, powerful con...

Please sign up or login with your details

Forgot password? Click here to reset