Improved Bayesian Compression

11/17/2017
by   Marco Federici, et al.
0

Compression of Neural Networks (NN) has become a highly studied topic in recent years. The main reason for this is the demand for industrial scale usage of NNs such as deploying them on mobile devices, storing them efficiently, transmitting them via band-limited channels and most importantly doing inference at scale. In this work, we propose to join the Soft-Weight Sharing and Variational Dropout approaches that show strong results to define a new state-of-the-art in terms of model compression.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2017

Soft Weight-Sharing for Neural Network Compression

The success of deep learning in numerous application domains created the...
research
05/18/2018

Neural Network Compression using Transform Coding and Clustering

With the deployment of neural networks on mobile devices and the necessi...
research
11/28/2019

Data-Driven Compression of Convolutional Neural Networks

Deploying trained convolutional neural networks (CNNs) to mobile devices...
research
01/25/2022

Bit-serial Weight Pools: Compression and Arbitrary Precision Execution of Neural Networks on Resource Constrained Processors

Applications of neural networks on edge systems have proliferated in rec...
research
06/04/2018

Dynamically Hierarchy Revolution: DirNet for Compressing Recurrent Neural Network on Mobile Devices

Recurrent neural networks (RNNs) achieve cutting-edge performance on a v...
research
11/23/2019

Compressing Representations for Embedded Deep Learning

Despite recent advances in architectures for mobile devices, deep learni...

Please sign up or login with your details

Forgot password? Click here to reset