Cramnet: Layer-wise Deep Neural Network Compression with Knowledge Transfer from a Teacher Network

04/11/2019
by   Jon Hoffman, et al.
0

Neural Networks accomplish amazing things, but they suffer from computational and memory bottlenecks that restrict their usage. Nowhere can this be better seen than in the mobile space, where specialized hardware is being created just to satisfy the demand for neural networks. Previous studies have shown that neural networks have vastly more connections than they actually need to do their work. This thesis develops a method that can compress networks to less than 10 accuracy, and without creating sparse networks that require special code to run.

READ FULL TEXT

page 17

page 26

page 27

page 28

page 33

page 34

research
05/24/2018

Multi-Task Zipping via Layer-wise Neuron Sharing

Future mobile devices are anticipated to perceive, understand and react ...
research
11/14/2017

Deep Rewiring: Training very sparse deep networks

Neuromorphic hardware tends to pose limits on the connectivity of deep n...
research
10/01/2015

Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding

Neural networks are both computationally intensive and memory intensive,...
research
04/21/2016

Deep Adaptive Network: An Efficient Deep Neural Network with Sparse Binary Connections

Deep neural networks are state-of-the-art models for understanding the c...
research
02/02/2022

Approximate Bisimulation Relations for Neural Networks and Application to Assured Neural Network Compression

In this paper, we propose a concept of approximate bisimulation relation...
research
01/23/2020

Chameleon: Adaptive Code Optimization for Expedited Deep Neural Network Compilation

Achieving faster execution with shorter compilation time can foster furt...

Please sign up or login with your details

Forgot password? Click here to reset