Multi-Task Zipping via Layer-wise Neuron Sharing

05/24/2018
by   Xiaoxi He, et al.
4

Future mobile devices are anticipated to perceive, understand and react to the world on their own by running multiple correlated deep neural networks on-device. Yet the complexity of these neural networks needs to be trimmed down both withinmodel and cross-model to fit in mobile storage and memory. Previous studies focus on squeezing the redundancy within a single neural network. In this work, we aim to reduce the redundancy across multiple models. We propose Multi-Task Zipping (MTZ), a framework to automatically merge correlated, pre-trained deep neural networks for cross-model compression. Central in MTZ is a layer-wise neuron sharing and incoming weight updating scheme that induces a minimal change in the error function. MTZ inherits information from each model and demands light retraining to re-boost the accuracy of individual tasks. Evaluations show that MTZ is able to fully merge the hidden layers of two VGG-16 networks with a 3.18 and CelebA, or share 39.61 increase in the test errors for both tasks. The number of iterations to retrain the combined network is at least 17.8x lower than that of training a single VGG-16 network.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2019

Disentangling Redundancy for Multi-Task Pruning

Can prior network pruning strategies eliminate redundancy in multiple co...
research
04/11/2019

Cramnet: Layer-wise Deep Neural Network Compression with Knowledge Transfer from a Teacher Network

Neural Networks accomplish amazing things, but they suffer from computat...
research
02/03/2019

MICIK: MIning Cross-Layer Inherent Similarity Knowledge for Deep Model Compression

State-of-the-art deep model compression methods exploit the low-rank app...
research
08/26/2016

Scalable Compression of Deep Neural Networks

Deep neural networks generally involve some layers with mil- lions of pa...
research
10/29/2021

Model Fusion of Heterogeneous Neural Networks via Cross-Layer Alignment

Layer-wise model fusion via optimal transport, named OTFusion, applies s...
research
04/19/2015

Compressing Neural Networks with the Hashing Trick

As deep nets are increasingly used in applications suited for mobile dev...
research
07/12/2023

DeepMapping: The Case for Learned Data Mapping for Compression and Efficient Query Processing

Storing tabular data in a way that balances storage and query efficienci...

Please sign up or login with your details

Forgot password? Click here to reset