LilNetX: Lightweight Networks with EXtreme Model Compression and Structured Sparsification

04/06/2022
by   Sharath Girish, et al.
0

We introduce LilNetX, an end-to-end trainable technique for neural networks that enables learning models with specified accuracy-rate-computation trade-off. Prior works approach these problems one at a time and often require post-processing or multistage training which become less practical and do not scale very well for large datasets or architectures. Our method constructs a joint training objective that penalizes the self-information of network parameters in a reparameterized latent space to encourage small model size while also introducing priors to increase structured sparsity in the parameter space to reduce computation. We achieve up to 50 model sparsity on ResNet-20 while retaining the same accuracy on the CIFAR-10 dataset as well as 35 ResNet-50 trained on ImageNet, when compared to existing state-of-the-art model compression methods. Code is available at https://github.com/Sharath-girish/LilNetX.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/15/2019

Model Compression by Entropy Penalized Reparameterization

We describe an end-to-end neural network weight compression approach tha...
research
07/05/2022

PoF: Post-Training of Feature Extractor for Improving Generalization

It has been intensively investigated that the local shape, especially fl...
research
06/21/2023

Efficient ResNets: Residual Network Design

ResNets (or Residual Networks) are one of the most commonly used models ...
research
03/13/2023

OTOV2: Automatic, Generic, User-Friendly

The existing model compression methods via structured pruning typically ...
research
08/12/2016

Learning Structured Sparsity in Deep Neural Networks

High demand for computation resources severely hinders deployment of lar...
research
10/13/2022

Parameter-Efficient Masking Networks

A deeper network structure generally handles more complicated non-linear...
research
01/18/2023

Local Learning with Neuron Groups

Traditional deep network training methods optimize a monolithic objectiv...

Please sign up or login with your details

Forgot password? Click here to reset