SlimNets: An Exploration of Deep Model Compression and Acceleration

08/01/2018
by   Ini Oguntola, et al.
0

Deep neural networks have achieved increasingly accurate results on a wide variety of complex tasks. However, much of this improvement is due to the growing use and availability of computational resources (e.g use of GPUs, more layers, more parameters, etc). Most state-of-the-art deep networks, despite performing well, over-parameterize approximate functions and take a significant amount of time to train. With increased focus on deploying deep neural networks on resource constrained devices like smart phones, there has been a push to evaluate why these models are so resource hungry and how they can be made more efficient. This work evaluates and compares three distinct methods for deep model compression and acceleration: weight pruning, low rank factorization, and knowledge distillation. Comparisons on VGG nets trained on CIFAR10 show that each of the models on their own are effective, but that the true power lies in combining them. We show that by combining pruning and knowledge distillation methods we can create a compressed network 85 times smaller than the original, all while retaining 96

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/24/2018

HAKD: Hardware Aware Knowledge Distillation

Despite recent developments, deploying deep neural networks on resource ...
research
06/09/2020

Knowledge Distillation: A Survey

In recent years, deep neural networks have been very successful in the f...
research
09/19/2020

Weight Distillation: Transferring the Knowledge in Neural Network Parameters

Knowledge distillation has been proven to be effective in model accelera...
research
02/20/2018

DeepThin: A Self-Compressing Library for Deep Neural Networks

As the industry deploys increasingly large and complex neural networks t...
research
02/03/2019

MICIK: MIning Cross-Layer Inherent Similarity Knowledge for Deep Model Compression

State-of-the-art deep model compression methods exploit the low-rank app...
research
07/03/2022

FasterAI: A Lightweight Library for Creating Sparse Neural Networks

FasterAI is a PyTorch-based library, aiming to facilitate the utilizatio...
research
07/20/2022

Model Compression for Resource-Constrained Mobile Robots

The number of mobile robots with constrained computing resources that ne...

Please sign up or login with your details

Forgot password? Click here to reset