A Survey of Model Compression and Acceleration for Deep Neural Networks

10/23/2017
by   Yu Cheng, et al.
0

Deep convolutional neural networks (CNNs) have recently achieved great success in many visual recognition tasks. However, existing deep convolutional neural network models are computationally expensive and memory intensive, hindering their deployment in devices with low memory resources or in applications with strict latency requirements. Therefore, a natural thought is to perform model compression and acceleration in deep networks without significantly decreasing the model performance. During the past few years, tremendous progresses have been made in this area. In this paper, we survey the recent advanced techniques for compacting and accelerating CNNs model developed. These techniques are roughly categorized into four schemes: parameter pruning and sharing, low-rank factorization, transfered/compact convolutional filters and knowledge distillation. Methods of parameter pruning and sharing will be described at the beginning, after that the other techniques will be introduced. For each scheme, we provide insightful analysis regarding the performance, related applications, advantages and drawbacks etc. Then we will go through a few very recent additional successful methods, for example, dynamic networks and stochastic depths networks. After that, we survey the evaluation matrix, main datasets used for evaluating the model performance and recent benchmarking efforts. Finally we conclude this paper, discuss remaining challenges and possible directions in this topic.

READ FULL TEXT
research
08/05/2022

Model Blending for Text Classification

Deep neural networks (DNNs) have proven successful in a wide variety of ...
research
05/08/2020

Pruning Algorithms to Accelerate Convolutional Neural Networks for Edge Applications: A Survey

With the general trend of increasing Convolutional Neural Network (CNN) ...
research
02/03/2018

A Survey on Acceleration of Deep Convolutional Neural Networks

Deep Neural Networks have achieved remarkable progress during the past f...
research
02/03/2018

Recent Advances in Efficient Computation of Deep Convolutional Neural Networks

Deep neural networks have evolved remarkably over the past few years and...
research
03/24/2020

A Survey of Methods for Low-Power Deep Learning and Computer Vision

Deep neural networks (DNNs) are successful in many computer vision tasks...
research
07/27/2020

ALF: Autoencoder-based Low-rank Filter-sharing for Efficient Convolutional Neural Networks

Closing the gap between the hardware requirements of state-of-the-art co...
research
11/21/2018

Graph-Adaptive Pruning for Efficient Inference of Convolutional Neural Networks

In this work, we propose a graph-adaptive pruning (GAP) method for effic...

Please sign up or login with your details

Forgot password? Click here to reset