Graph-Adaptive Pruning for Efficient Inference of Convolutional Neural Networks

11/21/2018
by   Mengdi Wang, et al.
0

In this work, we propose a graph-adaptive pruning (GAP) method for efficient inference of convolutional neural networks (CNNs). In this method, the network is viewed as a computational graph, in which the vertices denote the computation nodes and edges represent the information flow. Through topology analysis, GAP is capable of adapting to different network structures, especially the widely used cross connections and multi-path data flow in recent novel convolutional models. The models can be adaptively pruned at vertex-level as well as edge-level without any post-processing, thus GAP can directly get practical model compression and inference speed-up. Moreover, it does not need any customized computation library or hardware support. Finetuning is conducted after pruning to restore the model performance. In the finetuning step, we adopt a self-taught knowledge distillation (KD) strategy by utilizing information from the original model, through which, the performance of the optimized model can be sufficiently improved, without introduction of any other teacher model. Experimental results show the proposed GAP can achieve promising result to make inference more efficient, e.g., for ResNeXt-29 on CIFAR10, it can get 13X model compression and 4.3X practical speed-up with marginal loss of accuracy.

READ FULL TEXT
research
10/26/2019

Cross-Channel Intragroup Sparsity Neural Network

Modern deep neural network models generally build upon heavy over-parame...
research
09/30/2021

Prune Your Model Before Distill It

Unstructured pruning reduces a significant amount of weights of neural n...
research
08/31/2021

AIP: Adversarial Iterative Pruning Based on Knowledge Transfer for Convolutional Neural Networks

With the increase of structure complexity, convolutional neural networks...
research
06/10/2018

Smallify: Learning Network Size while Training

As neural networks become widely deployed in different applications and ...
research
05/07/2022

Automatic Block-wise Pruning with Auxiliary Gating Structures for Deep Convolutional Neural Networks

Convolutional neural networks are prevailing in deep learning tasks. How...
research
10/23/2017

A Survey of Model Compression and Acceleration for Deep Neural Networks

Deep convolutional neural networks (CNNs) have recently achieved great s...
research
03/12/2018

It was the training data pruning too!

We study the current best model (KDG) for question answering on tabular ...

Please sign up or login with your details

Forgot password? Click here to reset