BlockSwap: Fisher-guided Block Substitution for Network Compression

06/10/2019
by   Jack Turner, et al.
2

The desire to run neural networks on low-capacity edge devices has led to the development of a wealth of compression techniques. Moonshine is a simple and powerful example of this: one takes a large pre-trained network and substitutes each of its convolutional blocks with a selected cheap alternative block, then distills the resultant network with the original. However, not all blocks are created equally; for a required parameter budget there may exist a potent combination of many different cheap blocks. In this work, we find these by developing BlockSwap: an algorithm for choosing networks with interleaved block types by passing a single minibatch of training data through randomly initialised networks and gauging their Fisher potential. We show that block-wise cheapening yields more accurate networks than single block-type networks across a spectrum of parameter budgets. Code is available at https://github.com/BayesWatch/pytorch-blockswap.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/31/2021

1×N Block Pattern for Network Sparsity

Though network sparsity emerges as a promising direction to overcome the...
research
04/15/2021

AsymmNet: Towards ultralight convolution neural networks using asymmetrical bottlenecks

Deep convolutional neural networks (CNN) have achieved astonishing resul...
research
05/20/2018

Abstractive Text Classification Using Sequence-to-convolution Neural Networks

We propose a new deep neural network model and its training scheme for t...
research
09/08/2019

A Generalized Configuration Model with Degree Correlations and Its Percolation Analysis

In this paper we present a generalization of the classical configuration...
research
05/13/2021

HINet: Half Instance Normalization Network for Image Restoration

In this paper, we explore the role of Instance Normalization in low-leve...
research
01/06/2023

Codepod: A Namespace-Aware, Hierarchical Jupyter for Interactive Development at Scale

Jupyter is a browser-based interactive development environment that has ...
research
08/26/2021

Scalable and Modular Robustness Analysis of Deep Neural Networks

As neural networks are trained to be deeper and larger, the scalability ...

Please sign up or login with your details

Forgot password? Click here to reset