Training CNNs faster with Dynamic Input and Kernel Downsampling

10/15/2019
by   Zissis Poulos, et al.
0

We reduce training time in convolutional networks (CNNs) with a method that, for some of the mini-batches: a) scales down the resolution of input images via downsampling, and b) reduces the forward pass operations via pooling on the convolution filters. Training is performed in an interleaved fashion; some batches undergo the regular forward and backpropagation passes with original network parameters, whereas others undergo a forward pass with pooled filters and downsampled inputs. Since pooling is differentiable, the gradients of the pooled filters propagate to the original network parameters for a standard parameter update. The latter phase requires fewer floating point operations and less storage due to the reduced spatial dimensions in feature maps and filters. The key idea is that this phase leads to smaller and approximate updates and thus slower learning, but at significantly reduced cost, followed by passes that use the original network parameters as a refinement stage. Deciding how often and for which batches the downsmapling occurs can be done either stochastically or deterministically, and can be defined as a training hyperparameter itself. Experiments on residual architectures show that we can achieve up to 23 accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/29/2019

LeanConvNets: Low-cost Yet Effective Convolutional Neural Networks

Convolutional Neural Networks (CNNs) have become indispensable for solvi...
research
03/06/2018

MIMO Graph Filters for Convolutional Neural Networks

Superior performance and ease of implementation have fostered the adopti...
research
08/31/2016

Pruning Filters for Efficient ConvNets

The success of CNNs in various applications is accompanied by a signific...
research
05/05/2023

Compressing audio CNNs with graph centrality based filter pruning

Convolutional neural networks (CNNs) are commonplace in high-performing ...
research
06/01/2018

Targeted Kernel Networks: Faster Convolutions with Attentive Regularization

We propose Attentive Regularization (AR), a method to constrain the acti...
research
06/15/2018

Deep Learning Approximation: Zero-Shot Neural Network Speedup

Neural networks offer high-accuracy solutions to a range of problems, bu...
research
07/20/2020

Sorted Pooling in Convolutional Networks for One-shot Learning

We present generalized versions of the commonly used maximum pooling ope...

Please sign up or login with your details

Forgot password? Click here to reset