FreezeOut: Accelerate Training by Progressively Freezing Layers

06/15/2017
by   Andrew Brock, et al.
0

The early layers of a deep neural net have the fewest parameters, but take up the most computation. In this extended abstract, we propose to only train the hidden layers for a set portion of the training run, freezing them out one-by-one and excluding them from the backward pass. Through experiments on CIFAR, we empirically demonstrate that FreezeOut yields savings of up to 20 wall-clock time during training with 3 speedup without loss of accuracy for ResNets, and no improvement for VGG networks. Our code is publicly available at https://github.com/ajbrock/FreezeOut

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/28/2022

DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training

A standard hardware bottleneck when training deep neural networks is GPU...
research
12/12/2018

CFUN: Combining Faster R-CNN and U-net Network for Efficient Whole Heart Segmentation

In this paper, we propose a novel heart segmentation pipeline Combining ...
research
06/27/2017

Training a Fully Convolutional Neural Network to Route Integrated Circuits

We present a deep, fully convolutional neural network that learns to rou...
research
02/28/2019

FastFusionNet: New State-of-the-Art for DAWNBench SQuAD

In this technical report, we introduce FastFusionNet, an efficient varia...
research
11/27/2020

Progressively Stacking 2.0: A Multi-stage Layerwise Training Method for BERT Training Speedup

Pre-trained language models, such as BERT, have achieved significant acc...
research
05/10/2022

Accelerating the Training of Video Super-Resolution Models

Despite that convolution neural networks (CNN) have recently demonstrate...
research
01/28/2020

f-BRS: Rethinking Backpropagating Refinement for Interactive Segmentation

Deep neural networks have become a mainstream approach to interactive se...

Please sign up or login with your details

Forgot password? Click here to reset