Nesting Forward Automatic Differentiation for Memory-Efficient Deep Neural Network Training

09/22/2022
by   Cong Guo, et al.
0

An activation function is an element-wise mathematical function and plays a crucial role in deep neural networks (DNN). Many novel and sophisticated activation functions have been proposed to improve the DNN accuracy but also consume massive memory in the training process with back-propagation. In this study, we propose the nested forward automatic differentiation (Forward-AD), specifically for the element-wise activation function for memory-efficient DNN training. We deploy nested Forward-AD in two widely-used deep learning frameworks, TensorFlow and PyTorch, which support the static and dynamic computation graph, respectively. Our evaluation shows that nested Forward-AD reduces the memory footprint by up to 1.97x than the baseline model and outperforms the recomputation by 20

READ FULL TEXT
research
10/03/2018

Weighted Sigmoid Gate Unit for an Activation Function of Deep Neural Network

An activation function has crucial role in a deep neural network. A si...
research
10/27/2018

A Methodology for Automatic Selection of Activation Functions to Design Hybrid Deep Neural Networks

Activation functions influence behavior and performance of DNNs. Nonline...
research
11/27/2019

Optimal checkpointing for heterogeneous chains: how to train deep neural networks with limited memory

This paper introduces a new activation checkpointing method which allows...
research
02/01/2022

Few-Bit Backward: Quantized Gradients of Activation Functions for Memory Footprint Reduction

Memory footprint is one of the main limiting factors for large neural ne...
research
11/09/2017

Feed Forward and Backward Run in Deep Convolution Neural Network

Convolution Neural Networks (CNN), known as ConvNets are widely used in ...
research
06/08/2018

Learning in Integer Latent Variable Models with Nested Automatic Differentiation

We develop nested automatic differentiation (AD) algorithms for exact in...
research
11/16/2022

Separable PINN: Mitigating the Curse of Dimensionality in Physics-Informed Neural Networks

Physics-informed neural networks (PINNs) have emerged as new data-driven...

Please sign up or login with your details

Forgot password? Click here to reset