Feed-Forward On-Edge Fine-tuning Using Static Synthetic Gradient Modules

09/21/2020
by   Robby Neven, et al.
0

Training deep learning models on embedded devices is typically avoided since this requires more memory, computation and power over inference. In this work, we focus on lowering the amount of memory needed for storing all activations, which are required during the backward pass to compute the gradients. Instead, during the forward pass, static Synthetic Gradient Modules (SGMs) predict gradients for each layer. This allows training the model in a feed-forward manner without having to store all activations. We tested our method on a robot grasping scenario where a robot needs to learn to grasp new objects given only a single demonstration. By first training the SGMs in a meta-learning manner on a set of common objects, during fine-tuning, the SGMs provided the model with accurate gradients to successfully learn to grasp new objects. We have shown that our method has comparable results to using standard backpropagation.

READ FULL TEXT

page 8

page 12

research
08/18/2016

Decoupled Neural Interfaces using Synthetic Gradients

Training directed neural networks typically requires forward-propagating...
research
06/30/2022

FL-Tuning: Layer Tuning for Feed-Forward Network in Transformer

Prompt tuning is an emerging way of adapting pre-trained language models...
research
01/31/2022

Memory-Efficient Backpropagation through Large Linear Layers

In modern neural networks like Transformers, linear layers require signi...
research
05/05/2021

Perceptual Gradient Networks

Many applications of deep learning for image generation use perceptual l...
research
03/01/2017

Understanding Synthetic Gradients and Decoupled Neural Interfaces

When training neural networks, the use of Synthetic Gradients (SG) allow...
research
12/22/2017

Benchmarking Decoupled Neural Interfaces with Synthetic Gradients

Artifical Neural Network are a particular class of learning system model...
research
06/18/2020

Accelerating Training in Artificial Neural Networks with Dynamic Mode Decomposition

Training of deep neural networks (DNNs) frequently involves optimizing s...

Please sign up or login with your details

Forgot password? Click here to reset