Reducing the Computational Burden of Deep Learning with Recursive Local Representation Alignment

02/10/2020
by   Alexander Ororbia, et al.
1

Training deep neural networks on large-scale datasets requires significant hardware resources whose costs (even on cloud platforms) put them out of reach of smaller organizations, groups, and individuals. Backpropagation (backprop), the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize. Furthermore, it requires researchers to continually develop various tricks, such as specialized weight initializations and activation functions, in order to ensure a stable parameter optimization. Our goal is to seek an effective, parallelizable alternative to backprop that can be used to train deep networks. In this paper, we propose a gradient-free learning procedure, recursive local representation alignment, for training large-scale neural architectures. Experiments with deep residual networks on CIFAR-10 and the massive-scale benchmark, ImageNet, show that our algorithm generalizes as well as backprop while converging sooner due to weight updates that are parallelizable and computationally less demanding. This is empirical evidence that a backprop-free algorithm can scale up to larger datasets. Another contribution is that we also significantly reduce total parameter count of our networks by utilizing fast, fixed noise maps in place of convolutional operations without compromising generalization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/12/2020

RSO: A Gradient Free Sampling Based Approach For Training Deep Neural Networks

We propose RSO (random search optimization), a gradient free Markov Chai...
research
03/09/2021

Analytically Tractable Inference in Deep Neural Networks

Since its inception, deep learning has been overwhelmingly reliant on ba...
research
10/05/2014

Understanding Locally Competitive Networks

Recently proposed neural network activation functions such as rectified ...
research
06/04/2023

Random Feedback Alignment Algorithms to train Neural Networks: Why do they Align?

Feedback alignment algorithms are an alternative to backpropagation to t...
research
03/05/2023

Multiplexed gradient descent: Fast online training of modern datasets on hardware neural networks without backpropagation

We present multiplexed gradient descent (MGD), a gradient descent framew...
research
03/05/2018

Conducting Credit Assignment by Aligning Local Representations

The use of back-propagation and its variants to train deep networks is o...
research
07/07/2019

Towards Robust, Locally Linear Deep Networks

Deep networks realize complex mappings that are often understood by thei...

Please sign up or login with your details

Forgot password? Click here to reset