Training Deep Architectures Without End-to-End Backpropagation: A Brief Survey

01/09/2021
by   Shiyu Duan, et al.
0

This tutorial paper surveys training alternatives to end-to-end backpropagation (E2EBP) – the de facto standard for training deep architectures. Modular training refers to strictly local training without both the forward and the backward pass, i.e., dividing a deep architecture into several nonoverlapping modules and training them separately without any end-to-end operation. Between the fully global E2EBP and the strictly local modular training, there are "weakly modular" hybrids performing training without the backward pass only. These alternatives can match or surpass the performance of E2EBP on challenging datasets such as ImageNet, and are gaining increased attention primarily because they offer practical advantages over E2EBP, which will be enumerated herein. In particular, they allow for greater modularity and transparency in deep learning workflows, aligning deep learning with the mainstream computer science engineering that heavily exploits modularization for scalability. Modular training has also revealed novel insights about learning and may have further implications on other important research domains. Specifically, it induces natural and effective solutions to some important practical problems such as data efficiency and transferability estimation.

READ FULL TEXT
research
04/26/2017

Limits of End-to-End Learning

End-to-end learning refers to training a possibly complex learning syste...
research
09/25/2015

Training Deep Networks with Structured Layers by Matrix Backpropagation

Deep neural network architectures have recently produced excellent resul...
research
05/12/2020

Modularizing Deep Learning via Pairwise Learning With Kernels

By redefining the conventional notions of layers, we present an alternat...
research
05/24/2023

Block-local learning with probabilistic latent representations

The ubiquitous backpropagation algorithm requires sequential updates acr...
research
01/31/2021

PyTorch-Hebbian: facilitating local learning in a deep learning framework

Recently, unsupervised local learning, based on Hebb's idea that change ...
research
03/16/2022

Playing with blocks: Toward re-usable deep learning models for side-channel profiled attacks

This paper introduces a deep learning modular network for side-channel a...
research
02/24/2022

Exploiting Problem Structure in Deep Declarative Networks: Two Case Studies

Deep declarative networks and other recent related works have shown how ...

Please sign up or login with your details

Forgot password? Click here to reset