Adaptive Neuron Apoptosis for Accelerating Deep Learning on Large Scale Systems

10/03/2016
by   Charles Siegel, et al.
0

We present novel techniques to accelerate the convergence of Deep Learning algorithms by conducting low overhead removal of redundant neurons -- apoptosis of neurons -- which do not contribute to model learning, during the training phase itself. We provide in-depth theoretical underpinnings of our heuristics (bounding accuracy loss and handling apoptosis of several neuron types), and present the methods to conduct adaptive neuron apoptosis. Specifically, we are able to improve the training time for several datasets by 2-3x, while reducing the number of parameters by up to 30x (4-5x on average) on datasets such as ImageNet classification. For the Higgs Boson dataset, our implementation improves the accuracy (measured by Area Under Curve (AUC)) for classification from 0.88/1 to 0.94/1, while reducing the number of parameters by 3x in comparison to existing literature. The proposed methods achieve a 2.44x speedup in comparison to the default (no apoptosis) algorithm.

READ FULL TEXT

page 4

page 8

research
07/10/2018

An Adaptive Learning Method of Restricted Boltzmann Machine by Neuron Generation and Annihilation Algorithm

Restricted Boltzmann Machine (RBM) is a generative stochastic energy-bas...
research
01/05/2022

Improving Spiking Neural Network Accuracy Using Time-based Neurons

Due to the fundamental limit to reducing power consumption of running de...
research
02/22/2023

Unsupervised 3D Object Learning through Neuron Activity aware Plasticity

We present an unsupervised deep learning model for 3D object classificat...
research
05/19/2016

AMSOM: Adaptive Moving Self-organizing Map for Clustering and Visualization

Self-Organizing Map (SOM) is a neural network model which is used to obt...
research
06/23/2020

NeuralScale: Efficient Scaling of Neurons for Resource-Constrained Deep Neural Networks

Deciding the amount of neurons during the design of a deep neural networ...
research
11/19/2016

Learning the Number of Neurons in Deep Networks

Nowadays, the number of layers and of neurons in each layer of a deep ne...
research
02/01/2019

Compressing Gradient Optimizers via Count-Sketches

Many popular first-order optimization methods (e.g., Momentum, AdaGrad, ...

Please sign up or login with your details

Forgot password? Click here to reset