Training Spiking Deep Networks for Neuromorphic Hardware

by   Eric Hunsberger, et al.
University of Waterloo

We describe a method to train spiking deep networks that can be run using leaky integrate-and-fire (LIF) neurons, achieving state-of-the-art results for spiking LIF networks on five datasets, including the large ImageNet ILSVRC-2012 benchmark. Our method for transforming deep artificial neural networks into spiking networks is scalable and works with a wide range of neural nonlinearities. We achieve these results by softening the neural response function, such that its derivative remains bounded, and by training the network with noise to provide robustness against the variability introduced by spikes. Our analysis shows that implementations of these networks on neuromorphic hardware will be many times more power-efficient than the equivalent non-spiking networks on traditional hardware.


page 1

page 2

page 3

page 4


Spiking Deep Networks with LIF Neurons

We train spiking deep networks using leaky integrate-and-fire (LIF) neur...

Hyperdimensional Computing with Spiking-Phasor Neurons

Vector Symbolic Architectures (VSAs) are a powerful framework for repres...

Deep Rewiring: Training very sparse deep networks

Neuromorphic hardware tends to pose limits on the connectivity of deep n...

Generative models on accelerated neuromorphic hardware

The traditional von Neumann computer architecture faces serious obstacle...

Dictionary Learning with Accumulator Neurons

The Locally Competitive Algorithm (LCA) uses local competition between n...

Deep Spiking-UNet for Image Processing

U-Net, known for its simple yet efficient architecture, is widely utiliz...

Exploiting Oxide Based Resistive RAM Variability for Bayesian Neural Network Hardware Design

Uncertainty plays a key role in real-time machine learning. As a signifi...

Please sign up or login with your details

Forgot password? Click here to reset