Deep Learning in Memristive Nanowire Networks

03/03/2020
by   Jack D. Kendall, et al.
0

Analog crossbar architectures for accelerating neural network training and inference have made tremendous progress over the past several years. These architectures are ideal for dense layers with fewer than roughly a thousand neurons. However, for large sparse layers, crossbar architectures are highly inefficient. A new hardware architecture, dubbed the MN3 (Memristive Nanowire Neural Network), was recently described as an efficient architecture for simulating very wide, sparse neural network layers, on the order of millions of neurons per layer. The MN3 utilizes a high-density memristive nanowire mesh to efficiently connect large numbers of silicon neurons with modifiable weights. Here, in order to explore the MN3's ability to function as a deep neural network, we describe one algorithm for training deep MN3 models and benchmark simulations of the architecture on two deep learning tasks. We utilize a simple piecewise linear memristor model, since we seek to demonstrate that training is, in principle, possible for randomized nanowire architectures. In future work, we intend on utilizing more realistic memristor models, and we will adapt the presented algorithm appropriately. We show that the MN3 is capable of performing composition, gradient propagation, and weight updates, which together allow it to function as a deep neural network. We show that a simulated multilayer perceptron (MLP), built from MN3 networks, can obtain a 1.61 software-based network. This work represents, to the authors' knowledge, the first randomized nanowire architecture capable of reproducing the backpropagation algorithm.

READ FULL TEXT
research
08/17/2018

Optimizing Deep Neural Network Architecture: A Tabu Search Based Approach

The performance of Feedforward neural network (FNN) fully de-pends upon ...
research
01/30/2023

Equivariant Architectures for Learning in Deep Weight Spaces

Designing machine learning architectures for processing neural networks ...
research
05/31/2022

A comparative study of back propagation and its alternatives on multilayer perceptrons

The de facto algorithm for training the back pass of a feedforward neura...
research
06/23/2016

DropNeuron: Simplifying the Structure of Deep Neural Networks

Deep learning using multi-layer neural networks (NNs) architecture manif...
research
07/13/2018

Perceptrons from Memristors

Memristors, resistors with memory whose outputs depend on the history of...
research
04/08/2019

Nucleus Neural Network for Super Robust Learning

Artificial neural networks which model the neurons and connecting archit...
research
05/16/2019

Formal derivation of Mesh Neural Networks with their Forward-Only gradient Propagation

This paper proposes the Mesh Neural Network (MNN), a novel architecture ...

Please sign up or login with your details

Forgot password? Click here to reset