Energy-Efficient Time-Domain Vector-by-Matrix Multiplier for Neurocomputing and Beyond

11/29/2017
by   Mohammad Bavandpour, et al.
0

We propose an extremely energy-efficient mixed-signal approach for performing vector-by-matrix multiplication in a time domain. In such implementation, multi-bit values of the input and output vector elements are represented with time-encoded digital signals, while multi-bit matrix weights are realized with current sources, e.g. transistors biased in subthreshold regime. With our approach, multipliers can be chained together to implement large-scale circuits completely in a time domain. Multiplier operation does not rely on energy-taxing static currents, which are typical for peripheral and input/output conversion circuits of the conventional mixed-signal implementations. As a case study, we have designed a multilayer perceptron, based on two layers of 10x10 four-quadrant vector-by-matrix multipliers, in 55-nm process with embedded NOR flash memory technology, which allows for compact implementation of adjustable current sources. Our analysis, based on memory cell measurements, shows that at high computing speed the drain-induced barrier lowering is a major factor limiting multiplier precision to 6 bit. Post-layout estimates for a conservative 6-bit digital input/output NxN multiplier designed in 55 nm process, including I/O circuitry for converting between digital and time domain representations, show 7 fJ/Op for N>200, which can be further lowered well below 1 fJ/Op for more optimal and aggressive design.

READ FULL TEXT
research
08/07/2019

3D-aCortex: An Ultra-Compact Energy-Efficient Neurocomputing Platform Based on Commercial 3D-NAND Flash Memories

The first contribution of this paper is the development of extremely den...
research
01/26/2016

A network that learns Strassen multiplication

We study neural networks whose only non-linear components are multiplier...
research
06/23/2020

Inference with Artificial Neural Networks on the Analog BrainScaleS-2 Hardware

The neuromorphic BrainScaleS-2 ASIC comprises mixed-signal neurons and s...
research
06/23/2020

Inference with Artificial Neural Networks on Analog Neuromorphic Hardware

The neuromorphic BrainScaleS-2 ASIC comprises mixed-signal neurons and s...
research
07/16/2022

Spatz: A Compact Vector Processing Unit for High-Performance and Energy-Efficient Shared-L1 Clusters

While parallel architectures based on clusters of Processing Elements (P...
research
12/21/2020

A complete, parallel and autonomous photonic neural network in a semiconductor multimode laser

Neural networks are one of the disruptive computing concepts of our time...
research
08/20/2019

Comparing ternary and binary adders and multipliers

While many papers have proposed implementations of ternary adders and te...

Please sign up or login with your details

Forgot password? Click here to reset