Compiling Neural Networks for a Computational Memory Accelerator

03/05/2020
by   Kornilios Kourtis, et al.
0

Computational memory (CM) is a promising approach for accelerating inference on neural networks (NN) by using enhanced memories that, in addition to storing data, allow computations on them. One of the main challenges of this approach is defining a hardware/software interface that allows a compiler to map NN models for efficient execution on the underlying CM accelerator. This is a non-trivial task because efficiency dictates that the CM accelerator is explicitly programmed as a dataflow engine where the execution of the different NN layers form a pipeline. In this paper, we present our work towards a software stack for executing ML models on such a multi-core CM accelerator. We describe an architecture for the hardware and software, and focus on the problem of implementing the appropriate control logic so that data dependencies are respected. We propose a solution to the latter that is based on polyhedral compilation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/23/2020

hxtorch: PyTorch for ANNs on BrainScaleS-2

We present software facilitating the usage of the BrainScaleS-2 analog n...
research
01/28/2019

FPSA: A Full System Stack Solution for Reconfigurable ReRAM-based NN Accelerator Architecture

Neural Network (NN) accelerators with emerging ReRAM (resistive random a...
research
12/12/2012

Accelerating Inference: towards a full Language, Compiler and Hardware stack

We introduce Dimple, a fully open-source API for probabilistic modeling....
research
03/01/2021

Mitigating Edge Machine Learning Inference Bottlenecks: An Empirical Study on Accelerating Google Edge Models

As the need for edge computing grows, many modern consumer devices now c...
research
02/03/2021

Quantum Accelerator Stack: A Research Roadmap

This paper presents the definition and implementation of a quantum compu...
research
04/10/2020

Entropy-Based Modeling for Estimating Soft Errors Impact on Binarized Neural Network Inference

Over past years, the easy accessibility to the large scale datasets has ...
research
07/09/2018

XNOR Neural Engine: a Hardware Accelerator IP for 21.6 fJ/op Binary Neural Network Inference

Binary Neural Networks (BNNs) are promising to deliver accuracy comparab...

Please sign up or login with your details

Forgot password? Click here to reset