AIDA: Associative DNN Inference Accelerator

12/20/2018
by   Leonid Yavits, et al.
0

We propose AIDA, an inference engine for accelerating fully-connected (FC) layers of Deep Neural Network (DNN). AIDA is an associative in-memory processor, where the bulk of data never leaves the confines of the memory arrays, and processing is performed in-situ. AIDA area and energy efficiency strongly benefit from sparsity and lower arithmetic precision. We show that AIDA outperforms the state of art inference accelerator, EIE, by 14.5x (peak performance) and 2.5x (throughput).

READ FULL TEXT
research
07/27/2017

Tartan: Accelerating Fully-Connected and Convolutional Layers in Deep Learning Networks by Exploiting Numerical Precision Variability

Tartan (TRT), a hardware accelerator for inference with Deep Neural Netw...
research
08/10/2022

A Fresh Perspective on DNN Accelerators by Performing Holistic Analysis Across Paradigms

Traditional computers with von Neumann architecture are unable to meet t...
research
11/21/2022

BrainTTA: A 35 fJ/op Compiler Programmable Mixed-Precision Transport-Triggered NN SoC

Recently, accelerators for extremely quantized deep neural network (DNN)...
research
09/26/2022

FastStamp: Accelerating Neural Steganography and Digital Watermarking of Images on FPGAs

Steganography and digital watermarking are the tasks of hiding recoverab...
research
02/14/2018

Field-Programmable Deep Neural Network (DNN) Learning and Inference accelerator: a concept

An accelerator is a specialized integrated circuit designed to perform s...
research
02/26/2020

Graphcore C2 Card performance for image-based deep learning application: A Report

Recently, Graphcore has introduced an IPU Processor for accelerating mac...
research
06/16/2016

A 0.3-2.6 TOPS/W Precision-Scalable Processor for Real-Time Large-Scale ConvNets

A low-power precision-scalable processor for ConvNets or convolutional n...

Please sign up or login with your details

Forgot password? Click here to reset