MemSE: Fast MSE Prediction for Noisy Memristor-Based DNN Accelerators

05/03/2022
by   Jonathan Kern, et al.
0

Memristors enable the computation of matrix-vector multiplications (MVM) in memory and, therefore, show great potential in highly increasing the energy efficiency of deep neural network (DNN) inference accelerators. However, computations in memristors suffer from hardware non-idealities and are subject to different sources of noise that may negatively impact system performance. In this work, we theoretically analyze the mean squared error of DNNs that use memristor crossbars to compute MVM. We take into account both the quantization noise, due to the necessity of reducing the DNN model size, and the programming noise, stemming from the variability during the programming of the memristance value. Simulations on pre-trained DNN models showcase the accuracy of the analytical prediction. Furthermore the proposed method is almost two order of magnitude faster than Monte-Carlo simulation, thus making it possible to optimize the implementation parameters to achieve minimal error for a given power constraint.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/18/2021

Control Variate Approximation for DNN Accelerators

In this work, we introduce a control variate approximation technique for...
research
11/18/2022

Sharpness-Aware Training for Accurate Inference on Noisy DNN Accelerators

Energy-efficient deep neural network (DNN) accelerators are prone to non...
research
02/25/2020

TxSim:Modeling Training of Deep Neural Networks on Resistive Crossbar Systems

Resistive crossbars have attracted significant interest in the design of...
research
07/14/2023

Investigation of Deep Learning-Based Filtered Density Function for Large Eddy Simulation of Turbulent Scalar Mixing

The present investigation focuses on the application of deep neural netw...
research
07/29/2023

Improving Realistic Worst-Case Performance of NVCiM DNN Accelerators through Training with Right-Censored Gaussian Noise

Compute-in-Memory (CiM), built upon non-volatile memory (NVM) devices, i...
research
05/07/2022

Impact of L1 Batch Normalization on Analog Noise Resistant Property of Deep Learning Models

Analog hardware has become a popular choice for machine learning on reso...
research
04/13/2023

PowerGAN: A Machine Learning Approach for Power Side-Channel Attack on Compute-in-Memory Accelerators

Analog compute-in-memory (CIM) accelerators are becoming increasingly po...

Please sign up or login with your details

Forgot password? Click here to reset