Survey and Benchmarking of Machine Learning Accelerators

08/29/2019
by   Albert Reuther, et al.
0

Advances in multicore processors and accelerators have opened the flood gates to greater exploration and application of machine learning techniques to a variety of applications. These advances, along with breakdowns of several trends including Moore's Law, have prompted an explosion of processors and accelerators that promise even greater computational and machine learning capabilities. These processors and accelerators are coming in many forms, from CPUs and GPUs to ASICs, FPGAs, and dataflow accelerators. This paper surveys the current state of these processors and accelerators that have been publicly announced with performance and power consumption numbers. The performance and power values are plotted on a scatter graph and a number of dimensions and observations from the trends on this plot are discussed and analyzed. For instance, there are interesting trends in the plot regarding power consumption, numerical precision, and inference versus training. We then select and benchmark two commercially-available low size, weight, and power (SWaP) accelerators as these processors are the most interesting for embedded and mobile machine learning inference applications that are most applicable to the DoD and other SWaP constrained users. We determine how they actually perform with real-world images and neural network models, compare those results to the reported performance and power consumption values and evaluate them against an Intel CPU that is used in some embedded applications.

READ FULL TEXT
research
09/01/2020

Survey of Machine Learning Accelerators

New machine learning accelerators are being announced and released each ...
research
10/08/2022

AI and ML Accelerator Survey and Trends

This paper updates the survey of AI accelerators and processors from pas...
research
09/18/2021

AI Accelerator Survey and Trends

Over the past several years, new machine learning accelerators were bein...
research
07/07/2021

R2F: A Remote Retraining Framework for AIoT Processors with Computing Errors

AIoT processors fabricated with newer technology nodes suffer rising sof...
research
09/15/2019

Performance and Power Evaluation of AI Accelerators for Training Deep Learning Models

Deep neural networks (DNNs) have become widely used in many AI applicati...
research
12/04/2021

On the Implementation of Fixed-point Exponential Function for Machine Learning and Signal Processing Accelerators

The natural exponential function is widely used in modeling many enginee...
research
08/14/2019

Side-Channel Aware Fuzzing

Software testing is becoming a critical part of the development cycle of...

Please sign up or login with your details

Forgot password? Click here to reset