Graphcore C2 Card performance for image-based deep learning application: A Report

02/26/2020
by   Ilyes Kacher, et al.
0

Recently, Graphcore has introduced an IPU Processor for accelerating machine learning applications. The architecture of the processor has been designed to achieve state of the art performance on current machine intelligence models for both training and inference. In this paper, we report on a benchmark in which we have evaluated the performance of IPU processors on deep neural networks for inference. We focus on deep vision models such as ResNeXt. We report the observed latency, throughput and energy efficiency.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/20/2018

AIDA: Associative DNN Inference Accelerator

We propose AIDA, an inference engine for accelerating fully-connected (F...
research
05/28/2019

Energy Efficiency Features of the Intel Skylake-SP Processor and Their Impact on Performance

The overwhelming majority of High Performance Computing (HPC) systems an...
research
05/06/2023

Energy-Latency Attacks to On-Device Neural Networks via Sponge Poisoning

In recent years, on-device deep learning has gained attention as a means...
research
03/23/2020

Performance Evaluation of Low-Cost Machine Vision Cameras for Image-Based Grasp Verification

Grasp verification is advantageous for autonomous manipulation robots as...
research
04/05/2017

Best Practices for Applying Deep Learning to Novel Applications

This report is targeted to groups who are subject matter experts in thei...
research
06/16/2016

A 0.3-2.6 TOPS/W Precision-Scalable Processor for Real-Time Large-Scale ConvNets

A low-power precision-scalable processor for ConvNets or convolutional n...
research
05/25/2022

Ultra-compact Binary Neural Networks for Human Activity Recognition on RISC-V Processors

Human Activity Recognition (HAR) is a relevant inference task in many mo...

Please sign up or login with your details

Forgot password? Click here to reset