A JIT Compiler for Neural Network Inference

06/13/2019
by   Felix Thielke, et al.
0

This paper describes a C++ library that compiles neural network models at runtime into machine code that performs inference. This approach in general promises to achieve the best performance possible since it is able to integrate statically known properties of the network directly into the code. In our experiments on the NAO V6 platform, it outperforms existing implementations significantly on small networks, while being inferior on large networks. The library was already part of the B-Human code release 2018, but has been extended since and is now available as a standalone version that can be integrated into any C++14 code base.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/14/2020

A C Code Generator for Fast Inference and Simple Deployment of Convolutional Neural Networks on Resource Constrained Systems

Inference of Convolutional Neural Networks in time critical applications...
research
09/09/2023

A Further Study of Linux Kernel Hugepages on A64FX with FLASH, an Astrophysical Simulation Code

We present an expanded study of the performance of FLASH when using Linu...
research
04/11/2018

DLL: A Blazing Fast Deep Neural Network Library

Deep Learning Library (DLL) is a new library for machine learning with d...
research
02/26/2022

libRoadRunner 2.0: A High-Performance SBML Simulation and Analysis Library

Motivation: This paper presents libRoadRunner 2.0, an extensible, high-p...
research
06/27/2012

Utilizing Static Analysis and Code Generation to Accelerate Neural Networks

As datasets continue to grow, neural network (NN) applications are becom...
research
03/19/2018

A promise checked is a promise kept: Inspection Testing

Occasionally, developers need to ensure that the compiler treats their c...

Please sign up or login with your details

Forgot password? Click here to reset