HIR: An MLIR-based Intermediate Representation for Hardware Accelerator Description

02/27/2021
by   Kingshuk Majumder, et al.
0

The emergence of machine learning, image and audio processing on edge devices has motivated research towards power efficient custom hardware accelerators. Though FPGAs are an ideal target for energy efficient custom accelerators, the difficulty of hardware design and the lack of vendor agnostic, standardized hardware compilation infrastructure has hindered their adoption. This paper introduces HIR, an MLIR-based intermediate representation (IR) to describe hardware accelerator designs. HIR combines high level language features, such as loops and multi-dimensional tensors, with programmer defined explicit scheduling, to provide a high-level IR suitable for DSL compiler pipelines without compromising control over the micro-architecture of the accelerator. HIR's explicit schedules allow it to express fine-grained, synchronization-free parallelism and optimizations such as retiming and pipelining. Built as a dialect in MLIR, it draws from best IR practices learnt from communities like those of LLVM. While offering rich optimization opportunities and a high level abstraction, HIR enables sharing of optimizations, utilities and passes with software compiler infrastructure. Our implementation shows that the code generation time of the HIR code generator is on average 1112x lower than that of Xilinx Vivado HLS on a range of kernels without a compromise on the quality of the generated hardware. We believe that these are significant steps forward in the design of IRs for hardware synthesis and in equipping domain-specific languages with a productive and performing compilation path to custom hardware acceleration.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/19/2021

A Compiler Infrastructure for Accelerator Generators

We present Calyx, a new intermediate language (IL) for compiling high-le...
research
08/04/2023

Automatic multi-dimensional pipelining for high-level synthesis of dataflow accelerators

In recent years, there has been a surging demand for edge computing of i...
research
05/19/2021

Pure Tensor Program Rewriting via Access Patterns (Representation Pearl)

Tensor kernels in machine learning (ML) often correspond to pure mathema...
research
09/19/2022

Temporal Vectorization: A Compiler Approach to Automatic Multi-Pumping

The multi-pumping resource sharing technique can overcome the limitation...
research
02/09/2021

NOELLE Offers Empowering LLVM Extensions

Modern and emerging architectures demand increasingly complex compiler a...
research
10/12/2018

ISA Mapper: A Compute and Hardware Agnostic Deep Learning Compiler

Domain specific accelerators present new challenges and opportunities fo...
research
05/28/2022

TinyIREE: An ML Execution Environment for Embedded Systems from Compilation to Deployment

Machine learning model deployment for training and execution has been an...

Please sign up or login with your details

Forgot password? Click here to reset