Synthesizing Power and Area Efficient Image Processing Pipelines on FPGAs using Customized Bit-widths

03/06/2018
by   Vinamra Benara, et al.
0

High-level synthesis (HLS) has received significant attention in recent years, improving programmability for FPGAs. PolyMage is a domain-specific language (DSL) for image processing pipelines that also has a HLS backend to translate the input DSL into an equivalent circuit that can be synthesized on FPGAs, while leveraging an HLS suite. The data at each stage of a pipeline is stored using a fixed-point data type (α,β) where α and β denote the number of integral and fractional bits. The power and area savings while performing arithmetic operations on fixed-point data type is known to be significant over using floating point. In this paper, we first propose an interval-arithmetic based range analysis (α-analysis) algorithm to estimate the number of bits required to store the integral part of the data at each stage of an image processing pipeline. The analysis algorithm uses the homogeneity of pixel signals at each stage to cluster them and perform a combined range analysis. Secondly, we propose a novel compilation framework in which any range analysis algorithm, be it interval or affine arithmetic, can be deployed with ease. Thirdly, for estimating fractional bit requirements (β-analysis), we propose a simple and practical heuristic search algorithm, which makes very few profile passes, as against techniques such as simulated annealing proposed in prior work. The analysis algorithm attempts to minimize the number of fractional bits required at each stage while respecting an application specific quality metric. We evaluated our bit-width analysis algorithms on four image processing benchmarks with regard to quality, power and area metrics. We show that with negligible loss in quality, we obtain up to a factor of 3.8x and 6.2x improvements in power and area respectively, when compared with using floating-point data type.

READ FULL TEXT
research
04/04/2018

Training DNNs with Hybrid Block Floating Point

The wide adoption of DNNs has given birth to unrelenting computing requi...
research
12/08/2022

Customizing Number Representation and Precision

There is a growing interest in the use of reduced-precision arithmetic, ...
research
06/09/2022

AritPIM: High-Throughput In-Memory Arithmetic

Digital processing-in-memory (PIM) architectures are rapidly emerging to...
research
11/19/2022

Accuracy Boosters: Epoch-Driven Mixed-Mantissa Block Floating-Point for DNN Training

The unprecedented growth in DNN model complexity, size and the amount of...
research
12/04/2017

An 826 MOPS, 210 uW/MHz Unum ALU in 65 nm

To overcome the limitations of conventional floating-point number format...
research
06/03/2018

Deploying Customized Data Representation and Approximate Computing in Machine Learning Applications

Major advancements in building general-purpose and customized hardware h...
research
01/03/2020

Low-cost Stochastic Number Generators for Stochastic Computing

Stochastic unary computing provides low-area circuits. However, the requ...

Please sign up or login with your details

Forgot password? Click here to reset