daBNN: A Super Fast Inference Framework for Binary Neural Networks on ARM devices

08/16/2019
by   Jianhao Zhang, et al.
0

It is always well believed that Binary Neural Networks (BNNs) could drastically accelerate the inference efficiency by replacing the arithmetic operations in float-valued Deep Neural Networks (DNNs) with bit-wise operations. Nevertheless, there has not been open-source implementation in support of this idea on low-end ARM devices (e.g., mobile phones and embedded devices). In this work, we propose daBNN — a super fast inference framework that implements BNNs on ARM devices. Several speed-up and memory refinement strategies for bit-packing, binarized convolution, and memory layout are uniquely devised to enhance inference efficiency. Compared to the recent open-source BNN inference framework, BMXNet, our daBNN is 7× ∼ 23× faster on a single binary convolution, and about 6× faster on Bi-Real Net 18 (a BNN variant of ResNet-18). The daBNN is a BSD-licensed inference framework, and its source code, sample projects and pre-trained models are available on-line: https://github.com/JDAI-CV/dabnn.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/27/2017

BMXNet: An Open-Source Binary Neural Network Implementation Based on MXNet

Binary Neural Networks (BNNs) can drastically reduce memory size and acc...
research
12/05/2019

PhoneBit: Efficient GPU-Accelerated Binary Neural Network Inference Engine for Mobile Phones

Over the last years, a great success of deep neural networks (DNNs) has ...
research
07/28/2020

Optimization of XNOR Convolution for Binary Convolutional Neural Networks on GPU

Binary convolutional networks have lower computational load and lower me...
research
04/03/2023

Optimizing data-flow in Binary Neural Networks

Binary Neural Networks (BNNs) can significantly accelerate the inference...
research
03/07/2023

Run, Don't Walk: Chasing Higher FLOPS for Faster Neural Networks

To design fast neural networks, many works have been focusing on reducin...
research
11/11/2020

FastPathology: An open-source platform for deep learning-based research and decision support in digital pathology

Deep convolutional neural networks (CNNs) are the current state-of-the-a...
research
12/08/2022

Fast Parallel Exact Inference on Bayesian Networks: Poster

Bayesian networks (BNs) are attractive, because they are graphical and i...

Please sign up or login with your details

Forgot password? Click here to reset