Hardware-Efficient Structure of the Accelerating Module for Implementation of Convolutional Neural Network Basic Operation

11/07/2018
by   Aleksandr Cariow, et al.
0

This paper presents a structural design of the hardware-efficient module for implementation of convolution neural network (CNN) basic operation with reduced implementation complexity. For this purpose we utilize some modification of the Winograd minimal filtering method as well as computation vectorization principles. This module calculate inner products of two consecutive segments of the original data sequence, formed by a sliding window of length 3, with the elements of a filter impulse response. The fully parallel structure of the module for calculating these two inner products, based on the implementation of a naive method of calculation, requires 6 binary multipliers and 4 binary adders. The use of the Winograd minimal filtering method allows to construct a module structure that requires only 4 binary multipliers and 8 binary adders. Since a high-performance convolutional neural network can contain tens or even hundreds of such modules, such a reduction can have a significant effect.

READ FULL TEXT
research
04/12/2020

Minimal Filtering Algorithms for Convolutional Neural Networks

In this paper, we present several resource-efficient algorithmic solutio...
research
03/18/2017

Hardware-Efficient Schemes of Quaternion Multiplying Units for 2D Discrete Quaternion Fourier Transform Processors

In this paper, we offer and discuss three efficient structural solutions...
research
11/17/2020

FPGA deep learning acceleration based on convolutional neural network

In view of the large amount of calculation and long calculation time of ...
research
07/06/2017

Pipelined Parallel FFT Architecture

In this paper, an optimized efficient VLSI architecture of a pipeline Fa...
research
05/06/2020

Comparing quaternary and binary multipliers

We compare the implementation of a 8x8 bit multiplier with two different...
research
09/18/2017

Minimal Effort Back Propagation for Convolutional Neural Networks

As traditional neural network consumes a significant amount of computing...
research
07/07/2023

BlendNet: Design and Optimization of a Neural Network-Based Inference Engine Blending Binary and Fixed-Point Convolutions

This paper presents BlendNet, a neural network architecture employing a ...

Please sign up or login with your details

Forgot password? Click here to reset