DeepAI AI Chat
Log In Sign Up

Hardware-oriented Approximation of Convolutional Neural Networks

04/11/2016
by   Philipp Gysel, et al.
University of California-Davis
0

High computational complexity hinders the widespread usage of Convolutional Neural Networks (CNNs), especially in mobile devices. Hardware accelerators are arguably the most promising approach for reducing both execution time and power consumption. One of the most important steps in accelerator development is hardware-oriented model approximation. In this paper we present Ristretto, a model approximation framework that analyzes a given CNN with respect to numerical resolution used in representing weights and outputs of convolutional and fully connected layers. Ristretto can condense models by using fixed point arithmetic and representation instead of floating point. Moreover, Ristretto fine-tunes the resulting fixed point network. Given a maximum error tolerance of 1 The code for Ristretto is available.

READ FULL TEXT

page 1

page 2

page 3

page 4

05/20/2016

Ristretto: Hardware-Oriented Approximation of Convolutional Neural Networks

Convolutional neural networks (CNN) have achieved major breakthroughs in...
04/27/2020

A scalable and efficient convolutional neural network accelerator using HLS for a System on Chip design

This paper presents a configurable Convolutional Neural Network Accelera...
05/10/2016

CORDIC-based Architecture for Powering Computation in Fixed-Point Arithmetic

We present a fixed point architecture (source VHDL code is provided) for...
08/29/2018

FPGA Implementation of Convolutional Neural Networks with Fixed-Point Calculations

Neural network-based methods for image processing are becoming widely us...
07/24/2018

Method for Hybrid Precision Convolutional Neural Network Representation

This invention addresses fixed-point representations of convolutional ne...
01/06/2020

Stochastic Rounding: Algorithms and Hardware Accelerator

Algorithms and a hardware accelerator for performing stochastic rounding...
01/30/2023

Self-Compressing Neural Networks

This work focuses on reducing neural network size, which is a major driv...