Deep Learning Approximation: Zero-Shot Neural Network Speedup

06/15/2018
by   Michele Pratusevich, et al.
0

Neural networks offer high-accuracy solutions to a range of problems, but are costly to run in production systems because of computational and memory requirements during a forward pass. Given a trained network, we propose a techique called Deep Learning Approximation to build a faster network in a tiny fraction of the time required for training by only manipulating the network structure and coefficients without requiring re-training or access to the training data. Speedup is achieved by by applying a sequential series of independent optimizations that reduce the floating-point operations (FLOPs) required to perform a forward pass. First, lossless optimizations are applied, followed by lossy approximations using singular value decomposition (SVD) and low-rank matrix decomposition. The optimal approximation is chosen by weighing the relative accuracy loss and FLOP reduction according to a single parameter specified by the user. On PASCAL VOC 2007 with the YOLO network, we show an end-to-end 2x speedup in a network forward pass with a 5 be re-gained by finetuning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/20/2020

Learning Low-rank Deep Neural Networks via Singular Vector Orthogonality Regularization and Singular Value Sparsification

Modern deep neural networks (DNNs) often require high memory consumption...
research
04/02/2018

Subspace-Orbit Randomized Decomposition for Low-rank Matrix Approximation

An efficient, accurate and reliable approximation of a matrix by one of ...
research
04/01/2019

Deep Industrial Espionage

The theory of deep learning is now considered largely solved, and is wel...
research
05/04/2021

Deterministic matrix sketches for low-rank compression of high-dimensional simulation data

Matrices arising in scientific applications frequently admit linear low-...
research
05/13/2019

Fast Proper Orthogonal Decomposition Using Improved Sampling and Iterative Techniques for Singular Value Decomposition

Proper Orthogonal Decomposition (POD), also known as Principal component...
research
10/15/2019

Training CNNs faster with Dynamic Input and Kernel Downsampling

We reduce training time in convolutional networks (CNNs) with a method t...
research
02/10/2022

Forward Build Systems, Formally

Build systems are a fundamental part of software construction, but their...

Please sign up or login with your details

Forgot password? Click here to reset