Look-ups are not (yet) all you need for deep learning inference

07/12/2022
by   Calvin McCarter, et al.
0

Fast approximations to matrix multiplication have the potential to dramatically reduce the cost of neural network inference. Recent work on approximate matrix multiplication proposed to replace costly multiplications with table-lookups by fitting a fast hash function from training data. In this work, we propose improvements to this previous work, targeted to the deep learning inference setting, where one has access to both training data and fixed (already learned) model weight matrices. We further propose a fine-tuning procedure for accelerating entire neural networks while minimizing loss in accuracy. Finally, we analyze the proposed method on a simple image classification task. While we show improvements to prior work, overall classification accuracy remains substantially diminished compared to exact matrix multiplication. Our work, despite this negative result, points the way towards future efforts to accelerate inner products with fast nonlinear hashing methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/05/2018

Simple, Fast and Practicable Algorithms for Cholesky, LU and QR Decomposition Using Fast Rectangular Matrix Multiplication

This note presents fast Cholesky/LU/QR decomposition algorithms with O(n...
research
10/18/2022

Faster Matrix Multiplication via Asymmetric Hashing

Fast matrix multiplication is one of the most fundamental problems in al...
research
06/01/2023

Fast Matrix Multiplication Without Tears: A Constraint Programming Approach

It is known that the multiplication of an N × M matrix with an M × P mat...
research
06/25/2020

Constant-Depth and Subcubic-Size Threshold Circuits for Matrix Multiplication

Boolean circuits of McCulloch-Pitts threshold gates are a classic model ...
research
11/13/2015

Large Scale Artificial Neural Network Training Using Multi-GPUs

This paper describes a method for accelerating large scale Artificial Ne...
research
08/17/2020

Unitary Learning for Deep Diffractive Neural Network

Realization of deep learning with coherent diffraction has achieved rema...
research
12/14/2021

Training Multi-Layer Over-Parametrized Neural Network in Subquadratic Time

We consider the problem of training a multi-layer over-parametrized neur...

Please sign up or login with your details

Forgot password? Click here to reset