StrassenNets: Deep learning with a multiplication budget

12/11/2017
by   Michael Tschannen, et al.
0

A large fraction of the arithmetic operations required to evaluate deep neural networks (DNNs) are due to matrix multiplications, both in convolutional and fully connected layers. Matrix multiplications can be cast as 2-layer sum-product networks (SPNs) (arithmetic circuits), disentangling multiplications and additions. We leverage this observation for end-to-end learning of low-cost (in terms of multiplications) approximations of linear operations in DNN layers. Specifically, we propose to replace matrix multiplication operations by SPNs, with widths corresponding to the budget of multiplications we want to allocate to each layer, and learning the edges of the SPNs from data. Experiments on CIFAR-10 and ImageNet show that this method applied to ResNet yields significantly higher accuracy than existing methods for a given multiplication budget, or leads to the same or higher accuracy compared to existing methods while using significantly fewer multiplications. Furthermore, our approach allows fine-grained control of the tradeoff between arithmetic complexity and accuracy of DNN models. Finally, we demonstrate that the proposed framework is able to rediscover Strassen's matrix multiplication algorithm, i.e., it can learn to multiply 2 × 2 matrices using only 7 multiplications instead of 8.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2023

NeuralMatrix: Moving Entire Neural Networks to General Matrix Multiplication for Efficient Inference

In this study, we introduce NeuralMatrix, a novel framework that enables...
research
03/30/2022

Matrix Multiplication with Less Arithmetic Complexity and IO Complexity

After Strassen presented the first sub-cubic matrix multiplication algor...
research
11/06/2018

OverSketch: Approximate Matrix Multiplication for the Cloud

We propose OverSketch, an approximate algorithm for distributed matrix m...
research
12/18/2019

Pseudospectral Shattering, the Sign Function, and Diagonalization in Nearly Matrix Multiplication Time

We exhibit a randomized algorithm which given a square n× n complex matr...
research
08/18/2017

Improved Rectangular Matrix Multiplication using Powers of the Coppersmith-Winograd Tensor

In the past few years, successive improvements of the asymptotic complex...
research
07/20/2020

Effects of Approximate Multiplication on Convolutional Neural Networks

This paper analyzes the effects of approximate multiplication when perfo...
research
12/22/2022

EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models

With the advent of deep learning application on edge devices, researcher...

Please sign up or login with your details

Forgot password? Click here to reset