Faster Neural Network Training with Approximate Tensor Operations

05/21/2018
by   Menachem Adelman, et al.
0

We propose a novel technique for faster Neural Network (NN) training by systematically approximating all the constituent matrix multiplications and convolutions. This approach is complementary to other approximation techniques, requires no changes to the dimensions of the network layers, hence compatible with existing training frameworks. We first analyze the applicability of the existing methods for approximating matrix multiplication to NN training, and extend the most suitable column-row sampling algorithm to approximating multi-channel convolutions. We apply approximate tensor operations to training MLP, CNN and LSTM network architectures on MNIST, CIFAR-100 and Penn Tree Bank datasets and demonstrate 30 maintaining little or no impact on the test accuracy. Our promising results encourage further study of general methods for approximating tensor operations and their application to NN training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/13/2023

Provable Convergence of Tensor Decomposition-Based Neural Network Training

Advanced tensor decomposition, such as tensor train (TT), has been widel...
research
05/21/2018

AxTrain: Hardware-Oriented Neural Network Training for Approximate Inference

The intrinsic error tolerance of neural network (NN) makes approximate c...
research
05/10/2023

Compressing neural network by tensor network with exponentially fewer variational parameters

Neural network (NN) designed for challenging machine learning tasks is i...
research
10/29/2022

The isotropy group of the matrix multiplication tensor

By an isotropy group of a tensor t∈ V_1 ⊗ V_2⊗ V_3=V we mean the group o...
research
06/25/2021

Tensor-based framework for training flexible neural networks

Activation functions (AFs) are an important part of the design of neural...
research
06/22/2022

GACT: Activation Compressed Training for General Architectures

Training large neural network (NN) models requires extensive memory reso...
research
07/05/2023

Convolutions Through the Lens of Tensor Networks

Despite their simple intuition, convolutions are more tedious to analyze...

Please sign up or login with your details

Forgot password? Click here to reset