Tensorizing Neural Networks

09/22/2015
by   Alexander Novikov, et al.
0

Deep neural networks currently demonstrate state-of-the-art performance in several domains. At the same time, models of this class are very demanding in terms of computational resources. In particular, a large amount of memory is required by commonly used fully-connected layers, making it hard to use the models on low-end devices and stopping the further increase of the model size. In this paper we convert the dense weight matrices of the fully-connected layers to the Tensor Train format such that the number of parameters is reduced by a huge factor and at the same time the expressive power of the layer is preserved. In particular, for the Very Deep VGG networks we report the compression factor of the dense weight matrix of a fully-connected layer up to 200000 times leading to the compression factor of the whole network up to 7 times.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/21/2015

Compression of Fully-Connected Layer in Neural Network by Kronecker Product

In this paper we propose and study a technique to reduce the number of p...
research
03/14/2019

Tucker Tensor Layer in Fully Connected Neural Networks

We introduce the Tucker Tensor Layer (TTL), an alternative to the dense ...
research
12/27/2021

Learning Robust and Lightweight Model through Separable Structured Transformations

With the proliferation of mobile devices and the Internet of Things, dee...
research
06/05/2017

DeepIoT: Compressing Deep Neural Network Structures for Sensing Systems with a Compressor-Critic Framework

Recent advances in deep learning motivate the use of deep neutral networ...
research
05/25/2018

Tensorized Spectrum Preserving Compression for Neural Networks

Modern neural networks can have tens of millions of parameters, and are ...
research
10/28/2017

Trainable back-propagated functional transfer matrices

Connections between nodes of fully connected neural networks are usually...
research
01/29/2019

On the Expressive Power of Deep Fully Circulant Neural Networks

In this paper, we study deep fully circulant neural networks, that is de...

Please sign up or login with your details

Forgot password? Click here to reset