Compressing neural network by tensor network with exponentially fewer variational parameters

05/10/2023
by   Yong Qing, et al.
0

Neural network (NN) designed for challenging machine learning tasks is in general a highly nonlinear mapping that contains massive variational parameters. High complexity of NN, if unbounded or unconstrained, might unpredictably cause severe issues including over-fitting, loss of generalization power, and unbearable cost of hardware. In this work, we propose a general compression scheme that significantly reduces the variational parameters of NN by encoding them to multi-layer tensor networks (TN's) that contain exponentially-fewer free parameters. Superior compression performance of our scheme is demonstrated on several widely-recognized NN's (FC-2, LeNet-5, and VGG-16) and datasets (MNIST and CIFAR-10), surpassing the state-of-the-art method based on shallow tensor networks. For instance, about 10 million parameters in the three convolutional layers of VGG-16 are compressed in TN's with just 632 parameters, while the testing accuracy on CIFAR-10 is surprisingly improved from 81.14% by the original NN to 84.36% after compression. Our work suggests TN as an exceptionally efficient mathematical structure for representing the variational parameters of NN's, which superiorly exploits the compressibility than the simple multi-way arrays.

READ FULL TEXT
research
04/11/2019

Compressing deep neural networks by matrix product operators

A deep neural network is a parameterization of a multi-layer mapping of ...
research
12/22/2020

Residual Matrix Product State for Machine Learning

Tensor network (TN), which originates from quantum physics, shows broad ...
research
06/22/2022

GACT: Activation Compressed Training for General Architectures

Training large neural network (NN) models requires extensive memory reso...
research
02/15/2021

Neural Network Compression for Noisy Storage Devices

Compression and efficient storage of neural network (NN) parameters is c...
research
05/21/2018

Faster Neural Network Training with Approximate Tensor Operations

We propose a novel technique for faster Neural Network (NN) training by ...
research
10/02/2019

Towards Unifying Neural Architecture Space Exploration and Generalization

In this paper, we address a fundamental research question of significant...
research
03/31/2022

A unified theory of learning

Recently machine learning using neural networks (NN) has been developed,...

Please sign up or login with your details

Forgot password? Click here to reset