Tensor train decompositions on recurrent networks

06/09/2020
by   Alejandro Murua, et al.
0

Recurrent neural networks (RNN) such as long-short-term memory (LSTM) networks are essential in a multitude of daily live tasks such as speech, language, video, and multimodal learning. The shift from cloud to edge computation intensifies the need to contain the growth of RNN parameters. Current research on RNN shows that despite the performance obtained on convolutional neural networks (CNN), keeping a good performance in compressed RNNs is still a challenge. Most of the literature on compression focuses on CNNs using matrix product (MPO) operator tensor trains. However, matrix product state (MPS) tensor trains have more attractive features than MPOs, in terms of storage reduction and computing time at inference. We show that MPS tensor trains should be at the forefront of LSTM network compression through a theoretical analysis and practical experiments on NLP task.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/07/2017

Gated Recurrent Neural Tensor Network

Recurrent Neural Networks (RNNs), which are a powerful scheme for modeli...
research
08/21/2020

Kronecker CP Decomposition with Fast Multiplication for Compressing RNNs

Recurrent neural networks (RNNs) are powerful in the tasks oriented to s...
research
12/02/2016

Parameter Compression of Recurrent Neural Networks and Degradation of Short-term Memory

The significant computational costs of deploying neural networks in larg...
research
05/03/2022

Biometric Signature Verification Using Recurrent Neural Networks

Architectures based on Recurrent Neural Networks (RNNs) have been succes...
research
06/10/2020

Entanglement-Embedded Recurrent Network Architecture: Tensorized Latent State Propagation and Chaos Forecasting

Chaotic time series forecasting has been far less understood despite its...
research
05/11/2021

Tensor-Train Recurrent Neural Networks for Interpretable Multi-Way Financial Forecasting

Recurrent Neural Networks (RNNs) represent the de facto standard machine...
research
09/22/2022

A Case Report On The "A.I. Locked-In Problem": social concerns with modern NLP

Modern NLP models are becoming better conversational agents than their p...

Please sign up or login with your details

Forgot password? Click here to reset