Log In Sign Up

Computing low-rank approximations of large-scale matrices with the Tensor Network randomized SVD

by   Kim Batselier, et al.

We propose a new algorithm for the computation of a singular value decomposition (SVD) low-rank approximation of a matrix in the Matrix Product Operator (MPO) format, also called the Tensor Train Matrix format. Our tensor network randomized SVD (TNrSVD) algorithm is an MPO implementation of the randomized SVD algorithm that is able to compute dominant singular values and their corresponding singular vectors. In contrast to the state-of-the-art tensor-based alternating least squares SVD (ALS-SVD) and modified alternating least squares SVD (MALS-SVD) matrix approximation methods, TNrSVD can be up to 17 times faster while achieving the same accuracy. In addition, our TNrSVD algorithm also produces accurate approximations in particular cases where both ALS-SVD and MALS-SVD fail to converge. We also propose a new algorithm for the fast conversion of a sparse matrix into its corresponding MPO form, which is up to 509 times faster than the standard Tensor Train SVD (TT-SVD) method while achieving machine precision accuracy. The efficiency and accuracy of both algorithms are demonstrated in numerical experiments.


page 1

page 2

page 3

page 4


Randomized Quaternion Singular Value Decomposition for Low-Rank Approximation

Quaternion matrix approximation problems construct the approximated matr...

Faster Tensor Train Decomposition for Sparse Data

In recent years, the application of tensors has become more widespread i...

Performance of the low-rank tensor-train SVD (TT-SVD) for large dense tensors on modern multi-core CPUs

There are several factorizations of multi-dimensional tensors into lower...

A New High Performance and Scalable SVD algorithm on Distributed Memory Systems

This paper introduces a high performance implementation of Zolo-SVD algo...

A hybrid DEIM and leverage scores based method for CUR index selection

The discrete empirical interpolation method (DEIM) may be used as an ind...

Improved analysis of randomized SVD for top-eigenvector approximation

Computing the top eigenvectors of a matrix is a problem of fundamental i...

LazySVD: Even Faster SVD Decomposition Yet Without Agonizing Pain

We study k-SVD that is to obtain the first k singular vectors of a matri...