TensorCodec: Compact Lossy Compression of Tensors without Strong Data Assumptions

09/19/2023
by   Taehyung Kwon, et al.
0

Many real-world datasets are represented as tensors, i.e., multi-dimensional arrays of numerical values. Storing them without compression often requires substantial space, which grows exponentially with the order. While many tensor compression algorithms are available, many of them rely on strong data assumptions regarding its order, sparsity, rank, and smoothness. In this work, we propose TENSORCODEC, a lossy compression algorithm for general tensors that do not necessarily adhere to strong input data assumptions. TENSORCODEC incorporates three key ideas. The first idea is Neural Tensor-Train Decomposition (NTTD) where we integrate a recurrent neural network into Tensor-Train Decomposition to enhance its expressive power and alleviate the limitations imposed by the low-rank assumption. Another idea is to fold the input tensor into a higher-order tensor to reduce the space required by NTTD. Finally, the mode indices of the input tensor are reordered to reveal patterns that can be exploited by NTTD for improved approximation. Our analysis and experiments on 8 real-world datasets demonstrate that TENSORCODEC is (a) Concise: it gives up to 7.38x more compact compression than the best competitor with similar reconstruction error, (b) Accurate: given the same budget for compressed size, it yields up to 3.33x more accurate reconstruction than the best competitor, (c) Scalable: its empirical compression time is linear in the number of tensor entries, and it reconstructs each entry in logarithmic time. Our code and datasets are available at https://github.com/kbrother/TensorCodec.

READ FULL TEXT
research
02/09/2023

NeuKron: Constant-Size Lossy Compression of Sparse Reorderable Matrices and Tensors

Many real-world data are naturally represented as a sparse reorderable m...
research
08/29/2019

Multi-resolution Low-rank Tensor Formats

We describe a simple, black-box compression format for tensors with a mu...
research
04/22/2022

Scalable symmetric Tucker tensor decomposition

We study the best low-rank Tucker decomposition of symmetric tensors, ad...
research
05/29/2021

Cherry-Picking Gradients: Learning Low-Rank Embeddings of Visual Data via Differentiable Cross-Approximation

We propose an end-to-end trainable framework that processes large-scale ...
research
02/18/2018

Inductive Framework for Multi-Aspect Streaming Tensor Completion with Side Information

Low-rank tensor completion is a well-studied problem and has application...
research
09/18/2023

Multi-Dictionary Tensor Decomposition

Tensor decomposition methods are popular tools for analysis of multi-way...
research
07/13/2020

T-Basis: a Compact Representation for Neural Networks

We introduce T-Basis, a novel concept for a compact representation of a ...

Please sign up or login with your details

Forgot password? Click here to reset