Minimum Cost Loop Nests for Contraction of a Sparse Tensor with a Tensor Network

07/11/2023
by   Raghavendra Kanakagiri, et al.
0

Sparse tensor decomposition and completion are common in numerous applications, ranging from machine learning to computational quantum chemistry. Typically, the main bottleneck in optimization of these models are contractions of a single large sparse tensor with a network of several dense matrices or tensors (SpTTN). Prior works on high-performance tensor decomposition and completion have focused on performance and scalability optimizations for specific SpTTN kernels. We present algorithms and a runtime system for identifying and executing the most efficient loop nest for any SpTTN kernel. We consider both enumeration of such loop nests for autotuning and efficient algorithms for finding the lowest cost loop-nest for simpler metrics, such as buffer size or cache miss models. Our runtime system identifies the best choice of loop nest without user guidance, and also provides a distributed-memory parallelization of SpTTN kernels. We evaluate our framework using both real-world and synthetic tensors. Our results demonstrate that our approach outperforms available generalized state-of-the-art libraries and matches the performance of specialized codes.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/14/2018

Parallel Sparse Tensor Decomposition in Chapel

In big-data analytics, using tensor decomposition to extract patterns fr...
research
01/02/2020

A Parallel Sparse Tensor Benchmark Suite on CPUs and GPUs

Tensor computations present significant performance challenges that impa...
research
11/29/2021

An Asymptotic Cost Model for Autoscheduling Sparse Tensor Programs

While loop reordering and fusion can make big impacts on the constant-fa...
research
02/28/2018

Sparse Tensor Algebra Optimizations with Workspaces

This paper shows how to optimize sparse tensor algebraic expressions by ...
research
10/06/2019

Enabling Distributed-Memory Tensor Completion in Python using New Sparse Tensor Kernels

Tensor computations are increasingly prevalent numerical techniques in d...
research
07/10/2020

Distributed-Memory DMRG via Sparse and Dense Parallel Tensor Contractions

The Density Matrix Renormalization Group (DMRG) algorithm is a powerful ...
research
02/04/2018

Out-of-Core and Distributed Algorithms for Dense Subtensor Mining

How can we detect fraudulent lockstep behavior in large-scale multi-aspe...

Please sign up or login with your details

Forgot password? Click here to reset