A Massively Parallel Algorithm for the Approximate Calculation of Inverse p-th Roots of Large Sparse Matrices

10/30/2017
by   Michael Lass, et al.
0

We present the submatrix method, a highly parallelizable method for the approximate calculation of inverse p-th roots of large sparse symmetric matrices which are required in different scientific applications such as electronic structure methods. We follow the idea of Approximate Computing, allowing imprecision in the final result in order to be able to utilize the sparsity of the input matrix and to allow massively parallel execution. For an n x n matrix, the proposed algorithm allows to distribute the calculations over n nodes with only little communication overhead. The approximate result matrix exhibits the same sparsity pattern as the input matrix, allowing for efficient reuse of allocated data structures. We evaluate the algorithm with respect to the error that it introduces into calculated results, as well as its performance and scalability. We demonstrate that the error is relatively limited for well-conditioned matrices and discuss the execution time of the algorithm for increasing matrix sizes. Finally, we present a distributed implementation of the algorithm using MPI and OpenMP and demonstrate its scalability by running it on a high-performance compute cluster comprised of 1024 CPU cores.

READ FULL TEXT
research
06/19/2019

Sparse approximate matrix multiplication in a fully recursive distributed task-based parallel framework

In this paper we consider parallel implementations of approximate multip...
research
03/07/2017

Using Approximate Computing for the Calculation of Inverse Matrix p-th Roots

Approximate computing has shown to provide new ways to improve performan...
research
08/14/2017

PSelInv - A Distributed Memory Parallel Algorithm for Selected Inversion: the non-symmetric Case

This paper generalizes the parallel selected inversion algorithm called ...
research
07/02/2022

A Structured Sparse Neural Network and Its Matrix Calculations Algorithm

Gradient descent optimizations and backpropagation are the most common m...
research
05/16/2018

A Note on QR-Based Model Reduction: Algorithm, Software, and Gravitational Wave Applications

While the proper orthogonal decomposition (POD) is optimal under certain...
research
09/05/2022

Orthogonal layers of parallelism in large-scale eigenvalue computations

We address the communication overhead of distributed sparse matrix-(mult...
research
12/11/2019

DGEMM performance is data-dependent

The DGEMM function is a widely used implementation of the matrix product...

Please sign up or login with your details

Forgot password? Click here to reset