Reducing Communication in Algebraic Multigrid with Multi-step Node Aware Communication

04/11/2019
by   Amanda Bienz, et al.
0

Algebraic multigrid (AMG) is often viewed as a scalable 𝒪(n) solver for sparse linear systems. Yet, parallel AMG lacks scalability due to increasingly large costs associated with communication, both in the initial construction of a multigrid hierarchy as well as the iterative solve phase. This work introduces a parallel implementation of AMG to reduce the cost of communication, yielding an increase in scalability. Standard inter-process communication consists of sending data regardless of the send and receive process locations. Performance tests show notable differences in the cost of intra- and inter-node communication, motivating a restructuring of communication. In this case, the communication schedule takes advantage of the less costly intra-node communication, reducing both the number and size of inter-node messages. Node-centric communication extends to the range of components in both the setup and solve phase of AMG, yielding an increase in the weak and strong scalability of the entire method.

READ FULL TEXT

page 4

page 6

page 10

research
12/23/2016

Node Aware Sparse Matrix-Vector Multiplication

The sparse matrix-vector multiply (SpMV) operation is a key computationa...
research
08/26/2023

A Locality-Aware Sparse Dynamic Data Exchange

Parallel architectures are continually increasing in performance and sca...
research
02/14/2018

A co-located partitions strategy for parallel CFD-DEM couplings

In this work, a new partition-collocation strategy for the parallel exec...
research
06/25/2019

Parallel Performance of Algebraic Multigrid Domain Decomposition (AMG-DD)

Algebraic multigrid (AMG) is a widely used scalable solver and precondit...
research
12/19/2019

Scalable Resilience Against Node Failures for Communication-Hiding Preconditioned Conjugate Gradient and Conjugate Residual Methods

The observed and expected continued growth in the number of nodes in lar...
research
10/21/2019

Node-Aware Improvements to Allreduce

The MPI_Allreduce collective operation is a core kernel of many parallel...
research
01/11/2023

TAPS: Topology-Aware Intra-Operator Parallelism Strategy Searching Algorithm for Deep Neural Networks

TAPS is a Topology-Aware intra-operator Parallelism strategy Searching a...

Please sign up or login with your details

Forgot password? Click here to reset