Sphynx: a parallel multi-GPU graph partitioner for distributed-memory systems

05/02/2021
by   Seher Acer, et al.
0

Graph partitioning has been an important tool to partition the work among several processors to minimize the communication cost and balance the workload. While accelerator-based supercomputers are emerging to be the standard, the use of graph partitioning becomes even more important as applications are rapidly moving to these architectures. However, there is no distributed-memory parallel, multi-GPU graph partitioner available for applications. We developed a spectral graph partitioner, Sphynx, using the portable, accelerator-friendly stack of the Trilinos framework. In Sphynx, we allow using different preconditioners and exploit their unique advantages. We use Sphynx to systematically evaluate the various algorithmic choices in spectral partitioning with a focus on the GPU performance. We perform those evaluations on two distinct classes of graphs: regular (such as meshes, matrices from finite element methods) and irregular (such as social networks and web graphs), and show that different settings and preconditioners are needed for these graph classes. The experimental results on the Summit supercomputer show that Sphynx is the fastest alternative on irregular graphs in an application-friendly setting and obtains a partitioning quality close to ParMETIS on regular graphs. When compared to nvGRAPH on a single GPU, Sphynx is faster and obtains better balance and better quality partitions. Sphynx provides a good and robust partitioning method across a wide range of graphs for applications looking for a GPU-based partitioner.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/03/2022

Clustering-based Partitioning for Large Web Graphs

Graph partitioning plays a vital role in distributedlarge-scale web grap...
research
04/25/2023

Jet: Multilevel Graph Partitioning on GPUs

The multilevel heuristic is the dominant strategy for high-quality seque...
research
12/09/2022

Scalable Graph Convolutional Network Training on Distributed-Memory Systems

Graph Convolutional Networks (GCNs) are extensively utilized for deep le...
research
08/16/2019

Distributed Edge Partitioning for Trillion-edge Graphs

We propose Distributed Neighbor Expansion (Distributed NE), a parallel a...
research
07/06/2021

An MPI-based Algorithm for Mapping Complex Networks onto Hierarchical Architectures

Processing massive application graphs on distributed memory systems requ...
research
04/25/2022

Parallel coarsening of graph data with spectral guarantees

Finding coarse representations of large graphs is an important computati...
research
12/29/2017

Signed Graph Analysis for the Interpretation of Voting Behavior

In a signed graph, each link is labeled with either a positive or a nega...

Please sign up or login with your details

Forgot password? Click here to reset