BNS-GCN: Efficient Full-Graph Training of Graph Convolutional Networks with Partition-Parallelism and Random Boundary Node Sampling

03/21/2022
by   Cheng Wan, et al.
20

Graph Convolutional Networks (GCNs) have emerged as the state-of-the-art method for graph-based learning tasks. However, training GCNs at scale is still challenging, hindering both the exploration of more sophisticated GCN architectures and their applications to real-world large graphs. While it might be natural to consider graph partition and distributed training for tackling this challenge, this direction has only been slightly scratched the surface in the previous works due to the limitations of existing designs. In this work, we first analyze why distributed GCN training is ineffective and identify the underlying cause to be the excessive number of boundary nodes of each partitioned subgraph, which easily explodes the memory and communication costs for GCN training. Furthermore, we propose a simple yet effective method dubbed BNS-GCN that adopts random Boundary-Node-Sampling to enable efficient and scalable distributed GCN training. Experiments and ablation studies consistently validate the effectiveness of BNS-GCN, e.g., boosting the throughput by up to 16.2x and reducing the memory usage by up to 58 maintaining a full-graph accuracy. Furthermore, both theoretical and empirical analysis show that BNS-GCN enjoys a better convergence than existing sampling-based methods. We believe that our BNS-GCN has opened up a new paradigm for enabling GCN training at scale. The code is available at https://github.com/RICE-EIC/BNS-GCN.

READ FULL TEXT

page 3

page 4

page 5

research
03/20/2022

PipeGCN: Efficient Full-Graph Training of Graph Convolutional Networks with Pipelined Feature Communication

Graph Convolutional Networks (GCNs) is the state-of-the-art method for l...
research
03/30/2020

L^2-GCN: Layer-Wise and Learned Efficient Training of Graph Convolutional Networks

Graph convolution networks (GCN) are increasingly popular in many applic...
research
03/10/2021

Sampling methods for efficient training of graph convolutional networks: A survey

Graph Convolutional Networks (GCNs) have received significant attention ...
research
01/19/2021

Communication-Efficient Sampling for Distributed Training of Graph Convolutional Networks

Training Graph Convolutional Networks (GCNs) is expensive as it needs to...
research
06/25/2023

PolicyClusterGCN: Identifying Efficient Clusters for Training Graph Convolutional Networks

Graph convolutional networks (GCNs) have achieved huge success in severa...
research
05/27/2019

STAR-GCN: Stacked and Reconstructed Graph Convolutional Networks for Recommender Systems

We propose a new STAcked and Reconstructed Graph Convolutional Networks ...

Please sign up or login with your details

Forgot password? Click here to reset