DeepAI AI Chat
Log In Sign Up

PipeGCN: Efficient Full-Graph Training of Graph Convolutional Networks with Pipelined Feature Communication

by   Cheng Wan, et al.

Graph Convolutional Networks (GCNs) is the state-of-the-art method for learning graph-structured data, and training large-scale GCNs requires distributed training across multiple accelerators such that each accelerator is able to hold a partitioned subgraph. However, distributed GCN training incurs prohibitive overhead of communicating node features and feature gradients among partitions for every GCN layer during each training iteration, limiting the achievable training efficiency and model scalability. To this end, we propose PipeGCN, a simple yet effective scheme that hides the communication overhead by pipelining inter-partition communication with intra-partition computation. It is non-trivial to pipeline for efficient GCN training, as communicated node features/gradients will become stale and thus can harm the convergence, negating the pipeline benefit. Notably, little is known regarding the convergence rate of GCN training with both stale features and stale feature gradients. This work not only provides a theoretical convergence analysis but also finds the convergence rate of PipeGCN to be close to that of the vanilla distributed GCN training without any staleness. Furthermore, we develop a smoothing method to further improve PipeGCN's convergence. Extensive experiments show that PipeGCN can largely boost the training throughput (1.7x 28.5x) while achieving the same accuracy as its vanilla counterpart and existing full-graph training methods. The code is available at


page 1

page 2

page 3

page 4


Simple and Deep Graph Convolutional Networks

Graph convolutional networks (GCNs) are a powerful deep learning approac...

Old can be Gold: Better Gradient Flow can Make Vanilla-GCNs Great Again

Despite the enormous success of Graph Convolutional Networks (GCNs) in m...

FedGCN: Convergence and Communication Tradeoffs in Federated Training of Graph Convolutional Networks

Distributed methods for training models on graph datasets have recently ...

Communication-Efficient Sampling for Distributed Training of Graph Convolutional Networks

Training Graph Convolutional Networks (GCNs) is expensive as it needs to...

PolicyClusterGCN: Identifying Efficient Clusters for Training Graph Convolutional Networks

Graph convolutional networks (GCNs) have achieved huge success in severa...