GCNear: A Hybrid Architecture for Efficient GCN Training with Near-Memory Processing

11/01/2021
by   Zhe Zhou, et al.
0

Recently, Graph Convolutional Networks (GCNs) have become state-of-the-art algorithms for analyzing non-euclidean graph data. However, it is challenging to realize efficient GCN training, especially on large graphs. The reasons are many-folded: 1) GCN training incurs a substantial memory footprint. Full-batch training on large graphs even requires hundreds to thousands of gigabytes of memory to buffer the intermediate data for back-propagation. 2) GCN training involves both memory-intensive data reduction and computation-intensive features/gradients update operations. Such a heterogeneous nature challenges current CPU/GPU platforms. 3) The irregularity of graphs and the complex training dataflow jointly increase the difficulty of improving a GCN training system's efficiency. This paper presents GCNear, a hybrid architecture to tackle these challenges. Specifically, GCNear adopts a DIMM-based memory system to provide easy-to-scale memory capacity. To match the heterogeneous nature, we categorize GCN training operations as memory-intensive Reduce and computation-intensive Update operations. We then offload Reduce operations to on-DIMM NMEs, making full use of the high aggregated local bandwidth. We adopt a CAE with sufficient computation capacity to process Update operations. We further propose several optimization strategies to deal with the irregularity of GCN tasks and improve GCNear's performance. We also propose a Multi-GCNear system to evaluate the scalability of GCNear.

READ FULL TEXT
research
12/31/2019

GraphACT: Accelerating GCN Training on CPU-FPGA Heterogeneous Platforms

Graph Convolutional Networks (GCNs) have emerged as the state-of-the-art...
research
10/17/2021

MG-GCN: Scalable Multi-GPU GCN Training Framework

Full batch training of Graph Convolutional Network (GCN) models is not f...
research
03/10/2021

Sampling methods for efficient training of graph convolutional networks: A survey

Graph Convolutional Networks (GCNs) have received significant attention ...
research
11/10/2021

SPA-GCN: Efficient and Flexible GCN Accelerator with an Application for Graph Similarity Computation

While there have been many studies on hardware acceleration for deep lea...
research
03/05/2022

Scaling R-GCN Training with Graph Summarization

Training of Relation Graph Convolutional Networks (R-GCN) does not scale...
research
12/09/2022

Scalable Graph Convolutional Network Training on Distributed-Memory Systems

Graph Convolutional Networks (GCNs) are extensively utilized for deep le...
research
04/06/2022

Accelerating Backward Aggregation in GCN Training with Execution Path Preparing on GPUs

The emerging Graph Convolutional Network (GCN) has now been widely used ...

Please sign up or login with your details

Forgot password? Click here to reset