A Unified Lottery Ticket Hypothesis for Graph Neural Networks

02/12/2021
by   Tianlong Chen, et al.
0

With graphs rapidly growing in size and deeper graph neural networks (GNNs) emerging, the training and inference of GNNs become increasingly expensive. Existing network weight pruning algorithms cannot address the main space and computational bottleneck in GNNs, caused by the size and connectivity of the graph. To this end, this paper first presents a unified GNN sparsification (UGS) framework that simultaneously prunes the graph adjacency matrix and the model weights, for effectively accelerating GNN inference on large-scale graphs. Leveraging this new tool, we further generalize the recently popular lottery ticket hypothesis to GNNs for the first time, by defining a graph lottery ticket (GLT) as a pair of core sub-dataset and sparse sub-network, which can be jointly identified from the original GNN and the full dense graph by iteratively applying UGS. Like its counterpart in convolutional neural networks, GLT can be trained in isolation to match the performance of training with the full model and graph, and can be drawn from both randomly initialized and self-supervised pre-trained GNNs. Our proposal has been experimentally verified across various GNN architectures and diverse tasks, on both small-scale graph datasets (Cora, Citeseer and PubMed), and large-scale datasets from the challenging Open Graph Benchmark (OGB). Specifically, for node classification, our found GLTs achieve the same accuracies with 20 MACs saving on small graphs and 25 prediction, GLTs lead to 48 datasets, respectively, without compromising predictive performance. Codes available at https://github.com/VITA-Group/Unified-LTH-GNN.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/28/2022

You Can Have Better Graph Neural Networks by Not Training Weights at All: Finding Untrained GNNs Tickets

Recent works have impressively demonstrated that there exists a subnetwo...
research
05/03/2023

Rethinking Graph Lottery Tickets: Graph Sparsity Matters

Lottery Ticket Hypothesis (LTH) claims the existence of a winning ticket...
research
02/27/2023

IGB: Addressing The Gaps In Labeling, Features, Heterogeneity, and Size of Public Graph Datasets for Deep Learning Research

Graph neural networks (GNNs) have shown high potential for a variety of ...
research
06/18/2023

Graph Ladling: Shockingly Simple Parallel GNN Training without Intermediate Communication

Graphs are omnipresent and GNNs are a powerful family of neural networks...
research
04/06/2023

Graph Mixture of Experts: Learning on Large-Scale Graphs with Explicit Diversity Modeling

Graph neural networks (GNNs) have been widely applied to learning over g...
research
06/25/2020

Incremental Training of Graph Neural Networks on Temporal Graphs under Distribution Shift

Current graph neural networks (GNNs) are promising, especially when the ...
research
08/24/2021

Bag of Tricks for Training Deeper Graph Neural Networks: A Comprehensive Benchmark Study

Training deep graph neural networks (GNNs) is notoriously hard. Besides ...

Please sign up or login with your details

Forgot password? Click here to reset