Addressing the Impact of Localized Training Data in Graph Neural Networks

07/24/2023
by   Singh Akansha, et al.
0

Graph Neural Networks (GNNs) have achieved notable success in learning from graph-structured data, owing to their ability to capture intricate dependencies and relationships between nodes. They excel in various applications, including semi-supervised node classification, link prediction, and graph generation. However, it is important to acknowledge that the majority of state-of-the-art GNN models are built upon the assumption of an in-distribution setting, which hinders their performance on real-world graphs with dynamic structures. In this article, we aim to assess the impact of training GNNs on localized subsets of the graph. Such restricted training data may lead to a model that performs well in the specific region it was trained on but fails to generalize and make accurate predictions for the entire graph. In the context of graph-based semi-supervised learning (SSL), resource constraints often lead to scenarios where the dataset is large, but only a portion of it can be labeled, affecting the model's performance. This limitation affects tasks like anomaly detection or spam detection when labeling processes are biased or influenced by human subjectivity. To tackle the challenges posed by localized training data, we approach the problem as an out-of-distribution (OOD) data issue by by aligning the distributions between the training data, which represents a small portion of labeled data, and the graph inference process that involves making predictions for the entire graph. We propose a regularization method to minimize distributional discrepancies between localized training data and graph inference, improving model performance on OOD data. Extensive tests on popular GNN models show significant performance improvement on three citation GNN benchmark datasets. The regularization approach effectively enhances model adaptation and generalization, overcoming challenges posed by OOD data.

READ FULL TEXT
research
08/02/2021

Shift-Robust GNNs: Overcoming the Limitations of Localized Graph Training data

There has been a recent surge of interest in designing Graph Neural Netw...
research
06/07/2020

Bayesian Graph Neural Networks with Adaptive Connection Sampling

We propose a unified framework for adaptive connection sampling in graph...
research
05/21/2022

KGNN: Harnessing Kernel-based Networks for Semi-supervised Graph Classification

This paper studies semi-supervised graph classification, which is an imp...
research
10/07/2021

Label Propagation across Graphs: Node Classification using Graph Neural Tangent Kernels

Graph neural networks (GNNs) have achieved superior performance on node ...
research
07/16/2022

SizeShiftReg: a Regularization Method for Improving Size-Generalization in Graph Neural Networks

In the past few years, graph neural networks (GNNs) have become the de f...
research
04/07/2023

Distributional Signals for Node Classification in Graph Neural Networks

In graph neural networks (GNNs), both node features and labels are examp...
research
05/20/2022

On the Prediction Instability of Graph Neural Networks

Instability of trained models, i.e., the dependence of individual node p...

Please sign up or login with your details

Forgot password? Click here to reset