Deep Active Graph Representation Learning

10/30/2020
by   Yanqiao Zhu, et al.
28

Graph neural networks (GNNs) aim to learn graph representations that preserve both attributive and structural information. In this paper, we study the problem of how to select high-quality nodes for training GNNs, considering GNNs are sensitive to different training datasets. Active learning (AL), whose purpose is to find the most informative instances to maximize the performance of the model, is a promising approach to solve this problem. Previous attempts have combined AL with graph representation learning by designing several selection criteria to measure how informative a node is. However, these methods do not directly utilize both the rich semantic and structural information and are prone to select sparsely-connected nodes (i.e. nodes having few neighbors) and low-purity nodes (i.e. nodes having noisy inter-class edges), which are less effective for training GNN models. To address these problems, we present a Deep Active Graph Representation Learning framework (DAGRL), in which three novel selection criteria are proposed. Specifically, we propose to measure the uncertainty of nodes via random topological perturbation. Besides, we propose two novel representativeness sampling criteria, which utilize both the structural and label information to find densely-connected nodes with many intra-class edges, hence enhance the performance of GNN models significantly. Then, we combine these three criteria with time-sensitive scheduling in accordance to the training progress of GNNs. Furthermore, considering the different size of classes, we employ a novel cluster-aware node selection policy, which ensures the number of selected nodes in each class is proportional to the size of the class. Comprehensive experiments on three public datasets show that our method outperforms previous baselines by a significant margin, which demonstrates its effectiveness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/13/2023

SAILOR: Structural Augmentation Based Tail Node Representation Learning

Graph Neural Networks (GNNs) have achieved state-of-the-art performance ...
research
06/09/2022

ScatterSample: Diversified Label Sampling for Data Efficient Graph Neural Network Learning

What target labels are most effective for graph neural network (GNN) tra...
research
11/20/2022

Enhancing Intra-class Information Extraction for Heterophilous Graphs: One Neural Architecture Search Approach

In recent years, Graph Neural Networks (GNNs) have been popular in graph...
research
08/17/2023

Mitigating Semantic Confusion from Hostile Neighborhood for Graph Active Learning

Graph Active Learning (GAL), which aims to find the most informative nod...
research
06/13/2023

Finding the Missing-half: Graph Complementary Learning for Homophily-prone and Heterophily-prone Graphs

Real-world graphs generally have only one kind of tendency in their conn...
research
03/02/2022

Information Gain Propagation: a new way to Graph Active Learning with Soft Labels

Graph Neural Networks (GNNs) have achieved great success in various task...
research
02/24/2023

GraphSR: A Data Augmentation Algorithm for Imbalanced Node Classification

Graph neural networks (GNNs) have achieved great success in node classif...

Please sign up or login with your details

Forgot password? Click here to reset