Augmentations in Graph Contrastive Learning: Current Methodological Flaws Towards Better Practices

11/05/2021
by   Puja Trivedi, et al.
0

Graph classification has applications in bioinformatics, social sciences, automated fake news detection, web document classification, and more. In many practical scenarios, including web-scale applications, where labels are scarce or hard to obtain, unsupervised learning is a natural paradigm but it trades off performance. Recently, contrastive learning (CL) has enabled unsupervised computer vision models to compete well against supervised ones. Theoretical and empirical works analyzing visual CL frameworks find that leveraging large datasets and domain aware augmentations is essential for framework success. Interestingly, graph CL frameworks often report high performance while using orders of magnitude smaller data, and employing domain-agnostic augmentations (e.g., node or edge dropping, feature perturbations) that can corrupt the graphs' underlying properties. Motivated by these discrepancies, we seek to determine: (i) why existing graph CL frameworks perform well despite weak augmentations and limited data; and (ii) whether adhering to visual CL principles can improve performance on graph classification tasks. Through extensive analysis, we identify flawed practices in graph data augmentation and evaluation protocols that are commonly used in the graph CL literature, and propose improved practices and sanity checks for future research and applications. We show that on small benchmark datasets, the inductive bias of graph neural networks can significantly compensate for the limitations of existing frameworks. In case studies with relatively larger graph classification tasks, we find that commonly used domain-agnostic augmentations perform poorly, while adhering to principles in visual CL can significantly improve performance. For example, in graph-based document classification, which can be used for better web search, we show task-relevant augmentations improve accuracy by 20

READ FULL TEXT

page 4

page 8

page 11

research
07/24/2023

MARIO: Model Agnostic Recipe for Improving OOD Generalization of Graph Contrastive Learning

In this work, we investigate the problem of out-of-distribution (OOD) ge...
research
02/21/2022

Model-Agnostic Augmentation for Accurate Graph Classification

Given a graph dataset, how can we augment it for accurate graph classifi...
research
08/04/2023

From Fake to Hyperpartisan News Detection Using Domain Adaptation

Unsupervised Domain Adaptation (UDA) is a popular technique that aims to...
research
03/07/2022

Graph Neural Networks for Image Classification and Reinforcement Learning using Graph representations

In this paper, we will evaluate the performance of graph neural networks...
research
12/19/2020

A pipeline for fair comparison of graph neural networks in node classification tasks

Graph neural networks (GNNs) have been investigated for potential applic...
research
06/09/2021

Fairness-Aware Node Representation Learning

Node representation learning has demonstrated its effectiveness for vari...
research
10/25/2021

Generating artificial texts as substitution or complement of training data

The quality of artificially generated texts has considerably improved wi...

Please sign up or login with your details

Forgot password? Click here to reset