Structure-Aware DropEdge Towards Deep Graph Convolutional Networks

06/21/2023
by   Jiaqi Han, et al.
0

It has been discovered that Graph Convolutional Networks (GCNs) encounter a remarkable drop in performance when multiple layers are piled up. The main factor that accounts for why deep GCNs fail lies in over-smoothing, which isolates the network output from the input with the increase of network depth, weakening expressivity and trainability. In this paper, we start by investigating refined measures upon DropEdge – an existing simple yet effective technique to relieve over-smoothing. We term our method as DropEdge++ for its two structure-aware samplers in contrast to DropEdge: layer-dependent sampler and feature-dependent sampler. Regarding the layer-dependent sampler, we interestingly find that increasingly sampling edges from the bottom layer yields superior performance than the decreasing counterpart as well as DropEdge. We theoretically reveal this phenomenon with Mean-Edge-Number (MEN), a metric closely related to over-smoothing. For the feature-dependent sampler, we associate the edge sampling probability with the feature similarity of node pairs, and prove that it further correlates the convergence subspace of the output layer with the input features. Extensive experiments on several node classification benchmarks, including both full- and semi- supervised tasks, illustrate the efficacy of DropEdge++ and its compatibility with a variety of backbones by achieving generally better performance over DropEdge and the no-drop version.

READ FULL TEXT
research
08/22/2020

Tackling Over-Smoothing for General Graph Convolutional Networks

Increasing the depth of Graph Convolutional Networks (GCN), which in pri...
research
07/25/2019

The Truly Deep Graph Convolutional Networks for Node Classification

Existing Graph Convolutional Networks (GCNs) are shallow---the number of...
research
07/25/2019

DropEdge: Towards the Very Deep Graph Convolutional Networks for Node Classification

Existing Graph Convolutional Networks (GCNs) are shallow---the number of...
research
01/30/2022

Over-smoothing Effect of Graph Convolutional Networks

Over-smoothing is a severe problem which limits the depth of Graph Convo...
research
12/22/2021

SkipNode: On Alleviating Over-smoothing for Deep Graph Convolutional Networks

Over-smoothing is a challenging problem, which degrades the performance ...
research
09/14/2018

Adaptive Sampling Towards Fast Graph Representation Learning

Graph Convolutional Networks (GCNs) have become a crucial tool on learni...
research
03/30/2020

Revisiting "Over-smoothing" in Deep GCNs

Oversmoothing has been assumed to be the major cause of performance drop...

Please sign up or login with your details

Forgot password? Click here to reset