Diffusion-Jump GNNs: Homophiliation via Learnable Metric Filters

06/29/2023
by   Ahmed Begga, et al.
0

High-order Graph Neural Networks (HO-GNNs) have been developed to infer consistent latent spaces in the heterophilic regime, where the label distribution is not correlated with the graph structure. However, most of the existing HO-GNNs are hop-based, i.e., they rely on the powers of the transition matrix. As a result, these architectures are not fully reactive to the classification loss and the achieved structural filters have static supports. In other words, neither the filters' supports nor their coefficients can be learned with these networks. They are confined, instead, to learn combinations of filters. To address the above concerns, we propose Diffusion-jump GNNs a method relying on asymptotic diffusion distances that operates on jumps. A diffusion-pump generates pairwise distances whose projections determine both the support and coefficients of each structural filter. These filters are called jumps because they explore a wide range of scales in order to find bonds between scattered nodes with the same label. Actually, the full process is controlled by the classification loss. Both the jumps and the diffusion distances react to classification errors (i.e. they are learnable). Homophiliation, i.e., the process of learning piecewise smooth latent spaces in the heterophilic regime, is formulated as a Dirichlet problem: the known labels determine the border nodes and the diffusion-pump ensures a minimal deviation of the semi-supervised grouping from a canonical unsupervised grouping. This triggers the update of both the diffusion distances and, consequently, the jumps in order to minimize the classification error. The Dirichlet formulation has several advantages. It leads to the definition of structural heterophily, a novel measure beyond edge heterophily. It also allows us to investigate links with (learnable) diffusion distances, absorbing random walks and stochastic diffusion.

READ FULL TEXT
research
10/15/2021

Graph Neural Networks with Learnable Structural and Positional Representations

Graph neural networks (GNNs) have become the standard learning architect...
research
04/30/2022

Graph Anisotropic Diffusion

Traditional Graph Neural Networks (GNNs) rely on message passing, which ...
research
12/30/2020

Adaptive Graph Diffusion Networks with Hop-wise Attention

Graph Neural Networks (GNNs) have received much attention recent years a...
research
02/09/2022

Neural Sheaf Diffusion: A Topological Perspective on Heterophily and Oversmoothing in GNNs

Cellular sheaves equip graphs with "geometrical" structure by assigning ...
research
08/15/2022

Learnable Filters for Geometric Scattering Modules

We propose a new graph neural network (GNN) module, based on relaxations...
research
05/05/2020

Deep Lagrangian Constraint-based Propagation in Graph Neural Networks

Several real-world applications are characterized by data that exhibit a...

Please sign up or login with your details

Forgot password? Click here to reset