DeepAI AI Chat
Log In Sign Up

Learning Graph Neural Networks using Exact Compression

by   Jeroen Bollen, et al.

Graph Neural Networks (GNNs) are a form of deep learning that enable a wide range of machine learning applications on graph-structured data. The learning of GNNs, however, is known to pose challenges for memory-constrained devices such as GPUs. In this paper, we study exact compression as a way to reduce the memory requirements of learning GNNs on large graphs. In particular, we adopt a formal approach to compression and propose a methodology that transforms GNN learning problems into provably equivalent compressed GNN learning problems. In a preliminary experimental evaluation, we give insights into the compression ratios that can be obtained on real-world graphs and apply our methodology to an existing GNN benchmark.


page 1

page 2

page 3

page 4


The Logic of Graph Neural Networks

Graph neural networks (GNNs) are deep learning architectures for machine...

Triple Sparsification of Graph Convolutional Networks without Sacrificing the Accuracy

Graph Neural Networks (GNNs) are widely used to perform different machin...

Graph Neural Networks in Particle Physics: Implementations, Innovations, and Challenges

Many physical systems can be best understood as sets of discrete data wi...

GraphMDN: Leveraging graph structure and deep learning to solve inverse problems

The recent introduction of Graph Neural Networks (GNNs) and their growin...

Graph Neural Networks and 3-Dimensional Topology

We test the efficiency of applying Geometric Deep Learning to the proble...

RiskNet: Neural Risk Assessment in Networks of Unreliable Resources

We propose a graph neural network (GNN)-based method to predict the dist...

Pre-training Graph Neural Networks with Kernels

Many machine learning techniques have been proposed in the last few year...