DeepAI AI Chat
Log In Sign Up

Theory of Graph Neural Networks: Representation and Learning

04/16/2022
by   Stefanie Jegelka, et al.
0

Graph Neural Networks (GNNs), neural network architectures targeted to learning representations of graphs, have become a popular learning model for prediction tasks on nodes, graphs and configurations of points, with wide success in practice. This article summarizes a selection of the emerging theoretical results on approximation and learning properties of widely used message passing GNNs and higher-order GNNs, focusing on representation, generalization and extrapolation. Along the way, it summarizes mathematical connections.

READ FULL TEXT

page 1

page 2

page 3

page 4

04/29/2021

The Logic of Graph Neural Networks

Graph neural networks (GNNs) are deep learning architectures for machine...
05/10/2021

Optimization of Graph Neural Networks: Implicit Acceleration by Skip Connections and More Depth

Graph Neural Networks (GNNs) have been studied through the lens of expre...
04/20/2022

Simplicial Attention Networks

Graph representation learning methods have mostly been limited to the mo...
02/24/2020

Neural Message Passing on High Order Paths

Graph neural network have achieved impressive results in predicting mole...
09/20/2022

Neural Graph Databases

Graph databases (GDBs) enable processing and analysis of unstructured, c...
05/05/2021

Scalable Graph Neural Network Training: The Case for Sampling

Graph Neural Networks (GNNs) are a new and increasingly popular family o...
12/26/2022

Statistical Mechanics of Generalization In Graph Convolution Networks

Graph neural networks (GNN) have become the default machine learning mod...