Pytorch Code for the Lagrangian Propagation Graph Neural Network
In many real world applications, data are characterized by a complex structure, that can be naturally encoded as a graph. In the last years, the popularity of deep learning techniques has renewed the interest in neural models able to process complex patterns. In particular, inspired by the Graph Neural Network (GNN) model, different architectures have been proposed to extend the original GNN scheme. GNNs exploit a set of state variables, each assigned to a graph node, and a diffusion mechanism of the states among neighbor nodes, to implement an iterative procedure to compute the fixed point of the (learnable) state transition function. In this paper, we propose a novel approach to the state computation and the learning algorithm for GNNs, based on a constraint optimisation task solved in the Lagrangian framework. The state convergence procedure is implicitly expressed by the constraint satisfaction mechanism and does not require a separate iterative phase for each epoch of the learning procedure. In fact, the computational structure is based on the search for saddle points of the Lagrangian in the adjoint space composed of weights, neural outputs (node states), and Lagrange multipliers. The proposed approach is compared experimentally with other popular models for processing graphs.READ FULL TEXT VIEW PDF
Several real-world applications are characterized by data that exhibit a...
In this paper we study a constraint-based representation of neural netwo...
Semi-supervised node classification in graphs is a fundamental problem i...
The first provably efficient algorithm for learning graph neural network...
Graph Neural Networks (GNNs) require a relatively large number of labele...
Graph Drawing techniques have been developed in the last few years with ...
Graph neural networks (GNNs) are naturally distributed architectures for...
Pytorch Code for the Lagrangian Propagation Graph Neural Network
Due to their flexibility and approximation capabilities, the original processing and learning schemata of Neural Networks have been extended in order to deal with structured inputs. Based on the original feedforward model, able to process vectors of features as inputs, different architectures have been proposed to process sequences (Recurrent Neural Networks[Williams1989ALA]
), rasters of pixels (Convolutional Neural Networks[LeCun:1998:CNI:303568.303704]), directed acyclic graphs (Recursive Neural Networks [Goller96, Frasconi:1998:GFA:2325763.2326281]), and general graph structures (Graph Neural Networks [DBLP:journals/tnn/ScarselliGTHM09]
). All these models generally share the same learning mechanism based on the error BackPropagation (BP) through the network architecture, that allows the computation of the loss gradient with respect to the connection weights. When processing structured data the original BP schema is straightforwardly extended by the process ofunfolding that generates a network topology based on the current input structure by replicating a base neural network module (e.g. BP Through Time, BP Through Structure).
However, recently, some works [carreira2014distributed]
proposed a different approach to learning neural networks, where neural computations are expressed as constraints and the optimization is framed into the Lagrangian framework. These algorithms are naturally local and allow the learning of any computational structure, both acyclical or cyclical. The main drawback of these methods is that they are quite memory inefficient; in particular, they need to keep an extra-variable for each hidden neuron and for each example. This makes them inapplicable to large problems where BP is still the only viable option.
Graph Neural Networks (GNNs) [DBLP:journals/tnn/ScarselliGTHM09] exploit neural networks to learn how to encode nodes of a graph for a given task taking into account both the information local to each node and the whole graph topology. The learning process requires, for each epoch, an iterative diffusion mechanism up to convergence to a stable fixed point, that is computationally heavy. A maximum number of iterations can be defined but this limits the local encoding to a maximum depth of the neighborhood of each node. In this paper, we propose a new learning mechanism for GNNs based on a Lagrangian formulation that allows the embedding of the fixed point computation into the problem constraints. In the proposed scheme the network state representations and the weights are jointly optimized without the need of applying the fixed point relaxation procedure at each weight update epoch.
The paper is organized as follows. The next section reviews the main developments in both the Neural Network models for processing graphs and learning methods based on the Lagrangian approach. Section 3 introduces the basics of the GNN model, whereas in Section 4 the Lagrangian formulation of GNNs is described. Section 5 reports the experimental evaluation of the proposed constraint–based learning for GNNs. Finally, the conclusions are drawn in Section 6.
In many applications data are characterized by an underlying structure that lays on a non-Euclidean domain, i.e. graphs and manifolds. Whilst commonly addressed in relational learning, such domains have been initially not taken into account by popular machine learning techniques, that have been mostly devised for grid–like and Euclidean structured data[DBLP:journals/spm/BronsteinBLSV17]. Early machine learning approaches for structured data were designed for directed acyclic graphs [Sperduti:1997:SNN:2325755.2326105, Frasconi:1998:GFA:2325763.2326281], while a more general framework was introduced in [DBLP:journals/tnn/ScarselliGTHM09]. GNNs are able to directly deal with directed, undirected and cyclic graphs. The core idea is based on an iterative scheme of information diffusion among neighboring nodes, involving a propagation process aimed at reaching an equilibrium of the node states that represent a local encoding of the graph for a given task. The encoding is a computationally expensive process being based on the computation of the fixed point of the state transition function. Some proposals were aimed at simplifying this step, such as the scheme proposed in [DBLP:journals/corr/LiTBZ15]
that exploits gated recurrent units.
Recent approaches differ in the choice of neighborhood aggregation method and graph level pooling scheme, and can be categorized in two main areas. Spectral approaches exploit particular embeddings of the graph and the convolution operation defined in the spectral domain [DBLP:journals/corr/BrunaZSL13]. However, they are characterized by computational drawbacks caused by the eigen–decomposition of the graph Laplacian. Simplified approaches are based on smooth reparametrization [DBLP:journals/corr/HenaffBL15] or approximation of the spectral filters by a Chebyshev expansion [DBLP:conf/nips/DefferrardBV16]. Finally, in Graph Convolutional Networks (GCNs) [DBLP:conf/iclr/KipfW17], filters are restricted to operate in a 1-hop neighborhood of each node. Spatial methods, instead, exploit directly the graph topology, without the need of an intermediate representation. These approaches differ mainly in the definition of the aggregation operator used to compute the node states, that must be able to maintain weight sharing properties and to process nodes with different numbers of neighbors. The PATCHY-SAN [DBLP:conf/icml/NiepertAK16] model converts graph-structured data into a grid-structured representation, extracting and normalizing neighborhoods containing a fixed number of nodes. In [DBLP:conf/nips/DuvenaudMABHAA15] the model exploits a weight matrix for each node degree, whereas DCNNs [DBLP:conf/nips/AtwoodT16]
compute the hidden node representation convolving inputs channels with power series of the transition probability matrix, learning weights for each neighborhood degree. GraphSAGE[hamilton2017inductive] exploits different aggregation functions to merge the node neighborhood information. Deep GNNs [Bianchini2018] stack layers of GNNs to obtain a deep architecture. In the context of graph classification tasks, SortPooling [DBLP:conf/aaai/ZhangCNC18] uses a framework based on DGCNNs with a pooling strategy, that performs pooling by ordering vertices. Finally, the representational and discriminative power of GNN models were explored in [DBLP:journals/corr/abs-1810-00826], also introducing the novel GIN model.
A Lagrangian formulation of learning can be found in the seminal work of Yann LeCun [lecun1988theoretical], which studies a theoretical framework for Backpropagation. More recently, Carreira and Wang [carreira2014distributed] introduced the idea of training networks, transformed into a constraints-based representation, though an extension of the learnable parameters space. Their optimization scheme was based on quadratic penalties, aiming at an approximate solution of the problem afterwards refined by a post-processing phase. Differently, [taylor2016training] exploits closed-form solutions were most of the architectural constraints are softly enforced, and further additional variables are introduced to parametrize the neuron activations.
By framing the optimization of neural networks in the Lagrangian framework, where neural computations are expressed as constraints, their main goal is to obtain a local algorithm where computations of different layers can be carried out in parallel. On the contrary in the proposed approach, we use a novel mixed strategy. In particular, the majority of the computations still rely on Backpropagation while constraints are exploited only to express the diffusion mechanism. This allows to carry out both the optimization of the neural functions and the diffusion process at the same time, instead of alternating them into two distinct phases (as in [DBLP:journals/tnn/ScarselliGTHM09]), with a theoretical framework supporting this approach (Lagrangian optimization).
It has already been shown that algorithms on graphs can be effectively learned exploiting a constrained fixed-point formulation. For example, SSE [dai2018learning]
exploits the Reinforcement Learningpolicy iteration algorithm for the interleaved evaluation of the fixed point equation and the improvement of the transition and output functions. Our approach, starting from similar assumptions, exploits the unifying Lagrangian framework for learning both the transition and the output functions. Thus, by framing the optimization algorithm into a standard gradient descent/ascent scheme, we are allowed to use recent update rules (e.g. Adam) without the need to resort to ad-hoc moving average updates.
The term Graph Neural Network (GNN) refers to a general computational model, that exploits the processing and learning schemes of neural networks to process non Euclidean data, i.e. data organized as graphs.
Given an input graph , where is a finite set of nodes and collects the arcs, GNNs apply a two-phase computation on . In the encoding (or aggregation) phase the model computes a state vector for each node in by (iteratively) combining the states of neighboring nodes (i.e. nodes that are connected by an arc ). In the second phase, usually referred to as output (or readout), the latent representations encoded by the states stored in each node are exploited to compute the model output. The GNN can implement either a node-focused function, where an output is produced for each node of the input graph, or a graph-focused function, where the representations of all the nodes are aggregated to yield a single output for the whole input graph.
The GNN is defined by a pair of (learnable) functions, that respectively implement the state transition function required in the encoding phase and the output function exploited in the output phase, as follows:
where is the state of the node at iteration , is the set of the parents of node in , are the children of in , are the neighbors of the node in , is the feature vector available for node , and is the feature vector available for the arc 111With abuse of notation, we denote the set by . Similar definitions apply for , , and .. The vectors and collect the model parameters (the neural network weights) to be adapted during the learning procedure. Equations (2a) and (2b) are the two variants of the output function for node-focused or graph-focused tasks, respectively.
|Method: Function||Reference||Implementation of|
|GNN: Sum||Scarselli et al. [DBLP:journals/tnn/ScarselliGTHM09]|
|GIN: Sum||Xu et al. [DBLP:journals/corr/abs-1810-00826]||)|
|GCN: Mean||Kipf and Welling [DBLP:conf/iclr/KipfW17]||))|
|GraphSAGE: Max||Hamilton et al. [hamilton2017inductive]|
In Table 1, we show some possible choices of the function . It should be noted that this function may depend on a variable number of inputs, given that the nodes may have different degrees . Moreover, in general, the proposed implementations are invariant with respect to permutations of the nodes in , unless some predefined ordering is given for the neighbors of each node.
is the number of iterations of the state transition function applied before computing the output. The recursive application of the state transition function on the graph nodes yields a diffusion mechanism, whose range depends on . In fact, by stacking times the aggregation of 1-hop neighborhoods by , information of one node can be transferred to the nodes that are distant at most -hops. The number may be seen as the depth of the GNN and thus each iteration can be considered a different layer of the GNN. A sufficient number of layers is key to achieve a useful encoding of the input graph for the task at hand and, hence, the choice is problem–specific.
In the original GNN model [DBLP:journals/tnn/ScarselliGTHM09], eq. (1) is executed until convergence of the state representation, i.e. until . This scheme corresponds to the computation of the fixed point of the state transition function on the input graph. In order to guarantee the convergence of this phase, the transition function is required to be a contraction map.
Henceforth, for compactness, we denote the state transition function, applied to a node , with:
Basically, the encoding phase, through the iteration of , finds a solution to the fixed point problem defined by the constraint
In this case, the states encode the information contained in the whole graph. This diffusion mechanism is more general than executing only a fixed number of iterations (i.e. stacking a fixed number of layers). However, it can be computationally heavy and, hence, many recent GNN architectures apply only a fixed number of iterations for all nodes.
Neural network learning can be cast as a Lagrangian optimization problem by a formulation that requires the minimization of the classical data fitting loss (and eventually a regularization term) and the satisfaction of a set of architectural constraints that describe the computation performed on the data. Given this formulation, the solution can be computed by finding the saddle points of the associated Lagrangian in the space defined by the original network parameters and the Lagrange multipliers. The constraints can be exploited to enforce the computational structure that characterizes the GNN models.
The computation of Graph Neural Networks is driven by the input graph topology that defines the constraints among the computed state variables . In particular, the fixed point computation aims to solving eq. (4), that imposes a constraint between the node states and the way they are computed by the state transition function.
In the original GNN learning algorithm, the computation of the fixed point is required at each epoch of the learning procedure, as implemented by the iterative application of the transition function. Moreover, also the gradient computation requires us to take into account the relaxation procedure, by a backpropagation schema through the replicas of the state transition network exploited during the iterations for the fixed point computation. This procedure may be time consuming when the number of iterations for convergence to the fixed point is high (for instance in the case of large graphs).
We consider a Lagrangian formulation of the problem by adding free variables corresponding to the node states , such that the fixed point is directly defined by the constraints themselves, as
where is a function characterized by , such that the satisfaction of the constraints implies the solution of eq. (4). Apart from classical choices, like or , we can design different function shape (see Section 5.1), with desired properties. For instance, a possible implementations is , where is a parameter that can be used to allow tolerance in the satisfaction of the constraint. The hard formulation of the problem requires , but by setting to a small positive value it is possible to obtain a better generalization and tolerance to noise.
In the following, for simplicity, we will refer to a node-focused task, such that for some (or all) nodes of the input graph
, a target outputis provided as a supervision222For the sake of simplicity we consider only the case when a single graph is provided for learning. The extension for more graphs is straightforward for node-focused tasks, since they can be considered as a single graph composed by the given graphs as disconnected components.. If
is the loss function used to measure the target fitting approximation for node, the formulation of the learning task is:
where and are the weights of the MLPs implementing the state transition function and the output function, respectively, and is the set of the introduced free state variables.
This problem statement implicitly includes the definition of the fixed point of the state transition function in the optimal solution, since for any solution the constraints are satisfied and hence the computed optimal are solutions of eq. (4). As shown in the previous subsection, the constrained optimization problem of eq. (6) can be faced in the Lagrangian framework by introducing for each constraint a Lagrange multiplier , to define the Lagrangian function as:
where is the set of the Lagrangian multipliers. Finally, we can define the unconstrained optimization problem as the search for saddle points in the adjoint space as:
that can be solved by gradient descent with respect to the variables and gradient ascent with respect to the Lagrange multipliers , exploiting the Basic Differential Multiplier Method, introduced in [Platt] in the context of neural networks. We are interested in having a strong enforcement of the diffusion constraints, and common penalty-based methods are hard to tune and not always guaranteed to converge to the constraint satisfaction. BDMM could be seen as a simplified procedure that implements the principles behind the common Multiplier Methods, in order to enforce the hard fulfilment of the given constraints.
The gradient can be computed locally to each node, given the local variables and those of the neighboring nodes. In fact, the derivatives of the Lagrangian 333When parameters are vectors, the reported gradients should be considered element-wise. with respect to the considered parameters are:
where, , is its first derivative444The derivative is computed with respect to the same argument as in the partial derivative on the left side., , is its first derivative, and is its first derivative, and, finally, is the first derivative of . Being and implemented by feedforward neural networks, their derivatives are obtained easily by applying a classical backpropagation scheme, in order to optimize the Lagrangian function in the descent-ascent scheme, aiming at the saddle point, following [Platt].
We initialize the variables in and to zero, while the neural weights are randomly chosen. In particular, this differential optimization process consists of a gradient-descent step to update , and a gradient-ascent step to update , until we converge to the desired stationary point. Hence, the redefined differential equation system gradually fulfills the constraints, undergoing oscillations along the constraint subspace. To ease this procedure, we add the function, with the purpose of obtaining a more stable learning process.
Even if the proposed formulation adds the free state variables and the Lagrange multipliers , , there is no significant increase in the memory requirements since the state variables are also required in the original formulation and there is just a Lagrange multiplier for each node.
The diffusion mechanism of the state computation is enforced by means of the constraints. The learning algorithm is based on a mixed strategy where (i) Backpropagation is used to efficiently update the weights of the neural networks that implement the state transition and output functions, and, (ii) the diffusion mechanism evolves gradually by enforcing the convergence of the state transition function to a fixed point by virtue of the constraints. This last point is a novel approach in training Graph Neural Networks. In fact, in classical approaches, the encoding phase (see Section 3) is completely executed during the forward pass to compute the node states and, only after this phase is completed, the backward step is applied to update the weights of and . In the proposed scheme, both the neural network weights and the node state variables are simultaneously updated, forcing the state representation function towards a fixed point of in order to satisfy the constraints. In other words, the learning proceeds by jointly updating the function weights and by diffusing information among nodes, through their state, up to a stationary condition where both the objective function is minimized and the state transition function has reached a fixed point.
In our proposed algorithm, the diffusion process is turned itself into an optimization process that must be carried out both when learning and when making predictions. As a matter of fact, inference itself requires the diffusion of information through the graph, that, in our case, corresponds with satisfying the constraints of Eq. (5). For this reason, the testing phase requires a (short) optimization routine to be carried out, that simply looks for the satisfaction of Eq. (5) for test nodes, and it is implemented using the same code that is used to optimize Eq.(8), avoiding to update the previously learned state transition and output functions.
Common graph models exploit synchronous updates among all nodes and multiple iterations for the node state embedding, with a computational complexity for each parameter update , where is the number of iterations, the number of nodes and the number of edges. By simultaneously carrying on the optimization of neural models and the diffusion process, our scheme relies only on 1-hop neighbors for each parameter update, hence showing a computational cost of . From the memory cost viewpoint, the persistent state variable matrix requires space. However, it represents a much cheaper cost than most of GNN models, usually requiring space. In fact, those methods need to store all the intermediate state values of all the iterations, for a latter use in back-propagation.
The evaluation was carried out on two classes of tasks. Artificial tasks (Subgraph matching and Clique detection) are commonly exploited as benchmarks for GNNs, thus, allowing a direct comparison of the proposed constraint based optimization algorithm with respect to the original GNN learning scheme, on the same architecture. The second class of tasks consists of graph classification in the domains of social networks and bioinformatics. The goal is to compare the performances of the proposed approach, hereafter referred to as Lagrangian Propagation GNN (LP-GNN), that is based on a simpler model, with respect to deeper architectures such as Graph Convolutional Neural Networks.
With reference to Table 1, in our experiments we validated two formulations of the state transition function , with two different aggregation scheme. In particular:
Given a graph and a graph such that , the subgraph matching problem consists in finding the nodes of a subgraph which is isomorphic to . The task is that of learning a function , such that , when the node belongs to the given subgraph , otherwise . It is designed to identify the nodes in the input graph that belong to a single subgraph given a priori during learning. The problem of finding a given subgraph is common in many practical problems and corresponds, for instance, to finding a particular small molecule inside a greater compound. An example of a subgraph structure is shown in Fig. 1. Our dataset is composed of 100 different graphs, each one having 7 nodes. The number of nodes of the target subgraph is instead 3.
A clique is a complete graph, i.e. a graph in which each node is connected with all the others. In a network, overlapping cliques (i.e. cliques that share some nodes) are admitted. Clique localization is a particular instance of the subgraph matching problem, with being complete. However, the several symmetries contained in a clique makes the graph isomorphism test more difficult. Indeed, it is known that the graph isomorphism has polynomial time solutions only in absence of symmetries. A clique example is shown in Fig. 2. In the experiments, we consider a dataset composed by graphs having 7 nodes each, where the dimension of the maximal clique is 3 nodes.
We designed a batch of experiments on these two tasks aimed at validating our simple local optimization approach to constraint-based networks. In particular, we want to show that our optimization scheme can learn better transition and output functions than the corresponding GNN of [DBLP:journals/tnn/ScarselliGTHM09]. Moreover, we want to investigate the behaviour of the algorithm for different choices of the function , i.e. when changing how we enforce the state convergence constraints. In particular, we tested functions with different properties: -insensitive functions, i.e , unilateral functions, i.e. , and bilateral functions, i.e. (a function is either unilateral or bilateral). The considered functions are shown in Table 2.
Following the experimental setting of [DBLP:journals/tnn/ScarselliGTHM09]
, we exploited a training, validation and test set having the same size, i.e. 100 graphs each. We tuned the hyperparameters on the validation data, by selecting the node state dimension from the set, the dropout drop-rate from the set , the state transition function from and their number of hidden units from
. We used the Adam optimizer (TensorFlow). Learning rate for parametersand is selected from the set , and the learning rate for the variables and from the set .
We compared our model with the equivalent GNN in [DBLP:journals/tnn/ScarselliGTHM09], with the same number of hidden neurons of the and functions. For the comparison, we exploited the GNN Tensorflow implementation 555The framework is available at https://github.com/mtiezzi/gnn. The documentation is available at http://sailab.diism.unisi.it/gnn/ introduced in [rossi2018inductive]. Results are presented in Table 3.
|Avg # nodes||19.8||13.0||17.9||39.1||25.5||29.8|
|PatchySan||71.0 2.2||45.2 2.8||92.6 4.2||75.9 2.8||60.0 4.8||78.6 1.9|
|AWL||74.5 5.9||51.5 3.6||87.9 9.8||–||–||–|
|GIN||75.1 5.1||52.3 2.8||89.4 5.6||76.2 2.8||64.6 7.0||82.7 1.7|
|GNN||60.9 5.7||41.1 3.8||88.8 11.5||76.4 4.4||61.2 8.5||51.5 2.6|
|LP-GNN*||71.2 4.7||46.6 3.7||90.5 7.0||77.1 4.3||64.4 5.9||68.4 2.1|
We report the average accuracies and standard deviations for the graph classification benchmarks, evaluated on the test set, and we compare multiple GNN models. The proposed model is denoted as LP-GNN and marked with a star. Even though it exploits only shallow representation of nodes, our model performs, on average, on-par to other top models, setting a new state-of-the-art for the Proteins dataset.
Constraints characterized by unilateral functions usually offer better performances than equivalent bilateral constraints. This might be due to the fact that keeping constraints positive (as in unilateral constraints) provides a more stable learning process. Moreover, smoother constraints (i.e squared) or -insensitive constraints tend to perform slightly better than the hard versions. This can be due to the fact that as the constraints move closer to 0 they tend to give a small or null contribution, for squared and abs- respectively, acting as regularizers.
We used 6 graph classification benchmarks: 4 bioinformatics datasets (MUTAG, PTC, NCI1, PROTEINS) and 2 social network datasets (IMDB-BINARY, IMDB-MULTI) [yanardag2015deep], which are becoming popular for benchmarking GNN models. In the bioinformatic graphs, the nodes have categorical input labels (e.g. atom symbol). In the social networks, there are no input node labels. In this case, we followed what has been recently proposed in [DBLP:journals/corr/abs-1810-00826]
, i.e. using one-hot encodings of node degrees. Dataset statistics are summarized in Table4.
We compared the proposed Lagrangian Propagation GNN (LP-GNN) scheme with some of the state-of-the-art neural models for graph classification, such as Graph Convolutional Neural Networks. All the GNN-like models have a number of layers/iterations equal to 5. An important difference with these models is that, by using a different transition function at each iteration, at a cost of a much larger number of parameters, they have a much higher representational power. Even though our model could, in principle, stack multiple diffusion processes at different levels (i.e. different latent representation of the nodes) and, then, have multiple transition functions, we have not explored this direction in this paper. In particular, the models used in the comparison are: Diffusion-Convolutional Neural Networks (DCNN) [DBLP:conf/nips/AtwoodT16], PATCHY-SAN [DBLP:conf/icml/NiepertAK16], Deep Graph CNN (DGCNN) [DBLP:conf/aaai/ZhangCNC18], AWL [ivanov2018anonymous] , GIN-GNN [DBLP:journals/corr/abs-1810-00826], original GNN [DBLP:journals/tnn/ScarselliGTHM09]. Apart from original GNN, we report the accuracy as reported in the referred papers.
We followed the evaluation settings in [DBLP:conf/icml/NiepertAK16]. In particular, we performed 10-fold cross-validation and reported both the average and standard deviation of validation accuracies across the 10 folds within the cross-validation. The stopping epoch is selected as the epoch with the best cross-validation accuracy averaged over the 10 folds. We tuned the hyperparameters by searching: (1) the number of hidden units for both the and functions from the set ; (2) the state transition function from ; (3) the dropout ratio from ; (4) the size of the node state from ; (5) learning rates for both the , , and from . Results are shown in Table 4.
As previously stated, differently from the baseline models, our approach does not rely on a deep stack of layers based on differently learnable filters. Despite of this fact, the simple GNN model trained by the proposed scheme offers performances that, on average, are preferable or on-par to the ones obtained by more complex models that exploit a larger amount of parameters.
Moreover, it is interesting to note that for current GNN models, the role of the architecture depth is twofold. First, as it is common in deep learning, depth is used to perform a multi-layer feature extraction of node inputs. Secondly, it allows node information to flow through the graph fostering the realisation of a diffusion mechanism. Conversely, our model strictly splits these two processes. We believe this distinction to be a fundamental ingredient for a clearer understanding of which mechanism, between diffusion and node deep representation, is concurring in achieving specific performances. Indeed, in this paper, we show that the diffusion mechanism paired only with a simple shallow representation of nodes is sufficient to match performances of much deeper and complex networks.
We showed that formulation of the GNN learning task as a constrained optimization problem allows us to avoid the explicit computation of the fixed point needed to encode the graph. The proposed framework defines how to jointly optimize the model weights and the state representation without the need of separate phases. This approach simplifies the computational scheme of GNNs and allows us to incorporate alternative strategies in the fixed point optimization by the choice of the constraint function . As shown in the experimental evaluation, the appropriate functions may affect generalization and robustness to noise.
Future work will be devoted to explore systematically the properties of the proposed algorithm in terms of convergence and complexity. Moreover, we plan to extend the experimental evaluation to verify the algorithm behaviour with respect to either the characteristics of the input graphs, such as the graph diameter, the variability in the node degrees, the type of node and arc features or to the model architecture (f.i. type of the state transition function, of the constraint function, etc.). Furthermore, the proposed constraint-based scheme can be extended to all the other methods proposed in the literature that exploit more sophisticated architectures.
Finally, LP-GNN can be extended allowing the diffusion mechanism to take place at multiple layers allowing a controlled
integration of diffusion and deep feature extraction mechanisms.
This work was partly supported by the PRIN 2017 project RexLearn, funded by the Italian Ministry of Education, University and Research (grant no. 2017TWNMH2).