# Link prediction for partially observed networks

Link prediction is one of the fundamental problems in network analysis. In many applications, notably in genetics, a partially observed network may not contain any negative examples of absent edges, which creates a difficulty for many existing supervised learning approaches. We develop a new method which treats the observed network as a sample of the true network with different sampling rates for positive and negative examples. We obtain a relative ranking of potential links by their probabilities, utilizing information on node covariates as well as on network topology. Empirically, the method performs well under many settings, including when the observed network is sparse. We apply the method to a protein-protein interaction network and a school friendship network.

## Authors

• 10 publications
• 25 publications
• 20 publications
• ### ALPINE: Active Link Prediction using Network Embedding

Many real-world problems can be formalized as predicting links in a part...
02/04/2020 ∙ by Xi Chen, et al. ∙ 1

• ### Learning Edge Properties in Graphs from Path Aggregations

Graph edges, along with their labels, can represent information of funda...
03/11/2019 ∙ by Rakshit Agrawal, et al. ∙ 0

• ### Link Prediction via Deep Learning

Link prediction aims to infer the missing links or predicting future one...
10/10/2019 ∙ by Weiwei Gu, et al. ∙ 0

• ### Predictability limit of partially observed systems

Applications from finance to epidemiology and cyber-security require acc...
01/17/2020 ∙ by Andres Abeliuk, et al. ∙ 0

• ### Handling Class Imbalance in Link Prediction using Learning to Rank Techniques

We consider the link prediction problem in a partially observed network,...
11/13/2015 ∙ by Bopeng Li, et al. ∙ 0

• ### Link Prediction via Higher-Order Motif Features

02/08/2019 ∙ by Ghadeer Abuoda, et al. ∙ 0

• ### Beyond Observed Connections : Link Injection

In this paper, we proposed the link injection, a novel method that helps...
09/02/2020 ∙ by Jie Bu, et al. ∙ 0

##### This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

## 1 Introduction

A variety of data in many different fields can be described by networks. Examples include friendship and social networks, food webs, protein-protein interaction and gene regulatory networks, the World Wide Web, and many others.

One of the fundamental problems in network science is link prediction, where the goal is to predict the existence of a link between two nodes based on observed links between other nodes as well as additional information about the nodes (node covariates) when available (see [17], [16] and [7] for recent reviews). Link prediction has wide applications. For example, recommendation of new friends or connections for members is an important service in online social networks such as Facebook. In biological networks, such as protein-protein interaction and gene regulatory networks, it is usually time-consuming and expensive to test existence of links by comprehensive experiments; link prediction in these biological networks can provide specific targets for future experiments.

There are two different settings under which the link prediction problem is commonly studied. In the first setting, a snapshot of the network at time , or a sequence of snapshots at times , is used to predict new links that are likely to appear in the near future (at time ). In the second setting, the network is treated as static but not fully observed, and the task is to fill in the missing links in such a partially observed network. These two tasks are related in practice, since a network evolving over time can also be partially observed and a missing link is more likely to emerge in the future. From the analysis point of view, however, these settings are quite different; in this paper, we focus on the partially observed setting and do not consider networks evolving over time.

There are several types of methods for the link prediction problem in the literature. The first class of methods consists of unsupervised approaches based on various types of node similarities. These methods assign a similarity score to each pair of nodes and

, and higher similarity scores are assumed to imply higher probabilities of a link. Similarities can be based either on node attributes or solely on the network structure, such as the number of common neighbors; the latter are known as structural similarities. Typical choices of structural similarity measures include local indices based on common neighbors, such as the Jaccard index

[16] or the Adamic-Adar index [1], and global indices based on the ensemble of all paths, such as the Katz index [14] and the Leicht-Holme-Newman Index [15]. Comprehensive reviews of such similarity measures can be found in [16] and [17].

Another class of approaches to link prediction includes supervised learning methods that use both network structures and node attributes. These methods treat link prediction as a binary classification problem, where the responses are indicating whether there exists a link for a pair, and the predictors are covariates for each pair, which are constructed from node attributes. A number of popular supervised learning methods have been applied to the link prediction problem. For example, [2] and [4]

use the support vector machine with pairwise kernels, and

[8] compares the performance of several supervised learning methods. Other supervised methods use probabilistic models for incomplete networks to do link prediction, for example, the hierarchical structure models [6], latent space models [11], latent variable models [10, 18], and stochastic relational models [21].

Our approach falls in the supervised learning category, in the sense that we make use of both the node similarities and observed links. However, one difficulty in treating link prediction as a straightforward classification problem is the lack of certainty about the negative and positive examples. This is particularly true for negative examples (absent edges). In biological networks in particular, there may be no certain negative examples at all [3]. For instance, in a protein-protein interaction network, an absent edge may not mean that there is no interaction between the two proteins – instead, it may indicate that the experiment to test that interaction has not been done, or that it did not have enough sensitivity to detect the interaction. Positive examples could sometimes also be spurious – for example, high-throughput experiments can yield a large number of false positive protein-protein interactions [19]

. Here we propose a new link prediction method that allows for the presence of both false positive and false negative examples. More formally, we assume that the network we observe is the true network with independent observation errors, i.e., with some true edges missing and other edges recorded erroneously. The error rates for both kinds of errors are assumed unknown, and in fact cannot be estimated under this framework. However, we can provide rankings of potential links in order of their estimated probabilities, for node pairs with observed links as well as for node pairs with no observed links. These relative rankings rather than absolute probabilities of edges are sufficient in many applications. For example, pairs of proteins without observed interactions that rank highly could be given priority in subsequent experiments. To obtain these rankings, we utilize node covariates when available, and/or network topology based on observed links.

The rest of the paper is organized as follows. In Section 2, we specify our (rather minimal) model assumptions for the network and the edge errors. We propose link ranking criteria for both directed and undirected networks in Section 3. The algorithms used to optimize these criteria are discussed in Section 4. In Section 5 we compare performance of proposed criteria to other link prediction methods on simulated networks. In Section 6, we apply our methods to link prediction in a protein-protein interaction network and a school friendship network. Section 7 concludes with a summary and discussion of future directions.

## 2 The network model

A network with nodes (vertices) can be represented by an adjacency matrix , where

We will consider the link prediction problem for both undirected and directed networks. Therefore can be either symmetric (for undirected networks) or asymmetric (for directed networks).

In our framework, we distinguish between the adjacency matrix of the true underlying network , and its observed version . We assume that each

follows a Bernoulli distribution with

. Given the true network, we assume that the observed network is generated by

 P(Aij=1|ATrueij=1)=α,P(Aij=0|ATrueij=0)=β,

where and are the probabilities of correctly recording a true edge and an absent edge, respectively. Note that we assume that this probability is constant and does not depend on , , or . Then we have

 ~PijDef=P(Aij=1)=(α+β−1)Pij+(1−β). (1)

If the values of , and were known, then the probabilities of true edges conditional on the observed adjacency matrix could have been estimated as

 P(ATrueij=1|Aij=1) =αPij~Pij, (2) P(ATrueij=1|Aij=0) =(1−α)Pij1−~Pij. (3)

It is easy to check that both (2) and (3) are monotone increasing functions of . Taking (1) into account implies that they are also increasing functions of as long as . This gives us a crucial observation: if the goal is to obtain relative rankings of potential links, it is sufficient to estimate , and it is not necessary to know , and .

An important special case in this setting is . Then all the observed links are true positives, and we only need to provide a ranking for node pairs without observed links. This can be applied in recommender systems, for example, for recommending possible new friends in a social network. Another special case is when , which corresponds to all absent edges being true negatives. This setting can be used to frame the problem of investigating reliability of observed links, for example, in a gene regulatory network inferred from high-throughput gene expression data. An estimate of provides rankings for both these special cases and the general problem, and thus we focus on estimating for the rest of the paper.

In this section, we propose criteria for estimating the probabilities of edges in the observed network, , for both directed and undirected networks. The criteria rely on a symmetric matrix with , which describes the similarity between nodes and . The similarity matrix can be obtained from different sources, including node information, network topology, or a combination of the two. We will discuss choices of later in this section.

### 3.1 Link prediction for directed networks

First we consider directed networks.

The key assumption we make is that if two pairs of nodes are similar to each other, the probability of links within these two pairs are also similar. Specifically, in Figure 1, and are assumed close in value if node is similar to node and node is similar to node . For directed networks, we measure similarity of node pairs and by the product (see Figure 1), which implies two pairs are similar only if both pairs of endpoints are similar. This assumption should not to be confused with a different assumption made by many unsupervised link prediction methods, which assume that a link is more likely to exist between similar nodes, applicable to networks with assortative mixing. Assortative networks are common – a typical example is a social network, where people commonly tend to be friends with those of similar age, income level, race, etc. However, there are also networks with disassortative mixing, in which the assumption that similar pairs are more likely to be connected is no longer valid – for example, predators do not typically feed on each other in a food web. Our assumption, in contrast, is equally plausible for both assortative and disassortative networks, as well as more general settings, as it does not assume anything about the relationship between and .

Motivated by this assumption of similar probabilities of links for similar node pairs, we propose to estimate by

 ^f=argminf1n2n∑ij(Aij−fij)2+λn4n∑ii′jj′Wii′Wjj′(fij−fi′j′)2, (4)

where is a real-valued matrix, and is a tuning parameter. The first term is the usual squared error loss connecting the parameters with the observed network. The minimizer of its population version, i.e., is . The second term enforces our key assumption, penalizing the difference between and more if two node pairs and

are similar. The choice of the squared error loss is not crucial, and other commonly used loss functions could be considered instead, for example, the hinge loss or the negative log-likelihood. The main reason for choosing the squared error loss is computational efficiency, since it makes (

4) a quadratic problem; see more on this details in Section 4.

In some applications, we may have additional information about true positive and negative examples, i.e., some ’s may be known to be true 1’s and true 0’s, while others may be uncertain. This could happen, for example, when validation experiments have been conducted on a subset of a gene or protein network inferred from expression data. If such information is available, it makes sense to use it, and we can then modify criterion (4) as follows:

 argminf1∑nijEijn∑ijEij(Aij−fij)2+λn4n∑ii′jj′Wii′Wjj′(fij−fi′j′)2, (5)

where if it is known that , and 0 otherwise. This is similar to a semi-supervised criterion proposed in [13]. However, [13] did not consider the uncertainty in positive and negative examples, nor did they consider the undirected case which we discuss next. Since (5) only involves a partial sum of the loss function terms, we will refer to (5) as the partial-sum criterion and (4) as the full-sum criterion for the rest of the paper.

### 3.2 Link prediction for undirected networks

For undirected networks, our key assumption that and are close if two pairs and are similar needs to take into account that the direction no longer matters; thus the pairs are similar if either is similar to and is similar to , or if is similar to and is similar to (see Figure 2. Thus we need a new pair similarity measure that combines and . There are multiple options; for example, two natural combinations are

 S1=Wii′Wjj′+Wij′Wji′ and S2=max(Wii′Wjj′,Wij′Wji′).

Empirically, we found that performs better than for a range of real and simulated networks. The reason for this can be easily illustrated on the stochastic block model. The stochastic block model is a commonly used model for networks with communities, where the probability of a link only depends on the community labels of its two endpoints. Specifically, given community labels ,

’s are independent Bernoulli random variables with

 Pij=Scicj, (6)

where is a symmetric matrix, and is the number of communities in the network. Suppose we have the best similarity measure we can possibly hope to have based on the truth, , where is the indicator function. In that case, (6) implies if , whereas the sum of the weights would be misleading.

Using as the measure of pair similarity, we propose estimating for undirected networks by

 argminf 1n2n∑i

Similarly to the directed case, if we have information about true positive and negative examples, we can use a partial-sum criterion

 argminf 1∑ni

where if it is known that , otherwise .

### 3.3 Node similarity measures

The last component we need to specify is the node similarity matrix . One typical situation is when we have reasons to believe that the external node covariates are related to the structure of the network, in which case it is natural to use covariate information to construct . Though more complicated formats do exist, node covariates are typically represented by an matrix where is the value of variable on node . Then can be taken to be some similarity measure between the -th and -th rows of . For example, if contains only numerical variables and has been standardized, we can use the exponential decay kernel,

 Wii′=exp{−∥Xi−Xi′∥2σ2},

where is the Euclidean vector norm.

When node covariates are not available, node similarity is usually obtained from the topology of the observed network , i.e., is large if and have a similar pattern of connections with other nodes. For undirected networks, a simple choice of could be

 Wii′=|{k:Aik=Ai′k}|n , (9)

where denotes cardinality of a set. This particular measure turns out to be not very useful: since most real networks are sparse, most entries of any -th column will be 0, and thus most of ’s would be large. A more informative measure is the Jaccard index [16],

 Wii′=|N(i)∩N(i′)||N(i)∪N(i′)|, (10)

where is the set of neighbors of node .

The directed networks case is similar, except we need to count the in and the out links separately. The formulas corresponding to (9) and (10) become

 Wii′ =|{k:Aik=Ai′k}|2n+|{k:Aki=Aki′}|2n, Wii′ =|N1(i)∩N1(i′)|2|N1(i)∪N1(i′)|+|N2(i)∩N2(i′)|2|N2(i)∪N2(i′)|,

where and .

## 4 Optimization algorithms

The proposed link prediction criteria are convex and quadratic in parameters, and thus optimization is fairly straightforward. The obvious approach is to treat the matrix as a long vector with elements (or in the undirected case), and solve the linear system obtained by taking the first derivative of any criterion above with respect to this vector. However, solving a system of linear equations could be challenging for large-scale problems [5]; the number of parameters here is , and so the linear system requires memory. However, if is sparse, or sparsified by applying thresholding or some other similar method, then solving the linear system is the efficient choice.

If the matrix is not sparse, an iterative algorithm with sequential updates that only requires memory would be a better choice than solving the linear system. We propose an iterative algorithm following the idea of block coordinate descent [9, 20]. A block coordinate descent algorithm partitions the coordinates into blocks and iteratively optimizes the criterion with respect to each block while holding the other blocks fixed.

First, we derive the update equations for directed networks. Note (4) and (5) can be written in the general form

 Q=1n2n∑ijVij(Aij−fij)2+λn4n∑ii′jj′Wii′Wjj′(fij−fi′j′)2, (11)

where for (4) and for (5). For any matrix , let be the th row of . We treat as a block, and update iteratively. Define . Then

 ∑ijVij(fij−Aij)2 =∑i(fi⋅−Ai⋅)TVi(fi⋅−Ai⋅). (12)

Let be an diagonal matrix with . Then

 ∑jj′Wjj′(fij−fi′j′)2 =fTi⋅Dfi⋅−2fTi⋅Wfi′⋅+fTi′⋅Dfi′⋅ (13)

Plugging (12) and (13) into (11), and taking the first derivative of with respect to , we obtain

 ∂Q∂fi⋅ = 2n2(fi⋅−Ai⋅)Vi+ λ4n4⎡⎣Wii(Dfi⋅−Wfi⋅)+∑i′≠iWii′(Dfi⋅−Wfi′⋅)⎤⎦.

Solving with respect to , we obtain the updating formula

 f(t+1)i⋅ ←(n2Vi+2λ∑i′Wii′D−2λWiiW)−1⎛⎝n2Ai⋅Vi+2λ∑i′≠iWii′Wf(t)i′⋅⎞⎠, (15)

where is the value of at iteration .

This update is fast to compute but its derivation relies on the product form of and , and thus is not directly applicable in the undirected case, where is used as the similarity measure. However, we can still approximate with a product, using the fact that for , . Thus, for sufficiently large , we have

 [max(Wii′Wjj′,Wij′Wji′)]q≈(Wii′Wjj′)q+(Wij′Wji′)q. (16)

Further, is a monotone transformation of and can also serve as a similarity measure. Based on (16), we propose to substitute the following approximate criterion for undirected networks,

 Q = 1n2n∑i

where for the full sum criterion and for the partial sum criterion. By symmetry,

 n∑i

This is now in the same form as (11), with each term in the sum containing a product of and , and therefore (4) can be solved by block coordinate descent with an analogous updating equation as that in the directed network case.

In practice, we found that when is sparse or truncated to be sparse, solving the linear system can be much faster than the block coordinate descent method; however, when is dense and the number of nodes is reasonably large, the block coordinate descent method dominates directly solving linear equations.

## 5 Simulation studies

In this section, we test performance of our link prediction methods on simulated networks. In all cases, each network consists of nodes, and node ’s covariates

are independently generated from a multivariate normal distribution

with . Each is generated independently, with . We consider the following functions :

 (a)∑k(Xik−Xjk), (a′) ∑k(Xik−Xjk)−8, (b)2XTiXj/∥Xj∥, (b′) XTiXj/∥Xj∥−6, (c)∑k(Xik+Xjk), (c′) ∑k(Xik+Xjk)−8, (d)XTiXj, (d′) XTiXj−6.

The right hand column gives sparser versions of functions in the left hand column (subtracting a constant within the logit link functions lowers the overall degree), which we use to compare dense and sparse networks (the average degrees of all these networks are reported in Figures 3 and 4). Functions (a) and (b) are asymmetric in and , giving directed networks, while (c) and (d) are symmetric functions corresponding to undirected networks. Further, and are linear functions; is the projection model proposed in [11], under which the link probability is determined by the projection of onto the direction of , and is an undirected version of the projection model.

We also generate indicators ’s as independent Bernoulli variables taking values 1 and 0 with equal probability, and set . This setup corresponds to the “partially observed” network of the title, where all the observed edges are true but the missing edges may or may not be true 0s.

Since we have node covariates affecting the probabilities of links in this case, we define the similarity matrix by

 Wii′=exp{−∥Xi−Xi′∥2σ2},

where we choose . After truncating at 0.1, we optimize all criteria by solving linear equations, with chosen by 5-fold cross validation.

The performance of link prediction is evaluated on the “test” set . We report ROC curves, which only depend on the rankings of the estimates rather than their numerical values. Specifically, let be the ranking of on the test set in descending order. For any integer , we define false positives as pairs ranked within top but without links in the true network (), and true positives as pairs ranked within top with . Then the true positive rate (TPR) and the false positive rate (FPR) are defined by

 TPR(k)=|{(i,j):Eij=0,Rij≤k,ATrueij=1}||{(i,j):Eij=0,ATrueij=1}|, FPR(k)=|{(i,j):Eij=0,Rij≤k,ATrueij=0}||{(i,j):Eij=0,ATrueij=0}|.

The ROC curves showing the false positive rate vs. the true positive rate over a range of values are shown in Figures 3 (directed networks) and 4 (undirected networks). Each curve is the average of 20 replicates. We also show the ROC curve constructed from true ’s as a benchmark..

Overall, both the full sum and the partial sum criteria perform well. There is little difference between directed network models and their undirected versions. As expected, the partial sum criterion always gives better results since it has more information and only uses the true positive and negative examples for training. But its performance is quite comparable to the completely unsupervised full sum criterion, except perhaps for model . The gaps between the unsupervised full sum criterion and semi-supervised partial sum criterion become smaller for sparse networks, as the false negatives in the full sum are only a small proportion of the large number of true negatives in a sparse network. The ROC curve obtained from the true model in sparse networks is better than in the corresponding dense networks; this seemingly counter-intuitive finding is also explained by the large number of 0s in sparse networks. However, gaps between both our link prediction methods and the true model are larger in all the sparse networks than in their dense counterparts. This confirms the observation that a small number of positive examples in sparse networks makes the link prediction problem challenging.

## 6 Applications

### 6.1 The protein-protein interaction network

Our first application is to an undirected network containing yeast protein-protein interactions from [19]. This network was edited to contain only highly reliable interactions supported by multiple experiments [4], resulting in 984 protein nodes and 2438 edges, with the average node degree about 5. We take this verified network to be the true underlying network . [4] also constructed a matrix measuring similarities between proteins based on gene expression, protein localization, phylogenetic profiles and yeast two-hybrid data, which we use as the node similarity matrix for link prediction.

Here, we compare the full sum criterion (3.2), the partial sum criterion (3.2), and the latent variable model proposed by [10]. To test prediction, we generate indicators ’s as independent Bernoulli variables taking value 1 with probability , and set . We consider three different values of , , corresponding to different amounts of available information.

We use the block coordinate descent algorithm proposed in Section 4 to approximately optimize (3.2) and (3.2), with and chosen by cross-validation. The latent variable model depends on a tuning parameter , the dimension of the latent space. We fix since larger values of do not significantly change the performance in this example. We again use ROC curves to evaluate the link prediction performance on the set . Each ROC curve in Figure 5 is the average of 10 random realizations of ’s.

The semi-supervised criterion always performs better than the unsupervised criterion, as it should. Further, the semi-supervised criterion almost always outperforms the latent variable model, except for very small values of the false positive rate, and the fully unsupervised criterion also starts to outperform the latent variable model as the false positive rate increases. The latent variable model is also more sensitive to the sampling rate , with performance deteriorating for . This is because the model relies heavily on the structure of the network, and a low sampling rate may substantially distort the overall network topology. On the other hand, we use the node similarity matrix which depends only on the features of the proteins, and is thus unaffected by the sampling rate.

### 6.2 The school friendship network

This dataset is a school friendship network from the National Longitudinal Study of Adolescent Health (see

[12] for detailed information). This network contains 1011 high school students and 5459 directed links connecting students to their friends, as reported by the students themselves. The average degree of this network is also around . Here we test our two link prediction criteria, with the same settings for as in the protein example. Since the latent variable model of [10] is not applicable to directed networks, we omit it here. Due to lack of node covariates, we construct a network-based similarity by using the Jaccard index defined in (10). We again apply block coordinate descent to minimize the criteria with chosen by cross-validation, and report the average ROC curves over 10 realizations of ’s. As shown in Figure 6, both criteria perform fairly well for and , but fail for , as the sampling rate is too small for to capture the overall network topology. This does not happen in the protein-protein interactions network, since is constructed from covariates on proteins and is unaffected by sub-sampling.

## References

• [1] L. A. Adamic and E. Adar. Friends and neighbors on the web. Social Networks, 25(3):211–230, 2003.
• [2] A. Ben-Hur and W. S. Noble. Kernel methods for predicting protein-protein interactions. Bioinformatics, 21(Suppl. 1):i38–i46, 2005.
• [3] A. Ben-Hur and W. S. Noble. Choosing negative examples for the prediction of protein-protein interactions. BMC Bioinformatics, 7(Suppl 1):S2, 2006.
• [4] K. Bleakley, G. Biau, and J.-P. Vert. Supervised reconstruction of biological networks with local models. Bioinformatics, 23(13):i57–i65, 2007.
• [5] S. Boyd and L. Vandenberghe. Convex optimization. Cambridge University Press, New York, 2004.
• [6] A. Clauset, C. Moore, and M. E. J. Newman. Hierarchical structure and the prediction of missing links in networks. Nature, 453(7191):98–101, May 2008.
• [7] L. Getoor and C. P. Diehl. Link mining: A survey. ACM SIGKDD Explorations Newsletter, 7(2):3–12, 2005.
• [8] M. A. Hasan, V. Chaoji, S. Salem, and M. Zaki. Link prediction using supervised learning. In Workshop on Link Analysis, Counter-terrorism and Security (at SIAM Data Mining Conference), 2006.
• [9] C. Hildreth. A quadratic programming procedure. Naval Research Logistics Quarterly, 4(1):79¨C85, 1957.
• [10] P. D. Hoff. Modeling homophily and stochastic equivalence in symmetric relational data. In Advances in Neural Information Processing Systems, volume 19. MIT Press, Cambridge, MA, 2007.
• [11] P. D. Hoff, A. E. Raftery, and M. S. Handcock. Latent space approaches to social network analysis. Journal of the American Statistical Association, 97:1090–1098, 2002.
• [12] D. R. Hunter, S. M. Goodreau, and M. S. Handcock. Goodness of fit of social network models. J. Amer. Statist. Assoc., 103(481):248–258, 2008.
• [13] H. Kashima, T. Kato, Y. Yamanishi, M. Sugiyama, and K. Tsuda.