Attributed networks are ubiquitous in a variety of real-world applications. Data from many real-world domains can be represented as attributed networks, where nodes represent entities with attributes and edges express the interactions or relationships between entities. Different from plain networks where only structural information exists, attributed networks also contain rich features to provide more details to describe individual elements of the networks. For instance, in social networks, user profiles contain important information to describe users. In citation networks, paper abstracts can provide complementary information to the citation structures. In gene regulatory networks, gene sequence expressions are the attributes beside the interactions between molecular regulators. Due to the ubiquity of attributed networks, various data mining tasks on attributed networks have attracted an upsurge of interest such as community detection (falih2018community; li2018community; pei2015nonnegative), link prediction (barbieri2014follow; li2018streaming; brochier2019link), network embedding (huang2017label; huang2017accelerated; meng2019co), etc.
Anomaly detection is one of the most vital problems among these tasks on attributed networks because of its significant implications in a wide range of real-world applications including cyber attack detection in computer networks, fraud detection in finance and spammers discovery in social media, to name a few. It is more challenging to detect anomalies on attributed networks because both attributes and structures should be taken into consideration in order to detect anomalous nodes. An illustration is shown in Figure 1. The anomalous node is different from others because: 1) structurally it connects to all other nodes and 2) its attributes are significantly different from the majority.
Several approaches for anomaly detection on attributed networks have been proposed recently in the literature. Most of them aim at detecting anomalies in an unsupervised fashion because of the prohibitive cost for accessing the ground-truth anomalies (ding2019deep)
. They can be categorized into four types: community analysis, subspace selection, residual analysis and deep learning methods. Community analysis methods(gao2010community) detect anomalies by identifying the abnormality of current node with other nodes within the same community. Subspace selection approaches (perozzi2014focused) first learn a subspace for features and then discover anomalies in that learned subspace. Residual analysis methods (li2017radar; peng2018anomalous) explicitly model the residual information by reconstructing the input attributed network based on matrix factorization. Deep learning methods use deep neural networks to capture the nonlinearity of networks and detect anomalies in an unsupervised (ding2019deep) or supervised way (liang2018semi).
However, there are three major issues in existing approaches: sparsity and nonlinearity capturing, residual modeling, and network smoothing. Capturing sparsity and nonlinearity is important in anomaly detection on networks because real-world attributed networks are complex and non-linear. Previous shallow models such as non-negative matrix factorization (li2017radar; peng2018anomalous) fail to detect anomalies because of the incapability of modeling nonlinearity. Although residual modeling has been explored in previous studies (li2017radar; peng2018anomalous), the residual information has been modeled from the reconstruction error. Thus, they cannot be adaptively learned from the input networks. Smoothing networks, which is based on the homophily hypothesis (mcpherson2001birds), is a commonly used strategy to detect anomalies on networks, e.g., (ding2019deep). However, these methods are not in line with anomaly detection because they might over-smooth the node representations, and make anomalous nodes less distinguishable from the majority (li2019specae).
To tackle these issues, in this paper, we propose Residual Graph Convolutional Network (ResGCN), a novel approach for anomaly detection on attributed networks. ResGCN is capable of solving the above three problems as follows: (1) to capture the sparsity and nonlinearity of networks, ResGCN is based on GCN to model the attributed networks; (2) to model residual information, ResGCN learns residual directly from the input using a deep neural network; and (3) to prevent over-smoothing of node representations, ResGCN incorporates the attention mechanism based on learned residual information. Thus, the information propagation of anomalous nodes can be reduced. The contributions of this paper are summarized as follows:
We propose novel anomaly detection method named ResGCN. ResGCN captures the sparsity and nonlinearity of networks using GCN, learns the residual information using a deep neural network, and reduces the adverse effect from anomalous nodes using the residual-based attention mechanism.
We propose a residual information based anomaly ranking strategy and the residual information is learned from the input network instead of reconstruction errors.
We conduct extensive experiments on real-world attributed networks. Experimental results demonstrate the effectiveness of our proposed ResGCN in the task of anomaly detection w.r.t. different evaluation metrics.
The rest of this paper is organized as follows. Section 2 formally defines the problem of anomaly detection on attributed networks. Section 3 introduces the proposed ResGCN model for anomaly detection. Section 4 provides empirical evidence of ResGCN performance on anomaly detection in real-world networks w.r.t. different evaluation metrics. Section 5 briefly discusses related work on anomaly detection on attributed networks. Finally, we conclude in Section 6.
2 Problem Definition
We first summarize some notations and definitions used in this papers. Following the commonly used notations, we use bold uppercase characters for matrices, e.g.,
, bold lowercase characters for vectors, e.g.,, and normal lowercase characters for scalars, e.g., . The The row of a matrix is denoted by and element of matrix is denoted as . The Frobenius norm of a matrix is represented as and is the norm. In detail, the main symbols are listed in Table 1.
|number of edges|
|number of nodes|
|number of attributes|
|the trainable weight matrix in the layer|
|the latent representation matrix in the layer|
|the residual matrix in the layer|
|the trade-off parameter for reconstruction error|
|the residual parameter|
Attributed Networks. An attributed network consists of: (1) a set of nodes , where is the number of nodes; (2) a set of edges , where is the number of edges; and (3) the node attribute matrix , the row vector is the attribute of node .
The topological structure of attributed network can be represented by an adjacency matrix , where if there is an edge between node and node . Otherwise, . We focus on the undirected networks in this study and it is trivial to extend it to directed networks. The attribute of can be represented by an attribute matrix . Thus, the attributed network can be represented as . With these notations and definitions, same to previous studies (li2017radar; peng2018anomalous; ding2019deep), we formulate the task of anomaly detection on attributed networks:
Anomaly Detection on Attributed Networks. Given an attributed network , which is represented by the adjacency matrix and attribute matrix , the task of anomaly detection is to find a set of nodes that are rare and differ singularly from the majority reference nodes of the input network.
3 Proposed Method
In this section we first introduce the background of GCN. Next, we present the proposed model ResGCN in details. Then we analyze the complexity of ResGCN.
3.1 Graph Convolutional Networks
GCN learns node representations by passing and aggregating messages between neighboring nodes. Different types of GCN have been proposed recently (kipf2016semi; hamilton2017inductive), and we focus on one of the most widely used versions proposed in (kipf2016semi). Formally, a GCN layer is defined as
where is the latent representation of node in layer , is the set of neighbors of node , and is the layer-specific trainable weight matrix.kipf2016semi) (written as below). is the diagonal degree matrix of defined as where is the adjacency matrix of the input attributed network with self connections . Equivalently, we can rewrite GCN in a matrix form:
For the first layer, is the attribute matrix of the input network. Therefore, we have
The architecture of GCN can be trained end-to-end by incorporating task-specific loss functions. In the original study, GCN aims at semi-supervised classification task so the cross-entropy loss is evaluated by adding the softmax function as the output of the last layer.Formally, the overall cross-entropy error is evaluated on the graph for all the labeled samples:
where is the set of nodes with labels, is the number of classes, is the label and
is the prediction of GCN passing the hidden representation in the final layerto a softmax function.
Note that original GCN (kipf2016semi)
is designed for semi-supervised learning, our target is to detect anomalies in an unsupervised way. Therefore, the cross entropy loss for (semi-)supervised learning is not suitable in our problem settings. We will introduce our proposed loss function which is based on network reconstruction errors in the following section.
In this section, we present the proposed framework of ResGCN in details. ResGCN consists of four components: residual modeling, representation learning, network reconstruction and anomaly ranking. The architecture of this model is illustrated in Figure 2.
3.2.1 Residual Modeling
Although some previous studies explicitly model the residual information for anomaly detection on attributed networks, e.g., Radar (li2017radar) and ANOMALOUS (peng2018anomalous), these methods have two major limitations: (1) They are based on linear models, e.g., matrix factorization, so these shallow models are incapable of capturing the nonlinearity of networks. (2) The residual information has been modeled from the reconstruction error. Thus, they cannot be adaptively learned from the input networks. However, real-world networks are complex and residual information has different patterns in different datasets. Motivated by the study (dabkowski2017real), which proposes to learn the saliency map based on convolutional network, we propose to use a deep neural network to learn the residual by capturing the nonlinearity in ResGCN. Formally,
where is the input for the fully connected (FC) layer , and is the layer-specific trainable weight matrix which needs to be learned during the training of the model. The output of this network is the residual matrix, denoted as .
Another aim of the residual modeling component is to learn the attention weights to control the message passing in network representation based on the residual information. Similarly, we use FC layer which takes the residual matrix as input and the calculation is the same to Eq (5). Each output of the FC layer corresponds to the attention weights for each GCN layer shown in Figure 2. Therefore, the number of FC layers to learn the weights is equal to the number of GCN layers which will be presented below.
3.2.2 Representation Learning
The second component of ResGCN aims at learning representations of the input attributed network. Our proposed representation learning method can not only capture the sparsity and nonlinearity of networks but also prevent the information propagating of anomalies. In this component, we adopt GCN with attention which is based on the residual information modeled in the first component to learn the embeddings of nodes. To make the computations tractable, we follow (zhu2019robust) and assume all hidden representations of nodes are independent. Therefore, we can aggregate node neighbors as follows:
To prevent the information propagation from the anomalous nodes, we propose an attention mechanism based on the residual information modeled by the first component to assign different weights to neighbors. The reason is that it is intuitive the nodes with larger residual errors are more likely to be anomalies (li2017radar). Motivated by (zhu2019robust), we use the smooth exponential function to control the effect of residual information on weights. Formally, the weight is defined as
where are the attention weights of node in the layer and is a hyper-parameter. By taking the attention weights into account, the modified aggregated node neighbor representation can be written as:
where is the element-wise product. Then we apply learnable filters and non-linear activation function (ReLU used in this study) to in order to calculate . Formally the layer is defined as:
Equivalently, the matrix form is:
where . Similarly, for the first layer, we have
The output of the last GCN layer is the node embedding matrix .
3.2.3 Network Reconstruction
The target of the third component of ResGCN is to reconstruct the network which consists of structure reconstruction and attribute reconstruction. Both reconstructions are based on the latent representation learned in the representation learning component.
Let denote the reconstructed adjacency matrix. Following (ding2019deep; kipf2016variational), we use the inner product of the latent representations between two nodes to predict if an edge exists between them. Intuitively, if the latent representations of two nodes are similar, it is more likely that there is an edge between them. Formally, the prediction between two nodes and can represented as follows:
function is to convert the prediction as a probability value. Accordingly, the whole reconstructed network structure based on the latent representationscan be represented as follows:
Correspondingly, the reconstruction error for structure can be represented as:
To reconstruct the original attributes, DOMINANT (ding2019deep) uses another graph convolution layer as the decoder to reconstruct the attributes. However, considering that graph convolution is simply a special form of Laplacian smoothing and mixes the nodal features and its nearby neighbors (li2018deeper), we adopt the multi-layer perception as our decoder instead. Formally, let be the reconstructed attributes and the reconstruction process can be formalized as follows:
where denotes the number of FC layers and denotes n-layer perception which is composed with linear functions followed by non-linear activation function. By taking the residual into consideration, the attribute reconstruction is:
where is the residual parameter to control how much residual information we want to use in the attribute reconstruction error. This error is similar to (li2017radar; peng2018anomalous) which explicitly incorporate the residual information in attribute reconstruction.
Based on the structure and attribute reconstruction errors, we can propose the objective function of our proposed ResGCN model. To jointly learn the reconstruction errors, the objective function of ResGCN is defined as the weighted combination of two errors:
where is the trade-off parameter to control the importance of errors from structure and attributed reconstruction. By minimizing the objective function, we aim to approximate the input attributed network based on the latent representations. Different from previous studies which rank reconstruction errors to detect anomalous nodes (ding2019deep), in our proposed model, we rank the residual matrix for anomaly identification. Formally, the anomaly score for node is
Finally, the anomalies are the nodes with larger scores and we can detect anomalies according to the ranking of anomaly scores. This ranking strategy is superior to reconstruction error based methods because in our model the residual is explicitly learn from the data and implicitly updated by minimizing the reconstruction error. Therefore, it can better capture the anomaly of the data and less be adversely influenced by the noise from the model.
3.3 Complexity Analysis
The computational complexity of GCN is linear to the number of edges on the network. For a particular layer, the convolution operation is and its complexity is (ding2019deep), where is the number of non-zero elements in the adjacency matrix , is the dimensions of attributes, and is the number of feature maps of the weight matrix. For network reconstruction, we use link prediction to reconstruct the structure and multi-layer perception to reconstruct the attribute both of which are pairwise operations. Thus, the overall complexity is where is the summation of all feature maps across different layers.
In this section, we evaluate the effectiveness of our proposed ResGCN model on several real-world datasets and present experimental results in order to answer the following three research questions.
RQ1: Does ResGCN improve the anomaly detection performance on attributed networks?
RQ2: Is deep residual matrix ranking strategy effective in identifying anomalies?
RQ3: How do the parameters in ResGCN affect the anomaly detection performance?
In order to evaluate the effectiveness of our proposed method, we conduct experiments on two types of real-world attributed networks: data with and without ground-truth anomaly labels. All networks have been widely used in previous studies (li2017radar; peng2018anomalous; ding2019deep; gutierrez2019multi):
Networks with ground-truth anomaly labels: Amazon and Enron111https://www.ipd.kit.edu/mitarbeiter/muellere/consub/. Amazon is a co-purchase network (muller2013ranking). It contains 28 attributes for each node describing properties about online items including rating, price, etc. The anomalous nodes are defined as nodes having the tag amazonfail. Enron is an email network (metsis2006spam) where each node is an email with 20 attributes describing metadata of the email including content length, number of recipients, etc, and each edge indicates the email transmission between people. Spammers are labeled as the anomalies in Enron data. The details of these attributed networks are shown in Table 2.
Networks without ground-truth anomaly labels: BlogCatalog, Flickr and ACM222http://people.tamu.edu/~xhuang/Code.html. BlogCatalog is a blog sharing website where users are the nodes and following relations between users are edges. Each user is associated with a list of tags to describe themselves and their blogs, which are used as attributes. Flickr is an image hosting and sharing website. Similarly, users and user following relations are nodes and edges, respectively. Tags are the attributes. ACM is a citation network where each node is a paper and each edge indicates a citation relation between papers. Paper abstracts are used as attributes. The details of these attributed networks are shown in Table 3.
For the networks with labels, we directly use these provided labels to evaluate our method. For the data without labels, we need to manually inject anomalies for empirical evaluation. To make a fair comparison, we follow previous studies for anomaly injection (ding2019deep). In specific, two anomaly injection methods have been used to inject anomalies by perturbing topological structure and nodal attributes, respectively:
Structural anomalies: structural anomalies are generated by perturbing the topological structure of the network. It is intuitive that in real-world networks, small cliques are typically anomalous in which a small set of nodes are much more connected to each other than average (skillicorn2007detecting). Thus, we follow the method used in (ding2019deep; ding2019interactive) to generate some small cliques. In details, we randomly select nodes from the network and then make those nodes fully connected, and then all the nodes forming the clique are labeled as anomalies. cliques are generated repeatedly and totally there are structural anomalies.
Attribute anomalies: we inject an equal number of anomalies from structural perspective and attribute perspective. Same to (ding2019deep; song2007conditional), nodes are randomly selected as the attribute perturbation candidates. For each selected node , we randomly select another nodes from the network and calculate the Euclidean distance between and all the nodes. Then the node with largest distance is selected as and the attributes of node is changed to of node . The selected node is regarded as the attribute anomaly.
In the experiments, we set and set to 10, 15, and 20 for BlogCatalog, Flickr and ACM, respectively which are the same to (ding2019deep) in order to make the comparison with DOMINANT (ding2019deep). To facilitate the learning process, in our experiments, we follow (ding2019interactive)
to reduce the dimensionality of attributes using Principal Component Analysis (PCA) and the dimension is set to 20.
4.2 Evaluation Metrics
In the experiments, we use two evaluation metrics to validate the performance of these anomaly detection approaches:
: we use the area under the receiver operating characteristic curve (ROC-AUC) as the evaluation metric for anomaly detection as it has been widely used in previous studies(li2017radar; peng2018anomalous; ding2019deep; gutierrez2019multi). ROC-AUC can quantify the trade-off between true positive rate (TP) and false positive rate (FP) across different thresholds. The TP is defined as the detection rate, i.e. the rate of true anomalous nodes correctly identified as anomalous, whereas the FP is the false alarm rate, i.e. rate of normal nodes identified as anomalous (gutierrez2019multi).
Precision@K and Recall@K: Since we use the ranking strategy to detect anomalies, measures used in ranking-based tasks such as information retrieval and recommender systems can be utilized to evaluate the performance. In specific, we use Precision@K to measure the proportion of true anomalies that an approach discovered in its top K ranked nodes and Recall@K to measure the proportion of true anomalies that a method discovered in the total number of ground truth anomalies.
To demonstrate the effectiveness of our proposed framework in detecting anomalies, we compare the proposed ResGCN model with the following anomaly detection methods:
LOF (breunig2000lof) measures how isolated the object is with respect to the surrounding neighborhood and detects anomalies at the contextual level. LOF only considers nodal attributes.
AMEN (perozzi2016scalable) uses both attribute and network structure information to detect anomalous neighborhoods. Specifically, it analyzes the abnormality of each node from the ego-network point of view.
Radar (li2017radar) is an unsupervised anomaly detection framework for attributed networks. It detects anomalies whose behaviors are singularly different from the majority by characterizing the residuals of attribute information and its coherence with network information.
ANOMALOUS (peng2018anomalous) is a joint anomaly detection framework to optimize attribute selection and anomaly detection using CUR decomposition of matrix and residual analysis on attributed networks.
DOMINANT (ding2019deep) utilizes GCN to learn a low-dimensional embedding representations of the input attributed network and then reconstruct both the topological structure and nodal attributes with these representations. Anomalies are selected by ranking the reconstruction errors.
MADAN (gutierrez2019multi) is a multi-scale anomaly detection method. It uses the heat kernel as filtering operator to exploit the link with the Markov stability to find the context for anomalous nodes at all relevant scales of the network.
In the experiments, for our proposed ResGCN, we propose to optimize the loss function with Adam (kingma2014adam)
algorithm and train the proposed model for 100 epochs. We set the learning rate to 0.01. For representation learning, we use two GCN layers (64-neuron and 32-neuron), and for residual modeling, we use three FC layers (# neuron is equal to # features) to learn the residual matrix, two FC layers (both 64-neuron) to learn the attention weights for the GCN hidden representation and two FC layers (both 32-neuron) to learn the attention weights for the GCN embedding. For these baselines, we use the default parameters used in the original papers.
4.4 Experimental Results
We conduct experiments to evaluate the performance of our proposed model ResGCN by comparing it with several baselines on two different types of networks: networks with and without ground-truth anomaly labels. The experimental results w.r.t. ROC-AUC for networks with ground-truth labels are shown in Table 4. It can be observed from these results:
The proposed ResGCN model outperforms other baseline methods on Amazon data and achieves comparable result on Enron data. It demonstrates the effectiveness of ResGCN.
Deep models such as DOMINANT and residual analysis based methods such as Radar and ANOMALOUS are superior to traditional approaches such as LOF and AMEN. It further validates the effectiveness of deep models and residual modeling.
The experimental results w.r.t. Precision@K and Recall@K for networks without ground-truth labels are shown from Table 5 to Table 10 respectively. From these evaluation results, some conclusions can be drawn:
The proposed ResGCN model outperforms other baseline methods on all three attributed networks except Precision@50 on Flickr. It demonstrates the effectiveness of our method by combining residual modeling and deep representation learning using deep neural networks to detect anomalies.
Superiority of ResGCN to other approaches in Precision@K and Recall@K indicates our proposed model can not only achieve higher detection accuracy but also find more true anomalies within the ranking list of limited length.
Anomaly detection approaches using deep architecture achieve better performance including ResGCN and DOMINANT. It verifies the importance of nonlinearity modeling for anomaly detection on attributed networks.
The residual analysis based models, i.e., Radar and ANOMALOUS, although fail in capturing the nonlineariry of networks, achieve better performance than conventional approaches such as LOF. It demonstrates the rationality of explicit residual modeling in anomaly detection.
|ResGCN (Our Model)||0.710||0.660|
|ResGCN (Our Model)||0.848||0.860||0.670||0.483|
|ResGCN (Our Model)||0.780||0.830||0.875||0.680|
|ResGCN (Our Model)||0.812||0.780||0.675||0.573|
|ResGCN (Our Model)||0.143||0.299||0.456||0.483|
|ResGCN (Our Model)||0.088||0.187||0.393||0.458|
|ResGCN (Our Model)||0.079||0.148||0.235||0.309|
4.5 Ranking Strategy Analysis
One of the advantages of our proposed ResGCN is the deep residual modeling to capture the anomalous information. Therefore, different from DOMINANT (ding2019deep) which ranks the weighted combination of attribute and structure reconstruction errors to select the anomalous nodes, we rank the residual information for anomaly detection. In this section, we compare different ranking strategies for anomaly detection: (1) ranking attribute reconstruction error, (2) ranking structure reconstruction error, (3) ranking the weighted combination of attribute and structure reconstruction errors, and (4) ranking the residual matrix. The first three strategies have been used in (ding2019deep) and the last one has been used in Radar (li2017radar). The results of anomaly detection w.r.t. ROC-AUC on Amazon and Precision@100 and Recall@100 on BlogCatalog are shown in Figure 3.
From the results, it can be observed that:
ranking the residual matrix outperforms other ranking strategies on all the data w.r.t. different evaluation metrics except on Enron data. It demonstrates the effectiveness of residual modeling in ResGCN for anomaly detection.
By combining attribute and structure reconstruction errors, better detection performance can be achieved. This result indicates that both attributes and structures contain some useful information to detect anomalies.
An interesting observation is that attributes play a more important role in detecting anomalies than structures as ranking attribute reconstruction errors performs better than structure construction errors.
4.6 Parameter Analysis
There are different parameters in our proposed ResGCN model. Among them, there are two specific and important ones: (1) the trade-off parameter for structure and attribute reconstruction errors and (2) the residual parameter in the loss function in Eq (17). In this experiment, we investigate the impact of these two parameters separately. Specifically, we test the anomaly detection performance by ranging and from 0.0 to 1.0 on Amazon and BlogCatalog data respectively. The results for and are shown in Figure 4 and Figure 5 respectively.
From the results, it can be observed that:
The influence of shows different trends on different networks. For Amazon, the performance becomes much better when . For BlogCatalog, larger achieves better performance. The commonness is that it achieves the best performance when on both networks.
The impact of is similar on different networks, i.e., both Amazon and BlogCatalog prefer smaller . Empirically the best detection performance can be achieved when on Amazon and on BlogCatalog.
5 Related Work
Anomaly detection is one of the most important research questions in data mining and machine learning. There are different anomalies in different types of data, e.g., text(kannan2017outlier; ruff2019self), network (bhuyan2013network) and temporal data (gupta2013outlier). Earlier studies of anomaly detection on graphs mainly focused on structural anomalies, e.g., (noble2003graph) and (eberle2007discovering). However, compared to anomaly detection approaches on plain networks, anomaly detection on attributed networks is more challenging because both structures and attributes should be taken into consideration. In this section, we concentrate on the related work of anomaly detection on attributed networks.
Real-world networks often come with auxiliary attribute information, so recent years have witnessed an increasingly amount of efforts in detecting anomalies on attributed networks. Existing anomaly detection approaches on attributed networks can be categorized into several different types (ding2019deep): community analysis, subspace selection,residual analysis and deep learning methods.
CODA (gao2010community) focuses on community anomalies by simultaneously finding communities as well as spotting anomalies using a unified probabilistic model. AMEN (perozzi2016scalable) uses both attribute and network structure information to detect anomalous neighborhoods. Radar (li2017radar) detects anomalies whose behaviors are singularly different from the majority by characterizing the residuals of attribute information and its coherence with network information. ANOMALOUS (peng2018anomalous) is a joint anomaly detection framework to optimize attribute selection and anomaly detection using CUR decomposition of matrix and residual analysis on attributed networks. DOMINANT (ding2019deep) utilizes GCN to compress the input attributed network to succinct low-dimensional embedding representations and then reconstruct both the topological structure and nodal attributes with these representations. MADAN (gutierrez2019multi)
is a multi-scale anomaly detection method. It uses the heat kernel as filtering operator to exploit the link with the Markov stability to find the context for outlier nodes at all relevant scales of the network. For traditional anomaly detection methods on graphs, interested readers are referred to(akoglu2015graph) for detailed discussion.
With the popularity of network embedding techniques, which assigns nodes in a network to low-dimensional representations and these representations can effectively preserve the network structure (cui2018survey), learning anomaly aware network representations also attracts huge attentions. Recently, there are several studies taking both problems into consideration to learn anomaly aware network embedding in attributed networks (liang2018semi; zhou2018sparc; bandyopadhyay2019outlier; li2019specae; bandyopadhyay2020outlier). SEANO (liang2018semi) is a semi-supervised network embedding approach which learns a low-dimensional vector representation that systematically captures the topological proximity, attribute affinity and label similarity of nodes. SPARC (zhou2018sparc) is a self-paced framework for anomaly detection which gradually learns the rare category oriented network representation. ONE (bandyopadhyay2019outlier) jointly align and optimize the structures and attributes to generate robust network embeddings by minimizing the effects of outlier nodes. DONE and AdONE (bandyopadhyay2020outlier)
use two parallel autoencoders for link structure and attributes of the nodes respectively. By exploring the reconstruction errors for structures and attributes, the proposed methods can learn embedding and detect anomalies. Another related embedding methods aim to capture the uncertainties of learned representations, such as and DVNE(zhu2018deep) struc2gauss (pei2020struc2gauss)
Another related work is graph convolutional networks (GCNs). The original GCN (kipf2016semi) have been proposed to learn node representations by passing and aggregating messages between neighboring nodes. Different variants extend GCN have been proposed, e.g., introducing attention (velickovic2017graph), adding residual and jumping connections (xu2018representation) and disentangling node representations (ma2019disentangled).
In this paper, we proposed a novel graph convolutional network (GCN) with attention mechanism, ResGCN, on the problem of anomaly detection on attributed networks. ResGCN can effectively address the limitations of previous approaches. On one hand, as GCN handles the high-order node interactions with multiple layers of nonlinear transformations, ResGCN can capture the sparsity and nonlinearity of networks. On the other hand, the attention mechanism based on the explicit deep residual analysis can prevent anomalous nodes from propagating the abnormal information in the message passing process of GCN. Furthermore, ranking the residual information is employed to detect anomalies. The experimental results demonstrate the effectiveness of our proposed ResGCN model compared to state-of-the-art methods. In the future, we would like to investigate the extension of our model in dynamic and streaming networks.