Towards Data Poisoning Attack against Knowledge Graph Embedding

04/26/2019 ∙ by Hengtong Zhang, et al. ∙ Zhejiang University University at Buffalo 0

Knowledge graph embedding (KGE) is a technique for learning continuous embeddings for entities and relations in the knowledge graph.Due to its benefit to a variety of downstream tasks such as knowledge graph completion, question answering and recommendation, KGE has gained significant attention recently. Despite its effectiveness in a benign environment, KGE' robustness to adversarial attacks is not well-studied. Existing attack methods on graph data cannot be directly applied to attack the embeddings of knowledge graph due to its heterogeneity. To fill this gap, we propose a collection of data poisoning attack strategies, which can effectively manipulate the plausibility of arbitrary targeted facts in a knowledge graph by adding or deleting facts on the graph. The effectiveness and efficiency of the proposed attack strategies are verified by extensive evaluations on two widely-used benchmarks.



There are no comments yet.


page 1

page 2

page 3

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Knowledge graphs have become a critical resource for a large collection of real world applications, such as information extraction [Mintz et al.2009], question answering [Yih et al.2015] and recommendation system [Zhang et al.2016]. Due to its wide application domains, both academia and industry have spent considerable efforts on constructing large-scale knowledge graphs, such as YAGO [Hoffart et al.2013], NELL [Mitchell et al.2015], Freebase [Bollacker et al.2008], and Google Knowledge Graph111 In knowledge graphs, knowledge facts are usually stored as (head entity, relation, tail entity) triples. For instance, the fact triple (Albert Einstein, Profession, Scientist) means that Albert Einstein’s profession is a scientist.

Although such triples can effectively record abundant knowledge, their underlying symbolic nature makes them difficult to be directly fed to many machine learning models. Hence, knowledge graph embedding (KGE), which projects the symbolic entities and relations into continuous vector space, has quickly gained significant attention

[Nickel et al.2011, Lin et al.2015, Bordes et al.2013, Yang et al.2014b, Trouillon et al.2016]. These compact embeddings can preserve the inherent characteristics of entities and relations while enabling the use of these knowledge facts for a large variety of downstream tasks such as link prediction, question answering, and recommendation.

Despite the increasing success and popularity of Knowledge graph embeddings, their robustness has not been fully analyzed. In fact, many knowledge graphs are built upon unreliable or even public data sources. For instance, the well known Freebase harvests its data from various sources including individual, user-submitted wiki contributions222 The openness of such data unfortunately would make KGE vulnerable to malicious attacks. When being attacked, substantial unreliable or even biased knowledge graph embeddings would be generated, leading to serious impairment and financial loss of many downstream applications. Therefore, there is a strong need for the analysis of the vulnerability of knowledge graph embeddings.

In this paper, for the first time, we systemically investigate the vulnerability of KGE, through designing efficient adversarial attack strategies. Due to the unique characteristics of knowledge graph and its embedding models, existing adversarial attack methods on graph data [Zügner et al.2018, Sun et al.2018, Bojcheski and Günnemann2018] cannot be directly applied to attack KGE methods. First, they are all designed for homogeneous graphs, in which there is only a single type of nodes or links. However, in a knowledge graph, both the entities (nodes) and the relations (links) between entities are of different types. Second, existing attack methods for homogeneous graphs usually have strict requirements on the formulation of the targeted methods. For instance, the attack strategies proposed in [Sun et al.2018, Bojcheski and Günnemann2018] can only work for the embedding methods that can be transformed into matrix factorization. However, the KGE methods are diverse and may not be able to be transformed into matrix factorization problems.

In this paper, we introduce the first study on the vulnerability of KGE and propose a family of effective data poisoning attack strategies against KGE methods. Our proposed attack strategies can guide the adversary to manipulate the training set of KGE by adding and/or deleting some specific facts to promote or degrade the plausibility of specific targeted facts, which can potentially influence a large variety of applications that utilize the knowledge graph. The proposed strategies include both direct scheme which directly manipulates the embeddings of entities involved in the targeted facts and indirect scheme which utilizes other entities as proxies to achieve the attack goal. Empirically, we perform poisoning attack experiments against three most representative KGE methods on two common KGE datasets (FB15K, WN18), and verify the effectiveness of the proposed adversarial attack. Results show that the proposed strategies can dramatically worsen the link prediction results of targeted facts with only a small amount of changes to the graph needed.

2 Related Work

Knowledge Graph Embeddings. KGE as an emerging research topic has attracted tremendous interest. A large number of KGE models have been proposed to represent entities and relations in a knowledge graph with vectors or matrices. RESCAL [Nickel et al.2011], which is based on bi-linear matrix factorization, is one of the earliest KGE models. Then [Bordes et al.2013] introduces the first translation-based KGE method TransE. Given a fact ), composed of a relation () and two entities ( and ) in the knowledge graph, TransE learns vector representations of , , and (i.e., h, t and r) by compelling . Later, a large collection of variants, such as TransH [Wang et al.2014], TransR [Lin et al.2015], TransD [Ji et al.2015] and TransA [Xiao et al.2015], extend TransE by projecting the embedding vector into various spaces. On the other hand, DistMult [Yang et al.2014a] simplifies RESCAL by only using a diagonal matrix, and ComPlex [Trouillon et al.2016] extends DistMult into the complex number field. [Wang et al.2017] provides a comprehensive survey on these models. The attack strategy proposed in this paper can be used to attack most of the existing KGE models.

Data Poisoning Attack v.s. Evasion Attack. Data poisoning attacks, such as those in [Biggio et al.2012, Steinhardt et al.2017] are a family of adversarial attacks on machine learning methods. In these works, the attacker can access the training data of the learning algorithm, and has the power to manipulate a fraction of the training data in order to make the trained model meet certain desired objectives. Evasion attacks such a those in [Goodfellow et al.2014, Kurakin et al.2016] are another prevalent type of attack that may be encountered in adversarial settings. In the evasion setting, malicious samples are generated at test time to evade detection. In this paper, the proposed adversarial attack strategies against KGE methods can be categorized into the data poisoning attack setting.

Adversarial Attacks on Graphs There are limited existing works on adversarial attacks for graph learning tasks: node classification [Zügner et al.2018, Dai et al.2018], graph classification [Dai et al.2018], link prediction [Chen et al.2018] and node embedding [Sun et al.2018, Bojcheski and Günnemann2018]. The first work, introduced by [Zügner et al.2018] linearizes the graph convolutional network (GCN) [Kipf and Welling2016]

to derive the closed-form expression for the change in class probabilities for a given edge/feature perturbation and greedily pick the top perturbations that change the class probabilities.

[Dai et al.2018]

proposes a reinforcement learning based approach where the attack agent interacts with the targeted graph/node classifier to learn the policy of selecting the edge perturbations that fool the classifier.

[Chen et al.2018] adopts the fast gradient sign scheme to perform evasion attack against the link prediction task with GCN. [Sun et al.2018] and [Bojcheski and Günnemann2018] propose data poisoning attack

against factorization-based embedding methods on homogeneous graphs. They both formulate the poisoning attack as bi-level optimization problems. The former exploits the eigenvalue perturbation theory 

[Stewart1990], while the latter directly adopts iterative gradient method [Carlini and Wagner2017] to solve the problem. To the best of our knowledge, there is no existing investigation on adversarial attack for heterogeneous graphs, in which the links and/or nodes are of different types, like knowledge graphs. This paper sheds first light on this important problem that has not been studied yet.

3 Data Poisoning Attack against Knowledge Graph Embedding (KGE) Methods

Let us consider a knowledge graph , with a training set denoted as and a targeted fact triple that does not exist in the training set. The goal of the attacker is to manipulate the learned embeddings, which would degrade (or promote) the plausibility of measured by a specific fact plausibility scoring function . Without loss of generality, we focus on degrading the targeted fact. We also assume that the attacker has a limited attacking budget. Formally, the attack task is defined as follows:

Definition 1 (Problem Definition).

Consider a targeted fact triple that does not exist in the training set, we use to denote the embedding of the head entity , to denote the embedding of the tail entity and to denote the embedding of the relation from the original training set. Our task is to minimize the plausibility of , i.e., , by making perturbations (i.e., adding/deleting facts) on the training set. We assume the attacker has a given, fixed budget and is only capable of making perturbations.

Due to the discrete and combinatorial nature of the knowledge graph, solving this problem is highly challenging. Intuitively, in order to manipulate the plausibility of a specific targeted fact, we need to shift either the embedding vectors related to its entities or the embedding vectors/matrices related to its relations. However, in a knowledge graph, the number of facts that a relation type involves is much larger than the number of facts that an entity type involves. For instance, in the well-known knowledge graph Freebase, the number of entities is over 30 million, while the number of relation types is only 1345. This leads to the fact that the innate characteristics of each relation type is far more stable than that of entities and is difficult to be manipulated via a small number of modifications. Hence, in this paper, we focus on manipulating the plausibility of targeted facts from the perspective of entities. To achieve the attack goal, in the rest of this section, we propose a collection of effective yet efficient attack strategies.

3.1 Direct Attack

Given the unpolluted knowledge graph, the goal of direct attack is to determine a collection of perturbations (i.e., fact adding/deleting actions) to shift the embeddings of the entities involved in the targeted fact to minimize the plausibility of the targeted fact. First, we determine the optimal shifting direction that the entity’s embedding should move towards. Then we rank the possible perturbation actions by analyzing the training process of KGE models and designing scoring functions, which estimate the benefit of a perturbation, i.e., how much shifting can be achieved by this perturbation along the desired direction. We name the score as

perturbation benefit score and calculate such score for every possible perturbation. Finally, we conduct the Top- perturbations with highest perturbation benefit scores, where is the attack budget.

Suppose we want to degrade the plausibility of the fact . For simplicity, let’s focus on shifting the embedding of one of the entities in , say head entity , from to . Here, denotes the embedding shifting vector. With the norm of the embedding shifting vector constraint, the fastest direction of decreasing is opposite to its partial derivative with respect to . Let be the perturbation step size, the optimal embedding shifting vector is:


As mentioned in the problem definition, in order to shift by , the adversary is allowed to add perturbation facts to the knowledge graph or delete facts from the knowledge graph. Given the optimal embedding shifting vector , we then find a ranking of the all the perturbation (add or delete) candidates. We discuss the two schemes in detail as follows.

Direct Deleting Attack. Consider the uncontaminated training set, under the direct adversarial attack scheme, in order to shift the embedding of to , we need to select and delete one or more facts that directly involve entity . Intuitively, the fact to delete should have a great influence on the embedding of , while at the same time not hinder the process of shifting the embedding of to . To design a scoring criterion that captures these intuitions, let us look into the training process of KGE model. Consider the specific deletion candidate that involves . During training, the sum of the fact plausibility scores of the observed training samples is maximized. On one hand, the more plausible the fact is, the more it contributes to the final embedding of . Hence, the perturbation benefit score of deleting should be proportional to . On the other hand, if the plausibility of fact is large after is shifted to (i.e., is large), it means that the fact has a great positive impact on the embedding shifting and should not be deleted. Hence, the perturbation benefit score of deleting should be inversely proportional to . Formally, let the set of all the delete candidates be: , which intuitively denote the set of facts that involve as the head entity in the training set. The perturbation benefit score of deleting a specific perturbation fact can be estimated as:


where , , and denote the embeddings of , and , respectively, on the uncontaminated training set.

Direct Adding Attack. Now we discuss how to conduct direct adding perturbation. To shift the embedding of by , we just need to add new facts that involve to make plausible. The set of all the possible adding candidates can be denoted as , where denotes all the possible “relation-tail entity” combinations in the training set and stands for Cartesian product. In practice, for better efficiency, we can sample the possible “relation-tail entity” combinations from a subset of the knowledge graph facts. Formally, the perturbation benefit score of a specific candidate to add (i.e., ) can be estimated as:


, , and denote the embeddings of , and , respectively, on the uncontaminated training set.

3.2 Indirect Attack

Although the direct attack strategy is intuitive and effective, it is possible to be detected by data sanity check. In this section, we move on to introduce a more complicated yet more stealthy adversarial attack scheme, i.e., indirect attack. For indirect attack, instead of adding or deleting the facts that involve the entities in the targeted fact, we propose to perturb the facts that involve other entities in the knowledge graph and let the perturbation effect propagate to the targeted fact. For a better description, we provide the following toy example, which is used throughout this section.

Example 1.

Suppose we want to degrade the plausibility of the targeted fact via shifting the embedding of the targeted entity by , without loss of generality. Under indirect attack scheme, we perturb the facts that involve the K-hop neighbors of . These K-hop neighbors are called proxy entities. Then the entities between the K-hop neighbors (proxy entities) and are intermediate entities to propagate the influence of the perturbations to . The propagation path can be illustrated as follows:

where we use to denote the directional relation and use notation to denote the entities on the path. A specific can work as both the head entity and the tail entity. The notations in the path above are adopted in the rest of this section.

When the perturbations on the proxy entity cause an embedding shift on itself, the embeddings of its neighboring entities will also be influenced. The influence will propagate back to the embedding of the targeted entity ultimately.

However, finding the effective perturbations on the proxy entities, which are K-hop away from the targeted entity, is indeed a challenging task. The task involves two key problems: (1) given a specific propagation path, how can we determine the desired embedding shifting vectors on its intermediate entities and its proxy entity, in order to accomplish the embedding shifting goal on the targeted entity? (2) How do we select the propagation paths to propagate the influence of perturbation to the targeted entity? In the rest of this section, we discuss strategies to solve these key problems and propose a criterion to evaluate the benefit of an indirect perturbation (i.e., the perturbation benefit score).

For the first problem, given a specific path, in order to conduct a perturbation that makes the embedding of shift towards the desired direction (i.e., the direction of ), we decide the shifting goal for each entity on the path in a recurrent way. Suppose we want to shift by via the intermediate entities along the path specified in Example 1. The entity that directly influences is its neighbor and what we need to do is to determine the ideal embedding shifting vector on , so that the desired embedding shift on (i.e., ) is approached to the greatest extent. Formally, should satisfy:


where is the perturbation step size, denotes the embedding of , and denotes the embedding of . As a result, the embedding of will have a larger tendency to move towards than towards , during the training process on the contaminated training data. When is determined, we can further get the embedding shifting vector for , which are denoted as , respectively. This process is similar as above.

With the embedding shifting vectors on the proxy entities of each path determined, we calculate the scores and , defined in Eq. (2) and (3) for all the possible add/delete perturbations. These scores are later used to calculate the perturbation benefit score under indirect attack schemes.

1:Targeted fact , Neighbor hop , Targeted entity .
2:Exhaust all the -hop paths originating from
3:Exhaust all the possible perturbation candidates on the proxy entities of these -hop paths.
4:Calculate for according to Eq. (1).
5:for each path  do
6:     for each intermediate entity / proxy entity in path  do
7:         Calculate according to Eq. (4).
8:     end for
9:     Calculate the score for each perturbation on the current proxy entity according to Eq. (2), Eq. (3) and (5).
10:end for
11:Select perturbations with highest scores (i.e., ) and conduct the attack.
Algorithm 1 Indirect Attack

For the second problem, we look into the training objective function. Suppose we want to shift the embedding of via its neighbor , when the embedding shift on is . To estimate the influence of such embedding shift on , we isolate all the facts that involve in the training objective function, force a embedding shift on and ignore the negative sampling terms. Formally, the objective function becomes:, where stands for the set of all the observed facts, which involve except the fact , in the training set.

denotes the loss function for a single fact.

in indicates that the embedding of is already shifted. Clearly, if we fix the embeddings of all the relations and entities except , the impact of shifting to is highly correlated with the number of facts that involves , i.e., . That is to say, the more neighbors an entity has, the less it will be influenced by a specific perturbation on one of its neighbors.

Based on above discussions, we propose an empirical scoring function to evaluate the perturbation benefit score of every possible perturbation. We still consider the scenario specified in Example 1. Suppose we conduct an add/delete perturbation on the proxy entity . The perturbation benefit score of this indirect perturbation is defined as:


where stands for the maximum number of facts that involves each entity on the path. is the same as under add perturbation scheme and is the same as under delete perturbation scheme. is a trade-off parameter. The first term estimates the direct perturbation benefit of the perturbation in terms of shifting the proxy entity as desired. The second and the third term evaluate the capability of the intermediate entities on the path in terms of propagating the influence to the targeted entity. As the influence may be diluted by the facts that involve each entity on the path. A smaller averaged number of facts that involves each entity on the path indicates a larger capability of the path in terms of propagating the influence. Moreover, we also consider the maximum number of facts that involves each entity on the path. This is to avoid the case when some intermediate entities, whose embedding is difficult to shift, “block” the propagation path. With the perturbation benefit scoring function defined, we select the best perturbations to conduct the attack where denotes the attack budget. The overall workflow of indirect attack is illustrated in Algorithm 1.

3.3 Complexity Analysis

Finally, we report the complexity of the proposed strategies. Suppose the complexity of computing the plausibility score for each fact is and the complexity of computing the shifting vector is . For direct deleting attack scheme, the complexity is , where is the number of all the delete candidates defined in Section 3.1. For the direct adding attack scheme, the complexity is , where is the number of sampled adding candidates defined in Section 3.1. For indirect attack scheme, the computation involves the calculation of the shifting vector for each entity along the path and the calculation of the scores for all the possible perturbations on the entities of the path. Let the number of distinct paths be , the number of hops be , then the complexity is for indirect adding attack and for indirect deleting attack.

4 Experiments

In this section, we evaluate the proposed attack strategies under different settings on two benchmark datasets.

4.1 Datasets & Settings

Datasets: In this paper, we use two common KGE benchmark datasets for our experiment: FB15k and WN18. FB15k is a subset of Freebase, which is a large collaborative knowledge base consisting of a large number of real-world facts. WN18 is a subset of Wordnet 333, which is a large lexical knowledge graph. Both FB15k and WN18 are first introduced by [Bordes et al.2013]. The statistics of these two datasets are shown in Table 1. The training set and the test set of these two datasets are already fixed. We randomly sample 100 samples in the test set as the targeted facts for the proposed attack strategies.

Datasets #Relations #Entities #Train #Targeted Facts
WN18 18 40,943 141442 100
FB15K 1,345 14,951 483,142 100
Table 1: Statistics of the Two KGE Benchmark Datasets

Baseline & Targeted Models: Since there are no existing methods that can work under the setting of this paper, we compare the proposed attack schemes with several naive baseline strategies. Specifically, we design random-dd (random direct deleting), random-da (random direct adding), random-id (random indirect deleting), random-ia (random indirect adding) as comparison baselines for our proposed direct deleting attack, direct adding attack, indirect deleting attack, indirect adding attack, respectively. The difference between the baseline and its corresponding proposed methods is that the perturbation facts to add/delete on the targeted entities are randomly selected.

For the targeted models, we choose the most representative embedding methods TransE [Bordes et al.2013], TransR [Lin et al.2015] and RESCAL [Nickel et al.2011] as attack targets.

Metrics In order to evaluate the effectiveness of the proposed attack strategies. We compare the plausibility change of the targeted fact before and after the adversarial attack. Specifically, we follow the evaluation protocol of KGE models described in the previous works like [Bordes et al.2013]. Given a targeted fact , we remove the head or tail entity and then replace it with all the possible entities. We first compute plausibility scores of those corrupted facts and then rank them by descending order; the rank of the correct entity is stored. After that, we use MRR (Mean Reciprocal Rank) and H@10

(the proportion of correct entities ranked in top 10, for all correct entities.) as our evaluation metrics.

The smaller MRR and H@10 are on the contaminated dataset, the better the attack performance is.

Experiment Settings For the targeted KGE models, we use the standard implementation provided by THUNLP-OpenKE 444 [Han et al.2018]. The embedding dimension is fixed to 50. Other parameters of baseline methods are set according to their authors’ suggestions. For the proposed attack strategies, is fixed to 0.1. The parameter for indirect attack is fixed to 2. The attack models in this paper are all implemented via Python 3.7. The attack models are run on a laptop with 4 GB RAM, 2.7 GHz Intel Core i5 CPU.

4.2 Results and Analysis

In this section, we report and analyze the attack results of the proposed attack strategies under different settings. To avoid confusion, the performance of direct adding attack, direct deleting attack, indirect adding attack, and indirect deleting attack are reported separately in Table 234 and 5.

Clean random-da Direct Add
MRR H@10 MRR H@10 MRR H@10
TransE 0.26 0.49 0.26 0.50 0.23 0.42
FB15K TransR 0.24 0.52 0.25 0.51 0.21 0.41
RESCAL 0.19 0.42 0.20 0.40 0.17 0.39
TransE 0.39 0.70 0.30 0.68 0.21 0.53
WN18 TransR 0.44 0.73 0.41 0.71 0.22 0.51
RESCAL 0.41 0.72 0.44 0.69 0.30 0.57
Table 2: Overall Results of Direct Adding Attack
Clean random-dd Direct Delete
MRR H@10 MRR H@10 MRR H@10
TransE 0.26 0.49 0.26 0.54 0.19 0.37
FB15K TransR 0.24 0.52 0.25 0.49 0.18 0.41
RESCAL 0.19 0.42 0.19 0.38 0.13 0.30
TransE 0.39 0.70 0.36 0.71 0.11 0.26
WN18 TransR 0.44 0.73 0.43 0.68 0.11 0.24
RESCAL 0.41 0.72 0.40 0.67 0.02 0.05
Table 3: Overall Results of Direct Deleting Attack

Overall Attack Performance: Let us first discuss the performances of the direct attack schemes on two datasets. For the direct deleting attack scheme, we set the attack budget for each targeted fact to 4 and 1 on FB15K and WN18 dataset, respectively. For the direct deleting attack scheme, the attack budget for each targeted fact is merely 2 for both datasets. These budgets are low enough to make the whole attack process unnoticeable. From the results, we can clearly see that the plausibilities of these targeted facts significantly degrade as desired. From these results, we can conclude that these KGE models are quite vulnerable to even a small number of perturbations generated by well-designed attack strategies. For comparison, we have also tested the baseline methods random-da and random-dd, which cannot achieve satisfactory attack performances. This demonstrates the effectiveness of the proposed strategies. Moreover, we observe that the effectiveness of the proposed strategies is more significant on WN18 dataset than on FB15K dataset. This is because the average number of facts that each entity involves in WN18 dataset is significantly smaller than that in FB15K dataset. Hence, the graph structure of FB15K is more stable.

Then, let us move on to the discussion of indirect attack schemes. For the indirect adding attack, we set the attack budget for each targeted fact to 60 and 20 for FB15K and WN18 dataset, respectively. For the direct adding attack, the attack budgets for each targeted fact are set to 20 and 5 for FB15K and WN18 dataset, respectively. The reason why indirect attacks need more attack budgets to get comparable results is that only a small portion of the influence caused by the perturbations on proxy entities is propagated to the targeted entity. In contrast, nearly all of the influence of the perturbation is exerted on the targeted entity under direct attack schemes. Like direct attack schemes, these indirect attack schemes also demonstrate their effectiveness. For instance, under the indirect deleting attack scheme, the H@10 and MRR metrics of the targeted facts decrease by approximate 0.03 on FB15K dataset. Thus, the indirect deleting attack schemes can also be used in practices to make the attack process more stealthy.

Clean random-ia Indirect Add
MRR H@10 MRR H@10 MRR H@10
TransE 0.26 0.49 0.25 0.50 0.23 0.47
FB15K TransR 0.24 0.52 0.25 0.51 0.22 0.49
RESCAL 0.19 0.42 0.19 0.40 0.17 0.36
TransE 0.39 0.70 0.42 0.71 0.32 0.67
WN18 TransR 0.44 0.73 0.40 0.73 0.34 0.69
RESCAL 0.41 0.72 0.41 0.69 0.39 0.63
Table 4: Overall Results of Indirect Adding Attack
Clean random-id Indirect Delete
MRR H@10 MRR H@10 MRR H@10
TransE 0.26 0.49 0.27 0.50 0.22 0.44
FB15K TransR 0.24 0.52 0.25 0.53 0.21 0.48
RESCAL 0.19 0.42 0.20 0.36 0.16 0.34
TransE 0.39 0.70 0.44 0.74 0.35 0.68
WN18 TransR 0.44 0.73 0.45 0.74 0.41 0.71
RESCAL 0.41 0.72 0.42 0.70 0.38 0.64
Table 5: Overall Results of Indirect Deleting Attack

Efficiency Analysis: Finally, let us discuss the efficiency of the proposed attack strategies. In Table. 6, we report the time consumption for the proposed attack strategies to generate the perturbations for a single targeted fact on average. The attack budgets are the same as the cases reported in Table 2, 3, 4 and 5.

Direct Add Direct Delete Indirect Add Indirect Delete
Time 3.36s 0.13s 14.04s 1.22s
Table 6: Time Consumption for the Proposed Attack Schemes

From Table. 6, we can see that the proposed model takes less than 15 seconds on average to generate the perturbations for a single targeted fact. For the direct deleting attack scheme, the time cost is less than 1 second on average. These results show that the proposed attack strategies are quite efficient.

5 Conclusions

We present the first study on the vulnerability of existing KGE methods and propose a collection of data poisoning attack strategies for different attack scenarios. These attack strategies can be efficiently computed. Experiment results on two benchmark dataset demonstrate that the proposed strategies can effectively manipulate the plausibility of arbitrary facts in the knowledge graph with limited perturbations. As future work we aim to derive defence strategies for KGE models so that these models are more robust against adversarial attacks.