Efficient, Simple and Automated Negative Sampling for Knowledge Graph Embedding

10/24/2020
by   Yongqi Zhang, et al.
0

Negative sampling, which samples negative triplets from non-observed ones in knowledge graph (KG), is an essential step in KG embedding. Recently, generative adversarial network (GAN), has been introduced in negative sampling. By sampling negative triplets with large gradients, these methods avoid the problem of vanishing gradient and thus obtain better performance. However, they make the original model more complex and harder to train. In this paper, motivated by the observation that negative triplets with large gradients are important but rare, we propose to directly keep track of them with the cache. In this way, our method acts as a "distilled" version of previous GAN-based methods, which does not waste training time on additional parameters to fit the full distribution of negative triplets. However, how to sample from and update the cache are two critical questions. We propose to solve these issues by automated machine learning techniques. The automated version also covers GAN-based methods as special cases. Theoretical explanation of NSCaching is also provided, justifying the superior over fixed sampling scheme. Besides, we further extend NSCaching with skip-gram model for graph embedding. Finally, extensive experiments show that our method can gain significant improvements on various KG embedding models and the skip-gram model, and outperforms the state-of-the-art negative sampling methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/16/2018

NSCaching: Simple and Efficient Negative Sampling for Knowledge Graph Embedding

Knowledge Graph (KG) embedding is a fundamental problem in data mining r...
research
05/27/2017

word2vec Skip-Gram with Negative Sampling is a Weighted Logistic PCA

We show that the skip-gram formulation of word2vec trained with negative...
research
09/23/2018

Incorporating GAN for Negative Sampling in Knowledge Representation Learning

Knowledge representation learning aims at modeling knowledge graph by en...
research
02/19/2022

MixKG: Mixing for harder negative samples in knowledge graph

Knowledge graph embedding (KGE) aims to represent entities and relations...
research
10/26/2017

Improving Negative Sampling for Word Representation using Self-embedded Features

Although the word-popularity based negative sampler has shown superb per...
research
06/06/2019

Second-order Co-occurrence Sensitivity of Skip-Gram with Negative Sampling

We simulate first- and second-order context overlap and show that Skip-G...
research
06/21/2022

Comprehensive Analysis of Negative Sampling in Knowledge Graph Representation Learning

Negative sampling (NS) loss plays an important role in learning knowledg...

Please sign up or login with your details

Forgot password? Click here to reset