Revisiting Skip-Gram Negative Sampling Model with Regularization

04/01/2018
by   Cun Mu, et al.
0

We revisit skip-gram negative sampling (SGNS), a popular neural-network based approach to learning distributed word representation. We first point out the ambiguity issue undermining the SGNS model, in the sense that the word vectors can be entirely distorted without changing the objective value. To resolve this issue, we rectify the SGNS model with quadratic regularization. A theoretical justification, which provides a novel insight into quadratic regularization, is presented. Preliminary experiments are also conducted on Google's analytical reasoning task to support the modified SGNS model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/27/2017

word2vec Skip-Gram with Negative Sampling is a Weighted Logistic PCA

We show that the skip-gram formulation of word2vec trained with negative...
research
02/25/2015

Breaking Sticks and Ambiguities with Adaptive Skip-gram

Recently proposed Skip-gram model is a powerful method for learning high...
research
10/26/2017

Improving Negative Sampling for Word Representation using Self-embedded Features

Although the word-popularity based negative sampler has shown superb per...
research
04/13/2017

Incremental Skip-gram Model with Negative Sampling

This paper explores an incremental training strategy for the skip-gram m...
research
06/06/2019

Second-order Co-occurrence Sensitivity of Skip-Gram with Negative Sampling

We simulate first- and second-order context overlap and show that Skip-G...
research
11/20/2014

Linking GloVe with word2vec

The Global Vectors for word representation (GloVe), introduced by Jeffre...
research
12/03/2014

Skip-gram Language Modeling Using Sparse Non-negative Matrix Probability Estimation

We present a novel family of language model (LM) estimation techniques n...

Please sign up or login with your details

Forgot password? Click here to reset