Distributed Representations of Words and Phrases and their Compositionality

10/16/2013
by   Tomas Mikolov, et al.
0

The recently introduced continuous Skip-gram model is an efficient method for learning high-quality distributed vector representations that capture a large number of precise syntactic and semantic word relationships. In this paper we present several extensions that improve both the quality of the vectors and the training speed. By subsampling of the frequent words we obtain significant speedup and also learn more regular word representations. We also describe a simple alternative to the hierarchical softmax called negative sampling. An inherent limitation of word representations is their indifference to word order and their inability to represent idiomatic phrases. For example, the meanings of "Canada" and "Air" cannot be easily combined to obtain "Air Canada". Motivated by this example, we present a simple method for finding phrases in text, and show that learning good vector representations for millions of phrases is possible.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/18/2015

"The Sum of Its Parts": Joint Learning of Word and Phrase Representations with Autoencoders

Recently, there has been a lot of effort to represent words in continuou...
research
02/25/2015

Breaking Sticks and Ambiguities with Adaptive Skip-gram

Recently proposed Skip-gram model is a powerful method for learning high...
research
07/21/2016

Exploring phrase-compositionality in skip-gram models

In this paper, we introduce a variation of the skip-gram model which joi...
research
12/16/2014

Rehabilitation of Count-based Models for Word Vector Representations

Recent works on word representations mostly rely on predictive models. D...
research
12/10/2019

An Ensemble Method for Producing Word Representations for the Greek Language

In this paper we present a new ensemble method, Continuous Bag-of-Skip-g...
research
08/02/2016

Semantic Representations of Word Senses and Concepts

Representing the semantics of linguistic items in a machine-interpretabl...
research
11/20/2014

Linking GloVe with word2vec

The Global Vectors for word representation (GloVe), introduced by Jeffre...

Please sign up or login with your details

Forgot password? Click here to reset