Gaussian Word Embedding with a Wasserstein Distance Loss

08/21/2018
by   Chi Sun, et al.
0

Comparing with word embedding that based on the point representation, distribution-based word embedding shows more flexibility on expressing uncertainty and therefore, embeds a richer semantic information when representing words. While the Wasserstein distance provides a natural notion of dissimilarity with probability measures and has a closed form solution when measuring the distance between two Gaussian distributions. Therefore, with the aim of representing words in a high-efficient way, we propose to operate the Gaussian word embedding model with the loss function which is based on the Wasserstein distance. In addition, external information which is drawn from ConceptNet will be used to semi-supervise the results of the Gaussian word embedding. Thirteen datasets from the word similarity task, together with one from the word entailment task and six datasets from the downstream document classification task will be evaluated in this paper to testify our hypothesis.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/01/2020

Attention Word Embedding

Word embedding models learn semantically rich vector representations of ...
research
10/03/2016

Grounding the Lexical Sets of Causative-Inchoative Verbs with Word Embedding

Lexical sets contain the words filling the argument positions of a verb ...
research
11/12/2019

Learning Multi-Sense Word Distributions using Approximate Kullback-Leibler Divergence

Learning word representations has garnered greater attention in the rece...
research
09/21/2023

An Efficient Consolidation of Word Embedding and Deep Learning Techniques for Classifying Anticancer Peptides: FastText+BiLSTM

Anticancer peptides (ACPs) are a group of peptides that exhibite antineo...
research
06/20/2016

Uncertainty in Neural Network Word Embedding: Exploration of Threshold for Similarity

Word embedding, specially with its recent developments, promises a quant...
research
11/05/2015

Comparing Writing Styles using Word Embedding and Dynamic Time Warping

The development of plot or story in novels is reflected in the content a...
research
04/27/2017

Multimodal Word Distributions

Word embeddings provide point representations of words containing useful...

Please sign up or login with your details

Forgot password? Click here to reset