Bilingual Learning of Multi-sense Embeddings with Discrete Autoencoders

03/30/2016
by   Simon Šuster, et al.
0

We present an approach to learning multi-sense word embeddings relying both on monolingual and bilingual information. Our model consists of an encoder, which uses monolingual and bilingual context (i.e. a parallel sentence) to choose a sense for a given word, and a decoder which predicts context words based on the chosen sense. The two components are estimated jointly. We observe that the word representations induced from bilingual data outperform the monolingual counterparts across a range of evaluation tasks, even though crosslingual information is not available at test time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/10/2017

Making Sense of Word Embeddings

We present a simple yet effective approach for learning word sense embed...
research
06/25/2017

Beyond Bilingual: Multi-sense Word Embeddings using Multilingual Context

Word embeddings, which represent a word as a point in a vector space, ha...
research
06/15/2017

A Mixture Model for Learning Multi-Sense Word Embeddings

Word embeddings are now a standard technique for inducing meaning repres...
research
01/06/2017

Real Multi-Sense or Pseudo Multi-Sense: An Approach to Improve Word Representation

Previous researches have shown that learning multiple representations fo...
research
09/15/2018

CLUSE: Cross-Lingual Unsupervised Sense Embeddings

This paper proposes a modularized sense induction and representation lea...
research
03/12/2023

MWE as WSD: Solving Multiword Expression Identification with Word Sense Disambiguation

Recent work in word sense disambiguation (WSD) utilizes encodings of the...
research
12/14/2018

Detecting Reliable Novel Word Senses: A Network-Centric Approach

In this era of Big Data, due to expeditious exchange of information on t...

Please sign up or login with your details

Forgot password? Click here to reset