A Comparison of Word Embeddings for English and Cross-Lingual Chinese Word Sense Disambiguation

11/09/2016
by   Hong Jin Kang, et al.
0

Word embeddings are now ubiquitous forms of word representation in natural language processing. There have been applications of word embeddings for monolingual word sense disambiguation (WSD) in English, but few comparisons have been done. This paper attempts to bridge that gap by examining popular embeddings for the task of monolingual English WSD. Our simplified method leads to comparable state-of-the-art performance without expensive retraining. Cross-Lingual WSD - where the word senses of a word in a source language e come from a separate target translation language f - can also assist in language learning; for example, when providing translations of target vocabulary for learners. Thus we have also applied word embeddings to the novel task of cross-lingual WSD for Chinese and provide a public dataset for further benchmarking. We have also experimented with using word embeddings for LSTM networks and found surprisingly that a basic LSTM network does not work well. We discuss the ramifications of this outcome.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/04/2018

Neural Networks for Cross-lingual Negation Scope Detection

Negation scope has been annotated in several English and Chinese corpora...
research
10/16/2019

Meemi: A Simple Method for Post-processing Cross-lingual Word Embeddings

Word embeddings have become a standard resource in the toolset of any Na...
research
09/15/2018

CLUSE: Cross-Lingual Unsupervised Sense Embeddings

This paper proposes a modularized sense induction and representation lea...
research
03/04/2018

Concatenated p-mean Word Embeddings as Universal Cross-Lingual Sentence Representations

Average word embeddings are a common baseline for more sophisticated sen...
research
04/10/2023

On Evaluation of Bangla Word Analogies

This paper presents a high-quality dataset for evaluating the quality of...
research
11/22/2019

Multilingual Culture-Independent Word Analogy Datasets

In text processing, deep neural networks mostly use word embeddings as a...
research
09/21/2021

How Familiar Does That Sound? Cross-Lingual Representational Similarity Analysis of Acoustic Word Embeddings

How do neural networks "perceive" speech sounds from unknown languages? ...

Please sign up or login with your details

Forgot password? Click here to reset