SenseFitting: Sense Level Semantic Specialization of Word Embeddings for Word Sense Disambiguation

07/30/2019
by   Manuel Stoeckel, et al.
0

We introduce a neural network-based system of Word Sense Disambiguation (WSD) for German that is based on SenseFitting, a novel method for optimizing WSD. We outperform knowledge-based WSD methods by up to 25 state-of-the-art on the German sense-annotated dataset WebCAGe. Our method uses three feature vectors consisting of a) sense, b) gloss, and c) relational vectors to represent target senses and to compare them with the vector centroids of sample contexts. Utilizing widely available word embeddings and lexical resources, we are able to compensate for the lower resource availability of German. SenseFitting builds upon the recently introduced semantic specialization procedure Attract-Repel, and leverages sense level semantic constraints from lexical-semantic networks (e.g. GermaNet) or online social dictionaries (e.g. Wiktionary) to produce high-quality sense embeddings from pre-trained word embeddings. We evaluate our sense embeddings with a new SimLex-999 based similarity dataset, called SimSense, that we developed for this work. We achieve results that outperform current lemma-based specialization methods for German, making them comparable to results achieved for English.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/10/2017

Making Sense of Word Embeddings

We present a simple yet effective approach for learning word sense embed...
research
04/22/2023

Semantic Specialization for Knowledge-based Word Sense Disambiguation

A promising approach for knowledge-based Word Sense Disambiguation (WSD)...
research
05/14/2019

Sense Vocabulary Compression through the Semantic Knowledge of WordNet for Neural Word Sense Disambiguation

In this article, we tackle the issue of the limited quantity of manually...
research
04/27/2018

An Unsupervised Word Sense Disambiguation System for Under-Resourced Languages

In this paper, we present Watasense, an unsupervised system for word sen...
research
12/14/2019

Integrating Lexical Knowledge in Word Embeddings using Sprinkling and Retrofitting

Neural network based word embeddings, such as Word2Vec and GloVe, are pu...
research
04/02/2019

Using Multi-Sense Vector Embeddings for Reverse Dictionaries

Popular word embedding methods such as word2vec and GloVe assign a singl...
research
02/05/2017

Prepositions in Context

Prepositions are highly polysemous, and their variegated senses encode s...

Please sign up or login with your details

Forgot password? Click here to reset