Morphological Word Embeddings

07/04/2019
by   Ryan Cotterell, et al.
0

Linguistic similarity is multi-faceted. For instance, two words may be similar with respect to semantics, syntax, or morphology inter alia. Continuous word-embeddings have been shown to capture most of these shades of similarity to some degree. This work considers guiding word-embeddings with morphologically annotated data, a form of semi-supervised learning, encouraging the vectors to encode a word's morphology, i.e., words close in the embedded space share morphological features. We extend the log-bilinear model to this end and show that indeed our learned embeddings achieve this, using German as a case study.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/06/2017

The Interplay of Semantics and Morphology in Word Embeddings

We explore the ability of word embeddings to capture both semantic and m...
research
09/06/2018

Uncovering divergent linguistic information in word embeddings with lessons for intrinsic and extrinsic evaluation

Following the recent success of word embeddings, it has been argued that...
research
08/22/2019

ViCo: Word Embeddings from Visual Co-occurrences

We propose to learn word embeddings from visual co-occurrences. Two word...
research
04/10/2017

Exploring Word Embeddings for Unsupervised Textual User-Generated Content Normalization

Text normalization techniques based on rules, lexicons or supervised tra...
research
04/24/2017

A Trie-Structured Bayesian Model for Unsupervised Morphological Segmentation

In this paper, we introduce a trie-structured Bayesian model for unsuper...
research
02/25/2022

Morphology Without Borders: Clause-Level Morphological Annotation

Morphological tasks use large multi-lingual datasets that organize words...
research
06/06/2019

Derivational Morphological Relations in Word Embeddings

Derivation is a type of a word-formation process which creates new words...

Please sign up or login with your details

Forgot password? Click here to reset