Evaluation of Morphological Embeddings for the Russian Language

03/11/2021
by   Vitaly Romanov, et al.
0

A number of morphology-based word embedding models were introduced in recent years. However, their evaluation was mostly limited to English, which is known to be a morphologically simple language. In this paper, we explore whether and to what extent incorporating morphology into word embeddings improves performance on downstream NLP tasks, in the case of morphologically rich Russian language. NLP tasks of our choice are POS tagging, Chunking, and NER – for Russian language, all can be mostly solved using only morphology without understanding the semantics of words. Our experiments show that morphology-based embeddings trained with Skipgram objective do not outperform existing embedding model – FastText. Moreover, a more complex, but morphology unaware model, BERT, allows to achieve significantly greater performance on the tasks that presumably require understanding of a word's morphology.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/11/2021

Evaluation of Morphological Embeddings for English and Russian Languages

This paper evaluates morphology-based embeddings for English and Russian...
research
06/08/2016

A Joint Model for Word Embedding and Word Morphology

This paper presents a joint model for performing unsupervised morphologi...
research
03/02/2019

Predicting and interpreting embeddings for out of vocabulary words in downstream tasks

We propose a novel way to handle out of vocabulary (OOV) words in downst...
research
03/22/2019

LINSPECTOR: Multilingual Probing Tasks for Word Representations

Despite an ever growing number of word representation models introduced ...
research
11/30/2020

Modelling Verbal Morphology in Nen

Nen verbal morphology is remarkably complex; a transitive verb can take ...
research
08/30/2017

Paradigm Completion for Derivational Morphology

The generation of complex derived word forms has been an overlooked prob...
research
11/15/2020

Morphologically Aware Word-Level Translation

We propose a novel morphologically aware probability model for bilingual...

Please sign up or login with your details

Forgot password? Click here to reset