Evaluation method of word embedding by roots and affixes

06/24/2016
by   KeBin Peng, et al.
0

Word embedding has been shown to be remarkably effective in a lot of Natural Language Processing tasks. However, existing models still have a couple of limitations in interpreting the dimensions of word vector. In this paper, we provide a new approach---roots and affixes model(RAAM)---to interpret it from the intrinsic structures of natural language. Also it can be used as an evaluation measure of the quality of word embedding. We introduce the information entropy into our model and divide the dimensions into two categories, just like roots and affixes in lexical semantics. Then considering each category as a whole rather than individually. We experimented with English Wikipedia corpus. Our result show that there is a negative linear relation between the two attributes and a high positive correlation between our model and downstream semantic evaluation tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/28/2019

Evaluating Word Embedding Models: Methods and Experimental Results

Extensive evaluation on a large number of word embedding models for lang...
research
06/06/2017

Synergistic Union of Word2Vec and Lexicon for Domain Specific Semantic Similarity

Semantic similarity measures are an important part in Natural Language P...
research
06/06/2020

Quantum-like Generalization of Complex Word Embedding: a lightweight approach for textual classification

In this paper, we present an extension, and an evaluation, to existing Q...
research
03/13/2023

A Comprehensive Empirical Evaluation of Existing Word Embedding Approaches

Vector-based word representations help countless Natural Language Proces...
research
02/12/2017

Vector Embedding of Wikipedia Concepts and Entities

Using deep learning for different machine learning tasks such as image c...
research
02/22/2017

EVE: Explainable Vector Based Embedding Technique Using Wikipedia

We present an unsupervised explainable word embedding technique, called ...
research
05/12/2022

Noun2Verb: Probabilistic frame semantics for word class conversion

Humans can flexibly extend word usages across different grammatical clas...

Please sign up or login with your details

Forgot password? Click here to reset