A Rank-Based Similarity Metric for Word Embeddings

05/04/2018
by   Enrico Santus, et al.
0

Word Embeddings have recently imposed themselves as a standard for representing word meaning in NLP. Semantic similarity between word pairs has become the most common evaluation benchmark for these representations, with vector cosine being typically used as the only similarity metric. In this paper, we report experiments with a rank-based metric for WE, which performs comparably to vector cosine in similarity estimation and outperforms it in the recently-introduced and challenging task of outlier detection, thus suggesting that rank-based measures can improve clustering quality.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/27/2016

Testing APSyn against Vector Cosine on Similarity Estimation

In Distributional Semantic Models (DSMs), Vector Cosine is widely used t...
research
11/10/2016

Tracing metaphors in time through self-distance in vector spaces

From a diachronic corpus of Italian, we build consecutive vector spaces ...
research
05/19/2019

Correlation Coefficients and Semantic Textual Similarity

A large body of research into semantic textual similarity has focused on...
research
05/18/2021

WOVe: Incorporating Word Order in GloVe Word Embeddings

Word vector representations open up new opportunities to extract useful ...
research
08/22/2019

Unsupervised Lemmatization as Embeddings-Based Word Clustering

We focus on the task of unsupervised lemmatization, i.e. grouping togeth...
research
10/06/2020

Automatic Metaphor Interpretation Using Word Embeddings

We suggest a model for metaphor interpretation using word embeddings tra...
research
09/04/2017

Hypothesis Testing based Intrinsic Evaluation of Word Embeddings

We introduce the cross-match test - an exact, distribution free, high-di...

Please sign up or login with your details

Forgot password? Click here to reset