Contrastive Loss is All You Need to Recover Analogies as Parallel Lines

06/14/2023
by   Narutatsu Ri, et al.
0

While static word embedding models are known to represent linguistic analogies as parallel lines in high-dimensional space, the underlying mechanism as to why they result in such geometric structures remains obscure. We find that an elementary contrastive-style method employed over distributional information performs competitively with popular word embedding models on analogy recovery tasks, while achieving dramatic speedups in training time. Further, we demonstrate that a contrastive loss is sufficient to create these parallel structures in word embeddings, and establish a precise relationship between the co-occurrence statistics and the geometric structure of the resulting word embeddings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/08/2019

Deconstructing Word Embeddings

A review of Word Embedding Models through a deconstructive approach reve...
research
11/29/2019

Deconstructing and reconstructing word embedding algorithms

Uncontextualized word embeddings are reliable feature representations of...
research
07/21/2017

Mimicking Word Embeddings using Subword RNNs

Word embeddings improve generalization over lexical features by placing ...
research
09/05/2020

Bio-inspired Structure Identification in Language Embeddings

Word embeddings are a popular way to improve downstream performances in ...
research
05/13/2023

Frequency-aware Dimension Selection for Static Word Embedding by Mixed Product Distance

Static word embedding is still useful, particularly for context-unavaila...
research
11/01/2021

Interpretable contrastive word mover's embedding

This paper shows that a popular approach to the supervised embedding of ...
research
04/27/2022

Extremal GloVe: Theoretically Accurate Distributed Word Embedding by Tail Inference

Distributed word embeddings such as Word2Vec and GloVe have been widely ...

Please sign up or login with your details

Forgot password? Click here to reset