Rotate King to get Queen: Word Relationships as Orthogonal Transformations in Embedding Space

09/02/2019
by   Kawin Ethayarajh, et al.
0

A notable property of word embeddings is that word relationships can exist as linear substructures in the embedding space. For example, gender corresponds to w⃗o⃗m⃗a⃗n⃗ - m⃗a⃗n⃗ and q⃗u⃗e⃗e⃗n⃗ - k⃗i⃗n⃗g⃗. This, in turn, allows word analogies to be solved arithmetically: k⃗i⃗n⃗g⃗ - m⃗a⃗n⃗ + w⃗o⃗m⃗a⃗n⃗≈q⃗u⃗e⃗e⃗n⃗. This property is notable because it suggests that models trained on word embeddings can easily learn such relationships as geometric translations. However, there is no evidence that models exclusively represent relationships in this manner. We document an alternative way in which downstream models might learn these relationships: orthogonal and linear transformations. For example, given a translation vector for gender, we can find an orthogonal matrix R, representing a rotation and reflection, such that R(k⃗i⃗n⃗g⃗) ≈q⃗u⃗e⃗e⃗n⃗ and R(m⃗a⃗n⃗) ≈w⃗o⃗m⃗a⃗n⃗. Analogical reasoning using orthogonal transformations is almost as accurate as using vector arithmetic; using linear transformations is more accurate than both. Our findings suggest that these transformations can be as good a representation of word relationships as translation vectors.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/14/2017

Rotations and Interpretability of Word Embeddings: the Case of the Russian Language

Consider a continuous word embedding model. Usually, the cosines between...
research
02/13/2017

Offline bilingual word vectors, orthogonal transformations and the inverted softmax

Usually bilingual word vectors are trained "online". Mikolov et al. show...
research
04/20/2020

Learning Geometric Word Meta-Embeddings

We propose a geometric framework for learning meta-embeddings of words f...
research
08/11/2022

Word-Embeddings Distinguish Denominal and Root-Derived Verbs in Semitic

Proponents of the Distributed Morphology framework have posited the exis...
research
03/19/2022

From meaning to perception – exploring the space between word and odor perception embeddings

In this paper we propose the use of the Word2vec algorithm in order to o...
research
10/25/2020

Contextualized Word Embeddings Encode Aspects of Human-Like Word Sense Knowledge

Understanding context-dependent variation in word meanings is a key aspe...
research
10/11/2018

Towards Understanding Linear Word Analogies

A surprising property of word vectors is that vector algebra can often b...

Please sign up or login with your details

Forgot password? Click here to reset