Natural Alpha Embeddings

12/04/2019
by   Riccardo Volpi, et al.
0

Learning an embedding for a large collection of items is a popular approach to overcome the computational limitations associated to one-hot encodings. The aim of item embedding is to learn a low dimensional space for the representations, able to capture with its geometry relevant features or relationships for the data at hand. This can be achieved for example by exploiting adjacencies among items in large sets of unlabelled data. In this paper we interpret in an Information Geometric framework the item embeddings obtained from conditional models. By exploiting the α-geometry of the exponential family, first introduced by Amari, we introduce a family of natural α-embeddings represented by vectors in the tangent space of the probability simplex, which includes as a special case standard approaches available in the literature. A typical example is given by word embeddings, commonly used in natural language processing, such as Word2Vec and GloVe. In our analysis, we show how the α-deformation parameter can impact on standard evaluation tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2019

word2ket: Space-efficient Word Embeddings inspired by Quantum Entanglement

Deep learning natural language processing models often use vector word e...
research
01/07/2019

On the Dimensionality of Embeddings for Sparse Features and Data

In this note we discuss a common misconception, namely that embeddings a...
research
06/09/2021

Low-Dimensional Structure in the Space of Language Representations is Reflected in Brain Responses

How related are the representations learned by neural language models, t...
research
10/25/2018

Provable Gaussian Embedding with One Observation

The success of machine learning methods heavily relies on having an appr...
research
05/17/2016

Word2Vec is a special case of Kernel Correspondence Analysis and Kernels for Natural Language Processing

We show Correspondence Analysis (CA) is equivalent to defining Gini-inde...
research
08/02/2016

Exponential Family Embeddings

Word embeddings are a powerful approach for capturing semantic similarit...
research
11/15/2020

Link Prediction Using Hebbian Graph Embeddings

Methods and systems for generating link predictions are provided. In one...

Please sign up or login with your details

Forgot password? Click here to reset