On the Dimensionality of Embeddings for Sparse Features and Data

01/07/2019
by   Maxim Naumov, et al.
0

In this note we discuss a common misconception, namely that embeddings are always used to reduce the dimensionality of the item space. We show that when we measure dimensionality in terms of information entropy then the embedding of sparse probability distributions, that can be used to represent sparse features or data, may or not reduce the dimensionality of the item space. However, the embeddings do provide a different and often more meaningful representation of the items for a particular task at hand. Also, we give upper bounds and more precise guidelines for choosing the embedding dimension.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/04/2019

Natural Alpha Embeddings

Learning an embedding for a large collection of items is a popular appro...
research
02/13/2023

Sparse Dimensionality Reduction Revisited

The sparse Johnson-Lindenstrauss transform is one of the central techniq...
research
02/26/2020

AutoEmb: Automated Embedding Dimensionality Search in Streaming Recommendations

Deep learning based recommender systems (DLRSs) often have embedding lay...
research
11/17/2015

Learning the Dimensionality of Word Embeddings

We describe a method for learning word embeddings with data-dependent di...
research
08/24/2023

On the Consistency of Average Embeddings for Item Recommendation

A prevalent practice in recommender systems consists in averaging item e...
research
03/26/2018

On the Intrinsic Dimensionality of Face Representation

The two underlying factors that determine the efficacy of face represent...
research
03/01/2018

Pairwise Inner Product Distance: Metric for Functionality, Stability, Dimensionality of Vector Embedding

In this paper, we present a theoretical framework for understanding vect...

Please sign up or login with your details

Forgot password? Click here to reset