Evaluating Word Embedding Hyper-Parameters for Similarity and Analogy Tasks

04/11/2018
by   Maryam Fanaeepour, et al.
0

The versatility of word embeddings for various applications is attracting researchers from various fields. However, the impact of hyper-parameters when training embedding model is often poorly understood. How much do hyper-parameters such as vector dimensions and corpus size affect the quality of embeddings, and how do these results translate to downstream applications? Using standard embedding evaluation metrics and datasets, we conduct a study to empirically measure the impact of these hyper-parameters.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/12/2016

On the Convergent Properties of Word Embedding Methods

Do word embeddings converge to learn similar things over different initi...
research
06/19/2019

Considerations for the Interpretation of Bias Measures of Word Embeddings

Word embedding spaces are powerful tools for capturing latent semantic r...
research
04/19/2017

Redefining Context Windows for Word Embedding Models: An Experimental Study

Distributional semantic models learn vector representations of words thr...
research
08/26/2022

Multi-objective Hyper-parameter Optimization of Behavioral Song Embeddings

Song embeddings are a key component of most music recommendation engines...
research
11/06/2021

What augmentations are sensitive to hyper-parameters and why?

We apply augmentations to our dataset to enhance the quality of our pred...
research
05/03/2017

On the effectiveness of feature set augmentation using clusters of word embeddings

Word clusters have been empirically shown to offer important performance...
research
08/30/2019

Single Training Dimension Selection for Word Embedding with PCA

In this paper, we present a fast and reliable method based on PCA to sel...

Please sign up or login with your details

Forgot password? Click here to reset