Empirical Study of Diachronic Word Embeddings for Scarce Data

09/04/2019
by   Syrielle Montariol, et al.
0

Word meaning change can be inferred from drifts of time-varying word embeddings. However, temporal data may be too sparse to build robust word embeddings and to discriminate significant drifts from noise. In this paper, we compare three models to learn diachronic word embeddings on scarce data: incremental updating of a Skip-Gram from Kim et al. (2014), dynamic filtering from Bamler and Mandt (2017), and dynamic Bernoulli embeddings from Rudolph and Blei (2018). In particular, we study the performance of different initialisation schemes and emphasise what characteristics of each model are more suitable to data scarcity, relying on the distribution of detected drifts. Finally, we regularise the loss of these models to better adapt to scarce data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/22/2019

Learning dynamic word embeddings with drift regularisation

Word usage, meaning and connotation change throughout time. Diachronic w...
research
03/23/2017

Dynamic Bernoulli Embeddings for Language Evolution

Word embeddings are a powerful approach for unsupervised analysis of lan...
research
11/17/2015

Learning the Dimensionality of Word Embeddings

We describe a method for learning word embeddings with data-dependent di...
research
05/29/2020

InfiniteWalk: Deep Network Embeddings as Laplacian Embeddings with a Nonlinearity

The skip-gram model for learning word embeddings (Mikolov et al. 2013) h...
research
04/13/2017

Incremental Skip-gram Model with Negative Sampling

This paper explores an incremental training strategy for the skip-gram m...
research
07/19/2019

Exploring sentence informativeness

This study is a preliminary exploration of the concept of informativenes...
research
04/06/2019

Simple dynamic word embeddings for mapping perceptions in the public sphere

Word embeddings trained on large-scale historical corpora can illuminate...

Please sign up or login with your details

Forgot password? Click here to reset