TransDrift: Modeling Word-Embedding Drift using Transformer

06/16/2022
by   Nishtha Madaan, et al.
0

In modern NLP applications, word embeddings are a crucial backbone that can be readily shared across a number of tasks. However as the text distributions change and word semantics evolve over time, the downstream applications using the embeddings can suffer if the word representations do not conform to the data drift. Thus, maintaining word embeddings to be consistent with the underlying data distribution is a key problem. In this work, we tackle this problem and propose TransDrift, a transformer-based prediction model for word embeddings. Leveraging the flexibility of transformer, our model accurately learns the dynamics of the embedding drift and predicts the future embedding. In experiments, we compare with existing methods and show that our model makes significantly more accurate predictions of the word embedding than the baselines. Crucially, by applying the predicted embeddings as a backbone for downstream classification tasks, we show that our embeddings lead to superior performance compared to the previous methods.

READ FULL TEXT
research
05/12/2016

On the Convergent Properties of Word Embedding Methods

Do word embeddings converge to learn similar things over different initi...
research
04/05/2023

PWESuite: Phonetic Word Embeddings and Tasks They Facilitate

Word embeddings that map words into a fixed-dimensional vector space are...
research
11/29/2019

Deconstructing and reconstructing word embedding algorithms

Uncontextualized word embeddings are reliable feature representations of...
research
04/01/2021

Evaluating Neural Word Embeddings for Sanskrit

Recently, the supervised learning paradigm's surprisingly remarkable per...
research
08/21/2018

Downsampling Strategies are Crucial for Word Embedding Reliability

The reliability of word embeddings algorithms, i.e., their ability to pr...
research
02/24/2022

First is Better Than Last for Training Data Influence

The ability to identify influential training examples enables us to debu...
research
05/06/2021

Learning to Perturb Word Embeddings for Out-of-distribution QA

QA models based on pretrained language mod-els have achieved remarkable ...

Please sign up or login with your details

Forgot password? Click here to reset