Word2rate: training and evaluating multiple word embeddings as statistical transitions

04/16/2021
by   Gary Phua, et al.
0

Using pretrained word embeddings has been shown to be a very effective way in improving the performance of natural language processing tasks. In fact almost any natural language tasks that can be thought of has been improved by these pretrained embeddings. These tasks range from sentiment analysis, translation, sequence prediction amongst many others. One of the most successful word embeddings is the Word2vec CBOW model proposed by Mikolov trained by the negative sampling technique. Mai et al. modifies this objective to train CMOW embeddings that are sensitive to word order. We used a modified version of the negative sampling objective for our context words, modelling the context embeddings as a Taylor series of rate matrices. We show that different modes of the Taylor series produce different types of embeddings. We compare these embeddings to their similar counterparts like CBOW and CMOW and show that they achieve comparable performance. We also introduce a novel left-right context split objective that improves performance for tasks sensitive to word order. Our Word2rate model is grounded in a statistical foundation using rate matrices while being competitive in variety of language tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/17/2019

KaWAT: A Word Analogy Task Dataset for Indonesian

We introduced KaWAT (Kata Word Analogy Task), a new word analogy task da...
research
10/24/2020

Word Embeddings for Chemical Patent Natural Language Processing

We evaluate chemical patent word embeddings against known biomedical emb...
research
04/21/2018

Context-Attentive Embeddings for Improved Sentence Representations

While one of the first steps in many NLP systems is selecting what embed...
research
11/10/2017

Bayesian Paragraph Vectors

Word2vec (Mikolov et al., 2013) has proven to be successful in natural l...
research
04/11/2018

Word2Vec applied to Recommendation: Hyperparameters Matter

Skip-gram with negative sampling, a popular variant of Word2vec original...
research
08/01/2017

Learned in Translation: Contextualized Word Vectors

Computer vision has benefited from initializing multiple deep layers wit...
research
01/17/2023

Word Embeddings as Statistical Estimators

Word embeddings are a fundamental tool in natural language processing. C...

Please sign up or login with your details

Forgot password? Click here to reset