WOVe: Incorporating Word Order in GloVe Word Embeddings

05/18/2021
by   Mohammed Ibrahim, et al.
0

Word vector representations open up new opportunities to extract useful information from unstructured text. Defining a word as a vector made it easy for the machine learning algorithms to understand a text and extract information from. Word vector representations have been used in many applications such word synonyms, word analogy, syntactic parsing, and many others. GloVe, based on word contexts and matrix vectorization, is an ef-fective vector-learning algorithm. It improves on previous vector-learning algorithms. However, the GloVe model fails to explicitly consider the order in which words appear within their contexts. In this paper, multiple methods of incorporating word order in GloVe word embeddings are proposed. Experimental results show that our Word Order Vector (WOVe) word embeddings approach outperforms unmodified GloVe on the natural lan-guage tasks of analogy completion and word similarity. WOVe with direct concatenation slightly outperformed GloVe on the word similarity task, increasing average rank by 2 However, it greatly improved on the GloVe baseline on a word analogy task, achieving an average 36.34

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/06/2016

Neural-based Noise Filtering from Word Embeddings

Word embeddings have been demonstrated to benefit NLP tasks impressively...
research
12/16/2014

Rehabilitation of Count-based Models for Word Vector Representations

Recent works on word representations mostly rely on predictive models. D...
research
05/04/2018

A Rank-Based Similarity Metric for Word Embeddings

Word Embeddings have recently imposed themselves as a standard for repre...
research
08/14/2018

Syntree2Vec - An algorithm to augment syntactic hierarchy into word embeddings

Word embeddings aims to map sense of the words into a lower dimensional ...
research
07/07/2016

Representing Verbs with Rich Contexts: an Evaluation on Verb Similarity

Several studies on sentence processing suggest that the mental lexicon k...
research
10/04/2019

Investigating the Effectiveness of Representations Based on Word-Embeddings in Active Learning for Labelling Text Datasets

Manually labelling large collections of text data is a time-consuming, e...

Please sign up or login with your details

Forgot password? Click here to reset