Local Homology of Word Embeddings

10/24/2018
by   Tadas Temčinas, et al.
0

Topological data analysis (TDA) has been widely used to make progress on a number of problems. However, it seems that TDA application in natural language processing (NLP) is at its infancy. In this paper we try to bridge the gap by arguing why TDA tools are a natural choice when it comes to analysing word embedding data. We describe a parallelisable unsupervised learning algorithm based on local homology of datapoints and show some experimental results on word embedding data. We see that local homology of datapoints in word embedding data contains some information that can potentially be used to solve the word sense disambiguation problem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/28/2019

Evaluating Word Embedding Models: Methods and Experimental Results

Extensive evaluation on a large number of word embedding models for lang...
research
07/05/2018

A Review of Different Word Embeddings for Sentiment Classification using Deep Learning

The web is loaded with textual content, and Natural Language Processing ...
research
01/31/2021

Introduction of a novel word embedding approach based on technology labels extracted from patent data

Diversity in patent language is growing and makes finding synonyms for c...
research
06/10/2023

Universal Language Modelling agent

Large Language Models are designed to understand complex Human Language....
research
02/27/2020

The Spectral Underpinning of word2vec

word2vec due to Mikolov et al. (2013) is a word embedding method that is...
research
10/25/2018

Word Embedding based Edit Distance

Text similarity calculation is a fundamental problem in natural language...
research
03/13/2023

A Comprehensive Empirical Evaluation of Existing Word Embedding Approaches

Vector-based word representations help countless Natural Language Proces...

Please sign up or login with your details

Forgot password? Click here to reset