Can Network Embedding of Distributional Thesaurus be Combined with Word Vectors for Better Representation?

02/17/2018
by   Abhik Jana, et al.
0

Distributed representations of words learned from text have proved to be successful in various natural language processing tasks in recent times. While some methods represent words as vectors computed from text using predictive model (Word2vec) or dense count based model (GloVe), others attempt to represent these in a distributional thesaurus network structure where the neighborhood of a word is a set of words having adequate context overlap. Being motivated by recent surge of research in network embedding techniques (DeepWalk, LINE, node2vec etc.), we turn a distributional thesaurus network into dense word vectors and investigate the usefulness of distributional thesaurus embedding in improving overall word representation. This is the first attempt where we show that combining the proposed word representation obtained by distributional thesaurus embedding with the state-of-the-art word representations helps in improving the performance by a significant margin when evaluated against NLP tasks like word similarity and relatedness, synonym detection, analogy detection. Additionally, we show that even without using any handcrafted lexical resources we can come up with representations having comparable performance in the word similarity and relatedness tasks compared to the representations where a lexical resource has been used.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/18/2014

A Simple and Efficient Method To Generate Word Sense Representations

Distributed representations of words have boosted the performance of man...
research
02/06/2018

A Neurobiologically Motivated Analysis of Distributional Semantic Models

The pervasive use of distributional semantic models or word embeddings i...
research
05/31/2017

Are distributional representations ready for the real world? Evaluating word vectors for grounded perceptual meaning

Distributional word representation methods exploit word co-occurrences t...
research
05/25/2016

Integrating Distributional Lexical Contrast into Word Embeddings for Antonym-Synonym Distinction

We propose a novel vector representation that integrates lexical contras...
research
11/05/2019

Sparse Lifting of Dense Vectors: Unifying Word and Sentence Representations

As the first step in automated natural language processing, representing...
research
12/23/2017

A Framework for Enriching Lexical Semantic Resources with Distributional Semantics

We present an approach to combining distributional semantic representati...
research
11/26/2015

OntoSeg: a Novel Approach to Text Segmentation using Ontological Similarity

Text segmentation (TS) aims at dividing long text into coherent segments...

Please sign up or login with your details

Forgot password? Click here to reset