Category Enhanced Word Embedding

11/27/2015
by   Chunting Zhou, et al.
0

Distributed word representations have been demonstrated to be effective in capturing semantic and syntactic regularities. Unsupervised representation learning from large unlabeled corpora can learn similar representations for those words that present similar co-occurrence statistics. Besides local occurrence statistics, global topical information is also important knowledge that may help discriminate a word from another. In this paper, we incorporate category information of documents in the learning of word representations and to learn the proposed models in a document-wise manner. Our models outperform several state-of-the-art models in word analogy and word similarity tasks. Moreover, we evaluate the learned word vectors on sentiment analysis and text classification tasks, which shows the superiority of our learned word vectors. We also learn high-quality category embeddings that reflect topical meanings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/14/2017

Modeling Semantic Relatedness using Global Relation Vectors

Word embedding models such as GloVe rely on co-occurrence statistics fro...
research
10/09/2019

Word Embedding Visualization Via Dictionary Learning

Co-occurrence statistics based word embedding techniques have proved to ...
research
02/23/2019

Vector of Locally-Aggregated Word Embeddings (VLAWE): A novel document-level embedding

In this paper, we propose a novel representation for text documents base...
research
09/24/2015

Bilingual Distributed Word Representations from Document-Aligned Comparable Data

We propose a new model for learning bilingual word representations from ...
research
11/22/2015

On the Linear Algebraic Structure of Distributed Word Representations

In this work, we leverage the linear algebraic structure of distributed ...
research
11/21/2016

Unsupervised Learning for Lexicon-Based Classification

In lexicon-based classification, documents are assigned labels by compar...
research
11/11/2017

Interpretable probabilistic embeddings: bridging the gap between topic models and neural networks

We consider probabilistic topic models and more recent word embedding te...

Please sign up or login with your details

Forgot password? Click here to reset