Beyond Word Embeddings: Learning Entity and Concept Representations from Large Scale Knowledge Bases

01/01/2018
by   Walid Shalaby, et al.
0

Text representation using neural word embeddings has proven efficacy in many NLP applications. Recently, a lot of research interest goes beyond word embeddings by adapting the traditional word embedding models to learn vectors of multiword expressions (concepts/entities). However, current methods are limited to textual knowledge bases only (e.g., Wikipedia). In this paper, we propose a novel approach for learning concept vectors from two large scale knowledge bases (Wikipedia, and Probase). We adapt the skip-gram model to seamlessly learn from the knowledge in Wikipedia text and Probase concept graph. We evaluate our concept embedding models intrinsically on two tasks: 1) analogical reasoning where we achieve a state-of-the-art performance of 91 semantic analogies, 2) concept categorization where we achieve a state-of-the-art performance on two benchmark datasets achieving categorization accuracy of 100 study to extrinsically evaluate our model on unsupervised argument type identification for neural semantic parsing. We demonstrate the competitive accuracy of our unsupervised method and its ability to better generalize to out of vocabulary entity mentions compared to the tedious and error prone methods which depend on gazetteers and regular expressions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/08/2018

Exploration on Grounded Word Embedding: Matching Words and Images with Image-Enhanced Skip-Gram Model

Word embedding is designed to represent the semantic meaning of a word w...
research
10/11/2019

Finding Interpretable Concept Spaces in Node Embeddings using Knowledge Bases

In this paper we propose and study the novel problem of explaining node ...
research
07/11/2021

Document Embedding for Scientific Articles: Efficacy of Word Embeddings vs TFIDF

Over the last few years, neural network derived word embeddings became p...
research
02/10/2017

Learning Concept Embeddings for Efficient Bag-of-Concepts Densification

Explicit concept space models have proven efficacy for text representati...
research
05/04/2021

Large-scale Taxonomy Induction Using Entity and Word Embeddings

Taxonomies are an important ingredient of knowledge organization, and se...
research
02/22/2017

EVE: Explainable Vector Based Embedding Technique Using Wikipedia

We present an unsupervised explainable word embedding technique, called ...
research
12/10/2015

Measuring Semantic Relatedness using Mined Semantic Analysis

Mined Semantic Analysis (MSA) is a novel concept space model which emplo...

Please sign up or login with your details

Forgot password? Click here to reset