DeepAI AI Chat
Log In Sign Up

WikiDataSets : Standardized sub-graphs from WikiData

by   Armand Boschin, et al.
Télécom Paris

Developing new ideas and algorithms in the fields of graph processing and relational learning requires datasets to work with and WikiData is the largest open source knowledge graph involving more than fifty millions entities. It is larger than needed in many cases and even too large to be processed easily but it is still a goldmine of relevant facts and subgraphs. Using this graph is time consuming and prone to task specific tuning which can affect reproducibility of results. Providing a unified framework to extract topic-specific subgraphs solves this problem and allows researchers to evaluate algorithms on common datasets. This paper presents various topic-specific subgraphs of WikiData along with the generic Python code used to extract them. These datasets can help develop new methods of knowledge graph processing and relational learning.


page 1

page 2

page 3

page 4


MIRA: Multihop Relation Prediction in Temporal Knowledge Graphs

In knowledge graph reasoning, we observe a trend to analyze temporal dat...

EventKG+BT: Generation of Interactive Biography Timelines from a Knowledge Graph

Research on notable accomplishments and important events in the life of ...

Efficient Enumeration of Subgraphs and Induced Subgraphs with Bounded Girth

The girth of a graph is the length of its shortest cycle. Due to its rel...

Graph-augmented Learning to Rank for Querying Large-scale Knowledge Graph

Knowledge graph question answering (i.e., KGQA) based on information ret...

A Dataset for Hyper-Relational Extraction and a Cube-Filling Approach

Relation extraction has the potential for large-scale knowledge graph co...

Towards a GML-Enabled Knowledge Graph Platform

This vision paper proposes KGNet, an on-demand graph machine learning (G...