Rows from Many Sources: Enriching row completions from Wikidata with a pre-trained Language Model

04/14/2022
by   Carina Negreanu, et al.
2

Row completion is the task of augmenting a given table of text and numbers with additional, relevant rows. The task divides into two steps: subject suggestion, the task of populating the main column; and gap filling, the task of populating the remaining columns. We present state-of-the-art results for subject suggestion and gap filling measured on a standard benchmark (WikiTables). Our idea is to solve this task by harmoniously combining knowledge base table interpretation and free text generation. We interpret the table using the knowledge base to suggest new rows and generate metadata like headers through property linking. To improve candidate diversity, we synthesize additional rows using free text generation via GPT-3, and crucially, we exploit the metadata we interpret to produce better prompts for text generation. Finally, we verify that the additional synthesized content can be linked to the knowledge base or a trusted web source such as Wikipedia.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/03/2020

Towards Faithful Neural Table-to-Text Generation with Content-Matching Constraints

Text generation from a knowledge base aims to translate knowledge triple...
research
04/29/2020

ToTTo: A Controlled Table-To-Text Generation Dataset

We present ToTTo, an open-domain English table-to-text dataset with over...
research
04/04/2018

Abstractive Tabular Dataset Summarization via Knowledge BaseSemantic Embeddings

This paper describes an abstractive summarization method for tabular dat...
research
04/04/2018

Abstractive Tabular Dataset Summarization via Knowledge Base Semantic Embeddings

This paper describes an abstractive summarization method for tabular dat...
research
08/21/2020

Spatial Language Representation with Multi-Level Geocoding

We present a multi-level geocoding model (MLG) that learns to associate ...
research
02/09/2023

Few-Shot Table-to-Text Generation with Prompt Planning and Knowledge Memorization

Pre-trained language models (PLM) have achieved remarkable advancement i...
research
01/11/2020

PatentTransformer-2: Controlling Patent Text Generation by Structural Metadata

PatentTransformer is our codename for patent text generation based on Tr...

Please sign up or login with your details

Forgot password? Click here to reset