TURL: Table Understanding through Representation Learning

06/26/2020
by   Xiang Deng, et al.
0

Relational tables on the Web store a vast amount of knowledge. Owing to the wealth of such tables, there has been tremendous progress on a variety of tasks in the area of table understanding. However, existing work generally relies on heavily-engineered task specific features and model architectures. In this paper, we present TURL, a novel framework that introduces the pre-training/finetuning paradigm to relational Web tables. During pre-training, our framework learns deep contextualized representations on relational tables in an unsupervised manner. Its universal model design with pre-trained representations can be applied to a wide range of tasks with minimal task-specific fine-tuning. Specifically, we propose a structure-aware Transformer encoder to model the row-column structure of relational tables, and present a new Masked Entity Recovery (MER) objective for pre-training to capture the semantics and knowledge in large-scale unlabeled data. We systematically evaluate TURL with a benchmark consisting of 6 different tasks for table understanding (e.g., relation extraction, cell filling). We show that TURL generalizes well to all tasks and substantially outperforms existing methods in almost all instances.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/21/2020

Structure-aware Pre-training for Table Understanding with Tree-based Transformers

Tables are widely used with various structures to organize and present d...
research
12/08/2021

Improving Knowledge Graph Representation Learning by Structure Contextual Pre-training

Representation learning models for Knowledge Graphs (KG) have proven to ...
research
06/14/2021

GitTables: A Large-Scale Corpus of Relational Tables

The practical success of deep learning has sparked interest in improving...
research
02/04/2023

REaLTabFormer: Generating Realistic Relational and Tabular Data using Transformers

Tabular data is a common form of organizing data. Multiple models are av...
research
11/18/2021

SDCUP: Schema Dependency-Enhanced Curriculum Pre-Training for Table Semantic Parsing

Recently pre-training models have significantly improved the performance...
research
05/05/2022

Relational Representation Learning in Visually-Rich Documents

Relational understanding is critical for a number of visually-rich docum...
research
08/24/2021

Relation Extraction from Tables using Artificially Generated Metadata

Relation Extraction (RE) from tables is the task of identifying relation...

Please sign up or login with your details

Forgot password? Click here to reset