Text-to-Text Pre-Training for Data-to-Text Tasks

05/21/2020
by   Mihir Kale, et al.
0

We study the pre-train + fine-tune strategy for data-to-text tasks. Fine-tuning T5 achieves state-of-the-art results on the WebNLG, MultiWoz and ToTTo benchmarks. Moreover, the models are fully end-to-end and do not rely on any intermediate planning steps, delexicalization or copy mechanisms. T5 pre-training also enables stringer generalization, as evidenced by large improvements on out-of-domain test sets. We hope our work serves as a useful baseline for future research, as pre-training becomes ever more prevalent for data-to-text tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/14/2019

How to Fine-Tune BERT for Text Classification?

Language model pre-training has proven to be useful in learning universa...
research
08/15/2023

Handwritten Stenography Recognition and the LION Dataset

Purpose: In this paper, we establish a baseline for handwritten stenogra...
research
10/02/2020

Data Transfer Approaches to Improve Seq-to-Seq Retrosynthesis

Retrosynthesis is a problem to infer reactant compounds to synthesize a ...
research
09/11/2023

Examining the Effect of Pre-training on Time Series Classification

Although the pre-training followed by fine-tuning paradigm is used exten...
research
05/10/2022

UNITS: Unsupervised Intermediate Training Stage for Scene Text Detection

Recent scene text detection methods are almost based on deep learning an...
research
06/04/2019

Color Constancy Convolutional Autoencoder

In this paper, we study the importance of pre-training for the generaliz...
research
07/04/2022

Explore Faster Localization Learning For Scene Text Detection

Generally pre-training and long-time training computation are necessary ...

Please sign up or login with your details

Forgot password? Click here to reset