Few-shot NLG with Pre-trained Language Model

04/21/2019
by   Zhiyu Chen, et al.
0

Natural language generation (NLG) from structured data or knowledge is essential for many NLP research areas. While previous neural-based end-to-end approaches have made significant progress on several benchmarks, their data-hungry nature makes them hard to be widely adopted for real-world applications. Hence, in this work, we propose the new task of few-shot natural language generation. Motivated by how humans tend to summarize tabulated data, we propose a simple yet effective approach and show that it not only demonstrates strong performance but also provides good generalization across domains. Design on model architecture is based on two aspects: content selection/copying from input data, and language modeling to compose coherent sentences, which can be acquired from prior knowledge. Accordingly, we employ a pre-trained domain-independent language model to serve as the prior, while the content selection/copying can be learned with only a few in-domain training instances, thus attaining the few-shot learning objective. To demonstrate that our approach generalizes across domains, we curated table-to-text data from multiple domains. With just 200 training samples, across all domains, our approach outperforms the strongest baseline by an average of over 8.0 BLEU points. We will make our code and data publicly available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/31/2022

WHEN FLUE MEETS FLANG: Benchmarks and Large Pre-trained Language Model for Financial Domain

Pre-trained language models have shown impressive performance on a varie...
research
12/02/2021

LOGEN: Few-shot Logical Knowledge-Conditioned Text Generation with Self-training

Natural language generation from structured data mainly focuses on surfa...
research
08/31/2022

Few-Shot Learning for Clinical Natural Language Processing Using Siamese Neural Networks

Clinical Natural Language Processing (NLP) has become an emerging techno...
research
06/02/2022

Code Generation Tools (Almost) for Free? A Study of Few-Shot, Pre-Trained Language Models on Code

Few-shot learning with large-scale, pre-trained language models is a pow...
research
12/14/2022

MANTa: Efficient Gradient-Based Tokenization for Robust End-to-End Language Modeling

Static subword tokenization algorithms have been an essential component ...
research
10/23/2022

Code4Struct: Code Generation for Few-Shot Structured Prediction from Natural Language

Large Language Model (LLM) trained on the mixture of text and code has d...
research
05/19/2022

Self-augmented Data Selection for Few-shot Dialogue Generation

The natural language generation (NLG) module in task-oriented dialogue s...

Please sign up or login with your details

Forgot password? Click here to reset