Scalable Learning of Latent Language Structure With Logical Offline Cycle Consistency

05/31/2023
by   Maxwell Crouse, et al.
0

We introduce Logical Offline Cycle Consistency Optimization (LOCCO), a scalable, semi-supervised method for training a neural semantic parser. Conceptually, LOCCO can be viewed as a form of self-learning where the semantic parser being trained is used to generate annotations for unlabeled text that are then used as new supervision. To increase the quality of annotations, our method utilizes a count-based prior over valid formal meaning representations and a cycle-consistency score produced by a neural text generation model as additional signals. Both the prior and semantic parser are updated in an alternate fashion from full passes over the training data, which can be seen as approximating the marginalization of latent structures through stochastic variational inference. The use of a count-based prior, frozen text generation model, and offline annotation process yields an approach with negligible complexity and latency increases as compared to conventional self-learning. As an added bonus, the annotations produced by LOCCO can be trivially repurposed to train a neural text generation model. We demonstrate the utility of LOCCO on the well-known WebNLG benchmark where we obtain an improvement of 2 points against a self-learning parser under equivalent conditions, an improvement of 1.3 points against the previous state-of-the-art parser, and competitive text generation performance in terms of BLEU score.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/06/2021

Neural Data-to-Text Generation with LM-based Text Augmentation

For many new application domains for data-to-text generation, the main o...
research
04/25/2019

Neural Text Generation from Rich Semantic Representations

We propose neural models to generate high-quality text from structured r...
research
12/03/2019

AMR-to-Text Generation with Cache Transition Systems

Text generation from AMR involves emitting sentences that reflect the me...
research
12/02/2021

LOGEN: Few-shot Logical Knowledge-Conditioned Text Generation with Self-training

Natural language generation from structured data mainly focuses on surfa...
research
07/31/2017

The Code2Text Challenge: Text Generation in Source Code Libraries

We propose a new shared task for tactical data-to-text generation in the...
research
07/02/2021

Scarecrow: A Framework for Scrutinizing Machine Text

Modern neural text generation systems can produce remarkably fluent and ...
research
09/07/2023

Chasing Consistency in Text-to-3D Generation from a Single Image

Text-to-3D generation from a single-view image is a popular but challeng...

Please sign up or login with your details

Forgot password? Click here to reset