Logical Natural Language Generation from Open-Domain Tables

04/22/2020
by   Wenhu Chen, et al.
3

Neural natural language generation (NLG) models have recently shown remarkable progress in fluency and coherence. However, existing studies on neural NLG are primarily focused on surface-level realizations with limited emphasis on logical inference, an important aspect of human thinking and language. In this paper, we suggest a new NLG task where a model is tasked with generating natural language statements that can be logically entailed by the facts in an open-domain semi-structured table. To facilitate the study of the proposed logical NLG problem, we use the existing TabFact dataset <cit.> featured with a wide range of logical/symbolic inferences as our testbed, and propose new automatic metrics to evaluate the fidelity of generation models w.r.t. logical inference. The new task poses challenges to the existing monotonic generation frameworks due to the mismatch between sequence order and logical order. In our experiments, we comprehensively survey different generation architectures (LSTM, Transformer, Pre-Trained LM) trained with different algorithms (RL, Adversarial Training, Coarse-to-Fine) on the dataset and made following observations: 1) Pre-Trained LM can significantly boost both the fluency and logical fidelity metrics, 2) RL and Adversarial Training are trading fluency for fidelity, 3) Coarse-to-Fine generation can help partially alleviate the fidelity issue while maintaining high language fluency. The code and data are available at <https://github.com/wenhuchen/LogicNLG>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/30/2020

Logic2Text: High-Fidelity Natural Language Generation from Logical Forms

Previous works on Natural Language Generation (NLG) from structured data...
research
05/25/2022

PLOG: Table-to-Logic Pretraining for Logical Table-to-Text Generation

Logical table-to-text generation is a task that involves generating logi...
research
12/12/2021

Improving Logical-Level Natural Language Generation with Topic-Conditioned Data Augmentation and Logical Form Generation

Logical Natural Language Generation, i.e., generating textual descriptio...
research
05/23/2023

Generating Data for Symbolic Language with Large Language Models

While large language models (LLMs) bring not only performance but also c...
research
08/18/2023

BioMedGPT: Open Multimodal Generative Pre-trained Transformer for BioMedicine

Foundation models (FMs) have exhibited remarkable performance across a w...
research
06/11/2020

Teaching Pre-Trained Models to Systematically Reason Over Implicit Knowledge

To what extent can a neural network systematically reason over symbolic ...
research
04/18/2021

Flexible Operations for Natural Language Deduction

An interpretable system for complex, open-domain reasoning needs an inte...

Please sign up or login with your details

Forgot password? Click here to reset