Semantic Noise Matters for Neural Natural Language Generation

11/10/2019
by   Ondřej Dušek, et al.
0

Neural natural language generation (NNLG) systems are known for their pathological outputs, i.e. generating text which is unrelated to the input specification. In this paper, we show the impact of semantic noise on state-of-the-art NNLG models which implement different semantic control mechanisms. We find that cleaned data can improve semantic correctness by up to 97 omitting information, rather than hallucination.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/21/2020

Evaluating Semantic Accuracy of Data-to-Text Generation with Natural Language Inference

A major challenge in evaluating data-to-text (D2T) generation is measuri...
research
02/28/2022

The impact of lexical and grammatical processing on generating code from natural language

Considering the seq2seq architecture of TranX for natural language to co...
research
11/08/2019

A Good Sample is Hard to Find: Noise Injection Sampling and Self-Training for Neural Language Generation Models

Deep neural networks (DNN) are quickly becoming the de facto standard mo...
research
03/12/2022

On Information Hiding in Natural Language Systems

With data privacy becoming more of a necessity than a luxury in today's ...
research
01/25/2022

Semantic Code Classification for Automated Machine Learning

A range of applications for automatic machine learning need the generati...
research
08/30/2023

An xAI Approach for Data-to-Text Processing with ASP

The generation of natural language text from data series gained renewed ...
research
09/14/2019

ALTER: Auxiliary Text Rewriting Tool for Natural Language Generation

In this paper, we describe ALTER, an auxiliary text rewriting tool that ...

Please sign up or login with your details

Forgot password? Click here to reset