Underreporting of errors in NLG output, and what to do about it

08/02/2021
by   Emiel van Miltenburg, et al.
0

We observe a severe under-reporting of the different kinds of errors that Natural Language Generation systems make. This is a problem, because mistakes are an important indicator of where systems should still be improved. If authors only report overall performance metrics, the research community is left in the dark about the specific weaknesses that are exhibited by `state-of-the-art' research. Next to quantifying the extent of error under-reporting, this position paper provides recommendations for error identification, analysis and reporting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/22/2022

What and How Are We Reporting in HRI? A Review and Recommendations for Reporting Recruitment, Compensation, and Gender

Study reproducibility and generalizability of results to broadly inclusi...
research
08/06/2018

Crashing Privacy: An Autopsy of a Web Browser's Leaked Crash Reports

Harm to the privacy of users through data leakage is not an unknown issu...
research
02/09/2022

The Absurdity of Death Estimates Based on the Vaccine Adverse Event Reporting System

We demonstrate from first principles a core fallacy employed by a coteri...
research
04/28/2020

Showing Your Work Doesn't Always Work

In natural language processing, a recently popular line of work explores...
research
11/18/2020

Inspecting state of the art performance and NLP metrics in image-based medical report generation

Several deep learning architectures have been proposed over the last yea...
research
01/30/2021

The effect of differential victim crime reporting on predictive policing systems

Police departments around the world have been experimenting with forms o...
research
09/06/2019

Show Your Work: Improved Reporting of Experimental Results

Research in natural language processing proceeds, in part, by demonstrat...

Please sign up or login with your details

Forgot password? Click here to reset