Generation Challenges: Results of the Accuracy Evaluation Shared Task

08/12/2021
by   Craig Thomson, et al.
0

The Shared Task on Evaluating Accuracy focused on techniques (both manual and automatic) for evaluating the factual accuracy of texts produced by neural NLG systems, in a sports-reporting domain. Four teams submitted evaluation techniques for this task, using very different approaches and techniques. The best-performing submissions did encouragingly well at this difficult task. However, all automatic submissions struggled to detect factual errors which are semantically or pragmatically complex (for example, based on incorrect computation or inference).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2020

Shared Task on Evaluating Accuracy in Natural Language Generation

We propose a shared task on methodologies and algorithms for evaluating ...
research
09/21/2023

SemEval-2022 Task 7: Identifying Plausible Clarifications of Implicit and Underspecified Phrases in Instructional Texts

We describe SemEval-2022 Task 7, a shared task on rating the plausibilit...
research
04/25/2021

SemEval-2021 Task 6: Detection of Persuasion Techniques in Texts and Images

We describe SemEval-2021 task 6 on Detection of Persuasion Techniques in...
research
05/13/2022

LSCDiscovery: A shared task on semantic change discovery and detection in Spanish

We present the first shared task on semantic change discovery and detect...
research
11/10/2022

CREATIVESUMM: Shared Task on Automatic Summarization for Creative Writing

This paper introduces the shared task of summarizing documents in severa...
research
06/10/2019

Learning to combine Grammatical Error Corrections

The field of Grammatical Error Correction (GEC) has produced various sys...
research
10/29/2021

Overview of ADoBo 2021: Automatic Detection of Unassimilated Borrowings in the Spanish Press

This paper summarizes the main findings of the ADoBo 2021 shared task, p...

Please sign up or login with your details

Forgot password? Click here to reset