With a Little Help from the Authors: Reproducing Human Evaluation of an MT Error Detector

08/12/2023
by   Ondřej Plátek, et al.
0

This work presents our efforts to reproduce the results of the human evaluation experiment presented in the paper of Vamvas and Sennrich (2022), which evaluated an automatic system detecting over- and undertranslations (translations containing more or less information than the original) in machine translation (MT) outputs. Despite the high quality of the documentation and code provided by the authors, we discuss some problems we found in reproducing the exact experimental setup and offer recommendations for improving reproducibility. Our replicated results generally confirm the conclusions of the original study, but in some cases, statistically significant differences were observed, suggesting a high variability of human annotation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/01/2021

Detecting over/under-translation errors for determining adequacy in human translations

We present a novel approach to detecting over and under translations (OT...
research
06/03/2022

[Re] Badder Seeds: Reproducing the Evaluation of Lexical Methods for Bias Measurement

Combating bias in NLP requires bias measurement. Bias measurement is alm...
research
10/25/2022

Exploring Document-Level Literary Machine Translation with Parallel Paragraphs from World Literature

Literary translation is a culturally significant task, but it is bottlen...
research
05/29/2023

Improving BIM Authoring Process Reproducibility with Enhanced BIM Logging

This paper presents an enhanced building information modeling (BIM) logg...
research
05/17/2023

Bring More Attention to Syntactic Symmetry for Automatic Postediting of High-Quality Machine Translations

Automatic postediting (APE) is an automated process to refine a given ma...
research
06/17/2022

Automatic Correction of Human Translations

We introduce translation error correction (TEC), the task of automatical...
research
06/27/2019

Findings of the First Shared Task on Machine Translation Robustness

We share the findings of the first shared task on improving robustness o...

Please sign up or login with your details

Forgot password? Click here to reset