A Bayesian approach to translators' reliability assessment

03/14/2022
by   Marco Miccheli, et al.
0

Translation Quality Assessment (TQA) conducted by human translators is a process widely used, both in estimating the increasingly used Machine Translation performance and in finding an agreement between customers and translation providers in translation industry. While translation scholars are aware about the importance of having a reliable way to conduct the TQA process, it seems that there is limited literature facing the issue of reliability with a quantitative approach. Here we consider the TQA as a complex process, considering it from the physics of complex systems point of view, and we face the reliability issue with a Bayesian approach. Using a dataset of translation quality evaluations, in an error annotation setting, entirely produced by the Language Service Provider Translated Srl, we build two Bayesian models that parameterise the features involved in the TQA process, namely the translation difficulty, the characteristics of the translators involved in producing the translation and assessing its quality (reviewers). After validating the models in an unsupervised setting, showing that it is possible to get meaningful insights about translators even with just one review per translation job, we extract information about the translators and reviewers and we show that reviewers reliability cannot be taken for granted even if they are expert translators: the translator's expertise could induce also a cognitive bias when reviewing a translation produced by another translator. The most expert translators, though, show the highest level of consistency, both in the task of translating and in the one of assessing translation quality.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/15/2021

Measuring Uncertainty in Translation Quality Evaluation (TQE)

From both human translators (HT) and machine translation (MT) researcher...
research
09/22/2022

Approaching English-Polish Machine Translation Quality Assessment with Neural-based Methods

This paper presents our contribution to the PolEval 2021 Task 2: Evaluat...
research
05/05/2021

Translation Quality Assessment: A Brief Survey on Manual and Automatic Methods

To facilitate effective translation modeling and translation studies, on...
research
03/24/2023

Error Analysis Prompting Enables Human-Like Translation Evaluation in Large Language Models: A Case Study on ChatGPT

Generative large language models (LLMs), e.g., ChatGPT, have demonstrate...
research
07/07/2016

A Maturity Model for Public Administration as Open Translation Data Providers

Any public administration that produces translation data can be a provid...
research
07/25/2018

"Bilingual Expert" Can Find Translation Errors

Recent advances in statistical machine translation via the adoption of n...
research
03/10/2022

A new approach to calculating BERTScore for automatic assessment of translation quality

The study of the applicability of the BERTScore metric was conducted to ...

Please sign up or login with your details

Forgot password? Click here to reset