The JHU-Microsoft Submission for WMT21 Quality Estimation Shared Task

09/17/2021
by   Shuoyang Ding, et al.
0

This paper presents the JHU-Microsoft joint submission for WMT 2021 quality estimation shared task. We only participate in Task 2 (post-editing effort estimation) of the shared task, focusing on the target-side word-level quality estimation. The techniques we experimented with include Levenshtein Transformer training and data augmentation with a combination of forward, backward, round-trip translation, and pseudo post-editing of the MT output. We demonstrate the competitiveness of our system compared to the widely adopted OpenKiwi-XLM baseline. Our system is also the top-ranking system on the MT MCC metric for the English-German language pair.

READ FULL TEXT
research
10/09/2020

MLQE-PE: A Multilingual Quality Estimation and Post-Editing Dataset

We present MLQE-PE, a new dataset for Machine Translation (MT) Quality E...
research
11/07/2020

AlphaMWE: Construction of Multilingual Parallel Corpora with MWE Annotations

In this work, we present the construction of multilingual parallel corpo...
research
09/12/2021

Levenshtein Training for Word-level Quality Estimation

We propose a novel scheme to use the Levenshtein Transformer to perform ...
research
08/09/2019

UdS Submission for the WMT 19 Automatic Post-Editing Task

In this paper, we describe our submission to the English-German APE shar...
research
06/20/2023

Efficient Machine Translation Corpus Generation

This paper proposes an efficient and semi-automated method for human-in-...
research
10/19/2020

Incorporating Terminology Constraints in Automatic Post-Editing

Users of machine translation (MT) may want to ensure the use of specific...
research
09/01/2018

MS-UEdin Submission to the WMT2018 APE Shared Task: Dual-Source Transformer for Automatic Post-Editing

This paper describes the Microsoft and University of Edinburgh submissio...

Please sign up or login with your details

Forgot password? Click here to reset