Multilingual Machine Translation Systems from Microsoft for WMT21 Shared Task

11/03/2021
by   Jian Yang, et al.
0

This report describes Microsoft's machine translation systems for the WMT21 shared task on large-scale multilingual machine translation. We participated in all three evaluation tracks including Large Track and two Small Tracks where the former one is unconstrained and the latter two are fully constrained. Our model submissions to the shared task were initialized with DeltaLM[<https://aka.ms/deltalm>], a generic pre-trained multilingual encoder-decoder model, and fine-tuned correspondingly with the vast collected parallel data and allowed data sources according to track settings, together with applying progressive learning and iterative back-translation approaches to further improve the performance. Our final submissions ranked first on three tracks in terms of the automatic evaluation metric.

READ FULL TEXT

page 2

page 9

page 10

research
10/20/2022

The VolcTrans System for WMT22 Multilingual Machine Translation Task

This report describes our VolcTrans system for the WMT22 shared task on ...
research
09/17/2021

Back-translation for Large-Scale Multilingual Machine Translation

This paper illustrates our approach to the shared task on large-scale mu...
research
01/30/2023

KG-BERTScore: Incorporating Knowledge Graph into BERTScore for Reference-Free Machine Translation Evaluation

BERTScore is an effective and robust automatic metric for referencebased...
research
10/21/2022

SIT at MixMT 2022: Fluent Translation Built on Giant Pre-trained Models

This paper describes the Stevens Institute of Technology's submission fo...
research
11/20/2021

Data Processing Matters: SRPH-Konvergen AI's Machine Translation System for WMT'21

In this paper, we describe the submission of the joint Samsung Research ...
research
12/01/2022

CUNI Systems for the WMT22 Czech-Ukrainian Translation Task

We present Charles University submissions to the WMT22 General Translati...
research
05/14/2022

Pretraining Approaches for Spoken Language Recognition: TalTech Submission to the OLR 2021 Challenge

This paper investigates different pretraining approaches to spoken langu...

Please sign up or login with your details

Forgot password? Click here to reset