DeepAI AI Chat
Log In Sign Up

Facebook FAIR's WMT19 News Translation Task Submission

07/15/2019
by   Nathan Ng, et al.
0

This paper describes Facebook FAIR's submission to the WMT19 shared news translation task. We participate in two language pairs and four language directions, English <-> German and English <-> Russian. Following our submission from last year, our baseline systems are large BPE-based transformer models trained with the Fairseq sequence modeling toolkit which rely on sampled back-translations. This year we experiment with different bitext data filtering schemes, as well as with adding filtered back-translated data. We also ensemble and fine-tune our models on domain-specific data, then decode using noisy channel model reranking. Our submissions are ranked first in all four directions of the human evaluation campaign. On En->De, our system significantly outperforms other systems as well as human translations. This system improves upon our WMT'18 submission by 4.5 BLEU points.

READ FULL TEXT

page 1

page 2

page 3

page 4

10/15/2019

Facebook AI's WAT19 Myanmar-English Translation Task Submission

This paper describes Facebook AI's submission to the WAT 2019 Myanmar-En...
08/06/2021

Facebook AI WMT21 News Translation Task Submission

We describe Facebook's multilingual model submission to the WMT2021 shar...
06/09/2016

Edinburgh Neural Machine Translation Systems for WMT 16

We participated in the WMT 2016 shared news translation task by building...
10/29/2020

Tilde at WMT 2020: News Task Systems

This paper describes Tilde's submission to the WMT2020 shared task on ne...
10/10/2017

Confidence through Attention

Attention distributions of the generated translations are a useful bi-pr...