Cued@wmt19:ewc&lms

06/11/2019
by   Felix Stahlberg, et al.
0

Two techniques provide the fabric of the Cambridge University Engineering Department's (CUED) entry to the WMT19 evaluation campaign: elastic weight consolidation (EWC) and different forms of language modelling (LMs). We report substantial gains by fine-tuning very strong baselines on former WMT test sets using a combination of checkpoint averaging and EWC. A sentence-level Transformer LM and a document-level LM based on a modified Transformer architecture yield further gains. As in previous years, we also extract n-gram probabilities from SMT lattices which can be seen as a source-conditioned n-gram LM.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/14/2019

Microsoft Translator at WMT 2019: Towards Large-Scale Document-Level Neural Machine Translation

This paper describes the Microsoft Translator submissions to the WMT19 n...
research
11/22/2019

Improving N-gram Language Models with Pre-trained Deep Transformer

Although n-gram language models (LMs) have been outperformed by the stat...
research
06/29/2019

The CUED's Grammatical Error Correction Systems for BEA-2019

We describe two entries from the Cambridge University Engineering Depart...
research
01/14/2020

Faster Transformer Decoding: N-gram Masked Self-Attention

Motivated by the fact that most of the information relevant to the predi...
research
05/23/2022

Simple Recurrence Improves Masked Language Models

In this work, we explore whether modeling recurrence into the Transforme...
research
12/09/2010

MUDOS-NG: Multi-document Summaries Using N-gram Graphs (Tech Report)

This report describes the MUDOS-NG summarization system, which applies a...
research
04/13/2021

Can a Transformer Pass the Wug Test? Tuning Copying Bias in Neural Morphological Inflection Models

Deep learning sequence models have been successfully applied to the task...

Please sign up or login with your details

Forgot password? Click here to reset