Non-Autoregressive Translation with Layer-Wise Prediction and Deep Supervision

10/14/2021
by   Chenyang Huang, et al.
10

How do we perform efficient inference while retaining high translation quality? Existing neural machine translation models, such as Transformer, achieve high performance, but they decode words one by one, which is inefficient. Recent non-autoregressive translation models speed up the inference, but their quality is still inferior. In this work, we propose DSLP, a highly efficient and high-performance model for machine translation. The key insight is to train a non-autoregressive Transformer with Deep Supervision and feed additional Layer-wise Predictions. We conducted extensive experiments on four translation tasks (both directions of WMT'14 EN-DE and WMT'16 EN-RO). Results show that our approach consistently improves the BLEU scores compared with respective base models. Specifically, our best variant outperforms the autoregressive model on three translation tasks, while being 14.8 times more efficient in inference.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/14/2023

RenewNAT: Renewing Potential Translation for Non-Autoregressive Transformer

Non-autoregressive neural machine translation (NAT) models are proposed ...
research
06/18/2020

Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation

State-of-the-art neural machine translation models generate outputs auto...
research
01/27/2023

Candidate Soups: Fusing Candidate Results Improves Translation Quality for Non-Autoregressive Translation

Non-autoregressive translation (NAT) model achieves a much faster infere...
research
10/19/2022

Hybrid-Regressive Neural Machine Translation

In this work, we empirically confirm that non-autoregressive translation...
research
10/12/2022

Non-Autoregressive Machine Translation with Translation Memories

Non-autoregressive machine translation (NAT) has recently made great pro...
research
05/30/2018

Marian: Cost-effective High-Quality Neural Machine Translation in C++

This paper describes the submissions of the "Marian" team to the WNMT 20...
research
12/03/2022

The RoyalFlush System for the WMT 2022 Efficiency Task

This paper describes the submission of the RoyalFlush neural machine tra...

Please sign up or login with your details

Forgot password? Click here to reset