Modeling Context With Linear Attention for Scalable Document-Level Translation

10/16/2022
by   Zhaofeng Wu, et al.
0

Document-level machine translation leverages inter-sentence dependencies to produce more coherent and consistent translations. However, these models, predominantly based on transformers, are difficult to scale to long documents as their attention layers have quadratic complexity in the sequence length. Recent efforts on efficient attention improve scalability, but their effect on document translation remains unexplored. In this work, we investigate the efficacy of a recent linear attention model by Peng et al. (2021) on document translation and augment it with a sentential gate to promote a recency inductive bias. We evaluate the model on IWSLT 2015 and OpenSubtitles 2018 against the transformer, demonstrating substantially increased decoding speed on long sequences with similar or better BLEU scores. We show that sentential gating further improves translation quality on IWSLT.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/15/2021

Hierarchical Learning for Generation with Long Source Sequences

One of the challenges for current sequence to sequence (seq2seq) models ...
research
05/31/2021

G-Transformer for Document-level Machine Translation

Document-level MT models are still far from satisfactory. Existing work ...
research
10/29/2019

Big Bidirectional Insertion Representations for Documents

The Insertion Transformer is well suited for long form text generation d...
research
10/18/2020

Capturing Longer Context for Document-level Neural Machine Translation: A Multi-resolutional Approach

Discourse context has been proven useful when translating documents. It ...
research
04/21/2021

On User Interfaces for Large-Scale Document-Level Human Evaluation of Machine Translation Outputs

Recent studies emphasize the need of document context in human evaluatio...
research
01/26/2021

A Comparison of Approaches to Document-level Machine Translation

Document-level machine translation conditions on surrounding sentences t...
research
09/02/2019

Enhancing Context Modeling with a Query-Guided Capsule Network for Document-level Translation

Context modeling is essential to generate coherent and consistent transl...

Please sign up or login with your details

Forgot password? Click here to reset