Mask-Align: Self-Supervised Neural Word Alignment

12/13/2020
by   Chi Chen, et al.
0

Neural word alignment methods have received increasing attention recently. These methods usually extract word alignment from a machine translation model. However, there is a gap between translation and alignment tasks, since the target future context is available in the latter. In this paper, we propose Mask-Align, a self-supervised model specifically designed for the word alignment task. Our model parallelly masks and predicts each target token, and extracts high-quality alignments without any supervised loss. In addition, we introduce leaky attention to alleviate the problem of unexpected high attention weights on special tokens. Experiments on four language pairs show that our model significantly outperforms all existing unsupervised neural baselines and obtains new state-of-the-art results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/09/2022

Cross-Align: Modeling Deep Cross-lingual Interactions for Word Alignment

Word alignment which aims to extract lexicon translation equivalents bet...
research
09/04/2019

Jointly Learning to Align and Translate with Transformer Models

The state of the art in machine translation (MT) is governed by neural a...
research
04/02/2022

Accurate Online Posterior Alignments for Principled Lexically-Constrained Decoding

Online alignment in machine translation refers to the task of aligning a...
research
03/17/2022

Gaussian Multi-head Attention for Simultaneous Machine Translation

Simultaneous machine translation (SiMT) outputs translation while receiv...
research
01/01/2021

Bilingual Lexicon Induction via Unsupervised Bitext Construction and Word Alignment

Bilingual lexicons map words in one language to their translations in an...
research
09/23/2022

Extending Word-Level Quality Estimation for Post-Editing Assistance

We define a novel concept called extended word alignment in order to imp...
research
01/31/2019

Adding Interpretable Attention to Neural Translation Models Improves Word Alignment

Multi-layer models with multiple attention heads per layer provide super...

Please sign up or login with your details

Forgot password? Click here to reset