Saliency-driven Word Alignment Interpretation for Neural Machine Translation

06/25/2019
by   Shuoyang Ding, et al.
0

Despite their original goal to jointly learn to align and translate, Neural Machine Translation (NMT) models, especially Transformer, are often perceived as not learning interpretable word alignments. In this paper, we show that NMT models do learn interpretable word alignments, which could only be revealed with proper interpretation methods. We propose a series of such methods that are model-agnostic, are able to be applied either offline or online, and do not require parameter update or architectural change. We show that under the force decoding setup, the alignments induced by our interpretation method are of better quality than fast-align for some systems, and when performing free decoding, they agree well with the alignments induced by automatic alignment tools.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/19/2016

Modeling Coverage for Neural Machine Translation

Attention mechanism has enhanced state-of-the-art Neural Machine Transla...
research
11/26/2019

Neural Machine Translation with Explicit Phrase Alignment

While neural machine translation (NMT) has achieved state-of-the-art tra...
research
04/26/2022

When do Contrastive Word Alignments Improve Many-to-many Neural Machine Translation?

Word alignment has proven to benefit many-to-many neural machine transla...
research
01/27/2022

Learning How to Translate North Korean through South Korean

South and North Korea both use the Korean language. However, Korean NLP ...
research
09/04/2019

Jointly Learning to Align and Translate with Transformer Models

The state of the art in machine translation (MT) is governed by neural a...
research
03/31/2021

Leveraging Neural Machine Translation for Word Alignment

The most common tools for word-alignment rely on a large amount of paral...
research
01/31/2019

Adding Interpretable Attention to Neural Translation Models Improves Word Alignment

Multi-layer models with multiple attention heads per layer provide super...

Please sign up or login with your details

Forgot password? Click here to reset