Towards Opening the Black Box of Neural Machine Translation: Source and Target Interpretations of the Transformer

05/23/2022
by   Javier Ferrando, et al.
4

In Neural Machine Translation (NMT), each token prediction is conditioned on the source sentence and the target prefix (what has been previously translated at a decoding step). However, previous work on interpretability in NMT has focused solely on source sentence tokens attributions. Therefore, we lack a full understanding of the influences of every input token (source sentence and target prefix) in the model predictions. In this work, we propose an interpretability method that tracks complete input token attributions. Our method, which can be extended to any encoder-decoder Transformer-based model, allows us to better comprehend the inner workings of current NMT models. We apply the proposed method to both bilingual and multilingual Transformers and present insights into their behaviour.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/27/2019

Explicit Sentence Compression for Neural Machine Translation

State-of-the-art Transformer-based neural machine translation (NMT) syst...
research
10/21/2020

Analyzing the Source and Target Contributions to Predictions in Neural Machine Translation

In Neural Machine Translation (and, more generally, conditional language...
research
10/31/2018

You May Not Need Attention

In NMT, how far can we get without attention and without separate encodi...
research
05/21/2023

Explaining How Transformers Use Context to Build Predictions

Language Generation Models produce words based on the previous context. ...
research
12/22/2017

Source-side Prediction for Neural Headline Generation

The encoder-decoder model is widely used in natural language generation ...
research
09/13/2021

Attention Weights in Transformer NMT Fail Aligning Words Between Sequences but Largely Explain Model Predictions

This work proposes an extensive analysis of the Transformer architecture...
research
10/21/2020

Token Drop mechanism for Neural Machine Translation

Neural machine translation with millions of parameters is vulnerable to ...

Please sign up or login with your details

Forgot password? Click here to reset