ReSeTOX: Re-learning attention weights for toxicity mitigation in machine translation

05/19/2023
by   Javier García Gilabert, et al.
0

Our proposed method, ReSeTOX (REdo SEarch if TOXic), addresses the issue of Neural Machine Translation (NMT) generating translation outputs that contain toxic words not present in the input. The objective is to mitigate the introduction of toxic language without the need for re-training. In the case of identified added toxicity during the inference process, ReSeTOX dynamically adjusts the key-value self-attention weights and re-evaluates the beam search hypotheses. Experimental results demonstrate that ReSeTOX achieves a remarkable 57 quality of 99.5

READ FULL TEXT

page 7

page 19

research
08/10/2017

Neural Machine Translation Leveraging Phrase-based Models in a Hybrid Search

In this paper, we introduce a hybrid search for attention-based neural m...
research
10/03/2016

Learning to Translate in Real-time with Neural Machine Translation

Translating in real-time, a.k.a. simultaneous translation, outputs trans...
research
08/25/2018

Exploring Recombination for Efficient Decoding of Neural Machine Translation

In Neural Machine Translation (NMT), the decoder can capture the feature...
research
05/31/2021

Do Multilingual Neural Machine Translation Models Contain Language Pair Specific Attention Heads?

Recent studies on the analysis of the multilingual representations focus...
research
03/03/2019

Calibration of Encoder Decoder Models for Neural Machine Translation

We study the calibration of several state of the art neural machine tran...
research
11/10/2019

Modelling Bahdanau Attention using Election methods aided by Q-Learning

Neural Machine Translation has lately gained a lot of "attention" with t...
research
07/06/2017

Single-Queue Decoding for Neural Machine Translation

Neural machine translation models rely on the beam search algorithm for ...

Please sign up or login with your details

Forgot password? Click here to reset