Rank-One Editing of Encoder-Decoder Models

11/23/2022
by   Vikas Raunak, et al.
0

Large sequence to sequence models for tasks such as Neural Machine Translation (NMT) are usually trained over hundreds of millions of samples. However, training is just the origin of a model's life-cycle. Real-world deployments of models require further behavioral adaptations as new requirements emerge or shortcomings become known. Typically, in the space of model behaviors, behavior deletion requests are addressed through model retrainings whereas model finetuning is done to address behavior addition requests, both procedures being instances of data-based model intervention. In this work, we present a preliminary study investigating rank-one editing as a direct intervention method for behavior deletion requests in encoder-decoder transformer models. We propose four editing tasks for NMT and show that the proposed editing algorithm achieves high efficacy, while requiring only a single instance of positive example to fix an erroneous (negative) model behavior.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/17/2019

Hard but Robust, Easy but Sensitive: How Encoder and Decoder Perform in Neural Machine Translation

Neural machine translation (NMT) typically adopts the encoder-decoder fr...
research
02/27/2020

Echo State Neural Machine Translation

We present neural machine translation (NMT) models inspired by echo stat...
research
03/03/2019

Calibration of Encoder Decoder Models for Neural Machine Translation

We study the calibration of several state of the art neural machine tran...
research
05/30/2019

Unbabel's Submission to the WMT2019 APE Shared Task: BERT-based Encoder-Decoder for Automatic Post-Editing

This paper describes Unbabel's submission to the WMT2019 APE Shared Task...
research
09/13/2021

Attention Weights in Transformer NMT Fail Aligning Words Between Sequences but Largely Explain Model Predictions

This work proposes an extensive analysis of the Transformer architecture...
research
04/01/2020

Editable Neural Networks

These days deep neural networks are ubiquitously used in a wide range of...
research
12/17/2021

Transcribing Natural Languages for The Deaf via Neural Editing Programs

This work studies the task of glossification, of which the aim is to em ...

Please sign up or login with your details

Forgot password? Click here to reset