Flow-Adapter Architecture for Unsupervised Machine Translation

04/26/2022
by   Yihong Liu, et al.
0

In this work, we propose a flow-adapter architecture for unsupervised NMT. It leverages normalizing flows to explicitly model the distributions of sentence-level latent representations, which are subsequently used in conjunction with the attention mechanism for the translation task. The primary novelties of our model are: (a) capturing language-specific sentence representations separately for each language using normalizing flows and (b) using a simple transformation of these latent representations for translating from one language to another. This architecture allows for unsupervised training of each language independently. While there is prior work on latent variables for supervised MT, to the best of our knowledge, this is the first work that uses latent variables and normalizing flows for unsupervised MT. We obtain competitive results on several unsupervised MT benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/06/2018

Off-the-Shelf Unsupervised NMT

We frame unsupervised machine translation (MT) in the context of multi-t...
research
06/11/2019

Translating Translationese: A Two-Step Approach to Unsupervised Machine Translation

Given a rough, word-by-word gloss of a source language sentence, target ...
research
05/28/2020

Variational Neural Machine Translation with Normalizing Flows

Variational Neural Machine Translation (VNMT) is an attractive framework...
research
10/15/2018

(Self-Attentive) Autoencoder-based Universal Language Representation for Machine Translation

Universal language representation is the holy grail in machine translati...
research
04/23/2022

SIReN-VAE: Leveraging Flows and Amortized Inference for Bayesian Networks

Initial work on variational autoencoders assumed independent latent vari...
research
05/16/2018

Are BLEU and Meaning Representation in Opposition?

One of possible ways of obtaining continuous-space sentence representati...
research
05/23/2018

Amortized Context Vector Inference for Sequence-to-Sequence Networks

Neural attention (NA) is an effective mechanism for inferring complex st...

Please sign up or login with your details

Forgot password? Click here to reset