A Correlational Encoder Decoder Architecture for Pivot Based Sequence Generation

06/15/2016
by   Amrita Saha, et al.
0

Interlingua based Machine Translation (MT) aims to encode multiple languages into a common linguistic representation and then decode sentences in multiple target languages from this representation. In this work we explore this idea in the context of neural encoder decoder architectures, albeit on a smaller scale and without MT as the end goal. Specifically, we consider the case of three languages or modalities X, Z and Y wherein we are interested in generating sequences in Y starting from information available in X. However, there is no parallel training data available between X and Y but, training data is available between X & Z and Z & Y (as is often the case in many real world applications). Z thus acts as a pivot/bridge. An obvious solution, which is perhaps less elegant but works very well in practice is to train a two stage model which first converts from X to Z and then from Z to Y. Instead we explore an interlingua inspired solution which jointly learns to do the following (i) encode X and Z to a common representation and (ii) decode Y from this common representation. We evaluate our model on two tasks: (i) bridge transliteration and (ii) bridge captioning. We report promising results in both these applications and believe that this is a right step towards truly interlingua inspired encoder decoder architectures.

READ FULL TEXT
research
07/02/2019

Improving Robustness in Real-World Neural Machine Translation Engines

As a commercial provider of machine translation, we are constantly train...
research
10/13/2015

Bridge Correlational Neural Networks for Multilingual Multimodal Representation Learning

Recently there has been a lot of interest in learning common representat...
research
06/03/2014

Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation

In this paper, we propose a novel neural network model called RNN Encode...
research
04/14/2020

Balancing Training for Multilingual Neural Machine Translation

When training multilingual machine translation (MT) models that can tran...
research
05/09/2023

E2TIMT: Efficient and Effective Modal Adapter for Text Image Machine Translation

Text image machine translation (TIMT) aims to translate texts embedded i...
research
02/25/2022

Screening Gender Transfer in Neural Machine Translation

This paper aims at identifying the information flow in state-of-the-art ...
research
10/15/2018

(Self-Attentive) Autoencoder-based Universal Language Representation for Machine Translation

Universal language representation is the holy grail in machine translati...

Please sign up or login with your details

Forgot password? Click here to reset