Generative latent neural models for automatic word alignment

09/28/2020
by   Anh Khoa Ngo Ho, et al.
0

Word alignments identify translational correspondences between words in a parallel sentence pair and are used, for instance, to learn bilingual dictionaries, to train statistical machine translation systems or to perform quality estimation. Variational autoencoders have been recently used in various of natural language processing to learn in an unsupervised way latent representations that are useful for language generation tasks. In this paper, we study these models for the task of word alignment and propose and assess several evolutions of a vanilla variational autoencoders. We demonstrate that these techniques can yield competitive results as compared to Giza++ and to a strong neural network alignment system for two language pairs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/28/2020

Neural Baselines for Word Alignment

Word alignments identify translational correspondences between words in ...
research
04/30/2020

End-to-End Neural Word Alignment Outperforms GIZA++

Word alignment was once a core unsupervised learning task in natural lan...
research
09/28/2017

A Deep Neural Network Approach To Parallel Sentence Extraction

Parallel sentence extraction is a task addressing the data sparsity prob...
research
02/16/2018

Deep Generative Model for Joint Alignment and Word Representation

This work exploits translation data as a source of semantically relevant...
research
01/14/2016

Smoothing parameter estimation framework for IBM word alignment models

IBM models are very important word alignment models in Machine Translati...
research
07/10/2018

Latent Alignment and Variational Attention

Neural attention has become central to many state-of-the-art models in n...
research
01/23/2023

Noisy Parallel Data Alignment

An ongoing challenge in current natural language processing is how its m...

Please sign up or login with your details

Forgot password? Click here to reset