Latent Alignment and Variational Attention

07/10/2018
by   Yuntian Deng, et al.
0

Neural attention has become central to many state-of-the-art models in natural language processing and related domains. Attention networks are an easy-to-train and effective method for softly simulating alignment; however, the approach does not marginalize over latent alignments in a probabilistic sense. This property makes it difficult to compare attention to other alignment approaches, to compose it with probabilistic models, and to perform posterior inference conditioned on observed data. A related latent approach, hard attention, fixes these issues, but is generally harder to train and less accurate. This work considers variational attention networks, alternatives to soft and hard attention for learning latent variable alignment models, with tighter approximation bounds based on amortized variational inference. We further propose methods for reducing the variance of gradients to make these approaches computationally feasible. Experiments show that for machine translation and visual question answering, inefficient exact latent variable models outperform standard neural attention, but these gains go away when using hard attention based training. On the other hand, variational attention retains most of the performance gain but with training speed comparable to neural attention.

READ FULL TEXT
research
12/11/2018

Conditional Variational Autoencoder for Neural Machine Translation

We explore the performance of latent variable models for conditional tex...
research
05/16/2017

Learning Hard Alignments with Variational Inference

There has recently been significant interest in hard attention models fo...
research
10/07/2022

Latent Neural ODEs with Sparse Bayesian Multiple Shooting

Training dynamic models, such as neural ODEs, on long trajectories is a ...
research
09/22/2015

Learning Wake-Sleep Recurrent Attention Models

Despite their success, convolutional neural networks are computationally...
research
09/28/2020

Generative latent neural models for automatic word alignment

Word alignments identify translational correspondences between words in ...
research
06/18/2014

Exact Decoding on Latent Variable Conditional Models is NP-Hard

Latent variable conditional models, including the latent conditional ran...
research
01/18/2018

Overpruning in Variational Bayesian Neural Networks

The motivations for using variational inference (VI) in neural networks ...

Please sign up or login with your details

Forgot password? Click here to reset