Attention Strategies for Multi-Source Sequence-to-Sequence Learning

04/21/2017
by   Jindřich Libovický, et al.
0

Modeling attention in neural multi-source sequence-to-sequence learning remains a relatively unexplored area, despite its usefulness in tasks that incorporate multiple source languages or modalities. We propose two novel approaches to combine the outputs of attention mechanisms over each source sequence, flat and hierarchical. We compare the proposed methods with existing techniques and present results of systematic evaluation of those methods on the WMT16 Multimodal Translation and Automatic Post-editing tasks. We show that the proposed methods achieve competitive results on both tasks.

READ FULL TEXT
research
06/23/2016

CUNI System for WMT16 Automatic Post-Editing and Multimodal Translation Tasks

Neural sequence to sequence learning recently became a very promising pa...
research
11/12/2018

Input Combination Strategies for Multi-Source Transformer Decoder

In multi-source sequence-to-sequence tasks, the attention mechanism can ...
research
06/13/2017

An Exploration of Neural Sequence-to-Sequence Architectures for Automatic Post-Editing

In this work, we explore multiple neural architectures adapted for the t...
research
06/13/2016

Neural Associative Memory for Dual-Sequence Modeling

Many important NLP problems can be posed as dual-sequence or sequence-to...
research
10/04/2019

Modeling Confidence in Sequence-to-Sequence Models

Recently, significant improvements have been achieved in various natural...
research
04/08/2021

On Biasing Transformer Attention Towards Monotonicity

Many sequence-to-sequence tasks in natural language processing are rough...
research
05/23/2019

Copy this Sentence

Attention is an operation that selects some largest element from some se...

Please sign up or login with your details

Forgot password? Click here to reset