Synchronous Bidirectional Inference for Neural Sequence Generation

02/24/2019
by   Jiajun Zhang, et al.
0

In sequence to sequence generation tasks (e.g. machine translation and abstractive summarization), inference is generally performed in a left-to-right manner to produce the result token by token. The neural approaches, such as LSTM and self-attention networks, are now able to make full use of all the predicted history hypotheses from left side during inference, but cannot meanwhile access any future (right side) information and usually generate unbalanced outputs in which left parts are much more accurate than right ones. In this work, we propose a synchronous bidirectional inference model to generate outputs using both left-to-right and right-to-left decoding simultaneously and interactively. First, we introduce a novel beam search algorithm that facilitates synchronous bidirectional decoding. Then, we present the core approach which enables left-to-right and right-to-left decoding to interact with each other, so as to utilize both the history and future predictions simultaneously during inference. We apply the proposed model to both LSTM and self-attention networks. In addition, we propose two strategies for parameter optimization. The extensive experiments on machine translation and abstractive summarization demonstrate that our synchronous bidirectional inference model can achieve remarkable improvements over the strong baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/13/2019

Synchronous Bidirectional Neural Machine Translation

Existing approaches to neural machine translation (NMT) generate the tar...
research
06/23/2019

Sequence Generation: From Both Sides to the Middle

The encoder-decoder framework has achieved promising process for many se...
research
10/07/2021

Beam Search with Bidirectional Strategies for Neural Response Generation

Sequence-to-sequence neural networks have been widely used in language-b...
research
12/15/2021

Mask-combine Decoding and Classification Approach for Punctuation Prediction with real-time Inference Constraints

In this work, we unify several existing decoding strategies for punctuat...
research
08/16/2019

Attending to Future Tokens For Bidirectional Sequence Generation

Neural sequence generation is typically performed token-by-token and lef...
research
04/12/2022

InCoder: A Generative Model for Code Infilling and Synthesis

Code is seldom written in a single left-to-right pass and is instead rep...
research
01/20/2021

Towards the Right Direction in BiDirectional User Interfaces

Hundreds of millions of speakers of bidirectional (BiDi) languages rely ...

Please sign up or login with your details

Forgot password? Click here to reset