Recurrently Controlled Recurrent Networks

11/24/2018
by   Yi Tay, et al.
0

Recurrent neural networks (RNNs) such as long short-term memory and gated recurrent units are pivotal building blocks across a broad spectrum of sequence modeling problems. This paper proposes a recurrently controlled recurrent network (RCRN) for expressive and powerful sequence encoding. More concretely, the key idea behind our approach is to learn the recurrent gating functions using recurrent networks. Our architecture is split into two components - a controller cell and a listener cell whereby the recurrent controller actively influences the compositionality of the listener cell. We conduct extensive experiments on a myriad of tasks in the NLP domain such as sentiment analysis (SST, IMDb, Amazon reviews, etc.), question classification (TREC), entailment classification (SNLI, SciTail), answer selection (WikiQA, TrecQA) and reading comprehension (NarrativeQA). Across all 26 datasets, our results demonstrate that RCRN not only consistently outperforms BiLSTMs but also stacked BiLSTMs, suggesting that our controller architecture might be a suitable replacement for the widely adopted stacked architecture.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/28/2020

Learning Various Length Dependence by Dual Recurrent Neural Networks

Recurrent neural networks (RNNs) are widely used as a memory model for s...
research
01/25/2016

Long Short-Term Memory-Networks for Machine Reading

In this paper we address the question of how to render sequence-level ne...
research
10/06/2018

Co-Stack Residual Affinity Networks with Multi-level Attention Refinement for Matching Text Sequences

Learning a matching function between two text sequences is a long standi...
research
11/14/2019

Contextual Recurrent Units for Cloze-style Reading Comprehension

Recurrent Neural Networks (RNN) are known as powerful models for handlin...
research
01/03/2018

Gated Recurrent Networks for Seizure Detection

Recurrent Neural Networks (RNNs) with sophisticated units that implement...
research
01/03/2017

Shortcut Sequence Tagging

Deep stacked RNNs are usually hard to train. Adding shortcut connections...
research
03/24/2018

Multi-range Reasoning for Machine Comprehension

We propose MRU (Multi-Range Reasoning Units), a new fast compositional e...

Please sign up or login with your details

Forgot password? Click here to reset