Contextual Recurrent Neural Networks

02/09/2019
by   Sam Wenke, et al.
0

There is an implicit assumption that by unfolding recurrent neural networks (RNN) in finite time, the misspecification of choosing a zero value for the initial hidden state is mitigated by later time steps. This assumption has been shown to work in practice and alternative initialization may be suggested but often overlooked. In this paper, we propose a method of parameterizing the initial hidden state of an RNN. The resulting architecture, referred to as a Contextual RNN, can be trained end-to-end. The performance on an associative retrieval task is found to improve by conditioning the RNN initial hidden state on contextual information from the input sequence. Furthermore, we propose a novel method of conditionally generating sequences using the hidden state parameterization of Contextual RNN.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/20/2013

How to Construct Deep Recurrent Neural Networks

In this paper, we explore different ways to extend a recurrent neural ne...
research
10/03/2020

Cubic Spline Smoothing Compensation for Irregularly Sampled Sequences

The marriage of recurrent neural networks and neural ordinary differenti...
research
07/28/2023

Dynamic Analysis and an Eigen Initializer for Recurrent Neural Networks

In recurrent neural networks, learning long-term dependency is the main ...
research
08/29/2016

Optimizing Recurrent Neural Networks Architectures under Time Constraints

Recurrent neural network (RNN)'s architecture is a key factor influencin...
research
06/19/2023

Grammatical gender in Swedish is predictable using recurrent neural networks

The grammatical gender of Swedish nouns is a mystery. While there are fe...
research
06/29/2020

Incremental Training of a Recurrent Neural Network Exploiting a Multi-Scale Dynamic Memory

The effectiveness of recurrent neural networks can be largely influenced...
research
08/17/2022

Expressivity of Hidden Markov Chains vs. Recurrent Neural Networks from a system theoretic viewpoint

Hidden Markov Chains (HMC) and Recurrent Neural Networks (RNN) are two w...

Please sign up or login with your details

Forgot password? Click here to reset