Optimizing and Contrasting Recurrent Neural Network Architectures

10/16/2015
by   Ben Krause, et al.
0

Recurrent Neural Networks (RNNs) have long been recognized for their potential to model complex time series. However, it remains to be determined what optimization techniques and recurrent architectures can be used to best realize this potential. The experiments presented take a deep look into Hessian free optimization, a powerful second order optimization method that has shown promising results, but still does not enjoy widespread use. This algorithm was used to train to a number of RNN architectures including standard RNNs, long short-term memory, multiplicative RNNs, and stacked RNNs on the task of character prediction. The insights from these experiments led to the creation of a new multiplicative LSTM hybrid architecture that outperformed both LSTM and multiplicative RNNs. When tested on a larger scale, multiplicative LSTM achieved character level modelling results competitive with the state of the art for RNNs using very different methodology.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/26/2016

Multiplicative LSTM for sequence modelling

We introduce multiplicative LSTM (mLSTM), a recurrent neural network arc...
research
06/30/2019

Multiplicative Models for Recurrent Language Modeling

Recently, there has been interest in multiplicative recurrent neural net...
research
09/26/2016

Automatic Construction of a Recurrent Neural Network based Classifier for Vehicle Passage Detection

Recurrent Neural Networks (RNNs) are extensively used for time-series mo...
research
02/01/2020

Model Extraction Attacks against Recurrent Neural Networks

Model extraction attacks are a kind of attacks in which an adversary obt...
research
06/07/2021

The effect of phased recurrent units in the classification of multiple catalogs of astronomical lightcurves

In the new era of very large telescopes, where data is crucial to expand...
research
06/04/2019

Towards Task and Architecture-Independent Generalization Gap Predictors

Can we use deep learning to predict when deep learning works? Our result...
research
04/18/2017

Diagonal RNNs in Symbolic Music Modeling

In this paper, we propose a new Recurrent Neural Network (RNN) architect...

Please sign up or login with your details

Forgot password? Click here to reset