IGLOO: Slicing the Features Space to Represent Long Sequences

07/09/2018
by   Vsevolod Sourkov, et al.
0

We introduce a new neural network architecture, IGLOO, which aims at providing a representation for long sequences where RNNs fail to converge. The structure uses the relationships between random patches sliced out of the features space of some backbone 1 dimensional CNN to find a representation. This paper explains the implementation of the method and provides benchmark results commonly used for RNNs and compare IGLOO to other structures recently published. It is found that IGLOO can deal with sequences of up to 25,000 time steps. For shorter sequences it is also found to be effective and we find that it achieves the highest score in the literature for the permuted MNIST task. Benchmarks also show that IGLOO can run at the speed of the CuDNN optimised GRU or LSTM without being tied to any specific hardware.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/15/2016

Tunable Efficient Unitary Neural Networks (EUNN) and their application to RNNs

Using unitary (instead of general) matrices in artificial neural network...
research
03/11/2023

Resurrecting Recurrent Neural Networks for Long Sequences

Recurrent Neural Networks (RNNs) offer fast inference on long sequences ...
research
11/17/2019

Multi-Zone Unit for Recurrent Neural Networks

Recurrent neural networks (RNNs) have been widely used to deal with sequ...
research
09/12/2017

Parallelizing Linear Recurrent Neural Nets Over Sequence Length

Recurrent neural networks (RNNs) are widely used to model sequential dat...
research
10/03/2021

Graph Representation Learning for Spatial Image Steganalysis

In this paper, we introduce a graph representation learning architecture...
research
02/23/2018

Can Neural Networks Understand Logical Entailment?

We introduce a new dataset of logical entailments for the purpose of mea...
research
12/02/2019

Long Distance Relationships without Time Travel: Boosting the Performance of a Sparse Predictive Autoencoder in Sequence Modeling

In sequence learning tasks such as language modelling, Recurrent Neural ...

Please sign up or login with your details

Forgot password? Click here to reset