Neural Speed Reading with Structural-Jump-LSTM

03/20/2019
by   Christian Hansen, et al.
0

Recurrent neural networks (RNNs) can model natural language by sequentially 'reading' input tokens and outputting a distributed representation of each token. Due to the sequential nature of RNNs, inference time is linearly dependent on the input length, and all inputs are read regardless of their importance. Efforts to speed up this inference, known as 'neural speed reading', either ignore or skim over part of the input. We present Structural-Jump-LSTM: the first neural speed reading model to both skip and jump text during inference. The model consists of a standard LSTM and two agents: one capable of skipping single words when reading, and one capable of exploiting punctuation structure (sub-sentence separators (,:), sentence end symbols (.!?), or end of text markers) to jump ahead after reading a word. A comprehensive experimental evaluation of our model against all five state-of-the-art neural reading models shows that Structural-Jump-LSTM achieves the best overall floating point operations (FLOP) reduction (hence is faster), while keeping the same accuracy or even improving it compared to a vanilla LSTM that reads the whole text.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/28/2019

Leap-LSTM: Enhancing Long Short-Term Memory for Text Categorization

Recurrent Neural Networks (RNNs) are widely used in the field of natural...
research
04/23/2017

Learning to Skim Text

Recurrent Neural Networks are showing much promise in many sub-areas of ...
research
05/07/2018

Sentence-State LSTM for Text Representation

Bi-directional LSTMs are a powerful tool for text representation. On the...
research
02/15/2018

DR-BiLSTM: Dependent Reading Bidirectional LSTM for Natural Language Inference

We present a novel deep learning architecture to address the natural lan...
research
04/21/2021

AdaptiFont: Increasing Individuals' Reading Speed with a Generative Font Model and Bayesian Optimization

Digital text has become one of the primary ways of exchanging knowledge,...
research
02/29/2020

Depth-Adaptive Graph Recurrent Network for Text Classification

The Sentence-State LSTM (S-LSTM) is a powerful and high efficient graph ...
research
10/30/2016

Represent, Aggregate, and Constrain: A Novel Architecture for Machine Reading from Noisy Sources

In order to extract event information from text, a machine reading model...

Please sign up or login with your details

Forgot password? Click here to reset