Implicit Language Model in LSTM for OCR

05/23/2018
by   Ekraam Sabir, et al.
0

Neural networks have become the technique of choice for OCR, but many aspects of how and why they deliver superior performance are still unknown. One key difference between current neural network techniques using LSTMs and the previous state-of-the-art HMM systems is that HMM systems have a strong independence assumption. In comparison LSTMs have no explicit constraints on the amount of context that can be considered during decoding. In this paper we show that they learn an implicit LM and attempt to characterize the strength of the LM in terms of equivalent n-gram context. We show that this implicitly learned language model provides a 2.4% CER improvement on our synthetic test set when compared against a test set of random characters (i.e. not naturally occurring sequences), and that the LSTM learns to use up to 5 characters of context (which is roughly 88 frames in our configuration). We believe that this is the first ever attempt at characterizing the strength of the implicit LM in LSTM based OCR systems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/05/2015

Listen, Attend and Spell

We present Listen, Attend and Spell (LAS), a neural network that learns ...
research
11/06/2018

Evaluating the Ability of LSTMs to Learn Context-Free Grammars

While long short-term memory (LSTM) neural net architectures are designe...
research
07/31/2020

A Study on Effects of Implicit and Explicit Language Model Information for DBLSTM-CTC Based Handwriting Recognition

Deep Bidirectional Long Short-Term Memory (D-BLSTM) with a Connectionist...
research
11/05/2016

Neural Architecture Search with Reinforcement Learning

Neural networks are powerful and flexible models that work well for many...
research
07/20/2016

Sequence to sequence learning for unconstrained scene text recognition

In this work we present a state-of-the-art approach for unconstrained na...
research
08/31/2018

Neural DrugNet

In this paper, we describe the system submitted for the shared task on S...
research
08/31/2018

Do Language Models Understand Anything? On the Ability of LSTMs to Understand Negative Polarity Items

In this paper, we attempt to link the inner workings of a neural languag...

Please sign up or login with your details

Forgot password? Click here to reset