Boosting Handwriting Text Recognition in Small Databases with Transfer Learning

04/04/2018
by   José Carlos Aradillas, et al.
0

In this paper we deal with the offline handwriting text recognition (HTR) problem with reduced training datasets. Recent HTR solutions based on artificial neural networks exhibit remarkable solutions in referenced databases. These deep learning neural networks are composed of both convolutional (CNN) and long short-term memory recurrent units (LSTM). In addition, connectionist temporal classification (CTC) is the key to avoid segmentation at character level, greatly facilitating the labeling task. One of the main drawbacks of the CNNLSTM-CTC (CLC) solutions is that they need a considerable part of the text to be transcribed for every type of calligraphy, typically in the order of a few thousands of lines. Furthermore, in some scenarios the text to transcribe is not that long, e.g. in the Washington database. The CLC typically overfits for this reduced number of training samples. Our proposal is based on the transfer learning (TL) from the parameters learned with a bigger database. We first investigate, for a reduced and fixed number of training samples, 350 lines, how the learning from a large database, the IAM, can be transferred to the learning of the CLC of a reduced database, Washington. We focus on which layers of the network could be not re-trained. We conclude that the best solution is to re-train the whole CLC parameters initialized to the values obtained after the training of the CLC from the larger database. We also investigate results when the training size is further reduced. The differences in the CER are more remarkable when training with just 350 lines, a CER of 3.3 18.2 reduced. Similar good results are obtained from the Parzival database when trained with this reduced number of lines and this new approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/04/2020

Boosting offline handwritten text recognition in historical documents with few labeled lines

In this paper, we face the problem of offline handwritten text recogniti...
research
02/09/2016

Associative Long Short-Term Memory

We investigate a new method to augment recurrent neural networks with ex...
research
01/31/2021

Infant Cry Classification with Graph Convolutional Networks

We propose an approach of graph convolutional networks for robust infant...
research
03/30/2018

Learning Structure and Strength of CNN Filters for Small Sample Size Training

Convolutional Neural Networks have provided state-of-the-art results in ...
research
05/24/2019

Using Deep Networks and Transfer Learning to Address Disinformation

We apply an ensemble pipeline composed of a character-level convolutiona...
research
12/09/2020

Recurrence-free unconstrained handwritten text recognition using gated fully convolutional network

Unconstrained handwritten text recognition is a major step in most docum...
research
07/31/2020

A Study on Effects of Implicit and Explicit Language Model Information for DBLSTM-CTC Based Handwriting Recognition

Deep Bidirectional Long Short-Term Memory (D-BLSTM) with a Connectionist...

Please sign up or login with your details

Forgot password? Click here to reset