Scene Text Detection via Holistic, Multi-Channel Prediction

Recently, scene text detection has become an active research topic in computer vision and document analysis, because of its great importance and significant challenge. However, vast majority of the existing methods detect text within local regions, typically through extracting character, word or line level candidates followed by candidate aggregation and false positive elimination, which potentially exclude the effect of wide-scope and long-range contextual cues in the scene. To take full advantage of the rich information available in the whole natural image, we propose to localize text in a holistic manner, by casting scene text detection as a semantic segmentation problem. The proposed algorithm directly runs on full images and produces global, pixel-wise prediction maps, in which detections are subsequently formed. To better make use of the properties of text, three types of information regarding text region, individual characters and their relationship are estimated, with a single Fully Convolutional Network (FCN) model. With such predictions of text properties, the proposed algorithm can simultaneously handle horizontal, multi-oriented and curved text in real-world natural images. The experiments on standard benchmarks, including ICDAR 2013, ICDAR 2015 and MSRA-TD500, demonstrate that the proposed algorithm substantially outperforms previous state-of-the-art approaches. Moreover, we report the first baseline result on the recently-released, large-scale dataset COCO-Text.


page 1

page 2

page 3

page 4

page 7

page 8


Multi-Oriented Text Detection with Fully Convolutional Networks

In this paper, we propose a novel approach for text detec- tion in natur...

Soft-PHOC Descriptor for End-to-End Word Spotting in Egocentric Scene Images

Word spotting in natural scene images has many applications in scene und...

Scene Text Recognition from Two-Dimensional Perspective

Inspired by speech recognition, recent state-of-the-art algorithms mostl...

EAST: An Efficient and Accurate Scene Text Detector

Previous approaches for scene text detection have already achieved promi...

DGST : Discriminator Guided Scene Text detector

Scene text detection task has attracted considerable attention in comput...

Multi-Granularity Prediction for Scene Text Recognition

Scene text recognition (STR) has been an active research topic in comput...

PatchFCN for Intracranial Hemorrhage Detection

This paper studies the problem of detecting acute intracranial hemorrhag...

I Introduction

Textual information in natural scenes can be very valuable and beneficial in a variety of real-world applications, ranging from image search [46], human-computer interaction [24], to criminal investigation [1] and assistance for the blind [57]. In the past few years, scene text detection and recognition have received a lot of attention from both the computer vision community and document analysis community, and numerous inspiring ideas and methods [6, 9, 32, 49, 34, 31, 37, 2, 60, 15] have been proposed to tackle these problems.

However, localizing and reading text in uncontrolled environments (i.e., in the wild) are still overwhelmingly challenging, due to a number of factors, such as variabilities in text appearance, layout, font, language and style, as well as interferences from background clutter, noise, blur, occlusion, and non-uniform illumination. In this paper, we focus on the problem of scene text detection, which aims at predicting the presence of text, and if any, estimating the position and extent of each instance.

Previous methods mainly seek text instances (characters, words or text lines) in local regions, with sliding-window [23, 6, 50, 35, 16] or connected component extraction [9, 33, 54, 36] techniques. These algorithms have brought novel ideas into this field and constantly advanced the state-of-the-art. However, most of the existing algorithms spot text within local regions (up to text line level), making it nearly impossible to exploit context in wider scope, which can be critical for dealing with challenging situations. Consequently, they would struggle in hunting weak text instances and suppressing difficult false positives.

Moreover, almost all the previous methods (except for [54, 20, 59]) have focused on detecting horizontal or near-horizontal texts, overlooking non-horizontal ones. This largely limits the practicability and adaptability of these methods, since crucial information in regard to the scene might be embodied in such non-horizontal texts.

Fig. 1: Text regions predicted by the proposed text detection algorithm. In this work, scene text detection is casted as a semantic segmentation problem, which is conceptionally and functionally different from previous sliding-window or connected component based approaches.

We propose in this work a novel algorithm for scene text detection, which treats text detection as semantic segmentation problem [43]. This algorithm performs holistic, per-pixel estimation and produces dense maps, in which the properties of scene text are implied, as shown in Fig. 1. Nevertheless, simple two-class (text vs. non-text) semantic segmentation is not sufficient for scene text detection, since multiple text instances that are very close to each other may make it hard to separate each instance (see Fig. 2). We tackle this issue by taking into account the center and scale of individual characters as well as the linking orientation of nearby characters, in addition to the location of text regions.

When multiple text lines flock together and their orientations are unknown in advance, it is not trivial to identify and group each text line. Previous methods either simply assume that text lines are horizontal or near-horizontal, or use heuristics to perform text line grouping. In this paper, we propose to build a graph using predicted properties of characters (location, scale, linking orientation, etc) and form text lines with graph partition 


Fig. 2: An issue caused by simple two-class semantic segmentation. When several text lines are very close, the estimated text regions may stick together, making it difficult to identify each individual text line.

The proposed strategy is realized using the FCN framework [28]

, which was originally designed for semantic segmentation. This framework is chosen because it applies multi-scale learning and prediction, conforming to the multi-scale nature of text in natural scenes, and includes rich prior information of natural images, by pretraining on large volume of real-world data (ImageNet 

[8]). Building upon FCN, the proposed method is able to effectively detect text instances with high variability while coping with hard false alarms in real-world scenarios.

To evaluate the effectiveness and advantages of the proposed algorithm, we have conducted experiments on public datasets in this area, including ICDAR 2013 [22], ICDAR 2015 [21] and MSRA-TD500 [54]. The quantitative results demonstrate that the proposed method significantly outperforms previous state-of-the-art algorithms. Specifically, this is the first work111The algorithms evaluated in [48] were used in the process of data annotation, thus they cannot be considered as valid baselines. We learn about this via correspondence with the authors of [48]. that reports quantitative performance on the large-scale benchmark COCO-Text [48], which is much larger (63,686 natural images) and exhibits far more variability and complexity than previous datasets.

In summary, the contributions of this paper are four-folds:

(1) We cast scene text detection as a semantic segmentation problem and make holistic prediction in the detection procedure, in contrast to previous approaches which mainly make decision locally and thus cannot make full use of the contextual information in the whole image.

(2) This work simultaneously predicts the probability of text regions, characters and the relationship among adjacent characters in a unified framework, excavating more properties of scene text and endowing the system with the ability to detect multi-oriented and curved text.

(3) The algorithm substantially outperforms the prior arts on standard benchmarks in this field.

The rest of the paper is structured as follows: We review previous ideas and approaches in Sec. II. The main idea and details of the proposed algorithm are explained in Sec III and the experiments and comparisons are presented Sec. IV. Conclusion remarks are given in Sec V.

Ii Related Work

Scene text detection and recognition have been extensively studied for a few years in the computer vision community and document analysis community, and plenty of excellent works and effective strategies have been proposed [6, 9, 32, 38, 31, 37, 51, 2, 16, 59]. Comprehensive and detailed reviews can be found in [4, 19, 47, 55]. In this section, we will concentrate on text detection approaches that are most related to the proposed method.

Stroke Width Transform (SWT) [9], Maximally Stable Extremal Regions (MSER) [32] as well as their variants [54, 11, 34, 60] have been the mainstream in scene text detection. These methods generally hunt character candidates via edge detection or extreme region extraction. Different from such component-based approaches, Neumann and Matas [35] proposed to seek character strokes in a multi-scale sliding-window manner. Zhang et al. [62] presented a text detector that makes use of the local symmetry property of character groups. The work of Jaderberg et al. [15] adopted object proposal and regression techniques to spot words in natural images, drawing inspiration from R-CNN [10]. However, a common issue with these methods is that they all only use cues from local regions (up to text line level) in text detection. In contrast, the proposed algorithm makes decision in a much wider scope (up to whole image), and thus is able to take advantage of both short-range and long-range information in the images, which can be very useful in suppressing false alarms in complex scenes.

Most of the previous methods have focused on horizontal or near-horizontal text, with very few exceptions [54, 20, 59]. However, text in real-world situations can be in any orientation. The ignorance of non-horizontal text can be a severe drawback, since important information regarding the scene may be embodied in non-horizontal text. The proposed algorithm, directly inferring the orientation property of each text instance, can naturally and effortlessly deal with text of arbitrary directions.

In recent years, deep learning based methods for scene text detection 

[7, 16, 12, 15] have been very popular, due to their advantage in performance over conventional strategies [6, 9, 32, 56, 11]

. The approach proposed in this paper also utilizes deep convolutional neural networks to detect text with high variability and to eliminate false positives caused by complex background. The main difference lies in that the proposed approach works in a holistic fashion and produces global, pixel-wise prediction maps, while the other deep learning based methods essentially perform classification on local regions (image patches 

[7, 16] or proposals generated by other techniques [12, 15]).

This work is mainly inspired by the Holistically-Nested Edge Detection (HED) method proposed by Xie and Tu [52]. HED is an edge detection algorithm that adopts multi-scale and multi-level feature learning and performs holistic prediction. HED leverages the power of both Fully Convolutional Networks (FCN) [28] and Deeply-Supervised Nets (DSN), and obtains significantly enhanced performance on edge detection in natural images. There are mainly three reasons that we adopt HED as the base model of our text detection algorithm: (1) Text is highly correlated with edge, as pointed out in previous works [9, 26, 5]. (2) HED makes holistic, image-to-image prediction, which allows for the use of wide-scope and long-range contextual information to effectively suppress false positives. (3) HED can directly handle edges of different scales and orientations, which fit well the multi-scale and multi-orientation nature of text in scene images.

Iii Methodology

In this section, we will present the main idea, network architecture and details of the proposed method. Specifically, the pipeline for scene text detection is described in Sec. III-B, after an overview of the whole system (Sec III-A). The network architecture as well as the notation and formulation of the system are given in Sec. III-C and Sec. III-D, respectively. s

Iii-a Overview

Generally, the proposed algorithm follows the paradigm of the FCN [28] and HED framework [52], i.e., it infers properties of scene text in a holistic fashion by producing image-level, pixel-wise prediction maps. In this paper, we consider text regions (words or text lines), individual characters and their relationships (linking orientation between characters) as the three properties to be estimated at runtime, since these properties are effective for scene text detection and. Moreover, the ground truth of these properties can be easily obtained from standard benchmark datasets. After the pixel-wise prediction maps are generated, detections of scene text are formed by aggregating cues in these maps. As this idea is general, other useful properties regarding scene text (for instance, binary mask of character strokes and script type of text) can be readily introduced into this framework, which are expected to further improve the accuracy of scene text detection.

Iii-B Text Detection Pipeline

Fig. 3: Pipeline of the proposed algorithm. (a) Original image. (b) Prediction maps. From left to right: text region map, character map and linking orientation map. For better visualization, linking orientations are represented with color-coded lines and only those within text regions are shown. (c) Detections.

Unlike previous methods for scene text detection, which usually start from extracting connected components [9, 32, 11] or scanning multi-scale windows [6, 49, 16] within images, the proposed algorithm operates in an alternative way. As shown in Fig. 3, the pipeline is quite straightforward: The original image is fed into the trained model and three prediction maps, corresponding to text regions, characters and linking orientations of adjacent characters, are produced. Detections are formed by performing segmentation, aggregation and partition on the three maps. If required, word partition, the process of splitting text lines into individual words, is applied to the previously formed detections.

Iii-B1 Ground Truth Preparation

Fig. 4: Ground truth preparation. (a) Original image. (b) Ground truth polygons of text regions. (c) Ground truth polygons of characters. (d) Ground truth map for text regions. (e) Ground truth map for characters. (f) Ground truth map for linking orientations.

Since the problem of scene text detection has been converted to semantic segmentation and the base framework is FCN/HED, the ground truth should be converted to label maps that are compatible with FCN/HED. As depicted in Fig. 4, for each image three label maps are produced from the ground truth annotations. The label map for text regions is a binary map, in which foreground pixels (i.e., those within text regions) are marked as ’1’ and background pixels ’0’. The label map for characters is a binary map, in which foreground pixels (i.e., those within character regions) are marked as ’1’ and background pixels ’0’. To avoid the situation that the characters in the prediction map stuck together at runtime, the binary masks are shrunk to half of its original size in both dimensions.

The label map for linking orientations is a soft map, in which each foreground pixel is assigned with to a value within the range of . The orientations of the foreground pixels is assigned as the orientation of the corresponding ground truth polygons of text regions. In this work, we consider linking orientation in the range of . Linking orientations beyond this range is converted to it. All linking orientations are mapped to by shifting and normalization.

Iii-B2 Model Training

The prediction model is trained by feeding the training images and the corresponding ground truth maps into the network, which will be described in detail in Sec. III-C. The training procedure generally follows that of the HED method [52] and the differences will be explained in Sec. III-D1.

Iii-B3 Prediction Map Generation

With the trained model, maps that represent the information of text regions, characters and linking orientations are generated by inputting the original image into it. For details, see Sec. III-D2. Exemplar prediction maps generated by the trained model are demonstrated in Fig. 3 (b) and Fig. 5 (b), (c) and (d). Note that only the linking orientations within foreground regions are valid, so those in background regions (probability of being text region less than 0.5) are not shown.

Iii-B4 Detection Formation

Fig. 5: Detection formation. (a) Original image. (b) Prediction map of text regions. (c) Prediction map of characters. (d) Prediction map of linking orientations. (e) Text region (red rectangle). (f) Characters (green circles). The center and radius of the circles represent the location and scale of the corresponding characters. (g) Delaunay triangulation. (h) Graph partition. Blue lines: linkings retained. Black lines: linkings eliminated. (i) Detections.

Candidates of text regions and characters are hunted by segmenting the corresponding prediction map with adaptive thresholding (see Fig. 5 (e) and (f)). Note that since the scales of characters are shrunk to half of their original scales in the training phase, the estimated radii of the character candidates are multiplied with a factor of 2.

Then, character candidates within the same text regions are grouped into cliques and represented as graphs. Given a clique of characters , where is the character count, Delaunay triangulation [20] is applied to these characters. Delaunay triangulation provides an effective way to eliminate unnecessary linkings of distant characters. The triangulation is used to construct a graph , in which the vertexes are the characters and the edges model the similarities between pairs of characters.

The weight of edge is defined as:


where is the similarity between the pair of characters and . According to Eqn. 1, the weights of the linkings that not belonging to the triangulation are set to zero.

The similarity between and is the harmonic average of the spatial similarity and orientation similarity:


where and denote the spatial similarity and orientation similarity between and .

The spatial similarity defined as:


where is the distance of the centers of and , while is the average length of all the edges in the Delaunay triangulation .

The orientation similarity defined as:


where is the orientation of the line between the centers of and , while is the average value in the area between and in the linking orientation map (see 5 (d)). is an operator for computing the included angle (acute angle) of two orientations. This definition of orientation similarity rewards pairs that are in accordance with the predicted linking orientations while punishes those that violate such prediction.

To split groups of character candidates into text lines, a simple yet effective strategy proposed by Yin et al.[58] is adopted. Based on the graph , a Maximum Spanning Tree [39], , is constructed. Selecting and eliminating edges in is actually performing graph partition on , leading to segmentation of text lines. For example, eliminating any edge in will partition it into two parts, corresponding two text lines, while eliminating two edges will result in three text lines.

Since the number of text lines is unknown, it should be inferred in the procedure of text line segmentation. Under the assumption that text lines in real-world scenarios are in linear or near-linear form, a straightness measure function is defined:


where is the number of clusters (text lines), and

are the largest and second largest eigenvalues of the covariance matrix

. is computed using the coordinates of the centers of characters the in the th cluster. The optimal segmentation of text lines is achieved, when the value of function reaches its maximum.

However, this solution is not applicable to text lines with curved shapes (for example, those in Fig. 9), which they violate the linearity assumption. To tackle this problem, a threshold is introduced in this work. While performing text line segmentation, the edges with weight greater than will not be selected or eliminated. Since the proposed method is able to predict linking orientations of characters in straight text lines (see Fig. 3) as well as in curved text lines (see Fig. 9), this strategy works well on text lines with linear shapes and curved shapes.

Iii-B5 Post-Processing

In certain tasks, such as ICDAR 2013 [22] and ICDAR 2015[21], word partition is required, since text in images from these datasets is labelled in word level. We adopted the word partition method in [9], as it has proven to be simple and effective.

Iii-C Architecture

Fig. 6: Network architecture of the proposed algorithm. The base network is inherited from HED [52], which made surgery on the pretrained VGG-16 Net model [44].

As shown in Fig. 6, the architecture used in this work is based on that of HED [52], which made surgery on the pretrained 5-stage VGG-16 Net model [44]. Necessary modifications are applied to the architecture of HED, to accomplish the task of text detection. Specifically, three side-output layers, corresponding to text region, character and linking orientation predictions, are connected to the last convolutional layer of each stage. This means that conv1_2, conv2_2, conv3_3, conv4_3 and conv5_3 in the network have side-output layers. The side-output layer for text region prediction, character prediction and that for linking orientation prediction are learned independently. Note that from stage 2 to 5, upsampling operation, which is realized by in-network deconvolution, is needed to rescale the prediction maps to the same size of the original image. The outputs of the side layers from the 5 stages are fused to form overall predictions, which is accomplished by another convolutional layer, the same as in [52]. The loss of each stage is computed as described in Sec. III-D1.

Iii-D Formulation

The notation and formulation of the proposed method follow those of [52]. Here we will keep consistent with [52] and emphasis on the modifications and differences.

Iii-D1 Training Phase

Assume the training set is , where is the size of the training set, is the th original image and is the corresponding ground truth. Different from HED, in which the ground truth is a binary edge map, in this work is composed of three maps, i.e., , where is a binary map that indicates the presence of text regions at each pixel in the original image , is a binary map indicates the presence of characters (shrunk version are used in this work), and is a soft map that represents the linking orientations between adjacent characters in text regions. Note that the value of is valid only if , since linking orientation of adjacent characters is undefined in background. For simplicity, the index will be omitted hereafter, because the model treats each image independently.

The objective function is defined as the loss in the fused outputs, since the losses in lower side-output layers make little difference according to experiments:



is the loss function of the fused outputs.

and are the collection of parameters of all standard layers and those with the side-output layers and fuse layer, respectively. For more details, refer to [52].

In HED [52], each stage produces only one prediction map for edges. In this paper, however, each stage is used to generate three prediction maps: one for text regions, one for characters and one for linking orientations. Therefore, each stage is connected to two side-output layers, instead of one. Accordingly, the definition of the loss function is different. is a weighted sum of the corresponding channels for the three types of targets:


where , and are the losses in predicting text regions, characters and linking orientations, respectively. , and are parameters to control the relative contributions of these three types of loss functions and ++.

For an image, assume the ground truth is and the prediction maps produced by the fuse layer are , and , the loss function for text regions is similar with Eqn. (2) in [52]:


is a class-balancing parameter and it is defined as . denotes the count of pixels in non-text regions and denotes the total count of pixels. The definition of is similar with and thus is skipped here.

is defined as:


This loss function has a double-peak shape. When the difference (included angle) between the estimated orientation and true orientation is small, the loss is close to ; when it is near (or ), the loss approaches to ; when it goes beyond (or below ), the loss deceases since the difference turns around. According to the definition in Eqn. 9, orientation errors at background pixels will not be taken into account.

Iii-D2 Testing Phase

In the testing phase, the test image is mean-subtracted and fed to the trained model. The prediction maps for text regions, characters and linking orientations are obtained by taking the output of the fusion layer:


Different from HED, which used the average of the outputs of all the output-layers and the fusion layers, in this work we only employ the responses of the fusion layer. In practice, we found that the outputs of the side-output layers are likely to introduce noises and unimportant details into the prediction maps, which may be harmful to the task of scene text detection.

Iv Experiments and Discussions

We implemented the proposed algorithm using the code222 released by the authors of HED [52]

, which is based on the Caffe framework 

[18]. The proposed algorithm was evaluated on three standard benchmarks in this field and compared with other scene text detection methods. All the experiments were conducted on a regular server (2.6GHz 8-core CPU, 128G RAM, Tesla K40m GPU and Linux 64-bit OS) and the routine ran on a single GPU in each time.

Iv-a Datasets

The datasets used in the experiments will be introduced briefly:

ICDAR 2013. The ICDAR 2013 dataset333 is from the ICDAR 2013 Robust Reading Competition [22]. There are 229 natural images for training and 233 natural images for testing. All the text instances in this dataset are in English and are horizontally placed.

ICDAR 2015. The ICDAR 2015 dataset444 is from the Challenge 4 (Incidental Scene Text challenge) of the ICDAR 2015 Robust Reading Competition [21]. The dataset includes 1500 natural images in total, which are acquired using Google Glass. Different from the images from the previous ICDAR competitions [30, 41, 22]

, in which the text instances are well positioned and focused, the images from ICDAR 2015 are taken without user’s prior preference or intention, so the text instances are usually skewed or blurred.

MSRA-TD500. The MSRA Text Detection 500 Database (MSRA-TD500)555 is a benchmark dataset for assessing detection algorithms for text of different orientations, which is originally proposed in [54]. This dataset has 500 high-resolution natural scene images, in which text may be in varying directions and the language types include both Chinese and English. The training set consists of 300 images and the test set contains 200 images. This dataset is challenging due to the variability of text as well as the complexity of backgrounds.

COCO-Text. The COCO-Text666 is a newly released, large scale dataset for text detection and recognition in natural images. The original images are from the Microsoft COCO [27] dataset. In COCO-Text, 173,589 text instances from 63,686 images are annotated and each instance has 3 fine-grained text attributes. 43,686 images were chosen by the authors as training set, while 20,000 as validation/test set. This database is the largest benchmark in this area to date.

Iv-B Implementation Details

We directly made surgery on the network architecture of HED [52] to construct our model for scene text detection, and most parameters were inherited from it. In our model, the learning rate is adjusted to 1e-8 to avoid gradient explosion. Following [52], we also used VGG-16 Net [44] pre-trained on the ImageNet dataset [8]. and are used in all the experiments.

The training data are the union of the training images from the three datasets ICDAR 2013 (229 training images), ICDAR 2015 (1000 training images) and MSRA-TD500 (300 training images). These 1529 training images were first rescaled to have maximum dimension of 960 pixels with aspect ratio kept unchanged. Then we evenly rotated the images to 36 different angles (with 10 degree angle interval). The corresponding ground truth maps were generated as described in Sec. III and undergone the same augmentations as the original images. We found that the framework is insensitive to image scale, in accordance with [52]. Therefore, we did not resize the images (as well as the ground truth maps) to multiple scales. In testing, text detection is performed at multiple scales and the detections of different scales are fused to form the final results.

Iv-C Experiments and Discussions

Iv-C1 Qualitative Results

Fig. 7: Detection examples of the proposed algorithm. (a) ICDAR 2013. (b) ICDAR 2015. (c) MSRA-TD500. (d) COCO-Text

Fig. 7 illustrates a group of detection examples of the proposed algorithm on the four benchmark datasets. As can be seen, the proposed algorithm is able to handle text instances of different orientations, languages, fonts, colors and scales in diverse scenarios. Moreover, it is insensitive to non-uniform illumination, blur, local distractor and connected strokes, to some extent. These examples show the adaptability and robustness of the proposed method.

Iv-C2 Quantitative Results on ICDAR 2013

Algorithm Precision Recall F-measure
Proposed 0.8888 0.8022 0.8433
VGGMaxNet [13] 0.9218 0.7732 0.8410
Zhang et al. [63] 0.88 0.78 0.83
Zhang et al. [62] 0.88 0.74 0.80
Tian et al. [45] 0.85 0.76 0.80
Lu et al. [29] 0.89 0.70 0.78
iwrr2014 [61] 0.86 0.70 0.77
USTB TexStar [60] 0.88 0.66 0.76
Text Spotter [34] 0.88 0.65 0.74
CASIA NLPR [22] 0.79 0.68 0.73
Text Detector CASIA [42] 0.85 0.63 0.72
I2R NUS FAR [22] 0.75 0.69 0.72
I2R NUS [22] 0.73 0.66 0.69
TH-TextLoc [22] 0.70 0.65 0.67
TABLE I: Performances of different text detection methods evaluated on ICDAR 2013.

The text detection performance of the proposed algorithm as well as other methods evaluated on the ICDAR 2013 dataset are shown in Tab. I. The proposed method achieves the highest recall (0.8022) among all the methods. Specifically, the F-measure of the proposed algorithm is slightly better than that of previous state-of-the-art methods [13].

Note that on the ICDAR 2013 dataset the performance improvement of our algorithm over previous methods is not as obvious as on the ICDAR 2015 dataset (see Tab. II) and MSRA-TD500 dataset (see Tab. III). There are two main reasons: (1) The ICDAR 2013 dataset is much more time-honored (most images were from previous ICDAR competitions that can date back to 2003) and most text detection algorithms have saturated on it. (2) The main advantage of our algorithm lies in the capacity of handling multi-oriented text, but vast majority of the text instances in ICDAR 2013 are horizontal. So the advantage of our algorithm cannot be reflected when using ICDAR 2013 as benchmark.

Iv-C3 Quantitative Results on ICDAR 2015

Algorithm Precision Recall F-measure
Proposed 0.7226 0.5869 0.6477
Zhang et al. [63] 0.71 0.43 0.54
Stradvision-2 [21] 0.7746 0.3674 0.4984
Stradvision-1 [21] 0.5339 0.4627 0.4957
NJU [21] 0.7044 0.3625 0.4787
AJOU [25] 0.4726 0.4694 0.471
HUST-MCLAB [21] 0.44 0.3779 0.4066
Deep2Text-MO [60, 59] 0.4959 0.3211 0.3898
CNN MSER [21] 0.3471 0.3442 0.3457
TextCatcher-2 [21] 0.2491 0.3481 0.2904
TABLE II: Performances of different text detection methods evaluated on ICDAR 2015.

The text detection performance of the proposed method as well as other competing methods on the ICDAR 2015 dataset are shown in Tab. II. The proposed method achieves the highest recall (0.5869) and the second highest precision (0.7226). Specifically, the F-measure of the proposed algorithm is significantly better than that of previous state-of-the-art [63] (0.6477 vs. 0.54). This confirms the effectiveness and advantage of the proposed approach.

Iv-C4 Quantitative Results on MSRA-TD500

Algorithm Precision Recall F-measure
Proposed 0.7651 0.7531 0.7591
Zhang et al. [63] 0.83 0.67 0.74
Yin et al. [59] 0.81 0.63 0.71
Kang et al. [20] 0.71 0.62 0.66
Yin et al. [60] 0.71 0.61 0.66
Unified [53] 0.64 0.62 0.61
TD-Mixture [54] 0.63 0.63 0.60
TD-ICDAR [54] 0.53 0.52 0.50
Epshtein et al. [9] 0.25 0.25 0.25
Chen et al. [6] 0.05 0.05 0.05
TABLE III: Performances of different text detection methods evaluated on MSRA-TD500.

The performances of different text detection methods on the MSRA-TD500 dataset [54] are depicted in Tab. III. As can be seen, the proposed algorithm achieves the highest recall and F-measure and the second highest precision. Specially, the F-measure of the proposed algorithm significantly outperforms that of the prior art [63] (0.7591 vs. 0.74). The improvement in recall is even more obvious (0.7531 vs. 0.63).

Iv-C5 Quantitative Results on COCO-Text

Algorithm Precision Recall F-measure
Proposed 0.4323 0.271 0.3331
Baselines from [48]
A 0.8378 0.233 0.3648
B 0.8973 0.107 0.1914
C 0.1856 0.047 0.0747
TABLE IV: Performances of different text detection methods evaluated on COCO-Text.

The performance of the proposed method on the COCO-Text dataset [48] is depicted in Tab. IV777The numbers in this table have been modified, since the evaluation script for COCO-Text ( has been updated.. On this large, challenging benchmark, the proposed algorithm achieves 0.4323, 0.271 and 0.3331 in precision, recall and F-measure, respectively. As reference, the performances of the baseline methods from [48] are included. However, it is not possible to directly compare our method with these baselines, since they were used for labelling the ground truth in the data annotation procedure and they were evaluated on the full dataset (63686 image) instead of the validation set (20,000 images).

Algorithm Recall
Legible Illegible
Machine Hand Machine Hand
Proposed 0.3563 0.3109 0.0297 0.0567
Baselines from [48]
A 0.3401 0.1511 0.0409 0.0231
B 0.1616 0.1096 0.0069 0.0022
C 0.0709 0.0463 0.0028 0.0022
TABLE V: Performances on sub-categories in COCO-Text.

The performance of the proposed method on different sub-categories in the COCO-Text dataset is depicted in Tab. V888The numbers in this table have been modified, since the evaluation script for COCO-Text ( has been updated.. The proposed algorithm performs much better on legible, machine printed text than illegible, handwritten text. Likewise, the performances of the baseline methods from [48] are also included.

Iv-D Running time of Proposed Method

On 640x480 images, it take the proposed method about 0.42s on average to produce the prediction maps, when running on a K40m GPU. The subsequent processing consumes about 0.2s on CPU.

Iv-E Generalization Ability of Proposed Algorithm

Fig. 8: Generalization to text of different scripts. Even though the algorithm is only trained on examples of English and Chinese, it can generalize well to different types of scripts, such as Arabic, Hebrew, Korean and Russian. Original image are harvested from the Internet.
Fig. 9: Generalization to curved text. Curved examples are rarely seen in the training set, but the trained model can easily handle curved text. Original images are from [40].

As can be seen from Fig. 8 and Fig. 9, although the proposed method was not trained with examples of various languages nor curved instances, it generalize well to such scenarios. This verifies the strong generalization ability of the proposed algorithm.

Iv-F Limitations of Proposed Algorithm

The proposed method shows excellent capability in most scenarios, but it may fail in certain conditions. For example, it is found sensitive to serious blur and highlight. We believe if more examples and augmentation are introduced, these issues can be alleviated.

Moreover, the model size of the proposed method (about 56M) is still not compact enough and the speed without GPU (more than 14s for 640x480 images) is not sufficient, if we port the system to low-end PCs or mobile devices.

V Conclusions

In this paper, we have presented a novel algorithm for text detection in natural scene images, which is based on the HED model [52]. The proposed algorithm is fundamentally different from majority of the previous methods in that: (1) It approaches scene text detection via whole image semantic segmentation, instead of local component extraction [9, 33, 54, 36] or window-based classification. [50, 35, 16, 15], and thus is able to utilize more contextual information from larger scope. (2) It concurrently predicts the probability of text regions, characters and the linking orientation of nearby characters in a unified framework, while other methods [9, 33, 54, 16, 59] estimate such properties in separate stages. (3) It can directly spot multi-oriented and curved text from images, while most previous approaches [56, 2, 14] have focused on horizontal or near-horizontal text. Moreover, the experiments on standard benchmark datasets demonstrate that the proposed method achieves superior performance than other competing systems.

Future directions worthy of further investigation and exploration might include: (1) Architecture. The 5-stage architecture of HED leads to breakthrough in edge detection, but this architecture is not necessarily the best for scene text detection. We would modify the network architecture and seek better ones that are suitable for this task. (2) Capacity. Training with more detailed labels (e.g., binary masks of character shapes) will endow the system with the ability to explicitly segment out character shapes from original images, which may potentially facilitate the subsequent text recognition procedure. (3) Efficiency. Adopting acceleration techniques [17] could largely speed up the proposed method, making it more practical.

Vi Acknowledgements


  • [1] Text Recognition Algorithm Independent Evaluation (TRAIT).
  • [2] A. Bissacco, M. Cummins, Y. Netzer, and H. Neven. PhotoOCR: Reading text in uncontrolled conditions. In Proc. of ICCV, 2013.
  • [3] A. Buluç, H. Meyerhenke, I. Safro, P. Sanders, and C. Schulz. Recent advances in graph partitioning. In arXiv, 2015.
  • [4] D. Chen, J. Luettin, and K. Shearer. A survey of text detection and recognition in images and videos. Technical report, IDIAP, August 2000.
  • [5] H. Chen, S. S. Tsai, G. Schroth, D. M. Chen, R. Grzeszczuk, and B. Girod. Robust text detection in natural images with edge-enhanced maximally stable extremal regions. In Proc. of ICIP, 2011.
  • [6] X. Chen and A. Yuille. Detecting and reading text in natural scenes. In Proc. of CVPR, 2004.
  • [7] A. Coates, B. Carpenter, C. Case, S. Satheesh, B. Suresh, T. Wang, D. J. Wu, and A. Y. Ng. Text detection and character recognition in scene images with unsupervised feature learning. In Proc. of ICDAR, 2011.
  • [8] J. Deng, W. Dong, R. Socher, L. J. Li, K. Li, and L. Fei-Fei. Imagenet: A large-scale hierarchical image database. In Proc. of CVPR, 2009.
  • [9] B. Epshtein, E. Ofek, and Y. Wexler. Detecting text in natural scenes with stroke width transform. In Proc. of CVPR, 2010.
  • [10] R. Girshick, J. Donahue, T. Darrell, and J. Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proc. of CVPR, 2014.
  • [11] W. Huang, Z. Lin, J. Yang, and J. Wang. Text localization in natural images using stroke feature transform and text covariance descriptors. In Proc. of ICCV, 2013.
  • [12] W. Huang, Y. Qiao, and X. Tang. Robust scene text detection with convolution neural network induced mser trees. In Proc. of ECCV, 2014.
  • [13] M. Jaderberg, K. Simonyan, A. Vedaldi, and A. Zisserman. Synthetic data and artificial neural networks for natural scene text recognition. In NIPS Workshop on Deep Learning, 2014.
  • [14] M. Jaderberg, K. Simonyan, A. Vedaldi, and A. Zisserman. Deep structured output learning for unconstrained text recognition. In Proc. of ICLR, 2015.
  • [15] M. Jaderberg, K. Simonyan, A. Vedaldi, and A. Zisserman. Reading text in the wild with convolutional neural networks. IJCV, 2015.
  • [16] M. Jaderberg, A. Vedaldi, and A. Zisserman. Deep features for text spotting. In Proc. of ECCV, 2014.
  • [17] M. Jaderberg, A. Vedaldi, and A. Zisserman. Speeding up convolutional neural networks with low rank expansions. In Proc. of BMVC, 2014.
  • [18] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell. Caffe: Convolutional architecture for fast feature embedding. In Proc. of ACM International Conference on Multimedia, 2014.
  • [19] K. Jung, K. Kim, and A. Jain. Text information extraction in images and video: a survey”. PR, 37(5):977–997, 2004.
  • [20] L. Kang, Y. Li, and D. Doermann. Orientation robust text line detection in natural images. In Proc. of CVPR, 2014.
  • [21] D. Karatzas, L. Gomez-Bigorda, A. Nicolaou, S. Ghosh, A. Bagdanov, M. Iwamura, J. Matas, L. Neumann, V. R. Chandrasekhar, S. Lu, F. Shafait, S. Uchida, and E. Valveny. ICDAR 2015 competition on robust reading. In Proc. of ICDAR, 2015.
  • [22] D. Karatzas, F. Shafait, S. Uchida, M. Iwamura, L. G. i Bigorda, S. R. Mestre, J. Mas, D. F. Mota, J. A. Almazan, and L. P. de las Heras. ICDAR 2013 robust reading competition. In Proc. of ICDAR, 2013.
  • [23] K. I. Kim, K. Jung, and J. H. Kim.

    Texture-based approach for text detection in images using support vector machines and continuously adaptive mean shift algorithm.

    IEEE Trans. PAMI, 25(12):1631–1639, 2003.
  • [24] B. Kisacanin, V. Pavlovic, and T. S. Huang. Real-Time Vision for Human-Computer Interaction. Springer, Heidelberg, 2005.
  • [25] H. Koo and D. H. Kim. Scene text detection via connected component clustering and nontext filtering. IEEE Trans. on Image Processing, 22(6):2296–2305, 2013.
  • [26] S. H. Lee, K. J. M. S. Cho, and J. H. Kim. Scene text extraction with edge constraint and text collinearity link. In Proc. of ICPR, 2010.
  • [27] T. Y. Lin, M. Maire, S. Belongie, L. Bourdev, R. Girshick, J. Hays, P. Perona, D. Ramanan, C. L. Zitnick, and P. Dollár. Microsoft coco: Common objects in context. In arXiv, 2015.
  • [28] J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In Proc. of CVPR, 2015.
  • [29] S. Lu, T. Chen, S. Tian, J. H. Lim, and C. L. Tan. Scene text extraction based on edges and support vector regression. International Journal on Document Analysis and Recognition (IJDAR), 18(2):125–135, 2015.
  • [30] S. M. Lucas. Icdar 2005 text locating competition results. In Proc. of ICDAR, 2005.
  • [31] A. Mishra, K. Alahari, and C. V. Jawahar. Top-down and bottom-up cues for scene text recognition. In Proc. of CVPR, 2012.
  • [32] L. Neumann and J. Matas. A method for text localization and recognition in real-world images. In Proc. of ACCV, 2010.
  • [33] L. Neumann and J. Matas. Text localization in real-world images using efficiently pruned exhaustive search. In Proc. of ICDAR, 2011.
  • [34] L. Neumann and J. Matas. Real-time scene text localization and recognition. In Proc. of CVPR, 2012.
  • [35] L. Neumann and J. Matas. Scene text localization and recognition with oriented stroke detection. In Proc. of ICCV, 2013.
  • [36] L. Neumann and J. Matas. Efficient scene text localization and recognition with local character refinement. In Proc. of ICDAR, 2015.
  • [37] T. Novikova, , O. Barinova, P. Kohli, and V. Lempitsky.

    Large-lexicon attribute-consistent text recognition in natural images.

    In Proc. of ECCV, 2012.
  • [38] Y. Pan, X. Hou, and C. Liu. A hybrid approach to detect and localize texts in natural scene images. IEEE Trans. Image Processing, 20(3):800–813, 2011.
  • [39] S. Pemmaraju and S. Skiena. Computational Discrete Mathematics: Combinatorics and Graph Theory in Mathematica. Cambridge University Press, Cambridge, England, 2003.
  • [40] A. Risnumawan, P. Shivakumara, C. S. Chan, and C. L. Tan. A robust arbitrary text detection system for natural scene images. In Proc. of ESWA, 2014.
  • [41] A. Shahab, F. Shafait, and A. Dengel. ICDAR 2011 robust reading competition challenge 2: Reading text in scene images. In Proc. of ICDAR, 2011.
  • [42] C. Shi, C. Wang, B. Xiao, Y. Zhang, and S. Gao. Scene text detection using graph model built upon maximally stable extremal regions. Pattern Recognition Letters, 34(2):107–116, 2013.
  • [43] J. Shotton, M. Johnson, and R. Cipolla. Semantic texton forests for image categorization and segmentation. In Proc. of CVPR, 2008.
  • [44] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In Proc. of ICLR, 2014.
  • [45] S. Tian, Y. Pan, C. Huang, S. Lu, K. Yu, and C. L. Tan. Text flow: A unified text detection system in natural scene images. In Proc. of ICCV, 2015.
  • [46] S. Tsai, H. Chen, D. Chen, G. Schroth, R. Grzeszczuk, and B. Girod. Mobile visual search on printed documents using text and low bit-rate features. In Proc. of ICIP, 2011.
  • [47] S. Uchida. Text localization and recognition in images and video. Handbook of Document Image Processing and Recognition, pages 843–883, 2014.
  • [48] A. Veit, T. Matera, L. Neumann, J. Matas, and S. Belongie. Coco-text: Dataset and benchmark for text detection and recognition in natural images. In arXiv, 2016.
  • [49] K. Wang, B. Babenko, and S. Belongie. End-to-end scene text recognition. In Proc. of ICCV, 2011.
  • [50] K. Wang and S. Belongie. Word spotting in the wild. In Proc. of ECCV, 2010.
  • [51] J. J. Weinman, Z. Butler, D. Knoll, and J. Feild. Toward integrated scene text reading. IEEE Trans. on PAMI, 36(2):375–387, 2013.
  • [52] S. Xie and Z. Tu. Holistically-nested edge detection. In Proc. of ICCV, 2015.
  • [53] C. Yao, X. Bai, and W. Liu. A unified framework for multi-oriented text detection and recognition. IEEE Trans. Image Processing, 23(11):4737–4749, 2014.
  • [54] C. Yao, X. Bai, W. Liu, Y. Ma, and Z. Tu. Detecting texts of arbitrary orientations in natural images. In Proc. of CVPR, 2012.
  • [55] Q. Ye and D. Doermann. Text detection and recognition in imagery: A survey. IEEE Trans. PAMI, 37(7):1480–1500, 2014.
  • [56] C. Yi and Y. Tian. Text string detection from natural scenes by structure-based partition and grouping. IEEE Trans. Image Processing, 20(9):2594–2605, 2011.
  • [57] C. Yi and Y. Tian. Assistive text reading from complex background for blind persons. In Proc. of CBDAR, 2012.
  • [58] F. Yin and C. L. Liu. Handwritten text line extraction based on minimum spanning tree clustering. In Proc. of International Conference on Wavelet Analysis and Pattern Recognition, 2007.
  • [59] X. C. Yin, W. Y. Pei, J. Zhang, and H. W. Hao. Multi-orientation scene text detection with adaptive clustering. IEEE Trans. on PAMI, 37(9):1930–1937, 2015.
  • [60] X. C. Yin, X. Yin, K. Huang, and H. Hao. Robust text detection in natural scene images. IEEE Trans. on PAMI, 36(5):970–983, 2014.
  • [61] A. Zamberletti, L. Noce, and I. Gallo. Text localization based on fast feature pyramids and multi-resolution maximally stable extremal regions. In Proc. of ACCV workshop, 2014.
  • [62] Z. Zhang, W. Shen, C. Yao, and X. Bai. Symmetry-based text line detection in natural scenes. In Proc. of CVPR, 2015.
  • [63] Z. Zhang, C. Zhang, W. Shen, C. Yao, W. Liu, and X. Bai. Multi-oriented text detection with fully convolutional networks. In Proc. of CVPR, 2015.