Deeply Supervised Active Learning for Finger Bones Segmentation

05/07/2020 ∙ by Ziyuan Zhao, et al. ∙ 0

Segmentation is a prerequisite yet challenging task for medical image analysis. In this paper, we introduce a novel deeply supervised active learning approach for finger bones segmentation. The proposed architecture is fine-tuned in an iterative and incremental learning manner. In each step, the deep supervision mechanism guides the learning process of hidden layers and selects samples to be labeled. Extensive experiments demonstrated that our method achieves competitive segmentation results using less labeled samples as compared with full annotation.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 2

page 3

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

I Introduction

Image segmentation plays an important role in multiple applications of medical image analysis, e.g., medical diagnosis, surgical planning and pathological analysis [8803074, larson2017performance]

. Recently, deep learning based methods have achieved great success in image segmentation. However, different from segmenting natural scene images, medical image segmentation, such as hand bones segmentation, is challenging, because it is difficult to obtain medical image datasets with full annotation. On one hand, most biomedical image annotations require extensive domain knowledge and skills, which means only well trained experts can label the data. On another hand, manual labeling is tedious, costly and time-consuming. These impeded the development of deep learning in medical imaging.

To alleviate the aforementioned challenges, some self-supervised [SeseNet] or semi-supervised segmentation algorithms [NIPS2015_5858, 8834460] have been proposed, which leverage the model-generated labels (pseudo labels) to boost the performance of deep learning models iteratively. However, there is uncertainty about how to evaluate the quality of pseudo labels. Besides, in practice using a small number of accurate labels is more meaningful than using many noisy and uncertain pseudo labels. This motivates us to seek a solution to the problem of selecting data samples to be labeled for high-quality performance.

Accurate selection of informative and valuable samples to label involves the problem of Active Learning (AL). AL has been well researched on various vision tasks [7780682, wang2016cost, kading2015active, hasan2015context], but only a few studies have been reported on medical imaging. The method AIFT (active, incremental fine-tuning) proposed in [zhou2017fine]

integrates active learning and transfer learning (fine-tuning) into a single framework, which selects informative samples based on entropy and diversity. The deep active learning framework proposed in 

[yang_suggest]

first trains a set of Fully Convolutional Networks (FCNs) with little training data and then seeks informative samples based on the uncertainty among the trained FCNs and similarity between images. However, these methods based on uncertainty and similarity cannot be well embedded with final metrics. This further motivates us to study the following question:

Is it possible to improve the finger bones segmentation performance by guiding deep active learning directly?

In this work, we present a new framework that combines deep learning and active learning to reduce the annotation effort and improve the model performance by injecting deep supervision into hidden layers of the segmentation model. With a small number of training data, the U-Net is trained under deep supervision and then boosted iteratively. At the end of each stage, we exploit the deep supervision mechanism to extract useful information to decide which samples to label for the next batch. The U-Net is continuously fine-tuned by incrementally enlarging the training dataset with newly annotated samples. Experimental results show that the proposed method achieves comparable performance with less labeling cost as opposed to full annotation.

Fig. 1: Overall framework of our proposed method: the segmentation model is first trained using the labeled dataset, and the unlabeled dataset is applied on the trained networks to get multi-level segmentation outputs as the inputs of active learning process. In the active learning process, the dice coefficient between different segmentation outputs are calculated and averaged as the evaluation criteria for informative samples to label in next iteration. Informative samples are selected for annotation and the updated labeled dataset is used for the training process of the next batch.

Ii Related Work

Deep learning has dominated computer vision over the past few years, which now is a primary option for image segmentation, especially for medical image segmentation. So far, many work 

[7803544, long2015fully, Unet] in the literature adopt an “encoder-decoder” structure, in which, the resolution of the input is compressed and then, recovered by the encoder and the decoder sequentially. Among them, U-Net is commonly used in medical image segmentation, which concatenates multi-scale feature maps in the encoding stage with upsampled feature maps in the decoding stage. Such a design can be trained using very few annotated images and achieves outstanding performance in medical image segmentation.

However, different from other tasks, such as classification and regression, pixel-wise annotation is required for deep learning in medical image segmentation. Generally, only trained biomedical experts can annotate the data, which also needs extensive manual efforts. To address these issues, some deep learning frameworks [NIPS2015_5858, 8834460]

based on semi-supervised learning are developed for pseudo-labeling and finer segmentation. But these methods cannot well define how to select informative images for annotation. Therefore, several work based on active learning 

[zhou2017fine, yang_suggest] were proposed to seek the answer to the problem. Zhou et al. [zhou2017fine] proposed an active learning method for medical image segmentation called “AIFT”, in which, active learning and transfer learning are integrated to boost the entire framework for sample selection. Yang et al. [yang_suggest]

trained a set of FCNs to estimate the uncertainty among them as annotation criterion. These methods are based on uncertainty and similarity, which cannot be directly linked to the final evaluation. It is well noted that using some regularization techniques in the learning process, the information in hidden layers can be used to improve the model performance. For example, the deep supervision mechanism in 

[DOU201740] was introduced for liver segmentation, in which, the dense predictions are obtained from the hidden feature maps of some hidden layers for calculating classification errors, which can simultaneously speed up the optimization process and improve the discrimination capability. We speculate that the feature maps of hidden layers may provide guidance for active learning because of the spatial similarity between different feature maps. Therefore, we inject deep supervision into some hidden layers of U-Net and design a novel active learning strategy for finger bones segmentation.

Iii Methodology

Our active learning architecture is illustrated in Fig. 1, which can be summarized as follows. U-Net with deep supervision (deeply supervised U-Net) is adopted as the segmentation model, where additional multi-level deep supervision is injected into some hidden layers to improve the discrimination ability of the model. The model is trained on a small amount of labeled data firstly. In each iteration of training, the predicted masks of the remaining unlabeled data from hidden layers and final layers can be outputted by segmentation, which serves as the input of the active learning model. In the active learning model, the quality of the samples can be estimated by the consistency between the results of hidden layers and the final layer, and the indices of the informative samples are exported for expert annotation. To provide the ground truth of the informative samples, the process of expert annotation is simulated using the fully annotated dataset, then the training set is updated and the segmentation model is fine-tuned with the updated dataset iteratively and incrementally.

Iii-a Deeply Supervised U-Net

The segmentation model is shown in Fig. 1

. We follow the basic architecture of U-Net, The deep supervision mechanism is involved via the twentieth and twenty-third layers, namely lower layer and middle layer, in which, we upscale some low-level and middle-level feature maps of lower layer and middle layer respectively on the decoding stage using additional upsampling layers, and the softmax layer is applied on these layers as the final layer to obtain predicted masks for final loss calculation. Let

be the weights of the U-Net, and , ,

be the weights of three classifiers of lower layer, middle layer and final layer, respectively, then the cross-entropy loss function of a layer is defined as

(1)

where represents the training samples and

is the probability of target class label

corresponding to sample , in which, represents the index of the classifiers;

Finally, the total loss function can be defined as:

(2)

where , , are the weights of the associated classifiers. To control the strength of different terms, in our experiments, , and were specified as 0.1, 0.3 and 0.6, empirically.

In this way, different classifiers are simultaneously performed, which enforce the hidden layers to learn discriminative features without introducing many parameters.

Fig. 2: Example of radiographs in the dataset in different size, contrast and brightness: original images (first row), preprocessed images (second row) and binary finger bone masks (third row).

Iii-B Active Learning Model

The core of active learning is to define criteria to select informative samples. We propose a new measure of consistency in segmentation tasks for sample selection. Deep supervision in hidden layers can improve the discrimination ability of the model, and more accurate prediction from hidden layers also means better learned features, which may be similar to the output of the last layer. Moreover, different layers of a good model can perform well and consistently on high-informative samples. Therefore, the segmentation results of the hidden layers (lower layer and middle layer) are extracted to calculate the Dice’s Coefficient (DSC) with the results of the final layer, called L-DSC and M-DSC, respectively. The DSC is calculated as

(3)

where and represent the cardinalities of the two sets.

Finally, the mean of L-DSC and M-DSC serves as a proxy of the quality for each sample, which will be larger on informative samples. In another word, the images with larger mean of L-DSC and M-DSC will be selected to update the training process.

Iv Experiments

Iv-a Dataset and Evaluation Protocol

The dataset is from the 2017 RSNA Bone Age Prediction Challenge [larson2017performance], which includes hand radiographs ( for male and for female). Some samples are shown in Fig. 2. To clean the data, we first applied intensity normalization to improve the brightness and sharpness of bones. Then the images were cropped and resized to pixels. Considering the size of the dataset and manual costs, a small percentage of the dataset is sampled in our experiments to validate the effectiveness of the proposed method. For a better generalization of the segmentation, the dataset was grouped in 19 year-based age groups (0-1, 1-2, 2-3, …,18-19), then 11 cases were randomly selected from each year group to form the small subset of the training set () for experiments. Finally, the dataset is split into a training (139) /validation (20) /testing (50) set randomly.

To validate the effectiveness of the proposed pipeline, experiments were done in the 209 images with ground truth masks. We first trained a deeply supervised U-Net using 10 images randomly selected from the training set. After each step, according to the active learning model, the top 10 informative images with annotations are put into deeply supervised U-Net to update the training process. The process is called “DS-U-Net query”. Random query (randomly selecting 10 images to label) is also implemented using the same dataset for comparison.

Fig. 3:

The relationships between R-DSC and L-DSC (M-DSC, mean of them). On left side, let x be the R-DSC, y be the L-DSC (M-DSC, mean of them). On right side, DSCs are sorted by descending order, let x be the order of R-DSC, y be the order of e L-DSC (M-DSC, mean of them). We draw the scatter plots based on (x, y). Linear regression was carried out with the scatters on the right side. The regression coefficient is 0.80, 0.73 and 0.80 from top to bottom.

Iv-B Results and Discussions

To explore the relationships between real DSC (R-DSC) and L-DSC / M-DSC / mean of them, in the iterative training process, we randomly select 10 samples from each step and put them into the model for prediction. From Fig. 3, we can see that the L-DSC / M-DSC / Mean of them are positively correlated with the real DSC. Because there are slight changes and fluctuations of DSC over the training process, we sort the DSC by descending order in each iteration, and the scatter plots and trend lines between the order of L-DSC / M-DSC / mean of them and the order of the real DSC are drawn. There are strong relationships between these orders, which proves the feasibility of the proposed criteria in active learning. Therefore, the segmentation accuracy can be correctly estimated by the active learning model, and the most informative slices can be selected.

The experimental results on the testing data are shown in Fig. 4. We can see that out strategy is better than random query strategy, only using 43.16% to achieve comparable results as compared with full annotation, which benefits from the deep supervision on hidden layers. This proves that the deep supervision mechanism not only helps to improve the discrimination ability of the segmentation model, but also serve as a criterion for active learning.

Fig. 4: Results of experiments. This figure shows that our approach can get a comparable results only using 60 annotations (43.16% of the whole dataset).

V Conclusions and Future Work

In this paper, we propose a novel active learning strategy for finger bones segmentation. Auxiliary deep supervision is injected after certain hidden layers of a deep U-Net model, which not only help convergence, but also serves as a criterion for active learning. The deeply supervised U-Net is first initialized on a subset of training set. Then it is updated and fine-tuned iteratively and incrementally. In each iteration, informative samples are selected based on the consistency between final masks and masks from hidden layers. Experimental results show that the proposed method outperforms random query and obtains comparable results only using 43.16% samples as opposed to full annotation. Moreover, observations on the relationships between final masks and masks from hidden layers proves the effectiveness of the proposed active learning.

In the future research, we can implement it for semi-supervised learning and pseudo labeling. Deep supervision can provide guidance to the training process instead of introducing many parameters, and potentially reducing the cost of model design and implementation.

References