A Unified Mammogram Analysis Method via Hybrid Deep Supervision

08/31/2018 ∙ by Rongzhao Zhang, et al. ∙ 0

Automatic mammogram classification and mass segmentation play a critical role in a computer-aided mammogram screening system. In this work, we present a unified mammogram analysis framework for both whole-mammogram classification and segmentation. Our model is designed based on a deep U-Net with residual connections, and equipped with the novel hybrid deep supervision (HDS) scheme for end-to-end multi-task learning. As an extension of deep supervision (DS), HDS not only can force the model to learn more discriminative features like DS, but also seamlessly integrates segmentation and classification tasks into one model, thus the model can benefit from both pixel-wise and image-wise supervisions. We extensively validate the proposed method on the widely-used INbreast dataset. Ablation study corroborates that pixel-wise and image-wise supervisions are mutually beneficial, evidencing the efficacy of HDS. The results of 5-fold cross validation indicate that our unified model matches state-of-the-art performance on both mammogram segmentation and classification tasks, which achieves an average segmentation Dice similarity coefficient (DSC) of 0.85 and a classification accuracy of 0.89. The code is available at https://github.com/angrypudding/hybrid-ds.



There are no comments yet.


page 7

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Breast cancer is one of the top causes of cancer death in women. In 2017, it is estimated that there are 252,710 new diagnoses of invasive breast cancer among women in the United States, and approximately 40,610 women are expected to die from the disease

[2]. The detection of breast cancer in its early stage by mammography allows patients to get better treatments, and thus can effectively lower the mortality rate. Currently, mammogram screening is still based on experts’ reading, but this process is laborious and prone to error.

Computer-aided diagnosis (CADx) system is a potential solution to facilitate mammogram screening, and the research on automatic (or semi-automatic) mammogram analysis has been a focus in medical vision field. Given the fact that a mass only occupies a small region (typically  2%) of a whole mammogram (i.e. the “needle in a haystack” problem [9]), it is very hard to identify a mass from the whole image without introducing a large number of false positives. Therefore, traditionally, both hand-crafted feature based methods [1, 11]

and deep learning models

[4, 7] require manually extracted regions of interest (ROIs), which, however, affects their usefulness in clinical practice. Recently, Dhungel et al. [3] proposed a sophisticated framework integrating mass detection, segmentation and classification modules to do whole-image classification, which achieved state-of-the-art performance with minimal manual intervention (manually rejecting false positives after detection). Besides, Lotter et al. [9] proposed a 2-stage curriculum learning method to cope with the classification of whole mammograms, and Zhu et al. [14]

developed a sparse multi-instance learning (MIL) scheme to facilitate the end-to-end training of convolution neural networks (CNNs) for whole-image classification. Nevertheless, these methods either require manual intervention and multi-stage training, or only focus on the classification problem, while the accurate location and size of masses also play a critical role in a CADx system.

In this paper, we propose a CNN-based model with Hybrid Deep Supervision (Hybrid DS, HDS) to perform whole-mammogram classification and mass segmentation simultaneously. This model is based on a very deep U-Net [12] with residual connections [6] (U-ResNet) which has 45 convolutional layers in the main stream. To facilitate the multi-task training of the deep network and boost its performance, we extend deep supervision (DS) [8]

to Hybrid DS by introducing multi-task supervision into each auxiliary classifier in DS, and apply this scheme to the U-ResNet model. To evaluate the proposed method, we performed extensive experiments on a publicly available full-field digital mammographic (FFDM) dataset, i.e. INbreast

[10]. The results show that our model achieves state-of-the-art performance in both classification and segmentation metrics, and ablation studies are performed to demonstrate the efficacy of HDS scheme.

2 Method

Figure 1: Architecture of the U-ResNet model with Hybrid DS. Best viewed in color.

2.1 Motivation

Due to the very small size of masses, directly training deep CNN models for whole mammogram classification can lead to a severe overfitting problem, where the powerful model may easily memorize the patterns presented in the background area rather than learn the feature of masses, leading to poor generalization performance. To deal with this problem, we propose to employ both image-wise and pixel-wise labels to supervise the training process. The underlying assumption for this multi-task scheme is two-fold. First, since classification (whether there exist any masses in a mammogram) and segmentation (whether each pixel belongs to a mass) are highly correlated tasks, the features learned in one task should also be useful in the other; second, multi-task learning itself can serve as a regularization method as it prevents the training process from biasing towards either task. Therefore, we propose a multi-task CNN model trained with Hybrid DS to attack the whole-mammogram classification and segmentation problems.

2.2 Hybrid Deep Supervision

Similar to DS, Hybrid DS directly supervises intermediate layers via auxiliary classifiers to force the model to learn more discriminative features in early layers. Meanwhile, HDS extends DS by introducing multi-task classifiers into each supervision level. Formally, the optimization objective of HDS is defined as:


where denotes the training dataset, is the trainable parameters of the main stream, i.e. the U-ResNet model without multi-task supervision pathways, and respectively denote the parameters of the segmentation and classification parts of the multi-task paths, and is a constant that controls the relative importance of the classification loss. The third term to the right is a regularization term weighted by a hyper-parameter , and denotes L2-norm. The segmentation loss and classification loss are defined as follow:


where denotes the parameter in the first layers of the main stream, and are respectively the weights in the segmentation and the classification parts of the multi-task path associated with the -th layer, is the weights of the corresponding loss level, and is a set that contains the indices of the layers directly connected to multi-task supervision paths. is a segmentation cross-entropy cost function that returns the average loss across all pixel locations. is basically a cross-entropy cost function, which is made compatible with MIL scheme, and its definition will be detailed in the following section.

2.3 Deep Multi-Instance Learning

Conventionally, CNN models used in classification tasks includes at least 1 fully connected (FC) layers, which can only take fixed-size inputs. However, segmentation models, e.g. fully convolutional networks (FCNs), are usually trained on cropped image patches and tested on whole images, where the input size may vary. To integrate the two tasks into one unified framework, we convert the classification part of the model into a FCN manner. Thus, the classification part may take inputs of different sizes like a segmentation model, but its output also turns into a 2D probabilistic map, no longer a single value. If we map the pixels in such a 2D map back to nonoverlapping patches in the input image, the whole input image can then be regarded as a bag of patches (instances), thus the mammogram classification can be treated as a standard MIL problem. In this case, denoting the pixel values in a 2D probabilistic map as

, the mass probability of the input image

is then . Following the practice of Zhu et al. in [14], we define the classification cost for an input image as below:


where is the true label of image , and is the pixel value in the 2D probabilistic map. Since masses are sparse in mammograms, the summation of should be small. Therefore, a sparsity term (the second term to the right) is added to the cost function, which is weighted by .

2.4 Network Architecture

The architecture of the proposed neural network model is illustrated in Fig. 1. The model is basically a deep U-ResNet with 45

convolutional layers (1 convolution and 22 residual blocks), and multi-task supervision pathways are inserted into each scale level for Hybrid DS. We use max pooling in downsampling modules (except for the last downsampling layer in the classification part of each multi-task path, which employs average pooling), and bi-linearly upsample feature maps in upsampling layers. For those transition modules (i.e. downsample, upsample and concatenation), if the input and output channel dimensions are different,

convolutions are inserted before the operation to change the channel dimension. All max pooling layers have a stride of 2, and the stride of average pooling layers ranges from

to to ensure a total downsampling factor of for the output of each classification path (so the size of the output probabilistic map is in training and in testing). Similarly, All upsampling layers except for the ones in multi-task paths have a stride of 2, and those in multi-task paths range from to to ensure the output mask have the same size as the input image. Besides, Dropout [13] layers of rate 0.2 (for residual blocks with less than 128 channels) or 0.5 (for others) are inserted into each residual block.

3 Experiments and Results

3.0.1 Dataset

The proposed method was evaluated on a publicly available FFDM dataset, i.e. INbreast [10]. Among the 410 mammograms in INbreast dataset, 107 contain one or more masses, and totally contain 116 benign or malignant masses. In pre-processing, we removed the left or right blank regions by thresholding, resized the mammograms to , and then normalized each image to zero-mean and unit-std according to the statistics of training sets. During training, the whole image was randomly flipped vertically or horizontally, and patches of size were randomly sampled from it with 50% chance centered on a positive (mass) pixel. The classification label of a cropped patch was set to 1 if the patch contained any pixel from masses, and 0 otherwise. In our experiment, the whole dataset was uniformly divided into 5 folds (82 images per fold), and we used three of them for training, one for validation and one for testing. We first performed ablation study on one data split to demonstrate the efficacy of the proposed method, and then ran a 5-fold cross validation for a fairer comparison with existing methods.

3.0.2 Implementation Details

The proposed method was implemented with PyTorch v0.4.0 on a PC with one NVIDIA Titan Xp GPU. Stochastic gradient descent (SGD) method with a momentum of 0.9 was used to optimize the model, with an initial learning rate of 0.01 and decayed by 0.3 after 1000, 1800, 2400 and 2410 epochs. In all experiments, the model was trained for 2800 epochs, which took about 12.5 hours and was long enough for each configuration to converge. The model parameters were initialized by Kaiming method


. Other hyperparameters were set as follows: classification loss weight

(such that the segmentation and classification losses of each mini-batch were comparable in magnitude), weight decay , sparsity weight , and the weights of different supervision levels , where to were gradually decayed to very small values (i.e. 0.005) during training. The losses stemmed from inner layers of the U-Net were initially weighted higher to force the these layers to learn meaningful features, otherwise they tended to be ignored due to the difficulty in learning from low-resolution feature maps.

3.0.3 Metrics

We employed dice similarity coefficient (DSC), sensitivity (SE) and false positives per image (FPI) to evaluate the segmentation results. For classification, accuracy (ACC), area under ROC curve (AUC), score, precision (Prec) and recall (Recl) were reported.

Model Segmentation Classification
Multi-task Only 0.787 0.882 0.293 0.829 0.866 0.682 0.714 0.652
Cls + DS N/A N/A N/A 0.878 0.853 0.722 0.619 0.867
Seg + DS 0.802 0.910 0.183 0.842 0.890 0.723 0.810 0.654
Hybird DS 0.848 0.907 0.110 0.915 0.887 0.821 0.762 0.889
  • Classification results were retrieved from segmentation masks by assigning the largest activation across the output probabilistic map to the whole image.

Table 1: Ablation Study

3.0.4 Ablation Study

To investigate the efficacy of the proposed Hybrid DS scheme, a series of experiments were conducted on one data split. From Table 1, it can be observed that the Hybrid DS model achieves the best performance on several important metrics (e.g. DSC, FPI, ACC, , etc.), and also has high scores on others. HDS outperforms the baseline multi-task model by a large margin (0.848 vs 0.787 in DSC, 0.915 vs 0.829 in ACC), indicating that directly supervising intermediate layers is necessary for training such a deep model. Thanks to the sparse MIL [14] and DS schemes, the Cls+DS model performed well in classification, having an accuracy of 0.878. Meanwhile, HDS achieves even higher classification performance (e.g. ACC: 0.915) than Cls+DS, which evidences the benefit of employing extra pixel-wise supervision. Compared to Seg+DS, HDS achieves better DSC (0.848), accuracy (0.915), score (0.821) and precision (0.889), which we attribute to the extra image-wise supervision. Since image-wise supervision can force the network to look wider and to learn features based on the whole image (or at least a larger area), the network becomes less sensitive to local patterns that mimic masses and more robust in rejecting false positives, as has been validated by the much higher precision of HDS (0.889) than Seg+DS (0.654). Altogether, these experiments suggest that the proposed Hybrid DS scheme is a promising approach to improve deep model’s performance on the mammogram analysis problem.

Model Segmentation Classification
D. [3] 0.85 N/A 1.00 0.910.02 0.760.23 N/A N/A N/A
Z. [14] N/A N/A N/A 0.900.02 0.890.04 N/A N/A N/A
Ours 0.850.01 0.880.02 0.080.02 0.890.02 0.850.02 0.770.04 0.690.04 0.870.06
  • Calculated on correctly detected masses.

  • These detection false positives were manually rejected before further processing.

Table 2: Comparison with State-of-the-art Methods

3.0.5 Comparison with Existing Methods

To compare the proposed model with other mammogram analysis methods, we used 5-fold cross validation to evaluate it on the whole INbreast dataset. As shown in Table 2, our model matches the current state-of-the-art performance on both mass segmentation and classification tasks, achieving a high average segmentation DSC of 0.85 and a classification accuracy of 0.89. Besides, our method is fully automatic and easy to deploy, which takes whole mammograms as input and then outputs segmentation masks and image-wise labels simultaneously. In contrast, the method by Dhungel et al. [3] still requires manual intervention to reject false positives after mass detection, and the method by Zhu et al. [14] can only give a very rough location of identified masses. Qualitative segmentation results of our method on several typical testing images have been illustrated in Fig. 2.

Figure 2: Qualitative results. The first row is original mammograms. In the second row, red and green boundaries are ground truth delineation and automatic segmentation results, respectively. The fifth column contains a false negative lesion, and the last column has a false positive lesion. Best viewed in color.

4 Conclusion

In this paper, we have developed an end-to-end and unified framework for mammogram classification and segmentation. We seamlessly integrate the two tasks into one model by employing the novel Hybrid DS scheme, which not only inherits the merits of DS but also supports multi-task learning. With such a multi-task learning scheme, pixel-wise labels tell the model where to learn while image-wise labels force the network to make better use of contextual information. We conducted extensive experiments on the publicly available INbreast dataset, and the results show that our method matches the state-of-the-art performance on both segmentation and classification tasks. Ablation study demonstrates that pixel-wise and image-wise supervisions are mutually beneficial, and the proposed Hybrid DS can effectively boost the model’s performance on both tasks. Besides, our unified framework is inherently general, which can be easily extended to other medical vision problems.


  • [1] Buciu, I., Gacsadi, A.: Directional features for automatic tumor classification of mammogram images. Biomedical Signal Processing and Control 6(4), 370–378 (2011)
  • [2] DeSantis, C.E., Ma, J., Goding Sauer, A., Newman, L.A., Jemal, A.: Breast cancer statistics, 2017, racial disparity in mortality by state. CA: a cancer journal for clinicians 67(6), 439–448 (2017)
  • [3] Dhungel, N., Carneiro, G., Bradley, A.P.: A deep learning approach for the analysis of masses in mammograms with minimal user intervention. Medical image analysis 37, 114–128 (2017)
  • [4] Geras, K.J., Wolfson, S., Shen, Y., Kim, S., Moy, L., Cho, K.: High-resolution breast cancer screening with multi-view deep convolutional neural networks. arXiv preprint arXiv:1703.07047 (2017)
  • [5]

    He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE international conference on computer vision. pp. 1026–1034 (2015)

  • [6]

    He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 770–778 (2016)

  • [7] Kooi, T., Litjens, G., van Ginneken, B., Gubern-Mérida, A., Sánchez, C.I., Mann, R., den Heeten, A., Karssemeijer, N.: Large scale deep learning for computer aided detection of mammographic lesions. Medical image analysis 35, 303–312 (2017)
  • [8]

    Lee, C.Y., Xie, S., Gallagher, P., Zhang, Z., Tu, Z.: Deeply-supervised nets. In: Artificial Intelligence and Statistics. pp. 562–570 (2015)

  • [9] Lotter, W., Sorensen, G., Cox, D.: A multi-scale cnn and curriculum learning strategy for mammogram classification. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, pp. 169–177. Springer (2017)
  • [10] Moreira, I.C., Amaral, I., Domingues, I., Cardoso, A., Cardoso, M.J., Cardoso, J.S.: Inbreast: toward a full-field digital mammographic database. Academic radiology 19(2), 236–248 (2012)
  • [11]

    Pratiwi, M., Harefa, J., Nanda, S., et al.: Mammograms classification using gray-level co-occurrence matrix and radial basis function neural network. Procedia Computer Science

    59, 83–91 (2015)
  • [12] Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical image computing and computer-assisted intervention. pp. 234–241. Springer (2015)
  • [13]

    Srivastava, N., Hinton, G., Krizhevsky, A., Sutskever, I., Salakhutdinov, R.: Dropout: A simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research

    15(1), 1929–1958 (2014)
  • [14] Zhu, W., Lou, Q., Vang, Y.S., Xie, X.: Deep multi-instance networks with sparse label assignment for whole mammogram classification. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. pp. 603–611. Springer (2017)