1 Introduction
According to the International Agency for Research on Cancer, breast cancer accounts for of invasive cancers and of cancer-related deaths in women worldwide (Boyle et al., 2008). In the U.S., in women is expected to develop invasive breast cancer over the course of her lifetime DeSantis et al. (2014). Routine mammography is standard for preventive care and detection of breast cancer before biopsy. However, it is still a manual process, prone to human error due to high variability in mass appearance Ball and Bruce (2007) and low signal-to-noise ratio, and thus can cause unnecessary biopsies or worse, missed malignant masses. Furthermore, efficacy is often highly correlated with radiologist expertise and workload Elmore et al. (2009).
Convolutional Neural Networks (CNN) have achieved impressive results on computer visions tasks spanning classification
Krizhevsky et al. (2012), object detection Girshick et al. (2014), and segmentation Long et al. (2015). For breast mass diagnosis, deep learning techniques have been explored
Arevalo et al. (2016); Carneiro et al. (2015); Dhungel et al. (2015a, b); Wang et al. (2016), but never in a fully end-to-end manner (directly classifying from pixel space) because of the scarcity of available data and lack of interpretability. In this work, we successfully train end-to-end CNN architectures to directly classify breast masses as benign or malignant. We obtain state-of-the-art results using a combination of transfer learning, careful pre-processing and data augmentation. We analyze the effects of these modelling choices and furthermore show how we can provide interpretability to the model’s predictions.2 Related Work
While medical images differs significantly from natural images, traditional feature engineering techniques from computer vision such as scale-invariant feature transform (SIFT) and histogram of oriented gradients (HOG) have seen use and success when applied to medical images. More recently, deep learning-based approaches using CNNs have begun to achieve impressive performance on medical tasks such as chest pathology identification in X-Ray and CT Bar et al. (2015); van Ginneken et al. (2015), and thoraco-abdominal lymph node detection and interstitial lung disease classification Shin et al. (2016).
In the context of mammography, Dhungel et al. (2015a, b)
detect breast masses using a combination of R-CNN and random forests. Multiple works tackle the problem of breast lesion classification, but typically adopt a multi-stage approach.
Wang et al. (2016) extracts hand-engineered semantic (such as calcification) and textual features, and Carneiro et al. (2015) classifies a full mammogram by extracting features from each view of the breast and combining them to output a prediction. Arevalo et al. (2016) performs extensive pre-processing using domain knowledge before training a CNN. To the best of our knowledge, ours is the first work to directly classify pre-detected breast masses using CNN architectures.3 Dataset
In our experiments, we use the Digital Database for Screening Mammography (DDSM) Bowyer et al. (1996), a collaboratively maintained public dataset at the University of South Florida. It comprises approximately 2500 studies each containing both mediolateral oblique (MLO) and craniocaudal (CC) views of each breast. Each image is grayscale and accompanied by a mask specifying the region of the pre-segmented mass if present. Examples of benign and malignant masses are shown in Fig. 1.
We consider only mammograms which contain masses, resulting in 1820 images from 997 patients. We split these randomly by patient into training, validation and testing sets ( and of the full dataset), constraining the validation and test sets to be balanced.
|
|
4 Methods
We train three different CNN architectures for breast mass classification, and analyze the effect of a number of model choices. We describe these below.
4.1 CNN architectures
We evaluate three network architectures: a shallow CNN (the baseline model), an AlexNet Krizhevsky et al. (2012) and a GoogLeNet Szegedy et al. (2015). For both the AlexNet and GoogLeNet, we use the same base architecture as the original works but replace the last fully-connected (FC) layer to output classes. We also remove the two auxiliary classifiers from the GoogLeNet which we found impaired our training in practice.
The baseline model’s architecture is inspired by the early layers of AlexNet Krizhevsky et al. (2012)
. We additionally use batch normalization
Ioffe and Szegedy (2015). The network takes a image as input. It consists of convolutional blocks composed ofConvolutions - Batch Norm - ReLU - Max Pooling, with respectively
, and filters each, followed by FC layers of size , and. The final layer is a soft-max layer for binary classification. We use ReLU activation functions, Xavier
Glorot and Bengio (2010) weight initialization, and the Adam Kingma and Ba (2014) update rule with a base learning rate of and batch size .4.2 Aspects of model analysis
Transfer learning
We analyze the effect of initializing networks with pre-training on the ImageNet dataset
Russakovsky et al. (2015), and then fine-tuning on mammography images. For the AlexNet model, we initialize the convolutional layers with pre-trained weights and a smaller learning rate multiplier of , and randomly initialize the 3 FC layers. For the GoogLeNet, we use the same weight initialization scheme. We use a learning rate multiplier of for the layers before the Inception_5a module, for the Inception_5a and Inception_5b modules, and for the last FC layer for more aggressive learning.We train the AlexNet with Adam, base learning rate , and dropout rate . We train the GoogLeNet with Vanilla SGD, base learning rate , and dropout rate .
Mass context
The area surrounding a mass provides useful context for diagnosis. We explore two approaches for providing the network with context. In the first, our input to the network is the region including
pixels of fixed padding around the mass, providing a context size independent of mass dimensions (referred to as
Small Context). In the second approach we use proportional padding by extracting a region two times the size of the mass bounding box (referred to as Large Context).Data Augmentation
We study the impact of data augmentation to alleviate to the relatively small size of our training set, which is characteristic of many medical image datasets. We use rotation, cropping, and mirroring transformations to increase the effective size of our dataset (referred to as Aug). For each training image, we perform 5 random rotations and sample 5 random crops per rotation offline, effectively increasing training set size by a factor of 25. We also perform random mirroring at train time. These augmentations are justified since masses have no inherent orientation and their diagnosis is invariant to these transformations.
5 Results
We first present empirical analysis of our model design using the AlexNet base architecture, then show quantitative results of our best models. Finally we use techniques for visualizing saliency maps to provide interpretability of the model. All experiments are implemented with Caffe
Jia et al. (2014), and the analysis and results are presented on the validation and test sets, respectively.5.1 Empirical analysis
Effectiveness of transfer learning
CNN-based image representations learned on large-scale annotated datasets have proven to to be a useful form of pre-training that can be effectively transferred to other computer vision tasks with limited training data Oquab et al. (2014). More recently, low-level features learned from natural images have shown to be effective for medical image classification Bar et al. (2015); Shin et al. (2016); van Ginneken et al. (2015). In Table 1, we strongly confirm this claim by demonstrating that a fine-tuned AlexNet significantly outperforms our baseline model.
Model | Validation Accuracy |
---|---|
Baseline(Aug-Large Context) | |
AlexNet(Aug-Large Context) |
Influence of context
To understand the influence of context around masses, we fine-tune an AlexNet on two different datasets - one with fixed padding and the other with proportional padding. The results in Table 2 show that proportional padding contains greater signal for classification, and we consequently use this for the remaining experiments.
Model | Validation Accuracy |
---|---|
AlexNet(No Aug-Small Context) | |
AlexNet(No Aug-Large Context) |
Influence of data augmentation
Limited amounts of training data is a common bottleneck in machine learning applications to medical problems. We evaluate the utility of data augmentation schemes to increase the effective amount of training data and reduce overfitting. The training loss curves in Fig.
2 show that our data augmentation technique described in Sec. 4.2 successfully regularises the network and helps remedy the scarcity of data.
5.2 Performance
Our final results using the model choices described in Sec. 5.1 and all base architectures are presented in Table 3. The GoogLeNet outperforms the other models by a fair margin. It is also more suited for fine-tuning and less prone to overfitting due to its relatively small number of parameters, approximately million compared to million for AlexNet.
An important metric for diagnostic applications is maximizing recall, since the cost of a false negative (patient remaining undiagnosed) is much higer than a false positive (an additional biopsy). Our best model achieves recall at precision, outperforming human performance in a study that shows radiologist recall between and Elmore et al. (2009). This result is very promising for real-life use of such models in clinical practice.
Model | Accuracy | Precision | Recall | # Epochs |
---|---|---|---|---|
Baseline (Aug-Large Context) | ||||
AlexNet (Aug - Large Context) | ||||
GoogLeNet (Aug - Large Context) |
5.3 Interpretability
Deep learning models often lack interpretability and as such are hard to adopt for practical use in medical settings. Simonyan et al. (2013) describe a methodology to visualize saliency maps which show the regions of an image the network is sensitive to when making predictions. This is performed by computing the gradient of the image with respect to the unnormalized class scores. Regions with larger gradient indicate higher contribution to the prediction (brighter in Fig. 3). Both the AlexNet and GoogLeNet learn to attend to the edges of the mass, which is a high-signal criterion for diagnosis, while also paying attention to context.
![]() |
![]() |
6 Conclusion
In this work, we propose an end-to-end deep learning model to classify pre-detected breast masses from mammograms. We show how careful pre-processing, data augmentation and transfer learning can overcome the data bottleneck common to medical computer vision tasks, and additionally provide a method to give more interpretability to network predictions.
Our approach obtains state-of-the-art results, outperforming trained radiologists, and the interpretability enables more comfortable adoption in real-world settings. Future work includes exploring other architectures, and integration of attention mechanisms which are more difficult to train but could provide even more concrete interpretability.
Acknowledgements
We thank Justin Johnson for continuous feedback and guidance throughout this project as well as Serena Yeung for insightful comments on the draft. The authors also ackowledge the support of AWS Educate program for generously providing free instances with GPUs.
References
- Arevalo et al. (2016) J. Arevalo, F. A. González, R. Ramos-Pollán, J. L. Oliveira, and M. A. G. Lopez. Representation learning for mammography mass lesion classification with convolutional neural networks. Computer methods and programs in biomedicine, 127:248–257, 2016.
- Ball and Bruce (2007) J. E. Ball and L. M. Bruce. Digital mammographic computer aided diagnosis (cad) using adaptive level set segmentation. In 2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, pages 4973–4978. IEEE, 2007.
- Bar et al. (2015) Y. Bar, I. Diamant, L. Wolf, S. Lieberman, E. Konen, and H. Greenspan. Chest pathology detection using deep learning with non-medical training. In 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), pages 294–297. IEEE, 2015.
- Bowyer et al. (1996) K. Bowyer, D. Kopans, W. Kegelmeyer, R. Moore, M. Sallam, K. Chang, and K. Woods. The digital database for screening mammography. In Third international workshop on digital mammography, volume 58, page 27, 1996.
- Boyle et al. (2008) P. Boyle, B. Levin, et al. World cancer report 2008. IARC Press, International Agency for Research on Cancer, 2008.
- Carneiro et al. (2015) G. Carneiro, J. Nascimento, and A. P. Bradley. Unregistered multiview mammogram analysis with pre-trained deep learning models. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 652–660. Springer International Publishing, 2015.
- DeSantis et al. (2014) C. DeSantis, J. Ma, L. Bryan, and A. Jemal. Breast cancer statistics, 2013. CA: a cancer journal for clinicians, 64(1):52–62, 2014.
- Dhungel et al. (2015a) N. Dhungel, G. Carneiro, and A. P. Bradley. Automated mass detection in mammograms using cascaded deep learning and random forests. In Digital Image Computing: Techniques and Applications (DICTA), 2015 International Conference on, pages 1–8. IEEE, 2015a.
- Dhungel et al. (2015b) N. Dhungel, G. Carneiro, and A. P. Bradley. Deep learning and structured prediction for the segmentation of mass in mammograms. In International Conference on Medical Image Computing and Computer-Assisted Intervention, pages 605–612. Springer International Publishing, 2015b.
- Elmore et al. (2009) J. G. Elmore, S. L. Jackson, L. Abraham, D. L. Miglioretti, P. A. Carney, B. M. Geller, B. C. Yankaskas, K. Kerlikowske, T. Onega, R. D. Rosenberg, et al. Variability in interpretive performance at screening mammography and radiologists’ characteristics associated with accuracy 1. Radiology, 253(3):641–651, 2009.
-
Girshick et al. (2014)
R. Girshick, J. Donahue, T. Darrell, and J. Malik.
Rich feature hierarchies for accurate object detection and semantic
segmentation.
In
Proceedings of the IEEE conference on computer vision and pattern recognition
, pages 580–587, 2014. - Glorot and Bengio (2010) X. Glorot and Y. Bengio. Understanding the difficulty of training deep feedforward neural networks. In Aistats, volume 9, pages 249–256, 2010.
- Ioffe and Szegedy (2015) S. Ioffe and C. Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167, 2015.
- Jia et al. (2014) Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell. Caffe: Convolutional architecture for fast feature embedding. In Proceedings of the 22nd ACM international conference on Multimedia, pages 675–678. ACM, 2014.
- Kingma and Ba (2014) D. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Krizhevsky et al. (2012) A. Krizhevsky, I. Sutskever, and G. E. Hinton. ImageNet Classification with Deep Convolutional Neural Networks. In NIPS, 2012.
- Long et al. (2015) J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3431–3440, 2015.
- Oquab et al. (2014) M. Oquab, L. Bottou, I. Laptev, and J. Sivic. Learning and transferring mid-level image representations using convolutional neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1717–1724, 2014.
- Russakovsky et al. (2015) O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, et al. Imagenet large scale visual recognition challenge. International Journal of Computer Vision, 115(3):211–252, 2015.
- Shin et al. (2016) H.-C. Shin, H. R. Roth, M. Gao, L. Lu, Z. Xu, I. Nogues, J. Yao, D. Mollura, and R. M. Summers. Deep convolutional neural networks for computer-aided detection: Cnn architectures, dataset characteristics and transfer learning. IEEE transactions on medical imaging, 35(5):1285–1298, 2016.
- Simonyan et al. (2013) K. Simonyan, A. Vedaldi, and A. Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034, 2013.
- Szegedy et al. (2015) C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1–9, 2015.
- van Ginneken et al. (2015) B. van Ginneken, A. A. Setio, C. Jacobs, and F. Ciompi. Off-the-shelf convolutional neural network features for pulmonary nodule detection in computed tomography scans. In 2015 IEEE 12th International Symposium on Biomedical Imaging (ISBI), pages 286–289. IEEE, 2015.
- Wang et al. (2016) J. Wang, X. Yang, H. Cai, W. Tan, C. Jin, and L. Li. Discrimination of breast cancer with microcalcifications on mammography by deep learning. Scientific reports, 6:27327, 2016.
Comments
There are no comments yet.