Direct detection of pixel-level myocardial infarction areas via a deep-learning algorithm

06/10/2017 ∙ by Chenchu Xu, et al. ∙ 0

Accurate detection of the myocardial infarction (MI) area is crucial for early diagnosis planning and follow-up management. In this study, we propose an end-to-end deep-learning algorithm framework (OF-RNN ) to accurately detect the MI area at the pixel level. Our OF-RNN consists of three different function layers: the heart localization layers, which can accurately and automatically crop the region-of-interest (ROI) sequences, including the left ventricle, using the whole cardiac magnetic resonance image sequences; the motion statistical layers, which are used to build a time-series architecture to capture two types of motion features (at the pixel-level) by integrating the local motion features generated by long short-term memory-recurrent neural networks and the global motion features generated by deep optical flows from the whole ROI sequence, which can effectively characterize myocardial physiologic function; and the fully connected discriminate layers, which use stacked auto-encoders to further learn these features, and they use a softmax classifier to build the correspondences from the motion features to the tissue identities (infarction or not) for each pixel. Through the seamless connection of each layer, our OF-RNN can obtain the area, position, and shape of the MI for each patient. Our proposed framework yielded an overall classification accuracy of 94.35 results indicate the potential of our proposed method in aiding standardized MI assessments.



There are no comments yet.


page 5

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

There is a great demand for detecting the accurate location of a myocardial ischemia area for better myocardial infarction (MI) diagnosis. The use of magnetic resonance contrast agents based on gadolinium-chelates for visualizing the position and size of scarred myocardium has become ‘the gold standard’ for evaluating the area of the MI [1]. However, the contrast agents are not only expensive but also nephrotoxic and neurotoxic and, hence, could damage the health of humans. [2]

. In routine clinical procedures, and especially for early screening and postoperative assessment, visual assessment is one popular method, but it is subject to high inter-observer variability and is both subjective and non-reproducible. Furthermore, the estimation of the time course of the wall motion remains difficult even for experienced radiologists.

Therefore, computer-aided detection systems have been attempted in recent years to automatically analyze the left ventricle (LV) myocardial function quantitatively. This computerized vision can serve to simulate the brain of a trained physician s intuitive attempts at clinical judgment in a medical setting. Previous MI detection methods have been mainly based on information theoretic measures and Kalman filter approaches


, Bayesian probability model


, pattern recognition technique

[5][6], and biomechanical approaches [7]

. However, all of these existing methods still fail to directly and accurately identify the position and size of the MI area. More specifically, these methods have not been able to capture sufficient information to establish integrated correspondences between the myocardial motion field and MI area. More recently, unsupervised deep learning feature selection techniques have been successfully used to solve many difficult computer vision problems. The general concept behind deep learning is to learn hierarchical feature representations by first inferring simple representations and then progressively building up more complex representations from the previous level. This method has been successfully applied to the recognition and prediction of prostate cancer, Alzheimer s disease, and vertebrae and neural foramina stenosis


In this study, an end-to-end deep-learning framework has been developed for accurate and direct detection of infarction size at the pixel level using cardiac magnetic resonance (CMR) images. Our method s contributions and advantages are as follows: (1) for the first time, we propose an MI area detection framework at the pixel level that can give the physician the explicit position, size and shape of the infarcted areas; (2) a feature extraction architecture is used to establish solid correspondences between the myocardial motion field and MI area, which can help in understanding the complex cardiac structure and periodic nature of heart motion; and (3) a unified deep-learning framework can seamlessly fuse different methods and layers to better learn hierarchical feature representations and feature selection. Therefore, our framework has great potential for improving the efficiency of the clinical diagnosis of MI.

Figure 1: The architecture of OF-RNN: heart localization layers, motion statistical layers, and fully connected discriminate layers.

2 Methodology

As shown in Fig. 1, there are three function layers inside the OF-RNN. The heart localization layers can automatically detect the ROI, including the LV, and the motion statistical layers can generate motion features that accurately characterize myocardial physiologic and physical function, followed by the fully connected discriminate layers that use stacked auto-encoders and softmax classifiers to detect the MI area from motion features.
Heart localization layers. One FAST R-CNN [9]

is used here for the automatic detection of a region of interest (ROI) around the LV, to reduce the computational complexity and improve the accuracy. In this study, the first process of the heart localization layers is to generate category-independent region proposals. Afterward, a typical convolutional neural network model is used to produce a convolution feature map by input images. Then, for each object proposed, an ROI pooling layer extracts a fixed-length feature vector from the feature map. The ROI pooling layer uses max pooling to convert the features inside any valid region of interest into a small feature map with a fixed spatial extent of

, where and are layer hyper-parameters that are independent of any particular ROI. Finally, each feature vector is fed into a sequence of fully connected layers that branch into two sibling output layers, thereby generating a 6464 bounding-box for cropping the ROI image sequences, including the LV from CMR sequences.
Motion statistical layers. The motion statistical feature layers are used to extract time-series image motion features through ROI image sequences to understand the periodic nature of ghd heart motion. The local motion features are generated by LSTM-RNN, and the global motion features are generated by deep optical flow. Thus, in the first step, we attempt to compute the local motion features that are extracted from the ROI image sequence. For each ROI sequence, the input image of size 6464, represents a pixel coordinate of the image . A window of size 1111 is constructed for the overlapping neighborhoods, which has an intensity value that is representative of the feature of each on image . This approach results in the image sequence features being unrolled as vector for each pixel as input. Then, four layers of RNN [10] with LSTM cells layers are used to learn the input. Give the input layer at time , each time corresponds to each frame(t=J), which indicates that at frame , and for the hidden state frame of the previous time step ,the hidden and output layers for the current time step are computed as follows:


where , and are layers that represent the input, hidden, and output at each time step , respectively; and are the matrices that denote the weights between the input and hidden layers and between the hidden and output layers, respectively, and

denotes the activation function. The LSTM cell

[10] is designed to mitigate the vanishing gradient. In addition to the hidden layer vector , the LSTMs maintain a memory vector , an input gate , a forget gate , and an output gate ; These gates in the LSTMs are computed as follows:


where is the weight matrix, and is the dropout operator. The final memory cell and the final hidden state are given by


In the second step, we attempt to compute the global motion feature of the image sequence based on an optical flow algorithm [11] by the deep architecture. An optical flow can describe a dense vector field, where a displacement vector is assigned to each pixel, which points to where that pixel can be found in another image. Considering an adjacent frame, a reference image = () and a target image = (), the goal is to estimate the flow

that contains both horizontal and vertical components. We assume that the images are already smoothed by using a Gaussian filter with a standard deviation of

. The energy to be optimized is the weighted sum of a data term , a smoothness term , and a matching term :


Next, a procedure is developed to produce a pyramid of response maps, and we start from the optical flow constraint, assuming a constant brightness. A basic way to build a Data term and a Smoothness term is the following:


where is a robust penalizer;

is the tensor for channel

; and are the two balanced weights. The matching term encourages the flow estimation to be similar to a precomputed vector field , and a term has been added.


For any pixel of , is a measure of similarity between and . We have to be a patch size of N N(N4, 8, 16) from the first image centered at . We start with the bottom-level correlation maps, which are iteratively aggregated to obtain the upper levels. This aggregation consists of max-pooling, sub-sampling, computing a shifted average and non-linear rectification. In the end, for each image , a fully motion field is computed with reference to the next frame .
Fully connected discriminate layers. The fully connected discriminate layers are used to detect the MI area accurately from the local motion features and the global motion features. First, for each , we use image patches, say 33, by extracting the feature beginning from a point in the first frame and tracing in the following frame. We can thereby obtain while containing a 33 vector for displacement and a 33 vector for the orientation of for each frame. Second, we conduct a simple concatenation between the local image feature from the LSTM-RNN and the motion trajectories feature via optical flow, to establish a whole feature vector . Finally, an auto-encoder with three stacking layers is used for learning the

, followed by a softmax layer, which is used to determine whether

belongs to the MI area or not.

Figure 2: (a-b) Our predicted MI area (the green zone) can be a good fit for the ground truth (the yellow arrow) (c) our predicted MI area (the green zone) can be a good fit for the ground truth (the yellow dotted line).

3 Experimental results

Data acquisition. We collected the short axis image dataset and the corresponding enhanced images using gadolinium agents from 114 subjects in this study on a 3T CMR scanner. Each subject s short-axis image dataset consisted of 25 2D images (a cardiac cycle), a total of 43 apical, 37 mid-cavity and 34 basal short-axis image datasets for 114 subjects. The temporal resolution is 45.18.8 ms, and the short-axis planes are 8-mm thick. The delayed enhancement images were obtained approximately 20 min after intravenous injection of 0.2 mmol/kg gadolinium diethyltriaminepentaacetic acid. A cardiologist (with more than 10 years of experience) analyzed the delayed enhancement images and manually traced the MI area by the pattern of late gadolinium enhancement as the ground truth.
Implementation details.

We implemented all of the codes using Python and MATLAB R2015b on a Linux (Kylin 14.04) desktop computer with an Intel Xeon CPU E5-2650 and 32 GB DDR2 memory. The graphics card is an NVIDIA Quadro K600, and the deep learning libraries were implemented with Keras (Theano) with RMSProp solver. The training time was 373 minutes, and the testing time was 191 seconds for each subject (25 images).

Performance evaluation criteria. We used three types of criteria to measure the performance of the classifier: 1) the receiver operating characteristic (ROC) curve; 2) the precision-recall (PR) curve; 3) for pixel-level accuracy, we assessed the classifier performance with a 10-fold cross-validation test, and for segment-level accuracy, we used 2/3 data for training and the remaining data for testing.
Automatic localization of the LV. The experiment’s result shows that OF-RNN can obtain good localization of the LV. We achieve an overall classification accuracy of 96.49%, with a sensitivity of 94.39% and a specificity of 98.67%, in locating the LV in the heart localization layers. We used an architecture similar to the Zeiler and Fergus model to pre-train the network. Using selective search s quality mode, we sweep over 2k proposals per image. Our results for the ROI localization bounding-box from 2.85k CMR images were compared to the ground truth marked by the expert cardiologist. The ROCs and PRs curves are shown in Fig. 3 (a-b).
MI area detection. Our approach can also accurately detect the MI area, as shown in Fig. 2. The overall pixel classification accuracy is 94.35%, with a sensitivity of 91.23% and a specificity of 98.42%. We used the softmax classifier by fine-tuning the motion statistical layers to assess each pixel (as normal/abnormal). We also compared our results to 16 regional myocardial segments (depicted as normal/abnormal) by following the American Heart Association standards. The accuracy performance for the apical slices was an average of 99.2%; for the mid-cavity slices, it was an average of 98.1%; and for the basal slices, an average of 97.9%. The ROCs and PRs of the motion statistical layers are shown in Fig. 3 (a-b).

Figure 3: (a-b) ROCs and PRs show that our results have good classification performance. (c-d) ROCs and PRs for local motion features and global motion features. (e) The accuracy and time for various patch sizes.

Local and global motion statistical features. A combination of local and global motion statistical features has the potential to improve the results because the features influence one another through a shared representation. To evaluate the effect of motion features, we use local or global motion statistical features separately along with both motion features in our framework. Table 1 and Fig. 3 (c-d) show that the results that combine motion statistical features in our framework have better accuracy, sensitivity, and specificity in comparison to those that use only the local or global motion features, in another 10-fold cross-validation test.

local motion feature
global motion feature
accuracy 92.6% 87.3% 94.3%
sensitivity 86.5% 79.4% 91.2%
specificity 97.9 96.2% 98.4%
Table 1: Combined motion statistical features effectively improve the overall accuracy of our method

Size of patch. We use an NN patch to extract the local motion features from the whole image sequence. Because the displacements of the LV wall between two consecutive images are small (approximately 1 or 2 pixels/frame), it is necessary to adjust the size of the patch to capture sufficient local motion information. Fig. 3 (e) shows the accuracy and computational time of our framework, using from 33 to 1717 patches in one 10-fold cross-validation test. We find that the 1111 patch size in our framework can obtain better accuracy in a reasonable amount of time.
Performance of the LSTM-RNN. To evaluate the performance of the LSTM-RNN, we replaced the LSTM-RNN using SVMrbf, SAE-3, DBN-3, CNN and RNN in our deep learning framework, and we ran these different frameworks over 114 subjects using a 10-fold cross-validation test. Table 2

reports the classification performance by using the other five different learning strategies: the RNN, Deep Belief Networks (DBN), Convolutional Neural Network (CNN), SAE and Support Vector Machine with RBF kernel (SVMrbf). LSTM-RNN shows better accuracy and precision in all of the methods.

Figure 4: A pair of frames at the beginning of systole (a) and at the end of systole (b) were first displayed, followed by the visual results of our deep optical flow (c) and Horn and Schunck(HS) optical flow (d) at pixel precision.
Accuracy 80.9% 83.5%% 84.9% 83.7% 88.4% 94.3%
Precision 74.2% 75.5% 75.1% 76.5% 84.8% 91.3%
Table 2: LSTM-RNN works best in comparison with other models.

Performance of the optical flow. The purpose of the optical flow is to capture the global motion features. To evaluate the performance of our optical flow algorithm with a deep architecture, we used the average angular error (AAE) to evaluate our deep optical flow and other optical flow approaches. The other optical flow methods, including the Horn and Schunck method, pyramid Horn and Schunck method, intensity-based optical flow method, and phase-based optical flow method, can be found in [12]. The comparison results are shown in Table 3, and visual examples are illustrated in Fig. 4.

Horn and Schunck(HS) Pyramid HS Deep OF Intensity-based OF Phase-based OF
OF density 100% 100% 100% 55% 13%
AAE 12.69.2 7.43.4 5.72.3 5.74.1 5.53.9
Table 3: Deep optical flow (OF) can work better in comparison to other optical flow techniques in capturing global motion features.

4 Conclusions

We have, for the first time, developed and presented an end-to-end deep-learning framework for the detection of infarction areas at the pixel level from CMR sequences. Our experimental analysis was conducted on 114 subjects, and it yielded an overall classification accuracy of 94.35% at the pixel level. All of these results demonstrate that our proposed method can aid in the clinical diagnosis of MI assessments.


This work was supported in part by the Shenzhen Research and Innovation Funding (JCYJ20151030151431727, SGLH20150213143207911), the National Key Research and Development Program of China (2016YFC1300302, 2016YFC13017
00), the CAS President s International Fellowship for Visiting Scientists (2017VTA0011), the National Natural Science Foundation of China (No.61673020), the Provincial Natural Science Research Program of Higher Education Institutions of Anhui province (KJ2016A016) and the Anhui Provincial Natural Science Foundation (1708085QF143).


  • [1] örg Barkhausen, J., Ebert, W., Weinmann, H.J.: Imaging of myocardial infarction: comparison of magnevist and gadophrin-3 in rabbits. Journal of the American College of Cardiology 39(8), 392–1398 (2002).
  • [2] Wagner, A., Mahrholdt, H., Holly, T.: Contrastenhanced mri detects subendocardial myocardial infarcts that are missed by routine spect perfusion imaging. Lancet 361, 374–379, (2003).
  • [3] Shi, P., Liu, H.: Stochastic finite element framework for simultaneous estimation of cardiac kinematic functions and material parameters. Medical Image Analysis 7(4), 445–464 (2003).
  • [4] Wang, Z., Salah, M, B., Gu, B., Islam, A., Goela, A., Shuo, L.: Direct estimation of cardiac biventricular volumes with an adapted bayesian formulation. IEEE Transactions on Biomedical Engineering 61(4), 1251–1260 (2014).
  • [5] Afshin, M.: Assessment of regional myocardial function via statistical features in MR images. International Conference on Medical Image Computing and Computer-Assisted Intervention, 107–114 (2011).
  • [6] Zhen, X., Islam, A., Bhaduri, M.,Chan, I., Shuo, L.: Direct and simultaneous four-chamber volume estimation by multi-output regression. International Conference on Medical Image Computing and Computer-Assisted Intervention, 669–676 (2015).
  • [7] Wong, K.C.: Computer-aided infarction identification from cardiac CT images: a biomechanical approach with SVM. International Conference on Medical Image Computing and Computer-Assisted Intervention, 144–151 (2015).
  • [8] Cai, Y.: Multi-modal vertebrae recognition using transformed deep convolution network. Computerized Medical Imaging and Graphics 51, 11–19 (2016).
  • [9] Girshick, R.: Fast R-CNN. IEEE International Conference on Computer Vision, 1440–1448 (2015).
  • [10] Graves, Alex.: Supervised sequence labelling. In: Supervised Sequence Labelling with Recurrent Neural Networks, pp. 5–13. Springer, Heidelberg (2012).
  • [11] Revaud, J., Weinzaepfel, P., Harchaoui, Z., Schmid, C.: Deepmatching: Hierarchical deformable dense matching. International Journal of Computer Vision 120(3), 300–323 (2016).
  • [12] Fortun, Denis., Bouthemy, Patrick., Kervrann, Charles.: Optical flow modeling and computation: a survey. Computer Vision and Image Understanding 134, 1–21 (2015).