There has been increasing interest in modelling survival data using deep learning methods in medical research. Current approaches have focused on designing special cost functions to handle censored survival data. We propose a very different method with two steps. In the first step, we transform each subject's survival time into a series of jackknife pseudo conditional survival probabilities and then use these pseudo probabilities as a quantitative response variable in the deep neural network model. By using the pseudo values, we reduce a complex survival analysis to a standard regression problem, which greatly simplifies the neural network construction. Our two-step approach is simple, yet very flexible in making risk predictions for survival data, which is very appealing from the practice point of view. The source code is freely available at http://github.com/lilizhaoUM/DNNSurv.READ FULL TEXT VIEW PDF
We propose a novel approach for estimating mean survival time in the pre...
This manuscripts develops a new class of deep learning algorithms for
There is currently great interest in applying neural networks to predict...
Many online platforms have deployed anti-fraud systems to detect and pre...
In the following short article we adapt a new and popular machine learni...
Accurate models of patient survival probabilities provide important
In this paper, we consider survival analysis with right-censored data wh...
Recently, using deep neural networks to predict when an event of interest will happen has gained considerable attention. These studies are often characterized by incomplete observations, in particular right-censored data; e.g. patients may be lost to follow-up without experiencing the event. Thus, handling the censored data becomes the crucial aspect of these analyses.
developed a neural network assuming that survival time has a Weibull distribution. This parametric assumption is too restrictive in the analysis of large datasets. The semi-parametric Cox proportional hazards regression is the most widely used method for analyzing censored survival data. It studies the effects of the covariate variables on survival by estimating hazard functions, and it assumes that the hazard ratio for any two subjects is constant over time; that is, the proportional hazards (PH) assumption. Several authors have adapted the Cox model to the deep neural networks[Katzman:2018, Ching:2018, Zhu:2016]. They trained the deep neural network models by minimizing the negative partial likelihood defined under the PH assumption. However, this assumption is often questionable when the number of covariates is large, as every covariate needs to satisfy the PH assumption.
Another school of deep neural network modelling for survival data utilized the discrete-time survival model, including recent papers by [Gensheimer:2018, Fotso:2018, Lee:2018, Luck:2017, Giunchiglia:2018]. In this modelling framework, the follow-up time is divided into a set of fixed intervals. For each time interval, the conditional hazard/probability is estimated: the probability of failure in the interval, given that the individual has survived up to the beginning of the interval. Compared to the Cox model, the discrete-time survival model is more flexible as it does not rely on the PH assumption. Furthermore, it directly estimates the survival probability, which is often of direct interest to patients and physicians than the hazard function in the Cox model. However, the censored data in each interval are handled either by an oversimplified method that assumes individuals with a censoring time in the second half of an interval survived that interval [Gensheimer:2018]
, or by an overcomplicated method that uses a ranking loss function[Lee:2018] or a combination of a ranking loss and an isotonic regression [Luck:2017].
All existing deep neural network models referenced above require the development of a special cost function to handle the censored data, sometimes with a sophisticated network structure, such as a convolutional or recurrent neural network[Zhu:2016, Giunchiglia:2018]. In this article, we develop a very different approach. To circumvent the complexity introduced by censored data, we substitute the observed survival times by jackknife
pseudo observations, and then use these pseudo observations as a quantitative response variable in a regression analysis powered by the deep neural networks.
The remainder of this article is organized as follows. In section 2, we propose our two-step methods and address the covariate-dependent censoring problem. In section 3, we conduct simulation studies to evaluate our methods. In section 4, we apply our methods to two prospective, multicenter cohort studies for cardiovascular disease. We conclude this paper with a brief discussion in section 5.
Notation. Let be the survival time for subject , be the censoring time, be the observed survival time, and be the censoring indicator. We assume that survival times and censoring times are independent. Let denote the -dimensional covariates. We first assume that the censoring distribution does not depend on the covariates and then address the covariate-dependent problem in section 2.5.
The pseudo-observations approach [Andersen:2003, Klein:2005, Klein:2007, Andersen:2010] provides an efficient and straightforward way to study the association between the covariates and survival outcome in the presence of censoring.
If the outcome of interest is the survival probability a specific time , without incomplete data, we could directly model on () using a generalized linear model (GLM) with a logit link, for a binary outcome variable. In the presence of censoring, is not observed for all subjects. In this case the Kaplan-Meier (KM) estimator can be used to estimate the survival probability at any given time point. The KM estimator is approximately unbiased under independent censoring [Andersen:1993], which is a requirement for the validity of the pseudo-observations approach. Based on the jackknife idea, a pseudo survival probability is computed for each (censored and uncensored) subject. For the subject, the pseudo survival probability is computed by
where is the KM estimator of using all subjects and is the KM estimator using sample size of by eliminating the subject. Then () are used as a numeric response variable in the standard regression analysis, which is similar to model fit to if these values were observed.
[Klein:2005, Klein:2007, Klein:2008]
proposed computing a vector of pseudo survival probabilities at a finite number of time points (equally spread on the event time scale) for each subject, and then modelling these pseudo survival probabilities as a function of the covariates by the generalized estimating equation (GEE), with thecomplementary log-log link, which is equivalent to fitting a Cox model to the survival data. The regression coefficient estimates from this pseudo-based GEE is approximately consistent when the censoring distribution is independent of the covariates and of the survival times [Andersen:2003].
In this article, we adapt the pseudo-observations approach in a discrete-time survival framework. We first divide the follow-up time into intervals: ,, . For each interval, we compute the pseudo conditional survival probability: the probability of surviving the interval, given that the subject has survived the previous interval. For a given interval , all subjects who are still at risk is denoted by and the pseudo conditional probability of surviving is computed as
where is the KM estimator constructed using the remaining survival times for all patients still at-risk at time and is the KM estimator for all patients at-risk but the subject. For the first interval all subjects are at risk (i.e., ). If there is no censored data, the pseudo probability in each interval is either 0 or 1. With censoring, the pseudo probability is a real value; that is, it can be above 1 or below 0; see properties of the pseudo observations discussed in [Andersen:2010].
By using the discrete-time survival framework, we transform each subject’s observed survival time (censored or uncensored) into a series of pseudo conditional survival probabilities. Unlike the pseudo-based GEE, in which the pseudo probabilities are computed from the marginal survival function (i.e., defined from time zero), the pseudo probabilities in our approach are computed from the conditional survival function and they are independent. Thus, we avoid the complexity of considering the within-subject correlation as in the pseudo-based GEE.
Choosing the time intervals needs to be done carefully. It has been shown that as few as five time intervals equally spread on the event time scale worked quite well in most cases [Klein:2007]. One could also divide the time based on time points of clinical relevance; for example, patients’ prognosis at 5 and 10 years might be of particular interest to the investigators. In either case, the interval can not be too small or too large. If the interval is too large, data information is lost. Conversely, if the interval is too small, no or a few events would occur in the interval, leading to an inefficient analysis. As fewer subjects remain in the study at later follow-up time, the upper bound of the last interval, , should be reasonably smaller than the maximum follow-up time, so that we have sufficient information in the last interval. In this article, we set at the or the percentile of the empirical overall survival distribution.
We created a R function called getPseudoConditional to compute the pseudo conditional probabilities described above. The R code is available in the GitHub repository http://github.com/lilizhaoUM/DNNSurv/blob/master/getPseudoConditional.R. Table 1 is an example output for the first three subjects. Each subject has multiple pseudo probabilities and one for each time point, but the number of pseudo probabilities may vary as different subjects might have different follow-up times. The covariate, has the same value for the same subject. The outputs from getPseudoConditional are then used as inputs in a deep neural network model.
By using pseudo observations, the complex survival data analysis is reduced to a standard regression analysis with a single quantitative response variable. Thus, we directly employed the conventional cost function of the sum of squared errors for a regression problem in the deep neural network model. The input layer has nodes, including covariates and one variable for the time points (i.e., the second column in Table 1
). The output of the network is a single node, which estimates the survival probabilities at pre-specified time points. To constrain the probability between 0 and 1, we used the sigmoid activation function. We call this pseudo-based deep neural network model as DNNSurv and implemented it in the Keras library[chollet2015keras]
in R with Tensorflow[tensorflow2015-whitepaper] backend (code at http://github.com/lilizhaoUM/DNNSurv).
To determine the hyperparameters in DNNSurv, we used the Cardiovascular Health Study (CHS) data, which contains 5380 subjects (see details in the REAL APPLICATION section). We determined the network structure and hyperparameters by using 5-fold cross-validation (CV) on the 75% randomly selected training set and evaluated the performance on the remaining 25% holdout test set.
We first experimented networks with different numbers of hidden layers, activation functions, optimization algorithms and regularization approaches available in Keras. Specifically, we compared network structures with one hidden layer and two hidden layers (4, 8, 16, 32, 64, or 128 nodes in each layer), under both the dropout regularization [Srivastava:2014] (a dropout rate from 0.1 to 0.8 in each layer) and the ridge regularization [Ching:2018]
(a regularization penalty of 0.0001, 0.001, or 0.01 in each layer). For the activation function, we compared relu and tanh[chollet2015keras]chollet2015keras], with a learning rate of 0.001, 0.005 or 0.01 in both optimizers. Using 5-fold CV with the c-index as the performance metric, we found that a network with one or two hidden layers had similar performances; the dropout and ridge regularization had similar performances; tanh function had better performance than relu activation function, and Adam had better performance than rmsprop optimizer. Therefore, we decided on a network structure with two hidden layers, the tanh activation function in each hidden layer, the ridge regularizer and the Adam optimizer.
Next, we used a random Cartesian grid search [bergstra2012] by a 5-fold CV to further tune the above hyperparameters, including number of nodes in each hidden layer, the regularization parameter in the ridge regularizer and the learning rate of the Adam optimizer. Finally, we determined a set of default values for the hyperparamters (see details in Table S1), which worked well in all simulation studies (see SIMULATIONS section). We recommend that users start with the default hyperparamters and adjust them if necessary.
Given a set of input values, DNNSurv is able to predict the conditional survival probability in each interval, that is, for . The marginal survival probability, is calculated by multiplying the conditional survival probabilities up to the interval:
Thus, the discrete-time survival framework allows us to predict both the marginal and the conditional survival probability, or the complementary risks. Both the marginal and conditional estimates have important clinical implications. For example, a patient who is diagnosed with lung cancer might be interested in the probability of surviving one year (i.e., the marginal survival probability). If the patient has survived the first year, then he/she might be interested in the probability of surviving another year (i.e., the conditional survival probability).
KM estimates used in creating the pseudo values are subject to covariate-dependent censoring bias. In this case, we propose to use the inverse of probability of censoring weighted (IPCW) estimator for the survival function, denoted by , which has been successful at reducing the bias [Binder:2014, Xiang:2012]). We replace by in (2) to compute the IPCW pseudo conditional survival probabilities by
where and is the IPCW Nelson-Aalen estimator for the the cumulative hazard function and is estimated by
where is the observable counting process for subject is the at risk process for subject , and is the inverse of probability of censoring for subject at time . By assigning different weights for subjects based on their covariate values, is approximately unbiased if the censoring distribution is correctly specified [Binder:2014]. Calculation of the IPCW pseudo conditional survival probabilities is also implemented in the getPseudoConditional function.
To model the censoring time distribution, we consider the Cox model, but other models, such as accelerated failure time model (AFT) [wei1992] or deep neural networks, could work as well. In calculating the pseudo probabilities, we fit the censoring model once and then use the same censoring model for all subjects; that is, () remain the same in computing the pseudo probabilities in (3). Given these IPCW pseudo probabilities, we apply the same deep neural network model as described in section 2.3 for the survival probability estimation. We refer this model as DDNSurv_ipcw in this article.
We conducted simulation studies to evaluate our models in the case of independent and covariate-dependent censoring, respectively. We compared our models to Cox-nnet [Ching:2018], which is a neural network trained by minimizing the negative partial likelihood defined under the PH assumption; see comparisons of the hyperparameters between DNNSurv and Cox-nnet in Table S1. We chose Cox-nnet for comparison as it has undergone peer review and has well documented source code, which can be found at https://github.com/lanagarmire/cox-nnet.
To evaluate the model performance, we considered two metrics commonly used in survival analysis. The first metric is the time-dependent concordance index (c-index) [Harrell:1996], which measures of how well a model predicts the ordering of sample event times. The other metric is the Brier score [Gerds:2006], which evaluates the accuracy of a predicted survival probability at a given time point by measuring the average squared distances between the observed survival status and the predicted survival probability (the smaller the better). Cox-nnet does not provide an estimate of the survival probability as DNNSurv; instead, it outputs a log hazard ratio () for each subject. To compute the survival probability at a time point, we first computed the Nelson-Aalen estimate of the baseline survival function, , and then computed the survival probability for subject by .
Survival data were generated from an AFT model. Since the deep neural network model is able to approximate complex nonlinear functions, we considered the flexible random function generator in [friedman2001] for the mean function in the AFT model [henderson2017]. For all of the studies presented in this subsection, the number of variables is (i.e.,
’s were generated from a uniform distribution. Each is a Gaussian function of a randomly selected variables, of the variables; see details in [friedman2001] on the selection of these variables and the Gaussian function construction. The expected number of variables for each function was 4. By using a linear combination of 10 functions, the mean in the AFT model is a function of all, or nearly all, of the
variables with different strength of association with the survival outcome, and it also involves higher-order interactions between some of the variables. Finally, we generated the residuals in the AFT model from a gamma distribution with a shape parameter of 2 and a rate parameter of 1, resulting in a signal-to-noise ratio of 3.
We simulated 100 datasets and each dataset has a different mean function, generated from (4
). The censoring times were independently generated from an exponential distribution. A different rate parameter was chosen to obtain a censoring rate of 0.2, 0.4, or 0.6, which corresponds to light, moderate and heavy censoring, respectively. We generated 5000 observations, 75% of which were randomly chosen as training data and the remaining 25% were test data. We divided the time into six 10% intervals from approximately theto the percentile of the empirical overall survival distribution.
As shown in Figure 1, DNNSurv outperformed the Cox-nnet in the majority of the simulated datasets, as evidenced by the higher C-index and lower Brier scores in DNNSurv compared to Cox-nnet. It is worth noting that the performance of DNNSurv is even better when the censoring rate is high.
We performed a sensitivity analysis by reducing the 6 intervals into 3 intervals of 20% at the and percentiles of the empirical overall survival distribution. We found that the c-index and Brier scores remained almost the same (boxplots are shown in Supplementary Figure 1S).
First, we used the same model in section3.1 to generate survival times, but we generated covariate-dependent censoring times. Specifically, the censoring times were generated from a Cox model using the five most frequently selected variables in the functions in (4). In each simulated dataset, we controlled the censoring rate at approximately .
As shown in Figure 2, both DNNSurv and DDNSurv_ipcw performed significantly better than Cox-nnet, and DDNSurv_ipcw was slightly better than DDNSurv, as demonstrated by the larger c-index in DDNSurv_ipcw.
To further investigate the IPCW estimator, we set up a simple simulation study. In this study, survival data were generated from a Cox model with one covariate: and and was simulated from a standard normal distribution. The censoring model followed the same Cox model as the survival times, resulting in approximately censoring rate in each simulated dataset. We applied five models to each dataset: DDNSurv, DDNSurv_ipcw, Cox-nnet, pseudo-based GEE model as described in 2.1 and GEE_ipcw (i.e., the GEE using the IPCW pseudo values). In each simulation, we randomly selected subjects to train the model and predict survival probabilities for a separate set of subjects at and percentiles of the overall survival distribution.
Based on 100 simulations, we found that GEE_ipcw reduced the bias in the estimation of the regression coefficient in presence of covariate-dependent censoring. The estimate for averaged over 100 simulated datasets, was 0.783 with a mean squared error (MSE) of 0.05 in GEE, while the estimate was 1.002 with a MSE of 0.0026 in GEE_ipcw (remember the true ). Therefore, in correcting the bias in the GEE model, we further validated the IPCW method.
Figure 3 shows box plots of c-index over 100 simulations. It is interesting to note that all the studied methods have the same c-index. This could due to the fact that the rank-based c-index measure is not sensitive to the bias when the covariate effect is monotonic on survival.
The accuracy measure, Brier scores in Figure 3, shows more interesting findings: 1) GEE had the worst performance due to the bias, whereas GEE_ipcw had the best performance as it addressed the bias correctly, and the model matched the true data generating model; 2) DNNSurv_ipcw improved the prediction accuracy over the DNNSurv, and it was reasonably close to GEE_ipcw; 3) compared to GEE, DNNSurv was less sensitive to the violation of the independent censoring assumption; and 4) it is surprising to see that DNNSurv_ipcw performed better than Cox-nnet as the data were generated under the PH assumption.
We applied our DNNSurv to two prospective, multicenter cohort studies for coronary heart disease and stroke. Both studies are sponsored by the National Heart Lung and Blood Institute of the National Institutes of Health.
The Cardiovascular Health Study (CHS) was initiated in 1987 to determine the risk factors for development and progression of cardiovascular disease (CVD) in older adults, with an emphasis on subclinical measures. Detailed description of the study can be found in [DrPhilos:1993]
. The event of interest was time to CVD. The study has collected a large number of variables at baseline, including demographics (e.g., age, gender and race), family history of CVD, lab results and medication information, with the goal to identity important risk factors for the CVD event. We selected 29 predictor variables to make predictions of the CVD event; see a complete description of the variables in Table S2.
After excluding subjects with missing data in any of the selected predictor variables, we had subjects and 65.2% of which had CVD during the study period. We randomly selected 75% subjects as training data and the other 25% as test data. We then trained the data using DNNSurv and Cox-nnet. In DNNSurv, we used a model-averaging idea to obtain the risk predictions. Specifically, after the grid search, rather than using a single best set of hyperparameters to train the model, we used five best sets of hyperparameters to train the model five times, resulting in five models. Each model was then applied to the test data to predict the survival probability for each subject at each time point. The final estimated survival probability is the averaged survival probabilities from the five models. We found that by averaging five models, the prediction performance is more stable compared to using a single best model; for example, the prediction was less sensitive to initial values.
We predicted the survival probability in every year up to 15 years. In CHS data, an interval of one year is not very small, as the event rate (65.2%) is large and reasonable numbers of the event were observed in each interval. DNNSurv and Cox-nnet had the same model performance. The c-index is 0.705 in both methods and Brier scores are almost the same (see Table S3). Figure 4 shows the survival functions for some predictor variables, which were known to be important predictor variables for the CVD event. The estimated survival probabilities from DNNSurv are very similar to those from the Cox-nnet, which indicates that the proportional hazard assumption is reasonable in this dataset.
The second CVD study we considered is the Multi-Ethnic Study of Atherosclerosis (MESA) study, which enrolled subjects who were free from clinical cardiovascular disease from six communities in the United States in 20002002. Participants were followed for identification and characterization of cardiovascular disease events. Detailed description of the study can be found in [Bild:2002]. Similar to the CHS study, the event of interest was time to CVD. We have selected 30 variables to make predictions of the CVD event; see a complete description of the predictor variables in Table S4. After excluding subjects with missing data in any of the selected predictor variables, we had subjects and 5.4% of which had CVD during the study period. The event rate is very small, an interval of one year would be too small, as it is likely no event in some intervals. Thus, we predicted the survival probability only at 3 and 5 years, which are clinically meaningful landmark time points.
We randomly selected 75% subjects as training data and the other 25% as test data. Table 2 shows that DNNSurv and Cox-nnet had similar prediction performance. We also tried DNNSurv_ipcw with weights calculated based on four variables that were significantly associated with the censoring time using the Cox model. Both c-index and Brier score remained the same as the DNNSurv.
|year 3||year 5||year 3||year 5|
In this article, we develop a two-step approach for making risk predictions in survival analysis using a deep neural network model. The first step is to compute the jackknife pseudo survival probabilities in the discrete-time survival framework, and substitute the survival times by these pseudo probabilities in the deep neural networks to make risk predictions. The IPCW pseudo probabilities are used in case of the covariate-dependent censoring. By using the pseudo values, the analysis for censored survival data is reduced to a regression problem with a quantitative response variable, which greatly facilitates the use of deep learning methods. Standard deep neural networks can be directly applied, which avoids the difficulty of designing a special cost function for the censored data, as in the current methods.
We demonstrated superior performance of DNNSurv to Cox-nnet in all simulation studies. In real data analyses, these two methods had very similar performance. This could be due to the validity of the PH assumption in the real datasets. From another perspective, DNNSurv is a competing alternative to Cox-nnet when the data satisfies the PH assumption. Compared to Cox-nnet, DNNSurv is much simpler to implement, and it outputs survival or risk probabilities, which are often of direct interest to patients and physicians than the hazards in Cox-nnet.
The pseudo-observations approach offers a great opportunity to study right-censored survival data using deep neural networks. It can be generalized to analyze survival data with competing risks [Andersen:2003, Klein:2005, Klein:2008] and the restricted mean survival time (RMST) analysis [Wang:2018, Xiang:2012, Andersen:2004]. The proposed pseudo-value calculations in the discrete-time framework would allow risk predictions in the presence of competing risks. The RMST is often of great clinical interest in practice, especially in the presence of non-proportional hazard functions. However, the estimation procedure is often complicated as the RMST involves an integration of the survival function. Therefore, only a few parametric methods (such as Cox or GEE) are available for such analysis [Wang:2018, Xiang:2012, Andersen:2004]. By using the pseudo observations, the deep neural network model is directly applicable in estimating the RMST.
This Manuscript was prepared using the CHS and MESA data obtained from the National Heart, Lung, and Blood Institute (NHLBI) Biologic Specimen and Data Repository Information Coordinating Center and does not necessarily reflect the opinions or views of the NHLBI.CHS and MESA data were obtained from the National Heart, Lung, and Blood Institute (NHLBI) Biologic Specimen and Data Repository Information Coordinating Center (https://biolincc.nhlbi.nih.gov/)