# adversarial_timetoevent_modeling__4494d79a.pdf Adversarial Time-to-Event Modeling Paidamoyo Chapfuwa 1 Chenyang Tao 1 Chunyuan Li 1 Courtney Page 1 Benjamin Goldstein 1 Lawrence Carin 1 Ricardo Henao 1 Abstract Modern health data science applications leverage abundant molecular and electronic health data, providing opportunities for machine learning to build statistical models to support clinical practice. Time-to-event analysis, also called survival analysis, stands as one of the most representative examples of such statistical models. We present a deep-network-based approach that leverages adversarial learning to address a key challenge in modern time-to-event modeling: nonparametric estimation of event-time distributions. We also introduce a principled cost function to exploit information from censored events (events that occur subsequent to the observation window). Unlike most time-to-event models, we focus on the estimation of time-to-event distributions, rather than time ordering. We validate our model on both benchmark and real datasets, demonstrating that the proposed formulation yields significant performance gains relative to a parametric alternative, which we also propose. 1. Introduction Time-to-event modeling is one of the most widely used statistical analysis tools in biostatistics and, more broadly, health data science applications. For a given subject, these models estimate either a risk score or the time-to-event distribution, from a pre-specified point in time at which a set of covariates (predictors) are observed. In practice, the model is parameterized as a weighted, often linear, combination of covariates. Time is estimated parametrically or nonparametrically, the former by assuming an underlying time distribution and the latter as proportional to observed event times. These models have been widely used in risk profiling (Hippisley-Cox & Coupland, 2013; Cheng et al., 2013), treatment planning, and drug development (Fischl 1Duke University. Correspondence to: Paidamoyo Chapfuwa . Proceedings of the 35 th International Conference on Machine Learning, Stockholm, Sweden, PMLR 80, 2018. Copyright 2018 by the author(s). et al., 1987). Time-to-event modeling, and in a larger context, point processes, constitute the fundamental analytical tools in applications for which the future behavior of a system or individual is to be characterized statistically. The principal time-to-event modeling tool is the Cox Proportional Hazards (Cox-PH) model (Cox, 1992). Cox-PH is a semi-parametric model that assumes the effect of covariates is a fixed, time-independent, multiplicative factor on the hazard rate, which characterizes the instantaneous death rate of the surviving population. By optimizing a partial likelihood formulation, Cox-PH circumvents the difficulty of specifying the unknown, time-dependent, baseline hazard function. Consequently, Cox-PH results in point-estimates proportional to the event times. Further, estimation of Cox PH models depends heavily on event ordering and not the time-to-event itself, which is known to compromise the scalability of the estimation procedure to large datasets. This poor scaling behavior is manifested because the formulation is not amenable to stochastic training with minibatches. It is well accepted that the fixed-covariate-effects assumption made in Cox-PH is strong, and unlikely to hold in reality (Aalen, 1994). For instance, individual heterogeneity and other sources of variation, often likely to be dependent on time, are rarely measured or totally unobservable. This unobservable variation has been gradually recognized as a major concern in survival analysis and cannot be safely ignored (Collett, 2015; Aalen et al., 2001). When these sources of variation are independent of time, they can be modeled via fixed or random effects (Aalen, 1994; Hougaard, 1995). However, in cases for which they render the hazard rate timedependent, such variation is difficult to control, diagnose, or model parametrically. Cox-PH is known to be sensitive to such assumption violations (Aalen et al., 2001; Kleinbaum & Klein, 2010). Moreover, Cox-PH focuses on the estimation of the covariate effects rather than the survival time distribution, i.e., time-to-event prediction. The motivation behind Cox-PH and its shortcomings make it less appealing in applications where prediction is of highest importance. An alternative to the Cox-PH model is the Accelerated Failure Time (AFT) model (Wei, 1992). AFT makes the simplifying assumption that the effect of covariates either accelerates or delays the event progression, relative to a parametric Adversarial Time-to-Event Modeling baseline time-to-event distribution. However, by not making the baseline hazard a constant, as in standard Cox-PH, AFT is often a more reasonable assumption in clinical settings when predictions are important (Wei, 1992). AFT also encompasses a wide range of popular parametric proportional hazards models and proportional odds models, when the event baseline time distribution is specified properly (Klein & Moeschberger, 2005). Learning in AFT models falls into the category of maximum likelihood estimation, and therefore it scales well to large datasets, when trained via stochastic gradient descent. Further, AFT is also more robust to unobserved variation effects, relative to Cox-PH (Keiding et al., 1997). From a machine learning perspective, recent advances in deep learning are starting to transform clinical practice. Equipped with modern learning techniques and abundant data, machine-learning-driven diagnostic applications have surpassed human-expert performance in a wide array of health care applications (Cheng et al., 2013; 2016; Havaei et al., 2017; Djuric et al., 2017; Gulshan et al., 2016). However, applications involving time-to-event modeling have been largely under-explored. From the existing approaches, most focus on extending Cox-PH with nonlinear neuralnetwork-based covariate mappings (Katzman et al., 2016; Faraggi & Simon, 1995; Zhu et al., 2016), casting the timeto-event modeling as a discretized-time classification problem (Yu et al., 2011; Fotso, 2018), or introducing a nonlinear map between covariates and time via Gaussian processes (Fern andez et al., 2016; Alaa & van der Schaar, 2017). Interestingly, all of these approaches focus their applications toward relative risk, fixed-time risk (e.g., 1-year mortality) or competing events, rather than event-time estimation, which is key to individualized risk assessment. Generative Adversarial Networks (GANs) (Goodfellow et al., 2014) have recently demonstrated unprecedented potential for generative modeling, in settings where the goal is to estimate complex data distributions via implicit sampling. This is done by specifying a flexible generator function, usually a deep neural network, whose samples are adversarially optimized to match in distribution to those from real data. Succesful examples of GAN include generation of images (Radford et al., 2015; Salimans et al., 2016), text (Yu et al., 2017; Zhang et al., 2017) and data conditioned on covariates (Isola et al., 2016; Reed et al., 2016). However, ideas from adversarial learning are yet to be exploited for the challenging task that is time-to-event modeling. Previous work often represents time-to-event distributions using a limited family of parametric forms, i.e., log-normal, Weibull, Gamma, Exponential, etc. It is well understood that parametric assumptions are often violated in practice, largely because of the model is unable to capture unobserved (nuisance) variation. This fundamental shortcoming is one of the main reasons why non-parametric methods, e.g., Cox proportional hazards, are so popular. Adversarial learning leverages a representation that implicitly specifies a timeto-event distribution via sampling, rather than learning the parameters of a pre-specified distribution. Further, GANlearning penalizes unrealistic samples, which is a known issue in likelihood-based models (Karras et al., 2018). The work presented here seeks to improve the quality of the predictions in nonparametric time-to-event models. We propose a deep-network-based nonparametric time-to-event model called a Deep Adversarial Time-to-Event (DATE) model. Unlike existing approaches, DATE focuses on the estimation of time-to-event distributions, rather than event ordering, thus emphasizing predictive ability. Further, this is done while accounting for missing values, high-dimensional data and censored events. The key contributions associated with the DATE model are: (i) The first application of GANs to nonlinear and nonparametric time-to-event modeling, conditioned on covariates. (ii) A principled censored-eventaware cost function that is distribution-free and independent of time ordering. (iii) Improved uncertainty estimation via deep neural networks with stochastic layers. (iv) An alternative, parametric, non-adversarial time-to-event AFT model to be used as baseline in our experiments. (v) Results on benchmark and real data demonstrate that DATE outperforms its parametric counterpart by a substantial margin. 2. Background Time-to-event datasets are usually composed of three variables {xi, ti, li}N i=1 = D, the covariates xi = [xi1, ..., xip] Rp and event pairs (ti, li), where ti is the time-to-event of interest and li {0, 1} is a binary censoring indicator. Typically, li = 1 indicates the event is observed, alternatively, li = 0 indicates censoring at ti (when li = 0, an event, if it occurs, transpires after the observation window that ends at time ti). N denotes the size of the dataset. Time-to-event models characterize the survival function: S(t) = P(T > t) = 1 F(t) = exp R t 0h(s)ds , defined as the fraction of the population that survives up to time t, or the conditional hazards rate function h(t|x) (see below), defined as the instantaneous rate of occurrence of an event at time t given covariates x. We derive the relationship between h(t|x) and S(t) from standard definitions (Kleinbaum & Klein, 2010): h(t|x) = lim dt 0 P(t < T < t + dt|x) P(T > t|x)dt = f(t|x) S(t|x) , (1) where f(t|x) is the conditional survival density function and S(t|x) is formally the complement of the cumulative conditional density function F(t|x). To solve the hazard-rate Adversarial Time-to-Event Modeling differential equation (1), we establish the following relationship between f(t|x), h(t|x) and S(t|x) via the cumulative conditional hazard H(t|x) = R t 0 h(s|x)ds as S(t|x) = exp ( H(t|x)) , (2) f(t|x) = h(t|x)S(t|x) . (3) See Supplementary Material for examples of parametric time-to-event characterizations. Time-to-event models, e.g., Cox-PH and AFT, leverage the results in (2) and (3), to characterize the relationship between covariates x and timeto-event t, when estimating the conditional hazard function h(t|x). Two popular frameworks, Cox-PH and AFT, approach the estimation of h(t|x) using nonparametric and parametric techniques, respectively. Cox Proportional Hazard The Cox-PH (Cox, 1992) model is a semi-parametric, linear model where the conditional hazard function h(t|x) depends on time through the baseline hazard h0(t), and independent of covariates x as h(t|x) = h0(t) exp(x β) . (4) Provided with the N observation tuples in D, Cox-PH estimates the regression coefficients, β Rp, that maximize the partial likelihood (Cox, 1992): exp(x i β) P j:tj ti exp(x j β) , (5) where L(β) is independent of the baseline hazard in (4). Note also that (5) only depends on the ordering of ti, for i, . . . , N, and not their actual values. Cox-PH is nonparametric in that it estimates the ordering of the events, not their times, thus avoiding the need to specify a distribution for h0(t). Several techniques have been developed that assume a parametric distribution for h0(t), in order to estimate the actual time-to-event, however, not nearly as widely adopted as standard Cox-PH. See Bender et al. (2005) for specifications of h0(t) that result in an exponential, Weibull or Gompertz survival density functions. For example, when f(t) = λ exp( λt), i.e., exponential survival density, thus h0(t) = λ. Accelerated Failure Time The AFT model (Wei, 1992) is a popular alternative to the widely used Cox-PH model. In this model, similar to Cox-PH, it is assumed that h(t|x)) = ψ(x)h0(ψ(x)t), where ψ(x) is the total effect of covariates, x, usually through a linear relationship ψ(x) = exp( x β), where β represents the regression coefficients. If the conditional survival density function satisfies f(t|x) = ψ(x)f0(t), i.e., S(t) independent of x like in Cox-PH, then we can write log t = log(t0) log ψ(x) = ξ x β , (6) where t0 p0(t) is the unmoderated time, thus ξ characterizes the baseline survival density distribution. Note the similarity between (4) and (6) despite the differences in their motivation. Different choices of baseline distribution yield a variety of AFT distributions, including Weibull, log-normal, gamma and inverse Gaussian (Klein & Moeschberger, 2005). Intuitively, AFT assumes the effect of the covariates, ψ(x), accelerates or delays the life course, which is often meaningful in a clinical or pharmaceutical setting, and sometimes easier to interpret compared with Cox-PH (Wei, 1992). Empirical evidence has shown that AFT is more robust to missing values and misspecification of the survival function than Cox-PH (Keiding et al., 1997). Gaussian-process-based AFT models (Fine & Gray, 1999; Alaa & van der Schaar, 2017) have been used to model competing risk applications with success. 3. Adversarial Time-to-Event We develop a nonparametric model for p(t|x), where t is the (non-censored) time-to-event from the time at which covariates x were observed. More precisely, we learn the ability to sample from p(t|x) via approximation q(t|x). Further, we do so without specifying a distribution for the marginal (baseline survival distribution), p0(t), which in AFT is usually assumed log-normal. Like in Cox-PH and AFT, we assume that p0(t) is independent of covariates x. For censored events, li = 0, we wish the model to have a high likelihood for p(t > ti|xi), while for non-censored events, li = 1, we wish that the pairs {xi, ti} be consistent with data generated from p(t|x)p0(x), where p0(x) is the (empirical) marginal distribution for covariates, from which we can sample but whose explicit form is unknown. We consider a conditional generative adversarial network (GAN) (Goodfellow et al., 2014), in which we draw approximate samples from p(t|x), for li = 1, as t = Gθ(x, ϵ; l = 1), ϵ pϵ(ϵ) , (7) where pϵ(ϵ) is a simple distribution, e.g., isotropic Gaussian or uniform (discussed below). The generator, Gθ(x, ϵ; l = 1) is a deterministic function of x and ϵ, specified as a deep neural network with model parameters θ, that implicitly defines qθ(t|x, l = 1) in a nonparametric manner. We explicitly note that l = 1, to emphasize that all t drawn from this model are event times (non-censored times). Ideally the pairs {x, t} manifested from the model in (7) are indistinguishable from the observed data {x, t, l = 1} D, i.e., the non-censored samples. Let Dnc D and Dc D be the disjoint subsets of noncensored and censored data, respectively. Given a discriminator function Dφ(x, t) specified as a deep neural network with model parameters φ. The cost function based on the Adversarial Time-to-Event Modeling non-censored data has the following form: ℓ1(θ, φ; Dnc) = E(t,x) pnc[D(x, t)] (8) + Ex pnc,ϵ pϵ[1 D(x, Gθ(x, ϵ; l = 1))] , where pnc(t, x) is the empirical joint distribution responsible for Dnc, and the expectation terms are estimated through samples {t, x} pnc(t, x) and ϵ pϵ(ϵ) only. We seek to maximize ℓ1(θ, φ; Dnc) wrt discriminator parameters φ, while seeking to minimize it wrt generator parameters θ. For non-censored data, the formulation in (8) is the standard conditional GAN. We also leverage the censored data Dc to inform the parameters θ of generative model Gθ(x, ϵ; l = 0). We therefore consider the additional cost function ℓ2(θ; Dc) = E(t,x) pc,ϵ pϵ[max(0, t Gθ(x, ϵ; l = 0)] , (9) where max(0, ) encodes that Gθ(x, ϵ; l = 0) incurs no cost as long as the sampled time is larger than the censoring point. Further, pc(t, x) is the empirical joint distribution responsible for Dc, from which samples {t, x} are drawn to approximate the expectation in (9). Note that max(0, ) is one of many choices; smoothed or margin-based alternatives may be considered, but are not addressed here, for simplicity. For cases in which the proportion of observed events is low, the costs in (8) and (9) underrepresent the desire that timeto-events must be as close as possible to the ground truth, t. For this purpose, we also impose a distortion loss d( , ) ℓ3(θ; Dnc) = E(t,x) pnc[d(t, Gθ(x, ϵ; l = 1))] , (10) that penalizes Gθ(x, ϵ; l = 1) for not being close to the event time t for non-censored events only. In the experiments, we set d(a, b) = a b 1. The complete cost function is ℓ(θ, φ; D) = ℓ1(θ, φ; Dnc) + λ2ℓ2(θ; Dc) + λ3ℓ3(θ; Dnc) , (11) where {λ2, λ3} > 0 are tuning parameters controlling the trade-off between non-censored and censored cost functions relative to the discriminator objective in (8). In our experiments we set λ2 = λ3 = 1, provided that (9) and (10) are written in terms of expectations, thus already account for the proportion differences in Dc and Dnc. However, this may not be sufficient in heavily imbalanced cases or when the time domains for Dc and Dnc are very different. The cost function in (11) is optimized using stochastic gradient descent on minibatches from D. We maximize ℓ(θ, φ; D) wrt φ and minimize it wrt θ. The terms Figure 1. Effects of stochastic layers on uncertainty estimation on 10 randomly selected test-set subjects from the SUPPORT dataset. Ground truth times are denoted as t and box plots represent time-to-event distributions from a 2-layer model, where α = [α0, α1, α2] indicates whether the corresponding noise source, {ϵ0, ϵ1, ϵ2}, is active. For example α = [1, 0, 0] indicates noise on the input layer only. ℓ1(θ, φ; Dnc) and ℓ3(θ; Dnc) reward Gθ(x, ϵ; l = 1) if it synthesizes data that are consistent Dnc, and the term ℓ2(θ; Dc) encourages Gθ(x, ϵ; l = 1) to generate event times that are consistent with the data Dc, i.e., larger than censoring times. Time-to-event uncertainty The generator in (7) has a single source of stochasticity, ϵ, which in GAN-based models has been traditionally applied as input to the model, independent of covariates x. In a Multi-Layer Perceptron (MLP) architecture, h1 = g(W 10x + W 11ϵ), where h1 denotes the vector of layer-1 hidden units, g( ) is the activation function (RELU in the experiments), W 10 and W 11 are weight matrices for covariates and noise, respectively, and we have omitted the bias term for clarity. In a model with multiple layers, the noise term applied to the input tends to have a small effect on the distribution of sampled event times (see experiments). More specifically, samples from qθ(t|x) tend to have small variance. This results in a model with underestimated uncertainty, hence overconfident predictions. This is due to many factors, including compounding effects of activation nonlinearities, layer-wise regularizers (e.g., dropout), and cancelling terms when the support of the noise distribution is the real line (both positive and negative). Although the cost function in (11) rewards the generator for producing (non-censored) event times close to the ground-truth, thus in principle encouraging event time distributions to cover it, this rarely happens in practice. This issue is well-known in the GAN literature (Salimans et al., 2016). Adversarial Time-to-Event Modeling Here we take a simple approach consisting on adding sources of stochasticity to every layer of the generator as hj = g(W j0hj 1+W j1ϵj) , where j = 1, . . . , L and L is the number of layers. By doing this, we encourage increased coverage on the event times produced by the generator, without substantially changing the model or the learning procedure. In the experiments, we use a multivariate uniform distribution, ϵj Uniform(0, 1), for j = 1, . . . , L, over Gaussian to reduce cancelling effects. As we show empirically, this approach produces substantially better coverage compared to having noise only on the input layer and without the convergence issues associated with the additional stochasticity. In Figure 1 we illustrate the contribution of the noise on each layer to the distribution of event times. In this example, we show 10 test-set estimated time-to-event distributions using a 2-layer model with noise sources in all layers, including the input. We see that ground-truth times are nicely covered by the estimated distributions. Also, that the combination of noise sources, rather than any individual source, jointly contribute to the desired distribution coverage. Additional examples, including censored times, are found in the Supplementary Material. 4. Baseline AFT Model In the experiments, we will demonstrate the capabilities of the DATE model. However, since there are not many scalable nonlinear time-to-event models focused on event time estimation, below we present a parametric (non-adversarial) model to be used in the experiments as baseline. We do so with the goal of providing a fair comparison and, at the same time, to highlight the distinguishing factors of the DATE model. Starting from (6), we consider the following MLP-based log-normal AFT: log t = µβ(x) + ξ , ξ N(0, σ2 β(x)) , (12) where µβ(x) and σ2 β(x) are MLPs parameterized by β, representing the mean and variance of the log-transformed time-to-event as a function of covariates x. For convenience, we adopt the log-normal distribution for event time t, mainly because we found that it is considerably more stable during optimization, compared to other popular survival distributions, e.g., Weibull or Gamma. Note that (12) is very different from (7), where the generator implicitly defines the distribution qθ(t|x). The likelihood function of the log-normal AFT model in (12) for all events (censored and non-censored) is then i p(ti|xi) = Y i:li=1 fβ(ti|xi) Y i:li=0 Sβ(ti|xi) (13) i:li=1 φ(ν(ti, xi)) Y i:li=0 (1 Φ(ν(ti, xi))) , where φ( ) and Φ( ) are the Gaussian density and cumulative density functions, respectively, and ν(ti, xi) = (log ti µβ(xi))/σβ(xi). The likelihood in (13) is convenient, because it allows estimation of time-to-event, while seemly accounting for censored events. The latter comes as benefit of having a parametric model with closed-form cumulative density function. The cost function L(β; D) for the Deep Regularized AFT (DRAFT) model is written as: L(β; D) = log p(t|D) + ηR(β; D) , (14) where the first term is the negative log-likelihood loss from (13), η > 0 is a tuning parameters, and R(β; D) is a regularization loss that encourages event times to be properly ordered. Specifically, we use the following cost function adapted from Steck et al. (2008): R(β; D) = 1 j:tj>ti 1 + log σ(µβ(xj) µβ(xi)) where E is the set of all pairs {i, j} in {1, . . . , N} for which the second argument is observed, i.e., li = 1, and σ( ) is the sigmoid function. The cost function above is a lower bound on the Concordance Index (CI) (Harrell et al., 1984), which constitutes a difficult-to-optimize discrete objective, that is widely used as performance metric for survival analysis, precisely because it captures time-to-event order. Further, it is reminiscent of the partial-likelihood of Cox-PH in (5), but is more amenable to stochastic training. 5. Related Work Deep learning models, specifically MLPs, have been successfully integrated with Cox-PH-based objectives to improve risk estimation in time-to-event models. Faraggi & Simon (1995) proposed an neural-network-based model optimized using the standard partial-likelihood cost function from Cox-PH. Katzman et al. (2016) is similar to Faraggi & Simon (1995), but leverages modern deep learning techniques such as weight decay, batch normalization and dropout. Luck et al. (2017) replaced the partiallikelihood formulation in (5) with Efron s approximation (Efron, 1977) and an isotonic regression cost function adapted from Menon et al. (2012) to handle censored events. Zhu et al. (2016) proposed a time-to-event model for image covariates based on convolutional networks. From the Gaussian process literature, Fern andez et al. (2016) proposed a time-to-event model inspired by a Poisson process, where the nonlinear map between covariates and time is modeled as a Gaussian process on the Poisson rate. More recently, Alaa & van der Schaar (2017) proposed a deep multi-task Gaussian process model for survival analysis with competing risks, and learned via variational inference. Adversarial Time-to-Event Modeling Following a different path, other approaches recast the timeto-event problem as a classification task. Yu et al. (2011) proposed a linear model where (discretized) time is estimated using a sequence of dependent regressors. More recently, Fotso (2018) extended their approach to a nonlinear mapping of covariates using deep neural networks. Generative approaches have also been proposed to infer survival-time distributions with variational inference. Deep Survival Analysis (DSA) (Ranganath et al., 2016) specifies a latent model that leverages deep exponential family distribution, however their approach does not handle censored events. All the above methods focus on relative risk quantified as the CI on the ordering of event times, or fixed-time risk, e.g., 1-year mortality. However, relative risk is most useful when associated with covariate effects, which is difficult in nonlinear models based either on neural networks or Gaussian processes. Fixed-time risk, although very useful in practice, can be recast as a classification problem rather than a substantially more complex time-to-event model. Importantly, none of these approaches consider the task of time-to-event estimation, despite the fact that Gaussian process and generative approaches can be repurposed for such task. 6. Experiments The loss functions in (11) and (14) for DATE and DRAFT, respectively, are minimized via stochastic gradient descent. At test time, we draw 200 samples from (7) and (12) for DATE and DRAFT, respectively, and use medians for quantitative results requiring point estimates, i.e., ˆt = median({ts}200 s=1), where ts is a sample from the trained model. Detailed network architectures, optimization parameters and initialization settings are in the Supplementary Material. Tensor Flow code to replicate experiments can be found at https://github.com/ paidamoyo/adversarial_time_to_event. Comparison Methods For non-deep learning based models, we considered arguably the two most popular approaches to time-to-event modeling, namely, (regularized) Cox-Efron and RSF. For deep learning models, we considered DRAFT, which generalizes existing neural-networkbased methods by using both a parametric log-normal AFT objective and a non-parametric ordering cost function. Extending DRAFT to a mixture of log-normal distributions with different variances but shared mean, did not result in improved performance. We did not consider (variational) models for non-censored events only because learning from censored events is one of the main defining characteristics of time-to-event modeling (otherwise, the model essentially becomes non-negative regression). Further, in most practical situations the proportion of non-censored events is low, e.g., 24% in the EHR. Table 1. Summary of datasets used in experiments. EHR FLCHAIN SUPPORT SEER Events (%) 23.9 27.5 68.1 51.0 N 394,823 7,894 9,105 68,082 p (cat) 729 (106) 26 (21) 59 (31) 789 (771) Na N (%) 1.9 2.1 12.6 23.4 tmax 365 days 5,215 days 2,029 days 120 months Datasets Our model is evaluated on 4 diverse datasets: i) FLCHAIN: a public dataset introduced in a study to determine whether non-clonal serum immunoglobin free light chains are predictive of survival time (Dispenzieri et al., 2012). ii) SUPPORT: a public dataset introduced in a survival time study of seriously-ill hospitalized adults (Knaus et al., 1995). iii) SEER: a public dataset provided by the Surveillance, Epidemiology, and End Results Program. See (Ries et al., 2007) for details concerning the definition of the 10-year follow-up breast cancer subcohort used in our experiments. iv) EHR: a large study from Duke University Health System centered around inpatient visits due to comorbidities in patients with Type-2 diabetes. The datasets are summarized in Table 1, where p denotes the number of covariates to be analyzed, after one-hot-encoding for categorical (cat) variables. Events indicates the proportion of the observed events, i.e., those for which li = 1. NAN indicates the proportion of missing entries in the N p covariate matrix and tmax is the time range for both censored and non-censored events. Details about the public datasets: FLCHAIN, SUPPORT and SEER, including preprocessing procedures, can be found in the references provided above. EHR is a study designed to track primary care encounters of 19, 064 Type-2 diabetes patients over a period of 10 years (2007-2017). The purpose of the analysis is to predict diabetes-related causes of hospitalization within 1 year of an EHR-recorded primary care encounter. Data is processed and analyzed at the patient encounter-level. The total number of encounters is N = 394, 823. To avoid bias due to multiple encounters per patient, we split the training, validation and test sets so that a given patient can only be in one of the sets. The covariates, collected over a period of a year before the primary care encounter of interest, consist of a mixture of continuous and categorical summaries extracted from electronic health records: vitals and labs (minimum, maximum, count and mean values); comorbidities, medications and procedures ICD-9/10 codes (binary indicators and counts); and demographics (age, gender, race, language, smoking indicator, type of insurance coverage, as either continuous or categorical variables). 6.1. Qualitative results First, we visually compare the test-set time-to-event distributions by DATE and DRAFT on FLCHAIN data. In Figure 2 we show the top best (left) and worst (middle-left) Adversarial Time-to-Event Modeling Figure 2. Example test-set predictions on FLCHAIN data. Top best (left) and worst (middle-left) predictions on censored events, and top best (middle-right) and worst (right) predictions on non-censored events. Circles denote ground-truth events or censoring points, while box-plots represent distributions over 200 samples for both DATE and DRAFT. The horizontal dashed line represents the range (tmax = 5, 215 days) of the events. predictions on censored events, and the top best (middleright) and worst (right) predictions on non-censored events. Circles denote ground-truth events or censoring points, while box-plots represent distributions over 200 samples for both DATE and DRAFT models. We see that: (i) in nearly every case, DATE is more accurate than DRAFT. (ii) DRAFT tends to make predictions outside the event range (tmax = 5, 215 days), denoted as a horizontal dashed line. (iii) DRAFT tends to overestimate the variance of its predictions, approximately by one order of magnitude relative to DATE. This is not very surprising as DRAFT has an MLP dedicated to estimate, conditioned on the covariates, the variance of the time-to-event distribution. However, note that variances estimated well over the domain of the events (tmax) are not necessarily meaningful or desirable. Figures with similar findings for the other 3 datasets can be found in the Supplementary Material. To provide additional insight into the performance of DATE compared to DRAFT, we report the Normalized Relative Error (NRE) defined as (ˆt t)/tmax and min(0, ˆt t)/tmax for non-censored and censored events, respectively, where t, ˆt and tmax denote the ground-truth time-to-event, median time estimated (from samples) and event range, as indicated in Table 1. The NRE distribution provides a visual representation of the extent of test-set errors, while revealing whether the models are biased toward either overestimating (ˆt > t ) or underestimating (t > ˆt) the event times. Although models with unbiased NREs are naturally preferred, in most clinical applications where being conservative is important, overestimated time-to-events must be avoided as much as possible. Figure 3 shows NRE distributions for test-set non-censored events on SUPPORT and EHR data. We see that DRAFT results in a considerable amount of errors beyond the event range (|NRE| > 1), tmax = 120 months or tmax = 365 days for SUPPORT and EHR, respectively. Further, we see that the NRE distribution for DRAFT is heavily skewed toward NRE > 1, thus tending to overestimate event times. On the other hand, DATE produces errors substantially more concentrated around 0 Figure 3. Normalized Relative Error (NRE) distribution for SUP- PORT (top) and EHR (bottom), test-set non-censored events. The horizontal dashed lines represent the range of the events, tmax = 120 months and tmax = 365 days, respectively. and within |NRE| < 1, relative to DRAFT. This demonstrates the advantage of the adversarial method DATE over the likelihood-based method DRAFT in generating realistic samples. Similar results were observed on the other datasets for both censored and non-censored events. See Supplementary Material for additional figures. 6.2. Quantitative results Relative absolute error The perfromance of DATE is evaluated in terms of absolute error relative to the event range, i.e., |ˆt t|/tmax. For censored events, the relative error is defined as max(0, t ˆt)/tmax, to account for the fact that no error is made as long as t ˆt. Table 2 shows median and 50% empirical intervals for relative absolute errors on non-censored events, on all test-data. Results on censored data are small and comparable across approaches, and are thus presented in the Supplementary Material. Specifically, we see that DATE outperforms DRAFT in 3 our of 4 cases by a substantial margin, and is comparable on the SUPPORT data. For instance, on EHR data 75% of all DATE testset predictions have a relative absolute error less than 43% (approx. 156 days) which is substantially better than the 81% (approx. 295 days) by DRAFT. Missing data Since missing data are common in clinical data, e.g., SEER data contains 23.4% missing values, we also consider a modified version of DATE, where the generator in (7) takes the form t = Gθ(z, ϵ, l = 1), where z is modeled as an adversarial autoencoder (Dumoulin et al., 2016; Li et al., 2017; Pu et al., 2017) with an encoder/decoder Adversarial Time-to-Event Modeling Table 2. Median relative absolute errors (as percentages of tmax), on non-censored data. Ranges in parentheses are 50% empirical ranges over (median) test-set predictions. DATE DATE-AE DRAFT EHR 23.6(11.1,43.0) 24.5(12.4,44.0) 36.7(16.1,81.3) FLCHAIN 19.5(9.5,31.1) 19.3(8.9,32.4) 26.2(9.0,53.5) SUPPORT 2.7(0.4,16.1) 1.5(0.4,19.2) 2.0(0.2,35.3) SEER 18.6(8.3,34.1) 20.2(10.3,35.8) 23.7(9.9,51.2) pair specified similar to DATE. See Supplementary Material for additional details. This model, denoted in Table 2 as DATE-AE, does not require missing covariates to be imputed before hand, which is the case of DATE and DRAFT as specified originally. The results show no substantial performance improvement by DATE-AE, relative to DATE; results indicate that all of these approaches (DRAFT, DATE and DATE-AE) are robust to missing data. As a benchmark, we took FLCHAIN and SUPPORT, to then artificially introduced missing values ranging in proportion from 20% to 50%. Results in Supplementary Material support the idea of robustness, since all three approaches resulted in median relative absolute errors within 1% of those in Table 2. We also tried to quantify statistically the match between time-to-event samples generated from DATE and those from the empirical distribution of the data, using the distributionfree two-sample test based on Maximum Mean Discrepancy (MMD) proposed by Sutherland et al. (2017). Due to sample size limitations (number of non-censored events in the test-set) and high-variances on the p-value estimates, we could not reliably reject the hypothesis that real and DATE samples are drawn from the same distribution. We did confirm it for DRAFT, which is not surprising considering both qualitative and quantitative results discussed above. Concordance Index The concordance Index (CI) (Harrell et al., 1984), which quantifies the degree to which the order of the predicted times is consistent with the ground truth, is the most well-known performance metric in survival analysis. Although not the focus of our approach, we compared DATE to DATE-AE, DRAFT, Random Survival Forests (Ishwaran et al., 2008), and Cox-PH (with Efron s approximation (Efron, 1977)). We found all of these models to be largely comparable. The results, presented in the Supplementary Material, show DATE(-AE) and DRAFT being the best-performing models on EHR and SUPPORT, respectively. On SEER, DATE(-AE) and DRAFT outperform Cox-PH and RSF. Finally, on FLCHAIN, the smallest dataset, all methods perform about the same. Note that unlike Cox-PH and DRAFT in (5) and (14), DATE(-AE) does not explicitly encourage proper time-ordering on the objective function; it is consequently deemed a strength of the proposed GAN-based DATE model that it properly learns ordering, without needing to explicitly impose this condition when training. DATE does not have a clear advantage Table 3. Median of 95% intervals for all test-set time-to-event distributions on SUPPORT data. Ranges in parentheses are 50% empirical quantiles. Uniform(-1,1) Uniform(0,1) Gaussian(0,1) Non-censored All 60.0(3.9,176.5) 149.9(8.5,926.8) 37.9(3.5,237.4) Input 28.9(1.8,114.8) 22.4(1.5,91.2) 33.7(1.6,127.6) Output - 168.8(16.6,844.3) - Censored All 231.3(177.2,332.1) 1397.3(990.9,2000.1) 350.5(254.4,539.3) Input 137.3(99.4,205.0) 86.9(64.4,135.0) 155.8(106.7,229.3) Output - 1158.6(873.8,1670.4) - in terms of learning the correct order, however, we verified empirically that adding an ordering cost function, R(β; D) in (14), to DATE does not improve the results. Distribution coverage We now demonstrate that the DATE model, with noise sources on all layers, has timeto-event distributions with larger variances than versions of DATE with noise only on the input of the neural network. Table 3 shows the median of the 95% intervals for all test-set time-to-event distributions on SUPPORT data. DATE with Uniform(0,1) has larger variance and coverage compared to the other alternatives, while keeping relative absolute errors and CIs largely unchanged (see Supplementary Material for details). We did not run models with Uniform(-1,1) and Gaussian(0,1) only on the output layer, because from the other results presented above it is clear that these two options are not nearly as good as having Uniform(0,1) noise on all layers. Note also that we did not include DRAFT in these comparisons. DRAFT has naturally good coverage due to the variance of the time-to-event distributions being modeled independent for each observation as a function of the covariates (see for instance Figure 2). However, DRAFT has difficulties keeping good coverage while maintaining good performance, i.e,, small absolute relative error. 7. Conclusions We have presented an adversarially-learned time-to-event model that leverages a distribution-free cost function for censored events. The proposed approach extends GAN models to time-to-event modeling with censored data, and it is based on deep neural networks with stochastic layers. The model yields improved uncertainty estimation relative to alternative approaches. As a baseline model for our experiments, we also proposed a parametric AFT-based with a parametric log-normal distribution on the time of event. To the best of our knowledge, this work is the first to leverage adversarial learning to improve estimation of time-to-event distributions, conditioned on covariates. Experimental results on challenging time-to-event datasets showed that DATE, our adversarial solution, consistently outperforms DRAFT, its parametric (log-normal) counterpart. As future work, we will extend DATE to models with competing risks and longitudinally measured covariates. Adversarial Time-to-Event Modeling Acknowledgements The authors would like to thank the anonymous reviewers for their insightful comments. This research was supported in part by DARPA, DOE, NIH, ONR, NSF and SANOFI. The authors would also like to thank Shuyang Dai, Liqun Chen, Rachel Ballantyne Draelos, Xilin Cecilia Shi, Yan Zhao and Dr. Neha Pagidipati, for fruitful discussions. Aalen, O. O. Effects of frailty in survival analysis. Statistical Methods in Medical Research, 1994. Aalen, O. O., Gjessing, H. K., et al. Understanding the shape of the hazard rate: A process point of view (with comments and a rejoinder by the authors). Statistical Science, 2001. Alaa, A. M. and van der Schaar, M. Deep Multi-task Gaussian Processes for Survival Analysis with Competing Risks. In NIPS, 2017. Bender, R., Augustin, T., and Blettner, M. Generating survival times to simulate Cox proportional hazards models. Statistics in medicine, 2005. Cheng, J.-Z., Ni, D., Chou, Y.-H., Qin, J., Tiu, C.-M., Chang, Y.-C., Huang, C.-S., Shen, D., and Chen, C.-M. Computer-aided diagnosis with deep learning architecture: applications to breast lesions in US images and pulmonary nodules in CT scans. Scientific reports, 2016. Cheng, W.-Y., Yang, T.-H. O., and Anastassiou, D. Development of a prognostic model for breast cancer survival in an open challenge environment. Science translational medicine, 2013. Collett, D. Modelling survival data in medical research. CRC press, 2015. Cox, D. R. Regression models and life-tables. In Breakthroughs in statistics. 1992. Dispenzieri, A., Katzmann, J. A., Kyle, R. A., Larson, D. R., Therneau, T. M., Colby, C. L., Clark, R. J., Mead, G. P., Kumar, S., Melton, L. J., et al. Use of nonclonal serum immunoglobulin free light chains to predict overall survival in the general population. In Mayo Clinic Proceedings, 2012. Djuric, U., Zadeh, G., Aldape, K., and Diamandis, P. Precision histology: how deep learning is poised to revitalize histomorphology for personalized cancer care. npj Precision Oncology, 2017. Dumoulin, V., Belghazi, I., Poole, B., Lamb, A., Arjovsky, M., Mastropietro, O., and Courville, A. Adversarially learned inference. ar Xiv, 2016. Efron, B. The efficiency of Cox s likelihood function for censored data. JASA, 1977. Faraggi, D. and Simon, R. A neural network model for survival data. Statistics in medicine, 1995. Fern andez, T., Rivera, N., and Teh, Y. W. Gaussian processes for survival analysis. In NIPS, 2016. Fine, J. P. and Gray, R. J. A proportional hazards model for the subdistribution of a competing risk. JASA, 1999. Fischl, M. A., Richman, D. D., Grieco, M. H., Gottlieb, M. S., Volberding, P. A., Laskin, O. L., Leedom, J. M., Groopman, J. E., Mildvan, D., Schooley, R. T., et al. The efficacy of azidothymidine (AZT) in the treatment of patients with AIDS and AIDS-related complex. New England Journal of Medicine, 1987. Fotso, S. Deep neural networks for survival analysis based on a multi-task framework. ar Xiv, 2018. Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. Generative adversarial nets. In NIPS, 2014. Gulshan, V., Peng, L., Coram, M., Stumpe, M. C., Wu, D., Narayanaswamy, A., Venugopalan, S., Widner, K., Madams, T., Cuadros, J., et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA, 2016. Harrell, F. E., Lee, K. L., Califf, R. M., Pryor, D. B., and Rosati, R. A. Regression modelling strategies for improved prognostic prediction. Statistics in medicine, 1984. Havaei, M., Davy, A., Warde-Farley, D., Biard, A., Courville, A., Bengio, Y., Pal, C., Jodoin, P.-M., and Larochelle, H. Brain tumor segmentation with deep neural networks. Medical image analysis, 2017. Hippisley-Cox, J. and Coupland, C. Predicting risk of emergency admission to hospital using primary care data: derivation and validation of QAdmissions score. BMJ open, 2013. Hougaard, P. Frailty models for survival data. Lifetime data analysis, 1995. Ishwaran, H., Kogalur, U. B., Blackstone, E. H., and Lauer, M. S. Random survival forests. The annals of applied statistics, 2008. Isola, P., Zhu, J.-Y., Zhou, T., and Efros, A. A. Image-toimage translation with conditional adversarial networks. ar Xiv, 2016. Adversarial Time-to-Event Modeling Karras, T., Aila, T., Laine, S., and Lehtinen, J. Progressive growing of GANs for improved quality, stability, and variation. In ICLR, 2018. Katzman, J., Shaham, U., Bates, J., Cloninger, A., Jiang, T., and Kluger, Y. Deep survival: A deep cox proportional hazards network. ar Xiv, 2016. Keiding, N., Andersen, P. K., and Klein, J. P. The role of frailty models and accelerated failure time models in describing heterogeneity due to omitted covariates. Statistics in medicine, 1997. Klein, J. P. and Moeschberger, M. L. Survival analysis: techniques for censored and truncated data. Springer Science & Business Media, 2005. Kleinbaum, D. G. and Klein, M. Survival analysis. Springer, 2010. Knaus, W. A., Harrell, F. E., Lynn, J., Goldman, L., Phillips, R. S., Connors, A. F., Dawson, N. V., Fulkerson, W. J., Califf, R. M., Desbiens, N., et al. The SUPPORT prognostic model: objective estimates of survival for seriously ill hospitalized adults. Annals of internal medicine, 1995. Li, C., Liu, H., Chen, C., Pu, Y., Chen, L., Henao, R., and Carin, L. Alice: Towards understanding adversarial learning for joint distribution matching. In NIPS, 2017. Luck, M., Sylvain, T., Cardinal, H., Lodi, A., and Bengio, Y. Deep Learning for Patient-Specific Kidney Graft Survival Analysis. ar Xiv, 2017. Menon, A. K., Jiang, X. J., Vembu, S., Elkan, C., and Ohno Machado, L. Predicting accurate probabilities with a ranking loss. In ICML, 2012. Pu, Y., Chen, L., Dai, S., Wang, W., Li, C., and Carin, L. Symmetric Variational Autoencoder and Connections to Adversarial Learning. ar Xiv, 2017. Radford, A., Metz, L., and Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. ar Xiv, 2015. Ranganath, R., Perotte, A., Elhadad, N., and Blei, D. Deep survival analysis. In Machine Learning for Healthcare Conference, 2016. Reed, S., Akata, Z., Yan, X., Logeswaran, L., Schiele, B., and Lee, H. Generative adversarial text to image synthesis. ar Xiv, 2016. Ries, L. A. G., Young Jr, J. L., Keel, G. E., Eisner, M. P., Lin, Y. D., and Horner, M.-J. D. Cancer survival among adults: US SEER program, 1988 2001. Patient and tumor characteristics SEER Survival Monograph Publication, 2007. Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., and Chen, X. Improved techniques for training gans. In NIPS, 2016. Steck, H., Krishnapuram, B., Dehing-oberije, C., Lambin, P., and Raykar, V. C. On ranking in survival analysis: Bounds on the concordance index. In NIPS, 2008. Sutherland, D. J., Tung, H.-Y., Strathmann, H., De, S., Ramdas, A., Smola, A., and Gretton, A. Generative models and model criticism via optimized maximum mean discrepancy. ICLR, 2017. Wei, L.-J. The accelerated failure time model: a useful alternative to the Cox regression model in survival analysis. Statistics in medicine, 1992. Yu, C.-N., Greiner, R., Lin, H.-C., and Baracos, V. Learning patient-specific cancer survival distributions as a sequence of dependent regressors. In NIPS, 2011. Yu, L., Zhang, W., Wang, J., and Yu, Y. Seqgan: Sequence generative adversarial nets with policy gradient. In AAAI, 2017. Zhang, Y., Gan, Z., Fan, K., Chen, Z., Henao, R., Shen, D., and Carin, L. Adversarial feature matching for text generation. In ICML, 2017. Zhu, X., Yao, J., and Huang, J. Deep convolutional neural network for survival analysis with pathological images. In Bioinformatics and Biomedicine (BIBM), 2016 IEEE International Conference on, 2016.