# black_box_fdr__ba0a1b51.pdf Black Box FDR Wesley Tansey 1 2 Yixin Wang 1 3 David M. Blei 1 3 4 Raul Rabadan 2 Analyzing large-scale, multi-experiment studies requires scientists to test each experimental outcome for statistical significance and then assess the results as a whole. We present Black Box FDR (BB-FDR), an empirical-Bayes method for analyzing multi-experiment studies when many covariates are gathered per experiment. BB-FDR learns a series of black box predictive models to boost power and control the false discovery rate (FDR) at two stages of study analysis. In Stage 1, it uses a deep neural network prior to report which experiments yielded significant outcomes. In Stage 2, a separate black box model of each covariate is used to select features that have significant predictive power across all experiments. In benchmarks, BB-FDR outperforms competing state-of-the-art methods in both stages of analysis. We apply BB-FDR to two real studies on cancer drug efficacy. For both studies, BB-FDR increases the proportion of significant outcomes discovered and selects variables that reveal key genomic drivers of drug sensitivity and resistance in cancer. 1. Introduction High-throughput screening (HTS) techniques have fundamentally changed the landscape of modern biological experimentation. Rather than conducting just one experiment at a time, HTS enables scientists to perform hundreds of parallel experiments, each with different biological samples and different interventions. At the same time, HTS also enables scientists to gather rich contextual information about each experiment by profiling the samples under 1Data Science Institute, Columbia University, New York, NY, USA 2Department of Systems Biology, Columbia University Medical Center, New York, NY, USA 3Department of Statistics, Columbia University, New York, NY, USA 4Department of Computer Science, Columbia University, New York, NY, USA. Correspondence to: Wesley Tansey . Proceedings of the 35 th International Conference on Machine Learning, Stockholm, Sweden, PMLR 80, 2018. Copyright 2018 by the author(s). study using techniques like DNA sequencing. Thus, each HTS study produces a dataset of many experiments, where each experiment contains both an outcome variable and a high-dimensional feature set describing the context. Figure 1 shows a slice of the Genomics of Drug Sensitivity in Cancer (GDSC) dataset (Yang et al., 2012), an HTS study investigating how cancer cell lines respond to different cancer therapeutics. The left panel shows the relative response of 30 different cancer cell lines (C1, C2, . . . , C30) treated with the drug Nutlin-3. For each cell line, the treatment response (black triangles) is overlayed on top of the untreated control replicate distribution (gray box plots). Even when no drug is applied, each cell line still exhibits natural variation. The first goal in analyzing this data is therefore to address the question of whether a given cell line responded to the treatment. Concretely, we need to perform a hypothesis test for each cell line, where the null hypothesis is that the drug had no effect. Absent other information, this would be a classic multiple hypothesis testing (MHT) problem. But HTS studies such as GDSC differ from the classical setup by also producing a rich set of side information for each experiment. The right panel of Figure 1 shows a subset of the genomic profile for each cell line, with a black dot indicating the cell line has a mutation in that gene. Biologically, a mutated gene can lead to different phenotypic behavior that may cause sensitivity or resistance to a drug. Statistically, this means the likelihood of a cell line responding to treatment is a latent function of that cell line s genomic profile. Identifying which mutations are associated with treatment response could guide future experiments and development of new targeted therapies. Deriving scientific insight from patterns across experiments represents a second phase of hypothesis testing, where the null hypothesis is that a given gene is not associated with drug response. We term these two phases Stage 1 and Stage 2 and ask two scientifically-motivated statistical inference questions: Stage 1: How do we leverage the available side information in a HTS study to increase how many significant outcomes we can detect? Stage 2: Can we discover which variables are associated with significant outcomes, even when the underlying function is high-dimensional and nonlinear? Black Box FDR 6 4 2 0 2 4 Control and treatment response (z-scores) C30 C29 C28 C27 C26 C25 C24 C23 C22 C21 C20 C19 C18 C17 C16 C15 C14 C13 C12 C11 C10 C9 C8 C7 C6 C5 C4 C3 C2 C1 Cell line experiments C30 C29 C28 C27 C26 C25 C24 C23 C22 C21 C20 C19 C18 C17 C16 C15 C14 C13 C12 C11 C10 C9 C8 C7 C6 C5 C4 C3 C2 C1 Figure 1. Left: a subset of 30 cell line experiments from the Nutlin-3 case study in Section 5. Control replicates (gray box plots) and cell line responses (black triangles) are measured as z-scores relative to mean control values. Right: a subset of the corresponding genomic features for each experiment; black dots indicate a cell line has a recurrent mutation in the gene labeled on the x-axis. The goal in Stage 1 analysis is to select cell lines that showed a significant response (circled in blue). In Stage 2, the genomic features are analyzed to understand the mutations driving drug response (circled in orange). We answer both of these questions and propose Black Box FDR (BB-FDR), a method for analyzing multi-experiment studies with many covariates gathered per experiment. BBFDR uses the covariates to build a deep probabilistic model that predicts how likely a given experiment is to generate significant outcomes a priori. It uses this prior model to adaptively select significant outcomes in a manner that controls the overall false discovery rate (FDR) at a specified Stage 1 level. BB-FDR then builds a predictive model of each covariate to perform variable selection on the Stage 1 model while conserving a specified Stage 2 FDR threshold. We validate BB-FDR on both synthetic and real data. BBFDR outperforms other state-of-the-art Stage 1 methods in a series of benchmarks, including the recently-proposed Neural FDR (Xia et al., 2017). BB-FDR is also a more pragmatic choice compared to a fully-Bayesian approach: it scales trivially to thousands of covariates, can learn arbitrarily complex functions, and runs easily on a laptop. We apply BB-FDR to a real-world case study of two cancer drug screenings. BB-FDR finds more significant discoveries on the data and recovers an informative set of biologicallyplausible genes that may convey drug sensitivity and resistance in cancer. 2. Multiple testing and FDR control In the classical MHT setup, z = (z1, . . . , zn) are a set of independent observations of the outcomes of n experiments. For each experiment, a treatment is applied to a target and the treatment has either no effect (hi = 0) or some effect (hi = 1). If the treatment has no effect, the distribution of the test statistic is the null distribution f0(z); otherwise it follows an unknown alternative distribution f1(z). The null hypothesis for every experiment is that the test statistic was drawn from the null distribution: H0 : hi = 0. 2.1. False discovery rate control In most experiments of interest, it is impossible to determine hi with no error. For a given prediction ˆhi, we say it is a true positive or a true discovery if ˆhi = 1 = hi and a false positive or false discovery if ˆhi = 1 = hi. Let S = {i : hi = 1} be the set of observations for which the treatment had an effect and ˆS = {i : ˆhi = 1} be the set of predicted discoveries. We seek procedures that maximize the true positive rate (TPR) also known as power, while controlling the false discovery rate the expected proportion of the predicted discoveries that are actually false positives, FDR := E[FDP] , FDP = #{i : i ˆS\S} #{i : i ˆS} . (1) FDP in (1) is the false discovery proportion: the actual proportion of false positives in the predicted discovery set for a specific experiment. While ideally we would like to control the FDP, the randomness of the outcome variables makes this impossible in practice. Thus FDR is the typical error measure targeted in modern scientific analyses. Black Box FDR Figure 2. The graphical model for BB-FDR. 2.2. Related work Controlling FDR in multiple hypothesis testing has a long history in statistics and machine learning. The Benjamini Hochberg (BH) procedure (Benjamini & Hochberg, 1995) is the classic technique and still the most widely used in science. Many other methods have since been developed to take advantage of study-specific information to increase power. Recent examples include accumulation tests for ordering information (Li & Barber, 2017), the p-filter for grouping and test statistic dependency (Ramdas et al., 2017), FDR smoothing for spatial testing (Tansey et al., 2017), FDR-regression for low-dimensional covariates (Scott et al., 2015), and, most recently, Neural FDR for high-dimensional covariates (Xia et al., 2017). We consider high-dimensional covariates and compare against Neural FDR in Section 4. 3. Black Box FDR Consider a study with n independent experiments that produces a set of independent test statistics z = (z1, . . . , zn) corresponding to the outcome measurements, as in Section 2. However, now each experiment also has a vector of m covariates Xi = (Xi1, . . . , Xim) containing side information that may influence the outcome of that experiment. Specifically, whether the experiment comes from the null distribution hi = 0 or the alternative hi = 1 is allowed to depend arbitrarily on Xi . BB-FDR extends the empirical-Bayes two-groups model of Efron (2008) by building a hierarchical probabilistic model with experiment-specific priors modeled through a deep neural network. We first estimate the alternative distribution offline using predictive recursion (Newton, 2002) to estimate f1. This follows other recent extensions to the two-groups model (Scott et al., 2015; Tansey et al., 2017) and enjoys strong empirical performance and consistency guarantees (Tokdar et al., 2009). BB-FDR then focuses on modeling the experiment-specific prior, assuming the null and alternative distributions are fixed. 3.1. Stage 1: determining significant outcomes We model the test statistic as arising from a mixture model of two components, the null (f0) and the alternative (f1). An experiment-specific weight ci then models the prior probability of the test statistic coming from the alternative (i.e. the probability of the treatment having an effect a priori). We place a beta prior on each experiment-specific prior ci and model the parameters of the hyperprior with a black box function G parameterized by θ; in our implementation, G is a deep neural network. The complete model for BB-FDR is: zi hif1(zi) + (1 hi)f0(zi) hi Bernoulli(ci) ci Beta(ai, bi) ai, bi = Gθ,i(X) . We optimize θ by integrating out hi and maximizing the complete data log-likelihood, pθ(zi) = Z 1 0 (cif1(zi)+(1 ci)f0(zi))Beta(ci|Gθ,i(X))dci . (3) Figure 2 shows the BB-FDR graphical model. The beta prior is a departure from other two-groups extensions, which use a flatter hierarchy and learn a predictive model for ci (Scott et al., 2015; Tansey et al., 2017). We found the flat approach to be difficult to train, leading to a degenerate G that always predicts the global mean prior. A hierarchical prior improves training for two reasons. First, optimization is easier and more stable because the output of the function is two soft-plus activations. Compared to a sigmoid, this form leads to less saturated gradients. Second, the additional hierarchy allows the model to assign different degrees of confidence to each experiment, changing the model from homoskedastic to heteroskedastic. Finally, we found it important to enforce concavity of the beta distribution; we thus add 1 to both ai and bi. We fit the model in (2) with stochastic gradient descent (SGD) on an L2-regularized loss function, minimize θ R|θ| X i log pθ(zi) + λ ||Gθ(X)||2 F , (4) where || ||F is the Frobenius norm. In pilot studies, we found adding a small amount of L2-regularization prevented over-fitting at virtually no cost to statistical power. For computational purposes, we approximate the integral in (3) by a fine-grained numerical grid. Once the optimized parameters ˆθ are chosen, we calculate the posterior probability of each test statistic coming from Black Box FDR the alternative, ˆwi = pˆθ(hi = 1|zi) (5) cif1(zi)Beta(ci|Gˆθ,i(X)) cif1(zi) + (1 ci)f0(zi) dci . Assuming the posteriors are accurate, rejecting the ith hypothesis will produce 1 ˆwi false positives in expectation. Therefore we can maximize the total number of discoveries by a simple step down procedure. First, we sort the posteriors in descending order by the likelihood of the test statistics being drawn from the alternative. We then reject the first q hypotheses, where 0 q n is the largest possible index such that the expected proportion of false discoveries is below the FDR threshold. Formally, this procedure solves the optimization problem, maximize q q subject to Pq i=1(1 ˆwi) for a given FDR threshold α. By convention 0 The model in (2) (6) handles Stage 1 of the analysis. The black box model G uses the entire feature vector Xi of every experiment to predict the prior parameters over ci. The observations zi are then used to calculate the posterior probability ˆwi that the treatment had an effect. The selection procedure in (6) uses these posteriors to reject a maximum number of null hypotheses while conserving the FDR. 3.2. Stage 2: identifying important variables Using a flexible black box model for G in (2) provides a trade-off. On one hand, it enables BB-FDR to learn a rich class of functions for the relationship between the covariates and the test statistic. As we show in Section 4, this increases power in Stage 1 compared to a standard linear model. However, variable selection (Stage 2) is straightforward in linear models whereas black box models are by definition opaque. Understanding which variables deep learning models use to make predictions is an ongoing area of research in both machine learning (e.g. Elenberg et al., 2017) and specific scientific disciplines (e.g. Olden & Jackson, 2002; Giam & Olden, 2015, in ecology). As far as we are aware, there are currently no methods that provide variable selection with FDR control when the covariates may have arbitrary dependency structure. To overcome this challenge, BB-FDR uses conditional randomization tests (CRTs) (Candes et al., 2018). The idea of a CRT is to model each coordinate of the feature matrix X j using only the other coordinates X j. The conditional distribution P(X j|X j) then represents a valid null distribution for testing the hypothesis X j Z|X j, where Z is the test statistic. The corresponding p-value can be calculated by sampling from the conditional to approximate the true p-value, pj = E e X j P(X j|X j) h I h t(z, X) t(z, ( e X j, X j)) ii , where t is the test statistic of interest. Once the p-values have been estimated for all features, we can apply standard BH correction and report significant features. BB-FDR tests which features are associated with a change in the posterior probability of zi coming from the alternative. It uses the negative entropy of the posteriors as the test statistic, t(z, X) = X i ˆwi log ˆwi + X i (1 ˆwi) log(1 ˆwi) . (7) Intuitively, if a feature is useful in predicting treatment efficacy, it should reduce the overall entropy of the posterior. By definition, a feature sampled from the null adds no new information to the model; it cannot systematically reduce the entropy. For this procedure to retain frequentist consistency guarantees, both the conditional null distribution and the model of the prior must be the true distributions. In practice, one never has access to these and thus we estimate both; for the conditional null, we use gradient boosting trees (Chen & Guestrin, 2016). 4. Benchmarks We perform a series of benchmark studies to assess the performance of BB-FDR in both stages of inference. For each benchmark, we compare the power of BB-FDR to other state-of-the-art approaches. In all studies, we consider binary covariates and real-valued z-scores as test statistics. Across experiments, we found BB-FDR is particularly suitable for large samples: it outperforms competing methods in both stages while being more computationally efficient. We consider three different ground truth models for P(X), the joint distribution over the covariates, and P(h = 1|X), the prior probability of coming from the alternative distribution given the covariates: Constant: All covariates are sampled IID normal; the prior is independent of the covariates, with P(hi = 1|X) = 0.5. Linear: Covariates are sampled from a multivariate normal with full covariance matrix (i.e. conditionally linear); the prior is a linear function with IID standard normal coefficients for each covariate. Black Box FDR 10.0 7.5 5.0 2.5 0.0 2.5 5.0 7.5 10.0 z Null Well-Separated Poorly-Separated Figure 3. The two alternative densities used in our benchmarks. The well-separated (WS) density has little overlap with the null, making for a stronger signal. Nonlinear: Covariates and prior coefficients are generated similarly. We first drawing 20 IID uniform Bern(0.5) latent variables. For each covariate, 5 pairs of latent variables (ui, uj) are chosen and with equal probability are either ANDed or XORed together and multiplied by a draw from a standard normal; the latent weights are summed to get the final logit value for the covariate or coefficient. For each of the three ground truth models, we consider two different alternative distributions: Well-Separated (WS): A 3-component Gaussian mixture model, f1(z) = 0.48N( 2, 1) + 0.04N(0, 16) + 0.48N(2, 1) Poorly-Separated (PS): A single normal with high overlap with the null, f1(z) = N(0, 9). Figure 3 shows the densities used in our benchmarks. For each of the 6 combinations of the above scenarios, we run 100 independent trials. Each trial uses 50 covariates; for all trials with a non-constant prior, 25 of the variables are used in the true data generating distribution and the other 25 are null variables with no association with the outcome. To measure sample efficiency, we also vary the sample size from n = 100 to n = 50K. The target FDR threshold is set to 10% for both stages of inference. We compare BB-FDR to the classic Benjamini-Hochberg (BH) method (Benjamini & Hochberg, 1995), the recentlyproposed Neural FDR (Xia et al., 2017), and a fully Bayesian logistic regression model for ci in place of the black box prior in (2). For Neural FDR, we use the default recommended settings, including five random restarts and a ten-layer deep neural network. The fully-Bayesian method uses a standard normal prior on the weights and an inverse-Wishart prior on the variance, with weak hyperpriors. In the nonlinear scenario, we specify all possible pairwise interactions as the covariate set for the fully-Bayesian model to ensure it is well-specified. We fit the model using Polya-gamma sampling (Polson et al., 2013) with 5000 burn-in iterations and 1000 samples. For BB-FDR, we use a 50 200 200 2 network with Re LU activation; for training we use RMS-prop (Tieleman & Hinton, 2012) with dropout, learning rate 3 10 4, and batch size 100, and run for 50 epochs, with 3 folds to create 3 separate models as in Neural FDR; we set the λ regularization term to 10 4. 4.2. Stage 1 performance Figure 4 shows the results for the Stage 1 benchmarks, where the goal is to determine for which experiments the treatment had an effect. The four methods generally conserve FDR at the specified 10% threshold, though Neural FDR seems to systematically violate FDR in the lowsample regime. Across all experiments, we see that both BH and Neural FDR under-perform the two Bayesian methods. In the case of BH, this is straight-forward as it uses only the p-value from each experiment and has no notion of side information. Neural FDR, on the other hand, uses a deep neural network and several random restarts. There are a few possible reasons for its poor performance. First, the Neural FDR method was reported to be very difficult to train by the original authors, so it is possible that it is simply not finding good fits of the model. Second, BB-FDR assumes that the alternative distribution is conditionally independent of the prior; Neural FDR makes no such assumption and may lose some power as a result. Finally, Neural FDR was originally tested on 1and 2-dimensional problems against relatively weak baselines. Our benchmarks examine its performance in a higher-dimensional setting and with several uninformative features that may make fitting Neural FDR difficult. Since the fully-Bayesian method is well-specified in every benchmark, it serves as an oracle model to establish a reasonable upper bound on Stage 1 performance. However, the oracle power depends on the MCMC approximation of the posterior being well-mixed. As the sample size grows, the empirical-Bayes model used by BB-FDR gains an increasingly precise approximation to the true posterior. In the large-sample regime with a well-separated alternative, BB-FDR outperforms even the oracle. Furthermore, the fully-Bayesian method takes several hours to fit in the nonlinear scenarios; BB-FDR fits within a few minutes and can easily be run on a laptop. Black Box FDR 102 103 104 Samples Power and FDR BH Full-Bayes Neural FDR BB-FDR (a) Constant (PS) 102 103 104 Samples Power and FDR (b) Constant (WS) 102 103 104 Samples Power and FDR (c) Linear (PS) 102 103 104 Samples Power and FDR (d) Linear (WS) 102 103 104 Samples Power and FDR (e) Nonlinear (PS) 102 103 104 Samples Power and FDR (f) Nonlinear (WS) Figure 4. Hypothesis testing results on the synthetic datasets averaged over 100 trials at varying sample sizes on the two different alternative distributions. Solid lines show power; dashed lines show estimated FDR; the red horizontal line denotes the specified 10% FDR threshold. In general, the Benjamini-Hochberg and Neural FDR methods have lower power since they do not model the alternative. The fully-Bayesian method has high power in the low-to-moderate sample regime, but as the sample size grows the empirical-Bayes approach of BB-FDR becomes more effective. Black Box FDR 102 103 104 Samples Power and FDR Full-Bayes BB-FDR (a) Linear (PS) 102 103 104 Samples Power and FDR (b) Nonlinear (PS) Figure 5. Variable selection results at a 10% FDR threshold. In low sample regimes, the conditional null distribution used in the CRT procedure is poorly fit and results in violated FDR thresholds. At moderate-to-large samples, BB-FDR has higher power than the fully-Bayesian model and conserves FDR. 4.3. Stage 2 performance Neither BH nor Neural FDR provide support for detecting important features (Stage 2), so we could not compare against them. For the Bayesian linear regression, we take the 90% posterior credible interval over the coefficient value for each covariate. If the interval does not contain zero, we reject the null hypothesis and report it as a discovered feature; this approach is standard in the Bayesian literature (Gelman et al., 2014). Figure 5 presents the results of the variable selection benchmarks for the poorly-separated alternative distribution; results for the well-separated are similar. We omit the constant scenario, since there are no features to discover. In the small sample regime, the conditional distributions are poor estimators of the conditional null distribution for each feature. This leads to BB-FDR overestimating the number of signal features and violating the FDR threshold. As the sample size grows, the conditional null and black box prior become more accurate, leading to FDR control and higher power, respectively; in the large-sample regime, BB-FDR outperforms the fully-Bayesian approach. We conclude by noting that BB-FDR is competitive with the fully-Bayesian approach even when the latter is wellspecified. In practical data analysis scenarios, such as the cancer study we discuss next, we do not know the true prior function. It may easily contain many higher-order interaction terms that are prohibitive to consider explicitly in a fully-Bayesian model, making BB-FDR a pragmatic choice for real-world scientific datasets. 5. Cancer drug screening As a case study of how BB-FDR is useful in practice, we apply it to two high-throughput cancer drug screening studies (Lapatinib and Nutlin-3) from the Genomics of Drug Sensitivity in Cancer (GDSC) (Yang et al., 2012). For both drug studies, BB-FDR increases the number of Stage 1 discoveries over classical BH correction; results on Neural FDR were similar to BH and are omitted. In Stage 2, BB-FDR discovers biologically-plausible genes that may have a causal link to drug sensitivity and resistance. Experimental details are available in the supplement. 5.1. Analysis overview Analysis of the two drug studies broadly follows the two stages outlined in the motivating example in Section 1. The Stage 1 task is to determine, for a specific drug being tested on a specific cell line, whether the drug had any effect. As with any biological process, natural variation injects randomness at many levels of the experiment: how fast the cells grow, how each cell responds, etc. Thus Stage 1 requires performing statistical hypothesis testing to determine if the cell population after treatment is truly smaller than would be expected from a control (null) population. The inferential goal in Stage 2 is to gain scientific insight about which genes may be driving drug response. This involves building a statistical model of the relationship between the genomic profile of a cell line and its corresponding treatment response, then performing variable selection on the model. The selected genes form the basis for potential mechanisms of action and future experiments can be designed to test for a causal link or to investigate new drugs that better-target the proteins encoded by the genes. Black Box FDR Lapatinib Nutlin-3 BRCA1, BRCA2, CDK4 P300, FLCN, FLT3 FGFR2, KIT, MSH2 MET, KIT, MSH6 SETD2, TP53, BCR-ABL Table 1. Significant gene mutations identified by BB-FDR that are associated with sensitivity and resistance to each drug. Both lists align well with known genomic targets of Lapatinib and Nutlin-3. 5.2. Results Figure 6 shows the aggregate number of treatment effects discovered by both BH and BB-FDR. For both drugs, BBFDR provides approximately a 50% increase in Stage 1 discoveries compared to BH. The genomic profiles of the cell lines provide enough prior information that even some outcomes with a z-score above zero are still found to be significant. This flexibility is impossible with classical Stage 1 testing methods like BH that do not consider covariate information. Table 1 lists the genes reported by BB-FDR in Stage 2. Interpreting the quality of the results requires familiarity with genomics and cancer biology. Below, we briefly detail the scientific rational behind the biological plausibility of the Stage 2 results and refer the reader to Weinberg (2013) for a full review. 10 8 6 4 2 0 2 z Null Discoveries (a) BH on Lapatinib (117 discoveries) 10 8 6 4 2 0 2 z (b) BH on Nutlin-3 (151 discoveries) 10 8 6 4 2 0 2 z (c) BB-FDR on Lapatinib (181 discoveries) 10 8 6 4 2 0 2 z (d) BB-FDR on Nutlin-3 (222 discoveries) Figure 6. Discoveries found by BB-FDR on the two drugs, compared to the discoveries found by a naive BH approach. BB-FDR leverages the genomic profiling information of the cell lines to identify 50% more discoveries at the same 10% FDR threshold. Lapatinib has been approved for the treatment of HER2positive breast cancers. BB-FDR indicates that BRCA1 and BRCA2 are associated with responses to Lapatinib. Both are tumor suppressor genes that are seen mutated in more than 10% of breast cancers (BRCA stands for breast cancer ) and thus cancer type may represent a latent confounder for drug efficacy that induces a conditional dependence. Lapatinib targets over-expression of the gene ERBB2 which can be caused by a mutant CDK4 gene. FGFR2 and KIT are also commonly associated with breast cancers (Slattery et al., 2013; Zhu et al., 2014) and BRCA1 is known to induce inactivation of the tumor suppressor MSH2 (Atalay et al., 2002). Given Lapatinib s success as a breast cancer drug, the connection between all of the selected genes and breast cancer is a reassuring sign that BB-FDR selected biologically plausible features. Nutlin-3 is an inhibitor of the oncogene MDM2, which negatively-regulates TP53. When highly over-expressed, MDM2 can functionally inactivate TP53. By targeting MDM2, Nutlin-3 enables a non-mutated ( wild type ) TP53 to trigger apoptosis in cancer cells. However, if TP53 is mutated, Nutlin-3 will be ineffective and hence its mutation state is an important driver of Nutlin-3 sensitivity. When wild type TP53 is present, MET controls the fate of the cell (Sullivan et al., 2012), SETD2 functionality is required to activate TP53 (Carvalho et al., 2014), P300 mediates TP53 acetylation (Reed & Quelle, 2014), and BCR-ABL is a gene fusion that induces loss of TP53 (Pierce et al., 2000). These genes interact in complex, non-linear ways, yet BB-FDR is still able to identify them as important. Finally, FLCN is a tumor suppressor gene that can delay cell cycle like TP53 (Laviolette et al., 2013). The mechanism by which FLCN and TP53 are interrelated is currently unclear, representing a potential target for future experiments. 6. Conclusion We presented Black Box FDR (BB-FDR), an empirical Bayes method that increases statistical power in multiexperiment scientific studies when side information is available for each experiment. BB-FDR combines deep probabilistic modeling with recent multiple testing techniques to boost testing power without sacrificing interpretability. Benchmarks show that BB-FDR outperforms state-of-theart techniques, often substantially and under a wide array of experimental conditions. BB-FDR also finds more experimental discoveries two cancer drug screening datasets and provides scientific insight into the mechanisms associated with differential treatment response in cancer. This work was supported by a pilot grant from Columbia University, NIH U54 CA193313, ONR N00014-15-1-2209, ONR 133691-5102004, NIH 5100481-5500001084, NSF CCF-1740833, the Alfred P. Sloan Foundation, the John Simon Guggenheim Foundation, Facebook, Amazon, and IBM. Black Box FDR Atalay, A., Crook, T., Ozturk, M., and Yulug, I. G. Identification of genes induced by brca1 in breast cancer cells. Biochemical and biophysical research communications, 299(5):839 846, 2002. Benjamini, Y. and Hochberg, Y. Controlling the false discovery rate: a practical and powerful approach to multiple testing. Journal of the royal statistical society. Series B (Methodological), pp. 289 300, 1995. Candes, E., Fan, Y., Janson, L., and Lv, J. Panning for gold:model-xknockoffs for high dimensional controlled variable selection. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 2018. Carvalho, S., V ıtor, A. C., Sridhara, S. C., Martins, F. B., Raposo, A. C., Desterro, J. M., Ferreira, J., and de Almeida, S. F. Setd2 is required for dna double-strand break repair and activation of the p53-mediated checkpoint. Elife, 3, 2014. Chen, T. and Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, pp. 785 794. ACM, 2016. Efron, B. Microarrays, empirical bayes and the two-groups model. Statistical science, pp. 1 22, 2008. Elenberg, E., Dimakis, A. G., Feldman, M., and Karbasi, A. Streaming weak submodularity: Interpreting neural networks on the fly. In Advances in Neural Information Processing Systems, pp. 4047 4057, 2017. Gelman, A., Carlin, J. B., Stern, H. S., Dunson, D. B., Vehtari, A., and Rubin, D. B. Bayesian data analysis, volume 2. CRC press Boca Raton, FL, 2014. Giam, X. and Olden, J. D. A new r2-based metric to shed greater insight on variable importance in artificial neural networks. Ecological Modelling, 313:307 313, 2015. Laviolette, L. A., Wilson, J., Koller, J., Neil, C., Hulick, P., Rejtar, T., Karger, B., Teh, B. T., and Iliopoulos, O. Human folliculin delays cell cycle progression through late s and g2/m-phases: effect of phosphorylation and tumor associated mutations. Plo S one, 8(7):e66775, 2013. Li, A. and Barber, R. F. Accumulation tests for fdr control in ordered hypothesis testing. Journal of the American Statistical Association, 112(518):837 849, 2017. Newton, M. A. A nonparametric recursive estimator of the mixing distribution. Sankhya, Series A, 64:306 22, 2002. Olden, J. D. and Jackson, D. A. Illuminating the black box: a randomization approach for understanding variable contributions in artificial neural networks. Ecological modelling, 154(1-2):135 150, 2002. Pierce, A., Spooncer, E., Wooley, S., Dive, C., Francis, J. M., Miyan, J., Owen-Lynch, P. J., Dexter, T. M., and Whetton, A. D. Bcr-abl protein tyrosine kinase activity induces a loss of p53 protein that mediates a delay in myeloid differentiation. Oncogene, 19(48):5487, 2000. Polson, N. G., Scott, J. G., and Windle, J. Bayesian inference for logistic models using p olya gamma latent variables. Journal of the American statistical Association, 108(504):1339 1349, 2013. Ramdas, A., Barber, R. F., Wainwright, M. J., and Jordan, M. I. A unified treatment of multiple testing with prior knowledge. ar Xiv preprint ar Xiv:1703.06222, 2017. Reed, S. M. and Quelle, D. E. p53 acetylation: regulation and consequences. Cancers, 7(1):30 69, 2014. Scott, J. G., Kelly, R. C., Smith, M. A., Zhou, P., and Kass, R. E. False discovery rate regression: an application to neural synchrony detection in primary visual cortex. Journal of the American Statistical Association, 110(510): 459 471, 2015. Slattery, M. L., John, E. M., Stern, M. C., Herrick, J., Lundgreen, A., Giuliano, A. R., Hines, L., Baumgartner, K. B., Torres-Mejia, G., and Wolff, R. K. Associations with growth factor genes (fgf1, fgf2, pdgfb, fgfr2, nrg2, egf, erbb2) with breast cancer risk and survival: the breast cancer health disparities study. Breast cancer research and treatment, 140(3):587 601, 2013. Sullivan, K. D., Padilla-Just, N., Henry, R. E., Porter, C. C., Kim, J., Tentler, J. J., Eckhardt, S. G., Tan, A. C., De Gregori, J., and Espinosa, J. M. Atm and met kinases are synthetic lethal with nongenotoxic activation of p53. Nature chemical biology, 8(7):646, 2012. Tansey, W., Koyejo, O., Poldrack, R. A., and Scott, J. G. False discovery rate smoothing. Journal of the American Statistical Association, (just-accepted), 2017. Tieleman, T. and Hinton, G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA: Neural networks for machine learning, 4 (2):26 31, 2012. Tokdar, S., Martin, R., and Ghosh, J. Consistency of a recursive estimate of mixing distributions. The Annals of Statistics, 37(5A):2502 22, 2009. Weinberg, R. The biology of cancer. Garland science, 2013. Black Box FDR Xia, F., Zhang, M. J., Zou, J. Y., and Tse, D. Neuralfdr: Learning discovery thresholds from hypothesis features. In Advances in Neural Information Processing Systems, pp. 1540 1549, 2017. Yang, W., Soares, J., Greninger, P., Edelman, E. J., Lightfoot, H., Forbes, S., Bindal, N., Beare, D., Smith, J. A., Thompson, I. R., et al. Genomics of drug sensitivity in cancer (gdsc): a resource for therapeutic biomarker discovery in cancer cells. Nucleic acids research, 41(D1): D955 D961, 2012. Zhu, Y., Wang, Y., Guan, B., Rao, Q., Wang, J., Ma, H., Zhang, Z., and Zhou, X. C-kit and pdgfra gene mutations in triple negative breast cancer. International journal of clinical and experimental pathology, 7(7):4280, 2014.