# explaining_hyperparameter_optimization_via_partial_dependence_plots__45e35f03.pdf Explaining Hyperparameter Optimization via Partial Dependence Plots Julia Moosbauer , Julia Herbinger , Giuseppe Casalicchio, Marius Lindauer, Bernd Bischl Department of Statistics, Ludwig-Maximilians-University Munich, Munich, Germany Institute of Information Processing, Leibniz University Hannover, Hannover, Germany {julia.moosbauer, julia.herbinger, giuseppe.casalicchio, bernd.bischl}@stat.uni-muenchen.de lindauer@tnt.uni-hannover.de Automated hyperparameter optimization (HPO) can support practitioners to obtain peak performance in machine learning models. However, there is often a lack of valuable insights into the effects of different hyperparameters on the final model performance. This lack of explainability makes it difficult to trust and understand the automated HPO process and its results. We suggest using interpretable machine learning (IML) to gain insights from the experimental data obtained during HPO with Bayesian optimization (BO). BO tends to focus on promising regions with potential high-performance configurations and thus induces a sampling bias. Hence, many IML techniques, such as the partial dependence plot (PDP), carry the risk of generating biased interpretations. By leveraging the posterior uncertainty of the BO surrogate model, we introduce a variant of the PDP with estimated confidence bands. We propose to partition the hyperparameter space to obtain more confident and reliable PDPs in relevant sub-regions. In an experimental study, we provide quantitative evidence for the increased quality of the PDPs within sub-regions. 1 Introduction Most machine learning (ML) algorithms are highly configurable. Their hyperparameters must be chosen carefully, as their choice often impacts the model performance. Even for experts, it can be challenging to find well-performing hyperparameter configurations. Automated machine learning (Auto ML) systems and methods for automated HPO have been shown to yield considerable efficiency compared to manual tuning by human experts [Snoek et al., 2012]. However, these approaches mainly return a well-performing configuration and leave users without insights into decisions of the optimization process. Questions about the importance of hyperparameters or their effects on the resulting performance often remain unanswered. Not all data scientists trust the outcome of an Auto ML system due to the lack of transparency [Drozdal et al., 2020]. Consequently, they might not deploy an Auto ML model, despite all performance gains. Providing insights into the search process may help increase trust and facilitate interactive and exploratory processes: A data scientist could monitor the Auto ML process and make changes to it (e.g., restricting or expanding the search space) already during optimization to anticipate unintended results. Transparency, trust, and understanding of the inner workings of an Auto ML system can be increased by interpreting the internal surrogate model of an Auto ML approach. For example, BO trains a surrogate model to approximate the relationship between hyperparameter configurations and model performance. It is used to guide the optimization process towards the most promising regions of the hyperparameter space. Hence, surrogate models implicitly contain information about the influence of These authors contributed equally to this work. 35th Conference on Neural Information Processing Systems (Neur IPS 2021). hyperparameters. If the interpretation of the surrogate matches with a data scientist s expectation, confidence in the correct functioning of the system may be increased. If these do not match, it provides an opportunity to look either for bugs in the code or for new theoretical insights. We propose to analyze surrogate models with methods from IML to provide insights into the results of HPO. In the context of BO, typical choices for surrogate models are flexible, probabilistic blackbox models, such as Gaussian processes (GP) or random forests. Interpreting the effect of single hyperparameters on the performance of the model to be tuned is analogous to interpreting the feature effect of the black-box surrogate model. We focus on the PDP [Friedman, 2001], which is a widely-used method2 to visualize the average marginal effect of single features on a black-box model s prediction. When applied to surrogate models, they provide information on how a specific hyperparameter influences the estimated model performance. However, applying PDPs out of the box to surrogate models might lead to misleading conclusions. Efficient optimizers such as BO tend to focus on exploiting promising regions of the hyperparameter space while leaving other regions less explored. Therefore, a sampling bias in input space is introduced, which in turn can lead to a poor fit and biased interpretations in underexplored regions of the space. Contributions: We study the problem of sampling bias in experimental data produced by Auto ML systems and the resulting bias of the surrogate model and assess its implications on PDPs. We then derive an uncertainty measure for PDPs of probabilistic surrogate models. In addition, we propose a method that splits the hyperparameter space into interpretable sub-regions of varying uncertainty to obtain sub-regions with more reliable and confident PDP estimates. In the context of BO, we provide evidence for the usefulness of our proposed methods on a synthetic function and in an experimental study in which we optimize the architecture and hyperparameters of a deep neural network. Our Supplementary Material provides (A) more background related to uncertainty estimates, (B) notes on how our methods are applied to hierarchical hyperparameter spaces, (C) details on the experimental setup and more detailed results, (D) a link to the source code. Reproducibility and Open Science: The implementation of the proposed methods as well as reproducible scripts for the experimental analysis are provided in a public git-repository3. 2 Background and Related Work Recent research has begun to question whether the evaluation of an Auto ML system should be purely based on the generated models predictive performance without considering interpretability [Hutter et al., 2014a, Pfisterer et al., 2019, Freitas, 2019, Xanthopoulos et al., 2020]. Interpreting Auto ML systems can be categorized as (1) interpreting the resulting ML model on the underlying dataset, or (2) interpreting the HPO process itself. In this paper, we focus on the latter. Let c : ! R be a black-box cost function, mapping a hyperparameter configuration λ = (λ1, ..., λd) to the model error4 obtained by a learning algorithm with configuration λ. The hyperparameter space may be mixed, containing categorical and continuous hyperparameters. The goal of HPO is to find λ 2 arg minλ2 c(λ). Throughout the paper, we assume that a surrogate model ˆc : ! R is given as an approximation to c. If the surrogate is assumed to be a GP, ˆc(λ) is a random variable following a Gaussian posterior distribution. In particular, for any finite indexed family of hyperparameter configurations λ(1), ..., λ(k)" 2 k, the vector of estimated performance values is Gaussian with a posterior mean ˆ m = i=1,...,k and covariance ˆ K = λ(i), λ(j)" i,j=1,...,k. Hyperparameter Importance. Understanding which hyperparameters influence model performance can provide valuable insights into the tuning strategy [Probst et al., 2019]. To quantify relevance of hyperparameters, models that inherently quantify feature relevance e.g., GPs with ARD kernel [Neil, 1996] can be used as surrogate models. Hutter et al. [2014a] quantified the importance of hyperparameters based on a random forest fitted on data generated by BO, for which the importance of both the main and the interaction effects of hyperparameters was calculated by a functional ANOVA approach. Similarly, Sharma et al. [2019] quantified the hyperparameter importance of 2There exist various implementations [Greenwell, 2017, Pedregosa et al., 2011]), extensions [Greenwell et al., 2018, Goldstein et al., 2015] and applications [Friedman and Meulman, 2003, Cutler et al., 2007]. 3https://github.com/slds-lmu/paper_2021_xautoml 4Typically, the model error is estimated via cross-validation or hold-out testing. residual neural networks. These works highlight how useful it is to quantify the importance of hyperparameters. However, importance scores do not show how a specific hyperparameter affects the model performance according to the surrogate model. Therefore, we propose to visualize the assumed marginal effect of a hyperparameter. A model-agnostic interpretation method that can be used for this purpose is the PDP. PDPs for Hyperparameters. Let S {1, 2, ..., d} denote an index set of features, and let C = {1, 2, ..., d} \ S be its complement. The partial dependence (PD) function [Friedman, 2001] of c : ! R for hyperparameter(s) S is defined as5 c S(λS) := EλC [c(λ)] = c(λS, λC) d P(λC). (1) When analyzing the PDP of hyperparameters, we are usually interested in how their values λS impact model performance uniformly across the hyperparameter space. In line with prior work [Hutter et al., 2014a], we therefore assume P to be the uniform distribution over C. Computing c S(λS) exactly is usually not possible because c is unknown and expensive to evaluate in the context of HPO. Thus, the posterior mean ˆm of the probabilistic surrogate model ˆc(λ) is commonly used as a proxy for c. Furthermore, the integral may not be analytically tractable for arbitrary surrogate models ˆc. Hence, the integral is approximated by Monte Carlo integration, i.e., ˆc S (λS) = 1 n for a sample i=1,...,n P(λC). ˆm represents the marginal effect of λS for one specific instance i. Individual conditional expectation (ICE) curves [Goldstein et al., 2015] visualize the marginal effect of the i-th observation by plotting the value of ˆm against λS for a set of grid points6 λ(g) S 2 S, g 2 {1, ..., G}. Analogously, the PDP visualizes ˆc S(λS) against the grid points. Following from Eq. 2, the PDP visualizes the average over all ICE curves. In HPO, the marginal predicted performance is a related concept. Instead of approximating the integral via Monte Carlo, the integral over ˆc is computed exactly. Hutter et al. [2014a] propose an efficient approach to compute this integral for random forest surrogate models. Uncertainty Quantification in PDPs. Quantifying the uncertainty of PDPs provides additional information about the reliability of the mean estimator. Hutter et al. [2014a] quantified the model uncertainty specifically for random forests as surrogates in BO by calculating the standard deviation of the marginal predictions of the individual trees. However, this procedure is not applicable to general probabilistic surrogate models, such as the commonly used GP. There are approaches that quantify the uncertainty for ML models that do not provide uncertainty estimates out-of-the-box. Cafri and Bailey [2016] suggested a bootstrap approach for tree ensembles to quantify the uncertainties of effects based on PDPs. Another approach to quantify the uncertainty of PDPs is to leverage the ICE curves. For example, Greenwell [2017] implemented a method that marginalizes over the mean standard deviation of the ICE curves for each grid point. However, this approach quantifies the underlying uncertainty of the data at hand rather than the model uncertainty, as explained in Appendix A.1. A model-agnostic estimate based on uncertainty estimates for probabilistic models is missing so far. Subgroup PDPs. Recently, a new research direction concentrates on finding more reliable PDP estimates within subgroups of observations. Molnar et al. [2020] focused on problems in PDP estimation with correlated features. To that end, they apply transformation trees to find homogeneous subgroups and then visualize a PDP for each subgroup. Grömping [2020] looked at the same problem and also uses subgroup PDPs, where ICE curves are grouped regarding a correlated feature. Britton [2019] applied a clustering approach to group ICE curves to find interactions between features. However, none of these approaches aim at finding subgroups where reliable PDP estimates have low uncertainty. Additionally, to the best of our knowledge, nothing similar exists for analyzing experimental data created by HPO. 5To keep notation simple, we denote c(λ) as a function of two arguments (λS, λC) to differentiate components in the index set S from those in the complement. The integral shall be understood as a multiple integral of c where λj, j 2 C, are integrated out. 6Grid points are typically chosen as an equidistant grid or sampled from P(λS). The granularity G is chosen by the user. For categorical features, the granularity typically corresponds to the number of categories. Figure 1: Illustration of the sampling bias when optimizing the 2D Styblinski Tang function with BO and the Lower Confidence Bound (LCB) acquisition function a(λ) = ˆm(λ) + ˆs(λ) for = 0.1 (left) and = 2 (middle) vs. data sampled uniformly at random (right). Figure 2: The two horizontal cuts (left) yield two ICE curves (right) showing the mean prediction and uncertainty band against λ1 for ˆc with = 0.1 on the 2D Styblinski-Tang function. The upper ICE curve deviates more from the true effect (black) and shows a higher uncertainty. 3 Biased Sampling in HPO Visualizing the marginal effect of hyperparameters of surrogate models via PDPs can be misleading. We show that this problem is due to the sequential nature of BO, which generates dependent instances (i.e., hyperparameter configurations) and thereby introduces a sampling and a resulting model bias. To save computational resources in contrast to grid search or random search, efficient optimizers like BO tend to exploit promising regions of the hyperparameter space while other regions are less explored (see Figure 1). Consequently, predictions of surrogate models are usually more accurate with less uncertainty in well-explored regions and less accurate with high uncertainty in under-explored regions. This model bias also affects the PD estimate (see Figure 2). ICE curves may be biased and less confident if they are computed in poorly-learned regions where the model has not seen much data before. Under the assumption of uniformly distributed hyperparameters, poorly-learned regions are incorporated in the PD estimate with the same weight as well-learned regions. ICE curves belonging to regions with high uncertainty may obfuscate well-learned effects of ICE curves belonging to other regions when they are aggregated to a PDP. Hence, the model bias may also lead to a less reliable PD estimate. PDPs visualizing only the mean estimator of Eq. (2) do not provide insights into the reliability of the PD estimate and how it is affected by the described model bias. 4 Quantifying Uncertainty in PDPs Figure 3: PDPs (blue) with confidence bands for surrogates trained on data created by BO and LCB with = 0.1 (left), = 1 (middle) and uniform i.i.d. dataset (right) vs. the true PD (black). Pointwise uncertainty estimates of a probabilistic model provide insights into the reliability of the prediction ˆc(λ) for a specific configuration λ. This uncertainty directly correlates with how explored the region around λ is. Hence, including the model s uncertainty structure into the PD estimate enables users to understand in which regions the PDP is more reliable and which parts of the PDP must be cautiously interpreted.7 We now extend the PDP of Eq. (2) to probabilistic surrogate models ˆc (e.g., a GP). Let λS be a fixed grid point and i=1,...,n P (λC) a sample that is used to compute the Monte Carlo estimate of Eq. (2). The vector of predicted performances at the grid point λS is ˆc (λS) = i=1,...,n with (posterior) mean ˆ m (λS) := i=1,...,n and 7Note that we aim at representing model uncertainty in a PD estimate, and not the variability of the mean prediction (see Appendix A.1 for a more detailed justification). a (posterior) covariance ˆ K (λS) := i,j=1,...,n. Thus, ˆc S (λS) = is a random variable itself. The expected value of ˆc S (λS) corresponds to the PD of the posterior mean function ˆm at λS, i.e.: ˆm S (λS) = Eˆc [ˆc S (λS)] = Eˆc The variance of ˆc S (λS) is S(λS) = Vˆc [ˆc S (λS)] = Vˆc n2 1> ˆ K (λS) 1. (4) For the above estimate, it is important that the kernel is correctly specified such that the covariance structure is modeled properly by the surrogate model. Eq. (4) can be approximated empirically by treating the pairwise covariances as unknown, i.e.: ˆ K (λS)i,i . (5) In Appendix A.2, we show empirically that this approximation is less sensitive to kernel misspecifications. Please note that the variance estimate and the mean estimate can also be applied to other probabilistic models, such as GAMLSS8, transformation trees, or a random forest. An example for PDPs with uncertainty estimates is shown in Figure 3 for different degrees of a sampling bias. 5 Regional PDPs via Confidence Splitting As discussed in Section 3, (efficient) optimization may imply that the sampling is biased, which in turn can produce misleading interpretations when IML is naively applied. We now aim to identify subregions 0 of the hyperparameter space in which the PD can be estimated with high confidence, and separate those from sub-regions in which it cannot be estimated reliably. In particular, we identify sub-regions in which poorly-learned effects do not obfuscate the well-learned effects along each grid point, thereby allowing the user to draw conclusions with higher confidence. By partitioning the entire hyperparameter space through a tree-based approach into disjoint and interpretable sub-regions, a more detailed understanding of the sampling process and hyperparameter effects is achieved. Users can either study the hyperparameter effect of a (confident) sub-region individually or understand the exploration-exploitation sampling of HPO by considering the complete tree structure. The result of this procedure for a single split is shown in Figure 5. The PD estimate on the entire hyperparameter space is computed by sampling the Monte Carlo estimate (λ(i) C )i2N P(λC), N := {1, 2, ..., n}. We now introduce the PD estimate on a subregion 0 simply as (λ(i) C )i2N 0 only using N 0 = {i 2 N}λ(i)2 0. Since we are interested in the marginal effect of the hyperparameter(s) S at each λS 2 S, we will usually visualize the PD for the whole range S. Thus, all obtained sub-regions should be of the form 0 = S 0 C C. This corresponds to an average of ICE curves in the set i 2 N 0. The pseudo-code to partition a hyperparameter (sub-)space and corresponding sample (λ(i) C )i2N 2 C, N {1, ..., n}, into two child regions is shown in Algorithm 1. This splitting is recursively applied in a CART9-like procedure [Breiman et al., 1984b] to expand a full tree structure, with the usual stopping criteria (e.g., a maximum number of splits, a minimum size of a region, or a minimum improvement in each node). In each leaf node, the sub-regional PDP and its corresponding uncertainty estimate are computed by aggregating over all contained ICE curves. The criterion to evaluate a specific partitioning is based on the idea of grouping ICE curves with similar uncertainty structure. To be more exact, we evaluate the impurity of a PD estimate on a sub-region 0 with the help of the associated set of observations N 0 = {i 2 N}λ(i) C, also referred to as nodes, as follows: For each grid point λS, we use the L2 loss in L (λS, N 0) to evaluate how the 8Generalized additive models for location, scale and shape 9Classification and regression trees Figure 4: ICE curves of ˆs of λS for the left (green) and right (blue) sub-region after the first split. The darker lines represent the respective PDPs. The orange vertical line marks the value λS of the optimal configuration. λj < 6.9 λj 6.9 Figure 5: Example of two estimated PDPs (blue line) and 95% confidence bands after one partitioning step. The orange vertical line is the value of λS from the optimal configuration, the black curve is the true PD estimate c S(λS). uncertainty varies across all ICE estimates i 2 N 0 using ˆs2 S|N 0 (λS) := 1 |N 0| and aggregate the loss L (λS, N 0) over all grid points in RL2(N 0): L (λS, N 0) = and RL2(N 0) = S , N 0). (6) Algorithm 1: Tree-based Partitioning input: N for j 2 C do for Every split t on hyperparameter λj do l = {i 2 N}λ(i) r = {i 2 N}λ(i) j >t I(j, t) = RL2(N j,t l ) + RL2(N j,t r ) end for end for Choose 2 arg minj,t I(j, t) Return N j,t l and N j,t r for (j, t) = Hence, we measure the pointwise L2-distance between ICE curves of the variance function ˆs2(λS, λ(i) C ) and its PD estimate ˆs2 S|N 0 (λS) within a sub-region N 0. This seems reasonable, as ICE curves in well-explored regions of the search space should, on average, have a lower uncertainty than those in less-explored regions. However, since we only split according to hyperparameters in C but not in S, the partitioning does not cut off less explored regions w.r.t. λS. Thus, the chosen split criterion groups ICE curves of the uncertainty estimate such that we receive sub-regions associated with low costs c (and thus high relevance for a user) to be more confident in well-explored regions of λS and less confident in under-explored regions. Figure 4 shows that ICE curves of the uncertainty measure with high uncertainty over the entire range of λS are grouped together (right sub-region). Those with low uncertainty close to the optimal configuration of λS and increasing uncertainties for less suitable configurations are grouped together by curve similarities in the left sub-region. The respective PDPs are illustrated in Figure 5, where the confidence band in the left sub-region decreased compared to the confidence band of the global PDP especially for grid points close to the optimal value of λS. Hence, by grouping observations with similar ICE curves of the variance function, resulting sub-regional PDPs with confidence bands provide the user with the information of which sub-regions of C are well-explored and lead to more reliable PDP estimates. Furthermore, the user will know which ranges of λS can be interpreted reliably and which ones need to be regarded with caution. To sum up, the splitting procedure provides interpretable, disjoint sub-regions of the hyperparameter space. Based on the defined impurity measure, PDPs with high reliability can be identified and analyzed. In particular, the method provides more confident and reliable estimates in the sub-region containing the optimal configuration. Which PDPs are most interesting to explore depends on the question the user would like to answer. If the main interest lies in understanding the optimization and exploring the sampling process, a user might want to keep the number of sub-regions relatively low by performing only a few partitioning steps. Subsequently, one would investigate the overall structure of the sub-regions and the individual sub-regional PDPs. If users are more interested in interpreting hyperparameter effects only in the most relevant sub-region(s), they may want to split deeper and only look at sub-regions that are more confident than the global PDP. Due to the nature of the splitting procedure, the PDP estimate on the entire hyperparameter space is a weighted average of the respective sub-regional PDPs. Hence, the global PDP estimate is decomposed into several sub-regional PDP estimates. Furthermore, note that the proposed method does not assume a numeric hyperparameter space, since the uncertainty estimates as well as ICE and PDP estimates can also be calculated for categorical features. Thus, it is applicable to problems with mixed spaces as long as a probabilistic surrogate model and particularly its uncertainty estimates are available. In Appendix B we describe how our method is applied to hierarchical hyperparameter spaces. Since the proposed method is an instance of the CART algorithm, finding the optimal split for a categorical variable with q levels generally involves checking 2q subsets. This becomes computationally infeasible for high values of q. It remains an open question for future work if this can be sped by an optimal procedure as in regression with L2 loss [Fisher, 1958] and binary classification [Breiman et al., 1984a] or by a clever heuristic as for multiclass classification Wright and König [2019]. 6 Experimental Analysis In this section, we validate the effectiveness of the introduced methods. We formulate two main hypotheses: First, experimental data affected by the sampling bias lead to biased surrogate models and thus to unreliable and misleading PDPs. Second, the proposed partitioning allows us to identify an interpretable sub-region (around the optimal configuration) that yields a more reliable and confident PDP estimate. In a first experiment, we apply our methods to BO runs on a synthetic function. In this controlled setup, we investigate the validity of our hypotheses with regards to problems of different dimensionality and different degrees of sampling bias. In a second experiment, we evaluate our PDP partitioning in the context of HPO for neural networks on a variety of tabular datasets. We assess the sampling bias of the optimization design points by comparing their empirical distribution to a uniform distribution via Maximum Mean Discrepancy (MMD) [Gretton et al., 2012, Molnar et al., 2020], which is covered in more detail in the Appendix C.1. We measure the reliability of a PDP, i.e., the degree to which a user can rely on the estimate of the PD estimate, by comparing it to the true PD c S(λS) as defined in Eq. (1). More specifically, for every grid point λ(g) S , we compute the negative log-likelihood (NLL) of c S(λS) under the distribution of ˆc S (λS) pointwise for every grid point λ(g) S . The confidence of a PDP is illustrated by the width of its confidence bands ˆm S (λS) q1 /2 ˆs S (λS), with q1 /2 denoting the (1 /2)-quantile of a standard normal distribution. We measure the confidence by assessing ˆs S(λS) pointwise for every grid point. In particular, we consider the mean confidence (MC) across all grid points 1 the confidence at the grid point closest to ˆλS abbreviated by OC, with ˆλ being the best configuration evaluated by the optimizer. To evaluate the performance of the confidence splitting, we report the above metrics on the sub-region that contains the best configuration evaluated by the optimizer, assuming that this region is of particular interest for a user of HPO. PDPs are computed with regards to single features for G = 20 equidistant grid points and n = 1000 Monte Carlo samples. 6.1 BO on a Synthetic Function We consider the d-dimensional Styblinski-Tang function c : [ 5, 5]d ! R, λ 7! 1 2 for d 2 {3, 5, 8}. Since the PD is the same for each dimension i, we only present the effects of λ1. We performed BO with a GP surrogate model with a Matérn-3/2 kernel and the LCB acquisition function a(λ) = ˆm(λ)+ ˆs(λ) with different values 2 {0.1, 1, 5} to control the sampling bias. We compute the global PDP with confidence bands estimated according to Eq. (5) for the GP surrogate model ˆc that was fitted in the last iteration of BO. We ran Algorithm 1, and computed the PDP in the sub-region containing the optimal configuration. All computations were repeated 30 times. Further details on the setup are given in Appendix C.2.1. Figure 6: The figure presents the MC (left) and the NLL (right) for d 2 {3, 5, 8} for a high ( = 0.1), medium ( = 1), and low ( = 5) sampling bias across 30 replications. With a lower sampling bias, we obtain narrower confidence bands and a lower NLL. Table 1: The table shows the relative improvement of the MC and the NLL via Algorithm 1 with 1 and 3 splits, compared to the global PDP along with the sampling bias for a = 0.1 (high), = 2 (medium), and = 5 (low). Results are averaged across 30 replications. δ MC (%) δ NLL (%) d MMD nsp = 1 nsp = 3 nsp = 1 nsp = 3 3 low (0.18) 7.65 13.64 5.89 10.92 3 medium (0.51) 12.86 36.92 4.78 7.70 3 high (0.56) 16.52 34.84 2.77 -1.62 5 low (0.15) 6.63 15.45 2.82 6.05 5 medium (0.45) 19.67 37.28 4.05 7.80 5 high (0.53) 11.99 33.06 -3.86 -1.93 8 low (0.11) 3.58 9.67 0.84 2.40 8 medium (0.42) 8.86 23.03 1.51 3.30 8 high (0.56) 6.59 19.84 1.53 4.29 As presented in Figure 6, the PDPs for surrogate models trained on less biased data (measured by the MMD) yield lower values of the NLL, as well as lower values for the MC. Table 1 shows that a single tree-based split reduces the MC by up to almost 20%, and up to 37% when performing 3 partitioning steps. Additionally, the NLL improves with an increasing number of partitioning steps in most cases. The results on the synthetic functions support our second hypothesis that the tree-based partitioning improves the reliability in terms of the NLL and the confidence of the PD estimates. The improvement of the MC is higher for a medium to high sampling bias, compared to scenarios that are less affected by sampling bias. We observe that (particularly for high sampling bias) there are some outlier cases in which the NLL worsens. More detailed results are shown in Appendix C.3.1. 6.2 HPO on Deep Learning In a second experiment, we investigate HPO in the context of a surrogate benchmark [Eggensperger et al., 2015] based on the LCBench data [Zimmer et al., 2021]. For each of the 35 different Open ML [Vanschoren et al., 2013] classification tasks, LCBench provides access to evaluations of a deep neural network on 2000 configurations randomly drawn from the configuration space defined by Auto-Py Torch Tabular (see Table 5 in Appendix C.2). For each task, we trained a random forest as an empirical performance model that predicts the balanced validation error of the neural network for a given configuration. These empirical performance models serve as cheap to evaluate objective functions, which efficiently approximate the result of the real-world experiment of running a deep learning configuration on an LCBench instance. BO then acts on this empirical performance model as its objective10. For each task, we ran BO to obtain the optimal architecture and hyperparameter configuration. Again, we used a GP with a Matérn-3/2 kernel and LCB with = 1. Each BO run was allotted a budget of 200 objective function evaluations. We computed the PDPs and their confidences, which are estimated according to Eq. (5), based on the surrogate model ˆc after the final iteration. We performed tree-based partitioning with up to 6 splits based on a uniformly distributed dataset of size n = 1000. All computations were statistically repeated 30 times. Further details are provided in Appendix C.2.2. For the real-world data example, we focus on answering the second hypothesis, i.e., whether the treebased Algorithm 1 improves the reliability of the PD estimates. We compare the PDP in sub-regions after 6 splits with the global PDP. We computed the relative improvement of the confidence (MC and OC) and the NLL of the sub-regional PDP compared to the respective estimates for the global PDP. As shown in Table 2, the MC of the PDPs is on average reduced by 30% to 52%, depending on the hyperparameter. At the optimal configuration ˆλS, the improvement even increases to 50% 62%. Thus, PDP estimates for all hyperparameters are on average independent of the underlying dataset clearly more confident in the relevant sub-regions when compared to the global PD estimates, especially around the optimal configuration ˆλS. In addition to the MC, the NLL simultaneously improves. In Appendix C.3.2, we provide details regarding the evaluated metrics on the level of the dataset and demonstrate that our split criterion outperforms other impurity measures regarding MC 10Please note that the random forest is only used as a surrogate in order to construct an efficient benchmark objective, and not as a surrogate in the BO algorithm, where we use a GP. and OC. Furthermore, we emphasize in Appendix C.3.2 the significance of our results by providing a comparison to a naive baseline method. Figure 7: PDP (blue) and confidence band (grey) of the GP for hyperparameter max. number of units (batch size) on the left (right) side. The black line shows the PDP of the meta surrogate model representing the true PDP estimate. The orange vertical line marks the optimal configuration ˆλS. The relative improvements from the global PDP to the sub-regional PDP after 6 splits are for max. number of units (batch size): δ MC = 61.6% (28.4%), δ OC = 63.5% (62.2%), δ NLL = 48.6% (30.1%). Table 2: Relative improvement of MC, OC, and NLL on hyperparameter level. The table shows the respective mean (standard deviation) of the average relative improvement over 30 replications for each dataset and 6 splits. Hyperparameter δ MC (%) δ OC (%) δ NLL (%) Batch size 40.8 (14.9) 61.9 (13.5) 19.8 (19.5) Learning rate 50.2 (13.7) 57.6 (14.4) 17.9 (20.5) Max. dropout 49.7 (15.4) 62.4 (11.9) 17.4 (18.2) Max. units 51.1 (15.2) 58.6 (12.7) 24.6 (22.0) Momentum 51.7 (14.5) 58.3 (12.7) 19.7 (21.7) Number of layers 30.6 (16.4) 50.9 (16.6) 13.8 (32.5) Weight decay 36.3 (22.6) 61.0 (13.1) 11.9 (19.7) To further study our suggested method, we now highlight a few individual experiments. We chose one iteration of the shuttle dataset. On the two left plots of Figure 7, we see that the true PDP estimate for max. number of units is decreasing, while the globally estimated PDP trend is increasing and thus misleading. Although the confidence band already indicates that the PDP cannot be reliably interpreted on the entire hyperparameter space, it remains challenging to draw any conclusions from it. After performing 6 splits, we receive a confident and reliable PD estimate on an interpretable sub-region. The same plots are depicted for the hyperparameter batch size on the right part of Figure 7. This example illustrates that the confidence band might not always shrink uniformly over the entire range of λS during the partitioning, but often particularly around the optimal configuration ˆλS. 7 Discussion and Conclusion In this paper, we showed that partial dependence estimates for surrogate models fitted on experimental data generated by efficient hyperparameter optimization can be unreliable due to an underlying sampling bias. We extended PDPs by an uncertainty estimate to provide users with more information regarding the reliability of the mean estimator. Furthermore, we introduced a tree-based partitioning approach for PDPs, where we leverage the uncertainty estimator to decompose the hyperparameter space into interpretable, disjoint sub-regions. We showed with two experimental studies that we generate, on average, more confident and more reliable regional PDP estimates in the sub-region containing the optimal configuration compared to the global PDP. One of the main limitations of PDPs is that they bear the risk of providing misleading results if applied to correlated data in the presence of interactions, especially for nonparametric models [Grömping, 2020]. However, existing alternatives that visualize the global marginal effect of a feature such as accumulated local effect (ALE) plots [Apley and Zhu, 2020] do also not provide a fully satisfying solution to this problem [Grömping, 2020]. As a solution to this problem, Grömping [2020] suggests stratified PDPs by conditioning on a correlated and potentially interacting feature to group ICE curves. This idea is in the spirit of our introduced tree-based partitioning algorithm. However, in the context of BO we might assume the distribution in Eq. (1) to be uniform and therefore no correlations are present. Instead of correlated features, we are faced with a sampling bias (see Section 3) where we observe regions of varying uncertainty. Hence, instead of stratifying with respect to correlated features and aggregating ICE curves in regions with less correlated features, we stratify with respect to uncertainty and aggregate ICE curves in regions with low uncertainty variation. Nonetheless, it might be interesting to compare our approach with approaches based on the considerations made by Grömping [2020] or potentially improved ALE curves. Another limitation when using single-feature PDPs as in our examples is that hyperparameter interactions are not visible. However, two-way interactions can be visualized by plotting twodimensional PDPs within sub-regions. Another possibility to detect interactions is to look at ICE curves within the sub-regions. If the shape of ICE curves within a sub-region is very heterogeneous, it indicates that the hyperparameter under consideration interacts with one of the other hyperparameters. Hence, having the additional possibility to look at ICE curves of individual observations within a sub-region is an advantage compared to other global feature effect plots such as ALE plots [Apley and Zhu, 2020], as they are not defined on an observational level. While we mainly discussed GP surrogate models on a numerical hyperparameter space in our examples, our methods are applicable to a wide variety of distributional regression models and also for mixed and hierarchical hyperparameter spaces. We also considered in Appendix C.3.2 different impurity measures. While the one introduced in this paper performed best in our experimental settings, this impurity measure as well as other components are exchangeable within the proposed algorithm. In the future, we will study our method on more complex, hierarchical configuration spaces for neural architecture search. The proposed interpretation method is based on a surrogate and consequently does provide insights about what the Auto ML system has learned, which in turn allows plausibility checks and may increase trust in the system. To what extent this allows conclusions on the true underlying hyperparameter effects depends on the quality of the surrogate. How to efficiently perform model diagnostics to ensure a high surrogate quality before applying interpretability techniques is subject to future research. While we focused on providing better explanations without generating any additional experimental data, it might be interesting to investigate in future work how confidence and reliability of IML methods can be increased most efficiently when a user is allowed to conduct additional experiments. Overall, we believe that increasing interpretability of Auto ML will pave the way for human-centered Auto ML. Our vision is that users will be able to better understand the reasoning and the sampling process of Auto ML systems and thus can either trust and accept the results of the Auto ML system or interact with it in a feedback loop based on the gained insights and their preferences. How users can then best interact with Auto ML (beyond simple changes of the configuration space) will be left open for future research. Acknowledgments and Disclosure of Funding This work has been partially supported by the German Federal Ministry of Education and Research (BMBF) under Grant No. 01IS18036A. The authors of this work take full responsibilities for its content. D. W. Apley and J. Zhu. Visualizing the effects of predictor variables in black box supervised learning models. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 82(4):1059 1086, 2020. B. Bischl, J. Richter, J. Bossek, D. Horn, J. Thomas, and M. Lang. mlrmbo: A modular framework for model-based optimization of expensive black-box functions, 2018. L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone. Classification and Regression Trees. Wadsworth, 1984a. ISBN 0-534-98053-8. L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone. Classification and Regression Trees. Wadsworth, M. Britton. VINE: Visualizing statistical interactions in black box models. Co RR, abs/1904.00561, 2019. G. Cafri and B. A. Bailey. Understanding variable effects from black box prediction: Quantifying effects in tree ensembles using partial dependence. Journal of Data Science, 14(1):67 95, 2016. D. R. Cutler, T. C. Edwards Jr, K. H. Beard, A. Cutler, K. T. Hess, J. Gibson, and J. J. Lawler. Random forests for classification in ecology. Ecology, 88(11):2783 2792, 2007. J. Drozdal, J. Weisz, D. Wang, G. Dass, B. Yao, C. Zhao, M. Muller, L. Ju, and H. Su. Trust in Auto ML: Exploring information needs for establishing trust in automated machine learning systems. In Proceedings of the 25th International Conference on Intelligent User Interfaces, pages 297 307, 2020. K. Eggensperger, F. Hutter, H. Hoos, and K. Leyton-Brown. Efficient benchmarking of hyperparameter optimizers via surrogates. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 29, pages 1114 1120, 2015. W. D. Fisher. On grouping for maximum homogeneity. Journal of the American statistical Association, 53(284): 789 798, 1958. A. A. Freitas. Automated machine learning for studying the trade-off between predictive accuracy and inter- pretability. In Third IFIP International Cross-Domain Conference for Machine Learning and Knowledge Extraction (CD-MAKE 2019), volume 11713, pages 48 66. Springer, August 2019. J. H. Friedman. Greedy function approximation: A gradient boosting machine. Annals of Statistics, pages 1189 1232, 2001. J. H. Friedman and J. J. Meulman. Multiple additive regression trees with application in epidemiology. Statistics in medicine, 22(9):1365 1381, 2003. A. Goldstein, A. Kapelner, J. Bleich, and E. Pitkin. Peeking inside the black box: Visualizing statistical learning with plots of individual conditional expectation. Journal of Computational and Graphical Statistics, 24(1): 44 65, 2015. B. M. Greenwell. pdp: An R Package for Constructing Partial Dependence Plots. The R Journal, 9(1):421 436, B. M. Greenwell, B. C. Boehmke, and A. J. Mc Carthy. A simple and effective model-based variable importance measure. ar Xiv preprint ar Xiv:1805.04755, 2018. A. Gretton, K. M. Borgwardt, M. J. Rasch, B. Schölkopf, and A. Smola. A kernel two-sample test. Journal of Machine Learning Research, 13(25):723 773, 2012. U. Grömping. Model-agnostic effects plots for interpreting machine learning models. Report 1/2020, Reports in Mathematics, Physics and Chemistry. Department II, Beuth University of Applied Sciences Berlin, 2020. F. Hutter, H. H. Hoos, and K. Leyton-Brown. An efficient approach for assessing hyperparameter importance. In Proceedings of the 31th International Conference on Machine Learning, ICML, volume 32, pages 754 762. JMLR.org, 2014a. F. Hutter, L. Xu, H. H. Hoos, and K. Leyton-Brown. Algorithm runtime prediction: Methods & evaluation. Artificial Intelligence, 206:79 111, 2014b. J. Levesque, A. Durand, C. Gagné, and R. Sabourin. Bayesian optimization for conditional hyperparameter spaces. In 2017 International Joint Conference on Neural Networks, IJCNN 2017, Anchorage, AK, USA, May 14-19, 2017, pages 286 293. IEEE, 2017. doi: 10.1109/IJCNN.2017.7965867. URL https://doi.org/10. 1109/IJCNN.2017.7965867. C. Molnar, G. König, B. Bischl, and G. Casalicchio. Model-agnostic feature importance and effects with dependent features - A conditional subgroup approach. Co RR, abs/2006.04628, 2020. R. M. Neil. Bayesian Learning for Neural Networks. Springer-Verlag, Berlin, Heidelberg, 1996. ISBN 0387947248. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825 2830, 2011. F. Pfisterer, J. Thomas, and B. Bischl. Towards human centered Auto ML. Co RR, abs/1911.02391, 2019. P. Probst, A. Boulesteix, and B. Bischl. Tunability: Importance of hyperparameters of machine learning algorithms. Journal of Machine Learning Research, 20:53:1 53:32, 2019. A. Sharma, J. N. van Rijn, F. Hutter, and A. Müller. Hyperparameter importance for image classification by residual neural networks. In Discovery Science - 22nd International Conference, DS, volume 11828 of Lecture Notes in Computer Science, pages 112 126. Springer, 2019. J. Snoek, H. Larochelle, and R. P. Adams. Practical Bayesian optimization of machine learning algorithms. In Advances in Neural Information Processing Systems 25, pages 2960 2968, 2012. K. Swersky, D. Duvenaud, J. Snoek, F. Hutter, and M. A. Osborne. Raiders of the lost architecture: Kernels for bayesian optimization in conditional parameter spaces. ar Xiv: Machine Learning, 2014. J. Vanschoren, J. N. van Rijn, B. Bischl, and L. Torgo. Openml: networked science in machine learning. SIGKDD Explor., 15(2):49 60, 2013. M. N. Wright and I. R. König. Splitting on categorical predictors in random forests. Peer J, 7:e6339, 2019. I. Xanthopoulos, I. Tsamardinos, V. Christophides, E. Simon, and A. Salinger. Putting the human back in the Auto ML loop. In Proceedings of the Workshops of the EDBT/ICDT 2020 Joint Conference, volume 2578 of CEUR Workshop Proceedings. CEUR-WS.org, 2020. L. Zimmer, M. Lindauer, and F. Hutter. Auto-Py Torch Tabular: Multi-fidelity metalearning for efficient and robust Auto DL. IEEE TPAMI, 2021. Preprint via Early Access.