# divbo_diversityaware_cash_for_ensemble_learning__ddf5de80.pdf Div BO: Diversity-aware CASH for Ensemble Learning Yu Shen1, Yupeng Lu1, Yang Li4, Yaofeng Tu3, Wentao Zhang56, Bin Cui12 1Key Lab of High Confidence Software Technologies, Peking University, China 2Institute of Computational Social Science, Peking University (Qingdao), China 3 ZTE Corporation, China 4 Data Platform, TEG, Tencent Inc., China 5 Mila - Québec AI Institute 6HEC, Montréal, Canada 1{shenyu, xinkelyp, bin.cui}@pku.edu.cn, 2 tu.yaofeng@zte.com.cn 3thomasyngli@tencent.com 4wentao.zhang@mila.quebec The Combined Algorithm Selection and Hyperparameters optimization (CASH) problem is one of the fundamental problems in Automated Machine Learning (Auto ML). Motivated by the success of ensemble learning, recent Auto ML systems build post-hoc ensembles to output the final predictions instead of using the best single learner. However, while most CASH methods focus on searching for a single learner with the best performance, they neglect the diversity among base learners (i.e., they may suggest similar configurations to previously evaluated ones), which is also a crucial consideration when building an ensemble. To tackle this issue and further enhance the ensemble performance, we propose Div BO, a diversity-aware framework to inject explicit search of diversity into the CASH problems. In the framework, we propose to use a diversity surrogate to predict the pair-wise diversity of two unseen configurations. Furthermore, we introduce a temporary pool and a weighted acquisition function to guide the search of both performance and diversity based on Bayesian optimization. Empirical results on 15 public datasets show that Div BO achieves the best average ranks (1.82 and 1.73) on both validation and test errors among 10 compared methods, including post-hoc designs in recent Auto ML systems and state-of-the-art baselines for ensemble learning on CASH problems. 1 Introduction In recent years, machine learning has made great strides in various application areas, e.g., computer vision [14, 12], recommendation [36, 37], etc. However, it s often knowledge-intensive to develop customized solutions with promising performance, as the process includes selecting proper ML algorithms and tuning the hyperparameters. To reduce the barrier and facilitate the deployment of machine learning applications, the Auto ML community raises the Combined Algorithm Selection and Hyperparameters optimization (CASH) problem [39] and proposes several methods [33, 18, 15] to automate the optimization. While most Auto ML methods [35, 17, 2] for CASH focus on searching for the optimal performance of a single learner, it s widely acknowledged that ensembles of promising learners often outperform single ones [44, 3]. For example, He et al. [14] won first place in ILSRVC 2015 with an average of several learners. And ensemble strategies can be frequently found in the top solutions of Kaggle competitions [16, 4]. Motivated by those achievements, recent Auto ML systems (e.g., Auto-sklearn [11], Auto-Pytorch [50], Volcano ML [25]) build post-hoc ensembles based on all base learners from the entire optimization process and show better empirical results than using the best single learner. Despite the effectiveness of those post-hoc ensemble designs, the target of CASH methods is inconsistent with that of ensemble learning. In other words, a good ensemble should contain a pool of 36th Conference on Neural Information Processing Systems (Neur IPS 2022). base learners that are both well-performing and diverse with each other [43, 44], while most CASH methods only aim at searching for the best-performing configuration of learners. Figure 1 shows the diversity of the suggested learners from Auto-sklearn and Volcano ML after 200 iterations. We define the pair-wise diversity of two learners as the average regularized Euclidean distance of their probabilistic predictions which ranges from 0 to 1 (See Equation 3). And we show the minimum diversity of a suggested learner with other learners from the ensemble built at the 200th iteration. While most diversity values of the two Auto ML systems are around 0.05, they suggest learners whose predictions are often quite similar to one of the learners obtained from previous iterations. It implies a lack of diversity in previous post-hoc ensemble designs, and it also indicates that the performance of post-hoc ensembles might be further improved if diversity is also taken into consideration when suggesting new configurations to evaluate. However, the search of diversity is non-trivial. A simple diversity-inducing algorithm would degenerate the pool of learners, e.g., diversity can be easily increased by predicting all samples for the classes that are wrong and different from previous learners. Therefore, how to guide the search for both performance and diversity simultaneously in recent CASH algorithms is still an open question. 210 220 230 240 250 Iteration Minimum Diversity Volcano ML Auto-sklearn Figure 1: Minimum diversity of suggested learners after 200 iterations on quake. (20 runs) In this paper, we propose Div BO, a new algorithm framework that combines Bayesian optimization (BO) with an explicit search of diversity for classification. The contributions are summarized as: 1. To the best of our knowledge, this work is the first to enhance the ensemble performance of CASH methods by considering the diversity of suggested configurations. 2. To inject the search of diversity into CASH problems, we design 1) the diversity surrogate to predict the pair-wise diversity given two unseen configurations and 2) a BO-based framework that automatically balances the predictive performance and diversity of suggestions. 3. Empirical results show that the diversity surrogate achieves a better correlation with the groundtruth results compared with the performance surrogate used in BO. Compared with recent Auto ML systems, Div BO suggests configurations that are significantly more diverse with those in previous iterations (the red line in Figure 1). In addition, results on 15 public datasets show that Div BO achieves the best average rank (1.82 and 1.73) on both validation and test errors among 10 methods, including post-hoc designs in recent Auto ML systems and state-of-the-art methods for ensemble learning on CASH problems. 2 Preliminary and Related Work In this section, we review the related work and introduce the preliminary for our proposed method. Combined algorithm selection and hyperparameters (CASH). We first introduce the basic notations for the CASH problem. Let A be the set of K candidate algorithms A = {A1, ..., AK}. Each algorithm Ai has its corresponding hyperparameter space ΛAi. Given dataset D = {Dtrain, Dval} of a learning problem, the CASH problem aims to minimize the validation metric L by searching for the best joint configuration x = (A , λ ) trained on the training set, where A is the best algorithm and λ is its corresponding hyperparameters. For brevity, we use T(a,λ) to denote the learner constructed by the joint configuration x = (a, λ) and trained on Dtrain. The optimization objective can be formulated as: min a A,λ Λa L(T(a,λ), Dval). (1) The CASH problem is first introduced by Auto-WEKA [39] and solved by applying Bayesian optimization (BO) [35, 17, 2] on the entire search space. Each BO iteration follows three steps: 1) BO fits a probabilistic surrogate model based on observations D = {(x1, y1), ..., (xn 1, yn 1)}, in which xi is the ith evaluated joint configuration, and yi is its corresponding observed performance; 2) BO selects the most promising configuration xn by maximizing the acquisition function to balance exploration and exploitation; 3) BO evaluates the configuration xn (i.e., train the learner and obtain its validation performance), and augment the observations D. To avoid confusion, we refer to the surrogate of BO as the performance surrogate Mperf in the following section. In addition to BO, TPOT [31] proposes to use genetic programming to solve the CASH problem. The ADMM-based method [29] decomposes the problem into sub-problems and solves them using ADMM [5]. Rising bandit [23] continuously eliminates badly performing algorithms during optimization. FLAML [41] suggests configurations based on estimated costs of improvement. Ensemble-oriented CASH. Different from the original CASH, ensemble-oriented CASH aims to find an ensemble of m learners that minimizes the validation metric L, which is: min a1,...,am A,λ1 Λa1,...,λm Λam L(Ensemble(T(a1,λ1), ..., T(am,λm)), Dval). (2) To solve the problem, ensemble optimization [22] directly optimizes the target based on BO by using an ensemble pool with a fixed size. This strategy suffers from instability due to the risk of adding a bad configuration during optimization. Neural ensemble search (NES) [43] is proposed based on the regularized evolutionary algorithm [34]. The local search design in NES works well on low-dimensional problems like neural architecture search, but may encounter a bottleneck when applied to solve high-dimensional CASH problems. Rather than directly optimizing on Equation 2, recent Auto ML systems (e.g., Auto-sklearn [11], Volcano ML [25, 26], Auto-Pytorch [50]) adopt a specific ensemble strategy on the observation history after solving Equation 1, which we refer to as the post-hoc ensemble designs. While the optimization target is inconsistent, as shown in Section 1, there s still space to improve the performance of post-hoc ensembles. Inspired by the theoretical study that diversity can help improve the performance of the ensemble [21, 3, 13, 48], we follow the post-hoc ensemble designs and focus on generating a promising pool of learners for the final ensemble strategy. Ensemble selection. Ensemble strategies are methods that combine the predictions given a pool of learners, which are orthogonal to the direction of our method (i.e., suggesting a diverse pool of learners). Among different ensemble strategies (e.g., Bagging [9], Boosting [9, 30], Stacking [6]), we adopt the ensemble selection [7], which works empirically well with Auto ML as shown in previous study [11, 25, 43]. In short, ensemble selection starts from an empty ensemble and iteratively adds models from the pool with replacement to maximize the ensemble validation performance (with uniform weights). The pseudo-code is provided in Appendix A.1. 3 Diversity-aware Bayesian optimization (Div BO) In this section, we present Div BO our proposed diversity-aware CASH method for ensemble learning based on Bayesian optimization (BO). To inject the search of diversity into BO, we will answer the following two questions: 1) how to measure diversity and model the diversity of two unseen configurations, and 2) how to suggest configurations that are both well-performing and diverse with potential ones in the final ensemble. 3.1 Diversity Surrogate The existing diversity measures can be generally divided into pair-wise and nonpair-wise measures. The nonpair-wise diversity [47] directly measures a set of learners in the ensemble. While the ensemble may change during each iteration, it s difficult to model the diversity of an ensemble with a candidate configuration and multiple learners. Therefore, we use the pair-wise measures to simply learn the diversity of two given configurations. To this end, we follow the definition in previous research [44] that explicitly improves the diversity of neural networks and shows satisfactory empirical results. Let xi denote as the joint configuration of algorithm ai and hyperparameters λi. The diversity between two configurations (xi, xj) is the average Euclidean distance of class probabilities predicted on the validation set [44], which is: Div(xi, xj) = 2 2 1 |Dval| s Dval ||Txi(s) Txj(s)||2, (3) where Dval is the validation set, Txi is the learner corresponding to configuration xi and fitted on the training set, and Txi(s) is the predictive class probability on sample s. Obviously, the relationship between the diversity and configuration pair (xi, xj) is also a black-box function. Therefore, inspired by Bayesian optimization, we apply another surrogate to model this relationship. This surrogate, namely the diversity surrogate, takes a configuration pair as input and outputs the predictive mean and variance of the pair-wise diversity. Fitting. During each iteration, Div BO generates the training set for the diversity surrogate Mdiv by computing the diversity value of each pair of the observed configurations. Note that, (xi, xj) and (xj, xi) lead to the same diversity value, but the training set should include both of them to ensure symmetry. This leads to about |D|2 training samples, where |D| is the number of observations. To avoid fetching validation predictions by retraining, we store those predictions for all observations during optimization, which is a common trick applied in previous methods [11, 22, 43]. Implementation. Concretely, the diversity surrogate Mdiv in Div BO is an ensemble of several Light GBM models [19]. This implementation has the following two advantages: 1) The diversity surrogate gives predictions with uncertainty to balance exploration and exploitation. Concretely, the predictive mean and variance of the surrogate is obtained by computing the mean and variance of outputs generated by different Light GBM models. 2) The time complexity of fitting the surrogate is relatively low. The cost of training a Light GBM model is O(|D|2 log |D|), which is lower than O(|D|3) when fitting a Gaussian Process as the performance surrogate. In Section 4, we will further compare different implementations of diversity surrogates on their ability to fit the relationship. 3.2 Diversity-aware Framework Based on the diversity surrogate, we propose a diversity-aware framework to suggest configurations that are both well-performing and diverse. Before stepping into the design, consider a straightforward strategy that we choose the most diverse configuration with the observation history during each iteration. This simple strategy has two obvious drawbacks: 1) First, there is no need to suggest configurations that are diverse from badly-performing ones in the observation history we only expect the promising base learners in the final ensemble to be diverse with each other; 2) The diversity can be easily increased by suggesting a learner that predicts all samples for the classes that are wrong and different from previous learners. In this extreme case, the learner can not help improve the performance of the final ensemble. Therefore, how to control the diversity of the suggested configurations is non-trivial. In the following, we explain how Div BO tackles the two drawbacks. Suggesting diverse configurations. To tackle the first drawback, since which learners will appear in the final ensemble is unknown during optimization, Div BO proposes to use a temporary pool P to collect base learners that will probably be selected into the final ensemble. Concretely, the pool is built by applying ensemble selection to the observation history. When the optimization ends, this temporary pool is exactly the output of a post-hoc ensemble strategy (e.g., auto-sklearn[11]) that ends at the previous iteration, and thus the configurations in the temporary pool can be regarded as the potential ones in the final ensemble. Then, we define the diversity acquisition function αdiv of each unseen configuration x as its predicted diversity value with the most similar configuration in the temporary pool: αdiv(x) = 1 n=1 min θ P M n div(θ, x), (4) where P is the temporary configuration pool and Mdiv is the diversity surrogate. M n div(θ, x) is the nth sampled value from the output distribution of Mdiv given the pair (θ, x), and the final acquisition function is the average of N minimums via sampling. Note that, using the minimum diversity with learners in the temporary pool is more appropriate than the mean diversity. In extreme cases, when the algorithm may suggest a useless configuration that is exactly the same as a previously evaluated one, the minimum diversity is penalized to zero while the mean diversity is still larger than zero. By maximizing Equation 4, Div BO is able to suggest configurations that are diverse from other potential ones in the final ensemble during optimization. Combining predictive performance and diversity. As mentioned in the second drawback, optimizing diversity alone degenerates the predictive performance of the suggested configurations. Div BO tackles this issue by combining both the performance and diversity acquisition functions with a saturating weight. Since the performance and diversity acquisition values are of different scales, we propose to use the sum of ranking values instead of directly adding the output values. During each Algorithm 1: Algorithm framework of Div BO. Input: The search budget B, the architecture search space X, the ensemble size E, the training and validation set Dtrain, Dval. 1 Initialize observations as D = ; 2 while budget B does not exhaust do 3 if |D| < 5 then 4 Suggest a random configuration x X; 6 Fit performance surrogate Mperf and diversity surrogate Mdiv based on observations D; 7 Build a temporary pool of configurations as P = {θ1, ..., θE} = Ensemble Selection(D, Dval, E); 8 Compute the ranks of sampled configurations Rperf and Rdiv based on the performance and diversity surrogates Mperf, Mdiv and the temporary pool P; 9 Suggest a configuration x = arg min x X α(x) based on Equation 5; 10 Build and train the learner f x on Dtrain and evaluate its performance on Dval as y; 11 Update the observations D = D {( x, y)}; 12 Generate a pool of base configurations {θ1, ..., θE} = Ensemble Selection(D, Dval, E); 13 return the final ensemble Ensemble(Tθ1, ..., TθE). BO iteration, we sample sufficient configurations by random sampling from the entire space and local sampling on well-performed observed configurations [17]. Then we calculate the performance and diversity acquisition value for each sampled configuration. Based on these values, we further rank the sampled configurations and obtain the ranking value of xi as Rperf(xi) and Rdiv(xi) for the performance and diversity acquisition function, respectively. Finally, given a configuration xi, we define the weighted acquisition function α for Div BO as: α(xi) = Rperf(xi) + w Rdiv(xi), w = β(sigmoid(τt) 0.5), (5) where w is the weight for the diversity acquisition, t is the current number of BO iterations, β and τ are two hyperparameters. To match the intuition that the optimization should focus on performance in the beginning and gradually shift its attention to diversity, Div BO applies a saturating weight, where w ranges from [0, β) and τ controls the speed of approaching saturation. The goal of Div BO is to suggest configurations that minimize the function in Equation 5. Algorithm 1 illustrates the procedure of Div BO. During each iteration after initialization, Div BO 1) fits the performance and diversity surrogates based on observations (Line 6); 2) builds a temporary configuration pool by applying ensemble selection on the observation history (Line 7); 3) samples candidate configurations and compute their ranking values (Line 8); 4) suggests a configuration that minimizes the combined ranking value in Equation 5 (Line 9); 5) evaluates the suggested configuration on the validation set and then updates the observations (Lines 10-11). 3.3 Discussion In this section, we provide the discussion on Div BO as follows: Time complexity. As mentioned in Section 3.1, the time complexity of fitting the diversity surrogate is O(|D|2 log |D|), and the complexity of building a temporary pool is O(|Dval||D|), where |Dval| is the number of validation data samples, and |D| is the number of observations. For large datasets, we can prepare a constant subset of validation samples for ensemble selection. Therefore, the time complexity depends on the choice of performance surrogate in BO. Concretely, the complexity of Div BO for each iteration is O(|D|3) when using the Gaussian Process [35] and O(|D|2 log |D|) when using the probabilistic random forest [17]. Extension. As an abstract algorithm framework, the components of Div BO can be replaced to meet different requirements. Though this paper focuses on CASH problems for classification, Div BO can also be applied to regression problems by defining a new diversity function based on regression predictions instead of Equation 3. In addition, Div BO is independent of the choice of performance surrogate for Bayesian optimization, i.e., the performance surrogate can be replaced with state-of-the-art ones proposed for specific scenarios. Foundation. Like the diversity-driven methods in other scenarios [44, 32], Div BO does not target at optimizing Equation 2 directly. The foundation of its effectiveness lies in the claim that it is beneficial for ensembles to not only have promising base learners, but also more diversity in their predictions. The claim has been studied by extensive theoretical work [21, 3, 13, 48, 20], and please refer to previous work for more details. In the following section, we will empirically show that the ensemble generated by Div BO outperforms the state-of-the-art methods in real-world CASH problems. Limitations. The use of ensemble leads to higher inference latency than using the single best learner (approximately K times where K is the number of learners in the ensemble). This latency can be reduced with the aid of parallel computing if we have sufficient computational resources; As ensemble selection is fitted on the validation set, there s a risk of overfitting on the test set for small datasets; Div BO using Equation 3 as the diversity function can not directly support algorithms that only predict class labels (e.g., SVC). Though Div BO still works by converting the outputs to class probability (like [1, 0, ...]), other diversity functions can be developed to support those algorithms better. 4 Experiments In this section, we evaluate our proposed method on real-world CASH problems using public datasets. In the following, we list three main insights that we will investigate: 1) The diversity surrogate in Div BO can predict the diversity value of two unseen configurations well. 2) The Div BO framework outperforms the post-hoc designs used in recent Auto ML systems and other competitive baselines for ensemble learning, in terms of both validation and test performance. 3) The base learners in the ensemble given by Div BO show similar average performance but enjoy higher diversity than those from other post-hoc designs. 4.1 Experiment Setup Baselines. We compare the proposed Div BO with the following eight baselines Three CASH methods: 1) Random search (RS) [1]; 2) Bayesian optimization (BO) [17]; 3) Rising Bandit (RB) [23]; Two Auto ML methods proposed for ensemble learning: 4) Ensemble optimization (EO) [22]; 5) Neural ensemble search (NES) [43]; Three post-hoc designs: 6) Random search with post-hoc ensemble (RS-ES); 7) Bayesian optimization with post-hoc ensemble (BO-ES) [11]: the default strategy in Auto-sklearn; 8) Rising bandit with post-hoc ensemble (RB-ES) [25]: the default strategy in Volcano ML. While for Div BO, we also implement the variant without post-hoc ensemble, which we denote as Div BO- . Datasets and search space. While recent Auto ML systems differ in both search space and algorithm, to make a fair comparison of algorithms, we conduct the experiments on the same search space. Concretely, we slightly modify the search space of the Auto ML system Volcano ML [25]. The search space contains 100 hyperparameters in total, and the details of algorithms and feature engineering hyperparameters are provided in Appendix A.3. In addition, we use 15 public classification datasets that are collected from Open ML [40], whose number of samples ranges from 2k to 20k. More details about the datasets are provided in Appendix A.2. Basic settings. Each dataset is split into three sets, which are the training (60%), validation (20%), and test (20%) sets. To evaluate the diversity surrogate, we use the Kendall-tau rank correlation as the metric since we only care about the ranking relationship between two pairs during optimization. For comparison with other baselines on CASH problems, we report the best-observed validation error during optimization and the final test error. While it takes a different amount of time to evaluate the same configuration on different datasets, we use the evaluation iterations as the unit of budget. Following Volcano ML [25] where each baseline evaluates approximately 250 configurations, we set the number of maximum iterations to 250. The evaluation of each method on each dataset is repeated 10 times, and we report the mean std. result by default. The hyperparameters β and τ are set to 0.05 and 0.2 in Div BO. We provide more implementation details for other baselines and the sensitivity analysis in Appendix A.4 and A.5, respectively. 50 100 150 200 250 Number of Observations Kendall Tau 50 100 150 200 250 Number of Observations Kendall Tau Figure 2: Kendall Tau correlation of different diversity surrogates with standard deviations shaded. -p refers to the performance surrogate while the other three are diversity surrogates. 0 50 100 150 200 250 Number of Observations Average Validation Rank RS BO RB Div BOEO NES RS-ES BO-ES RB-ES Div BO Figure 3: Average validation rank of 10 methods during optimization across 15 datasets. Note that ranks are a relative measure of performance (the rank of all methods add up to 55), and an improvement in validation error in one method may influence the rank of another one. 4.2 Evaluation of Diversity Surrogate To show the soundness of Div BO, we first provide an analysis of the fitting capability of different diversity surrogates during optimization. We take Light GBM (LGB) [19], XGBoost (XGB) [8], and probabilistic random forest (PRF) as the candidates for diversity surrogate, where all three tree-based alternatives share the same time complexity O(|D|2 log(|D|)). We evaluate 300 randomly chosen configurations, among which up to 250 configurations are used to fit the surrogate, and the left 50 are used for surrogate evaluation. Figure 2 shows the Kendall Tau correlation between the surrogate predictive means across different runs and the ground-truth results on two different datasets. We observe that, in general, the correlation of the diversity surrogate improves as the observations increase. Among the three alternatives, XGBoost performs slightly better than PRF, and Light GBM performs better than the other two alternatives. Remarkably, Light GBM achieves a strong Kendall Tau correlation of 0.65 and 0.58 on quake and wind when fitted with 250 observations, respectively. We also evaluate the performance surrogate (PRF-p) used in BO. Its correlation over the number of observations is much lower than the diversity surrogate, and the correlation is only 0.26 and 0.19 on quake and wind when fitted with 250 observations. The reason is that the search space for CASH problems is too large, and the performance surrogate can not be fitted well using limited observations D. While the number of pair-wise samples is much larger than non-pair-wise samples (i.e., |D|2 vs. |D|), the diversity surrogate captures more information from the observations. As a result, the diversity surrogate can fit the diversity relationship between two configurations well and enjoys a 0 50 100 150 200 250 Number of Observations Best Observed Val Error 0 50 100 150 200 250 Number of Observations Best Observed Val Error 0 50 100 150 200 250 Number of Observations Average Test Rank (c) Average test rank EO NES RS-ES BO-ES RB-ES Div BO Figure 4: Figures (a) and (b): Best observed validation error with standard deviations shaded on two datasets. Figure (c): Average test rank of post-hoc ensemble designs across 15 datasets. relatively strong correlation given a limited budget. In the following, we apply Light GBM as the diversity surrogate and evaluate Div BO on real-world Auto ML tasks. 4.3 Evaluation of Div BO Performance Analysis. In this section, we evaluate Div BO on 15 real-world CASH problems. Figure 3 shows the rank of validation error of all compared methods during optimization. We get four observations from the figure: 1) Ensemble learning indeed helps improve the performance of CASH results. Though Rising Bandit (RB) is competitive, the rank of non-ensemble methods (RS, BO, Div BO-) at the 250th iteration is only 8.72, 7.21, and 8.22 among 10 compared methods, respectively; 2) Among methods with ensemble learning, the post-hoc designs outperform the other designs for ensemble learning (NES, EO). Note that, random search with ensemble selection (RS-ES) is a strong baseline. It achieves a rank of 4.30 at the 250th iteration, which is better than that of NES (5.82) and EO (6.65). The reason is that the evolutionary algorithm in NES is not suitable for large search space while the performance of EO fluctuates if a poor learner is added into its fixed ensemble during optimization; 3) Among post-hoc designs, Div BO outperforms the other ones. While the second-best baseline achieves a rank of 2.58 at the 250th iteration, the rank of Div BO is 1.82, and it consistently outperforms the other baselines after about 100 iterations. 4) We also find that Div BOperforms worse than BO without ensemble learning. The reason is that, the consideration of diversity is useless when searching for the single best classifier. But with the aid of the performance surrogate, Div BOstill performs better than random search. To demonstrate the validation error on specific datasets, we also plot the best-observed validation errors during optimization for six methods with ensemble learning in Figures 4(a) and 4(b). Similar to the trend of rank, Div BO exceeds the second-best baseline RB-ES after about 100 iterations. We further compare the number of required iterations for Div BO to achieve the same best validation errors as other methods. Concretely, Div BO achieves 1.52-1.54x and 1.67-2.53x speedups relative to RB-ES and BO-ES on the two datasets. We provide ablation study on weight scheduling and comparison with other intuitive designs in Appendix A.5. Table 1 demonstrates the test errors and average rank on 15 datasets. We observe that the rank value of Div BO on test error is sometimes not consistent with that on validation error. The reason is that the distributions of the validation and test set are not exactly the same [18]. In other words, the configuration with the best validation error may not be the one with the best test error. But overall, post-hoc ensemble designs outperform other methods for ensemble learning, among which Div BO achieves the best test error on 10 out of 15 datasets, and its average rank is 1.73. The second best baseline is RB-ES, which performs the best on 3 out of 15 datasets, and its average rank is 2.90. The test rank of 4 post-hoc ensemble designs during optimization is also presented in Figure 4(c). To check whether the improvement of Div BO is statistically significant, we conduct the Wilcoxon signed-rank test on each dataset given two methods. The difference is significant when the value p 0.05. We count the number of datasets if 1) the mean error of Div BO is lower, and the difference is statistically significant (B); 2) the difference is not statistically significant (S); and 3) the mean error of Div BO is higher, and the difference is statistically significant (W). The results are presented in Table 2. Though RB-ES is a strong baseline, we observe that Div BO performs no worse than RB-ES on 12 datasets and better on 8 datasets. In Appendix A.5, we provide additional experiments to study the effects of different ensemble strategies and ensemble sizes. Table 1: Test error (%) with standard deviations and the average rank across different datasets. Method amazon_employee bank32nh cpu_act cpu_small eeg elevators house_8L pol CASH Methods RS 5.32 0.08 18.43 0.79 6.17 0.45 7.57 0.41 6.58 0.52 10.22 0.68 11.32 0.26 1.65 0.12 BO 5.26 0.06 18.46 0.50 5.69 0.38 7.76 0.92 5.47 1.35 10.25 0.87 11.20 0.19 1.59 0.50 RB 5.27 0.08 18.33 0.32 5.70 0.31 7.53 0.35 4.70 1.05 9.77 0.20 11.12 0.15 1.39 0.03 Methods for Ensemble Learning EO 5.19 0.27 18.47 0.53 5.85 0.62 7.42 0.45 3.54 0.76 10.34 0.61 11.14 0.21 1.43 0.13 NES 5.32 0.38 18.21 0.32 5.59 0.20 7.28 0.84 2.68 0.73 9.75 0.30 11.28 0.48 1.74 0.35 Post-hoc Designs RS-ES 5.29 0.15 18.30 0.66 5.70 0.26 7.21 0.30 4.45 0.22 9.51 0.28 11.21 0.38 1.39 0.15 BO-ES 5.25 0.15 18.41 0.39 5.50 0.47 7.16 0.28 3.55 0.78 9.61 0.36 11.06 0.33 1.35 0.18 RB-ES 5.21 0.11 18.07 0.57 5.58 0.20 7.08 0.22 2.86 0.92 10.01 0.15 10.81 0.27 1.36 0.16 Div BO 5.16 0.09 18.35 0.32 5.37 0.23 7.04 0.29 3.26 0.84 9.40 0.28 10.80 0.22 1.34 0.17 Method pollen puma32H quake satimage spambase wind 2dplanes Average Rank CASH Methods RS 51.53 0.53 10.59 1.26 48.17 1.88 10.00 0.72 7.14 1.23 14.66 0.58 7.21 0.08 8.37 BO 49.64 3.39 10.42 0.81 46.88 2.16 9.21 0.99 6.45 0.84 14.52 0.60 7.15 0.06 6.47 RB 49.79 1.15 11.21 0.39 47.98 1.56 9.50 0.94 6.71 1.05 14.11 0.25 7.20 0.05 6.37 Methods for Ensemble Learning EO 49.01 2.10 9.74 1.60 46.88 1.48 9.42 1.11 6.41 0.68 14.62 0.48 7.13 0.07 5.77 NES 51.56 1.53 10.63 0.54 46.42 1.15 8.66 0.95 6.23 0.68 14.25 0.50 7.11 0.08 5.00 Post-hoc Designs RS-ES 49.69 1.63 10.58 0.73 46.79 1.57 9.35 0.73 6.45 0.23 14.34 0.47 7.11 0.12 5.27 BO-ES 48.91 1.75 9.27 1.13 46.10 2.52 9.10 0.87 6.38 0.64 14.04 0.53 7.07 0.08 3.13 RB-ES 49.58 1.34 7.85 0.43 46.70 1.34 8.55 1.36 6.12 0.36 13.98 0.45 7.20 0.08 2.90 Div BO 49.25 1.35 8.07 0.99 45.55 1.37 8.71 1.25 5.91 0.45 13.93 0.42 7.00 0.08 1.73 Table 2: Counts of datasets when Div BO performs statistically better (B), the same (S), and worse (W) than compared three baselines. (a) vs. RS-ES B S W Div BO 13 2 0 (b) vs. BO-ES B S W Div BO 12 1 2 (c) vs. RB-ES B S W Div BO 8 4 3 Diversity Analysis. Finally, we analyze the optimization process of BO-ES, RB-ES, and Div BO. In Table 3, we show the validation errors of learners during optimization. Without ensemble, the single learner suggested by Div BOperforms worse than BO and RB. Note that, this does not mean that Div BO suggests bad configurations. We randomly evaluate 300 configurations from the search space. The mean result of those diverse configurations is better than 88% of the random configurations. To show how the diversity during the search process affects the ensemble, we use the update times of the temporary pool as a metric and the results are shown in Table 4. Since the temporary pool is built in the same way as the final ensemble, if the temporary pool changes, the configuration suggested at the previous iteration is included in the pool. Therefore, a change of the temporary pool at least indicates the suggested configuration affects the current ensemble. However, though the pool changes, the performance may not be improved due to the greedy mechanism, and thus we count the effective update times (i.e., the pool changes and the validation error of the ensemble decreases). As the pool updates very frequently in the beginning, we only calculate the mean effective update times of Div BO, RB-ES, and BO-ES on all datasets during the last 50 and 100 iterations. The pool is relatively stable in the last 50 iterations, which also indicates a budget of 250 iterations is sufficient for the datasets. We observe that, on average, Div BO will improve the temporary pool at least once in the last 50 iterations. While the difference between BO-ES and Div BO is the diversity part, we attribute this frequency gain to the use of diversity during the search process. In Figure 5, given the total budget of 250 evaluation iterations, we plot the minimum diversity of the suggested learners (solid lines) after 200 iterations. The minimum diversity is defined as the diversity value (Equation 3) with the most similar learner in the ensemble built at the previous iteration. The minimum diversity of suggestions given by BO-ES and RB-ES is around 0.05, which indicates that the configuration suggestion is similar to one of the configurations in the previous ensemble. While for Div BO, the diversity is around 0.23, which is much higher than those of BO-ES and RB-ES. 0 2 4 6 8 10 12 14 16 18 0 2 4 6 8 10 12 14 16 18 (a) RB-ES Disagreement: 0.13 0 2 4 6 8 10 12 14 0 2 4 6 8 10 12 14 (b) BO-ES Disagreement: 0.09 0 2 4 6 8 10 12 14 16 0 2 4 6 8 10 12 14 16 (c) Div BO Disagreement: 0.27 Figure 6: Pair-wise diversity of learners in the post-hoc ensemble on quake. The learners in each method are numbered in order of observation. Table 3: Val errors (%) of single learners during optimization. BO-ES RB-ES Div BO Val Errors (%) 44.03 2.58 43.85 2.49 44.33 2.67 Table 4: Effective pool updates during optimization. BO-ES RB-ES Div BO Counts (last 100) 1.8 2.1 3.4 Counts (last 50) 0.6 0.8 1.5 210 220 230 240 250 Number of Observations Minimum Diversity RB-ES-d RB-ES-p BO-ES-d BO-ES-p Div BO-d Div BO-p Average Base Learner Errors Figure 5: Minimum diversity of suggested learners (solid) and average error (dash) of base learners in the ensemble on quake. In addition, we plot the average validation error of base learners in the current ensemble (dash lines) in Figure 5. The average error at the 250th iteration given by the three post-hoc methods are quite similar (around 44.60%), which shows that applying the diversity surrogate in Div BO will not degenerate the pool of base learners. We further plot the pair-wise diversity between unique learners in the final ensemble in Figure 6 and present the average pair-wise predictive disagreement [38] in the caption. Generally, the larger disagreement is, the more diverse the learners in the ensemble are. Since ensemble selection selects base learners with replacement, the number of unique learners in the ensemble is different in each independent run. Though the average performance of base learners in the three post-hoc methods are similar, for RB-ES and BO-ES, the base learners are quite similar to each other. While for Div BO, we find that the learners are more diverse. Specifically, the learners that are found later (with a larger number) are more diverse with each other. The reason is that the diversity surrogate becomes more precise when fitted with more observations, and thus it can suggest more diverse configurations. This observation indicates that Div BO is able to generate a more diverse ensemble while ensuring the performance of base learners. 5 Conclusion In this paper, we introduced Div BO, a diversity-aware framework based on Bayesian optimization to solve CASH problems for ensemble learning. In Div BO, we proposed to use a diversity surrogate to model the relationship between two configurations, and combined the ranking values of the performance and diversity surrogates with a saturating weight. Through empirical study, we showed that the prediction of the diversity surrogate achieves a satisfactory correlation with ground-truth results, and Div BO outperforms post-hoc designs in recent Auto ML systems and other baselines for ensemble learning in CASH problems. Acknowledgments This work is supported by NSFC (No. 61832001) and ZTE-PKU Joint Laboratory for Foundation Software. Yang Li and Bin Cui are the corresponding authors. [1] J. Bergstra and Y. Bengio. Random search for hyper-parameter optimization. Journal of Machine Learning Research, 13(Feb):281 305, 2012. [2] J. S. Bergstra, R. Bardenet, Y. Bengio, and B. Kégl. Algorithms for hyper-parameter optimization. In Advances in neural information processing systems, pages 2546 2554, 2011. [3] Y. Bian and H. Chen. When does diversity help generalization in classification ensembles? IEEE Transactions on Cybernetics, 2021. [4] C. S. Bojer and J. P. Meldgaard. Kaggle forecasting competitions: An overlooked learning opportunity. International Journal of Forecasting, 37(2):587 603, 2021. [5] S. Boyd, N. Parikh, E. Chu, B. Peleato, J. Eckstein, et al. Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine learning, 3(1):1 122, 2011. [6] L. Breiman. Stacked regressions. Machine learning, 24(1):49 64, 1996. [7] R. Caruana, A. Niculescu-Mizil, G. Crew, and A. Ksikes. Ensemble selection from libraries of models. In Proceedings of the twenty-first international conference on Machine learning, page 18, 2004. [8] T. Chen and C. Guestrin. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, pages 785 794. ACM, 2016. [9] T. G. Dietterich. Ensemble methods in machine learning. In International workshop on multiple classifier systems, pages 1 15. Springer, 2000. [10] G. FAN, B. LI, Q. HAN, R. JIAO, and G. QU. Robust lane detection and tracking based on machine vision. ZTE Communications, 18(4):69 77, 2021. [11] M. Feurer, A. Klein, K. Eggensperger, J. Springenberg, M. Blum, and F. Hutter. Efficient and robust automated machine learning. In Advances in neural information processing systems, pages 2962 2970, 2015. [12] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. Advances in neural information processing systems, 27, 2014. [13] L. K. Hansen and P. Salamon. Neural network ensembles. IEEE transactions on pattern analysis and machine intelligence, 12(10):993 1001, 1990. [14] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. [15] X. He, K. Zhao, and X. Chu. Automl: A survey of the state-of-the-art. Knowledge-Based Systems, 212:106622, 2021. [16] T. Hoch. An ensemble learning approach for the kaggle taxi travel time prediction challenge. In DC@ PKDD/ECML, 2015. [17] F. Hutter, H. H. Hoos, and K. Leyton-Brown. Sequential model-based optimization for general algorithm configuration. In International Conference on Learning and Intelligent Optimization, pages 507 523. Springer, 2011. [18] F. Hutter, L. Kotthoff, and J. Vanschoren. Automated machine learning: methods, systems, challenges. Springer Nature, 2019. [19] G. Ke, Q. Meng, T. Finley, T. Wang, W. Chen, W. Ma, Q. Ye, and T.-Y. Liu. Lightgbm: A highly efficient gradient boosting decision tree. Advances in neural information processing systems, 30, 2017. [20] L. I. Kuncheva and C. J. Whitaker. Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Machine learning, 51(2):181 207, 2003. [21] Y. Le Cun, Y. Bengio, and G. Hinton. Deep learning. nature, 521(7553):436 444, 2015. [22] J.-C. Lévesque, C. Gagné, and R. Sabourin. Bayesian hyperparameter optimization for ensemble learning. In Proceedings of the Thirty-Second Conference on Uncertainty in Artificial Intelligence, pages 437 446, 2016. [23] Y. Li, J. Jiang, J. Gao, Y. Shao, C. Zhang, and B. Cui. Efficient automatic cash via rising bandits. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 4763 4771, 2020. [24] Y. Li, Y. Shen, J. Jiang, J. Gao, C. Zhang, and B. Cui. Mfes-hb: Efficient hyperband with multi-fidelity quality measurements. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 8491 8500, 2021. [25] Y. Li, Y. Shen, W. Zhang, J. Jiang, B. Ding, Y. Li, J. Zhou, Z. Yang, W. Wu, C. Zhang, et al. Volcanoml: speeding up end-to-end automl via scalable search space decomposition. Proceedings of the VLDB Endowment, 14(11):2167 2176, 2021. [26] Y. Li, Y. Shen, W. Zhang, C. Zhang, and B. Cui. Efficient end-to-end automl via scalable search space decomposition. The VLDB Journal, 2022. [27] S. Limin, Z. Qiang, L. Shuang, and L. C. Harold. Balanced discriminative transfer feature learning for visual domain adaptation. ZTE Communications, 18(4):78 83, 2021. [28] C. Liu, B. P. Chamberlain, and E. J. Mc Coy. What is the value of experimentation and measurement? Data Science and Engineering, 5(2):152 167, 2020. [29] S. Liu, P. Ram, D. Vijaykeerthy, D. Bouneffouf, G. Bramble, H. Samulowitz, D. Wang, A. Conn, and A. Gray. An admm based framework for automl pipeline configuration. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 4892 4899, 2020. [30] M. Moghimi, S. J. Belongie, M. J. Saberian, J. Yang, N. Vasconcelos, and L.-J. Li. Boosted convolutional neural networks. In BMVC, volume 5, page 6, 2016. [31] R. S. Olson and J. H. Moore. Tpot: A tree-based pipeline optimization tool for automating machine learning. In Workshop on automatic machine learning, pages 66 74. PMLR, 2016. [32] I. Partalas, G. Tsoumakas, and I. P. Vlahavas. Focused ensemble selection: A diversity-based method for greedy ensemble selection. In ECAI, pages 117 121, 2008. [33] Y. Quanming, W. Mengshuo, J. E. Hugo, G. Isabelle, H. Yi-Qi, L. Yu-Feng, T. Wei-Wei, Y. Qiang, and Y. Yang. Taking human out of learning applications: A survey on automated machine learning. ar Xiv preprint ar Xiv:1810.13306, 2018. [34] E. Real, A. Aggarwal, Y. Huang, and Q. V. Le. Regularized evolution for image classifier architecture search. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 4780 4789, 2019. [35] J. Snoek, H. Larochelle, and R. P. Adams. Practical bayesian optimization of machine learning algorithms. In Advances in neural information processing systems, pages 2951 2959, 2012. [36] X. Su and T. M. Khoshgoftaar. A survey of collaborative filtering techniques. Advances in artificial intelligence, 2009, 2009. [37] F. Sun, J. Liu, J. Wu, C. Pei, X. Lin, W. Ou, and P. Jiang. Bert4rec: Sequential recommendation with bidirectional encoder representations from transformer. In Proceedings of the 28th ACM international conference on information and knowledge management, pages 1441 1450, 2019. [38] E. K. Tang, P. N. Suganthan, and X. Yao. An analysis of diversity measures. Machine learning, 65(1):247 271, 2006. [39] C. Thornton, F. Hutter, H. H. Hoos, and K. Leyton-Brown. Auto-weka: Combined selection and hyperparameter optimization of classification algorithms. In Proceedings of the 19th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 847 855, 2013. [40] J. Vanschoren, J. N. Van Rijn, B. Bischl, and L. Torgo. Openml: networked science in machine learning. ACM SIGKDD Explorations Newsletter, 15(2):49 60, 2014. [41] C. Wang, Q. Wu, M. Weimer, and E. Zhu. Flaml: a fast and lightweight automl library. Proceedings of Machine Learning and Systems, 3:434 447, 2021. [42] L. Yang, Z. Zhang, and S. Hong. Diffusion models: A comprehensive survey of methods and applications. ar Xiv preprint ar Xiv:2209.00796, 2022. [43] S. Zaidi, A. Zela, T. Elsken, C. C. Holmes, F. Hutter, and Y. Teh. Neural ensemble search for uncertainty estimation and dataset shift. Advances in Neural Information Processing Systems, 34, 2021. [44] W. Zhang, J. Jiang, Y. Shao, and B. Cui. Efficient diversity-driven ensemble for deep neural networks. In 2020 IEEE 36th International Conference on Data Engineering (ICDE), pages 73 84. IEEE, 2020. [45] W. Zhang, J. Jiang, Y. Shao, and B. Cui. Snapshot boosting: a fast ensemble framework for deep neural networks. Science China Information Sciences, 63(1):1 12, 2020. [46] W. Zhang, Y. Shen, Z. Lin, Y. Li, X. Li, W. Ouyang, Y. Tao, Z. Yang, and B. Cui. Pasca: A graph neural architecture search system under the scalable paradigm. In Proceedings of the ACM Web Conference 2022, pages 1817 1828, 2022. [47] Z.-H. Zhou. Ensemble methods: foundations and algorithms. CRC press, 2012. [48] Z.-H. Zhou, J. Wu, and W. Tang. Ensembling neural networks: many could be better than all. Artificial intelligence, 137(1-2):239 263, 2002. [49] D.-H. Zhu, X.-Y. Dai, and J.-J. Chen. Pre-train and learn: Preserving global information for graph neural networks. Journal of Computer Science and Technology, 36(6):1420 1430, 2021. [50] L. Zimmer, M. Lindauer, and F. Hutter. Auto-pytorch: multi-fidelity metalearning for efficient and robust autodl. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(9):3079 3090, 2021. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Section 3.3. (c) Did you discuss any potential negative societal impacts of your work? [No] (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] We have carefully read the guidelines. 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] We have a detailed implementation instruction in Appendix A.4. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] See Table 1 and Figure 4. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See Appendix A.4. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [N/A] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [Yes] We use the widely used and public data on Open ML. The datasets are used for evaluation in previous work. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [No] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]