# derandomized_smoothing_for_decision_stump_ensembles__080ade3e.pdf (De-)Randomized Smoothing for Decision Stump Ensembles Miklós Z. Horváth , Mark Niklas Müller , Marc Fischer, Martin Vechev Department of Computer Science ETH Zurich Switzerland mihorvat@ethz.ch, {mark.mueller,marc.fischer,martin.vechev}@inf.ethz.ch Tree-based models are used in many high-stakes application domains such as finance and medicine, where robustness and interpretability are of utmost importance. Yet, methods for improving and certifying their robustness are severely under-explored, in contrast to those focusing on neural networks. Targeting this important challenge, we propose deterministic smoothing for decision stump ensembles. Whereas most prior work on randomized smoothing focuses on evaluating arbitrary base models approximately under input randomization, the key insight of our work is that decision stump ensembles enable exact yet efficient evaluation via dynamic programming. Importantly, we obtain deterministic robustness certificates, even jointly over numerical and categorical features, a setting ubiquitous in the real world. Further, we derive an MLE-optimal training method for smoothed decision stumps under randomization and propose two boosting approaches to improve their provable robustness. An extensive experimental evaluation on computer vision and tabular data tasks shows that our approach yields significantly higher certified accuracies than the state-of-the-art for tree-based models. We release all code and trained models at https://github.com/eth-sri/drs. 1 Introduction Tree-based models have long been a favourite for making decisions in high-stakes domains such as medicine and finance, due to their interpretability and exceptional performance on tabular data [1]. However, recent results have highlighted that tree-based models are, similarly to other machine learning models [2, 3], also highly susceptible to adversarial examples [4 6], raising concerns about their use in high-stakes domains where errors can have dire consequences. While the robustness of neural models has received considerable attention [7 21], the challenge of obtaining robustness guarantees for ensembles of tree-based models has only been investigated recently [4, 22, 23]. However, these initial works only consider numerical features and are based on worst-case approximations, which do not scale well to the difficult ℓp-norm setting. This Work In this work, we address this challenge and present DRS, a novel (De-)Randomized Smoothing approach, for constructing robust tree-based models with deterministic ℓp-norm guarantees while supporting both categorical and numerical variables. Unlike prior work, our method is based on Randomized Smoothing (RS) [24], an approach that obtains robustness guarantees by evaluating a general base model under an input randomization ϕ(x). However, in contrast to standard applications of RS, which use costly and imprecise approximations via sampling and only obtain probabilistic certificates, we leverage the structure of decision stump ensembles to compute their exact output Equal contribution 36th Conference on Neural Information Processing Systems (Neur IPS 2022). 0.0 0.5 1.0 0.0 0.5 1.0 0.0 ρ 1(0) Natural Prediction success probabilitypy class 0 class 1 (a) Meta-Stumps (b) PDFs (c) Ensemble PDF (d) Ensemble CDF Figure 1: Given an ensemble of 3 meta-stumps fi (piecewise constant univariate functions), each operating on a different feature xi of an input x, we calculate the probability of every output under input randomization (a) to obtain a distribution over their outputs (b). We aggregate these individual PDFs via dynamic programming to obtain the probability distribution over the ensemble s outputs (c). We can then compute the corresponding CDF (d) to evaluate the smoothed stump ensemble exactly. distributions for a given input randomization scheme and thus obtain deterministic certificates. Our key insight is that this distribution can be efficiently computed by aggregating independent distributions associated with the individual features used by the ensemble. We illustrate this idea in Fig. 1: In (a), we show an ensemble of decision stumps over three features (x1, x2, x3), aggregated to piecewise constant functions over one feature each (discussed in Section 3) and evaluated under the input randomization ϕ(x), here a Gaussian. We can compute the independent probability density functions of their outputs (PDFs) (shown in (b)) directly, by evaluating the (Gaussian) cumulative density function (CDF) over the constant regions. Aggregating the individual PDFs (discussed in Section 3), we can efficiently compute the exact PDF (c) and CDF (d) of the ensemble s output. To evaluate and certify the smoothed model, we can now simply look up the median prediction and success probability, respectively, in the CDF, without requiring sampling. DRS combines ℓp-norm certificates over numerical features, computed as described above, with an efficient worst-case analysis for ℓ0-perturbations of categorical features in order to, for the first time, provide joint certificates. To train models amenable to certification with DRS, we propose a robust MLE optimality criterion for training individual stumps and two boosting schemes targeting the certified robustness of the whole ensemble. We show empirically that DRS significantly improves on the state-of-the-art, increasing certified accuracies on established benchmarks up to 4-fold. Main Contributions Our key contributions are: DRS, a novel and efficient (De-)Randomized Smoothing approach for robustness certification, enabling joint deterministic certificates over numerical and categorical variables (Section 3). A novel MLE optimality criterion for training decision stumps robust under input randomization and two boosting approaches for certifiably robust stump ensembles (Section 4). An extensive empirical evaluation, demonstrating the effectiveness of our approach and establishing a new state-of-the-art in a wide range of settings (Section 5). 2 Background on Randomized Smoothing For a given base model F : Rd [C], classifying inputs to one of C Z 2 classes, Randomized Smoothing (RS) is a method to construct a classifier G: Rd [C] with robustness guarantees. For a randomization scheme ϕ: Rd Rd, we define the success probability py :=Px ϕ(x)[F(x )=y] and G(x):=arg maxc [C] py. Depending on the choice of ϕ, we obtain different certificates of the form: Theorem 2.1 (Adapted from Cohen et al. [24], Yang et al. [25]). If P(F(ϕ(x)) = y) := py py and py > 0.5, then G(x + δ) = y for all δ satisfying δ p < R with R := ρ(py). Table 1: Randomized Smoothing guarantees. ϕ(x) R := ρ(py) ℓ1 x + Unif([ λ, λ]d) 2λ(py 1 ℓ2 x + N(0, σI) σΦ 1(py) In particular, we present two instantiations that we utilize throughout this paper in Table 1, where Φ 1 is the inverse Gaussian CDF. Similar results, yielding other ℓp-norm certificates, can be derived for a wide range of input randomization schemes [25, 26]. Note that, by using more information than just py, e.g., pc for the runner-up class c, tighter certificates can be obtained [24, 27]. Once py is computed, we can directly calculate the certifiable radius R := ρ(py). For a broader overview of variants of Randomized Smoothing, please refer to Section 6. For most choices of F and ϕ, the exact success probability py can not be computed efficiently. Thus a lower bound py is estimated with confidence 1 α (typically α = 10 3) using Monte Carlo sampling and the Neyman-Pearson lemma [28]. Not only is this extremely computationally expensive, as typically 100 000 samples have to be evaluated per data point, but this also severely limits the maximum certifiable radius (see Fig. 6) and only yields probabilistic guarantees. Additionally, if the number of samples is not sufficient for the statistical test, the procedure will abstain from classifying. In the following, we will show how considering a specific class of models F allows us to compute the success probability py exactly, overcoming these drawbacks, and thus invoke ρ(py) to compute deterministic certificates over larger radii, orders of magnitude faster than RS. 3 (De-)Randomized Smoothing for Decision Stump Ensembles Tree-based models such as decision stump ensembles often combine exceptional performance on tabular data [1] with good interpretability, making them ideal for many real-world high-stakes applications. Here, we propose a (De-)Randomized Smoothing approach, DRS, to equip them with deterministic robustness guarantees. For this, we first revisit decision stump ensembles and then show that their structure permits an exact evaluation under isotropic input randomization schemes, such as those discussed in Section 2. Finally, we propose joint certification over numerical and categorical variables, as many practical tabular datasets often contain both variable types. Stump Ensembles We define a decision stump as fm(x) = γl,m + (γr,m γl,m)1xjm>vm, with leaf predictions γl,m, γr,m [0, 1], split position vm, and split variable jm. We construct unweighted ensembles, particularly suitable for Smoothing [29], of M such stumps f M : Rd 7 [0, 1] as f M(x) := 1 m=1 fm(x), (1) and treat them as a binary classifiers 1 f M(x)>0.5. While our approach is extensible to multi-class classification by replacing the scalar leaf predictions γ with prediction-vectors, assigning a score per class, we focus on the binary case in this work. Smoothed Stump Ensemble We now define a smoothed stump ensemble g M along the lines of Randomized Smoothing as discussed in Section 2, by evaluating f M not only on the original input x but rather on a whole distribution of x ϕ(x): f1(xi) γl,1 γr,1 γr,1 f2(xi) γl,2 γl,2 γr,2 fi(xi) γi,1 Figure 2: A meta-stump constructed from two stumps. g M(x) := Px ϕ(x)[ f M(x ) > 0.5]. In this work, we consider randomization schemes ϕ(x) that are (i) isotropic, i.e., the dimensions of x ϕ(x) are independently distributed, and (ii) permit an efficient computation of their marginal cumulative distribution functions (CDF). This includes a wide range of distributions, e.g., the Gaussian and Uniform distributions used in Table 1 and others commonly used for RS [25]. By denoting the model CDF as FM,x(z) = Px ϕ(x)[ f(x ) z], we can alternatively define g M as g M(x) := 1 FM,x(0.5), which will become useful later. For a label y {0, 1} we obtain the success probability py = |y FM,x(0.5)| of predicting y for a sample from ϕ(x). Algorithm 1 Stump Ensemble PDF computation via Dynamic Programming function COMPUTEPDF({(Γ, v)i}d i=1, x, ϕ) pdf[i][t] = 0 for t [M + 1], i [d] pdf[0][0] = 1 For 0 stumps all probability mass is on 0 for i = 1 to d do for j = 1 to Mi do for t = 0 to M + 1 Γi,j do pdf[i][t + Γi,j] = pdf[i][t + Γi,j] + pdf[i 1][t] Px i ϕ(x)[vi,j 1 < x i vi,j] Meta-Stumps To evaluate py exactly as illustrated in Fig. 1, we group the stumps constituting an ensemble by their split variable jm to obtain one meta-stump fi per feature i. The key idea is that outputs of these meta-stumps are now independently distributed under isotropic input randomization (illustrated in Fig. 1 (b)), allowing us to aggregate them efficiently later on. We showcase this in Fig. 2, where two stumps (f1 and f2) are combined into the meta-stump fi. Formally, we have m Ii fm(x), Ii := {m [M] | jm = i}, (2) define Mi = |Ii| and rewrite our ensemble as f M(x) := 1 M Pd i=1 fi(xi). Every meta-stump can be represented by its split positions vi,j, sorted such that vi,j vi,j+1, and its predictions γi,j = Pj 1 t=1 γr,t + P|Ii| t=j γl,m on each of the resulting |Ii| + 1 regions, written as (γ, v)i. CDF Computation Now we leverage the independence of our meta-stumps output distributions under an isotropic input randomization scheme ϕ to compute the PDF of their ensemble efficiently via dynamic programming (DP) (illustrated in Fig. 1 (c) and explained below). Given its PDF, we can trivially compute the ensemble s CDF FM,x, allowing us to evaluate the smoothed model exactly (illustrated in Fig. 1 (d)). This efficient CDF computation constitutes the core of DRS. In more detail, we observe that the PDF of a stump ensemble is the convex sum of exponentially many (O((maxi Ii)d)) Dirac-delta distributions. To avoid this exponential blow-up, we discretize all leaf predictions γ to a grid of values (typically = 100), when constructing the smoothed model g M. For each γi,j, we define a corresponding Γi,j {0, . . . , Mi } such that γi,j = Γi,j . Now, we construct a DP-table, where every entry pdf[i][t] corresponds to the weight of the Dirac-delta associated with an output of t M after considering the first i meta-stumps (in any arbitrary but fixed order). We show the PDF computation in Algorithm 1 and provide an intuition below. We initialize pdf[0][*] by allocating all probability mass to t = 0 (pdf[0][0]=1). Now, we compute pdf[i][*] from pdf[i-1][*] by accounting for the effect of the ith meta-stump as follows: The weight of the Dirac-delta at t after considering i meta-stumps is exactly the sum over the weights of the Dirac-deltas at t-Γi,j after i-1 meta-stumps, weighted with the probability pi,j := Px i ϕ(x)[vi,j 1 < x i vi,j] of the ith meta stump predicting Γi,j. We compute pi,j as the probability of the randomized x i lying between vi,j 1 and vi,j (padded with and on the left and right, respectively), as illustrated in Fig. 2. After termination, the last line of the DP-table pdf[d][*] contains the full PDF (see Fig. 1(c)). Formally we summarize this in the theorem below, delaying a formal proof to App. A.1: Theorem 3.1. For z [0, 1], FM,x(z) = P z M t=0 pdf[d][t] describes the exact CDF and thus success probability py = Px ϕ(x)[ f M(x ) = y] = |y FM,x(0.5)| for y {0, 1}. Note that the presented algorithm is slightly simplified, and we actually only have to track the range of non-zero entries of one row of the DP-table. This allows us to compute the full PDF and thus certificates for smoothed stump ensembles very efficiently, e.g., taking only around 1.2 s total for the MNIST 2 VS. 6 task (around 2.000 data points and over 500 stumps). Certification Recall from Section 2 that, given the success probability py, robustness certification for ℓp-norm bounded perturbations reduces to computing the maximal certifiable robustness radius R = ρ(py). For all popular ℓp-norms, ρ (and its inverse ρ 1; used shortly) can be either evaluated symbolically [24, 25] or precomputed efficiently [30, 31], such that the core challenge of certification becomes computing (a lower bound to) py, which we solve efficiently via Theorem 3.1. Alternatively, for a given target radius r, we need to check whether py ρ 1(r) by equivalently calculating g M,r(x) = F 1 M,x(z) z = 1 ρ 1(r) if y = 1 ρ 1(r) if y = 0 , (3) and checking g M,r(x) > 0.5. This corresponds to asserting that class y is predicted at least z of the time. Here, the inverse CDF F 1 M,x(z) can be efficiently evaluated using the step-wise FM,x computed via Theorem 3.1. We will see in Section 4 that this view is useful when training stump ensembles for certifiability. Finally, we want to highlight that this approach can be used with all common randomization schemes yielding certificates for different ℓp-norm bounded adversaries. Categorical Variables & Joint Certificates For practical applications, it is essential to handle both numerical and categorical features jointly. To consider a categorical feature xi {1, . . . , di} in our stump ensemble, we construct a di-ary stump fi : [di] [0, 1] returning a value γi,j corresponding to each of the di categorical values and treated as a meta-stump with Mi = 1 for normalization. To provide certificates in this setting, we propose a novel scheme combining an arbitrary ℓp-norm certificate of radius rp over all numerical features, computed as discussed above, with an ℓ0 certificate of radius r0 over all categorical features C, computed using an approach adapted from Wang et al. [23]. Conceptually, we compute the worst-case effect of every individual categorical variable independently, greedily aggregate these worst-case effects, and account for them in our ensemble s CDF. Given a meta-stump s prediction on a concrete sample qi = fi(xi) as well as its maximal and minimal output ui and li, respectively, we compute the maximum and minimum perturbation effect to δi = ui qi M and δi = li qi M , respectively. Given the set of categorical features C, we can compute the worst-case effect when perturbing at most r0 samples as δr0 = max R δi, s.t.|R| r0, R C 0.3 0.5 0.7 z Figure 3: CDF shifted by the effect of categorical feature perturbations. by greedily picking the r0 largest δi. For δr0 we proceed analogously. Shifting the CDF, computed as above, by δ and δ for samples with labels y = 0 and y = 1, respectively, before computing the success probability py, allows us to account for the worst-case categorical perturbations exactly. We illustrate this for a sample with y = 0 in Fig. 3, where we show the CDFs obtained by all possible perturbations of at most r0 categorical variables, bounded to the right by those obtained by shifting the original by δr0. Note that here no smoothing over the categorical variables is done or required, making inference trivial. 4 Training for and with (De-)Randomized Smoothing To obtain large certified radii via smoothing, the base model has to be robust to the chosen randomization scheme. To train robust decision stump ensembles, we propose a robust MLE optimality criterion for individual stumps (Section 4.1) and two boosting schemes for whole ensembles (Section 4.2). 4.1 Independently MLE-Optimal Stumps To train an individual stump fm(x) = γl,m + (γr,m γl,m)1xjm>vm, its split feature jm, split position vm, and leaf predictions γl,m, γr,m have to be determined. We choose them in an MLEoptimal fashion with respect to the randomization scheme ϕ, starting with vm, as follows: We consider the probabilities pl,i(vm) = Px ϕ(xi)[x jm vm] and pr,i = 1 pl,i(vm) of x i lying to the left or the right of vm, respectively, under the input randomization scheme ϕ. To avoid clutter, we drop the explicit dependence on vm in the following. For an i.i.d. dataset with n samples (xi, yi) (X, Y), we define the probabilities py j = 1 {i|yi=y} pj,i of picking the j {l, r} leaf, conditioned on the target label, and pj = p0 j + p1 j as their sum to compute the entropy impurity Hentropy [32] as Hentropy = X j {l,r} pj X py j pj log We then choose the vm approximately minimizing Hentropy via line-search. After fixing vm this way, we compute the MLE-optimal leaf predictions γϕ,MLE l and γϕ,MLE r as: γϕMLE l , γϕMLE r = arg max γl,γr P[Y | ϕ(X), fm] = arg max γl,γr i=1 Ex ϕ(xi) [log P[yi | x , fm]] = arg max γl,γr i {i|yi=0} pl,i log(1 γl) + pr,i log(1 γr) i {i|yi=1} pl,i log(γl) + pr,i log(γr) = arg max γl,γr p0 l log(1 γl) + p0 r log(1 γr) + p1 l log(γl) + p1 r log(γr), where the second line is obtained by splitting the sum over samples by class and explicitly computing the expectation. We solve the maximization problem by setting the first derivatives γl and γr of our optimization objective to zero and checking its Hessian to confirm that γϕMLE l = p1 l p1 l + p0 l γϕMLE r = p1 r p1r + p0r (4) are indeed maxima. We show in App. A.2 that γϕ,MLE l , γϕ,MLE l , and vm are even jointly MLE-optimal, when vm is chosen as the exact instead of an approximate minimizer of the entropy impurity. Ensembling To train an ensemble of independently MLE-optimal decision stumps, we sequentially train one stump for every feature jm [d] and construct an ensemble with equal weights, rejecting stumps with an entropy impurity Hentropy above a predetermined threshold. 4.2 Boosting Stump Ensembles for Certifiable Robustness 0.0 0.1 0.5 1.0 z Clean Prediction Cert. Prediction at r(q = 0.9) Figure 4: Inverse CDF F 1 M Decision stumps trained this way maximize the expected likelihood under the chosen randomization scheme. Assuming (due to the law of large numbers) a Gaussian output distribution, this corresponds to optimizing for the median output, which determines the clean prediction. However, certified correctness at a given radius r is determined by the prediction y (x, r) = F 1 m 1,x(z(r)) at the z(r) := |y ρ 1(r)| percentile of the output distribution. Where we call y the certifiable prediction, as certification is now equivalent to checking y = 1y (x,r)>0.5 (Eq. (3)). This difference is illustrated in Fig. 4, where the clean prediction is correct (class 1) while the certifiable prediction is incorrect. To align our training objective better with certified accuracy, we propose two novel boosting schemes along the lines of the popular TREEBOOST [33] and ADABOOST [34]. Gradient Boosting for Certifiable Robustness The key idea of gradient boosting is to compute the gradient of a loss function with respect to an ensemble s outputs and then add a model to the ensemble that makes a prediction along this gradient direction. Implementing this idea, we adapt TREEBOOST [33] to propose ROBTREEBOOST: At a high level, we add stumps to the ensemble, which aim to predict the residual between the target label and the current certifiable prediction. Concretely, to add the mth stump to our ensemble, we begin by computing the current ensemble s certifiable predictions y (r) at a target radius r and then defining the pseudo labels y = y y (r) as the residual between the target labels y and the certifiable predictions y (r). This yields a regression problem, which we tackle by choosing a feature jm and split threshold vm (approximately) minimizing MSE impurity under input randomization before computing γl,m and γr,m as approximate minimizers of the cross-entropy loss over the whole ensemble. Please see App. A.3 for a more detailed discussion of ROBTREEBOOST. Table 2: Natural accuracy (NAC) [%] and certified accuracy (CA) [%] with respect to ℓ1and ℓ2-norm bounded perturbations. Results for Wang et al. [23] as reported by them. Larger is better. Perturbation Dataset Radius r Standard Training Wang et al. [23] Ours (Independent) Ours (Boosting) NAC NAC CA NAC CA NAC CA BREASTCANCER 1.0 99.3 98.5 64.2 100.0 81.0 100.0 83.9 DIABETES 0.05 74.7 72.7 68.2 76.0 69.5 77.9 72.1 FMNIST-SHOES 0.5 95.0 87.6 67.8 85.8 83.3 87.2 84.2 MNIST 1 VS. 5 1.0 99.1 95.5 83.8 96.6 94.1 99.3 98.1 MNIST 2 VS. 6 1.0 96.0 92.3 66.5 96.3 93.9 96.6 94.1 BREASTCANCER 0.7 99.3 91.2 60.6 100.0 75.2 100.0 82.5 DIABETES 0.05 74.7 - - 77.3 68.2 79.9 71.4 FMNIST-SHOES 0.4 95.0 75.5 51.5 86.8 81.2 91.0 84.5 MNIST 1 VS. 5 0.8 99.1 95.6 63.4 95.8 91.6 99.2 96.3 MNIST 2 VS. 6 0.8 96.0 86.3 23.0 96.3 89.6 96.3 89.6 Adaptive Boosting for Certifiable Robustness The key idea of adaptive boosting is to build an ensemble by iteratively training models, weighted based on their error rate, while adapting sample weights based on whether they are classified correctly. We build on ADABOOST [34] to propose ROBADABOOST: We construct an ensemble of K stump ensembles via hard voting, where every ensemble is weighted based on its certifiable accuracy. To train a new ensemble, we increase the weights of all samples that are currently not classified certifiably correctly at a given radius r. We choose stump ensembles instead of individual stumps as base classifiers because single stumps often can not reach the success probabilities under input randomization required for certification. To compute the certifiable radius for such an ensemble FK, we compute the certifiable radii Rk of the individual stump ensembles f k M, sort them in decreasing order such that Rk Rk+1 and obtain the largest radius Rk such that the weights of the first k ensembles sum up to more than half of the total weights. Please see App. A.4 for a more detailed discussion of ROBADABOOST. 5 Experimental Evaluation In this section, we empirically demonstrate the effectiveness of DRS in a wide range of settings. We show that DRS significantly outperforms the current state-of-the-art for certifying tree-based models on established benchmarks, using only numerical features (Section 5.1), before highlighting its novel ability to obtain joint certificates on a set of new benchmarks (Section 5.2). Finally, we perform an ablation study, investigating the effect of DRS s key components (Section 5.3). Experimental Setup We implement our approach in Py Torch [35] and evaluate it on Intel Xeon Gold 6242 CPUs and an NVIDIA RTX 2080Ti. We compare to prior work on the DIABETES [36], BREASTCANCER [37], FMNIST-SHOES [38], MNIST 1 VS. 5 [39], and MNIST 2 VS. 6 [39] datasets and are the first to provide joint certificates of categorical and numerical features, demonstrated on the ADULT [37] and CREDIT [37] datasets. For a more detailed description of the experimental setup, please refer to App. B. 5.1 Certification for Numerical Features In Table 2, we compare the certified accuracies obtained via DRS on ensembles of independently MLE optimal stumps (Independent) or boosted stump ensembles (Boosting) to the current state-ofthe-art, Wang et al. [23], and standard training [40] using established benchmarks [23]. Independently MLE Optimal Stumps We first consider stump ensembles trained without boosting as described in Section 4.1 and observe that DRS obtains higher certified accuracies in all settings and higher natural accuracies in most. For example, on MNIST 2 VS. 6, we increase the certified accuracy at an ℓ2 radius of r2 = 0.8 from 23.0% to 89.6%, almost quadrupling it compared to Wang et al. [23], while also improving natural accuracy from 86.3% to 96.3%. Boosting for Certified Accuracy Leveraging the boosting techniques introduced in Section 4.2, ROBTREEBOOST for BREASTCANCER and DIABETES and ROBADABOOST for FMNIST-SHOES, MNIST 1 VS. 5, and MNIST 2 VS. 6, we increase certifiable and natural accuracies even further in most settings. For example, compared to our independently trained stump ensemble, we improve the certified accuracy for MNIST 1 VS. 5 at an ℓ1-radius of r1 = 1.0 from 94.1% to 98.1% and for BREASTCANCER at an ℓ2-radius of r2 = 0.7 from 75.2% to 82.5%. Table 3: Balanced certified accuracy (BCA) [%] under joint ℓ0and ℓ2-perturbations of categorical and numerical features, respectively, depending on whether model uses categorical and/or numerical features. The balanced natural accuracy is the BCA at radius r = 0.0. Larger is better. Dataset Categorical Features ℓ0 Radius r0 BCA without Numerical Features BCA with Numerical Features at ℓ2 Radius r2 0.00 0.25 0.50 0.75 1.00 1.25 1.50 no - - 74.9 65.7 42.4 27.4 14.5 8.9 5.1 0 76.6 77.5 73.9 68.1 63.3 48.7 40.7 35.2 1 57.4 66.0 61.7 53.9 47.4 34.3 26.6 21.8 2 33.5 51.4 46.2 37.5 29.3 21.5 17.1 13.4 3 8.9 36.7 31.4 24.1 15.4 10.3 8.1 5.7 no - - 56.1 44.5 33.3 17.7 9.7 7.2 5.0 0 70.7 74.1 70.3 67.3 59.7 57.1 54.9 53.4 1 48.2 52.7 47.7 41.7 38.3 37.1 35.1 34.7 2 26.4 29.3 26.0 23.8 19.2 16.8 13.5 13.0 3 7.8 13.6 10.3 7.8 4.9 4.4 3.9 3.4 5.2 Joint Certificates for Categorical and Numerical Features 0.0 0.5 1.0 1.5 2.0 ℓ2 Radius r2 Balanced Certified Accuracy [%] r0 = 0 r0 = 1 r0 = 2 r0 = 3 no categorical features Figure 5: Effect of ℓ0-perturbations on ℓ2robustness for CREDIT. In Table 3, we compare models using only numerical, only categorical, or both types of features with regards to their balanced certified accuracy (BCA) (accounting for class frequency) at different combinations of ℓ2and ℓ0-radii for numerical and categorical features, respectively. We observe that models using both categorical and numerical features perform notably better on clean data, highlighting the importance of utilizing and thus also certifying them in combination. Moreover, categorical features make the model significantly more robust to ℓ2 perturbations, e.g., at ℓ2-radii 0.75, they improve certified accuracies, even when 2 categorical features (of only 8 and 7 for ADULT and CREDIT, respectively) are adversarially perturbed. We visualize this in Fig. 5, showing BCA over ℓ2-perturbation radius and confirm that the model utilizing only numerical features (dotted line) loses accuracy much quicker with perturbation magnitude than the model leveraging categorical variables (solid lines). As we are the first to tackle this setting, we do not compare to other methods but provide more detailed experiments in App. C.1. 5.3 Ablation Study We first illustrate the effectiveness of our derandomization approach, before demonstrating the benefit of training with our MLE optimality criterion and investigating the effect of the noise level on DRS. 0.0 1.0 2.0 3.0 4.0 ℓ2 Radius r2 Certified Accuracy [%] n = 100 n = 1 000 n = 10 000 n = 100 000 DRS (ours) Figure 6: DRS vs. RS with various sample counts n on MNIST 1 VS. 5. (De-)Randomized vs Randomized Smoothing In Fig. 6, we compare DRS, (dotted line) and sampling-based RS (solid lines), w.r.t. certified accuracy over ℓ2 radii. We observe that the sampling-based estimation of the success probability in RS significantly limits the obtained certifiable radii. While this effect is particularly pronounced for small sample counts n, increasing the maximum certifiable radius, visible as the sudden drop in certifiable accuracy, requires an exponentially increasing number of samples, making the certification of large radii intractable. DRS, in contrast, can compute exact success probabilities and thus deterministic guarantees for much larger radii, yielding a 33.1% increase in ACR compared to using n = 100 000 samples. Additionally, DRS is multiple orders of magnitude faster than RS, here, only requiring approximately 6.45 10 4 s per sample. For more extensive experiments, please refer to App. C.2. Table 4: Comparison of training with the exact distribution (MLE), randomly perturbed data (Sampling), or clean data (Default) on BREASTCANCER for σ = 1. Method ACR Certified Accuracy [%] at Radius r 0.0 0.25 0.5 0.75 MLE (Ours) 0.675 100.0 97.1 86.1 30.7 Sampling 0.567 99.3 95.6 75.2 8.8 Default 0.356 26.3 25.5 25.5 25.5 MLE Optimality Criterion In Table 4, we evaluate our robust MLE optimality criterion (MLE) by comparing it to the standard entropy criterion applied to samples drawn from the input randomization scheme (Sampling) or the clean data (Default). We observe that the ensemble trained on the clean data (Default) suffers from a mode collapse when evaluated under noise. In contrast, both approaches considering the input randomization perform much better, with our robust MLE approach outperforming sampling by a significant margin, especially at large radii. For more extensive experiments, please refer to App. C.3. 0.0 4.0 8.0 12.0 16.0 ℓ1 Radius r1 Certified Accuracy [%] λ = 0.5 λ = 1.0 λ = 2.0 λ = 4.0 λ = 8.0 λ = 16.0 Figure 7: Comparing DRS for various noise levels λ on MNIST 1 VS. 5. Effect of Noise Level In Fig. 7, we compare the certified accuracy over ℓ1-radii for a range of different noise magnitudes λ and ensembles of independently MLE optimal stumps. We observe that at large perturbation magnitudes, we obtain stumps that think outside the (hyper-)box , i.e., choose splits outside of the original data range, making their ensembles exceptionally robust, even at large radii. In particular, we obtain a certifiable accuracy of 87.3% at radius r1 = 4.0, while the state-ofthe-art achieves only 83.8% at r1 = 1.0 [23]. We provide more experiments for varying noise magnitudes in App. C.4. 6 Related Work (De-)Randomized Smoothing Probabilistic certification methods [41, 42, 24] are a popular approach for obtaining robustness certificates for a wide range of tasks [31, 43 46], threat models [25 27, 30, 31, 43, 47 53], and robustness-accuracy trade-offs [54]. These methods follow the general blueprint discussed in Section 2 and consider arbitrary base classifiers, though specially trained [55 57]. While recent work [29, 58] has found ensembles to be particularly suitable base classifiers, they use neural networks and can thus, in contrast to our work, not leverage their structure. Specifically designed randomization schemes [51, 53] enable efficient enumeration and thus a deterministic certificate for, e.g., patch attacks or ℓ1-norm perturbations. In contrast to these approaches, we permit arbitrary isotropic continuous randomization schemes, allowing us to leverage comprehensive results on RS to obtain robustness guarantees against a wide range of ℓp-norm bounded adversaries [25]. Certification and Training of Tree-Based Models In the setting of ℓ robustness, where every feature can be perturbed independently, various methods have been proposed to train [4, 22, 59 61] and certify [22, 62 64] robust decision trees and stumps. However, ℓ robust models are still vulnerable to other ℓp norm perturbations [65, 66], which cover many realistic perturbations better and are the focus of this work. There, the admissible perturbation of one feature depends on the perturbations of all others, making the above approaches leveraging their independence not applicable. On the other hand, Kantchelian et al. [67] discuss complete robustness certification of tree ensembles in the ℓp-norm setting via MILP. However, this approach is intractable in most settings due to its Co NP-complete complexity. Wang et al. [23] propose an efficient but incomplete DP-based certification algorithm for stump ensembles based on over-approximating the maximum perturbation effect in the ℓp-norm setting. While similarly fast as our approach, we show empirically in Section 5 that DRS obtains significantly stronger certificates. Wang et al. [23] further introduce an incomplete certification algorithm for tree ensembles, which is based on computing the distance between the pre-image of all trees leaves and the original sample. As they report significantly worse results using this approach than with stump ensembles, we omit a detailed comparison. 7 Limitations and Societal Impact Limitations While able to handle arbitrary stump ensembles, and being extensible to arbitrary decision trees (see App. D), DRS can not handle arbitrary ensembles of decision trees. However, as these have been shown to be significantly more sensitive to ℓp-norm perturbations than stump ensembles [23], we believe this limitation to be of little practical relevance. Further, like all Smoothingbased approaches, we construct a smoothed model from a base classifier and only obtain robustness guarantees for the former. In contrast to standard Randomized Smoothing approaches, we can, however, evaluate the smoothed model exactly and efficiently. Societal Impact As our contributions improve certified accuracy and certification radii while retaining high natural accuracy, they could help make real-world AI systems more robust and thus generally amplify both any positive or negative societal effects. Further, while we achieve state-of-the-art results, these may not be sufficient to guarantee robustness in real-world deployment and could give practitioners a false sense of security, leading to them relying more on our models than is justified. 8 Conclusion We propose DRS, a (De-)Randomized Smoothing approach to robustness certification, enabling joint deterministic certificates over numerical and categorical variables for decision stump ensembles by leveraging their structure to compute their exact output distributions for a given input randomization scheme. The key insight enabling this is that this output distribution can be efficiently computed by aggregating independent distributions associated with the individual features used by the ensemble. We additionally propose a robust MLE optimality criterion for training individual decision stumps and two boosting schemes improving an ensemble s certifiable accuracy. Empirically, we demonstrate that DRS significantly outperforms the state-of-the-art for tree-based models in a wide range of settings, obtaining up to 4-fold improvements in certifiable accuracy. [1] R. Shwartz-Ziv and A. Armon, Tabular data: Deep learning is not all you need, Information Fusion, vol. 81, 2022. [2] B. Biggio, I. Corona, D. Maiorca, B. Nelson, N. Srndic, P. Laskov, G. Giacinto, and F. Roli, Evasion attacks against machine learning at test time, in Machine Learning and Knowledge Discovery in Databases - European Conference, ECML PKDD 2013, Prague, Czech Republic, September 23-27, 2013, Proceedings, Part III, vol. 8190, 2013. [3] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. J. Goodfellow, and R. Fergus, Intriguing properties of neural networks, in Proc. of ICLR, 2014. [4] H. Chen, H. Zhang, D. S. Boning, and C. Hsieh, Robust decision trees against adversarial examples, in Proc. of ICML, vol. 97, 2019. [5] F. Cartella, O. Anunciação, Y. Funabiki, D. Yamaguchi, T. Akishita, and O. Elshocht, Adversarial attacks for tabular data: Application to fraud detection and imbalanced data, in Proceedings of the Workshop on Artificial Intelligence Safety 2021 (Safe AI 2021) co-located with the Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI 2021), Virtual, February 8, 2021, vol. 2808, 2021. [6] Y. Mathov, E. Levy, Z. Katzir, A. Shabtai, and Y. Elovici, Not all datasets are born equal: On heterogeneous tabular data and adversarial examples, Knowl. Based Syst., vol. 242, 2022. [7] G. Singh, T. Gehr, M. Püschel, and M. T. Vechev, An abstract domain for certifying neural networks, Proc. ACM Program. Lang., vol. 3, no. POPL, 2019. [8] K. Xu, Z. Shi, H. Zhang, Y. Wang, K. Chang, M. Huang, B. Kailkhura, X. Lin, and C. Hsieh, Automatic perturbation analysis for scalable certified robustness and beyond, in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. [9] T. Gehr, M. Mirman, D. Drachsler-Cohen, P. Tsankov, S. Chaudhuri, and M. T. Vechev, AI2: safety and robustness certification of neural networks with abstract interpretation, in 2018 IEEE Symposium on Security and Privacy, SP 2018, Proceedings, 21-23 May 2018, San Francisco, California, USA, 2018. [10] S. Wang, K. Pei, J. Whitehouse, J. Yang, and S. Jana, Efficient formal safety analysis of neural networks, in Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS 2018, December 3-8, 2018, Montréal, Canada, 2018. [11] T. Weng, H. Zhang, H. Chen, Z. Song, C. Hsieh, L. Daniel, D. S. Boning, and I. S. Dhillon, Towards fast computation of certified robustness for relu networks, in Proc. of ICML, vol. 80, 2018. [12] E. Wong and J. Z. Kolter, Provable defenses against adversarial examples via the convex outer adversarial polytope, in Proc. of ICML, vol. 80, 2018. [13] G. Singh, T. Gehr, M. Mirman, M. Püschel, and M. T. Vechev, Fast and effective robustness certification, in Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS 2018, December 3-8, 2018, Montréal, Canada, 2018. [14] M. N. Müller, G. Makarchuk, G. Singh, M. Püschel, and M. T. Vechev, PRIMA: general and precise neural network certification via scalable convex hull approximations, Proc. ACM Program. Lang., vol. 6, no. POPL, 2022. [15] V. Tjeng, K. Y. Xiao, and R. Tedrake, Evaluating robustness of neural networks with mixed integer programming, in Proc. of ICLR, 2019. [16] S. Dathathri, K. Dvijotham, A. Kurakin, A. Raghunathan, J. Uesato, R. Bunel, S. Shankar, J. Steinhardt, I. J. Goodfellow, P. Liang, and P. Kohli, Enabling certification of verificationagnostic networks via memory-efficient semidefinite programming, in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. [17] R. Ehlers, Formal verification of piece-wise linear feed-forward neural networks, in Automated Technology for Verification and Analysis - 15th International Symposium, ATVA 2017, Pune, India, October 3-6, 2017, Proceedings, vol. 10482, 2017. [18] M. Mirman, T. Gehr, and M. T. Vechev, Differentiable abstract interpretation for provably robust neural networks, in Proc. of ICML, vol. 80, 2018. [19] M. Balunovic and M. T. Vechev, Adversarial training and provable defenses: Bridging the gap, in Proc. of ICLR, 2020. [20] A. Raghunathan, J. Steinhardt, and P. Liang, Certified defenses against adversarial examples, in Proc. of ICLR, 2018. [21] S. Gowal, K. Dvijotham, R. Stanforth, R. Bunel, C. Qin, J. Uesato, R. Arandjelovic, T. A. Mann, and P. Kohli, On the effectiveness of interval bound propagation for training verifiably robust models, Ar Xiv preprint, vol. abs/1810.12715, 2018. [22] M. Andriushchenko and M. Hein, Provably robust boosted decision stumps and trees against adversarial attacks, in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019. [23] Y. Wang, H. Zhang, H. Chen, D. S. Boning, and C. Hsieh, On lp-norm robustness of ensemble decision stumps and trees, in Proc. of ICML, vol. 119, 2020. [24] J. M. Cohen, E. Rosenfeld, and J. Z. Kolter, Certified adversarial robustness via randomized smoothing, in Proc. of ICML, vol. 97, 2019. [25] G. Yang, T. Duan, J. E. Hu, H. Salman, I. P. Razenshteyn, and J. Li, Randomized smoothing of all shapes and sizes, in Proc. of ICML, vol. 119, 2020. [26] D. Zhang, M. Ye, C. Gong, Z. Zhu, and Q. Liu, Black-box certification with randomized smoothing: A functional optimization based framework, in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. [27] K. D. Dvijotham, J. Hayes, B. Balle, Z. Kolter, C. Qin, A. György, K. Xiao, S. Gowal, and P. Kohli, A framework for robustness certification of smoothed classifiers using f-divergences, in Proc. of ICLR, 2020. [28] J. Neyman and E. S. Pearson, Ix. on the problem of the most efficient tests of statistical hypotheses, Philosophical Transactions of the Royal Society of London. Series A, Containing Papers of a Mathematical or Physical Character, vol. 231, no. 694-706, 1933. [29] M. Z. Horváth, M. N. Müller, M. Fischer, and M. T. Vechev, Boosting randomized smoothing with variance reduced classifiers, in Proc. of ICLR, 2022. [30] G. Lee, Y. Yuan, S. Chang, and T. S. Jaakkola, Tight certificates of adversarial robustness for randomly smoothed classifiers, in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019. [31] A. Bojchevski, J. Klicpera, and S. Günnemann, Efficient robustness certificates for discrete data: Sparsity-aware randomized smoothing for graphs, images and more, in Proc. of ICML, vol. 119, 2020. [32] B. Bustos, D. A. Keim, D. Saupe, T. Schreck, and D. V. Vranic, Using entropy impurity for improved 3d object similarity search, in Proceedings of the 2004 IEEE International Conference on Multimedia and Expo, ICME 2004, 27-30 June 2004, Taipei, Taiwan, 2004. [33] J. H. Friedman, Greedy function approximation: a gradient boosting machine, Annals of statistics, 2001. [34] Y. Freund and R. E. Schapire, A decision-theoretic generalization of on-line learning and an application to boosting, J. Comput. Syst. Sci., vol. 55, no. 1, 1997. [35] A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, A. Desmaison, A. Köpf, E. Yang, Z. De Vito, M. Raison, A. Tejani, S. Chilamkurthy, B. Steiner, L. Fang, J. Bai, and S. Chintala, Pytorch: An imperative style, high-performance deep learning library, in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019. [36] J. W. Smith, J. E. Everhart, W. C. Dickson, W. C. Knowler, and R. S. Johannes, Using the adap learning algorithm to forecast the onset of diabetes mellitus, in Annual Symposium on Computer Application in Medical Care, 1988. [37] D. Dua and C. Graff, UCI machine learning repository, 2017. [38] H. X. an, Fashion-mnist: a novel image dataset for benchmarking machine learnin, Ar Xiv preprint, vol. abs/1708.07747, 2017. [39] Y. Le Cun and C. Cortes, MNIST handwritten digit database, 1998. [40] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vander Plas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay, Scikit-learn: Machine learning in python, Co RR, vol. abs/1201.0490, 2012. [Online]. Available: http://arxiv.org/abs/1201.0490 [41] B. Li, C. Chen, W. Wang, and L. Carin, Certified adversarial robustness with additive noise, in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019. [42] M. Lécuyer, V. Atlidakis, R. Geambasu, D. Hsu, and S. Jana, Certified robustness to adversarial examples with differential privacy, in 2019 IEEE Symposium on Security and Privacy, SP 2019, San Francisco, CA, USA, May 19-23, 2019, 2019. [43] Z. Gao, R. Hu, and Y. Gong, Certified robustness of graph classification against topology attack with randomized smoothing, in IEEE Global Communications Conference, GLOBECOM 2020, Virtual Event, Taiwan, December 7-11, 2020, 2020. [44] P. Chiang, M. J. Curry, A. Abdelkader, A. Kumar, J. Dickerson, and T. Goldstein, Detection as regression: Certified object detection with median smoothing, in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. [45] M. Fischer, M. Baader, and M. T. Vechev, Scalable certified segmentation via randomized smoothing, in Proc. of ICML, vol. 139, 2021. [46] J. Jia, X. Cao, B. Wang, and N. Z. Gong, Certified robustness for top-k predictions against adversarial perturbations via randomized smoothing, in Proc. of ICLR, 2020. [47] M. Fischer, M. Baader, and M. T. Vechev, Certified defense to image transformations via randomized smoothing, in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. [48] L. Li, M. Weber, X. Xu, L. Rimanic, B. Kailkhura, T. Xie, C. Zhang, and B. Li, Tss: Transformation-specific smoothing for robustness certification, in ACM CCS, 2021. [49] B. Wang, J. Jia, X. Cao, and N. Z. Gong, Certified robustness of graph neural networks against adversarial structural perturbation, in KDD 21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, Singapore, August 14-18, 2021, 2021. [50] J. Schuchardt, A. Bojchevski, J. Klicpera, and S. Günnemann, Collective robustness certificates, in International Conference on Learning Representations, 2021. [51] A. Levine and S. Feizi, Robustness certificates for sparse adversarial attacks by randomized ablation, in The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 2020. [52] , (de)randomized smoothing for certifiable defense against patch attacks, in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. [53] , Improved, deterministic smoothing for l1 certified robustness, in Proc. of ICML, vol. 139, 2021. [54] M. Z. Horváth, M. N. Müller, M. Fischer, and M. Vechev, Robust and accurate - compositional architectures for randomized smoothing, in ICLR 2022 Workshop on Socially Responsible Machine Learning, 2022. [55] J. Jeong and J. Shin, Consistency regularization for certified robustness of smoothed classifiers, in Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. [56] R. Zhai, C. Dan, D. He, H. Zhang, B. Gong, P. Ravikumar, C. Hsieh, and L. Wang, MACER: attack-free and scalable robust training via maximizing certified radius, in Proc. of ICLR, 2020. [57] H. Salman, J. Li, I. P. Razenshteyn, P. Zhang, H. Zhang, S. Bubeck, and G. Yang, Provably robust deep learning via adversarially trained smoothed classifiers, in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019. [58] Z. Yang, L. Li, X. Xu, B. Kailkhura, T. Xie, and B. Li, On the certified robustness for ensemble models and beyond, in The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022, 2022. [59] S. Calzavara, C. Lucchese, G. Tolomei, S. A. Abebe, and S. Orlando, Treant: training evasionaware decision trees, Data Min. Knowl. Discov., vol. 34, no. 5, 2020. [60] D. Vos and S. Verwer, Efficient training of robust decision trees against adversarial examples, in Proc. of ICML, vol. 139, 2021. [61] J.-Q. Guo, M.-Z. Teng, W. Gao, and Z.-H. Zhou, Fast provably robust decision trees and boosting, in Proceedings of the 39th International Conference on Machine Learning, 2022. [62] H. Chen, H. Zhang, S. Si, Y. Li, D. S. Boning, and C. Hsieh, Robustness verification of tree-based models, in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019. [63] F. Ranzato and M. Zanella, Abstract interpretation of decision tree ensemble classifiers, in The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, 2020. [64] J. Törnblom and S. Nadjm-Tehrani, An abstraction-refinement approach to formal verification of tree ensembles, in Computer Safety, Reliability, and Security - SAFECOMP 2019 Workshops, ASSURE, DECSo S, SASSUR, STRIVE, and WAISE, Turku, Finland, September 10, 2019, Proceedings, vol. 11699, 2019. [65] L. Schott, J. Rauber, M. Bethge, and W. Brendel, Towards the first adversarially robust neural network model on MNIST, in Proc. of ICLR, 2019. [66] F. Tramèr and D. Boneh, Adversarial training and robustness for multiple perturbations, in Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 2019. [67] A. Kantchelian, J. D. Tygar, and A. D. Joseph, Evasion and hardening of tree ensemble classifiers, in Proc. of ICML, vol. 48, 2016. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] We substantiate all claims theoretically (Section 3 and Section 4) or empirically (Section 5), as appropriate. (b) Did you describe the limitations of your work? [Yes] We describe the limitations in Section 7. (c) Did you discuss any potential negative societal impacts of your work? [Yes] We discuss societal impacts in Section 7. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] See Section 3, Section 4, and App. A. (b) Did you include complete proofs of all theoretical results? [Yes] See Section 3, Section 4, and App. A. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] We will include them as supplemental material and will release them to the public upon publication. All datasets we use are publicly available. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] We provide full training details in App. B. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] Both our training and certification methods are deterministic (up to floating point errors), thus our main experiments do not contain error bars. However, in App. C we report numerous error bars with respect to the data split via 5-fold cross-validation. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] We provide resource type details in Section 5 and timing details in App. B. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [Yes] (c) Did you include any new assets either in the supplemental material or as a URL? [N/A] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]