# multiscale_diffusion_denoised_smoothing__52161c16.pdf Multi-scale Diffusion Denoised Smoothing Jongheon Jeong Jinwoo Shin Korea Advanced Institute of Science and Technology (KAIST) Daejeon, South Korea {jongheonj, jinwoos}@kaist.ac.kr Along with recent diffusion models, randomized smoothing has become one of a few tangible approaches that offers adversarial robustness to models at scale, e.g., those of large pre-trained models. Specifically, one can perform randomized smoothing on any classifier via a simple denoise-and-classify pipeline, so-called denoised smoothing, given that an accurate denoiser is available such as diffusion model. In this paper, we present scalable methods to address the current trade-off between certified robustness and accuracy in denoised smoothing. Our key idea is to selectively apply smoothing among multiple noise scales, coined multi-scale smoothing, which can be efficiently implemented with a single diffusion model. This approach also suggests a new objective to compare the collective robustness of multi-scale smoothed classifiers, and questions which representation of diffusion model would maximize the objective. To address this, we propose to further finetune diffusion model (a) to perform consistent denoising whenever the original image is recoverable, but (b) to generate rather diverse outputs otherwise. Our experiments show that the proposed multi-scale smoothing scheme, combined with diffusion fine-tuning, not only allows strong certified robustness at high noise scales but also maintains accuracy close to non-smoothed classifiers. Code is available at https://github.com/jh-jeong/smoothing-multiscale. 1 Introduction Arguably, one of the important lessons in modern deep learning is the effectiveness of massive data and model scaling [38, 74], which enabled many breakthroughs in recent years [8, 53, 55, 7]. Even with the largest amount of data available on the web and computational budgets, however, the worst-case behaviors of deep learning models are still challenging to regularize. For example, large language models often leak private information in training data [10], and one can completely fool superhuman-level Go agents with few meaningless moves [40]. Such unintentional behaviors can be of critical concerns in deploying deep learning into real-world systems, and the risk has been increasing as the capability of deep learning continues to expand. In this context, adversarial robustness [61, 9] has been a seemingly-close milestone towards reliable deep learning. Specifically, neural networks are even fragile to small, imperceptible scales of noise when it comes to the worst-case, and consequently there have been many efforts to obtain neural networks that are robust to such noise [47, 3, 75, 15, 65]. Although it is a reasonable premise that we humans have an inherent mechanism to correct against adversarial examples [21, 26], yet so far the threat still persists in the context of deep learning, e.g., even in recent large vision-language models [48], and is known as hard to avoid without a significant reduction in model performance [66, 64]. Randomized smoothing [41, 15], the focus in this paper, is currently one of a few techniques that has been successful in obtaining adversarial robustness from neural networks. Specifically, it constructs a smoothed classifier by taking a majority vote from a base classifier, e.g., a neural network, over random Gaussian input noise. The technique is notable for its provable guarantees on adversarial 37th Conference on Neural Information Processing Systems (Neur IPS 2023). (a) Cascaded smoothing (b) Diffusion calibration Figure 1: An overview of the proposed approaches, (a) cascaded randomized smoothing (Section 3.2 and (b) diffusion calibration (Section 3.3) to attain a better trade-off between accuracy and certified robustness in randomized smoothing, upon on the recent diffusion denoised smoothing scheme [11]. robustness, i.e., it is a certified defense [44], and its scalability to arbitrary model architectures. For example, it was the first certified defense that could offer adversarial robustness on Image Net [56]. The scalablity of randomized smoothing has further expanded by Salman et al. [58], which observed that randomized smoothing can be applied to any pre-trained classifiers by prepending a denoiser model, dubbed denoised smoothing. Combined with the recent diffusion-based models [31], denoised smoothing could provide the current state-of-the-arts in ℓ2-certified robustness [11, 69, 44]. Despite its desirable properties, randomized smoothing in practice is also at odds with model accuracy [15], similarly to other defenses [47, 76], which makes it burdening to be applied in the real-world. For example, the variance of Gaussian noise, or smoothing factor, is currently a crucial hyperparameter to increase certified robustness at the cost of accuracy. The fundamental trade-off between accuracy and adversarial robustness has been well-evidenced in the literature [66, 75, 64], but it has been relatively under-explored whether they may or may not be applicable in the context of randomized smoothing. For example, Tramèr et al. [64] demonstrate that pursuing ε-uniform adversarial robustness in neural networks may increase their vulnerability against invariance attacks, i.e., some semantics-altering perturbations possible inside the ε-ball: yet, it is still unclear whether such a vulnerability would be inevitable at non-uniformly robust models such as smoothed classifiers. In attempt to understand the accuracy-robustness trade-off in randomized smoothing, efforts have been made to increase the certified robustness a given smoothed classifier can provide, e.g., through a better training method [58, 73, 35 37]. For example, Salman et al. [57] have employed adversarial training [47] in the context of randomized smoothing, and Jeong and Shin [35] have proposed to train a classifier with consistency regularization. Motivated by the trade-off among different smoothing factors, some works have alternatively proposed to perform smoothing with input-dependent factors [1, 13, 67]: unfortunately, subsequent works [60, 36] have later shown that such schemes do not provide valid robustness certificates, reflecting its brittleness in overcoming the trade-off. Contribution. In this paper, we develop a practical method to overcome the current frontier of accuracy-robustness trade-off in randomized smoothing, particularly upon the architectural benefits that the recent diffusion denoised smoothing [11] can offer. Specifically, we first propose to aggregate multiple smoothed classifiers of different smoothing factors to obtain their collective robustness (see Figure 1(a)), which leverages the scale-free nature of diffusion denoised smoothing. In this way, the model can decide which smoothed classifier to use for each input, maintaining the overall accuracy. Next, we fine-tune a diffusion model for randomized smoothing through the denoise-and-classify pipeline (see Figure 1(b)), as an efficient alternative of the full fine-tuning of (potentially larger) pretrained classifiers. Here, we identify the over-confidence of diffusion models, e.g., to a specific class given certain backgrounds, as a major challenge towards accurate-yet-robust randomized smoothing, and design a regularization objective to mitigate the issue. In our experiments, we evaluate our proposed schemes on CIFAR-10 [39] and Image Net [56], two of standard benchmarks for certified ℓ2-robustness, particularly considering practical scenarios of applying diffusion denoised smoothing to large pre-trained models such as CLIP [53]. Overall, the results consistently highlight that (a) the proposed multi-scale smoothing scheme upon the recent diffusion denoised smoothing [11] can significantly improve the accuracy of smoothed inferences while maintaining their certified robustness at larger radii, and (b) our fine-tuning scheme of diffusion models can additively improve the results - not only in certified robustness but also in accuracy - by simply replacing the denoiser model without any further adaptation. For example, we could improve certifications from a diffusion denoised smoothing based classifier by 30.6% 72.5% in clean accuracy, while also improving its certified robustness at ε = 2.0 by 12.6% 14.1%. We observe that the collective robustness from our proposed multi-scale smoothing does not always correspond to their individual certified robustness, which has been a major evaluation in the literature, but rather to their calibration across models: which opens up a new direction to pursue for a practical use of randomized smoothing. 2 Preliminaries Adversarial robustness and randomized smoothing. For a given classifier f : X Y, where x X Rd and y Y := {1, , K}, adversarial robustness refers to the behavior of f in making consistent predictions at the worst-case perturbations under semantic-preserving restrictions. Specifically, for samples from a data distribution (x, y) pdata(x, y), it requires f(x + δ) = y for every perturbation δ that a threat model defines, e.g., an ℓ2-ball δ 2 ε. One of ways to quantify adversarial robustness is to measure the following average minimum-distance of adversarial perturbations [9], i.e., R(f; pdata) := E(x,y) pdata minf(x ) =y ||x x||2 . The essential challenge in achieving adversarial robustness stems from that evaluating this (and further optimizing on it) is usually infeasible. Randomized smoothing [41, 15] bypasses this difficulty by constructing a new classifier ˆf from f instead of letting f to directly model the robustness: specifically, it transforms the base classifier f with a certain smoothing measure, where in this paper we focus on the case of Gaussian distributions N(0, σ2I): ˆf(x) := arg max c Y Pδ N(0,σ2I) [f(x + δ) = c] . (1) Then, the robustness of ˆf at (x, y), namely R( ˆf; x, y), can be lower-bounded in terms of the certified radius R( ˆf, x, y), e.g., Cohen et al. [15] showed that the following bound holds, which is tight for ℓ2-adversarial threat models: R( ˆf; x, y) σ Φ 1(p ˆ f(x, y)) =: R( ˆf, x, y), where p ˆ f(x, y) := Pδ[f(x + δ) = y], (2) provided that ˆf(x) = y, otherwise R( ˆf; x, y) := 0.1 Here, we remark that the formula for certified radius is essentially a function of p ˆ f, which is the accuracy of f(x + δ) over δ. Denoised smoothing. Essentially, randomized smoothing requires f to make accurate classification of Gaussian-corrupted inputs. A possible design of f in this regard is to concatenate a Gaussian denoiser, say denoise( ), with any standard classifier fstd, so-called denoised smoothing [58]: f(x + δ) := fstd(denoise(x + δ)). (3) Under this design, an ideal denoiser denoise( ) should accurately recover x from x + δ, i.e., denoise(x + δ) x (in terms of their semantics to perform classification) with high probability of δ N(0, σ2I). Denoised smoothing offers a more scalable framework for randomized smoothing, considering that (a) standard classifiers (rather than those specialized to Gaussian noise) are nowadays easier to obtain in the paradigm of large pre-trained models, and (b) the recent developments in diffusion models [31] has supplied denoisers strong enough for the framework. In particular, Lee [42] has firstly explored the connection between diffusion models and randomized smoothing; Carlini et al. [11] has further observed that latest diffusion models combined with a pre-trained classifier provides a state-of-the-art design of randomized smoothing. Diffusion models. In principle, diffusion models [59, 31] aims to generate a given data distribution pdata(x) via an iterative denoising process from a Gaussian noise ˆx T N(0, T 2I) for a certain T > 0. Specifically, it first assumes the following diffusion process which maps pdata to N(0, T 2I): dxt = µ(xt, t)dt+σ(t)dwt, where t [0, T], and wt denotes the standard Brownian motion. Based 1Φ denotes the cumulative distribution function of N(0, 12). on this, diffusion models first train a score model sϕ(x, t) log pt(x) via score matching [34], and use the model to solve the probabilistic flow from ˆx T N(0, T 2I) to x0 for sampling. The score estimator sϕ(x, t) is often parametrized by a denoiser D(x; σ(t)) in practice, viz., log pt(x) = (D(x; σ(t)) x)/σ(t)2, which establishes its close relationship with denoised smoothing. Consider a classification task from X to Y where training data D = {(xi, yi)} pdata(x, y) is available, and let f : X Y be a classifier. We denote ˆfσ to be the smoothed classifier of f with respect to the smoothing factor σ > 0, as defined by (1). In this work, we aim to better understand how the accuracy-robustness trade-off of ˆfσ occurs, with a particular consideration of the recent denoised smoothing scheme. Generally speaking, the trade-off implies the following: for a given model, there exists a sample that the model gets wrong as it is optimized for (adversarial) robustness on another sample. In this respect, we start by taking a closer look at what it means by a model gets wrong, particularly when it is from randomized smoothing. 3.1 Over-smoothing and over-confidence in randomized smoothing Consider a smoothed classifier ˆfσ, and suppose there exists a sample (x, y) where ˆfσ makes an error: i.e., ˆfσ(x) = y. Our intuition here is to separate possible scenarios of ˆfσ(x) making an error into two distinct cases, based on the prediction confidence of ˆfσ(x). Specifically, we define the confidence of a smoothed classifier ˆfσ at x based on the definition of randomized smoothing (1) and (2): p ˆ fσ(x) := max y p ˆ fσ(x, y) = max y Pδ N(0,σ2I)[f(x + δ) = y]. (4) Intuitively, this notion of smoothed confidence measures how consistent the base classifier f is in classifying x + δ over Gaussian noise δ N(0, σ2I). In cases when f is modeled by denoised smoothing, achieving high confidence requires the denoiser D to accurately bounce-back a given noisy image x + δ into one that falls into class y with high probability over δ. Given a smoothed confidence p := p ˆ fσ(x), we propose to distinguish two cases of model errors, which are both peculiar to randomized smoothing, by considering a certain threshold p0 on p. Namely, we interpret an error of ˆfσ at x either as (a) p p0: the model is over-smoothing, or (b) p > p0: the model is having an over-confidence on the input: 1. Over-smoothing (p p0): On one hand, it is unavoidable that the mutual information I(x + δ; y) between input and its class label absolutely decrease from smoothing with larger variance σ2, although using larger σ can increase the maximum certifiable radius of ˆfσ in practice (2). Here, a well-calibrated smoothed classifier should output a prediction close to the uniform distribution across Y, leading to a low prediction confidence. In terms of denoised smoothing, the expectation is clearer: as x + δ gets closer to the pure Gaussian noise, the denoiser D should generate more diverse outputs hence in Y as well. Essentially, this corresponds to an accuracy-robustness trade-off from choosing a specific σ for a given (e.g., information-theoretic) capacity of data. 2. Over-confidence (p > p0): On the other hand, it is also possible for a model ˆfσ to be incorrect but with a high confidence. Compared to the over-smoothing case, this scenario rather signals a miscalibration and corresponds to a trade-off from model biases: even across smoothed models with a fixed σ, the balance between accuracy and robustness can be different depending on how each model assigns robustness in its decision boundary per-sample basis. When viewed in terms of denoised smoothing, this occurrence can reveal an implicit bias of the denoiser function D, e.g., that of diffusion models. For example, a denoiser might be trained to adapt to some spurious cues in training data, e.g., their backgrounds, as also illustrated in Figure 1(b). In the subsequent sections, Section 3.2 and 3.3, we introduce two methods to exhibit better accuracyrobustness trade-off in randomized smoothing, each of which focuses on the individual scenarios of over-smoothing and over-confidence, respectively. Specifically, Section 3.2 proposes to use a cascaded inference of multiple smoothed classifiers across different smoothing factors to mitigate the limit of using a single smoothing factor. Next, in Section 3.3, we propose to calibrate diffusion models to reduce its over-confidence particularly in denoised smoothing. 3.2 Cascaded randomized smoothing To overcome the trade-off between accuracy and certified robustness from over-smoothing, i.e., from choosing a specific σ, we propose to combine multiple smoothed classifiers with different σ s. In a nutshell, we design a pipeline of smoothed inferences that each input (possibly with different noise resilience) can adaptively select which model to use for its prediction. Here, the primary challenge is to make it correct , so that the proposed pipeline does not break the existing statistical guarantees on certified robustness that each smoothed classifier makes. Specifically, we now assume K distinct smoothing factors, say 0 < σ1 < < σK, and their corresponding smoothed classifiers of f, namely ˆfσ1, , ˆfσK. For a given input x, our desiderata is (a) to maximize robustness certification at x based on the smoothed inferences available from the individual models, say p ˆ fσ1(x), , p ˆ fσK (x), while (b) minimizing the access to each of the models those require a separate Monte Calro integration in practice. In these respects, we propose a simple predict-or-abstain policy, coined cascaded randomized smoothing:2 casc(x; { ˆfσi}K i=1) := ˆfσK(x) if p ˆ fσK (x) > p0, casc(x; { ˆfσi}K 1 i=1 ) if p ˆ fσK (x) p0 and K > 1, ABSTAIN otherwise, where ABSTAIN / Y denotes an artificial class to indicate undecidable . Intuitively, the pipeline starts from computing ˆfσK(x), the model with highest σ, but takes its output only if its (smoothed) confidence p ˆ fσK (x) exceeds a certain threshold p0:3 otherwise, it tries a smaller noise scale, say σK 1 and so on, applying the same abstention policy of p ˆ fσ (x) p0. In this way, it can early-stop the computation at higher σ if it is confident enough, so it can maintain higher certified robustness, while avoiding unnecessary accesses to other models of smaller σ. Next, we ask whether this pipeline can indeed provide a robustness certification: i.e., how much one can ensure casc(x + δ) = casc(x) in its neighborhood δ. Theorem 3.1 below shows that one can indeed enjoy the most certified radius from ˆfσk where casc(x) =: ˆy halts, as long as the preceding models are either keep abstaining or output ˆy over δ:4 Theorem 3.1. Let ˆfσ1, , ˆfσK : X Y be smoothed classifiers with 0 < σ1 < < σK. Suppose casc(x; { ˆfσi}K i=1) =: ˆy Y halts at ˆfσk for some k. Consider any p and pi,c [0, 1] that satisfy the following: (a) p p ˆ fσk (x, ˆy), and (b) pk ,c p ˆ fσk (x, c) for k > k and c Y. Then, it holds that casc(x + δ; { ˆfσi}K i=1) = ˆy for any δ < R, where: , min y =ˆy k >k σk Φ 1 1 pk ,y Overall, the proposed multi-scale smoothing scheme (and Theorem 3.1) raise the importance of abstaining well in randomized smoothing: if a smoothed classifier can perfectly detect and abstain its potential errors, one could overcome the trade-off between accuracy and certified robustness by joining a more accurate model afterward. The option to abstain in randomized smoothing was originally adopted to make its statistical guarantees correct in practice. Here, we extend this usage to also rule out less-confident predictions for a more conservative decision making. As discussed in Section 3.1, now the over-confidence becomes a major challenge in this matter: e.g., such samples can potentially bypass the abstention policy of cascaded smoothing, which motivates our fine-tuning scheme presented in Section 3.3. Certification. We implement our proposed cascaded smoothing to make statistically consistent predictions across different noise samples, considering a certain significance level α (e.g., α = 0.001): in a similar fashion as Cohen et al. [15]. Roughly speaking, for a given input x, it makes predictions only when the (1 α)-confidence interval of p ˆ f(x) does not overlap with p0 upon n i.i.d. noise samples (otherwise it abstains). The more details can be found in Appendix D.2. 2We also discuss several other possible (and more sophisticated) designs in Appendix E. 3In our experiments, we simply use p0 = 0.5 for all σ s. See Appendix C.3 for an ablation study with p0. 4The proof of Theorem 3.1 is provided in Appendix D.1. 3.3 Calibrating diffusion models through smoothing Next, we move on to the over-confidence issue in randomized smoothing: viz., ˆfσ often makes errors with a high confidence to wrong classes. We propose to fine-tune a given ˆfσ to make rather diverse outputs when it misclassifies, i.e., towards a more calibrated ˆfσ. By doing so, we aim to cast the issue of over-confidence as that of over-smoothing: which can be easier to handle in practice, e.g., by abstaining. In this work, we particularly focus on fine-tuning only the denoiser model D in the context of denoised smoothing, i.e., f := fstd D for a standard classifier fstd: this offers a more scalable approach to improve certified robustness of pre-trained models, given that fine-tuning the entire classifier model in denoised smoothing can be computationally prohibitive in practice. Specifically, given a base classifier f := fstd D and training data D, we aim to fine-tune D to improve the certified robustness of ˆfσ. To this end, we leverage the confidence information of the backbone classifier fstd fairly assuming it as an oracle - which became somewhat reasonable given the recent off-the-shelf models available - and apply different losses depending on the confidence information per-sample basis. We propose two losses in this matter: (a) Brier loss for either correct or under-confident samples, and (b) anti-consistency loss for incorrect, over-confident samples. Brier loss. For a given training sample (x, y), we adopt the Brier (or squared ) loss [6] to regularize the denoiser function D to promote the confidence of fstd(D(x + δ)) towards y, which can be beneficial to increase the smoothed confidence p ˆ fσ(x) that impacts the certified robustness at x. Compared to the cross-entropy (or log ) loss, a more widely-used form in such purpose, we observe that the Brier loss can be favorable in such fine-tuning of D through fstd, in a sense that the loss is less prone to over-optimize the confidence at values closer to 1. Here, an important detail is that we do not apply the regularization to incorrect-yet-confident samples: i.e., whenever fstd(D(x + δ)) = y and pstd(δ) := maxc Fstd,c(D(x + δ)) > p0, where Fstd is the soft prediction of fstd. This corresponds to the case when D(x + δ) rather outputs a realistic off-class sample, which will be handled by the anti-consistency loss we propose. Overall, we have: LBrier(x, y) := Eδ[1[ˆyδ = y or pstd(δ) p0] Fstd(D(x + δ)) ey 2], (7) where we denote ˆyδ := f(x + δ) and ey is the y-th unit vector in R|Y|. Anti-consistency loss. On the other hand, the anti-consistency loss aims to detect whether the sample is over-confident, and penalizes it accordingly. The challenge here is that identifying over-confidence in a smoothed classifier requires checking for p ˆ fσ(x) > p0 (4), which can be infeasible during training. We instead propose a simpler condition to this end, which only takes two independent Gaussian noise, say δ1, δ2 N(0, σ2I). Specifically, we identify (x, y) as over-confident whenever (a) ˆy1 := f(x + δ1) and ˆy2 := f(x + δ2) match, while (b) they are incorrect, i.e., ˆy1 = y. Intuitively, such a case signals that the denoiser D is often making a complete flip to the semantics of x + δ (see Figure 1(b) for an example), which we aim to penalize. A care should be taken, however, considering the possibility that D(x + δ) indeed generates an in-distribution sample that falls into a different class: in this case, penalizing it may result in a decreased robustness of that sample. In these respects, our design of anti-consistency loss forces the two samples simply to have different predictions, by keeping at least one prediction as the original, while penalizing the counterpart. Denoting p1 := Fstd(D(x + δ1)) and p2 := Fstd(D(x + δ2)), we again apply the squared loss on p1 and p2 to implement the loss design, as the following: LAC(x, y) := 1[ˆy1 = ˆy2 and ˆy1 = y] ( p1 sg(p1) 2 + p2 2), (8) where sg( ) denotes the stopping gradient operation. Overall objective. Combining the two proposed losses, i.e., the Brier loss and anti-consistency loss, defines a new regularization objective to add upon any pre-training objective for the denoiser D: L(D) := LDenoiser + λ (LBrier + α LAC), (9) where λ, α > 0 are hyperparameters. Here, α denotes the relative strength of LAC over LBrier in its regularization.5 Remark that increasing α would give more penalty on over-confident samples, which would lead the model to make more abstentions: therefore, this results in an increased accuracy particularly in cascaded smoothing (Section 3.2). 5In our experiments, we simply use α = 1 unless otherwise specified. Carlini et al. (σ=0.25) Carlini et al. (σ=0.50) Carlini et al. (σ=1.00) + Cascading (ours) Certified accuracy Radius 0 1 2 3 (a) CIFAR-10 Carlini et al. (σ=0.25) Carlini et al. (σ=0.50) Carlini et al. (σ=1.00) + Cascading (ours) Certified accuracy Radius 0 1 2 3 (b) Image Net Figure 2: Comparison of certified test accuracy with Carlini et al. [11] on the effectiveness of cascaded smoothing. Gaussian Consistency Carlini et al. (2023) + Calibration (ours) Certified accuracy Radius 0 1 2 3 Figure 3: Comparison of certified test accuracy on CIFAR-10 at σ {0.25, 0.50, 1.00}. Table 1: Comparison of (a) certified accuracy, (b) empricial accuracy, and (c) average certified radius (ACR) on CIFAR-10. We set our result bold-faced whenever it achieves the best upon baselines. CIFAR-10, Vi T-B/16@224 Empirical accuracy (%) Certified accuracy at ε (%) σ Training C10 C10-C C10.1 0.0 0.5 1.0 1.5 2.0 ACR Gaussian [15] 92.7 88.3 76.7 87.0 55.3 0.479 Consistency [35] 89.5 84.7 67.6 84.9 60.7 0.517 Carlini et al. [11] 92.8 88.8 67.6 89.5 59.8 0.527 + Calibration (ours) 94.5 89.4 69.2 90.3 61.9 0.535 Gaussian [15] 87.4 81.8 71.0 69.9 45.9 24.4 6.0 0.543 Consistency [35] 80.7 75.0 59.6 67.7 49.2 31.2 13.1 0.617 Carlini et al. [11] 88.8 84.9 64.3 75.3 50.6 31.0 14.3 0.646 + Calibration (ours) 88.6 84.2 63.7 75.1 53.3 32.5 15.3 0.664 (0.25-0.50) + Cascading (ours) 92.5 89.1 78.5 85.1 53.3 32.9 14.9 0.680 Gaussian [15] 75.6 72.4 68.3 42.2 27.9 16.9 9.0 4.0 0.426 Consistency [35] 65.8 61.3 53.6 43.1 31.1 22.3 14.6 8.7 0.513 Carlini et al. [11] 87.3 83.2 73.0 48.6 32.5 21.9 11.8 6.4 0.530 + Calibration (ours) 83.3 79.3 60.9 52.6 36.7 24.9 16.1 8.8 0.577 (0.25-1.00) + Cascading (ours) 90.1 86.9 74.5 79.6 40.5 25.0 16.2 8.8 0.645 4 Experiments We verify the effectiveness of our proposed schemes, (a) cascaded smoothing and (b) diffusion calibration, mainly on CIFAR-10 [39] and Image Net [56]: two standard datasets for an evaluation of certified ℓ2-robustness. We provide the detailed experimental setups, e.g., training, datasets, hyperparameters, computes, etc., in Appendix B. Baselines. Our evaluation mainly compares with diffusion denoised smoothing [11], the current state-of-the-art methodology in randomized smoothing. We additionally compare with two other training baselines from the literature, by considering models with the same classifier architecture but without the denoising step of Carlini et al. [11]. Specifically, we consider (a) Gaussian training [15], which trains a classifier with Gaussian augmentation; and (b) Consistency [35], which additionally regularizes the variance of predictions over Gaussian noise in training. We select the baselines assuming practical scenarios where the training cost of the classifier side is crucial: other existing methods for smoothed classifiers often require much more costs, e.g., 8 times over Gaussian [57, 73]. Setups. We follow Carlini et al. [11] for the choice of diffusion models: specifically, we use the 50M-parameter 32 32 diffusion model from Nichol and Dhariwal [52] for CIFAR-10, and the 552Mparameter 256 256 unconditional model from Dhariwal and Nichol [18] for Image Net. For the classifier side, we use Vi T-B/16 [63] pre-trained via CLIP [53] throughout our experiments. For uses we fine-tune the model on each of CIFAR-10 and Image Net via FT-CLIP [19], resulting in classifiers that achieve 98.1% and 85.2% in top-1 accuracy on CIFAR-10 and Image Net, respectively. Following the prior works, we mainly consider σ {0.25, 0.50, 1.00} for smoothing in our experiments. Evaluation metrics. We consider two popular metrics in the literature when evaluating certified robustness of smoothed classifiers: (a) the approximate certified test accuracy at r: the fraction of the test set which CERTIFY [15] classifies correctly with the radius larger than r without abstaining, Table 2: Comparison of (a) certified accuracy, (b) empricial accuracy, and (c) average certified radius (ACR) on Image Net. We set our result bold-faced whenever it achieves the best upon baselines. Image Net, Vi T-B/16@224 Empirical accuracy (%) Certified accuracy at ε (%) Training σ IN-1K IN-R IN-A 0.0 0.5 1.0 1.5 2.0 ACR Carlini et al. [11] 0.25 80.9 69.3 35.3 78.8 61.4 0.517 Carlini et al. [11] 0.50 79.5 67.5 32.3 65.8 52.2 38.6 23.6 0.703 + Cascading (ours) (0.25-0.50) 83.5 69.6 41.3 75.0 52.6 39.0 22.8 0.720 + Calibration (ours) (0.25-0.50) 83.8 69.8 41.7 76.6 54.6 39.8 23.0 0.743 Carlini et al. [11] 1.00 77.2 64.3 32.0 30.6 25.8 20.6 17.0 12.6 0.538 + Cascading (ours) (0.25-1.00) 82.6 69.8 40.5 69.0 42.4 26.6 19.0 14.6 0.752 + Calibration (ours) (0.25-1.00) 83.2 69.5 40.8 72.5 44.0 27.5 19.9 14.1 0.775 and (b) the average certified radius (ACR) [73]: the average of certified radii on the test set Dtest while assigning incorrect samples as 0: viz., ACR := 1 |Dtest| P (x,y) Dtest[CR(f, σ, x) 1 ˆ f(x)=y], where CR( ) denotes the certified radius that CERTIFY returns. Throughout our experiments, we use n = 10, 000 noise samples to certify robustness for both CIFAR-10 and Image Net. We follow [15] for the other hyperparameters to run CERTIFY, namely by n0 = 100, and α = 0.001. In addition to certified accuracy, we also compare the empirical accuracy of smoothed classifiers. Here, we define empirical accuracy by the fraction of test samples those are either (a) certifiably correct, or (b) abstained but correct in the clean classifier: which can be a natural alternative especially at denoised smoothing. For this comparison, we use n = 100 to evaluate empirical accuracy. Unlike the evaluation of Carlini et al. [11], however, we do not compare cascaded smoothing with the envelop accuracy curve over multiple smoothed classifiers at σ {0.25, 0.50, 1.00}: although the envelope curve can be a succinct proxy to compare methods, it can be somewhat misleading and unfair to compare the curve directly with an individual smoothed classifier. This is because the curve does not really construct a concrete classifier on its own: it additionally assumes that each test sample has prior knowledge on the value of σ {0.25, 0.5, 1.0} to apply, which is itself challenging to infer. This is indeed what our proposal of cascaded smoothing addresses. 4.1 Results Cascaded smoothing. Figure 2 visualizes the effect of cascaded smoothing we propose in the plots of certified accuracy, and the detailed results are summarized in Table 1 and 2 as + Cascading . Overall, we observe that the certified robustness that cascaded smoothing offers can be highly desirable over the considered single-scale smoothed classifiers. Compared to the single-scale classifiers at highest σ, e.g., σ = 1.0 in Table 1 and 2, our cascaded classifiers across σ {0.25, 0.50, 1.00} absolutely improve the certified accuracy at all the range of ε by incorporating more accurate predictions from classifiers, e.g., of lower σ {0.25, 0.50}. On the opposite side, e.g., compared to Carlini et al. [11] at σ = 0.25, the cascaded classifiers provide competitive certified clean accuracy (ε = 0), e.g., 89.5% vs. 85.1% on CIFAR-10, while being capable of offering a wider range of robustness certificates. Those considerations are indeed reflected quantitatively in terms of the improvements in ACRs. The existing gaps in the clean accuracy are in principle due to the errors in higher-σ classifiers: i.e., a better calibration, to let them better abstain, could potentially reduce the gaps. Diffusion calibration. Next, we evaluate the effectiveness of our proposed diffusion fine-tuning scheme: + Calibration in Table 1 and 2 report the results on CIFAR-10 and Image Net, respectively, and Figure 3 plots the CIFAR-10 results for each of σ {0.25, 0.50, 1.00}. Overall, on both CIFAR10 and Image Net, we observe that the proposed fine-tuning scheme could uniformly improve certified accuracy across the range considered, even including the clean accuracy. This confirms that the tuning could essentially improve the accuracy-robustness trade-off rather than simply moving along itself. As provided in Table 3, we remark that simply pursuing only the Brier loss (7) may achieve a better ACR in overall, but with a decreased accuracy: it is the role of the anti-consistency loss (8) to balance between the two, consequently to achieve a better trade-off afterwards. Empirical accuracy. In practical scenarios of adopting smoothed classifiers for inference, it is up to users to decide how to deal with the abstained inputs. Here, we consider a possible candidate of simply outputting the standard prediction instead for such inputs: in this way, the output could be noted as an uncertified prediction, while possibly being more accurate, e.g., for in-distribution Table 3: Comparison of ACR and certified accuracy of cascaded smooth classifiers on CIFAR-10 over different training and ablations. We use σ {0.25, 0.5, 1.0} for cascaded smoothing. Bold and underline indicate the best and runner-up, respectively. Model reported in Table 1 is marked as grey. Cascaded, σ {0.25, 0.5, 1.0} Certified accuracy at ε (%) Training ACR 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 Gaussian [15] 0.470 70.7 46.9 31.7 23.7 16.8 12.4 9.2 6.5 Consistency [35] 0.548 57.6 42.7 32.3 26.3 22.4 18.1 14.6 11.1 Carlini et al. [11] 0.579 80.5 52.8 37.4 28.0 21.7 16.8 11.8 8.5 + Cross-entropy 0.605 77.9 52.8 39.6 29.1 22.2 18.7 13.5 9.5 + Brier loss 0.666 77.6 55.7 41.7 32.7 26.8 21.7 16.3 11.9 + Anti-consist. (α = 1.0) 0.645 79.6 53.5 40.5 31.2 25.0 20.1 16.2 11.4 + Anti-consist. (α = 2.0) 0.625 80.1 54.6 39.2 30.6 24.0 18.5 14.5 9.9 + Anti-consist. (α = 1.0) 0.566 83.9 55.1 37.8 25.8 19.8 14.5 11.5 8.1 Table 4: Comparison of ACR and certified accuracy (%) on CIFAR-10 across two fine-tuning schemes: diffusion fine-tuning ( Diff. ; ours) and classifier fine-tuning ( Class. ) [11]. We use σ = 0.5 for this experiment. Fine-tuning Certified accuracy at ε (%) Diff. Class. ACR 0.00 0.25 0.50 0.75 1.00 1.25 0.639 75.3 62.0 50.6 39.3 31.0 22.9 0.664 75.1 64.1 53.3 42.1 32.5 25.0 0.664 76.0 64.7 52.7 42.1 32.4 24.4 0.673 75.2 64.7 53.5 43.8 32.5 25.8 Table 5: Comparison of model error rates on CIFAR-10 decomposed into (a) over-smoothing (p p0) and (b) overconfidence (p > p0), also with ACR. Error rates (%, ) Model (σ = 1.0) p p0 p > p0 ACR ( ) Carlini et al. [11] 43.8 7.6 0.498 Cascading 5.0 14.5 0.579 + Anti-consist. 4.7 11.4 0.566 + Brier loss 3.6 16.8 0.645 inputs. Specifically, in Table 1 and 2, we consider a fixed standard classifier of CLIP-finetuned Vi T-B/16 on CIFAR-10 (or Image Net), and compare the empirical accuracy of smoothed classifiers on CIFAR-10, -10-C [27] and -10.1 [54] (or Image Net, -R [28], and -A [30]). Overall, the results show that our smoothed models could consistently outperform even in terms of empirical accuracy while maintaining high certified accuracy, i.e., they abstain only when necessary. We additionally observe in Appendix C.2 that the empirical accuracy of our smoothed model can be further improved by considering an ensemble with the clean classifier: e.g., the ensemble improves the accuracy of our cascaded classifier (σ {0.25, 0.5}) on CIFAR-10-C by 88.8% 95.0%, even outperforming the accuracy of the standard classifier of 93.4%. 4.2 Ablation study In Table 3, we compare our result with other training baselines as well as some ablations for a component-wise analysis, particularly focusing on their performance in cascaded smoothing across σ {0.25, 0.50, 1.00} on CIFAR-10. Here, we highlight several remarks from the results, and provide the more detailed study, e.g., the effect of p0 (5), in Appendix C.3. Cascading from other training. Gaussian and Consistency in Table 3 report the certified robustness of cascaded classifiers where each of single-scale models is individually trained by the method. Even while their (certified) clean accuracy of σ = 0.25 models are competitive with those of denoising-based models, their collective accuracy significantly degraded after cascading, and interestingly the drop is much more significant on Consistency , although it did provide more robustness at larger ε. Essentially, for a high clean accuracy in cascaded smoothing the individual classifiers should make consistent predictions although their confidence may differ: the results imply that individual training of classifiers, without a shared denoiser, may break this consistency. This supports an architectural benefit of denoised smoothing for a use as cascaded smoothing. Cross-entropy vs. Brier. + Cross-entropy in Table 3 considers an ablation of the Brier loss (7) where the loss is replaced by the standard cross-entropy: although it indeed improves ACR compared to Carlini et al. [11] and achieves the similar clean accuracy with + Brier loss , its gain in overall certified robustness is significantly inferior to that of Brier loss as also reflected in the worse ACR. The superiority of the squared loss instead of the log loss suggests that it may not be necessary to optimize the confidence of individual denoised image toward a value strictly close to 1, which is reasonable considering the severity of noise usually considered for randomized smoothing. Anti-consistency loss. The results marked as + Anti-consist. in Table 3, on the other hand, validates the effectiveness of the anti-consistency loss (8). Compared to + Brier loss , which is equivalent to the case when α = 0.0 in (9), we observe that adding anti-consistency results in a slight decrease in ACR but with an increase in clean accuracy. The increased accuracy of cascaded smoothing indicates that the fine-tuning could successfully reduce over-confidence, and this tendency continues at larger α = 2.0. Even with the decrease in ACR over the Brier loss, the overall ACRs attained are still superior to others, confirming the effectiveness of our proposed loss in total (9). Classifier fine-tuning. In Table 4, we compare our proposed diffusion fine-tuning (Section 3.3) with another possible scheme of classifier fine-tuning, of which the effectiveness has been shown by Carlini et al. [11]. Overall, we observe that fine-tuning both classifiers and denoiser can bring their complementary effects to improve denoised smoothing in ACR, and the diffusion fine-tuning itself could obtain a comparable gain to the classifier fine-tuning. The (somewhat counter-intuitive) effectiveness of diffusion fine-tuning confirms that denoising process can be also biased as well as classifiers to make over-confident errors. We further remark two aspects where classifier fine-tuning can be less practical, especially when it comes with the cascaded smoothing pipeline we propose: (a) To perform cascaded smoothing at multiple noise scales, classifier fine-tuning would require separate runs of training for optimal models per scale, also resulting in multiple different models to load - which can be less scalable in terms of memory efficiency. (b) In a wider context of denoised smoothing, the classifier part is often assumed to be a model at scale, even covering cases when it is a black-box model, e.g., public APIs. Classifier fine-tuning, in such cases, can become prohibitive or even impossible. With respect to (a) and (b), denoiser fine-tuning we propose offers a more efficient inference architecture: it can handle multiple noise scales jointly with a single diffusion model, while also being applicable to the extreme scenario when the classifier model is black-box. Over-smoothing and over-confidence. As proposed in Section 3.1, errors from smoothed classifiers can be decomposed into two: i.e., p p0 for over-smoothing, and p > p0 over-confidence, respectively. In Table 5, we further report a detailed breakdown of the model errors on CIFAR-10, assuming σ = 1.0. Overall, we observe that over-smoothing can be the major source of errors especially at such a high noise level, and our proposed cascading dramatically reduces the error. Although we find cascading could increase over-confidence from accumulating errors through multiple inferences, our diffusion fine-tuning could alleviate the errors jointly reducing the over-smoothing as well. 5 Conclusion and Discussion Randomized smoothing has been traditionally viewed as a somewhat less-practical approach, perhaps due to its cost in inference time and impact on accuracy. In another perspective, to our knowledge, it is currently one of a few existing approaches that is prominent in pursuing adversarial robustness at scale, e.g., in a paradigm where training cost scales faster than computing power. This work aims to make randomized smoothing more practical, particularly concerning on scalable scenarios of robustifying large pre-trained classifiers. We believe our proposals in this respect, i.e., cascaded smoothing and diffusion calibration, can be a useful step towards building safer AI-based systems. Limitation. A practical downside of randomized smoothing is in its increased inference cost, mainly from the majority voting procedure per inference. Our method also essentially possesses this practical bottleneck, and the proposed multi-scale smoothing scheme may further increase the cost from taking multiple smoothed inferences. Yet, we note that randomized smoothing itself is equipped with many practical axes to reduce its inference cost by compensating with abstention: for example, one can reduce the number of noise samples, e.g., to n = 100. It would be an important future direction to explore practices for a better trade-off between the inference cost and robustness of smoothed classifiers, which could eventually open up a feasible way to obtain adversarial robustness at scale. Broader impact. Deploying deep learning based systems into the real-world, especially when they are of security-concerned [12, 72], poses risks for both companies and customers, and we researchers are responsible to make this technology more reliable through research towards AI safety [2, 29]. Adversarial robustness, that we focus on in this work, is one of the central parts of this direction, but one should also recognize that adversarial robustness is still a bare minimum requirement for reliable deep learning. The future research should also explore more diverse notions of AI Safety to establish a realistic sense of security for practitioners, e.g., monitoring and alignment research, to name a few. Acknowledgments and Disclosure of Funding This work was partly supported by Center for Applied Research in Artificial Intelligence (CARAI) grant funded by Defense Acquisition Program Administration (DAPA) and Agency for Defense Development (ADD) (UD230017TD), and by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No.2019-0-00075, Artificial Intelligence Graduate School Program (KAIST)). We are grateful to the Center for AI Safety (CAIS) for generously providing compute resources that supported a significant portion of the experiments conducted in this work. We thank Kyuyoung Kim and Sihyun Yu for the proofreading of our manuscript, and Kyungmin Lee for the initial discussions. We also thank the anonymous reviewers for their valuable comments to improve our manuscript. [1] Motasem Alfarra, Adel Bibi, Philip HS Torr, and Bernard Ghanem. Data dependent randomized smoothing. In Uncertainty in Artificial Intelligence, pages 64 74. PMLR, 2022. [2] Dario Amodei, Chris Olah, Jacob Steinhardt, Paul Christiano, John Schulman, and Dan Mané. Concrete problems in AI safety. ar Xiv preprint ar Xiv:1606.06565, 2016. [3] Anish Athalye, Nicholas Carlini, and David Wagner. Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. In Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 274 283, Stockholmsmässan, Stockholm Sweden, 10 15 Jul 2018. PMLR. URL http://proceedings.mlr.press/ v80/athalye18a.html. [4] Mislav Balunovic and Martin Vechev. Adversarial training and provable defenses: Bridging the gap. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum? id=SJx SDxr KDr. [5] Carlo Bonferroni. Teoria statistica delle classi e calcolo delle probabilita. Pubblicazioni del R Istituto Superiore di Scienze Economiche e Commericiali di Firenze, 8:3 62, 1936. [6] Glenn W Brier. Verification of forecasts expressed in terms of probability. Monthly weather review, 78(1): 1 3, 1950. [7] Anthony Brohan, Yevgen Chebotar, Chelsea Finn, Karol Hausman, Alexander Herzog, Daniel Ho, Julian Ibarz, Alex Irpan, Eric Jang, Ryan Julian, et al. Do as I can, not as I say: Grounding language in robotic affordances. In 6th Annual Conference on Robot Learning, 2022. [8] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. In Advances in Neural Information Processing Systems, volume 33, pages 1877 1901, 2020. [9] Nicholas Carlini, Anish Athalye, Nicolas Papernot, Wieland Brendel, Jonas Rauber, Dimitris Tsipras, Ian Goodfellow, and Aleksander Madry. On evaluating adversarial robustness. ar Xiv preprint ar Xiv:1902.06705, 2019. [10] Nicholas Carlini, Florian Tramèr, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, et al. Extracting training data from large language models. In 30th USENIX Security Symposium, pages 2633 2650, 2021. [11] Nicholas Carlini, Florian Tramer, Krishnamurthy Dj Dvijotham, Leslie Rice, Mingjie Sun, and J Zico Kolter. (Certified!!) adversarial robustness for free! In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=JLg5a HHv7j. [12] Rich Caruana, Yin Lou, Johannes Gehrke, Paul Koch, Marc Sturm, and Noemie Elhadad. Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 1721 1730, 2015. [13] Chen Chen, Kezhi Kong, Peihong Yu, Juan Luque, Tom Goldstein, and Furong Huang. Insta-RS: Instancewise randomized smoothing for improved robustness and accuracy. ar Xiv preprint ar Xiv:2103.04436, 2021. [14] Charles J Clopper and Egon S Pearson. The use of confidence or fiducial limits illustrated in the case of the binomial. Biometrika, 26(4):404 413, 1934. [15] Jeremy Cohen, Elan Rosenfeld, and Zico Kolter. Certified adversarial robustness via randomized smoothing. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 1310 1320, Long Beach, California, USA, 09 15 Jun 2019. PMLR. URL http: //proceedings.mlr.press/v97/cohen19c.html. [16] Francesco Croce and Matthias Hein. Provable robustness against all adversarial lp-perturbations for p 1. In International Conference on Learning Representations, 2020. [17] Francesco Croce, Maksym Andriushchenko, Vikash Sehwag, Edoardo Debenedetti, Nicolas Flammarion, Mung Chiang, Prateek Mittal, and Matthias Hein. Robust Bench: a standardized adversarial robustness benchmark. ar Xiv preprint ar Xiv:2010.09670, 2020. [18] Prafulla Dhariwal and Alexander Nichol. Diffusion models beat GANs on image synthesis. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 8780 8794. Curran Associates, Inc., 2021. URL https://proceedings.neurips.cc/paper_files/paper/2021/file/ 49ad23d1ec9fa4bd8d77d02681df5cfa-Paper.pdf. [19] Xiaoyi Dong, Jianmin Bao, Ting Zhang, Dongdong Chen, Gu Shuyang, Weiming Zhang, Lu Yuan, Dong Chen, Fang Wen, and Nenghai Yu. CLIP itself is a strong fine-tuner: Achieving 85.7% and 88.0% top-1 accuracy with Vi T-B and Vi T-L on Image Net. ar Xiv preprint ar Xiv:2212.06138, 2022. [20] Francisco Eiras, Motasem Alfarra, M. Pawan Kumar, Philip H. S. Torr, Puneet K. Dokania, Bernard Ghanem, and Adel Bibi. ANCER: Anisotropic certification via sample-wise volume maximization, 2021. [21] Gamaleldin Elsayed, Shreya Shankar, Brian Cheung, Nicolas Papernot, Alexey Kurakin, Ian Goodfellow, and Jascha Sohl-Dickstein. Adversarial examples that fool both computer vision and time-limited humans. Advances in Neural Information Processing Systems, 31, 2018. [22] Marc Fischer, Maximilian Baader, and Martin Vechev. Certified defense to image transformations via randomized smoothing. Advances in Neural information processing systems, 33:8404 8417, 2020. [23] Timon Gehr, Matthew Mirman, Dana Drachsler-Cohen, Petar Tsankov, Swarat Chaudhuri, and Martin Vechev. Ai2: Safety and robustness certification of neural networks with abstract interpretation. In IEEE Symposium on Security and Privacy, 2018. [24] Leo A Goodman. On simultaneous confidence intervals for multinomial proportions. Technometrics, 7(2): 247 254, 1965. [25] Sven Gowal, Krishnamurthy Dj Dvijotham, Robert Stanforth, Rudy Bunel, Chongli Qin, Jonathan Uesato, Relja Arandjelovic, Timothy Mann, and Pushmeet Kohli. Scalable verified training for provably robust image classification. In IEEE/CVF International Conference on Computer Vision, pages 4842 4851, 2019. [26] Chong Guo, Michael Lee, Guillaume Leclerc, Joel Dapello, Yug Rao, Aleksander Madry, and James Dicarlo. Adversarially trained neural representations are already as robust as biological neural representations. In International Conference on Machine Learning, pages 8072 8081. PMLR, 2022. [27] Dan Hendrycks and Thomas Dietterich. Benchmarking neural network robustness to common corruptions and perturbations. In International Conference on Learning Representations, 2019. URL https:// openreview.net/forum?id=HJz6ti Cq Ym. [28] Dan Hendrycks, Steven Basart, Norman Mu, Saurav Kadavath, Frank Wang, Evan Dorundo, Rahul Desai, Tyler Zhu, Samyak Parajuli, Mike Guo, Dawn Song, Jacob Steinhardt, and Justin Gilmer. The many faces of robustness: A critical analysis of out-of-distribution generalization. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 8340 8349, October 2021. [29] Dan Hendrycks, Nicholas Carlini, John Schulman, and Jacob Steinhardt. Unsolved problems in ML safety. ar Xiv preprint ar Xiv:2109.13916, 2021. [30] Dan Hendrycks, Kevin Zhao, Steven Basart, Jacob Steinhardt, and Dawn Song. Natural adversarial examples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15262 15271, June 2021. [31] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840 6851, 2020. [32] Miklós Z. Horváth, Mark Niklas Mueller, Marc Fischer, and Martin Vechev. Boosting randomized smoothing with variance reduced classifiers. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=m Hu2v Ids_-b. [33] Miklós Z. Horváth, Mark Niklas Müller, Marc Fischer, and Martin Vechev. Robust and accurate compositional architectures for randomized smoothing. In ICLR 2022 Workshop on Socially Responsible Machine Learning, 2022. [34] Aapo Hyvärinen and Peter Dayan. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005. [35] Jongheon Jeong and Jinwoo Shin. Consistency regularization for certified robustness of smoothed classifiers. Advances in Neural Information Processing Systems, 33:10558 10570, 2020. [36] Jongheon Jeong, Sejun Park, Minkyu Kim, Heung-Chang Lee, Doguk Kim, and Jinwoo Shin. Smooth Mix: Training confidence-calibrated smoothed classifiers for certified robustness. In Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=nl EQMVBD359. [37] Jongheon Jeong, Seojin Kim, and Jinwoo Shin. Confidence-aware training of smoothed classifiers for certified robustness. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 8005 8013, 2023. doi: 10.1609/aaai.v37i7.25968. URL https://ojs.aaai.org/index.php/AAAI/ article/view/25968. [38] Jared Kaplan, Sam Mc Candlish, Tom Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. Scaling laws for neural language models, 2020. [39] Alex Krizhevsky. Learning multiple layers of features from tiny images. Technical report, Department of Computer Science, University of Toronto, 2009. [40] Li-Cheng Lan, Huan Zhang, Ti-Rong Wu, Meng-Yu Tsai, I-Chen Wu, and Cho-Jui Hsieh. Are Alpha Zerolike agents robust to adversarial perturbations? In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https: //openreview.net/forum?id=y Z_Jl Za OCzv. [41] Mathias Lecuyer, Vaggelis Atlidakis, Roxana Geambasu, Daniel Hsu, and Suman Jana. Certified robustness to adversarial examples with differential privacy. In 2019 IEEE Symposium on Security and Privacy (SP), pages 656 672. IEEE, 2019. [42] Kyungmin Lee. Provable defense by denoised smoothing with learned score function. In ICLR Workshop on Security and Safety in Machine Learning Systems, 2021. [43] Alexander J Levine and Soheil Feizi. Improved, deterministic smoothing for l1 certified robustness. In International Conference on Machine Learning, pages 6254 6264. PMLR, 2021. [44] Linyi Li, Xiangyu Qi, Tao Xie, and Bo Li. So K: Certified robustness for deep neural networks. ar Xiv preprint ar Xiv:2009.04131, 2020. [45] Linyi Li, Jiawei Zhang, Tao Xie, and Bo Li. Double sampling randomized smoothing. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato, editors, Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 13163 13208. PMLR, 17 23 Jul 2022. URL https://proceedings.mlr. press/v162/li22aa.html. [46] Chizhou Liu, Yunzhen Feng, Ranran Wang, and Bin Dong. Enhancing certified robustness via smoothed weighted ensembling, 2020. [47] Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=r Jz IBf ZAb. [48] Chengzhi Mao, Scott Geng, Junfeng Yang, Xin Wang, and Carl Vondrick. Understanding zero-shot adversarial robustness for large-scale models. In International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=P4b XCaw Ri5J. [49] Matthew Mirman, Timon Gehr, and Martin Vechev. Differentiable abstract interpretation for provably robust neural networks. In International Conference on Machine Learning, volume 80, pages 3578 3586, 10 15 Jul 2018. URL http://proceedings.mlr.press/v80/mirman18b.html. [50] Jeet Mohapatra, Ching-Yun Ko, Tsui-Wei Weng, Pin-Yu Chen, Sijia Liu, and Luca Daniel. Higherorder certification for randomized smoothing. Advances in Neural Information Processing Systems, 33: 4501 4511, 2020. [51] Mark Niklas Mueller, Mislav Balunovic, and Martin Vechev. Certify or predict: Boosting certified robustness with compositional architectures. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=USCNapootw. [52] Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 8162 8171. PMLR, 18 24 Jul 2021. URL https://proceedings.mlr.press/v139/nichol21a.html. [53] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, pages 8748 8763. PMLR, 2021. [54] Benjamin Recht, Rebecca Roelofs, Ludwig Schmidt, and Vaishaal Shankar. Do CIFAR-10 classifiers generalize to CIFAR-10?, 2018. [55] Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10684 10695, 2022. [56] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. Image Net Large Scale Visual Recognition Challenge. International Journal of Computer Vision, 115(3):211 252, 2015. doi: 10.1007/s11263-015-0816-y. [57] Hadi Salman, Jerry Li, Ilya Razenshteyn, Pengchuan Zhang, Huan Zhang, Sebastien Bubeck, and Greg Yang. Provably robust deep learning via adversarially trained smoothed classifiers. In Advances in Neural Information Processing Systems 32, pages 11289 11300. Curran Associates, Inc., 2019. [58] Hadi Salman, Mingjie Sun, Greg Yang, Ashish Kapoor, and J Zico Kolter. Denoised smoothing: A provable defense for pretrained classifiers. Advances in Neural Information Processing Systems, 33:21945 21957, 2020. [59] Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning, pages 2256 2265. PMLR, 2015. [60] Peter Súkeník, Aleksei Kuvshinov, and Stephan Günnemann. Intriguing properties of input-dependent randomized smoothing. In Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 20697 20743. PMLR, 17 23 Jul 2022. URL https://proceedings.mlr.press/v162/sukeni-k22a.html. [61] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. Intriguing properties of neural networks, 2014. [62] Antonio Torralba, Rob Fergus, and William T. Freeman. 80 million tiny images: A large data set for nonparametric object and scene recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 30(11):1958 1970, 2008. doi: 10.1109/TPAMI.2008.128. [63] Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Herve Jegou. Training data-efficient image transformers & distillation through attention. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 10347 10357. PMLR, 18 24 Jul 2021. URL https://proceedings.mlr.press/v139/touvron21a.html. [64] Florian Tramèr, Jens Behrmann, Nicholas Carlini, Nicolas Papernot, and Jörn-Henrik Jacobsen. Fundamental tradeoffs between invariance and sensitivity to adversarial perturbations. In International Conference on Machine Learning, pages 9561 9571. PMLR, 2020. [65] Florian Tramèr, Nicholas Carlini, Wieland Brendel, and Aleksander Madry. On adaptive attacks to adversarial example defenses. In Advances in Neural Information Processing Systems, volume 33, 2020. [66] Dimitris Tsipras, Shibani Santurkar, Logan Engstrom, Alexander Turner, and Aleksander Madry. Robustness may be at odds with accuracy. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=Syx Ab30c Y7. [67] Lei Wang, Runtian Zhai, Di He, Liwei Wang, and Li Jian. Pretrain-to-finetune adversarial training via sample-wise randomized smoothing, 2021. URL https://openreview.net/forum?id=Te1a Z2my PIu. [68] Eric Wong and Zico Kolter. Provable defenses against adversarial examples via the convex outer adversarial polytope. In International Conference on Machine Learning, volume 80, pages 5286 5295, 10 15 Jul 2018. URL http://proceedings.mlr.press/v80/wong18a.html. [69] Chaowei Xiao, Zhongzhu Chen, Kun Jin, Jiongxiao Wang, Weili Nie, Mingyan Liu, Anima Anandkumar, Bo Li, and Dawn Song. Dense Pure: Understanding diffusion models for adversarial robustness. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview. net/forum?id=p7hv OJ6Gq0i. [70] Greg Yang, Tony Duan, J Edward Hu, Hadi Salman, Ilya Razenshteyn, and Jerry Li. Randomized smoothing of all shapes and sizes. In International Conference on Machine Learning, pages 10693 10705. PMLR, 2020. [71] Zhuolin Yang, Linyi Li, Xiaojun Xu, Bhavya Kailkhura, Tao Xie, and Bo Li. On the certified robustness for ensemble models and beyond. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=t Ua4REj Gj Tf. [72] Ekim Yurtsever, Jacob Lambert, Alexander Carballo, and Kazuya Takeda. A survey of autonomous driving: Common practices and emerging technologies. IEEE Access, 8:58443 58469, 2020. [73] Runtian Zhai, Chen Dan, Di He, Huan Zhang, Boqing Gong, Pradeep Ravikumar, Cho-Jui Hsieh, and Liwei Wang. MACER: Attack-free and scalable robust training via maximizing certified radius. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum? id=r Jx1Na4Fwr. [74] Xiaohua Zhai, Alexander Kolesnikov, Neil Houlsby, and Lucas Beyer. Scaling Vision Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12104 12113, 2022. [75] Hongyang Zhang, Yaodong Yu, Jiantao Jiao, Eric Xing, Laurent El Ghaoui, and Michael Jordan. Theoretically principled trade-off between robustness and accuracy. In Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 7472 7482, Long Beach, California, USA, 09 15 Jun 2019. PMLR. URL http: //proceedings.mlr.press/v97/zhang19p.html. [76] Huan Zhang, Hongge Chen, Chaowei Xiao, Sven Gowal, Robert Stanforth, Bo Li, Duane Boning, and Cho Jui Hsieh. Towards stable and efficient training of verifiably robust neural networks. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum?id=Skxuk1r Fw B. [77] Huan Zhang, Shiqi Wang, Kaidi Xu, Linyi Li, Bo Li, Suman Jana, Cho-Jui Hsieh, and J Zico Kolter. General cutting planes for bound-propagation-based neural network verification. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=5ha AJAcofjc. A Related Work In the context of aggregating multiple smoothed classifiers, several works have previously explored on ensembling smoothed classifiers of the same noise scale [46, 32, 71], in attempt to boost their certified robustness. Our proposed multi-scale framework for randomized smoothing considers a different setup of aggregating smoothed classifiers of different noise scales, with a novel motivation of addressing the current accuracy-robustness trade-off in randomized smoothing. More in this respect, Mueller et al. [51] have considered a selective inference scheme between a certification network (for robustness) and a core network (for accuracy), and Horváth et al. [33] have adapted the framework into randomized smoothing. Our approach explores an orthogonal direction, and can be viewed as directly improving the certification network here: in this way, we could offer improvements not only in empirical but also in certified accuracy as a result. Independently to our method, there have been approaches to further tighten the certified lower-bound that randomized smoothing guarantees [50, 70, 43, 45, 69]. Mohapatra et al. [50] have shown that higher-order information of smoothed classifiers, e.g., its gradient on inputs, can tighten the certification of Gaussian-smoothed classifiers, beyond that only utilizes the zeroth-order information of smoothed confidence. In the context of denoised smoothing [58, 11], Xiao et al. [69] have recently proposed a multi-step, multi-round scheme for diffusion models to improve accuracy of denoising at randomized smoothing. Such approaches could be also incorporated to our method to further improve certification. Along with the developments of randomized smoothing, there have been also other attempts in certifying deep neural networks against adversarial examples [23, 68, 49, 25, 76, 16, 4, 77], where a more extensive survey on the field can be found in Li et al. [44]. B Experimental Details B.1 Datasets CIFAR-10 [39] consist of 60,000 images of size 32 32 pixels, 50,000 for training and 10,000 for testing. Each of the images is labeled to one of 10 classes, and the number of data per class is set evenly, i.e., 6,000 images per each class. By default, we do not apply any data augmentation but the input normalization with mean [0.5, 0.5, 0.5] and standard deviation [0.5, 0.5, 0.5], following the standard training configurations of diffusion models. The full dataset can be downloaded at https://www.cs.toronto.edu/~kriz/cifar.html. CIFAR-10-C [27] are collections of 75 replicas of the CIFAR-10 test datasets (of size 10,000), which consists of 15 different types of common corruptions each of which contains 5 levels of corruption severities. Specifically, the datasets includes the following corruption types: (a) noise: Gaussian, shot, and impulse noise; (b) blur: defocus, glass, motion, zoom; (c) weather: snow, frost, fog, bright; and (d) digital: contrast, elastic, pixel, JPEG compression. In our experiments, we evaluate test errors on CIFAR-10-C for models trained on the clean CIFAR-10 datasets, where the error values are averaged across different corruption types per severity level. The full datasets can be downloaded at https://github.com/hendrycks/robustness. CIFAR-10.1 [54] is a reproduction of the CIFAR-10 test set that are separately collected from Tiny Images dataset [62]. The dataset consists of 2,000 samples for testing, and designed to minimize distribution shift relative to the original CIFAR-10 dataset in their data creation pipelines. The full dataset can be downloaded at https://github.com/modestyachts/CIFAR-10.1, where we use the v6 version for our experiments among those given in the repository. Image Net [56], also known as ILSVRC 2012 classification dataset, consists of 1.2 million highresolution training images and 50,000 validation images, which are labeled with 1,000 classes. As a data pre-processing step, we apply a 256 256 center cropping for both training and testing images after re-scaling the images to have 256 in their shorter edges, making it compatible with the preprocessing of ADM [18], the backbone diffusion model. Similarly to CIFAR-10, all the images are normalized with mean [0.5, 0.5, 0.5] and standard deviation [0.5, 0.5, 0.5]. A link for downloading the full dataset can be found in http://image-net.org/download. Image Net-R [28] consists of 30,000 images of various artistic renditions for 200 (out of 1,000) Image Net classes: e.g., art, cartoons, deviantart, graffiti, embroidery, graphics, origami, paintings, patterns, plastic objects, plush objects, sculptures, sketches, tattoos, toys, video game renditions, and so on. To perform an evaluation of Image Net classifiers on this dataset, we apply masking on classifier logits for the 800 classes those are not in Image Net-R. The full dataset can be downloaded at https://github.com/hendrycks/imagenet-r. Image Net-A [30] consists of 7,500 images of 200 Image Net classes those are adversarially filtered, by collecting natural images that causes wrong predictions from a pre-trained Res Net-50 model. To perform an evaluation of Image Net classifiers on this dataset, we apply masking on classifier logits for the 800 classes those are not in Image Net-A. The full dataset can be downloaded at https://github.com/hendrycks/natural-adv-examples. B.2 Training CLIP fine-tuning. We follow the training configuration suggested by FT-CLIP [19] in finetuning the CLIP Vi T-B/16@224 model, where the full script is specified at https://github.com/ Light DXY/FT-CLIP. The same configuration is applied to all the datasets considered for fine-tuning, namely CIFAR-10, Image Net, and Gaussian-corrupted versions of CIFAR-10 (for baselines such as Gaussian and Consistency ), which can be done by resizing the datasets into 224 224 as a data pre-processing. We fine-tune all these tested models for 50 epochs on each of the considered datasets. Diffusion fine-tuning. Similarly to the CLIP fine-tuning, we follow the training details given by Nichol and Dhariwal [52] in fine-tuning diffusion models, which are specified in https://github. com/openai/improved-diffusion. More concretely, here we fine-tune each pre-trained diffusion model by resuming the training following its configuration, but with an added regularization term we propose upon the original training loss. We use the 50M-parameter 32 32 diffusion model from Nichol and Dhariwal [52] for CIFAR-10,6 and the 552M-parameter 256 256 unconditional model from Dhariwal and Nichol [18] for Image Net7 to initialize each fine-tuning run. We fine-tune each model for 50K training steps, using batch size 128 and 64 for CIFAR-10 and Image Net, respectively. B.3 Hyperparameters Unless otherwise noted, we use p0 = 0.5 for cascaded smoothing throughout our experiments. We mainly consider two configurations of cascaded smoothing: (a) σ {0.25, 0.50, 1.00}, and (b) σ {0.25, 0.50}, each of which denoted as (0.25-1.00) and (0.25-0.50) in tables, respectively. For diffusion calibration, on the other hand, we use α = 1.0 by default. We use λ = 0.01 on CIFAR-10 and λ = 0.005 on Image Net, unless noted: we halve the regularization strength λ on Image Net considering its batch size used in the fine-tuning, i.e., 128 (for CIFAR-10) vs. 64 (for Image Net). In other words, we follow λ = 0.01 (batch_size/128). Among the baselines considered, Consistency [35] requires two hyperparameters: the coefficient for the consistency term η and the entropy term γ. Following those considered by Jeong and Shin [35], we fix γ = 0.5 and η = 5.0 throughout our experiments. B.4 Computing infrastructure Overall, we conduct our experiments with a cluster of 8 NVIDIA V100 32GB GPUs and 8 instances of a single NVIDIA A100 80GB GPU. All the CIFAR-10 experiments are run on a single NVIDIA A100 80GB GPU, including both the diffusion fine-tuning and the smoothed inference procedures. For the Image Net experiments, we use 8 NVIDIA V100 32GB GPUs per run. In our computing environment, it takes e.g., 10 seconds and 5 minutes per image (we use n = 10, 000 for each inference) to perform a single pass of smoothed inference on CIFAR-10 and Image Net, respectively. For a single run of fine-tuning diffusion models, we observe 1.5 days of training cost on CIFAR-10 with a single NVIDIA A100 80GB GPU, and that of 4 days on Image Net using a cluster of 8 NVIDIA V100 32GB GPUs, to run 50K training steps. At inference time, our proposed cascaded smoothing may introduce an extra computational overhead as the cost for the increased accuracy, i.e., from taking multiple times of smoothed inferences. Nevertheless, remark that the actual overhead per input in practice can be different depending on the early-stopping result of cascaded smoothing. For example, in our experiments, we observe that 6https://github.com/openai/improved-diffusion 7https://github.com/openai/guided-diffusion Carlini et al. (2023) IDRS (E[σ] 0.263) Ours (σ {0.2, 0.25, 0.3}) Ours (σ {0.25, 0.5}) Certified accuracy Radius 0 0.5 1.0 1.5 Figure 4: Comparison of certified test accuracy with (a) Carlini et al. [11] at σ = 0.25, (b) Input-Dependent Randomized Smoothing (IDRS) proposed by Súkeník et al. [60]. For (b), we adopt the per-sample σ values of CIFAR-10 test set of Súkeník et al. [60]. p0 = 0.5 p0 = 0.6 p0 = 0.7 p0 = 0.8 p0 = 0.9 σx ABSTAIN 0.25 0.50 1.00 Figure 5: Comparison of histograms across p0 [0.5, 1.0) on the distribution of σ s where each of CIFAR-10 test samples halts from cascaded smoothing. Each of smoothed inferences is performed with n = 10, 000 noise samples and α = 0.001. the average inference time to process the CIFAR-10 test set via cascaded smoothing is only 1.5 compared to standard (single-scale) smoothed inferences, even considering σ {0.25, 0.5, 1.0}. C Additional Results C.1 Comparison with input-dependent smoothing In a sense that the proposed cascaded smoothing results in different smoothing factors σ per-sample basis, our approach can be also viewed in the context of input-dependent randomized smoothing [1, 67, 13, 20], a line of attempts to improve certified robustness by applying different smoothing factor σ conditioned on input. From being more flexible in assigning σ, however, these approaches commonly require more assumptions for their certification, making them hard to compare directly upon the standard evaluation protocol, e.g., that we considered in our experiments. For example, Súkeník et al. [60] have proved that such a scheme of freely assigning σ does not provide valid robustness certificates in the standard protocol, and [1, 20] in response have suggested to further assume that all the previous test samples during evaluation can be memorized to predict future samples, which can affect the practical relevance of the overall evaluation protocol. To address the issue, Súkeník et al. [60] have also proposed a fix of [1] to discard its test-time dependency, by assuming a more strict restriction on the variation of σ. Nevertheless, they also report that the benefit of input-dependent σ then becomes quite limited, supporting the challenging nature of input-dependent smoothing. As shown in Figure 4 that additionally compares our method with [60], we indeed confirm that cascaded smoothing offers significantly wider robustness certificates with a better accuracy trade-off. This is unprecedented to the best of our knowledge, which could be achieved from our completely new approach of data-dependent smoothing: i.e., by combining multiple smoothed classifiers through abstention, moving away from the previous attempts that are focused on obtaining a single-step smoothing. C.2 Clean ensemble for empirical accuracy In this appendix, we demonstrate that predictions from smoothed classifiers can not only useful to obtain adversarial robustness, but also to improve out-of-distribution robustness of clean classifiers, which suggests a new way to utilize randomized smoothing. Specifically, consider a smoothed classifier ˆf, and its smoothed confidences at x, say p ˆ f(y|x) for y Y. Next, we consider prediction confidences of a clean prediction, namely pf(y|x). Remark that pf(y|x) can be interpreted in terms of logits, namely log pf(x|y), of a neural network based model f as follows: pf(y|x) = pf(x, y) pf(x) = p(y) pf(x|y) P y Y pf(x, y). (10) Note that the prior p(y) is usually assumed to be the uniform distribution U(Y). Here, we consider a variant of this prediction, by replacing the prior p(y) by a mixture of p(y) and a smoothed prediction Table 6: Comparison of per-corruption empirical accuracy of clean ( FT-CLIP ) and smoothed classifiers (Others) on CIFAR-10-C. For each corruption type, we report accuracy averaged over 5 levels of severity, and for each level we use 1,000 uniform subsamples to compute the accuracy. FT-CLIP [19] 98.1 83.7 88.7 96.4 96.8 85.7 95.0 96.1 96.1 96.0 97.0 97.8 97.5 94.7 91.0 88.5 93.4 Carlini et al. [11] 88.8 85.8 86.9 88.7 87.3 83.6 86.8 86.0 85.2 83.6 78.9 87.8 75.7 85.4 86.4 85.8 84.9 + Casc. + Calib. 92.5 89.0 90.1 91.5 90.7 87.3 89.6 90.7 89.9 89.4 86.0 91.1 81.2 89.7 90.4 89.3 89.1 + Clean prior 98.0 90.9 93.0 96.5 96.8 89.9 95.2 96.3 96.2 96.6 96.8 97.7 97.4 95.3 94.2 91.6 95.0 Table 7: Comparison of empirical accuracy and certified accuracy of cascaded smooth classifiers on CIFAR-10 across ablations on λ. We use σ {0.25, 0.5, 1.0} for cascaded smoothing. CIFAR-10 Empirical Certified accuracy at ε (%) Setups Clean 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 α = 1.0 84.1 79.9 54.4 39.6 30.9 24.2 18.9 13.9 9.3 α = 2.0 84.3 80.7 53.9 38.9 29.7 23.9 18.7 13.0 9.6 α = 4.0 85.1 80.9 52.8 39.1 29.5 23.2 17.3 12.4 9.4 α = 8.0 86.1 81.5 53.2 38.5 29.0 22.0 17.1 13.0 8.8 α = 1.0 82.9 79.6 53.5 40.5 31.2 25.0 20.1 16.2 11.4 α = 2.0 84.8 80.1 54.6 39.2 30.6 24.0 18.5 14.5 9.9 α = 4.0 85.1 80.4 53.6 39.1 29.5 22.8 17.8 13.1 9.5 α = 8.0 86.1 81.5 54.5 39.0 27.7 21.4 17.3 12.9 8.8 α = 1.0 80.8 77.2 53.9 39.9 31.4 24.8 20.1 15.5 11.6 α = 2.0 82.7 78.0 53.9 39.9 30.7 24.1 19.8 15.1 11.1 α = 4.0 82.1 78.3 52.6 39.1 30.1 22.6 17.4 13.7 10.0 α = 8.0 83.7 78.2 51.1 36.9 28.0 22.3 17.9 13.1 9.1 Table 8: Comparison of empirical accuracy and certified accuracy of cascaded smooth classifiers on CIFAR-10 across ablations on p0. We use σ {0.25, 0.5, 1.0} for cascaded smoothing. CIFAR-10 Empirical Certified accuracy at ε (%) Setups Clean 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 p0 = 0.50 84.8 81.1 52.8 37.2 27.9 21.7 16.8 11.8 8.5 p0 = 0.55 87.9 82.6 53.5 35.6 26.9 19.8 14.0 10.1 7.1 p0 = 0.60 90.7 83.9 52.7 34.5 23.9 16.9 11.9 8.3 6.4 p0 = 0.65 92.5 83.4 52.1 34.0 22.6 14.4 9.8 7.0 5.5 p0 = 0.70 94.1 83.0 52.6 31.6 19.2 12.2 8.2 6.3 4.7 p0 = 0.75 95.2 81.9 52.2 29.1 16.5 9.5 6.7 5.2 3.7 p0 = 0.80 96.0 79.5 47.6 25.0 14.0 8.0 5.9 4.6 2.4 p0 = 0.85 96.6 76.3 45.0 22.8 12.5 7.1 4.7 2.8 2.2 p0 = 0.90 97.2 72.1 42.5 20.0 11.7 4.6 2.8 1.9 1.0 p0 = 0.95 97.8 65.5 38.6 16.7 7.2 2.2 1.7 0.7 0.0 p(ˆy) := p ˆ f(y|x): in this way, one could use ˆf(x) as an additional prior to refine the clean confidence in cases when the two sources of confidence do not match. More specifically, we consider: log pf(ˆy|x) := log pβ ˆ f(y|x) + log pf(x|y) + C, where (11) log pβ ˆ f(y|x) := (1 β) log p ˆ f(y|x) + β p(y). (12) where C denotes the normalizing constant, and β > 0 is a hyperparameter which we fix as β = 0.1 in our experiment. In Table 6, we compare the accuracy of the proposed inference with the accuracy of standard classifiers derived from FT-CLIP [19] fine-tuned on CIFAR-10: overall, we observe that this ensemble of clean prediction with cascaded smoothed predictions could improve the corruption accuracy of FT-CLIP on CIFAR-10-C, namely by 93.4% 95.0%, while maintaining its clean accuracy. The gain can be considered significant in a sense that the baseline accuracy of 93.4% is already at the state-of-the-art level, e.g., according to Robust Bench [17]. CIFAR-10 Classes Plane Car Bird Cat Deer Dog Frog Horse Ship Truck Figure 6: Comparison of per-class histograms of critical σ: the minimum value of σ for each input where the smoothed confidence at true class falls below 0.5. We use the whole CIFAR-10 training samples for the analysis. Each image is smoothed with n = 1, 000 samples of Gaussian noise, while iterating σ within the range σ [0.0, 2.0]. Images with critical σ larger than 2.0 are truncated. C.3 Ablation study Effect of λ. In Table 7, we jointly ablate the effect of tuning λ {0.001, 0.01, 0.1} upon the choices of α {1.0, 2.0, 4.0, 8.0}, by comparing its certified accuracy of cascaded smoothing as well as their empirical accuracy on CIFAR-10 test set. Overall, we observe that increasing λ has an effect of improving robustness at large radii, with a slight decrease in the clean accuracy as well as the empirical accuracy, where using higher α could compensate this at some extent. Although we generally observe effectiveness of the proposed regularization for a wide range of λ, using higher values of λ, e.g., λ = 0.1, often affects the base diffusion model and results in a degradation in accuracy without a significant gain in its certified robustness. Effect of p0. Table 8, on the other hand, reports the effect of p0 we introduced for cascaded smoothing. Recall that we use p0 = 0.5 throughout our experiments by default. We observe that using a bit of higher values of p0, e.g., p0 = 0.6 could be beneficial to improve the clean accuracy of cascaded smoothing, at the cost of robustness at larger radii: for example, it improves the certified accuracy at r = 0.0 by 81.1 83.9. The overall certified robustness starts to decrease as p0 further increases, due to its effect of increasing abstention rate. One could still observe the consistent improvement in the empirical accuracy as p0 increases, i.e., by also considering abstain as a valid prediction, i.e., using high values of p0 could be useful in certain scenarios. We think a more detailed choice of p0 per σ can further boost the performance of cascaded smoothing, which is potentially an important practical consideration. Critical σ. Although it is not required for cascaded smoothing to guarantee the stability of smoothed confidences across σ (e.g., outside of {0.25, 0.5, 1.0} in our experiments), we observe that smoothed confidences from denoised smoothing usually interpolate smoothly between different σ s, at least for the diffusion denoised smoothing pipeline. In this respect, we consider a new concept of critical σ for each sample, by measuring the threshold of σ where its confidence goes below p0 = 0.5. Figure 6 examines this measure on CIFAR-10 training samples and plots its distribution for each class as histograms. Interestingly, we observe that the distributions are often significantly biased, e.g., the Ship class among CIFAR-10 classes relatively obtains much higher critical σ - possibly due to its peculiar background of, e.g., ocean. Per-σ distribution from cascading. To further verify that per-sample σ s from cascaded smoothing are diverse enough to achieve better robustness-accuracy trade-off, we examine in Figure 5 the actual histograms of the resulting σ {ABSTAIN, 0.25, 0.50, 1.00} on CIFAR-10, also varying p0 {0.5, 0.6, 0.7, 0.8, 0.9}. Overall, we observe that (a) the distributions of σx indeed widely cover the total range of σ for every p0 s considered, and (b) using higher p0 can further improve diversity of the distribution while its chance to be abstained can be also increased. D Technical Details D.1 Proof of Theorem 3.1 Our proof is based on the fact that any smoothed classifier ˆf composed with Φ 1( ) is always Lipschitz-continuous, which is shown by Salman et al. [57]: Lemma D.1 (followed by Salman et al. [57]). Let Φ(a) := 1 2π R a exp 1 2s2 ds be the cumulative distribution function of standard Gaussian. For any function f : Rd [0, 1], the map x 7 σ Φ 1(p ˆ fσ(x)) is 1-Lipschitz for any σ > 0. Given Lemma D.1, the robustness certificate by Cohen et al. [15] in (2) (of the main text) can be viewed as computing the region of p ˆ fσ(x + δ; y) > 0.5, i.e., where it is guaranteed to return y over other (binary) smoothed classifiers p ˆ fσ(x + δ; y ) of y = y . In general, as in our proposed cascaded smoothing, one can consider a more strict region defined by p ˆ fσ(x + δ; y) > p0 for some p0 0.5. Here, the region corresponds to a policy that ˆfσ abstains whenever p ˆ fσ(x) p0. Accordingly, the certified radius one can obtain from this policy becomes: Rp0( ˆfσ; x, y) := σ Φ 1(p ˆ fσ(x, y)) Φ 1(p0) . (13) With this notation, we restate Theorem 3.1 into a tighter form to show, which do not consider neither p nor p upon the knowledge of p ˆ fσk and implies Theorem 3.1 as a consequence: Theorem 3.1 (restated). Let ˆfσ1, , ˆfσK : X Y be smoothed classifiers with 0 < σ1 < < σK. Suppose casc(x; { ˆfσi}K i=1) =: ˆy Y halts at ˆfσk for some k. Then, it holds that casc(x + δ; { ˆfσi}K i=1) = ˆy for δ < R, where: R := min n Rp0( ˆfσk; x, ˆy), R o , and R := min y =ˆy k >k n min n 0, Rp0( ˆfσk ; x, y) oo . (14) Proof. By the definition of casc(x), one can ensure casc(x + δ) = ˆy if the following holds: (a) ˆfσk(x + δ) = ˆy with p ˆ fσk (x + δ, ˆy) > p0, and (b) p ˆ fσk (x + δ) < p0 or ˆfσk (x + δ) = ˆy, for every k = k + 1, , K. Remark that the condition (a) is satisfied by δ within δ < Rp0( ˆfσk; x, ˆy). Specifically, one can consider the certifiable region of δ at ˆfσk as follows: Rk = {δ X : δ < Rp0( ˆfσk; x, ˆy)}. (15) For the condition (b), on the other hand, note that the condition is equivalent to that: (b ) p ˆ fσk (x + δ; y) < p0 for y = ˆy, and for every k = k + 1, , K. Fix k {k + 1, , K}. From the Lipschitzedness of x 7 σ Φ 1(p ˆ fσ(x; y)) by Lemma D.1, it is clear to show that δ s inside the following region Rk satisfies (b ): n δ X : δ < min n 0, Rp0( ˆfσk ; x, y) oo (16) = δ X : δ < min y =ˆy n min n 0, Rp0( ˆfσk ; x, y) oo . (17) Lastly, the result is followed by considering the intersection of Rk s which represent a certifiable region where casc(x + δ; { ˆfσi}K i=1) = ˆy: R := Rk ( k k+1Rk ) = {δ X : δ < R}. (18) Once we have the restated version of Theorem 3.1 above, the original statement follows from an observation that the new radius by introducing either p or p, say R, always lower-bounds R of the above: i.e., R R for any choices of p and p, thus certifiable for the region as well. D.2 Certification and prediction For practical uses of cascaded smoothing, where the exact values of p ˆ fσk (x, y) are not available, one has to estimate the smoothed confidence values as per Theorem 3.1, i.e., by properly lower-bound (or upper-bound) the confidence values. More concretely, one is required to obtain (a) p ˆ fσk (x) at the halting stage k, and (b) p ˆ fσk (x, y) for k = k + 1, , K and y Y. For both estimations, we use Monte Carlo algorithm with n i.i.d. Gaussian samples, say {δi}, and perform a majority voting across {fσ(x + δi)} to obtain a histogram on Y of n trials: (a) To estimate a lower-bound p ˆ fσk (x), we follow the statistical procedure by Cohen et al. [15], namely CERTIFY(n, n0, α), which additionally takes two hyper-parameters n0 and α: here, n0 denotes the number of samples to initially make a guess to estimate a smoothed prediction, and α is the significance level of the Clopper-Pearson confidence interval [14]. Concretely, it sets p ˆ fσk as the lower confidence level of p ˆ fσk with coverage 1 α. (b) For p ˆ fσk (x, y), on the other hand, we first obtain voting counts from n Gaussian trials, say c(σk ) x , and adopt Goodman confidence interval to bound multinomial proportions [24] with the significance level α. Specifically, one can set p ˆ fσk (x, y) for y Y by the (1 α)-upper level of the (multinomial) confidence interval. This can be implemented by using, e.g., multinomial_proportions_confint of the statsmodels library. We make an additional treatment on the significance level α: unless the pipeline halts at the first stage K, cascaded smoothing makes a (K k + 1)-consecutive testing of confidence intervals to make a prediction. In attempt to maintain the overall significance level of this prediction as α, one can apply the Bonferroni correction [5] on the significance level of individual testing, specifically by considering α (K k+1) as the adjusted significance level (instead of using α) at the stage k. E Other Possible Designs In this appendix, we present two alternative designs of multi-scale randomized smoothing we have also considered other than cascaded smoothing proposed in the main text. Specifically, here we additionally propose two different policies of aggregating multiple smoothed classifiers, each of which dubbed as the (a) max-radius policy, and (b) focal smoothing, respectively. Overall, both design do provide valid certified guarantees combining the accuracy-robustness trade-offs across multiple smoothing factors, often attaining a better robustness even compared to cascaded smoothing when n, the number of noise samples, is large enough. The common drawback in the both approaches here, however, is the necessity to access every smoothed confidence, which can be crucial in practice in both terms of efficiency and accuracy. Specifically, for given K smoothed classifiers, this implies the exact K-times increase in the inference cost, which makes it much costly on average compared to cascaded smoothing. The need to access and compare with K smoothed confidence in practice also demands more accurate estimations of the confidence values to avoid abstaining. For example, these approaches can significantly degrade the certified accuracy in practice where n cannot be large unlike the certification phase: which again makes our cascaded smoothing more favorable over the approaches presented here. E.1 Max-radius policy Consider K smoothed classifiers, say ˆfσ1, , ˆfσK : X Y, where 0 < σ1 < < σK. The max-radius policy, say max R( ), can be another simple policy one can consider for given { ˆfσi}: max R(x; { ˆfσi}K i=1) := ˆfσ (x), where σ := arg max σi σi Φ 1 p ˆ fσi(x) . (19) Again, given the Lipschitzedness associated with the smoothed confidences p ˆ fσ(x, y), it is easy to check the following robustness certificate for max R(x; { ˆfσi}K i=1): Theorem E.1. Let ˆfσ1, , ˆfσK : X Y be smoothed classifiers with 0 < σ1 < < σK. If max R(x; { ˆfσi}K i=1) =: ˆy Y, it holds that max R(x + δ; { ˆfσi}K i=1) = ˆy for δ < R, where: σ Φ 1 p ˆ fσ (x) max y =ˆy i=1, ,K σi Φ 1 p ˆ fσi(x; y) Remark that the certified radius given by (20) recovers the guarantee of Cohen et al. [15] if K = 1. E.2 Focal smoothing In this method, coined focal smoothing, we take a somewhat different approach to achieve multi-scale smoothing. Here, now we assume two different levels of smoothing factor, say σ0 > σ1 > 0, and consider inputs that contains noise with spatially different magnitudes. Specifically, we consider a binary mask M {0, 1}d and an input corrupted by the following: focal(x, M; σ0, σ1) := x + (σ0 (1 M) + σ1 M) ϵ, (21) where denotes the element-wise product and ϵ N(0, I). In this way, each of corrupted input ˆx gets a hint to recover the original content of x by observing a less-corrupted region marked by M. Randomized smoothing over such anisotropic noise has been recently studied [22, 20]: in a nutshell, smoothing over noises specified by (21) could produce an ellipsoid-shaped certified region with two axes of length σ0Φ 1(p) and σ1Φ 1(p) for its smoothed confidence p, respectively. In terms of certified radius, however, this implies that one can still certify only the region limited by the shorter axes (of length σ1Φ 1(p)). It turns out that such a limitation can be avoided by considering an ensemble of K multiple orthogonal masks, say M1, , MK, so that the shorter axes of ellipsoids from the masks do not overlap each other. Specifically, we observe the following: Theorem E.2. Let M1, , MK {0, 1}d be K orthogonal masks and σ0 > σ1 > 0 be smoothing factors, and pk(x) be the smoothed confidence from inputs corrupted by focal(x, Mk; σ0, σ1), for a given binary classifier fk : Rd [0, 1]. Define ˆp := 1 k pk, and suppose ˆp(x) > 0.5. Then, it is certified that ˆp(x + δ) > 0.5 for δ < Rb , where b is the solution of the following optimization: minimize b Rb := 1 ak t2bk + 1 subject to X bk 0, k = 1, 2, , K, where ak := σ0Φ 1(pk) for k = 1, , K, and t2 := σ2 0 σ2 1 1 . In Algorithm 1, we propose an efficient 1-dimensional grid-search based optimization to solve (22), based on observations given as Lemma E.3 and E.4: each of which identifies the condition of bk when it attains either the maximum value or zero, respectively. By solving the optimization, one could obtain a certified radius R that the aggregated smoothing can guarantee consistent prediction. Lemma E.3. Consider the constrained optimization (22) with respect to b = (b1, , b K). Let Ma := {m [K] : am = maxk ak} and b be an optimal solution of (22). Then, it holds that b m = b m > 0 for any m, m Ma. Proof. First, we show that there exists at least one m Ma where b m is nonzero, i.e., b m > 0. Suppose the contrary, i.e., b k = 0 for all k Ma. Denote the corresponding objective value as R . Given the condition P k bk = 1, there always exists i / M where bi > 0. Now, consider swapping the allocations of bm and bi for some m Ma and denote the corresponding objective value R . Algorithm 1 Focal smoothing: A grid-search based optimization of (22) Require: ak > 0, k = 1, 2, . . . , K, tolerance ε > 0 Ensure: optimal budget assignment b , certified radius R 1: Find the maximum value amax = maxk ak 2: Identify the index set M such that ak = amax for k M 3: Set a grid of values for bm in the interval (0, 1]. 4: for each grid value of bm do 5: for k = 1, 2, . . . , K do 6: if k / M then 7: bk 0 if ak am (t2bm+1)3/2 else 1 t2 2/3 (t2bm + 1) 1 8: end if 9: end for 10: if | PK k=1 bk 1| ε then 11: R 1 K PK k=1 ak t2bk+1 12: Update b k, R bk, R if R < R 13: end if 14: end for Then, we have: t2bj + 1 + aj t2bk + 1 ak t2bk + 1 aj p t2bj + 1 1 t2bk + 1 1 t2bk + 1 1 p t2bk + 1 1 p given that ak > aj and 1 t2bk+1 = 1. This means R < R , contradicting the optimality of the initial solution b with bk = 0 for all k Ma. Thus, there must exist at least one b m > 0 for some m Ma. Next, we show b m = b m for any m, m Ma. Suppose there exist such m, m Ma where b m = b m . Due to the symmetry, one can construct another optimal allocation b by swapping the two allocations b m and b m , as it does not affect the optimality. Now consider another allocation b by averaging the two allocations: bm = bm = b m + b m 2 . (24) Let the corresponding objective value R. Note that R is strictly convex, which it implies that R < R , contradicting the optimality of the initial solution b . Thus, it holds b m = b m for any m, m Ma. Lemma E.4. Consider the constrained optimization (22) with b as an optimal solution. Denote B := b m = b m > 0 for all m, m Ma. Then, b k = 0 if and only if ak A (t2B+1)3/2 . Proof. Suppose b k = 0. From the Karush Kuhn Tucker (KKT) stationarity condition, we have akt2 K + λ µk = 0. Since λ, µk 0, we obtain λ akt2 K . Now, consider the stationarity condition for m M: At2 t2B+1)3 + λ = 0. Substituting the lower bound for λ, we have ak A (t2B+1)3/2 . Conversely, suppose ak A (t2B+1)3/2 . Then, b k > 0 contradict to that b k = A 2/3 (t2B + 1) 1 0, hence b k = 0.