# improving_robustness_using_generated_data__e74aca03.pdf Improving Robustness using Generated Data Sven Gowal*, Sylvestre-Alvise Rebuffi*, Olivia Wiles, Florian Stimberg, Dan Calian and Timothy Mann Deep Mind, London {sgowal,sylvestre}@deepmind.com Recent work argues that robust training requires substantially larger datasets than those required for standard classification. On CIFAR-10 and CIFAR-100, this translates into a sizable robust-accuracy gap between models trained solely on data from the original training set and those trained with additional data extracted from the 80 Million Tiny Images dataset (80M-TI). In this paper, we explore how generative models trained solely on the original training set can be leveraged to artificially increase the size of the original training set and improve adversarial robustness to ℓp norm-bounded perturbations. We identify the sufficient conditions under which incorporating additional generated data can improve robustness, and demonstrate that it is possible to significantly reduce the robust-accuracy gap to models trained with additional real data. Surprisingly, we show that even the addition of non-realistic random data (generated by Gaussian sampling) can improve robustness. We evaluate our approach on CIFAR-10, CIFAR-100, SVHN and TINYIMAGENET against ℓ and ℓ2 norm-bounded perturbations of size ϵ = 8/255 and ϵ = 128/255, respectively. We show large absolute improvements in robust accuracy compared to previous state-of-the-art methods. Against ℓ normbounded perturbations of size ϵ = 8/255, our models achieve 66.10% and 33.49% robust accuracy on CIFAR-10 and CIFAR-100, respectively (improving upon the state-of-the-art by +8.96% and +3.29%). Against ℓ2 norm-bounded perturbations of size ϵ = 128/255, our model achieves 78.31% on CIFAR-10 (+3.81%). These results beat most prior works that use external data. 1 Introduction Neural networks are being deployed in a wide variety of applications ranging from ranking content on the web [15] to autonomous driving [5] via medical diagnostics [22]. It has become increasingly important to ensure that deployed models are robust and generalize to various input perturbations. Unfortunately, the addition of imperceptible adversarial perturbations can cause neural networks to make incorrect predictions [9, 10, 27, 44, 64]. There has been a lot of work on understanding and generating adversarial perturbations [1, 4, 10, 64], and on building defenses that are robust to such perturbations [27, 49, 60, 82]. We note that while robustness and invariance to input perturbations is crucial to the deployment of machine learning models in various applications, it can also have broader negative impacts to society such as hindering privacy [63] or increasing bias [68]. The adversarial training procedure proposed by Madry et al. [49] feeds adversarially perturbed examples back into the training data. It is widely regarded as one of the most successful method to train robust deep neural networks [30], and it has been augmented in different ways with changes in the attack procedure [25], loss function [50, 82] or model architecture [76, 85]. We highlight the works by Carmon et al. [11], Najafiet al. [51], Uesato et al. [72], Zhai et al. [80] who simultaneously proposed the use of additional unlabeled external data. While the addition of external data helped boost robust accuracy by a large margin, progress in the setting without additional data has slowed 35th Conference on Neural Information Processing Systems (Neur IPS 2021). 2018 2019 2020 2021 Robust test accuracy Madry et al. (44.04%) Rice et al. (53.42%) (Early stopping) Zhang et al. (53.08%) Gowal et al. (57.20%) Ours (66.10%) Figure 1: Robust accuracy of models against AUTOATTACK [16] on CIFAR-10 with ℓ perturbations of size 8/255 displayed in publication order. Our method explores how generated data can be used to improve robust accuracy by +8.96% without using any additional external data. This constitutes the largest jump in robust accuracy in this setting. Our best model reaches a robust accuracy of 66.10% against AA+MT [30]. Figure 2: Overview of our approach. Our method initially trains a generative model and a non-robust classifier. The non-robust classifier is used to provide pseudo-labels to the generated data. Finally, generated and original training data are combined to train a robust classifier. (see Fig. 1). On CIFAR-10 [42] against ℓ perturbations of size ϵ = 8/255, the best known model obtains a robust accuracy of 65.87% when using additional data. The same model obtains a robust accuracy of 57.14% without this data [30]. As a result, we ask ourselves whether it is possible to leverage the information contained in the original training set to a greater extent. This manuscript challenges the status-quo. To the contrary of standard training where it is widely believed that generative models lack diversity and that the samples they produce cannot be used to train better classifiers [59], we demonstrate both theoretically and experimentally that these generated samples can be used to improve robustness (using the approach described in Fig. 2 and Sec. 3.3). We make the following contributions: We demonstrate in Sec. 3.2 that it is possible to use low-quality random inputs (sampled from a conditional Gaussian fit of the training data) to improve robust accuracy on CIFAR-10 against ℓ perturbations of size ϵ = 8/255 (+0.93% on a WRN-28-10) and provide a justification and sufficient conditions in Sec. 4. We leverage higher quality generated inputs (i.e., inputs generated by generative models solely trained on the original data), and study four recent generative models: the Denoising Diffusion Probabilistic Model (DDPM) [36], Style GAN2 [40], Big GAN [7] and the Very Deep Variational Auto-Encoder (VDVAE) [14] (Sec. 5). We show that DDPM samples cover most closely the real data distribution (as measured by the distance to the test set in the Inception feature space). Using images generated by the DDPM allows us to reach a robust accuracy of 66.10% on CIFAR-10 against ℓ perturbations of size ϵ = 8/255 (an improvement of +8.96% upon the stateof-the-art). Notably, our best CIFAR-10 models beat all techniques that use additional data (see Sec. 6) and constitutes one of the largest improvements ever made in the setting without additional data. As a consequence, we demonstrate that it is possible to avoid the use of 80M-TI [65] which has been withdrawn due to presence of offensive images.1 2 Related work Adversarial ℓp norm-bounded attacks. Since Biggio et al. [4], Szegedy et al. [64] observed that neural networks which achieve high accuracy are highly vulnerable to adversarial examples, the art of crafting increasingly sophisticated adversarial examples has received a lot of attention. Goodfellow et al. [27] proposed the Fast Gradient Sign Method (FGSM) which generates adversarial examples 1https://groups.csail.mit.edu/vision/Tiny Images/ with a single normalized gradient step. It was followed by R+FGSM [67], which adds a randomization step, and the Basic Iterative Method (BIM) [44], which takes multiple smaller gradient steps. Adversarial training as a defense. Adversarial training [49] is widely regarded as one of the most successful methods to train deep neural networks robust to such attacks. It has received significant attention and various modifications have emerged [25, 50, 76]. A notable work is TRADES [82], which balances the trade-off between standard and robust accuracy, and achieved state-of-the-art performance against ℓ norm-bounded perturbations on CIFAR-10. More recently, the work from Rice et al. [60] studied robust overfitting and demonstrated that improvements similar to TRADES could be obtained more easily using classical adversarial training with early stopping. Finally, Gowal et al. [30] highlighted how different hyper-parameters (such as network size and model weight averaging) affect robustness. Data-driven augmentations. Works, such as Auto Augment [18] and related Rand Augment [19], learn augmentation policies directly from data. These methods are tuned to improve standard classification accuracy and have been shown to work well on multiple datasets. Deep Augment [34] explores how perturbations of the parameters of pre-trained image-to-image models can be used to generate augmented datasets that provide increased robustness to common corruptions [32]. Similarly, generative models can be used to create novel views of images [37, 39, 57] by manipulating them in latent space. When optimized and used during training, these novel views reduce the impact of spurious correlations and improve accuracy [28, 73]. Most recently, Laidlaw et al. [45] proposed an adversarial training method based on bounding a neural perceptual distance (i.e., an approximation of the true perceptual distance). While these works make significant contributions towards improving generalization and robustness to semantic perturbations, they do not improve robustness to ℓp normbounded perturbations. Robustness to ℓp norm-bounded perturbations using generative modeling. Finally, we highlight works, such as Defense-GAN [61] or ME-Net [77], which leverage data modeling techniques to create stronger defenses against ℓp norm-bounded attacks. Unfortunately, these techniques are not as robust as they seem and are broken by adaptive attacks [2, 16, 66]. Overall, to the best of our knowledge, there is little [48] to no evidence that data augmentations or generative models can be used to improve robustness to ℓp norm-bounded attacks. In fact, generative models mostly lack diversity and it is widely believed that the samples they produce cannot be used to train classifiers to the same accuracy than those trained on original datasets [59]. We differentiate ourselves from earlier works by leveraging additional generated samples for training rather than modifying the defense procedure, and by establishing sufficient conditions under which such samples improve robustness. 3 Adversarial training using generated data The rest of this manuscript is organized as follows. In this section, we provide an overview of adversarial training, demonstrate using a motivational example that low-quality generated data can be leveraged to improve robustness to adversarial examples, and describe our method. In Sec. 4, we detail sufficient conditions that explain why generated samples can improve robustness and explore the limitations of our approach. In Sec. 5, we analyze four complementary and recent generative models in the context of our method. Finally, we provide experimental results in Sec. 6. 3.1 Adversarial training For classification tasks, Madry et al. [49] propose to find model parameters θ that minimize the adversarial risk: arg min θ E(x,y) D max δ S [f(x + δ; θ) = y] (1) where D is a data distribution over pairs of examples x and corresponding labels y, f( ; θ) is a model parametrized by θ, [ ] is the Iverson bracket notation and corresponds to the 0 1 loss, and S defines the set of allowed perturbations. For ℓp norm-bounded perturbations of size ϵ, the perturbation set is defined as Sp = {δ | δ p ϵ}. Hence, for ℓ norm-bounded perturbations S = S and for ℓ2 norm-bounded perturbations S = S2. In the rest of this manuscript, we use ϵp to denote ℓp norm-bounded perturbations of size ϵ (e.g., ϵ = 8/255). 0% 20% 40% 60% 80% 100% Proportion of real data ( ) Robust test accuracy conditional Gaussian-fit Figure 3: Low-quality random inputs can improve robustness. Panel (a) shows the robust test accuracy (against AA+MT [30]) of a WRN-28-10 against ϵ = 8/255 on CIFAR-10 trained with additional data randomly sampled from a class-conditional Gaussian fit of the training data. We compare how the proportion of original CIFAR-10 and generated images affects robustness (0% means generated samples only, while 100% means original CIFAR-10 train set only). Panel (b) shows some of the class-conditional Gaussian samples that are used during training. In practice, given a training set Dtrain, the adversarial training procedure replaces the 0 1 loss with the cross-entropy loss lce and is formulated as arg min θ E(x,y) Dtrain max δ S lce(f(x + δ; θ), y) . (2) 3.2 Generated data can improve robust generalization Data augmentations can reduce the generalization error of standard (non-robust) training [18, 19, 23, 81]. However, to the contrary of standard training, augmentations beyond random flips and crops [31] such as Cutout [23], mixup [81], Auto Augment [18] or Rand Augment [19] have been unsuccessful in the context of adversarial training [30, 60, 75]. The gap in robust accuracy between models trained with and without additional data suggests that common augmentation techniques, which tend to produce augmented views that are close to the original image they augment, are intrinsically limited in their ability to improve robust generalization. In other words, augmented samples are diverse (if the training set is diverse), but not complementary to the training set. This phenomenon is particularly exacerbated when training adversarially robust models which are known to require an amount of data polynomial in the number of input dimensions [62]. We hypothesize that, to improve robust generalization, it is critical to use additional training samples (augmented or generated) that are diverse and that complement the original training set (in the sense that these new samples should ideally come from the same underlying distribution as the training set but should not duplicate the training set). To test this hypothesis, we propose to use samples generated from a simple class-conditional Gaussian fit of the training data. By construction, such samples (shown in Fig. 3(b)) are extremely blurry but diverse. We proceed by fitting a multivariate Gaussian to each set of 5K training images corresponding to each class in CIFAR-10. For each class, we sample 100K images resulting in a new dataset of 1M datapoints (no further filtering is applied). In Fig. 3(a), we show the performance of various robust models trained by decreasing the proportion of real samples present in each batch from 100% (original data only) to 0% (generated data only). Decreasing this proportion reduces the importance of the original data. We observe that all proportions between 50% and 90% provide improvements in robust accuracy. Most surprisingly, the optimal proportion of 80% provides an absolute improvement of +0.93%, which is an improvement comparable in size to the ones provided by model weight averaging or TRADES [30]. As we show in Sec. 4, the drop in robust accuracy for proportions below 50% is expected in the capacity-limited regime. This experiment directly motivates our method. Given access to a pre-trained non-robust classifier f NR and an unconditional generative model approximating the true data distribution D by a distribution ˆD, we would like to train a robust classifier f( ; θ) parameterized by θ. We propose the following optimization problem: arg min θ α E (x,y) Dtrain max δ S lce(f(x + δ; θ), y) +(1 α) E x ˆ D max δ S lce(f(x + δ; θ), f NR(x)) (3) where α corresponds to a mixing factor that blends examples from the training set with those that are generated. When α is set to one, our method reverts back to the original adversarial training formulation in Eq. 2. When α is set to zero, our method only uses generated samples with their corresponding pseudo-labels. In practice, for efficiency, rather than generating samples on-the-fly, we pre-generate samples offline. Hence, both the original training set Dtrain and generated set ˆD contain a finite number of samples. We have the advantage, however, to be able to generate significantly more samples than present in the original training set. In App. B, we evaluate how varying the number of generated samples impacts adversarial robustness. Overall, the complete method is described in Fig. 2 and is composed of three steps: (i) it starts by training the non-robust classifier and generative model on the original training set (for CIFAR-10, that corresponds to 50K images only); (ii) then, the generated dataset is constructed by drawing samples from the generative model and pseudo-labeling them using the non-robust classifier; (iii) finally, the robust classifier is trained using both the original training set and the generated dataset using Eq. 3. 4 Randomness might be enough In this section, we formalize our notation (Sec. 4.1), and provide three sufficient conditions that explain why generated data can improve robustness (Sec. 4.2). In summary, (i) the pre-trained, non-robust classifier f NR used for pseudo-labeling must be accurate, (ii) the likelihood of sampling examples that are adversarial to this non-robust classifier must be low, and (iii) the generative model must be able to sample images from the true data distribution with non-zero probability. Given a ground-truth function f , we would like to find optimal parameters θ for f( ; θ ) that minimize the adversarial risk, θ = arg min θ Ex D max δ S [f(x + δ; θ) = f (x)] , (4) without access to the true data distribution D or the ground-truth classifier f . As such, we replace the distribution D with an approximated distribution ˆD (from a generative model) and use a pre-trained non-robust classifier f NR instead of f (see Sec. 3.3). This results in sub-optimal parameters ˆθ = arg min θ Ex ˆ D max δ S [f(x + δ; θ) = f NR(x)] . (5) We introduce the unknown probability measure µ corresponding to the true data distribution D and defined over the set of inputs A Rn (where n is the input dimensionality), as well as the known probability measure ˆµ corresponding to the approximated distribution ˆD. The set of relevant inputs X A (i.e., the set of realistic images for which we would like to enforce robustness) is the support of µ such that µ(X) = 1 and W X, µ(W) > 0 if W is non-empty. We assume that each input x X can be assigned a label y = f (x) where f : X 7 Y is the ground-truth classifier (only valid for realistic images) and Y 2Z is the set of labels. Finally, given a perturbation set S, we restrict labels such that there exists no realistic image within the perturbation set of another that has a different label; i.e., for x X, for all δ {δ S|x + δ X} we have f (x) = f (x + δ). 4.2 Limitations and sufficient conditions To understand the limitations of our approach, it is useful to think about idealized sufficient conditions that would allow the sub-optimal parameters ˆθ to approach the performance of the optimal parameters θ . First, we concentrate on the capacity-limited regime and later extrapolate to the infinite-capacity, infinite-compute regime to gain more insights. The first sufficient condition concerns the pre-trained non-robust classifier and holds for both regimes. Condition 1 (accurate non-robust classifier). The pre-trained non-robust classifier f NR : A 7 Y must be accurate on all realistic inputs x X: x X, f NR(x) = f (x). Indeed, if we had access to the true distribution D, Eq. 4 and Eq. 5 could be made equal by setting f (x) = f NR(x). 2 In the capacity-limited regime, when Cond. 1 is satisfied, the problem reduces to a robust generalization problem. This problem is widely studied [3, 21, 70] and one can show that the adversarial risk is bounded by the Wasserstein distance between the training distribution ˆD and true data distribution D (under mild assumptions) [46]. In other words, as ˆD approaches D, we expect the robust accuracy of f( ; ˆθ ) to approach the one of f( ; θ ). This intuitively leads to the second sufficient condition and Prop. 1. Condition 2 (accurate approximated distribution). The approximated data distribution ˆD and true data distribution D must be equivalent: µ(W) = ˆµ(W) for all measurable subset W X. Proposition 1 (capacity-limited regime). Cond. 1 and Cond. 2 are sufficient conditions that allow the sub-optimal parameters ˆθ to match the performance of the optimal parameters θ . Together, Cond. 1 and 2 provide sufficient conditions for the capacity-limited regime (see proof in Sec. E.1). Cond. 2 (and associated bounds from [46]) generally indicates that the robust accuracy of the classifier f( ; ˆθ ) should increase as the quality of the generative model that provides the approximated distribution ˆD improves. However, these two conditions do not provide a satisfying answer when it comes to understanding why seemingly random data can help improve robustness (as demonstrated in Sec. 3.2). To help our understanding, it is worth analyzing the consequence of increasing the capacity of f. In particular, in the infinite-capacity regime, Cond. 2 can be relaxed and replaced by the following two conditions, and Prop. 1 becomes Prop. 2. Condition 3 (unlikely adversarial examples). It is not possible to sample a point x ˆD outside the realistic set X such that it is adversarial to f NR: ˆµ(W) = 0 on the measurable subset W = {x + δ | x X, δ S, f NR(x + δ) = f NR(x)}. Condition 4 (sufficient coverage). The likelihood of any finite sample in the set of realistic inputs X obtained from ˆD should be non-zero under the measure ˆµ: ˆµ(W) > 0 for all open measurable subsets W X. Proposition 2 (infinite-capacity regime). Cond. 1, Cond. 3 and Cond. 4 are sufficient conditions that allow the sub-optimal parameters ˆθ to match the performance of the optimal parameters θ when the model f has infinite capacity. Cond. 3 enforces that labels are non-conflicting within the perturbation set of a realistic input,3 while Cond. 4 guarantees that realistic inputs appear with enough frequency during training. Together, Cond. 1, 3 and 4 do not only provide sufficient conditions for the infinite-capacity regime (see proof in Sec. E.1), but also explain why samples generated by a simple class-conditional Gaussian-fit can be used to improve robustness. Indeed, they imply that it is not necessary to have access to either the true data distribution or a perfect generative model when given enough compute and capacity. However, when compute and capacity are limited, it is critical that the optimization in Eq. 5 focuses on realistic inputs and that the distribution ˆD be as close as possible to the true distribution D. In practice, this translates to the fact that better generative models (such as DDPM) can be used to achieve better robustness. We have relegated a discussion about the theoretical impact of the mixing factor α in Sec. E.2. Briefly stated, increasing α improves the realism of training samples (since the training samples mostly come from the original training set), but comes at the cost of a reduction in complementarity with the training set. 2In Sec. 6, we show that sub-optimal parameters ˆθ that improve upon those obtained by Eq. 2 can be obtained even when the non-robust classifier f NR is not perfect. In our experiments, we use a classifier that achieves 95.68% on the CIFAR-10 test set. 3Unless the generative model is trained to produce adversarial examples, random sampling is unlikely to produce images that are adversarial [8]. In fact, even strong black-box adversarial attacks require thousands of model queries to find adversarial examples. Table 1: Complementarity and coverage of augmented and generated samples. We sample 10K images from the train set and various different generative models. For each sample in each set, we find its closest neighbor in Inception feature space (obtained after the pooling layer). To estimate complementarity, we report the proportion of samples with a nearest neighbor in either the train set, test set or the sampled set itself. To estimate coverage, we report the proportion of unique neighbors in the train and test set. We also include the IS and FID computed from 50K samples from each set and the robust accuracy obtained by a WRN-28-10 models trained on 1M samples (Sec. 6). COMPLEMENTARITY COVERAGE INCEPTION METRICS ROBUST SETUP TRAIN TEST SELF TRAIN TEST IS FID ACCURACY mixup [81] 90.34% 3.91% 5.75% 90.43% 45.61% 9.33 0.22 7.71 Class-conditional Gaussian-fit 0.13% 0.22% 99.65% 12.36% 12.24% 3.64 0.03 117.62 55.37% VDVAE [14] 11.97% 12.14% 75.89% 34.20% 33.76% 6.88 0.05 26.44 55.51% Big GAN [7] 14.97% 14.81% 70.22% 38.86% 39.06% 9.73 0.07 13.78 55.99% Style GAN2 [40] 28.13% 27.22% 44.65% 50.16% 48.29% 10.04 0.11 2.57 58.17% DDPM [36] 29.29% 29.17% 41.54% 49.07% 49.10% 9.50 0.14 3.15 60.73% 5 Generative models The derivations from Sec. 4 and the experiment performed in Sec. 3.2 strongly suggest that generative models, which are capable of creating novel images [54], are viable augmentation candidates for adversarial training. Generative models considered in this work. In this work, we limit ourselves to generative models that are solely trained on the original train set, as we focus on how to improve robustness in the setting without external data. We consider four recent and fundamentally different models: (i) Big GAN [7]: one of the first large-scale application of Generative Adversarial Networks (GANs) which produced significant improvements in Frechet Inception Distance (FID) and Inception Score (IS) on CIFAR-10 (as well as on IMAGENET); (ii) VDVAE [14]: a hierarchical Variational Auto Encoder (VAE) which outperforms alternative VAE baselines; (iii) Style GAN2 [40]: an improved version of Style GAN which borrows interesting properties from the style transfer literature; and (iv) DDPM [36]: a diffusion probabilistic model based on Langevin dynamics that reaches state-of-the-art FID on CIFAR-10.4 As we have done for the simpler class-conditional Gaussian-fit, for each model, we sample 100K images per class, resulting in 1M images in total (see App. D for details). Samples are shown in App. D. Analysis of complementary and coverage. In Table 1, we evaluate how close Cond. 2 and Cond. 4 are to be satisfied in practice. To do so, we sample 10K images from each generative model. We also sample 10K images from the CIFAR-10 training set, and apply mixup to them as a point of comparison.5 We observe that mixup achieves a similar IS to the Big GAN and DDPM models. In the left-most set of three columns, for each augmented or generated sample, we report whether its closest neighbor in the Inception6 feature space belongs to the train set, test set or the generated set itself (more details are available in App. D). An ideal generative model should create samples that are equally likely to be close to images from each set. We observe that mixup tends to produce samples that are too close to the original train set and that lack complementarity, potentially explaining its limited usefulness in terms of improving adversarial robustness. Meanwhile, generated samples (including those from the class-conditional Gaussian-fit) are much more likely to be close to images of the test set. We also observe that the DDPM neighbor distribution matches more closely the ideal uniform distribution. Images generated by Big GAN and VDVAE tend to have their nearest neighbor among themselves which indicates that these samples are either far from the train and test distributions or produce overly similar samples. Images generated by Style GAN2, which reach an FID of 2.57 and IS of 10.07 that are better than the DDPM scores, have a slightly worse neighbor distribution (indicating a slight memorization of the training set). The middle two columns measure the ratio of unique neighbors that are matched in the train and test set. This provides a rough approximation 4We use VDVAE, Style GAN2 and DDPM checkpoints available online and train our own Big GAN. 5According to prior work, mixup is unable to improve robust accuracy beyond the one obtained with random cropping/flipping when using early stopping [60]. Table 5 in the appendix shows more data augmentations. 6Using the LPIPS [84] feature space (see Table 4 in the appendix) provides similar conclusions. 36% 49% 61% 74% 89% 93% 96% Accuracy of non-robust classifier Robust accuracy (CIFAR-10) Res Net-18 WRN-28-10 WRN-70-16 (a) Condition 1 0% 1% 2% 5% 10%20%50%100% Proportion of Big GAN samples Robust accuracy (Big GAN) Res Net-18 WRN-28-10 WRN-70-16 (b) Condition 2 0 1 2 3 4 5 6 7 8 9 10 # classes seen from Big GAN Robust accuracy (Big GAN) Res Net-18 WRN-28-10 WRN-70-16 (c) Condition 4 Figure 4: Impact of violations of the sufficient conditions detailed in Sec. 4. We report the robust test accuracy (against AA+MT [30]) when training different model architectures against ϵ = 8/255. In panel (a), non-robust classifiers with different clean acccuracies are used for pseudo-labeling. In panel (b), we vary the mixture of training samples from a class-conditional Gaussian and a Big GAN distribution while the test distribution is the Big GAN distribution. In panel (c), we fix the proportion of samples from the class-conditional Gaussian to 99% and increase the number of classes from the Big GAN distribution seen during training (thus increasing coverage). of coverage. We observe a similar trend where samples from the DDPM seem to provide a better coverage of the true data distribution. Note that, while these numbers rely on an inaccurate distance measure (i.e., Euclidean distance in Inception feature space) and should be taken with a grain of salt, they correlate well with the results obtained from our experiments. For example, models trained with Style GAN2 samples obtain a lower robust accuracy than those trained with DDPM samples despite obtaining better FID and IS. 6 Experiments The experimental setup is explained in App. A. We use Residual Networks (Res Nets) and Wide Res Nets (WRNs) [31, 79] with Swish/Si LU [33] activations. We use stochastic weight averaging [38] with a decay rate of 0.995. For adversarial training, we use TRADES [82] with 10 Projected Gradient Descent (PGD) steps. We train for 400 CIFAR-10-equivalent epochs with a batch size of 1024 (i.e., 19K steps). We evaluate our models against AUTOATTACK [16] and MULTITARGETED [29], which is denoted AA+MT [30]. For comparison, we trained ten WRN-28-10 models on CIFAR-10 (without additional generated samples) against ϵ = 8/255. The resulting robust accuracy is 54.44 0.39%, thus showing a relatively low variance. Furthermore, as we will see, our best models are well clear of the threshold for statistical significance. On CIFAR-10 against ϵ = 8/255 without additional generated samples a Res Net-18 achieves a robust accuracy of 50.64% and a WRN-70-16 achieves 57.14%. Unless stated otherwise, all results pertain to CIFAR-10. 6.1 Sufficient conditions The first set of experiments probes how violations of Cond. 1, 2 and 4 impact robustness against ϵ = 8/255 (violations to Cond. 3 have an impact equivalent to those of Cond. 1). All experiments are summarized in Fig. 4 where we train models with increasing capacity. Non-robust classifier accuracy. In Fig. 4(a), we train models using 1M samples generated by the DDPM and vary the accuracy of the the pre-trained, non-robust classifier f NR. We evaluate the robust accuracy obtained on CIFAR-10 test set. We observe that robustness improves as the accuracy of f NR increases. Notably, even with the 74.47%-accurate non-robust classifier, the WRN-28-10 and WRN-70-16 obtain robust accuracies of 58.15% and 59.83%, respectively, and already improve upon the state-of-the-art (57.14% at the time of writing). Thus, validating that, in practice, it is not necessary to have access to a perfect non-robust classifier. Quality of the generative models. To analyze how the quality of the generative model influences robustness, we use the Big GAN to model the true data distribution. During training, we use samples generated from a mixture of the class-conditional Gaussian and Big GAN distributions; during testing, we evaluate on a separate subset of 10K unseen Big GAN samples. To probe Cond. 2, we 54.44% 55.99% 0% 20% 40% 60% 80% 100% Proportion of real data ( ) Robust test accuracy VDVAE Big GAN Style GAN DDPM Figure 5: Robust test accuracy obtained by training a WRN-28-10 against ϵ = 8/255 on CIFAR-10 when using additional data produced by different generative models. We compare how the ratio between original and generated images (i.e., α) affects robustness (0% means generated samples only, 100% means CIFAR-10 train set only). Table 2: Clean (without perturbations) and robust (under adversarial attack) accuracy obtained by different models (we pick the worst accuracy obtained by either AUTOATTACK or AA+MT). The accuracies are reported on the full test sets. For CIFAR-10, we test against ϵ = 8/255 and ϵ2 = 128/255. For CIFAR-100, SVHN and TINYIMAGENET, we test against ϵ = 8/255. * This model is trained for 2000 epochs on 100M samples. MODEL DATASET NORM CLEAN ROBUST Wu et al. [75] (WRN-34-10) 85.36% 56.17% Gowal et al. [30] (WRN-70-16) 85.29% 57.14% Ours (DDPM) (WRN-28-10) 85.97% 60.73% Ours (DDPM) (WRN-70-16) 86.94% 63.58% Ours (100M DDPM)* (Res Net-18) 87.35% 58.50% Ours (100M DDPM)* (WRN-28-10) 87.50% 63.38% Ours (100M DDPM)* (WRN-70-16) 88.74% 66.10% Wu et al. [75] (WRN-34-10) CIFAR-10 ℓ2 88.51% 73.66% Gowal et al. [30] (WRN-70-16) 90.90% 74.50% Ours (DDPM) (WRN-28-10) 90.24% 77.37% Ours (DDPM) (WRN-70-16) 90.83% 78.31% Cui et al. [20] (WRN-34-10) CIFAR-100 ℓ 60.64% 29.33% Gowal et al. [30] (WRN-70-16) 60.86% 30.03% Ours (DDPM) (WRN-28-10) 59.18% 30.81% Ours (DDPM) (WRN-70-16) 60.46% 33.49% Ours (without DDPM) (WRN-28-10) SVHN ℓ 92.87% 56.83% Ours (DDPM) (WRN-28-10) 94.15% 60.90% Ours (without DDPM) (WRN-28-10) TINYIMAGENET ℓ 51.56% 21.56% Ours (DDPM) (WRN-28-10) 60.95% 26.66% change the proportion of training samples from the class-conditional Gaussian. In effect, decreasing the proportion of such samples skews the mixed generative model (modeled by the mixture of Gaussian and Big GAN distributions) to produce more samples from the true distribution (modeled by the Big GAN distribution), thereby closing the gap between the approximated distribution ˆD and true distribution D. As expected, Fig. 4(b) demonstrates that, given enough capacity, models can significantly reduce the adversarial risk. Relationship between coverage and capacity. Similarly to Fig. 4(b), we use samples generated from a mixture of the class-conditional Gaussian and Big GAN distributions; during testing, we evaluate on a separate subset of 10K unseen Big GAN samples. To probe Cond. 4, we keep the proportion of samples from the class-conditional Gaussian distribution fixed at 99% and use the remaining 1% to include Big GAN samples corresponding to either 0, 1, ...or 10 classes (thereby increasing coverage). In other words, the coverage of the true data distribution D (given by the Big GAN) increases as the number of seen classes increases. However, the approximated distribution ˆD remains different from the true data distribution even when the coverage reaches all classes (as the proportion of Gaussian samples is fixed to 99%). We observe in Fig. 4(c) that the robust accuracy of models with lower capacity improves less drastically yielding a gap of 17.37% at full coverage between the Res Net-18 and WRN-70-16 models. This observation confirms that, with enough coverage, model capacity can compensate for the lack of a perfect generative model. Discussion. Overall, Fig. 4(b) shows that when Cond. 2 is satisfied, the difference between models reduces and capacity takes a secondary role (since all models can bring their adversarial risk close to zero). Fig. 4(c) shows that when Cond. 4 is satisfied (and Cond. 2 is not), capacity matters as we observe that larger models benefit more from increased coverage. Both figures point to the fact that the quality of the generative model becomes less important when the capacity of the classifiers increases (as long as coverage is sufficient). 6.2 State-of-the-art robust accuracy Effect of mixing factor (α). As done in Sec. 3.3, we vary the proportion α of original images in each batch for all generated datasets. Fig. 5 explores a wide range of proportions while training a WRN-28-10 against ϵ = 8/255 on CIFAR-10. Samples from all models improve robustness when mixed optimally, but only samples from the Style GAN2 and DDPM improve robustness significantly (+3.73% and +6.29%, respectively). It is also interesting to observe that, in the case of the DDPM, using 1M generated images is better than using the 50K images from the original train set only. While this may seem surprising, it can easily be explained if we assume that the DDPM produces many more high-quality, high-diversity images than the limited set of images present in the original data (c.f. [62]). We also observe that the optimal mixing factor is different for different generative models. Indeed, increasing α reduces the gap to the true data distribution at the cost of less complementarity with the original train set (see Sec. E.2). CIFAR-10. Table 2 shows the performance of models trained with 1M samples generated by the DDPM on CIFAR-10 against ϵ = 8/255 and ϵ2 = 128/255. Irrespective of their size, models trained with 1M DDPM samples surpass the current state-of-the-art in robust accuracy by a large margin (+6.44% and +3.81%). When using 100M DDPM samples (and training for 2000 epochs), we reach 66.10% robust accuracy against ϵ = 8/255 which constitutes an improvement of +8.96% over the state-of-the-art. In this setting, our smallest model (Res Net-18) surpasses state-of-the-art results obtained by much larger models (e.g., WRN-70-16). Most remarkably, despite not using any external data, against ϵ = 8/255, our best model beats all Robust Bench [17] entries that used external data (see Table 6 in the appendix). Generalization to other datasets (CIFAR-100, SVHN and TINYIMAGENET). Finally, to evaluate the generality of our approach, we evaluate it on CIFAR-100, SVHN [53] and TINYIMAGENET [26]. We train two new DDPM on the train set of CIFAR-100 and SVHN and sample 1M images from each. For TINYIMAGENET, we use a DDPM trained on IMAGENET [24] at 64 64 resolution and restricted samples to the 200 classes of TINYIMAGENET. The results are shown in Table 2. On CIFAR-100, our best model reaches a robust accuracy of 33.49% and improves noticeably upon the state-of-the-art by +3.46% (in the setting that does not use any external data). On SVHN, in the same table, we compare models trained without and with DDPM samples. Again, the addition of DDPM samples significantly improves robustness, with the robust accuracy improving by by +4.07%. On TINYIMAGENET, the improvement is +5.10%. 7 Conclusion. Using generative models, we posit and demonstrate that generated samples provide a greater diversity of augmentations that allow adversarial training to go well beyond the current state-of-the-art. Our work provides novel insights into the effect of diversity and complementarity on robustness, which we hope can further our understanding of robustness. All our models and generated datasets are available online at https://github.com/deepmind/deepmind-research/tree/master/adversarial_robustness. [1] A. Athalye and I. Sutskever. Synthesizing robust adversarial examples. Int. Conf. Mach. Learn., 2018. [2] A. Athalye, N. Carlini, and D. Wagner. Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. Int. Conf. Mach. Learn., 2018. [3] A. N. Bhagoji, D. Cullina, and P. Mittal. Lower bounds on adversarial robustness from optimal transport. Adv. Neural Inform. Process. Syst., 2019. [4] B. Biggio, I. Corona, D. Maiorca, B. Nelson, N. Šrndi c, P. Laskov, G. Giacinto, and F. Roli. Evasion attacks against machine learning at test time. In Machine Learning and Knowledge Discovery in Databases, 2013. [5] M. Bojarski, D. Del Testa, D. Dworakowski, B. Firner, B. Flepp, P. Goyal, L. D. Jackel, M. Monfort, U. Muller, J. Zhang, X. Zhang, J. Zhao, and K. Zieba. End to end learning for self-driving cars. NIPS Deep Learning Symposium, 2016. [6] J. Bradbury, R. Frostig, P. Hawkins, M. J. Johnson, C. Leary, D. Maclaurin, and S. Wanderman Milne. JAX: composable transformations of Python+Num Py programs, 2018. URL http: //github.com/google/jax. [7] A. Brock, J. Donahue, and K. Simonyan. Large scale gan training for high fidelity natural image synthesis. Int. Conf. Learn. Represent., 2018. [8] T. Brunner, F. Diehl, M. T. Le, and A. Knoll. Guessing smart: Biased sampling for efficient black-box adversarial attacks. ar Xiv preprint ar Xiv:1812.09803, 2019. [9] N. Carlini and D. Wagner. Adversarial examples are not easily detected: Bypassing ten detection methods. In Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security, pages 3 14. ACM, 2017. [10] N. Carlini and D. Wagner. Towards evaluating the robustness of neural networks. In 2017 IEEE Symposium on Security and Privacy, 2017. [11] Y. Carmon, A. Raghunathan, L. Schmidt, J. C. Duchi, and P. S. Liang. Unlabeled data improves adversarial robustness. In Adv. Neural Inform. Process. Syst., 2019. [12] H. Chang, T. D. Nguyen, S. K. Murakonda, E. Kazemi, and R. Shokri. On adversarial bias and the robustness of fair machine learning. ar Xiv preprint ar Xiv:2006.08669, 2020. [13] D. Chen, N. Yu, Y. Zhang, and M. Fritz. Gan-leaks: A taxonomy of membership inference attacks against generative models. ar Xiv preprint ar Xiv:1909.03935, 2020. [14] R. Child. Very deep vaes generalize autoregressive models and can outperform them on images. Int. Conf. Learn. Represent., 2021. [15] P. Covington, J. Adams, and E. Sargin. Deep neural networks for You Tube recommendations. In Proceedings of the 10th ACM Conference on Recommender Systems, 2016. [16] F. Croce and M. Hein. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. ar Xiv preprint ar Xiv:2003.01690, 2020. [17] F. Croce, M. Andriushchenko, V. Sehwag, N. Flammarion, M. Chiang, P. Mittal, and M. Hein. Robustbench: a standardized adversarial robustness benchmark. ar Xiv preprint ar Xiv:2010.09670, 2020. [18] E. D. Cubuk, B. Zoph, D. Mane, V. Vasudevan, and Q. V. Le. Autoaugment: Learning augmentation policies from data. IEEE Conf. Comput. Vis. Pattern Recog., 2019. [19] E. D. Cubuk, B. Zoph, J. Shlens, and Q. V. Le. Randaugment: Practical automated data augmentation with a reduced search space. IEEE Conf. Comput. Vis. Pattern Recog., 2020. [20] J. Cui, S. Liu, L. Wang, and J. Jia. Learnable boundary guided adversarial training. ar Xiv preprint ar Xiv:2011.11164, 2020. [21] D. Cullina, A. N. Bhagoji, and P. Mittal. Pac-learning in the presence of adversaries. In Adv. Neural Inform. Process. Syst., 2018. [22] J. De Fauw, J. R. Ledsam, B. Romera-Paredes, S. Nikolov, N. Tomasev, S. Blackwell, H. Askham, X. Glorot, B. O Donoghue, D. Visentin, G. v. d. Driessche, B. Lakshminarayanan, C. Meyer, F. Mackinder, S. Bouton, K. Ayoub, R. Chopra, D. King, A. Karthikesalingam, C. O. Hughes, R. Raine, J. Hughes, D. A. Sim, C. Egan, A. Tufail, H. Montgomery, D. Hassabis, G. Rees, T. Back, P. T. Khaw, M. Suleyman, J. Cornebise, P. A. Keane, and O. Ronneberger. Clinically applicable deep learning for diagnosis and referral in retinal disease. In Nature Medicine, 2018. [23] T. De Vries and G. W. Taylor. Improved regularization of convolutional neural networks with cutout. ar Xiv preprint ar Xiv:1708.04552, 2017. [24] P. Dhariwal and A. Nichol. Diffusion models beat GANs on image synthesis. ar Xiv preprint ar Xiv:2105.05233, 2021. [25] Y. Dong, F. Liao, T. Pang, H. Su, J. Zhu, X. Hu, and J. Li. Boosting Adversarial Attacks with Momentum. IEEE Conf. Comput. Vis. Pattern Recog., 2018. [26] A. K. Fei-Fei Li and J. Johnson. The Tiny Image Net dataset. 2017. URL https://www.kaggle.com/c/ tiny-imagenet. [27] I. J. Goodfellow, J. Shlens, and C. Szegedy. Explaining and harnessing adversarial examples. Int. Conf. Learn. Represent., 2015. [28] S. Gowal, C. Qin, P.-S. Huang, T. Cemgil, K. Dvijotham, T. Mann, and P. Kohli. Achieving Robustness in the Wild via Adversarial Mixing with Disentangled Representations. ar Xiv preprint ar Xiv:1912.03192, 2019. [29] S. Gowal, J. Uesato, C. Qin, P.-S. Huang, T. Mann, and P. Kohli. An Alternative Surrogate Loss for PGD-based Adversarial Testing. ar Xiv preprint ar Xiv:1910.09338, 2019. [30] S. Gowal, C. Qin, J. Uesato, T. Mann, and P. Kohli. Uncovering the limits of adversarial training against norm-bounded adversarial examples. ar Xiv preprint ar Xiv:2010.03593, 2020. [31] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. IEEE Conf. Comput. Vis. Pattern Recog., 2016. [32] D. Hendrycks and T. Dietterich. Benchmarking neural network robustness to common corruptions and perturbations. In Int. Conf. Learn. Represent., 2018. [33] D. Hendrycks and K. Gimpel. Gaussian error linear units (gelus). ar Xiv preprint ar Xiv:1606.08415, 2016. [34] D. Hendrycks, S. Basart, N. Mu, S. Kadavath, F. Wang, E. Dorundo, R. Desai, T. Zhu, S. Parajuli, M. Guo, D. Song, J. Steinhardt, and J. Gilmer. The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization. ar Xiv preprint ar Xiv:2006.16241, 2020. URL https://arxiv.org/pdf/2006.16241. [35] T. Hennigan, T. Cai, T. Norman, and I. Babuschkin. Haiku: Sonnet for JAX, 2020. URL http://github.com/deepmind/dm-haiku. [36] J. Ho, A. Jain, and P. Abbeel. Denoising diffusion probabilistic models. Adv. Neural Inform. Process. Syst., 2021. [37] E. Härkönen, A. Hertzmann, J. Lehtinen, and S. Paris. GANSpace: Discovering Interpretable GAN Controls. ar Xiv preprint ar Xiv:2004.02546, 2020. [38] P. Izmailov, D. Podoprikhin, T. Garipov, D. Vetrov, and A. G. Wilson. Averaging Weights Leads to Wider Optima and Better Generalization. Uncertainty in Artificial Intelligence, 2018. [39] A. Jahanian, L. Chai, and P. Isola. On the "steerability" of generative adversarial networks. In Int. Conf. Learn. Represent., 2019. [40] T. Karras, S. Laine, M. Aittala, J. Hellsten, J. Lehtinen, and T. Aila. Analyzing and improving the image quality of stylegan. In IEEE Conf. Comput. Vis. Pattern Recog., 2020. [41] D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. Int. Conf. Learn. Represent., 2015. [42] A. Krizhevsky, V. Nair, and G. Hinton. The CIFAR-10 dataset. 2014. URL http://www.cs.toronto. edu/kriz/cifar.html. [43] S. Kumar, V. Bitorff, D. Chen, C. Chou, B. Hechtman, H. Lee, N. Kumar, P. Mattson, S. Wang, T. Wang, Y. Xu, and Z. Zhou. Scale mlperf-0.6 models on google tpu-v3 pods. ar Xiv preprint ar Xiv:1909.09756, 2019. [44] A. Kurakin, I. Goodfellow, and S. Bengio. Adversarial examples in the physical world. ICLR workshop, 2016. [45] C. Laidlaw, S. Singla, and S. Feizi. Perceptual adversarial robustness: Defense against unseen threat models. In Int. Conf. Learn. Represent., 2021. [46] J. Lee and M. Raginsky. Minimax statistical learning with wasserstein distances. Adv. Neural Inform. Process. Syst., 2018. [47] I. Loshchilov and F. Hutter. SGDR: stochastic gradient descent with warm restarts. In Int. Conf. Learn. Represent., 2017. [48] D. Madaan, J. Shin, and S. J. Hwang. Learning to generate noise for robustness against multiple perturbations. ar Xiv preprint ar Xiv:2006.12135, 2020. [49] A. Madry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu. Towards deep learning models resistant to adversarial attacks. Int. Conf. Learn. Represent., 2018. [50] M. Mosbach, M. Andriushchenko, T. Trost, M. Hein, and D. Klakow. Logit Pairing Methods Can Fool Gradient-Based Attacks. ar Xiv preprint ar Xiv:1810.12042, 2018. [51] A. Najafi, S.-i. Maeda, M. Koyama, and T. Miyato. Robustness to adversarial perturbations in learning from incomplete data. Adv. Neural Inform. Process. Syst., 2019. [52] Y. Nesterov. A method of solving a convex programming problem with convergence rate o(1/k2). In Sov. Math. Dokl, 1983. [53] Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A. Y. Ng. Reading digits in natural images with unsupervised feature learning. 2011. [54] Open AI. Dall-e: Creating images from text, 2021. URL https://openai.com/blog/dall-e. [55] T. Pang, X. Yang, Y. Dong, H. Su, and J. Zhu. Bag of tricks for adversarial training. ar Xiv preprint ar Xiv:2010.00467, 2020. [56] T. Pang, X. Yang, Y. Dong, K. Xu, H. Su, and J. Zhu. Boosting Adversarial Training with Hypersphere Embedding. Adv. Neural Inform. Process. Syst., 2020. [57] A. Plumerault, H. L. Borgne, and C. Hudelot. Controlling generative models with continuous factors of variations. In Int. Conf. Learn. Represent., 2020. [58] B. T. Polyak. Some methods of speeding up the convergence of iteration methods. USSR Computational Mathematics and Mathematical Physics, 1964. [59] S. Ravuri and O. Vinyals. Classification Accuracy Score for Conditional Generative Models. ar Xiv preprint ar Xiv:1905.10887, 2019. [60] L. Rice, E. Wong, and J. Z. Kolter. Overfitting in adversarially robust deep learning. Int. Conf. Mach. Learn., 2020. [61] P. Samangouei, M. Kabkab, and R. Chellappa. Defense-gan: Protecting classifiers against adversarial attacks using generative models. In Int. Conf. Learn. Represent., 2018. [62] L. Schmidt, S. Santurkar, D. Tsipras, K. Talwar, and A. Madry. Adversarially Robust Generalization Requires More Data. Adv. Neural Inform. Process. Syst., 2018. [63] L. Song, R. Shokri, and P. Mittal. Privacy risks of securing machine learning models against adversarial examples. In Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, pages 241 257, 2019. [64] C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. Goodfellow, and R. Fergus. Intriguing properties of neural networks. Int. Conf. Learn. Represent., 2014. [65] A. Torralba, R. Fergus, and W. T. Freeman. 80 million tiny images: a large dataset for nonparametric object and scene recognition. IEEE Trans. Pattern Anal. Mach. Intell., 2008. [66] F. Tramer, N. Carlini, W. Brendel, and A. Madry. On adaptive attacks to adversarial example defenses. ar Xiv preprint ar Xiv:2002.08347, 2020. [67] F. Tramèr, A. Kurakin, N. Papernot, I. Goodfellow, D. Boneh, and P. Mc Daniel. Ensemble Adversarial Training: Attacks and Defenses. ar Xiv preprint ar Xiv:1705.07204, 2017. [68] F. Tramèr, J. Behrmann, N. Carlini, N. Papernot, and J.-H. Jacobsen. Fundamental Tradeoffs between Invariance and Sensitivity to Adversarial Perturbations. ar Xiv preprint ar Xiv:2002.04599, 2020. [69] D. Tsipras, S. Santurkar, L. Engstrom, A. Turner, and A. Madry. Robustness may be at odds with accuracy. ar Xiv preprint ar Xiv:1805.12152, 2018. [70] Z. Tu, J. Zhang, and D. Tao. Theoretical analysis of adversarial learning: A minimax approach. In Adv. Neural Inform. Process. Syst. 2019. [71] J. Uesato, B. O Donoghue, A. v. d. Oord, and P. Kohli. Adversarial Risk and the Dangers of Evaluating Against Weak Attacks. Int. Conf. Mach. Learn., 2018. [72] J. Uesato, J.-B. Alayrac, P.-S. Huang, R. Stanforth, A. Fawzi, and P. Kohli. Are labels required for improving adversarial robustness? Adv. Neural Inform. Process. Syst., 2019. [73] E. Wong and J. Z. Kolter. Learning perturbation sets for robust machine learning. In Int. Conf. Learn. Represent., 2021. [74] B. Wu, J. Chen, D. Cai, X. He, and Q. Gu. Do wider neural networks really help adversarial robustness? ar Xiv preprint ar Xiv:2010.01279, 2021. [75] D. Wu, S.-t. Xia, and Y. Wang. Adversarial weight perturbation helps robust generalization. Adv. Neural Inform. Process. Syst., 2020. [76] C. Xie, Y. Wu, L. van der Maaten, A. Yuille, and K. He. Feature denoising for improving adversarial robustness. IEEE Conf. Comput. Vis. Pattern Recog., 2019. [77] Y. Yang, G. Zhang, D. Katabi, and Z. Xu. Me-net: Towards effective adversarial robustness with matrix estimation. In Int. Conf. Mach. Learn., 2019. [78] S. Yun, D. Han, S. J. Oh, S. Chun, J. Choe, and Y. Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. Int. Conf. Comput. Vis., 2019. [79] S. Zagoruyko and N. Komodakis. Wide residual networks. Brit. Mach. Vis. Conf., 2016. [80] R. Zhai, T. Cai, D. He, C. Dan, K. He, J. Hopcroft, and L. Wang. Adversarially Robust Generalization Just Requires More Unlabeled Data. ar Xiv preprint ar Xiv:1906.00555, 2019. [81] H. Zhang, M. Cisse, Y. N. Dauphin, and D. Lopez-Paz. mixup: Beyond empirical risk minimization. Int. Conf. Learn. Represent., 2018. [82] H. Zhang, Y. Yu, J. Jiao, E. P. Xing, L. E. Ghaoui, and M. I. Jordan. Theoretically Principled Trade-off between Robustness and Accuracy. Int. Conf. Mach. Learn., 2019. [83] J. Zhang, J. Zhu, G. Niu, B. Han, M. Sugiyama, and M. Kankanhalli. Geometry-aware instancereweighted adversarial training. In Int. Conf. Learn. Represent., 2021. [84] R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang. The unreasonable effectiveness of deep features as a perceptual metric. ar Xiv preprint ar Xiv:1801.03924, 2018. [85] D. Zoran, M. Chrzanowski, P.-S. Huang, S. Gowal, A. Mott, and P. Kohl. Towards Robust Image Classification Using Sequential Attention Models. IEEE Conf. Comput. Vis. Pattern Recog., 2019. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] We claim that adversarial robustness can be improved using generated data. Experiments demonstrate that this is possible and that the improvements are significant (Sec. 6). (b) Did you describe the limitations of your work? [Yes] The sufficient conditions detailed in Sec. 4 provide an idealized situation that we degrade in the experimental section. In particular, the method is intrinsically limited by the accuracy of the non-robust classifier f NR as well as the ability to generate realistic inputs. (c) Did you discuss any potential negative societal impacts of your work? [Yes] We highlight potential pitfalls in the introduction and related work. We summarize these and highlights them more broadly in App. G (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] We provide two set of sufficient conditions. The first set, composed of Cond. 1 and Cond. 2, applies to the limited-capacity and limited-compute regime, while the second set, composed of Cond. 1, Cond. 3 and Cond. 2, applies to the infinite-capacity and infinite-compute regime. Other assumptions are detailed in Sec. 4.1. (b) Did you include complete proofs of all theoretical results? [Yes] Proof are included in the main text and rely on either demonstrating that Eq. 4 and Eq. 5 can be made equal when Cond. 1 and Cond. 2 are satisfied (limited-capacity) or showing that the combination of Cond. 1, Cond. 3 and Cond. 4 yields a perfectly robust classifier on the subset of inputs that includes the set of realistic inputs (infinite-capacity regime). 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] The training and evaluation code is available at https://github.com/deepmind/deepmind-research/tree/master/ adversarial_robustness. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] All details are in App. A. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [No] Adversarial training is compute-intensive and the negative-impact of running more experiments outweighed the benefits. We use the same evaluation setup than proposed in [30]. That is, we always trained two models for each hyper-parameter setup and chose the best model according to a separate validation set of 1024 images (the average degradation in robustness between both models measured on a subset of 10 hyper-parameters setups is -0.12% in absolute robust accuracy). We also evaluated all models with one of the strongest set of adversarial attacks. Additionally, our baseline model was trained ten times resulting in a standard deviation of 0.39% and all reported improvements in robust accuracy are well beyond 2 standard deviations. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] All details are in App. A. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] We use CIFAR10 [42], CIFAR-100 [42], SVHN [53] and TINYIMAGENET [26]. (b) Did you mention the license of the assets? [No] Please refer to citations for details on licensing. All datasets are available for non-commercial use. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] We generate additional data using generative models. These are available online at https://github.com/deepmind/deepmind-research/tree/master/adversarial_robustness. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]