# adversarial_unlearning_reducing_confidence_along_adversarial_directions__10fdfb7d.pdf Adversarial Unlearning: Reducing Confidence Along Adversarial Directions Amrith Setlur1, Benjamin Eysenbach1 Virginia Smith1 Sergey Levine2 1 Carnegie Mellon University 2 UC Berkeley Supervised learning methods trained with maximum likelihood objectives often overfit on training data. Most regularizers that prevent overfitting look to increase confidence on additional examples (e.g., data augmentation, adversarial training), or reduce it on training data (e.g., label smoothing). In this work we propose a complementary regularization strategy that reduces confidence on self-generated examples. The method, which we call RCAD (Reducing Confidence along Adversarial Directions), aims to reduce confidence on out-of-distribution examples lying along directions adversarially chosen to increase training loss. In contrast to adversarial training, RCAD does not try to robustify the model to output the original label, but rather regularizes it to have reduced confidence on points generated using much larger perturbations than in conventional adversarial training. RCAD can be easily integrated into training pipelines with a few lines of code. Despite its simplicity, we find on many classification benchmarks that RCAD can be added to existing techniques (e.g., label smoothing, Mix Up training) to increase test accuracy by 1 3% in absolute value, with more significant gains in the low data regime. We also provide a theoretical analysis that helps to explain these benefits in simplified settings, showing that RCAD can provably help the model unlearn spurious features in the training data. 1 Introduction Supervised learning techniques typically consider training models to make accurate predictions on fresh test examples drawn from the same distribution as training data. Unfortunately, it is well known that maximizing the likelihood of the training data alone may result in overfitting. Prior work broadly considers two approaches to combat this issue. Some methods train on additional examples, e.g., generated via augmentations [53, 10, 66] or adversarial updates [14, 36, 4]. Others modify the objective by using alternative losses and/or regularization terms (e.g., label smoothing [56, 39], Mix Up [69], robust objectives [65, 22]). In effect, these prior approaches either make the model s predictions more certain on new training examples or make the distribution over potential models less certain. Existing regularization methods can be seen as providing a certain inductive bias for the model, e.g., the model s weights should be small (i.e., weight decay), the model s predictions should vary linearly between training examples (i.e., Mix Up). In this paper we identify a different inductive bias: the model s predictions should be less confident on out-of-distribution inputs that look nothing like the training examples. We turn this form of inductive bias into a simple regularizer, whose benefits are complementary to existing regularization strategies. To instantiate such a method, we must be able to sample out-of-distribution examples. For this, we propose a simple approach: generating adversarial examples [38, 14] using very large step sizes (orders-of-magnitude larger than traditional adversarial Correspondence can be sent to asetlur@cs.cmu.edu. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Cross entropy Cross entropy Cross entropy Entropy Adversarial images (small step) Original images Original images Self-generated images by RCAD (large step) Adversarial training RCAD regularization CIFAR-100 Tiny Imagenet 60 Test Accuracy DA (Shorten et al.) LS (Szegedy et al.) FGSM (Goodfellow et al.) ME-ADA (Zhao et al.) RCAD (Ours) Figure 1: Reducing confidence along adversarial directions (RCAD) is a simple and efficient regularization technique to improve test performance. (Left) For RCAD, examples are generated by taking a large step (10 typical for adversarial examples) along the gradient direction. We see that generated images thus look very different from the original, with accentuated spurious components responsible for the model s flipped predictions on adversarial images. (Right) RCAD achieves greater test accuracy than data augmentation (DA), label smoothing (LS), and methods that minimize cross-entropy on adversarial examples: adversarial training via FGSM [14] and ME-ADA [73]. training [36]). Hence, we first perturb the training points using the training loss gradient, and then maximize predictive entropy on these adversarially perturbed examples. In contrast to adversarial training, our method does not try to robustify the model to output the original label, but rather regularizes it to have reduced confidence on examples generated via a large step size (Figure 1a). As shown in Figure 1b, this can lead to significant improvements in in-distribution test accuracy unlike adversarial training [36, 37, 5, 4], which tends to decrease in-distribution test performance [48, 70, 61]. Compared with semi-supervised learning methods [15, 64], our method does not require an additional unlabeled dataset (it generates one automatically), and different from prior works [64, 74] it also trains the model to be less confident on these examples. As the self-generated samples in our procedure no longer resemble in-distribution data, we are uncertain of their labels and train the model to predict a uniform distribution over labels on them, following the principle of maximum entropy [25]. A few prior works [56, 43, 52, 11] have also considered using the same principle to prevent the memorization of training examples, but only increase entropy on iid sampled in-distribution labeled/unlabeled data. In contrast, we study the effect of maximizing entropy on self-generated out-of-distribution examples along adversarial directions. The main contribution of this work is a training procedure we call RCAD: Reducing Confidence along Adversarial Directions. RCAD improves test accuracy (for classification) and log-likelihood (for regression) across multiple supervised learning benchmarks. Importantly, we find that the benefits of RCAD are complementary to prior methods: Combining RCAD with alternative regularizers (e.g., augmentation [53, 66], label smoothing [56], Mix Up training [69]) further improves performance. Our method requires adding 5 lines of code and is computationally efficient (with training time at most 1.3 standard). We provide a theoretical analysis that helps to explain RCAD s benefits in a simplified setting, showing that RCAD can unlearn spurious features in the training data, thereby improving accuracy on unseen examples. 2 Related Work Below we survey common regularization techniques in machine learning, as well as other methods that utilize entropy maximization in training for various purposes (differing from our own). Data augmentation. A common strategy to improve test accuracy (particularly on image tasks) is to augment the training data with corruptions [53] or surface level variations [10, 66] (e.g., rotations, random crops). Some methods [37, 36, 14] further augment data with imperceptibly different adversarial examples or interpolated samples [69]. Others [46, 42] sample new examples from generative models that model the marginal density of the input distribution. Also related are semisupervised methods, which assume access to an extra set of unlabeled data and train the model to have more confident predictions on them [15, 37]. All these methods minimize entropy [16] on the augmented or unlabeled data to improve generalization [69, 73]. In contrast, our method maximizes entropy on perturbed samples along the adversarial direction. Table 1: Regularization objectives: We summarize prior works that employ adversarial examples or directly regularize model s predictions pw(y | x) along with the scalar hyperparameters (in [ ]) associated with each. name objective cross entropy minw P x,y ˆ D log pw(y | x) label smoothing [ϵ] [39] minw P x,y ˆ D (1 ϵ) log pw(y | x) + P y =y ϵ |Y| 1 log pw(y | x) Adv. training [α] [36] minw P x,y ˆ D log pw(y | x α sign( x log pw(y | x))) ME-ADA [α, β] [73] minw P (x,y) ˆ D ˆ D log pw(y | x) where, for a distance metric Cw:(X Y) (X Y) 7 R ˆ D {( x, y) | x supx0 log pw(x0 | y) + αHw(x0) βCw((x0, y), (x, y)), (x, y) ˆ D} RCAD (ours) [α, λ] minw P x,y ˆ D ( log pw(y | x) λ Hw(x α x log pw(y | x))) Adversarial training. Deep learning models are vulnerable to worst-case perturbations that can flip the model s predictions [14, 50]. Adversarial training was proposed to improve robustness to such attacks by reducing worst-case loss in small regions around the training samples [36, 33, 48]. More recently, Zhao et al. [73] proposed maximum entropy adversarial data augmentation (MEADA), which uses an information bottleneck principle to identify worst-case perturbations that both maximize training loss and predictive entropy. Similar to adversarial training ME-ADA still minimizes cross-entropy to output the same label on new points. There are two key differences between above methods and RCAD; (i) rather than minimizing cross entropy loss on adversarial examples, we increase model s uncertainty on the self-generated examples; and (ii) we take much larger steps so large that unlike adversarial images, the generated example is no longer similar to the original one (Figure 1a). While prior work has successfully used adversarial examples to improve OOD detection [2] or adversarial robustness [36, 33, 48], we show that these changes allow our method to improve the standard test accuracy, a metric that adversarial training typically makes worse [48, 60, 70]. Further, RCAD has a lower (at most 1/5th) computational cost compared to multistep adversarial training procedures [73, 70]. Robust objectives. In order to improve robustness to noise and outliers in the data, a common approach is to modify the objective by considering risk averse loss functions [e.g., 22, 49, 57] or incorporating regularization terms such as the l2/l1 norms [31, 59]. Our method is similar to these approaches in that we propose a new regularization term. However, whereas most regularization terms are applied to the model s weights or activations, ours is directly applied to the model s predictions. Another effective loss function for classification problems is label smoothing [56, 8, 26], which uniformly increases model s predictive uncertainty in a region around training samples [12]. In contrast, RCAD increases entropy only on examples generated along the adversarial direction that has been shown to comprise of spurious features [23, 7], thereby unlearning them. Entropy maximization. Finally, we note that our work builds upon prior work that draws connections between entropy maximization in supervised [52, 11, 43] and reinforcement learning [17]. For example, Pereyra et al. [43] apply a hinge form of confidence penalty directly on training data which is similar to label smoothing in principle and performance (Table 2 in [43]). Other works like [11, 52] also adapt the principle of entropy maximization but do so either on additional unlabeled data or a subset of the training samples. More recently [45] show that maximizing entropy on interpolated samples from the same class improves out-of-distribution uncertainty quantification. While we also minimize cross-entropy loss on training data, in contrast to the above we maximize entropy on samples generated along the adversarial direction. Our experimental results also span a wider set of benchmarks and presents significant gains (+1 3%) complementary to methods like label smoothing. We also theoretically analyze our objective for the class of linear predictors and show how RCAD can mitigate vulnerability to spurious correlations. 3 RCAD: Reducing Confidence Along Adversarial Directions We now introduce our regularization technique for reducing confidence along adversarial directions (RCAD). This section describes the objective and an algorithm for optimizing it; Section 5 presents a more formal discussion of RCAD and in Section 4 we provide our empirical study. Notation. We are given a training dataset ˆD (x(i), y(i)) N i=1 where x(i) X, yi Y, are sampled iid from a joint distribution D over X Y. We use ˆD to denote both the training dataset and an empirical measure over it. The aim is to learn a parameterized distribution pw(y | x), w W where the learnt model is typically obtained by maximizing the log-likelihood over ˆD. Such a solution is referred to as the maximum likelihood estimate (MLE): ˆwmle arg maxw W E ˆ D log pw(y | x). In the classification setting, we measure the performance of any learned solution ˆw using its accuracy on the test (population) data: ED 1(arg maxy p ˆ w(y | x) = y) . Solely optimizing the log-likelihood over the training data can lead to poor test accuracy, since the estimate ˆwmle can overfit on noise in ˆD and fail to generalize to unseen samples. Many algorithms aim to mitigate overfitting by either designing suitable loss functions replacing the log-likelihood objective, or by augmenting the training set with additional data (see Section 2). Our main contribution is a new data-dependent regularization term that will depend not just on the model parameters but also on the training dataset. We reduce confidence on out-of-distribution samples that are obtained by perturbing the original inputs along the direction that adversarially maximizes training loss. In the paragraphs that follow we describe the methodology and rationale behind our objective that uses the following definition of model s predictive entropy when Y is a discrete set: y Y pw(y | x) log pw(y | x). In cases where Y is continuous, for example in regression tasks, we will use the differential form of predictive entropy: R Y pw(y | x) log pw(y | x) dy. RCAD: Reducing Confidence along Adversarial Directions def rcad_loss(x, y, α, λ): loss = model(x).log_prob(y) x_adv = x + α * loss.grad(x) entropy = model(x_adv).entropy() return loss - λ * entropy Reducing Confidence Along Adversarial Directions. The key idea behind RCAD is that models should not only make accurate predictions on the sampled data, but also make uncertain predictions on examples that are very different from training data. We use directions that adversarially maximize the training loss locally around the training points to construct these out-of-distribution examples that are different from the training data. This is mainly because adversarial directions have been known to comprise of spurious features [23] and we want to regularize the model in a way that makes it uncertain on these features. We first describe how these examples are generated, and then describe how we train the model to be less confident. We generate an out-of-distribution example x by taking a large gradient of the MLE objective with respect to the input x, using a step size of α > 0: x x α x log pw(y | x) (1) We train the model to make unconfident predictions on these self-generated examples by maximizing the model s predictive entropy. We add this entropy term Hw ( x) to the standard MLE objective, weighted by scalar λ > 0, yielding the final RCAD objective: ˆwrcad arg max w W E ˆ D [log pw(y | x) + λ Hw (x α x log pw(y | x))] (2) In adversarial training, adversarial examples are generated by solving a constrained optimization problem [36, 48]. Using a first-order approximation, the adversarial example x generated by one of the simplest solvers [14] has a closed form resembling Equation 1. We note that RCAD is different from the above form of adversarial training in two ways. First, RCAD uses a much larger step size (10 larger), so that the resulting example no longer resembles the training examples (Figure 1a). Second, whereas adversarial training updates the model to be more confident on the new example, RCAD trains the model to be less confident. Our experiments in Section 4 (Figure 4b) show that these differences are important for improving test accuracy. Informal understanding of RCAD. For image classification, image features that are pure noise and independent of the true label can still be spuriously correlated with the labels for a finite set of iid drawn examples [54]. Such features are usually termed spurious features [72]. Overparameterized neural networks have a tendency to overfit on spurious features in the training examples [68, 51]. In order to generate unrealistic out-of-distribution samples, we pick examples that are far away from the true data point along the adversarial direction because (i) adversarial directions are comprised of noisy features that are spuriously correlated with the label on a few samples [23]; and (ii) maximizing entropy on samples with an amplified feature forces the model to quickly unlearn that feature. By DA LS Cut Out+Mix Mix Up 28 Test Accuracy CIFAR-100-2k (a) base base + RCAD DA LS Cut Out+Mix Mix Up 58 68 CIFAR-100-10k (b) base base + RCAD DA LS Cut Out+Mix Mix Up 74 86 CIFAR-100 (c) base base + RCAD DA LS Cut Out+Mix Mix Up 94 Test Accuracy CIFAR-10 (d) base base + RCAD DA LS Cut Out+Mix Mix Up 64 72 Tiny Imagenet (e) base base + RCAD DA LS Cut Out+Mix Mix Up 97.0 99.0 SVHN (f) base base + RCAD Figure 2: Main results on supervised image classification benchmarks: We plot the mean test accuracy and 95% confidence intervals over 10 independent runs for models trained with base methods: Data Augmentation (DA), Label Smoothing (LS), Cut Out/Cut Mix (Cut Out+Mix) augmentation, Mix Up and compare them with the test accuracies of the models trained with the RCAD objective in Equation 2, in addition to the base methods. using a large step size α we exacerbate the spurious features in our self-generated example x. When we maximize predictive entropy over our generated examples we then force the model to unlearn these spurious correlations. Hence, the trained model can generalize better and achieve higher prediction performance on unseen test examples. In Section 5 we build on this informal understanding to present a more formal analysis on the benefits of RCAD. 4 Experiments: Does RCAD Improve Test Performance? Our experiments aim to study the effect that RCAD has on test accuracy, both in comparison to and in addition to existing regularization techniques and adversarial methods, so as to understand the degree to which its effect is complementary to existing methods. Benchmark datasets. We use six image classification benchmarks. In addition to CIFAR-10, CIFAR-100 [30], SVHN [41] and Tiny Imagenet [34], we modify CIFAR-100 by randomly subsampling 2,000 and 10,000 training examples (from the original 50,000) to create CIFAR-100-2k and CIFAR-100-10k. These smaller datasets allow us to study low-data settings where we expect the generalization gap to be larger. CIFAR-100(-2k/10k) share the same test sets. If the validation split is not provided by the benchmark, we hold out 10% of our training examples for validation. Implementation details2. Unless specified otherwise, we train all methods using the Res Net-18 [19] backbone, and to accelerate training loss convergence we clip gradients in the l2 norm (at 1.0) [71, 18]. We train all models for 200 epochs and use SGD with an initial learning rate of 0.1 and Nesterov momentum of 0.9, and decay the learning rate by a factor of 0.1 at epochs 100, 150 and 180 [10]. We select the model checkpoint corresponding to the epoch with the best accuracy on validation samples as the final model representing a given training method. For all datasets (except CIFAR-100-2k and CIFAR-100-10k for which we used 32 and 64 respectively) the methods were trained with a batch size of 128. For details on algorithm specific hyperparameter choices refer to Appendix B. Baselines. The primary aim of our experiments is to study whether entropy maximization along the adversarial direction shrinks the generalization gap. Hence, we explore baselines commonplace in deep learning that either (i) directly constrain the model s predictive distribution on observed samples (label smoothing [39]) or (ii) implicitly regularize the model by training on additional images generated in an adversarial manner. For the first, our main comparisons are to label smoothing [56]), standard data augmentation [53], cutout data augmentation [66, 10], and Mix Up [69] training. For the second, we compare with adversarial training[36] that uses FGSM [14] to perturb the inputs. Additionally, we compare RCAD with two recent approaches that use adversarial examples in different ways: adversarial data augmentation (ADA) [62] and maximum entropy adversarial data augmentation (ME-ADA) [73]. We summarize these baselines in Table 1. 2Code for this work can be found at https://github.com/ars22/RCAD-regularizer. Tiny Imagenet CIFAR-10 CIFAR-100 SVHN -4.0% Test acc. improvement over ERM Adv. training LS on Adv ( = 0.02) LS on Adv ( = 0.5) ME-ADA RCAD ME-ADA+LS RCAD+LS 0.5k 1k 2k 5k 10k 50k (full) Training data size (in 1000s) Test acc. improvement over ERM adversarial training ME-ADA Figure 3: (Left) How should we use adversarial examples to improve test accuracy? We compare RCAD with adversarial baselines including label smoothing on adversarial samples (LS on Adv), measuring the improvement in test accuracy relative to ERM. We also compare RCAD with ME-ADA when combined with label smoothing. (Right) RCAD is more effective in low-data regime. We compare the test accuracy improvement over ERM for baselines: Label Smoothing (LS), Data Augmentation (DA), adversarial training, and ME-ADA with RCAD + LS on different sub-samples of CIFAR-100 training set (0.5k 50k). We find RCAD achieves the largest gains in the low data regime. In both plots, we plot the mean and 95% confidence intervals over 10 independent runs. 4.1 How much does RCAD improve test accuracy in comparison and in addition to other regularization methods? Figure 2 presents the main empirical findings of RCAD over six image classification benchmarks and across four baseline regularizers. Across all datasets, we observe that training with the RCAD objective improves the test accuracy over the baseline method in 22/24 cases. The effects are more pronounced on datasets with fewer training samples. For example, on CIFAR-100-2k, adding RCAD on top of label smoothing boosts the performance by 3%. These results also show that the benefits of RCAD are complementary to prior methods RCAD + Mix Up outperforms Mix Up, and RCAD + augmentation/smoothing outperforms both. We test for statistical significance using a 1-sided p-value, finding that p 1e 3 in 22/24 comparisons. In summary, these results show that our proposed regularizer is complementary to prior regularization techniques. 4.2 How effectively does RCAD improve test accuracy compared to adversarial training? Our next set of experiments compares different ways of using adversarial examples. RCAD maximizes the model s predictive entropy on unlabeled examples along the adversarial direction (obtained with a large step-size 0.5). In contrast, other methods we compare against (Adversarial Training [36], ADA [62], ME-ADA [62]) minimize the cross entropy loss on examples obtained by adversarially perturbing the inputs without changing the original labels (using a much smaller step size 0.05). Additionally, we look at the baseline that performs label smoothing on these adversarial samples (LS on Adv) a relaxed version of RCAD. We evaluate all methods by measuring their test accuracy improvement over empirical risk minimization (ERM), noting that some of these methods were proposed to optimize robustness, a different metric. We show results in Figure 3a and note that RCAD outperforms adversarial training and LS on Adv with small step-sizes (α = 0.02), ADA and ME-ADA by significant margins on all benchmarks. We numerically verify that RCAD statistically outperforms the best baseline ME-ADA by computing 1-sided p-values (p=0.009 on CIFAR-10, p < 1e 4 on others). Next, we look at LS on Adv with large α = 0.5. Since this is the same value of α used by RCAD for most datasets, this method is equivalent to performing label smoothing on examples generated by RCAD. Since label smoothing is a relaxed form of entropy maximization it is not surprising that the performance of this method is similar to (if not better than) RCAD on a few benchmarks. Finally, when both RCAD and ME-ADA are equipped with label smoothing, both methods improve, but the benefit of RCAD persists (p = 0.0478 on CIFAR-10, p < 1e 4 on others). 4.3 How effective is RCAD in the low training data regime? Since supervised learning methods are more susceptible to overfitting when training samples are limited [13], we analyze the effectiveness of RCAD in the low-data regime. To verify this, we sample Test acc. improvement over ERM ME-ADA (WRN) RCAD (WRN) 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 1.1 1.2 RCAD step size ( ) Test accuracy Negative log-likelihood (NLL) on test data Method Boston Concrete Energy Wine Yacht ERM 2.85 0.04 ME-ADA 3.01 0.04 RCAD (ours) 2.60 Figure 4: RCAD can be used with larger networks and is also effective on regression tasks. (Left) We compare the improvements in test accuracy (over ERM) for RCAD and ME-ADA when all methods use Wide Res Net 28-10 [67]. (Center) For CIFAR-100-2k, we show the effect on test accuracy of the step size α that RCAD takes while generating out-of-distribution examples along the adversarial direction. (Right) We compare the negative log-likelihood on test samples for RCAD with baselines ERM and ME-ADA on five regression datasets from the UCI repository. For the above we show the mean and 95% confidence intervals across 10 runs. small training datasets of size 0.5k, 1k, 2k, 5k and 10k from the 50,000 training samples in CIFAR100. The test set for each is the same as that of CIFAR-100. We train the baseline regularizers and RCAD on each of these datasets and compare the observed improvements in test accuracies relative to ERM. We show our results for this experiment in Figure 3b. Straight away we observe that RCAD yields positive improvements in test accuracy for any training dataset size. In contrast, in line with the robust overfitting phenomena described in Raghunathan et al. [48, 47], Zhang et al. [70], adversarial training (with FGSM [14]) is found to be hurtful in some settings. Next we observe that compared to typical regularizers for supervised learning like label smoothing and data augmentation, the regularization effect of RCAD has a stronger impact as the size of the training data reduces. Notice that RCAD has a steeper upward trend (right left) compared to the next best method label smoothing, while outperforming each baseline in terms of absolute performance values. 4.4 Additional Results and Ablations In addition to the experiments below on wider architectures, effect of step size α, and regression tasks, we conduct more analysis and experiments for RCAD (e.g., evaluating its robustness to adversarial perturbations and distribution shifts). For these and other experimental details refer to Appendix B, C. Results on a larger architecture. We compare the test accuracies of RCAD and ME-ADA (most competitive baseline from Res Net-18 experiments) when trained with the larger backbone Wide Res Net 28-10 [67] (WRN) on CIFAR-100 and its derivatives. We plot these test accuracies relative to ERM trained with WRN in Figure 4a. Clearly, the benefit of RCAD over ME-ADA still persists albeit with slightly diminished absolute performance differences compared to Res Net-18 in Figure 3a. Effect of step size α on test accuracy. In Figure 4b we plot the test accuracy of RCAD on CIFAR100-2k as we vary the step size α. We see that RCAD is effective only when the step size is sufficiently large (> 0.5), so that new examples do not resemble the original ones. When the step size is small (< 0.5), the perturbations are small and the new example is indistinguishable from the original, as in adversarial training. Our theory in Section 5 also agrees with this observation. Results on regression tasks. On five regression datasets from the UCI repository we compare the performance of RCAD with ERM and ME-ADA in terms of test negative log-likelihood (NLL) ( is better). To be able to compute NLL, we model the predictive distribution pw(y | x) as a Gaussian and each method outputs two parameters (mean, variance of pw(y | x)) at each input x. For the RCAD regularizer we use the differential form of entropy. Results are shown in Figure 4c from which we see that RCAD matches/improves over baselines on 4/5 regression datasets. 5 Analysis: Why Does RCAD Improve Test Performance? In this section, we present theoretical results that provide a more formal explanation as to why we see a boost in test accuracy when we regularize a classifier using RCAD. We analyze the simplified problem of learning l2-regularized linear predictors in a fully specified binary classification setting. Specifically, our analysis will show that, while linear predictors trained with standard ERM can have y = 1 y = + 1 ERM RCAD ( = 0.70) RCAD ( = 0.01) 0 25 50 75 100 125 150 175 200 Epochs Individual weight norms True feature wt. norm |w1| Noisy feature wt. norm ||w2||2 ERM Training RCAD Training ( = 0.7) 0 25 50 75 100 125 150 175 200 Epochs Test Accuracy Train Accuracy ERM Training RCAD Training ( = 0.7) Figure 5: RCAD corrects ERM solution: (Left) We plot the training data and high-dimensional decision boundaries for all three estimates ˆwerm and ˆwrcad with α = 0.01, 0.70. Here, we plot the linear boundary projected onto the set of planes {x1, xi}i=9 i=2 spanned by the true feature x1 and noisy features x2, . . . , x8. Across the ERM and RCAD training iterations, we plot the weight norms along the true feature |w1| and noisy dimensions w2 2 in the (Center) plot, and on the (Right) we plot the train/test accuracies. a high dependence on spurious features, further optimization with the RCAD objective will cause the classifier to unlearn these spurious features. The intuition behind our result can be summarized as follows: when RCAD generates examples after taking a large step along adversarial directions, a majority of the generated examples end up lying close to the decision boundary along the true feature, and yet a portion of them would significantly depend on the noisy spurious features. Thus, when RCAD maximizes the predictive entropy on these new examples, the classifier weights that align with the spurious components are driven toward smaller values. Detailed proofs for the analysis that follows can be found in Appendix A. Setup. We consider a binary classification problem with a joint distribution D over (x, y) X {0, 1} where X Rd+1 with x = [x1, x2] , x1 R, x2 Rd and for some β > 0, D is: y Unif{ 1, 1}, x1 N(β y, σ2 1), x2 N(0, Σ2), and let Σ σ2 1 0 0 Σ2 We borrow this setup from Chen et al. [7] which analyzes self-training/entropy minimization on unlabeled out-of-distribution samples. While their objective is very different from RCAD, their setup is relevant for our analysis since it captures two kinds of features: x1, which is the true univariate feature that is predictive of the label y, and x2, which is the high dimensional noise, that is not correlated with the label under the true distribution D, but w.h.p. is correlated with the label on finite sampled training dataset ˆD = {(x(i), y(i))}n i=1 Dn. We wish to learn the class of homogeneous linear predictors W {w Rd+1 : w, x = w1 x1 + w2 x2, w 2 1}. For mathematical convenience we make three modifications to the RCAD objective in Equation 2 that are common in works analyzing logistic classifiers and entropy based objectives [55, 7]: (i) we minimize the margin loss lγ(w; (x, y)) max(0, γ y w, x ) (instead of the negative log-likelihood) over D with some γ > 0; (ii) we consider the constrained optimization problem in Equation 4 as opposed to the Lagrangian form of RCAD in Equation 2 [35]; and (iii) we use Lemma 5.1 and replace Hw(x+α xlγ(w; (x, y))) with exp( | w, x+α xlγ(w; (x, y)) |). Note that the optimal w W that has the best test accuracy and lowest population margin loss is given by w [1, 0, . . . , 0] . max w M ˆ D(w) E ˆ D exp( | w, x + α xlγ(w; (x, y)) |) s.t. E ˆ D 1(lγ(w; (x, y)) > 0) ρ/2, w W (4) Lemma 5.1 ([6, 55], informal). Hw(x) = Hbin((1 + exp( w, x )) 1) where Hbin(p) is the entropy of Bern(p) distribution. Thus Hw(x) exp( | w, x |), as both exhibit same tail behavior. We analyze the RCAD estimate ˆwrcad w(T ) returned after T iterations of projected gradient ascent on the non-convex objective in Equation 4, initialized with w(0) satisfying the constraints in Equation 4. Given the projection ΠS onto the convex set S, and learning rate η > 0, the update rule for w(t) is as follows: w(t+1) = w(t) + η w(t)M ˆ D(w), and w(t+1) = Π w 2 1 w(t+1). Since the initialization w(0) satisfies the constraint in Equation 4, it has a low margin loss on training samples and using margin based generalization gaps [27] we can conclude w.h.p. w(0) has low test error ( ρ). This, in turn tells us that w(0) should have learnt the true feature x1 to some extent (Lemma 5.2). Intuitively, if the classifier is not trained at all, then adversarial directions would be less meaningful since they may include a higher component of the true feature x1, as opposed to noisy x2. To obtain w(0), we simply minimize loss lγ on ˆD using projected (onto w 2 1) gradient descent and the ERM solution ˆwerm serves as the initialization w(0) for RCAD s gradient ascent iterations. Note that ˆwerm still significantly depends on spurious features (see Figure 5). Furthermore, this dependence is unavoidable and worse when n is small, or when the noise dimension d is large. Lemma 5.2 (true feature is partially learnt before RCAD). If w(0) satisfies constraint in Equation 4, and when n > β2+log(1/δ) Σ op+ Σ γ2ρ2 , with probability 1 δ over ˆD, w(0) 1 erfc 1(2ρ) 2σmin( Σ) β . We are now ready to present our main results in Theorem 5.3 which states that after T = O(log(1/ϵ)) gradient ascent iterations of optimizing the RCAD objective ˆwrcad is ϵ close to w , since it unlearns the spurious feature ( w(T ) 2 2 ϵ) and amplifies the true feature (|w(T ) 1 | 1 ϵ2). Note that Theorem 5.3 suggests a minimum step size α in arriving at a sufficient condition for RCAD to unlearn spuriousness and improve test performance This is also in line with our empirical findings (Figure 4b). Since α = Θ(γ), most of the points generated by RCAD lie close to the decision boundary along x1, except for the ones with noisy features that are correlated with the classifier weights. Theorem 5.3 ( ˆwrcad w ). If β = Ω( Σ op) and c0, K > 0, such that β > α > max(γ, Σ op) and γ + c0 σmin( Σ) β, then with n > β2+log(1/δ) Σ op+ Σ 2σmin( Σ) 2 + log 1/δ ϵ4 , after T = O (log(1/ϵ)) gradient ascent iterations, |w(T ) 1 | 1 ϵ2 and w(T ) 2 2 ϵ, with probability 1 δ over D . The key idea behind our guarantee above is an inductive argument. We show that, if the class separation β and step size α are sufficiently larger than any noise variance, then RCAD monotonically increases the weight norm along the true feature and monotonically decreases it along the noisy dimensions with each gradient ascent update. This is because at any given instant the gradient of RCAD objective MD(w) with respect to w1 always points in the same direction, while the one with respect to w2 is sufficiently anti-correlated with the direction of w2. We formalize this argument in Lemma 5.4. It is also easy to see why the update will improve test accuracy monotonically, thus satisfying the constraint in Equation 4 with high probability. Lemma 5.4 (w1, w2 update). If α, β, γ and sample size n satisfy the noise conditions in Theorem 5.3, then MD(w) w(t) 1 , w(t) 1 > 0, and | w(t+1) 1 | > |w(t) 1 |. On the other hand, c1 > 0 such that w(t) 2 MD(w(t)), w(t) 2 c1 w(t) 2 2 2. Consequently η, such that w(t+1) 2 2/ w(t) 2 2 < 1. Empirically validating our theoretical results in the toy setup. Now, using the same setup as our theoretical study, we check if taking the ERM solution and updating it with RCAD truly helps in unlearning the spurious feature x2 and amplifying the true feature x1. With d = 8, γ = 2.0, β = 0.5, σ1 = 0.1, Σ = Id we collect training dataset ˆD of size 20 according to the data distribution in Equation 3. First, we obtain the ERM solution ˆwerm by optimizing the margin loss on ˆD for 100 projected gradient descent iterations. Then, with ˆwerm as initialization we optimize the RCAD objective in Equation 4 for 100 additional projected gradient ascent iterations to obtain RCAD estimate ˆwrcad. For RCAD, we try both small (α = 0.01) and large (α = 0.7) values of the step size. Using the results of this study in Figure 5 we shall now verify our main theoretical claims: (i) RCAD unlearns spurious features learnt by ERM (Theorem 5.3): RCAD training with large step size α = 0.7 corrects ERM s decision boundaries across all noisy dimensions i.e., ˆwrcad has almost no dependence on noisy dimensions {xi}i=8 i=2 (Figure 5a). On the other hand, the ERM solution has a significant non-zero component across all noisy dimensions (except x7). (ii) RCAD is helpful only when step size is large enough (Theorem 5.3): In Figure 5a we see that RCAD estimate with small step size α = 0.01 has a higher dependence on spurious features compared to ERM. Thus optimizing RCAD with a very small α leads to an even poor solution than ERM. This is because, the perturbed data point still has a significant component of the true feature, which the model is forced to unlearn when RCAD tries to maximize entropy over the perturbed point. (iii) The weight norms |w1| and w2 2 change monotonically through RCAD iterations (Lemma 5.4): In Figure 5b we see that the norm along the true feature |w1| increases monotonically through RCAD iterations, whereas w2 2 decreases monotonically, until |w1| 1 and w2 2 0. Thus, RCAD successfully recovers optimal w . Based on this, we also see the test accuracy increase monotonically through RCAD iterations; reaching 100% by the end of it (Figure 5c). In contrast, since the ERM solution depends on spurious features, its accuracy does not improve beyond 70% in the first 100 iterations. We find that training ERM for more iterations only improves training accuracy. The analysis above explains why RCAD improves test performance in the linear setting and it is inline with our intuition of RCAD s ability to unlearn spurious features. To check if our intuition also generalizes to deeper networks we further study RCAD s behaviour in a non-linear setting using a different toy example in Appendix D. Furthermore, while our analysis is restricted to linear classifiers, linear models can provide a rough proxy for neural network learning dynamics via the neural tangent kernel (NTK) approximation [24]. This strategy has been used in a number of prior works [40, 58, 3] and extending the analysis to the NTK setting may be possible in future work. 6 Conclusion In this paper we propose RCAD, a regularization technique that maximizes a model s predictive entropy on out-of-distribution examples. These samples lie along the directions that adversarially maximize the loss locally around the training points. Our experiments on image classification benchmarks show that RCAD not only improves test accuracy by a significant margin, but that it can also seamlessly compose with prior regularization methods. We find that RCAD is particularly effective when learning from limited data. Finally, we present analyses in a simplified setting to show that RCAD can help the model unlearn noisy features. Some current limitations of our work are that RCAD slows training by 30% and our theoretical analysis is limited to the linear case, which would be interesting directions to address/expand on in future work particularly in light of the significant empirical benefits of RCAD for improving generalization. Acknowledgements. The authors would like to thank Oscar Li, Ziyu Liu, Saurabh Garg at Carnegie Mellon University and members of the RAIL lab at UC Berkeley for helpful feedback and discussion. [1] Arjovsky, M., Bottou, L., Gulrajani, I., and Lopez-Paz, D. (2020). Invariant risk minimization. stat, 1050:27. [2] Besnier, V., Bursuc, A., Picard, D., and Briot, A. (2021). Triggering failures: Out-of-distribution detection by learning from local adversarial attacks in semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 15701 15710. [3] Cao, Y. and Gu, Q. (2019). Generalization bounds of stochastic gradient descent for wide and deep neural networks. Advances in neural information processing systems, 32. [4] Carlini, N., Athalye, A., Papernot, N., Brendel, W., Rauber, J., Tsipras, D., Goodfellow, I., Madry, A., and Kurakin, A. (2019). On evaluating adversarial robustness. ar Xiv preprint ar Xiv:1902.06705. [5] Carlini, N. and Wagner, D. (2017). Towards evaluating the robustness of neural networks. In 2017 ieee symposium on security and privacy (sp), pages 39 57. IEEE. [6] Chen, T., Zhang, Z., Liu, S., Chang, S., and Wang, Z. (2020a). Robust overfitting may be mitigated by properly learned smoothening. In International Conference on Learning Representations. [7] Chen, Y., Wei, C., Kumar, A., and Ma, T. (2020b). Self-training avoids using spurious features under domain shift. Advances in Neural Information Processing Systems, 33:21061 21071. [8] Church, K. W. and Gale, W. A. (1991). A comparison of the enhanced good-turing and deleted estimation methods for estimating probabilities of english bigrams. Computer Speech & Language, 5(1):19 54. [9] Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. (2009). Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248 255. Ieee. [10] De Vries, T. and Taylor, G. W. (2017). Improved regularization of convolutional neural networks with cutout. ar Xiv preprint ar Xiv:1708.04552. [11] Dubey, A., Gupta, O., Raskar, R., and Naik, N. (2018). Maximum-entropy fine grained classification. Advances in neural information processing systems, 31. [12] Gao, Y., Wang, W., Herold, C., Yang, Z., and Ney, H. (2020). Towards a better understanding of label smoothing in neural machine translation. In Proceedings of the 1st Conference of the Asia Pacific Chapter of the Association for Computational Linguistics and the 10th International Joint Conference on Natural Language Processing, pages 212 223. [13] Goodfellow, I., Bengio, Y., and Courville, A. (2016). Deep learning. MIT press. [14] Goodfellow, I. J., Shlens, J., and Szegedy, C. (2014). Explaining and harnessing adversarial examples. ar Xiv preprint ar Xiv:1412.6572. [15] Grandvalet, Y. and Bengio, Y. (2004). Semi-supervised learning by entropy minimization. Advances in neural information processing systems, 17. [16] Grathwohl, W., Wang, K.-C., Jacobsen, J.-H., Duvenaud, D., Norouzi, M., and Swersky, K. (2019). Your classifier is secretly an energy based model and you should treat it like one. ar Xiv preprint ar Xiv:1912.03263. [17] Haarnoja, T., Zhou, A., Abbeel, P., and Levine, S. (2018). Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In International conference on machine learning, pages 1861 1870. PMLR. [18] Hardt, M., Recht, B., and Singer, Y. (2016). Train faster, generalize better: Stability of stochastic gradient descent. In International Conference on Machine Learning, pages 1225 1234. PMLR. [19] He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778. [20] Heinze-Deml, C. and Meinshausen, N. (2017). Conditional variance penalties and domain shift robustness. ar Xiv preprint ar Xiv:1710.11469. [21] Hendrycks, D. and Dietterich, T. (2019). Benchmarking neural network robustness to common corruptions and perturbations. ar Xiv preprint ar Xiv:1903.12261. [22] Hertz, J., Krogh, A., and Palmer, R. G. (2018). Introduction to the theory of neural computation. CRC Press. [23] Ilyas, A., Santurkar, S., Tsipras, D., Engstrom, L., Tran, B., and Madry, A. (2019). Adversarial examples are not bugs, they are features. Advances in neural information processing systems, 32. [24] Jacot, A., Gabriel, F., and Hongler, C. (2018). Neural tangent kernel: Convergence and generalization in neural networks. Advances in neural information processing systems, 31. [25] Jaynes, E. T. (1957). Information theory and statistical mechanics. Physical review, 106(4):620. [26] Johnson, W. E. (1932). Probability: The deductive and inductive problems. Mind, 41(164):409 423. [27] Kakade, S. M., Sridharan, K., and Tewari, A. (2008). On the complexity of linear prediction: Risk bounds, margin bounds, and regularization. Advances in neural information processing systems, 21. [28] Kalimeris, D., Kaplun, G., Nakkiran, P., Edelman, B., Yang, T., Barak, B., and Zhang, H. (2019). Sgd on neural networks learns functions of increasing complexity. Advances in Neural Information Processing Systems, 32:3496 3506. [29] Koltchinskii, V. and Panchenko, D. (2002). Empirical margin distributions and bounding the generalization error of combined classifiers. The Annals of Statistics, 30(1):1 50. [30] Krizhevsky, A. et al. (2009). Learning multiple layers of features from tiny images. Citeseer. [31] Krogh, A. and Hertz, J. A. (1992). A simple weight decay can improve generalization. In Advances in neural information processing systems, pages 950 957. [32] Kschischang, F. R. (2017). The complementary error function. Online, April. [33] Kurakin, A., Goodfellow, I., and Bengio, S. (2016). Adversarial machine learning at scale. ar Xiv preprint ar Xiv:1611.01236. [34] Le, Y. and Yang, X. (2015). Tiny imagenet visual recognition challenge. CS 231N, 7(7):3. [35] Li, D. (1995). Zero duality gap for a class of nonconvex optimization problems. Journal of Optimization Theory and Applications, 85(2):309 324. [36] Madry, A., Makelov, A., Schmidt, L., Tsipras, D., and Vladu, A. (2017). Towards deep learning models resistant to adversarial attacks. ar Xiv preprint ar Xiv:1706.06083. [37] Miyato, T., Maeda, S.-i., Koyama, M., and Ishii, S. (2018). Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE transactions on pattern analysis and machine intelligence, 41(8):1979 1993. [38] Moosavi-Dezfooli, S.-M., Fawzi, A., and Frossard, P. (2016). Deepfool: a simple and accurate method to fool deep neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2574 2582. [39] Müller, R., Kornblith, S., and Hinton, G. (2019). When does label smoothing help? ar Xiv preprint ar Xiv:1906.02629. [40] Nagarajan, V. and Kolter, J. Z. (2019). Uniform convergence may be unable to explain generalization in deep learning. Advances in Neural Information Processing Systems, 32. [41] Netzer, Y., Wang, T., Coates, A., Bissacco, A., Wu, B., and Ng, A. Y. (2011). Reading digits in natural images with unsupervised feature learning. Citeseer. [42] Odena, A. (2016). Semi-supervised learning with generative adversarial networks. ar Xiv preprint ar Xiv:1606.01583. [43] Pereyra, G., Tucker, G., Chorowski, J., Kaiser, Ł., and Hinton, G. (2017). Regularizing neural networks by penalizing confident output distributions. ar Xiv preprint ar Xiv:1701.06548. [44] Peters, J., Bühlmann, P., and Meinshausen, N. (2016). Causal inference by using invariant prediction: identification and confidence intervals. Journal of the Royal Statistical Society. Series B (Statistical Methodology), pages 947 1012. [45] Pinto, F., Yang, H., Lim, S.-N., Torr, P., and Dokania, P. K. (2021). Mix-maxent: Improving accuracy and uncertainty estimates of deterministic neural networks. In Neur IPS 2021 Workshop on Distribution Shifts: Connecting Methods and Applications. [46] Poursaeed, O., Jiang, T., Yang, H., Belongie, S., and Lim, S.-N. (2021). Robustness and generalization via generative adversarial training. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 15711 15720. [47] Raghunathan, A., Xie, S. M., Yang, F., Duchi, J., and Liang, P. (2020). Understanding and mitigating the tradeoff between robustness and accuracy. In International Conference on Machine Learning, pages 7909 7919. PMLR. [48] Raghunathan, A., Xie, S. M., Yang, F., Duchi, J. C., and Liang, P. (2019). Adversarial training can hurt generalization. ar Xiv preprint ar Xiv:1906.06032. [49] Rockafellar, R. T., Uryasev, S., et al. (2000). Optimization of conditional value-at-risk. Journal of risk, 2:21 42. [50] Sabour, S., Cao, Y., Faghri, F., and Fleet, D. J. (2016). Adversarial manipulation of deep representations. In ICLR (Poster). [51] Sagawa, S., Raghunathan, A., Koh, P. W., and Liang, P. (2020). An investigation of why overparameterization exacerbates spurious correlations. In International Conference on Machine Learning, pages 8346 8356. PMLR. [52] Saito, K., Kim, D., Sclaroff, S., Darrell, T., and Saenko, K. (2019). Semi-supervised domain adaptation via minimax entropy. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 8050 8058. [53] Shorten, C. and Khoshgoftaar, T. M. (2019). A survey on image data augmentation for deep learning. Journal of Big Data, 6(1):1 48. [54] Singla, S. and Feizi, S. (2021). Salient imagenet: How to discover spurious features in deep learning? In International Conference on Learning Representations. [55] Soudry, D., Hoffer, E., Nacson, M. S., Gunasekar, S., and Srebro, N. (2018). The implicit bias of gradient descent on separable data. The Journal of Machine Learning Research, 19(1):2822 2878. [56] Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. (2016). Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2818 2826. [57] Tamar, A., Glassner, Y., and Mannor, S. (2015). Optimizing the cvar via sampling. In Twenty Ninth AAAI Conference on Artificial Intelligence. [58] Tancik, M., Srinivasan, P., Mildenhall, B., Fridovich-Keil, S., Raghavan, N., Singhal, U., Ramamoorthi, R., Barron, J., and Ng, R. (2020). Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems, 33:7537 7547. [59] Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society: Series B (Methodological), 58(1):267 288. [60] Tsipras, D., Santurkar, S., Engstrom, L., Turner, A., and Madry, A. (2018). Robustness may be at odds with accuracy. ar Xiv preprint ar Xiv:1805.12152. [61] Tsipras, D., Santurkar, S., Engstrom, L., Turner, A., and Madry, A. (2019). Robustness may be at odds with accuracy. ar Xiv: Machine Learning. [62] Volpi, R., Namkoong, H., Sener, O., Duchi, J., Murino, V., and Savarese, S. (2018). Generalizing to unseen domains via adversarial data augmentation. ar Xiv preprint ar Xiv:1805.12018. [63] Wainwright, M. J. (2019). High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge University Press. [64] Wang, D., Shelhamer, E., Liu, S., Olshausen, B., and Darrell, T. (2020). Tent: Fully test-time adaptation by entropy minimization. ar Xiv preprint ar Xiv:2006.10726. [65] Wang, X., Jiang, Y., Huang, M., and Zhang, H. (2013). Robust variable selection with exponential squared loss. Journal of the American Statistical Association, 108(502):632 643. [66] Yun, S., Han, D., Oh, S. J., Chun, S., Choe, J., and Yoo, Y. (2019). Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6023 6032. [67] Zagoruyko, S. and Komodakis, N. (2016). Wide residual networks. ar Xiv preprint ar Xiv:1605.07146. [68] Zhang, C., Bengio, S., Hardt, M., Recht, B., and Vinyals, O. (2021a). Understanding deep learning (still) requires rethinking generalization. Communications of the ACM, 64(3):107 115. [69] Zhang, H., Cisse, M., Dauphin, Y. N., and Lopez-Paz, D. (2018). mixup: Beyond empirical risk minimization. In International Conference on Learning Representations. [70] Zhang, H., Yu, Y., Jiao, J., Xing, E., El Ghaoui, L., and Jordan, M. (2019a). Theoretically principled trade-off between robustness and accuracy. In International Conference on Machine Learning, pages 7472 7482. PMLR. [71] Zhang, J., He, T., Sra, S., and Jadbabaie, A. (2019b). Why gradient clipping accelerates training: A theoretical justification for adaptivity. ar Xiv preprint ar Xiv:1905.11881. [72] Zhang, Y., Gong, M., Liu, T., Niu, G., Tian, X., Han, B., Schölkopf, B., and Zhang, K. (2021b). Adversarial robustness through the lens of causality. ar Xiv preprint ar Xiv:2106.06196. [73] Zhao, L., Liu, T., Peng, X., and Metaxas, D. (2020). Maximum-entropy adversarial data augmentation for improved generalization and robustness. Advances in Neural Information Processing Systems, 33:14435 14447. [74] Zhou, A. and Levine, S. (2021). Training on test data with bayesian adaptation for covariate shift. ar Xiv preprint ar Xiv:2109.12746. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Sec. 6. (c) Did you discuss any potential negative societal impacts of your work? [No] While supervised learning might be used in applications with both positive and negative societal impacts, our empirical and theoretical contributions do not dictate these potential applications. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] We provide them in the Appendix. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] We document these in the Appendix. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] For our results in Section 4, we provide error bars for the estimated test performance. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] Details in Appendix. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] We use public datasets and cite the creators. (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [No] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]