# do_input_gradients_highlight_discriminative_features__886085d2.pdf Do Input Gradients Highlight Discriminative Features? Harshay Shah Microsoft Research India harshay@google.com Prateek Jain Microsoft Research India prajain@google.com Praneeth Netrapalli Microsoft Research India pnetrapalli@google.com Post-hoc gradient-based interpretability methods [1, 2] that provide instancespecific explanations of model predictions are often based on assumption (A): magnitude of input gradients gradients of logits with respect to input noisily highlight discriminative task-relevant features. In this work, we test the validity of assumption (A) using a three-pronged approach: 1. We develop an evaluation framework, Diff ROAR, to test assumption (A) on four image classification benchmarks. Our results suggest that (i) input gradients of standard models (i.e., trained on original data) may grossly violate (A), whereas (ii) input gradients of adversarially robust models satisfy (A) reasonably well. 2. We then introduce Block MNIST, an MNIST-based semi-real dataset, that by design encodes a priori knowledge of discriminative features. Our analysis on Block MNIST leverages this information to validate as well as characterize differences between input gradient attributions of standard and robust models. 3. Finally, we theoretically prove that our empirical findings hold on a simplified version of the Block MNIST dataset. Specifically, we prove that input gradients of standard one-hidden-layer MLPs trained on this dataset do not highlight instance-specific signal coordinates, thus grossly violating (A). Our findings motivate the need to formalize and test common assumptions in interpretability in a falsifiable manner [3]. We believe that the Diff ROAR framework and Block MNIST datasets serve as sanity checks to audit interpretability methods; code and data available at https://github.com/harshays/inputgradients. 1 Introduction Interpretability methods that provide instance-specific explanations of model predictions are often used to identify biased predictions [4], debug trained models [5], and aid decision-making in highstakes domains such as medical diagnosis [6, 7]. A common approach for providing instance-specific explanations is feature attribution. Feature attribution methods rank or score input coordinates, or features, in the order of their purported importance in model prediction; coordinates achieving the top-most rank or score are considered most important for prediction, whereas those with the bottom-most rank or score are considered least important. Input gradient attributions. Ranking input coordinates based on the magnitude of input gradients is a fundamental feature attribution technique [8, 1] that undergirds well-known methods such as Smooth Grad [2] and Integrated Gradients [9]. Given instance x and a trained model θ with prediction ˆy on x, the input gradient attribution scheme (i) computes the input gradient x Logitθ(x, ˆy) of the logit 2 of the predicted label ˆy and (ii) ranks the input coordinates in decreasing order of their input gradient magnitude. Below we explicitly characterize the underlying intuitive assumption behind input gradient attribution methods: Part of the work completed after joining Google Research India 2In Appendix C, we show that our results also hold for input gradients taken w.r.t. the loss 35th Conference on Neural Information Processing Systems (Neur IPS 2021). Class 0 Class 1 Top block Bottom block (a) Block MNIST images (b) Standard Resnet18 gradients (c) Standard MLP gradients (d) 2 Robust Resnet18 gradients (e) 2 Robust MLP gradients Figure 1: Experiments on Block MNIST dataset. (a) Four representative images from class 0 & class 1 in Block MNIST dataset; every image consists of a signal and null block that are randomly placed as the top or bottom block. The signal block, containing the MNIST digit, determines the image class. The null block, containing the square patch, does not encode any information of the image class. For these four images, subplots (b-e) show the input gradients of standard Resnet18, standard MLP, ℓ2 robust Resnet18 (ϵ=2) and ℓ2 robust MLP (ϵ=4) respectively. The plots clearly show that input gradients of standard Block MNIST models highlight the signal block and the non-discriminative null block, thereby violating (A). In contrast, input gradients of adversarially robust models exclusively highlight the signal block, suppress the null block, and satisfy (A). Please see Section 5 for details. Assumption (A): Coordinates with larger input gradient magnitude are more relevant for model prediction compared to coordinates with smaller input gradient magnitude. Sanity-checking attribution methods. Several attribution methods [10] are based on input gradients and explicitly or implicitly assume an appropriately modified version of (A). For example, Integrated Gradients [9] aggregate input gradients of linearly interpolated points, Smooth Grad [2] averages input gradients of points perturbed using gaussian noise, and Guided Backprop [11] modifies input gradients by zeroing out negative values at every layer during backpropagation. Surprisingly, unlike vanilla input gradients, popular methods that output attributions with better visual quality fail simple sanity checks that are indeed expected out of any valid attribution method [12, 13]. On the other hand, while vanilla input gradients pass simple sanity checks, Hooker et al. [14] suggest that they produce estimates of feature importance that are no better than a random designation of feature importance. Do input gradients satisfy assumption (A)? Since (A) is necessary for input gradients attributions to accurately reflect model behavior, we introduce an evaluation framework, Diff ROAR, to analyze whether input gradient attributions satisfy assumption (A) on real-world datasets. While Diff ROAR adopts the remove-and-retrain (ROAR) methodology [14], Diff ROAR is more appropriate for testing the validity of assumption (A) because it directly compares top-ranked features against bottom-ranked features. We apply Diff ROAR to evaluate input gradient attributions of MLPs & CNNs trained on multiple image classification datasets. Consistent with the message in Hooker et al. [14], our experiments indicate that input gradients of standard models (i.e., trained on original data) can grossly violate (A) (see Section 4). Furthermore, we also observe that unlike standard models, adversarially trained models [15] that are robust to ℓ2 and ℓ perturbations satisfy (A) in a consistent manner. Probing input gradient attributions using Block MNIST. Our empirical findings mentioned above strongly suggest that standard models grossly violate (A). However, without knowledge of groundtruth discriminative features learned by models trained on real data, conclusively testing (A) remains elusive. In fact, this is a key shortcoming of the remove-and-retrain (ROAR) framework. So, to further verify and better understand our empirical findings, we introduce an MNIST-based semi-real dataset, Block MNIST, that by design encodes a priori knowledge of ground-truth discriminative features. Block MNIST is based on the principle that for different inputs, discriminative and non-discriminative features may occur in different parts of the input. For example, in an object classification task, the object of interest can occur in different parts of the image (e.g., top-left, center, bottom-right etc.) for different images. As shown in Figure 1(a), Block MNIST images consist of a signal block and a null block that are randomly placed at the top or bottom. The signal block contains the MNIST digit that determines the class of the image, whereas the null block, contains a square patch with two diagonals that has no information about the label. This a priori knowledge of ground-truth discriminative features in Block MNIST data allows us to (i) validate our empirical findings vis-a-vis input gradients of standard and robust models (see fig. 1) and (ii) identify feature leakage as a reason that potentially explains why input gradients violate (A) in practice. Here, feature leakage refers to the phenomenon wherein given an instance, its input gradients highlight the location of discriminative features in the given instance as well as in other instances that are present in the dataset. For example, consider the first Block MNIST image in fig. 1(a), in which the signal is placed in the bottom block. For this image, as shown in fig. 1(b,c), input gradients of standard models incorrectly highlight the top block because there are other instances in the Block MNIST dataset which have signal in the top block. Rigorously demonstrating feature leakage. In order to concretely verify as well as understand feature leakage more thoroughly, we design a simplified version of Block MNIST that is amenable to theoretical analysis. On this dataset, we first rigorously demonstrate that input gradients of standard one-hidden-layer MLPs exhibit feature leakage in the infinite-width limit and then discuss how feature leakage results in input gradient attributions that clearly violate assumption (A). Paper organization: Section 2 discusses related work and section 3 presents our evaluation framework, Diff ROAR, to test assumption (A). Section 4 employs Diff ROAR to evaluate input gradient attributions on four image classification datasets. Section 5 analyzes Block MNIST data to differentially characterize input gradients of standard and robust models using feature leakage. Section 6 provides theoretical results on a simplified version on Block MNIST that shed light on how feature leakage results in input gradients that violate assumption (A). Our code, along with the proposed datasets, is publicly available at https://github.com/harshays/inputgradients. 2 Related work Due to space constraints, we only discuss directly related work and defer the rest to Appendix A. Sanity checks for explanations. Several explanation methods that provide feature attributions are often primarily evaluated using inherently subjective visual assessments [1, 2]. Unsurprisingly, recent sanity checks show that sole reliance on visual assessment is misleading, as attributions can lack fidelity and inaccurately reflect model behavior. Adebayo et al. [12] and Kindermans et al. [13] show that unlike input gradients [8], other popular methods guided backprop [16], gradient input [17], integrated gradients [9] output explanations which lack fidelity on image data, as they remain invariant to model and label randomization. Similarly, Yang and Kim [18] use custom image datasets to show that several explanation methods are more likely to produce false positive explanations than vanilla input gradients. Moreover, several explanation methods based on modified backpropagation do not pass basic sanity checks [19, 20, 21]. To summarize, well-known gradient-based attribution methods that seek to mitigate gradient saturation [9, 22], discontinuity [23], and visual noise [16] surprisingly fare worse than vanilla input gradients on multiple sanity checks. Evaluating explanation fidelity. The black-box nature of neural networks necessitates frameworks that evaluate the fidelity or correctness of post-hoc explanations without knowledge of ground-truth features learned by trained models. Modification-based evaluation frameworks [24, 25, 26] gauge explanation fidelity by measuring the change in model performance after masking input coordinates that a given explanation method considers most (or least) important. However, due to distribution shifts induced by input modifications, one cannot conclusively attribute changes in model performance to the fidelity of instance-specific explanations [27]. The remove-and-retrain (ROAR) framework [14] accounts for distribution shifts by evaluating the performance of models retrained on train data masked using post-hoc explanations. Surprisingly, contrary to findings obtained via sanity checks, experiments with the ROAR framework show that multiple attribution methods, including vanilla input gradients, are no better than model-independent random attributions that lack explanatory power [14]. Therefore, motivated by the central role of vanilla input gradients in attribution methods, we augment the ROAR framework to understand when and why input gradients violate assumption (A). Effect of adversarial robustness. Adversarial training [15] not only leads to robustness to ℓp adversarial attacks [28], but also leads to perceptually-aligned feature representations [29], and improved visual quality of input gradients [30]. Recent works hypothesize that adversarial training improves the visual quality of input gradients by suppressing irrelevant features [31] and promoting sparsity and stability [32] in explanations. Kim et al. [33] use the ROAR framework to conjecture that adversarial training tilts input gradients to better align with the data manifold. In this work, we use experiments on real-world data and theory on data with features known a priori in order to differentially characterize input gradients of standard and robust models vis-a-vis assumption (A). 3 Diff ROAR evaluation framework In this section, we introduce our evaluation framework, Diff ROAR, to probe the extent to which instance-specific explanations, or feature attributions, highlight discriminative features in practice. Specifically, our framework, Diff ROAR, builds upon the remove-and-retrain (ROAR) methodology [14] to test whether feature attribution methods satisfy assumption (A) on real-world datasets. Setting. We consider the standard classification setting; Each data point (x(i), y(i)), where instance x(i) Rd and label y(i) Y for some label set Y, is drawn independently from a distribution D on Rd Y. Given dataset {(x(i), y(i))} where i [n]:= {1, , n}, x(i) j denotes the jth coordinate of x(i). Note that we also refer to the d coordinates of instance x(i) as features interchangeably. Attribution schemes. A feature attribution scheme A : Rd {σ : σ is a permutation of [d]} maps a d-dimensional instance x to a permutation, or ordering, A(x) : [d] [d] of its coordinates. For example, the input gradient attribution scheme takes as input instance x & predicted label ˆy and outputs an ordering [d] that ranks coordinates in decreasing order of their input gradient magnitude. That is, coordinate j is ranked ahead of coordinate k if the magnitude of the jth coordinate of x Logitθ(x, ˆy) is larger than that of the kth coordinate. Unmasking schemes. Given instance x and a subset S [d] of coordinates, the unmasked instance x S zeroes out all coordinates that are not in subset S: x S j = xj if j S and 0 if j / S. An unmasking scheme A : Rd {S : S [d]} simply maps instance x to a subset A(x) [d] of coordinates that can be used to obtain unmasked instance x A(x). Any attribution scheme A naturally induces top-k and bottom-k unmasking schemes, Atop k and Abot k , which output k coordinates with the top-most and bottom-most attributions in A(x) respectively. In other words, given attribution scheme A and level k, the top-k and bottom-k unmasking schemes, Atop k and Abot k , can be defined as follows: Top-k unmasked image Original image Figure 2: Pictorial example of a top25% unmasked image. Atop k (x):= {A(x)j : j k} , Abot k (x):= {A(x)j : d k < j d} . For example, Figure 2 depicts an image x and its top-k unmasked variant x Atop k (x). In this case, the attribution scheme A assigns higher rank to pixels in the foreground. So, the top25% unmasking operation, x Atop 25%(x), highlights the monkey by retaining pixels with top-25% attribution ranks and zeroing out the remaining pixels that correspond to the green background. Predictive power of unmasking schemes. The predictive power of an unmasking scheme A with respect to model architecture M (e.g., resnet18) can be defined as the best classification accuracy that can be attained by training a model with architecture M on unmasked instances that are obtained via unmasking scheme A. More formally, it can defined as follows: Pred Power M(A) := sup f M,f:Rd Y ED h 1 h f(x A(x)) = y ii . Due to masking-induced distribution shifts, models with architecture M that are trained using original data cannot be plugged in to estimate Pred Power M(A). The ROAR framework [14] sidesteps this issue by retraining models on unmasked data, as similar model architectures tend to learn similar classifiers [34, 35, 36, 37]. Therefore, we employ the ROAR framework to estimate Pred Power M(A) in two steps. First, we use unmasking scheme A to obtain unmasked train and test datasets that comprise data points of the form (x A(x), y). Then, we retrain a new model with the same architecture M on unmasked train data and evaluate its accuracy on unmasked test data. Diff ROAR evaluation metric to test assumption (A). Recall that an attribution scheme A maps an instance x to a permutation of its coordinates that reflects the order of estimated importance in model prediction. An attribution scheme that satisfies assumption (A) must place coordinates that are more important for model prediction higher up in the the attribution order. More formally, given attribution scheme A, architecture M and level k, we define Diff ROAR as the difference between the predictive power of top-k and bottom-k unmasking schemes, Atop k and Abot k : Diff ROARM(A, k) = Pred Power M(Atop k ) Pred Power M(Abot k ) (1) Interpreting the Diff ROAR metric. The sign of the Diff ROAR metric indicates whether the given attribution scheme satisfies or violates assumption (A). For example, Diff ROARM(A, ) < 0 implies that A violates assumption (A) , as coordinates with higher attribution ranks have worse predictive power with respect to architecture M. Similarly, the magnitude of the Diff ROAR metric quantifies the extent to which the ordering in attribution scheme A separates the most and least discriminative coordinates into two disjoint subsets. For example, a random attribution scheme Ar, which outputs attributions Ar(x) chosen uniformly at random from all permutations of [d], neither highlights nor suppresses discriminative features; E[Diff ROARM(Ar, k)] = 0 for any architecture M. On testing assumption (A). To verify (A) for a given attribution scheme A, it is necessary to evaluate whether input coordinates with higher attribution rank are more important for model prediction than coordinates with lower rank. Consequently, the ROAR-based metric in Hooker et al. [14], which essentially computes the top-k predictive power, is not sufficient to test whether attribution methods satisfy assumption (A). Therefore, as discussed above, Diff ROAR tests (A) by comparing the top-k predictive power, Pred Power M(Atop k ), to the bottom-k predictive power, Pred Power M(Abot k ), using multiple values of k. 4 Testing assumption (A) on image classification benchmarks In this section, we use Diff ROAR to evaluate whether input gradient attributions of standard and adversarially robust MLPs and CNNs trained on four image classification benchmarks satisfy assumption (A). We first summarize the experiment setup and then describe key empirical findings. Datasets and models. We consider four benchmark image classification datasets: SVHN [38], Fashion MNIST [39], CIFAR-10 [40] and Image Net-10 [41]. Image Net-10 is an open-sourced variant (https://github.com/Madry Lab/robustness/) of Imagenet [41], with 80, 000 images grouped into 10 super-classes. Image Net-10 enables us to test assumption (A) on Imagenet without the computational overload of training models on the 1000-way ILSVRC classification task [42]. We evaluate input gradient attributions of standard and adversarially trained two-hidden-layer MLPs and Resnets [43]. We obtain ℓ2 and ℓ ϵ-robust models with perturbation budget ϵ using PGD adversarial training [15]. Unless mentioned otherwise, we train models using stochastic gradient descent (SGD), with momentum 0.9, batch size 256, ℓ2 regularization 0.0005 and initial learning rate 0.1 that decays by a factor of 0.75 every 20 epochs. Additionally, we use standard data augmentation and train models for at most 500 epochs, stopping early if cross-entropy loss on training data goes below 0.001. Appendix C.1 provides additional details about the datasets and trained models.3 Estimating Diff ROAR on real data. We compute the evaluation metric, Diff ROARM(A, k), on real datasets in four steps, as follows. First, we train a standard or robust model with architecture M on the original dataset and obtain its input gradient attribution scheme A. Second, as outlined in Section 3, we use attribution scheme A and level k (i.e., fraction of pixels to be unmasked) to extract the top-k and bottom-k unmasking schemes: Atop k and Abot k . Third, we apply Atop k and Abot k on the original train & test datasets to obtain top-k and bottom-k unmasked datasets respectively. Finally, to compute Diff ROARM(A, k) via eq. (1), we estimate top-k and bottom-k predictive power, Pred Power M(Atop k ) and Pred Power M(Abot k ), by retraining new models with architecture M on topk and bottom-k unmasked datasets respectively. Also, note that we (a) average the Diff ROAR metric over five runs for each model and unmasking fraction or level k and (b) unmask individual image pixels without grouping them channel-wise. Experiment setup. Now, we analyze the Diff ROAR metric as a function of the unmasking fraction k {5, 10, 20, . . . , 100}% in order to evaluate whether input gradient attributions of models trained on four image classification benchmarks satisfy assumption (A). In particular, as shown in Figure 3, we use Diff ROAR to analyze input gradients of standard and adversarially robust two-hidden-layer MLPs on SVHN & Fashion MNIST, Resnet18 on Image Net-10, and Resnet50 on CIFAR-10. In order to calibrate our findings, we compare input gradient attributions of these models to two natural baselines: model-agnostic random attributions and input-agnostic attributions of linear models. Input gradients of standard models. Input gradient attributions of standard MLPs trained on SVHN satisfy assumption (A), as the Diff ROAR metric in Figure 3(a) is positive for all values of level k < 100%. However, in Figure 3(b), the Diff ROAR curves of standard MLPs trained on 3Code publicly available at https://github.com/harshays/inputgradients 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Level k: fraction of unmasked pixels Diff ROAR metric at level k (a) Input gradients of MLPs on SVHN standard, ϵ = 0 ℓ2 robust, ϵ = 0.10 ℓ2 robust, ϵ = 0.25 ℓ robust, ϵ = 1 255 ℓ robust, ϵ = 2 255 linear model random attribution 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Level k: fraction of unmasked pixels Diff ROAR metric at level k (c) Input gradients of Resnet18 on Image Net-10 standard, ϵ = 0 ℓ robust, ϵ = 2 255 ℓ2 robust, ϵ = 1.0 ℓ robust, ϵ = 4 255 linear model random attribution 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Level k: fraction of unmasked pixels Diff ROAR metric at level k (b) Input gradients of MLPs on Fashion MNIST standard, ϵ = 0.00 ℓ robust, ϵ = 4 255 ℓ robust, ϵ = 16 255 ℓ2 robust, ϵ = 0.50 ℓ2 robust, ϵ = 1 linear model random attribution 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Level k: fraction of unmasked pixels Diff ROAR metric at level k (d) Input gradients of Resnet50 on CIFAR-10 standard, ϵ = 0 ℓ2 robust, ϵ = 0.25 ℓ2 robust, ϵ = 0.50 ℓ robust, ϵ = 8 255 linear model random attribution Figure 3: Diff ROAR plots for input gradient attributions of standard and adversarially robust twohidden-layer MLPs on (a) SVHN & (b) Fashion MNIST, (c) Resnet18 on Image Net-10 and (d) Resnet50 on CIFAR-10. Subplot (a) indicates that adversarially robust MLPs consistently and considerably outperform standard MLPs on the Diff ROAR metric for all k < 100%. Subplot (b) shows that for most unmasking fractions k, standard MLPs trained on Fashion MNIST, unlike robust MLPs, fare no better than model-agnostic random attributions and input-agnostic attributions of linear models. Subplots (c) and (d) show that when k < 40%, standard Resnet models trained on CIFAR10 and Image Net-10 grossly violate (A), thereby implying that coordinates with top-most gradient attribution rank have worse predictive power than coordinates with bottom-most rank. In stark contrast, input gradients of Resnets that are robust to ℓ2 and ℓ adversarial perturbations satisfy assumption (A) reasonably well. We observe that increasing the perturbation budget ϵ during adversarial training amplifies the magnitude of Diff ROAR for every k across all four image classification benchmarks. Fashion MNIST indicate that input gradient attributions, consistent with findings in Hooker et al. [14], can fare no better than model-agnostic random attributions and input-agnostic attributions of linear models vis-a-vis assumption (A). Furthermore, and rather surprisingly, the shaded area in Figure 3(c) and Figure 3(d) shows that when level k < 40%, Diff ROAR curves of standard Resnets trained on CIFAR-10 and Imagenet-10 are consistently negative and perform considerably worse than model-agnostic and input-agnostic baseline attributions. These results strongly suggest that on CIFAR-10 and Imagenet-10, input gradients of standard Resnets grossly violate assumption (A) and suppress discriminative features. In other words, coordinates with larger gradient magnitude have worse predictive power than coordinates with smaller gradient magnitude. Input gradients of robust models. Models that are ϵ-robust to ℓ2 and ℓ adversarial perturbations fare considerably better than standard models on the Diff ROAR metric. For example, in Figure 3(a), when level k equals 10%, robust MLPs trained on SVHN outperform standard MLPs on the Diff ROAR metric by roughly 10-30%. The Diff ROAR curves of adversarially robust MLPs in Figure 3 are positive at every level k < 100%, which strongly suggests that input gradient attributions of robust MLPs satisfy assumption (A). Similarly, robust resnet50 models trained on CIFAR-10 and Image Net-10 satisfy assumption (A) reasonably well and, unlike standard resnet50 models, starkly highlight discriminative features. Furthermore, we observe that increasing the perturbation budget ϵ in ℓ2 or ℓ PGD adversarial training [15] amplifies the magnitude of Diff ROAR across k and for all four datasets. That is, the adversarial perturbation budget ϵ determines the extent to which input gradients differentiates the most and least discriminative coordinates into two disjoint subsets. Additional results. In Appendix C, we show that our Diff ROAR results are robust to choice of model architecture & SGD hyperparameters during retraining and also hold for input gradients taken with respect to cross-entropy loss. Additionally, while Diff ROAR without retraining gives qualitatively similar results, they are not as consistent across architectures as with retraining, particularly for small unmasking fraction k that induce non-trivial distribution shifts. 5 Analyzing input gradient attributions using Block MNIST data To verify whether input gradients satisfy assumption (A) more thoroughly, we introduce and perform experiments on Block MNIST, an MNIST-based dataset that by design encodes a priori knowledge of ground-truth discriminative features. Block MNIST dataset design: The design of the Block MNIST dataset is based on two intuitive properties of real-world object classification tasks: (i) for different images, the object of interest may appear in different parts of the image (e.g., top-left, bottom-right); (ii) the object of interest and the rest of the image often share low-level patterns such as edges that are not informative of the label on their own. We replicate these aspects in Block MNIST instances, which are vertical concatenations of two 28 28 signal and null image blocks that are randomly placed at the top or bottom with equal probability. The signal block is an MNIST image of digit 0 or digit 1, corresponding to class 0 or 1 of the Block MNIST image respectively. On the other hand, the null block in every Block MNIST image, independent of its class, contains a square patch made of two horizontal, vertical, and slanted lines, as shown in Figure 1(a). It is important to note that unlike the MNIST signal block that is fully predictive of the class, the non-discriminative null block contains no information about the class. Standard as well as adversarially robust models trained on Block MNIST data attain 99.99% test accuracy, thereby implying that model predictions are indeed based solely on the signal block for any given instance. We further verify this by noting that the predictions of trained model remain unchanged on almost every instance even when all pixels in the null block are set to zero. Do standard and robust models satisfy (A)? As discussed above, unlike the null block that has no task-relevant information, the MNIST digit in the signal block entirely determines the class of any given Block MNIST image. Therefore, in this setting, we can restate assumption (A) as follows: Do input gradient attributions highlight the signal block over the null block? Surprisingly, as shown in Figure 1(b,c), input gradient attributions of standard MLP and Resnet18 models highlight the signal block as well as the non-discriminative null block. In stark contrast, subplots (d) and (e) in Figure 1 show that input gradient attributions of ℓ2 robust MLP and Resnet18 models exclusively highlight MNIST digits in the signal block and clearly suppress the square patch in the null block. These results validate our findings on real-world datasets by showing that unlike standard models, adversarially robust models satisfy (A) on Block MNIST data. Feature leakage hypothesis: Recall that the discriminative signal block in Block MNIST images is randomly placed at the top or bottom with equal probability. Given our results in Figure 1, we hypothesize that when discriminative features vary across instances (e.g., signal block at top vs. bottom), input gradients of standard models not only highlight instance-specific features but also leak discriminative features from other instances. We term this hypothesis feature leakage. Class 0 Class 1 (a) Block MNIST-Top Class 0 Class 1 (b) Standard Resnet18 Class 0 Class 1 (c) Standard MLP Figure 4: (a) In Block MNIST-Top images, the signal & null blocks are fixed at the top & bottom respectively. In contrast to results on Block MNIST in fig. 1, input gradients of standard (b) Resnet18 and (c) MLP trained on Block MNIST-Top highlight discriminative features in the signal block, suppress the null block, and satisfy (A). To test our hypothesis, we leverage the modular structure in Block MNIST to construct a slightly modified version, Block MNIST-Top, wherein the location of the MNIST signal block is fixed at the top for all instances (see fig. 4). In this setting, in contrast to results on Block MNIST, input gradients of standard Resnet18 and MLP models trained on Block MNIST-Top satisfy assumption (A). Specifically, when the signal block is fixed at the top, input gradient attributions in Figure 4(b, c) clearly highlight the signal block and suppress the null block, thereby supporting our feature leakage hypothesis. Based on our Block MNIST experiments, we believe that understanding how adversarial robustness mitigates feature leakage is an interesting direction for future work. Additional results. In Appendix D.1, we (i) visualize input gradients of several Block MNIST and Block MNIST-Top images, (ii) introduce a quantitative proxy metric to compare feature leakage between standard and robust models, (iii) show that our findings are fairly robust to the choice and number of classes in Block MNIST data, and (iv) evaluate feature leakage in five feature attribution methods. We also provide experiments that falsify hypotheses vis-a-vis input gradients and assumption (A) that we considered in addition to feature leakage. 6 Feature leakage in input gradient attributions To understand the extent of feature leakage more thoroughly, we introduce a simplified version of the Block MNIST dataset that is amenable to theoretical analysis. We rigorously show that input gradients of standard one-hidden-layer MLPs do not differentiate instance-specific features from other task-relevant features that are not pertinent to the given instance. Dataset: Given dimension of each block ed, feature vector u R e d with u = 1, number of blocks d and noise parameter η, we will construct input instances of dimension ed d. More concretely, a sample (x, y) R e d d { 1} from the distribution D is generated as follows: y = 1 with probability 0.5 and x = [ηg1, ηg2, . . . , yu + ηgj, . . . , ηgd] with j chosen at random from [d/2] (2) where each gi R e d is drawn uniformly at random from the unit ball. For simplicity, we take d to be even so that d/2 is an integer. We can think of each x as a concatenation of d ed-dimensional blocks {x1, . . . , xd}. The first d/2 blocks, {1, . . . , d/2}, are task-relevant, as every example (x, y) contains an instance-specific signal block xi = yu + ηgi for some i [d/2] that is informative of its label y. Given instance x, we use j (x) to denote the unique instance-specific signal block such that xj (x) = yu + ηgj (x). On the other hand, noise blocks {d/2 + 1, . . . , d} do not contain task-relevant signal for any instance x. At a high level, the instance-specific signal block j (x) and noise blocks {d/2 + 1, . . . , d} in instance x correspond to the discriminative MNIST digit and the null square patch in Block MNIST images respectively. For example, each row in Figure 5(a) illustrates an instance x where d = 10, ed = 1, η = 0 and u = 1. Model: We consider one-hidden layer MLPs with Re LU nonlinearity in the infinite-width limit. More concretely, for a given width m, the network is parameterized by R Rm e d d, b Rm and w Rm. Given an input instance (x, y) R e dd { 1}, the output score (or logit) f and cross-entropy (CE) loss L are given by: f((w, R, b), x):= w, φ (Rx + b) , L((w, R, b), (x, y)):= log (1 + exp ( y f((w, R, b), x))) . where φ(t):= max(0, t) denotes the Re LU function. A remarkable set of recent results [44, 45, 46, 47] show that as m , the training procedure is equivalent to gradient descent (GD) on an infinite dimensional Wasserstein space. In the Wasserstein space, the network can be interpreted as a probability distribution ν over R R e d d R with output score f and cross entropy loss L defined as: f(ν, x):=E(w,r,b) ν [wφ ( r, x + b)] , L(ν, (x, y)):= log (1 + exp ( y f(ν, x))) . (3) Theoretical analysis: Our approach leverages the recent result in Chizat and Bach [48], which shows that if GD in the Wasserstein space W2 R R e dd R on ED [L(ν, (x, y))] converges, it does so to a max-margin classifier given by: ν := arg max ν P(Sd e d+1) min (x,y) D y f(ν, x), (4) where Sd e d+1 denotes the surface of the Euclidean unit ball in R e dd+2, and P Sd e d+1 denotes the space of probability distributions over Sd e d+1. Intuitively, our main result shows that on any data point (x, y) D, the input gradient magnitude of the max-margin classifier ν is equal over all task-relevant blocks {1, . . . , d/2} and zero on the remaining noise blocks {d/2 + 1, . . . , d}. Theorem 1. Consider distribution D (2) with η < 1 10d. There exists a max-margin classifier ν for D in Wasserstein space (i.e., training both layers of FCN with m ) given by (4), such that for all (x, y) D: (i) ( x L(ν , (x, y)))j = c > 0 for every j [d/2] and (ii) ( x L(ν , (x, y)))j = 0 for every j {d/2 + 1, , d}, where ( x L(ν , (x, y)))j denotes the jth block of the input gradient x L(ν , (x, y)). Theorem 1 guarantees the existence of a max-margin classifier such that the input gradient magnitude for any given instance is (i) a non-zero constant on each of the first d/2 task-relevant blocks, and 1 2 3 4 5 6 7 8 9 10 Input coordinates x1 x2 x3 x4 x5 x6 x7 x8 x9 x10 Input data points Task-relevant Noise (a) Input data 1 2 3 4 5 6 7 8 9 10 Input coordinates Input gradient attributions Task-relevant Noise (b) Standard linear model 1 2 3 4 5 6 7 8 9 10 Input coordinates Input gradient attributions Task-relevant Noise (c) Standard MLP 1 2 3 4 5 6 7 8 9 10 Input coordinates Input gradient attributions Task-relevant Noise (d) 1 Robust MLP Figure 5: Input gradients of linear models and standard & robust MLPs trained on data from eq. (2) with d = 10, ed = 1, η = 0 and u = 1. (a) Each row in corresponds to an instance x, and the highlighted coordinate denotes the signal block j (x) & label y. (b) Linear models suppress noise coordinates but lack the expressive power to highlight instance-specific signal j (x), as their input gradients in subplot (b) are identical across all examples. (c) Despite the expressive power to highlight instance-specific signal coordinate j (x), input gradients of standard MLPs exhibit feature leakage (see Theorem 1) and violate (A) as well. (d) In stark contrast, input gradients of adversarially trained MLPs suppress feature leakage and starkly highlight instance-specific signal coordinates j (x). (ii) equal to zero on the remaining d/2 noise blocks that do not contain any information about the label. However, input gradients fail at highlighting the unique instance-specific signal block over the remaining task-relevant blocks. This clearly demonstrates feature leakage, as input gradients for any given instance also highlight task-relevant features that are, in fact, not specific to the given instance. Therefore, input gradients of standard one-hidden-layer MLPs do not highlight instance-specific discriminative features and grossly violate assumption (A). In Appendix F, we present additional results that demonstrate that adversarially trained one-hidden-layer MLPs can suppress feature leakage and satisfy assumption (A). Empirical results: Now, we supplement our theoretical results by evaluating input gradients of linear models as well as standard & robust one-hidden-layer Re LU MLPs with width 10000 on the dataset shown in Figure 5. Note that all models obtain 100% test accuracy on this linearly separable dataset, a simplified version of Block MNIST that is obtained via eq. 2 with d = 10, ed = 1, η = 0 and u = 1. Due to insufficient expressive power, linear models have input-agnostic gradients that suppress all five noise coordinates, but do not differentiate the instance-specific signal coordinate from the remaining task-relevant coordinates. Consistent with Theorem 1, even standard MLPs, which are expressive enough to have input gradients that correctly highlight instance-specific coordinates, apply equal weight on all five task-relevant coordinates and violate (A) due to feature leakage. On the other hand, Figure 5(c) shows that the same MLP architecture, if robust to ℓ adversarial perturbations with norm 0.35, satisfies (A) by clearly highlighting the instance-specific signal coordinate over all other noise and task-relevant coordinates 7 Discussion and conclusion In this work, we took a three-pronged approach to investigate the validity of a key assumption made in several popular post-hoc attribution methods: (A) coordinates with larger input gradient magnitude are more relevant for model prediction compared to coordinates with smaller input gradient magnitude. Through (i) evaluation on real-world data using our Diff ROAR framework, (ii) empirical analysis on Block MNIST data that encodes information of ground-truth discriminative features, and (iii) a rigorous theoretical study, we present strong evidence to suggest that standard models do not satisfy assumption (A). In contrast, adversarially robust models satisfy (A) in a consistent manner. Furthermore, our analysis in Section 5 and Section 6 indicates that feature leakage sheds light on why input gradients of standard models tend to violate (A). We provide additional discussion in Appendix B. This work exclusively focused on vanilla" input gradients due to their fundamental significance in feature attribution. A similarly thorough investigation that analyzes other commonly-used attribution methods is an interesting avenue for future work. Another interesting avenue for further analyses is to understand how adversarial training mitigates feature leakage in input gradient attributions. [1] Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. ar Xiv preprint ar Xiv:1312.6034, 2013. [2] Daniel Smilkov, Nikhil Thorat, Been Kim, Fernanda Viégas, and Martin Wattenberg. Smoothgrad: removing noise by adding noise. ar Xiv preprint ar Xiv:1706.03825, 2017. [3] Matthew L. Leavitt and Ari S. Morcos. Towards falsifiable interpretability research. Ar Xiv, abs/2010.12016, 2020. [4] Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. "why should I trust you?": Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, August 13-17, 2016, pages 1135 1144, 2016. [5] Julius Adebayo, Michael Muelly, Ilaria Liccardi, and Been Kim. Debugging tests for model explanations. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 700 712. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper/2020/file/ 075b051ec3d22dac7b33f788da631fd4-Paper.pdf. [6] John R Zech, Marcus A Badgeley, Manway Liu, Anthony B Costa, Joseph J Titano, and Eric Karl Oermann. Variable generalization performance of a deep learning model to detect pneumonia in chest radiographs: a cross-sectional study. PLo S medicine, 15(11):e1002683, 2018. [7] Gregor Stiglic, Primoz Kocbek, Nino Fijacko, Marinka Zitnik, Katrien Verbert, and Leona Cilar. Interpretability of machine learning-based prediction models in healthcare. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 10(5):e1379, 2020. [8] David Baehrens, Timon Schroeter, Stefan Harmeling, Motoaki Kawanabe, Katja Hansen, and Klaus-Robert Müller. How to explain individual classification decisions. The Journal of Machine Learning Research, 11:1803 1831, 2010. [9] Mukund Sundararajan, Ankur Taly, and Qiqi Yan. Axiomatic attribution for deep networks. In International Conference on Machine Learning, pages 3319 3328. PMLR, 2017. [10] W James Murdoch, Chandan Singh, Karl Kumbier, Reza Abbasi-Asl, and Bin Yu. Definitions, methods, and applications in interpretable machine learning. Proceedings of the National Academy of Sciences, 116(44):22071 22080, 2019. [11] J Springenberg, Alexey Dosovitskiy, Thomas Brox, and M Riedmiller. Striving for simplicity: The all convolutional net. In ICLR (workshop track), 2015. [12] Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, and Been Kim. Sanity checks for saliency maps. Advances in Neural Information Processing Systems, 31: 9505 9515, 2018. [13] Pieter-Jan Kindermans, Sara Hooker, Julius Adebayo, Maximilian Alber, Kristof T Schütt, Sven Dähne, Dumitru Erhan, and Been Kim. The (un) reliability of saliency methods. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning, pages 267 280. Springer, 2019. [14] Sara Hooker, D. Erhan, P. Kindermans, and Been Kim. A benchmark for interpretability methods in deep neural networks. In Neur IPS, 2019. [15] Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=r Jz IBf ZAb. [16] Jost Tobias Springenberg, Alexey Dosovitskiy, Thomas Brox, and Martin Riedmiller. Striving for simplicity: The all convolutional net. ar Xiv preprint ar Xiv:1412.6806, 2014. [17] Avanti Shrikumar, Peyton Greenside, Anna Shcherbina, and Anshul Kundaje. Not just a black box: Learning important features through propagating activation differences. ar Xiv preprint ar Xiv:1605.01713, 2016. [18] Mengjiao Yang and Been Kim. Benchmarking attribution methods with relative feature importance. ar Xiv preprint ar Xiv:1907.09701, 2019. [19] Leon Sixt, Maximilian Granz, and Tim Landgraf. When explanations lie: Why many modified BP attributions fail. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 9046 9057. PMLR, 13 18 Jul 2020. URL http: //proceedings.mlr.press/v119/sixt20a.html. [20] Chih-Kuan Yeh, Cheng-Yu Hsieh, Arun Suggala, David I Inouye, and Pradeep K Ravikumar. On the (in)fidelity and sensitivity of explanations. In Advances in Neural Information Processing Systems, volume 32, pages 10967 10978, 2019. [21] Marco Ancona, Enea Ceolini, Cengiz Öztireli, and Markus Gross. Towards better understanding of gradient-based attribution methods for deep neural networks. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=Sy21R9JAW. [22] Mukund Sundararajan, Ankur Taly, and Qiqi Yan. Gradients of counterfactuals. ar Xiv preprint ar Xiv:1611.02639, 2016. [23] Avanti Shrikumar, Peyton Greenside, and Anshul Kundaje. Learning important features through propagating activation differences. In International Conference on Machine Learning, pages 3145 3153. PMLR, 2017. [24] Sebastian Bach, Alexander Binder, Grégoire Montavon, Frederick Klauschen, Klaus-Robert Müller, and Wojciech Samek. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLOS ONE, 10(7):1 46, 07 2015. doi: 10.1371/journal.pone. 0130140. URL https://doi.org/10.1371/journal.pone.0130140. [25] W. Samek, A. Binder, G. Montavon, S. Lapuschkin, and K. Müller. Evaluating the visualization of what a deep neural network has learned. IEEE Transactions on Neural Networks and Learning Systems, 28(11):2660 2673, 2017. doi: 10.1109/TNNLS.2016.2599820. [26] Leila Arras, Ahmed Osman, Klaus-Robert Müller, and Wojciech Samek. Evaluating recurrent neural network explanations. In Proceedings of the 2019 ACL Workshop Blackbox NLP: Analyzing and Interpreting Neural Networks for NLP, pages 113 126, Florence, Italy, August 2019. Association for Computational Linguistics. doi: 10.18653/v1/W19-4813. URL https://www.aclweb.org/anthology/W19-4813. [27] Richard Tomsett, D. Harborne, S. Chakraborty, Prudhvi Gurram, and A. Preece. Sanity checks for saliency metrics. Ar Xiv, abs/1912.01451, 2020. [28] Florian Tramer, Nicholas Carlini, Wieland Brendel, and Aleksander Madry. On adaptive attacks to adversarial example defenses. ar Xiv preprint ar Xiv:2002.08347, 2020. [29] Shibani Santurkar, Dimitris Tsipras, Brandon Tran, Andrew Ilyas, Logan Engstrom, and Aleksander Madry. Image synthesis with a single (robust) classifier. Advances in Neural Information Processing Systems, 32, 2019. [30] A. Ross and Finale Doshi-Velez. Improving the adversarial robustness and interpretability of deep neural networks by regularizing their input gradients. In AAAI, 2018. [31] Beomsu Kim, Junghoon Seo, Seunghyun Jeon, Jamyoung Koo, J. Choe, and Taegyun Jeon. Why are saliency maps noisy? cause of and solution to noisy saliency maps. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), pages 4149 4157, 2019. [32] P. Chalasani, J. Chen, A. Chowdhury, S. Jha, and X. Wu. Concise explanations of neural networks using adversarial training. In ICML, 2020. [33] Beomsu Kim, Junghoon Seo, and Taegyun Jeon. Bridging adversarial robustness and gradient interpretability. Ar Xiv, abs/1903.11626, 2019. [34] Guy Hacohen, Leshem Choshen, and Daphna Weinshall. Let s agree to agree: Neural networks share classification order on real datasets. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 3950 3960. PMLR, 13 18 Jul 2020. URL http://proceedings.mlr.press/v119/hacohen20a.html. [35] Yixuan Li, Jason Yosinski, Jeff Clune, Hod Lipson, and John Hopcroft. Convergent learning: Do different neural networks learn the same representations? In Dmitry Storcheus, Afshin Rostamizadeh, and Sanjiv Kumar, editors, Proceedings of the 1st International Workshop on Feature Extraction: Modern Questions and Challenges at NIPS 2015, volume 44 of Proceedings of Machine Learning Research, pages 196 212, Montreal, Canada, 11 Dec 2015. PMLR. URL http://proceedings.mlr.press/v44/li15convergent.html. [36] Dimitris Kalimeris, Gal Kaplun, Preetum Nakkiran, Benjamin Edelman, Tristan Yang, Boaz Barak, and Haofeng Zhang. Sgd on neural networks learns functions of increasing complexity. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips.cc/paper/2019/file/ b432f34c5a997c8e7c806a895ecc5e25-Paper.pdf. [37] Harshay Shah, Kaustav Tamuly, Aditi Raghunathan, Prateek Jain, and Praneeth Netrapalli. The pitfalls of simplicity bias in neural networks. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 9573 9585. Curran Associates, Inc., 2020. URL https://proceedings.neurips. cc/paper/2020/file/6cfe0e6127fa25df2a0ef2ae1067d915-Paper.pdf. [38] Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y Ng. Reading digits in natural images with unsupervised feature learning. 2011. [39] Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. ar Xiv preprint ar Xiv:1708.07747, 2017. [40] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009. [41] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Image Net: A Large-Scale Hierarchical Image Database. In CVPR09, 2009. [42] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. Image Net Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 115(3):211 252, 2015. doi: 10.1007/s11263-015-0816-y. [43] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. ar Xiv preprint ar Xiv:1512.03385, 2015. [44] Song Mei, Andrea Montanari, and Phan-Minh Nguyen. A mean field view of the landscape of two-layer neural networks. Proceedings of the National Academy of Sciences, 115(33): E7665 E7671, 2018. [45] Lénaïc Chizat and Francis Bach. On the global convergence of gradient descent for overparameterized models using optimal transport. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, pages 3040 3050, 2018. [46] Grant M Rotskoff and Eric Vanden-Eijnden. Trainability and accuracy of neural networks: An interacting particle system approach. ar Xiv preprint ar Xiv:1805.00915, 2018. [47] Justin Sirignano and Konstantinos Spiliopoulos. Mean field analysis of deep neural networks. Mathematics of Operations Research, 2021. [48] Lénaïc Chizat and Francis Bach. Implicit bias of gradient descent for wide two-layer neural networks trained with the logistic loss. In Conference on Learning Theory, pages 1305 1338. PMLR, 2020. [49] A. Ghorbani, Abubakar Abid, and James Y. Zou. Interpretation of neural networks is fragile. In AAAI, 2019. [50] Juyeon Heo, Sunghwan Joo, and T. Moon. Fooling neural network interpretations via adversarial model manipulation. In Neur IPS, 2019. [51] Ann-Kathrin Dombrowski, Maximilian Alber, Christopher J. Anders, Marcel Ackermann, Klaus-Robert Müller, and Pan Kessel. Explanations can be manipulated and geometry is to blame. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d Alché Buc, Emily B. Fox, and Roman Garnett, editors, Neur IPS, pages 13567 13578, 2019. URL http://dblp.uni-trier.de/db/conf/nips/nips2019.html#Dombrowski AAAMK19. [52] Naman Bansal, Chirag Agarwal, and Anh M Nguyen. Sam: The sensitivity of attribution methods to hyperparameters. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 11 21, 2020. [53] Mayank Singh, Nupur Kumari, P. Mangla, Abhishek Sinha, V. Balasubramanian, and Balaji Krishnamurthy. Attributional robustness training using input-gradient spatial alignment. In ECCV, 2020. [54] H. Lakkaraju, Nino Arsov, and Osbert Bastani. Robust and stable black box explanations. In ICML, 2020. [55] E. Chu, Deb Roy, and Jacob Andreas. Are visual explanations useful? a case study in model-inthe-loop prediction. Ar Xiv, abs/2007.12248, 2020. [56] Forough Poursabzi-Sangdeh, D. Goldstein, J. Hofman, Jennifer Wortman Vaughan, and H. Wallach. Manipulating and measuring model interpretability. Ar Xiv, abs/1802.07810, 2018. [57] Danish Pruthi, Bhuwan Dhingra, Livio Baldini Soares, M. Collins, Zachary C. Lipton, Graham Neubig, and William W. Cohen. Evaluating explanations: How much do explanations from the teacher aid students? Ar Xiv, abs/2012.00893, 2020. [58] Peijie Chen, Chirag Agarwal, and Anh Nguyen. The shape and simplicity biases of adversarially robust imagenet-trained cnns. ar Xiv preprint ar Xiv:2006.09373, 2020. [59] Osman Semih Kayhan and Jan C van Gemert. On translation invariance in cnns: Convolutional layers can exploit absolute spatial location. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14274 14285, 2020. [60] Thomas Tanay and Lewis Griffin. A boundary tilting persepective on the phenomenon of adversarial examples. ar Xiv preprint ar Xiv:1608.07690, 2016. [61] Andrew Ilyas, Shibani Santurkar, Dimitris Tsipras, Logan Engstrom, Brandon Tran, and Aleksander Madry. Adversarial examples are not bugs, they are features. ar Xiv preprint ar Xiv:1905.02175, 2019. [62] Ali Shafahi, W Ronny Huang, Christoph Studer, Soheil Feizi, and Tom Goldstein. Are adversarial examples inevitable? ar Xiv preprint ar Xiv:1809.02104, 2018. [63] Sébastien Bubeck, Yin Tat Lee, Eric Price, and Ilya Razenshteyn. Adversarial examples from computational constraints. In International Conference on Machine Learning, pages 831 840. PMLR, 2019. [64] Hadi Salman, Andrew Ilyas, Logan Engstrom, Ashish Kapoor, and Aleksander Madry. Do adversarially robust imagenet models transfer better? ar Xiv preprint ar Xiv:2007.08489, 2020. [65] Logan Engstrom, Andrew Ilyas, Shibani Santurkar, Dimitris Tsipras, Brandon Tran, and Aleksander Madry. Adversarial robustness as a prior for learned representations. ar Xiv preprint ar Xiv:1906.00945, 2019. [66] Adi Shamir, Odelia Melamed, and Oriel Ben Shmuel. The dimpled manifold model of adversarial examples in machine learning. ar Xiv preprint ar Xiv:2106.10151, 2021. [67] Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks (2013). ar Xiv preprint ar Xiv:1311.2901, 2013. [68] Lénaïc Chizat. Personal communication, 2021.