# gse_groupwise_sparse_and_explainable_adversarial_attacks__bc286829.pdf Published as a conference paper at ICLR 2025 GSE: GROUP-WISE SPARSE AND EXPLAINABLE ADVERSARIAL ATTACKS Shpresim Sadiku1,2, Moritz Wagner1,2 & Sebastian Pokutta1,2 1Department for AI in Society, Science, and Technology, Zuse Institute Berlin, Germany 2Institute of Mathematics, Technische Universität Berlin, Germany {sadiku,wagner,pokutta}@zib.de Sparse adversarial attacks fool deep neural networks (DNNs) through minimal pixel perturbations, often regularized by the ℓ0 norm. Recent efforts have replaced this norm with a structural sparsity regularizer, such as the nuclear group norm, to craft group-wise sparse adversarial attacks. The resulting perturbations are thus explainable and hold significant practical relevance, shedding light on an even greater vulnerability of DNNs. However, crafting such attacks poses an optimization challenge, as it involves computing norms for groups of pixels within a non-convex objective. We address this by presenting a two-phase algorithm that generates group-wise sparse attacks within semantically meaningful areas of an image. Initially, we optimize a quasinorm adversarial loss using the 1/2 quasinorm proximal operator tailored for non-convex programming. Subsequently, the algorithm transitions to a projected Nesterov s accelerated gradient descent with 2 norm regularization applied to perturbation magnitudes. Rigorous evaluations on CIFAR-10 and Image Net datasets demonstrate a remarkable increase in groupwise sparsity, e.g., 50.9% on CIFAR-10 and 38.4% on Image Net (average case, targeted attack). This performance improvement is accompanied by significantly faster computation times, improved explainability, and a 100% attack success rate. 1 INTRODUCTION Deep neural networks (DNNs) are susceptible to adversarial attacks, where input perturbations deceive the network into producing incorrect predictions (Carlini & Wagner, 2017; Athalye et al., 2018; Zhou et al., 2020; Zhang et al., 2020). These attacks pose serious security risks in real-world systems and raise questions about the robustness of neural classifiers (Stutz et al., 2019). Investigating adversarial attacks is crucial for diagnosing and strengthening DNN vulnerabilities, especially in areas where such attacks have proven effective, including image classification (Chen et al., 2020; Li et al., 2020), image captioning (Xu et al., 2019b), image retrieval (Bai et al., 2020; Feng et al., 2020), question answering (Liu et al., 2020a), autonomous driving (Liu et al., 2019), automatic checkout (Liu et al., 2020b), face recognition (Dong et al., 2019), face detection (Li et al., 2019), etc. While many methods for crafting adversarial examples focus on ℓp neighbourhoods with p 1, recent research has explored the intriguing case of p = 0, leading to sparse adversarial attacks. The prevailing approaches for generating sparse adversarial attacks involve solving ℓ0 formulated problems, employing greedy single-pixel selection (Su et al., 2019), local search techniques (Narodytska & Kasiviswanathan, 2016), utilizing evolutionary algorithms (Croce & Hein, 2019), or relaxing ℓ0 via the ℓ1 ball and applying various algorithms to handle these structures (Carlini & Wagner, 2017; Modas et al., 2019). However, most of these methods only minimize the number of modified pixels and do not constrain the location and the magnitude of the changed pixels. The perturbed pixels can thus exhibit substantial variations in intensity or colour compared to their surroundings, rendering them easily visible (Su et al., 2019). This has motivated the necessity of imposing structure to sparse adversarial attacks by generating group-wise sparse perturbations that are targeted to the main objective in the image (Xu et al., 2019a; Zhu et al., 2021; Imtiaz et al., 2022; Kazemi et al., 2023). Fig. 1 illustrates successful group-wise sparse adversarial examples generated by our proposed algorithm (GSE). This way the generated perturbations are also explainable, i.e., they perturb semantically meaningful pixels in the images, thus enhancing human interpretability. Published as a conference paper at ICLR 2025 Unlike traditional attacks that appear as noise to humans but as features to DNNs (Ilyas et al., 2019), this approach bridges the gap between human perception and machine interpretation. GSE perturbed Changed pixels Figure 1: Adversarial attacks generated by our algorithm. The top row depicts a targeted attack of target label water bottle , and the bottom row depicts an untargeted attack. Contributions. 1 We devise a two-phase approach for creating group-wise sparse adversarial attacks. First, we employ non-convex regularization to select pixels for perturbation, leveraging a combination of the proximal gradient method from (Li & Lin, 2015) and a novel proximity-based pixel selection technique. Then, we apply Nesterov s accelerated gradient descent (NAG) on the chosen pixel coordinates to complete the attack. 2 Our GSE (Group-wise Sparse and Explainable) attacks outperform state-of-the-art (SOTA) methods on CIFAR-10 and Image Net datasets, requiring significantly fewer perturbations. We achieve a 50.9% increase in group-wise sparsity on CIFAR-10 and a 38.4% increase on Image Net (average case, targeted attack), all while maintaining a 100% attack success rate. 3 Through a quantitative assessment of the alignment between perturbations and salient image regions, we underscore the value of groupwise sparse perturbations for explainability analysis. Specifically, our GSE attacks provide superior explainability over SOTA techniques in both group-wise sparse and traditional sparse attack domains. 1.1 RELATED WORK Recent works have introduced methods for generating group-wise sparse and explainable adversarial attacks. Str Attack (Xu et al., 2019a) represents a structured sparse and explainable adversarial attack, relying on the alternating direction method of multipliers (ADMM). It enforces group-wise sparsity through a dynamic sliding mask designed to extract spatial structures from the image. Similarly, SAPF (Fan et al., 2020) leverages ℓp Box ADMM for integer programming to jointly optimize binary selection factors, continuous perturbation magnitudes of pixels, and image sub-regions. FWnuclgroup (Kazemi et al., 2023), abbreviated as FWnucl, is another structured sparse adversarial attack. It quantifies the proximity of perturbations to benign images by utilizing the nuclear group norm, capitalizing on the convexity of nuclear group norm balls through the application of a Frank-Wolfe (FW) optimization scheme (Frank & Wolfe, 1956). Homotopy-Attack (Zhu et al., 2021) is a sparse adversarial attack that operates based on the non-monotone accelerated proximal gradient algorithm (nm APG) (Li & Lin, 2015). It can incorporate group-wise sparsity by segmenting the image pixels using the SLIC superpixel algorithm and applying regularization using the resulting 2, 0 norm . We empirically compare our approach to the aforementioned attacks. New research combines adversarial examples and model explanations, merging key concepts from both areas. On datasets like MNIST (Le Cun et al., 1998), (Ignatiev et al., 2019) illustrates a hitting set duality between adversarial examples and model explanations, while (Xu et al., 2019a) shows the correspondence of attack perturbations with discriminative image features. Our attack strategy perturbs regions neighbouring already perturbed pixels, resulting in a shift of the most susceptible pixels of an image due to our initial sparse perturbations. We conduct empirical analysis to examine the overlap between our attacks and salient image regions. To highlight the crucial role of group-wise sparse perturbations in enhancing explainability, we also incorporate SOTA sparse adversarial attacks (Modas et al., 2019; Croce & Hein, 2019) into our experiments. 2 ADVERSARIAL ATTACK FORMULATION Let X = [Imin, Imax]M N C represent feasible images, where M and N are the image dimensions (height and width) and C is the number of color channels. Let x X be a benign image with label l N, and t N be a target label, where t = l. Additionally, let L : X N R be a classification loss function, such as the cross-entropy loss, tailored for a given classifier C. We introduce a distortion Published as a conference paper at ICLR 2025 Algorithm 1 Forward-Backward Splitting Attack Require: Image x X, target label t, loss function L, sparsity parameter λ > 0, step sizes αk, number of iterations K. 1: Initialize w(0) 0. 2: for k 0, ..., K 1 do 3: w(k+1) proxαkλ p p w(k) αk w(k)L(x + w(k), t) {Definition in Eq. (3).} 4: end for 5: return ˆw = w(K) function D : RM N C R 0. The goal of a targeted adversarial attack is to find an image xadv to which C assigns the target label t and that is in close proximity to x according to the function D. In summary, formulating a targeted adversarial example for the input x can be framed as min w RM N C L(x + w, t) + λD(w), (1) where λ > 0 is a regularization parameter. The untargeted adversarial attack formulation can be found in Appendix A. Note that Eq. (1) is a general framework, and a specific distortion function D must be defined to generate adversarial examples in practice. 2.1 1/2 QUASINORM REGULARIZATION Sparse adversarial attacks can be generated using quasinorm-regularized methods (Wang et al., 2021), where the distortion function in Eq. (1) is set to D( ) := p p , leading to the optimization problem min w RM N C L(x + w, t) + λ w p p, (2) for 0 < p < 1. Here w p = (P i |wi|p)1/p is only a quasinorm since subadditivity is not satisfied for p < 1. Once Eq. (2) is solved, yielding ˆw arg minw RM N C L(x + w, t) + λ w p p, the adversarial example is given by xadv = clip X (x+ ˆw). To obtain ˆw, we employ the forward-backward splitting algorithm (Li & Lin, 2015), as detailed in Algorithm 1. In their work, (Cao et al., 2013) derive a closed-form solution for the proximal operator of p p proxλ p p(w) := arg min y RM N C 1 2λ y w 2 2 + y p p, (3) 2. Given that p p is separable, by (Beck, 2017, Theorem 6.6) it is sufficient to deduce the characterization of proxλ p p in Eq. (3) when MNC = 1. Each component is thus given by h proxλ p p(w) i ϕ2λ(wi) = arccos 54 4 (2λ) 2 3 , S = {i : |wi| > g(2λ)}. (5) While the present solution to Eq. (2) for p = 1 2 can generate highly sparse adversarial perturbations, these perturbations tend to be of large magnitude, making them easily perceptible (Fan et al., 2020). Moreover, this approach does not guarantee that the perturbations will affect the most critical pixels in the image. Our primary objective is to generate group-wise sparse adversarial attacks that are of low magnitude and targeted at the most important regions of the image. 2.2 GROUP-WISE SPARSE ADVERSARIAL ATTACKS OF LOW MAGNITUDE We propose a two-phase method to generate meaningful group-wise sparse adversarial examples with minimal ℓ2 norm perturbation, and enhanced explainability as a natural byproduct (see Sec. 3.4.2). Step 1: Determine which pixel coordinates to perturb by tuning coefficients λ. This step finds the most relevant group-wise sparse pixels to perturb by combining the 1/2 quasinorm regularization Published as a conference paper at ICLR 2025 of Sec. 2.1 with a novel heuristic approach on λ. Instead of a single λ, we consider a vector of tradeoff parameters λ RM N C 0 for the 1/2 quasinorm regularization term, allowing us to adjust each entry individually. Due to Eq. (4), we can formally define the proximal operator to a vector λ RM N C 0 of tradeoff parameters h proxλ p p(w) i i,j,c := h proxλi,j,c p p(w) i i,j,c . (6) In the first ˆk iterations, constituting Step 1, we reduce λi,j,: for pixels located in proximity to already perturbed pixels (Eq. (10)), thereby making them more susceptible to perturbation. Adjust Lambda. After computing iterate w(k) by forward-backward splitting with Nesterov momentum (line 4 in Algorithm 2), we update λ(k) by first building a mask c=1 |w(k)|:,:,c {0, 1}M N, (7) to identify the perturbed pixels. Next, we perform a 2D convolution on m using a square Gaussian blur kernel K Rn n with appropriate padding, yielding a matrix M = m K [0, 1]M N, (8) where entries with indices close to non-zero entries of m are non-zero. Appendix B properly defines this operation. From the blurred perturbation mask M, we compute a matrix M RM N using M ij = M ij + 1, if M ij = 0, q, else, (9) for 0 < q 1. Finally, we increase (decrease) the tradeoff parameters for pixels situated far from (close to) those previously perturbed by computing new tradeoff parameters λ(k+1) i,j,: = 1 M i,j λ(k) i,j,:, where i {1, ..., M} and j {1, ..., N}. (10) We thus expand regions with reduced tradeoff parameters over a specified ˆk iterations, identifying the most relevant group-wise sparse pixels at coordinates (i, j), where λ(ˆk) i,j,c is lower than the initial value λ(0) i,j,c. Unlike existing methods, ours operates independently of predefined structures like pixel partitions. This first step acts only as a heuristic to select the key group-wise sparse pixels, with the proper optimization problem formulated and solved over these selected coordinates in the next step. Step 2: Solve a low magnitude adversarial attack problem only over the selected coordinates. In the remaining iterations, our algorithm formulates a simplified optimization problem over the pixel coordinates V selected in Step 1 given by min w V L(x + w, t) + µ w 2, (11) where µ > 0 (same for all pixels) is a tradeoff parameter controlling ℓ2 norm perturbation magnitude. The subspace V is defined as V := span({ei,j,c | λ(ˆk) i,j,c < λ(0) i,j,c}) RM N C, (12) spanned by standard unit vectors ei,j,c. The projection onto V operates as [PV (w)]i,j,c = wi,j,c, if ei,j,c V, 0, otherwise. (13) Thus, the projection PV zeros out entries not in V while leaving the others unchanged. We apply projected Nesterov s accelerated gradient descent (NAG) to solve Eq. (11), enabling generation of meaningful group-wise sparse perturbations of low magnitude. In the next iterations, perturbations only affect pixels at coordinates (i, j) where λ(ˆk) i,j,: is less than the initial λ. For most perturbed pixels, Published as a conference paper at ICLR 2025 Algorithm 2 GSE: Group-Wise Sparse and Explainable Attack Require: Image x X, target label t, loss function L, regularization parameters λ, µ > 0, step size σ > 0, numbers of iterations ˆk, K, sequence αk. 1: Initialize w(0) 0, λ(0) λ1, define f(w) := L(x + w, t) + µ w 2 2: for k 0, ..., ˆk 1 do 3: w(k+1) proxσλ(k) p p w(k) σ w(k) f(w(k)) 4: w(k+1) (1 αk) w(k+1) + αk w(k) 5: λ(k+1) Adjust Lambda(λ(k), w(k+1)) 6: end for 7: for k = ˆk, ..., K 1 do 8: w(k+1) w(k) σ w(k) f(w(k)) 9: w(k+1) PV (1 αk) w(k+1) + αk w(k) {Definition in Eq. (13).} 10: end for 11: return ˆw = w(K) all channels will be affected, as ei,j,c V across all channels c. This process is succinctly outlined in Algorithm 2, where we use a sequence αk as in (Nesterov, 1983) β0 = 0, βk = 1 1 + 4β2 k 1 , αk = 1 βk+1 (1 βk) . (14) In our tests, we initially perform a section search to find the maximum λ where w(1) = 0. Another section search then determines the appropriate λ for a successful attack. The projected NAG in Algorithm 2 (solving Eq. (11)) converges as NAG solving an unconstrained problem, since the projection PV acts only within subspace V (see Appendix C for proof). 3 EXPERIMENTS We demonstrate the effectiveness of our proposed GSE attack for crafting group-wise sparse and explainable adversarial perturbations. After outlining experimental setup and defining the relevant comparing metrics, Sec. 3.4 compares our approach with multiple previous SOTA group-wise sparse adversarial attacks. Additionally, Sec. 3.4 covers ablation studies regarding the explainability of such attacks, visualizations, empirical time costs, and success against adversarially trained networks. 3.1 DATASETS We experiment on CIFAR-10 (Krizhevsky et al., 2009) and Image Net (Deng et al., 2009) datasets, analyzing DNNs on 10k randomly selected images from both validation sets. For the classifier C on CIFAR-10, we train a Res Net20 model (He et al., 2016) for 600 epochs using stochastic gradient descent, with an initial learning rate of 0.01, reduced by a factor of 10 after 100, 250, and 500 epochs. We set the weight decay to 10 4, momentum to 0.9, and batch size to 512. For Image Net, we employ a Res Net50 (He et al., 2016) and a more robust transformer model, Vi T_B_16 (Dosovitskiy et al., 2020), both with default weights from the torchvision library. All tests are conducted on a machine with an NVIDIA A40 GPU, and our codes, 10k image indices from the Image Net validation dataset, and target labels for targeted Image Net tests are available at https://github.com/wagnermoritz/GSE. 3.2 EVALUATION METRICS Consider n images x(1), ..., x(n) and the corresponding perturbations w(1), ..., w(n), where x(i), x(i) + w(i) X. Among these, let ms n denote the number of successfully generated adversarial examples x(i) + w(i). Then, the attack success rate (ASR) is simply defined as ASR = ms/n. Assume that the first ms adversarial examples are the successful ones. We compute the average number of changed pixels (ACP) via ACP = 1 ms MN i=1 m(i) 0, (15) Published as a conference paper at ICLR 2025 Table 1: Our attack vs. Homotopy-Attack and SAPF on a Res Net20 classifier for CIFAR-10. Perturbations for every attack were computed for all images sequentially. Due to the extensive computation time of Homotopy-Attack, we tested on a limited sample size of 100. Untargeted Targeted Attack ASR ACP ANC ℓ2 d2,0 Time ASR ACP ANC ℓ2 d2,0 Time GSE (Ours) 100% 42.5 1.88 0.76 180 2.93s 100% 107 2.56 1.13 301 4.08s Homotopy 100% 55.3 2.78 0.64 256 499s 100% 113 4.27 1.05 394 754s SAPF 100% 87.3 4.59 0.41 324 250s 100% 109 4.87 0.78 346 277s Table 2: Untargeted attacks on Res Net20 classifier for CIFAR-10, and Res Net50 and Vi T_B_16 classifiers for Image Net. Tested on 10k images of each dataset. Attack ASR ACP ANC ℓ2 d2,0 CIFAR-10 Res Net20 GSE (Ours) 100% 41.7 1.66 0.80 177 Str Attack 100% 118 7.50 1.02 428 FWnucl 94.6% 460 1.99 2.01 594 Image Net Res Net50 GSE (Ours) 100% 1629 8.42 1.50 3428 Str Attack 100% 7265 15.3 2.31 11693 FWnucl 47.4% 13760 3.79 1.81 16345 Image Net Vi T_B_16 GSE (Ours) 100% 941 5.11 1.95 1964 Str Attack 100% 3589 10.8 2.03 8152 FWnucl 57.9% 7515 5.67 3.04 9152 where the perturbation masks m(i) are defined as in Eq. (7). We utilize the ACP instead of the 0-norm since all tested group-wise sparse attacks minimize the ACP. To determine the number of perturbed pixel clusters in x(i) + w(i), we create a mask m(i) as in Eq. (7) and treat it as a graph where adjacent 1-entries are connected nodes. The number of clusters in the perturbation is equal to the number of disjoint connected subgraphs, which we determine by depth-first search. Averaging this number over all successful examples gives us the average number of clusters (ANC). The value w 2,0 for a perturbation w, as suggested in (Zhu et al., 2021), heavily relies on the pixel partitioning method. Instead, we assess the group-wise sparsity of a perturbation by considering all np by np pixel patches of an image, rather than only a subset resulting from partitioning. Let w RM N C, np < M, N, and let G = {G1, ..., Gk} be a set containing the index sets of all overlapping np by np patches in w. Then we define d2,0(w) := |{i : w Gi 2 = 0, i = 1, ..., k}| . (16) Let x and w represent a vectorized image and its corresponding perturbation in Rd, respectively. We measure the explainability of perturbations using the interpretability score (IS) (Xu et al., 2019a), derived from the adversarial saliency map (ASM) (Papernot et al., 2016b). For an image x of true class l and a target class t, ASM(x, l, t) Rd indicates the importance of each feature for classification. The IS is then defined as IS(w, x, l, t) = B(x, l, t) w 2 w 2 , [B(x, l, t)]i = 1, if [ASM(x, l, t)]i > ν, 0, otherwise, (17) where ν is some percentile of the entries of ASM(x, l, t). Note that, when IS(w, x, l, t) approaches 1, the perturbation primarily targets pixels crucial for the class prediction of the model. For a visual representation of salient image regions, we use the class activation map (CAM) (Zhou et al., 2016). Recall that CAM identifies class-specific discriminative image regions, aiding in visually explaining adversarial perturbations. See Appendix D for more details on ASM, IS, and CAM. We evaluate all attacks on these metrics in the untargeted and in the targeted setting. For CIFAR-10, the targeted attacks are performed with respect to each wrong label. For the evaluation of the attacks on Image Net, we choose ten target labels for each image by randomly choosing ten distinct numbers a1, ..., a10 {1, ..., 999} and defining the target labels ti for an image with the true label l by ti = l + ai mod 1000. In the targeted setting, as suggested in (Fan et al., 2020), we give three versions of each metric: best, average, and worst case. Published as a conference paper at ICLR 2025 Table 3: Targeted attacks performed on Res Net20 classifier for CIFAR-10, and Res Net50 and Vi T_B_16 classifiers for Image Net. Tested on 1k images from each dataset, 9 target labels for CIFAR-10 and 10 target labels for Image Net. Best case Average case Worst case Attack ASR ACP ANC ℓ2 d2,0 ASR ACP ANC ℓ2 d2,0 ASR ACP ANC ℓ2 d2,0 CIFAR-10 Res Net20 GSE (Ours) 100% 29.6 1.06 0.68 137 100% 86.3 1.76 1.13 262 100% 162 3.31 1.57 399 Str Attack 100% 78.4 4.56 0.79 352 100% 231 10.1 1.86 534 100% 406 15.9 4.72 619 FWnucl 100% 283 1.18 1.48 515 85.8% 373 2.52 2.54 564 40.5% 495 4.27 3.36 609 Image Net Res Net50 GSE (Ours) 100% 3516 5.89 2.16 5967 100% 12014 14.6 2.93 16724 100% 21675 22.8 3.51 29538 Str Attack 100% 6579 7.18 2.45 9620 100% 15071 18.0 3.97 20921 100% 26908 32.1 6.13 34768 FWnucl 31.1% 9897 3.81 2.02 11295 7.34% 19356 7.58 3.17 26591 0.0% N/A N/A N/A N/A Image Net Vi T_B_16 GSE (Ours) 100% 916 3.35 2.20 1782 100% 2667 7.72 2.87 4571 100% 5920 14.3 3.60 9228 Str Attack 100% 3550 7.85 2.14 5964 100% 8729 17.2 3.50 13349 100% 16047 27.4 5.68 22447 FWnucl 53.2% 5483 4.13 2.77 6718 11.2% 6002 9.73 3.51 7427 0.0% N/A N/A N/A N/A 50 60 70 80 90 (a) Image Net Vi T_B_16 50 60 70 80 90 (b) CIFAR-10 Res Net20 GSE (Ours) Str Attack FWnucl PGD0 Sparse-RS SAPF Figure 2: IS vs. percentile ν for targeted versions of GSE vs. five other attacks. Evaluated on an Image Net Vi T_B_16 classifier (a), and CIFAR-10 Res Net20 classifier (b). Tested on 1k images from each dataset, 9 target labels for CIFAR-10 and 10 target labels for Image Net. 3.3 ATTACK CONFIGURATIONS To find hyperparameters q, σ, µ, and ˆk of our algorithm, we run a grid search where ACP + ANC is the objective and ASR = 1.0 the constraint. The exact hyperparameters can be found in Appendix D.1. Regarding Str Attack, we modify the authors implementation to be compatible with Py Torch, employing the parameters recommended in (Xu et al., 2019a), specifically those from Appendix F. For FWnucl, we implement the nuclear group norm attack in Py Torch, setting ε = 5. We adjust the implementation of Homotopy-Attack to support group-wise sparsity following (Zhu et al., 2021) and use their recommended parameter settings. For SAPF, we use hyperparameters from (Fan et al., 2020) and choose the sparsity parameter to minimize perturbed pixels while retaining a success rate of 100%. We run all the attacks for a total of 200 iterations. 3.4 RESULTS As shown in Tab. 1, our method significantly outperforms the Homotopy-Attack and SAPF in both targeted and untargeted attacks on CIFAR-10. While Homotopy-Attack and SAPF surpass our method only slightly in the 2 norm perturbation magnitude, this metric is of secondary importance in the group-wise sparse attack setting. In contrast, our method excels in increasing the overall sparsity and group-wise sparsity, all while substantially reducing the cluster count and the computation time. Due to Homotopy-Attack s much slower speed and inability to process batches of images in parallel, we exclude it from further experiments. Additionally, as we couldn t replicate SAPF s Image Net results due to floating-point overflow during the ADMM update, we also exclude SAPF. We proceed with the more recent FWnucl and Str Attack in further tests. Published as a conference paper at ICLR 2025 Figure 3: Visual comparison of successful untargeted adversarial instances generated by our attack, Str Attack, and FWnucl. Adversarial examples are shown in the top row, perturbed pixels highlighted in red in the middle row, and the perturbations in the bottom row. The target model is a Res Net50. Perturbations are enhanced for visibility. 3.4.1 EMPIRICAL PERFORMANCE In Tab. 2 we present untargeted attack results on CIFAR-10 and Image Net datasets. Notably, our method and Str Attack attain a 100% ASR in both cases. Furthermore, our algorithm significantly outperforms other attacks in terms of ACP. Specifically, our algorithm achieves an average sparsity of 4.1% on CIFAR-10 and 0.5% on Image Net when attacking a Res Net50 classifier. Additionally, our method demonstrates exceptionally high group-wise sparsity, with an increase of 70.2% on CIFAR-10 and 70.7% on Image Net when attacking a Res Net50 classifier, outperforming the current SOTA. The results for targeted attacks are presented in Tab. 3, where our method and Str Attack achieve a perfect ASR in both cases. However, FWnucl results in a 0% worst-case ASR on the Image Net dataset indicating no worst-case results for FWnucl on Image Net. Our algorithm yields the sparsest perturbations, changing on average only 8.4% of the pixels in CIFAR-10 images and 13.4% of the pixels in Image Net images when attacking a Res Net50 classifier. Moreover, our method achieves remarkable group-wise sparsity, with a significant decrease of d2,0 by 50.9% on CIFAR-10 and 20% GSE example GSE Perturbation Figure 4: Targeted adversarial examples generated by GSE. The target is airship for the first two rows, and golf cart for the last two rows. The attacked model is a VGG19. Perturbations are enhanced for visibility. on Image Net when attacking a Res Net50 classifier, while maintaining the lowest magnitude of attacks measured by the 2 norm. While GSE is outperformed by a SOTA attack (FWnucl) only in the number of clusters for Image Net, this remains insignificant considering FWnucl s high ACP and low ASR compared to our method. Furthermore, replicating experiments with a Vi T_B_16 classifier in Tab. 2 and Tab. 3 demonstrates our algorithm s substantial margin over SOTA, achieving a 73.8% (55.6%) increase in sparsity and 76% (38.4%) in group-wise sparsity in the untargeted (targeted) setting, while also attaining the lowest perturbation norm. Additional results can be found in Appendix E.1. 3.4.2 EXPLAINABILITY Figure 2 illustrates the IS metric for the targeted attacks across various percentiles of ASM scores for samples from the CIFAR-10 and Image Net datasets. Notably, the figure highlights that our method consistently achieves higher IS scores compared to other group-wise sparse attacks, especially for higher percentiles ν. This shows that Published as a conference paper at ICLR 2025 Table 4: Comparison of empirical attack computation time per image. In every experiment, all attacks utilize a batch size of 100 images. Untargeted Targeted CIFAR-10 Image Net CIFAR-10 Image Net Attack Res Net20 Res Net50 Vi T_B_16 Res Net20 Res Net50 Vi T_B_16 GSE (Ours) 0.13s 3.18s 7.05s 0.16s 4.42s 10.3s Str Attack 0.97s 15.2s 33.6s 1.28s 18.5s 34.2s FWnucl 0.25s 9.93s 26.0s 0.32s 11.6s 26.2s Table 5: Comparison of group-wise sparse adversarial attacks for an adversarially trained Res Net50 Image Net classifier. Tested on 10k Image Net instances. Attack ASR ACP ANC ℓ2 d2,0 Image Net Res Net50 GSE (Ours) 98.5% 14246 9.17 13.9 20990 Str Attack 100% 16163 13.5 13.2 22097 FWnucl 35.1% 26542 8.83 15.6 25956 the perturbations generated by GSE are more focused on the most salient regions of the image. To emphasize group-wise sparsity s pivotal role in enhancing explainability, we include SOTA sparse adversarial attacks like PGD0 (Croce & Hein, 2019) and Sparse-RS (Croce et al., 2022), all of which are outperformed by the group-wise sparse attack techniques considered. See Appendix E.2 for additional results. We display the visualizations for untargeted group-wise sparse adversarial examples in Fig. 3, while Fig. 4 presents those for targeted group-wise sparse adversarial examples. Using the CAM technique (Zhou et al., 2016), we demonstrate the alignment between our algorithm s perturbations and localized, class-specific discriminative regions within the images. The generated perturbations effectively encompass the most discriminative areas of the objects, a testament to our algorithm s impressive achievement of explainability. In the perturbation depictions, unperturbed pixels are shown in grey since the perturbations are not required to be non-negative. 3.4.3 SPEED COMPARISON Tab. 4 summarizes the runtime performance of our algorithm vs. SOTA methods. Notably, GSE exhibits significantly faster performance compared to both FWnucl and Str Attack. This speed advantage stems partially from the method used to enforce group-wise sparsity. While Str Attack calculates the Euclidean norm of each pixel group in every iteration and integrates it as regularization using ADMM, FWnucl computes a solution for a nuclear group-norm LMO in each iteration. In contrast, our attack initially computes a solution for the 1/2 quasinorm proximal operator in the first ˆk iterations. Subsequently, the attack transitions to projected NAG with ℓ2 norm regularization, which is less computationally costly than the methods employed by both Str Attack and FWnucl. 3.4.4 GSE AGAINST ADVERSARIALLY TRAINED NETWORKS In this section, we present further results on the performance of GSE against an adversarially trained DNN, following the method outlined in (Madry et al., 2018), a standard approach for improving DNN robustness to adversarial attacks. Tab. 5 shows the results of group-wise sparse attacks on an adversarially trained Res Net50, using PGD projected onto a 2 norm ball of radius 3, from Madry Lab s Robustness package (Engstrom et al., 2019), tested on the Image Net dataset. Overall, GSE shows superior tradeoffs, achieving competitive accuracy while altering the fewest pixels (ACP), maintaining a favorable ratio of perturbed pixels (ACP) to clusters (ANC), and recording the lowest d2,0. Although FWnucl slightly outperforms GSE in ANC, this is due to its excessive number of perturbed pixels (ACP). Str Attack only marginally surpasses GSE in attack success rate (ASR) and ℓ2 norm. These results suggest that GSE generates perturbations that adversarially robust models struggle to defend against effectively. Published as a conference paper at ICLR 2025 4 CONCLUSION We introduced GSE , a novel algorithm for generating group-wise sparse and explainable adversarial attacks. Our approach is rooted in proximal gradient methods for non-convex programming, featuring additional control over changed pixels, and the use of projected NAG technique to solve optimization problems. Extensive experiments validate that GSE excels in producing group-wise sparse adversarial perturbations, all while simultaneously exhibiting the highest level of sparsity and the shortest computation time. Moreover, GSE excels over existing approaches in terms of quantitative metrics for explainability and offers transparency for visualizing the vulnerabilities inherent in DNNs. This endeavour not only establishes a new benchmark for the research community to evaluate the robustness of deep learning algorithms but also suggests a simple defense strategy: employing the adversarial examples generated by GSE in adversarial training (Madry et al., 2018). For more sophisticated solutions, we advocate further exploration. We anticipate no adverse ethical implications or future societal consequences from our research. ACKNOWLEDGEMENTS This research was partially supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) as part of Germany s Excellence Strategy through the Berlin Mathematics Research Center MATH+ (EXC-2046/1, Project ID: 390685689) and the Research Campus MODAL funded by the German Federal Ministry of Education and Research (BMBF grant number 05M14ZAM). Anish Athalye, Logan Engstrom, Andrew Ilyas, and Kevin Kwok. Synthesizing robust adversarial examples. In International conference on machine learning, pp. 284 293. PMLR, 2018. Jiawang Bai, Bin Chen, Yiming Li, Dongxian Wu, Weiwei Guo, Shu-tao Xia, and En-hui Yang. Targeted attack for deep hashing based retrieval. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part I 16, pp. 618 634. Springer, 2020. Amir Beck. First-order methods in optimization. SIAM, 2017. Wenfei Cao, Jian Sun, and Zongben Xu. Fast image deconvolution using closed-form thresholding formulas of lq (q= 12, 23) regularization. Journal of visual communication and image representation, 24(1):31 41, 2013. Nicholas Carlini and David Wagner. Towards evaluating the robustness of neural networks. In 2017 IEEE symposium on security and privacy (sp), pp. 39 57. IEEE, 2017. Weilun Chen, Zhaoxiang Zhang, Xiaolin Hu, and Baoyuan Wu. Boosting decision-based black-box adversarial attacks with random sign flip. In European Conference on Computer Vision, pp. 276 293. Springer, 2020. Francesco Croce and Matthias Hein. Sparse and imperceivable adversarial attacks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4724 4732, 2019. Francesco Croce, Maksym Andriushchenko, Naman D Singh, Nicolas Flammarion, and Matthias Hein. Sparse-rs: a versatile framework for query-efficient sparse black-box adversarial attacks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 6437 6445, 2022. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248 255. Ieee, 2009. Yinpeng Dong, Hang Su, Baoyuan Wu, Zhifeng Li, Wei Liu, Tong Zhang, and Jun Zhu. Efficient decision-based black-box adversarial attacks on face recognition. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 7714 7722, 2019. Published as a conference paper at ICLR 2025 Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. Logan Engstrom, Andrew Ilyas, Hadi Salman, Shibani Santurkar, and Dimitris Tsipras. Robustness (python library), 2019. URL https://github.com/Madry Lab/robustness. Yanbo Fan, Baoyuan Wu, Tuanhui Li, Yong Zhang, Mingyang Li, Zhifeng Li, and Yujiu Yang. Sparse adversarial attack via perturbation factorization. In Proceedings of European Conference on Computer Vision, pp. 35 50. Springer, 2020. Yan Feng, Bin Chen, Tao Dai, and Shu-Tao Xia. Adversarial attack on deep product quantization network for image retrieval. In Proceedings of the AAAI conference on Artificial Intelligence, volume 34, pp. 10786 10793, 2020. Marguerite Frank and Philip Wolfe. An algorithm for quadratic programming. Naval Research Logistics Quarterly, 3, 1956. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Alexey Ignatiev, Nina Narodytska, and Joao Marques-Silva. On relating explanations and adversarial examples. Advances in neural information processing systems, 32, 2019. Andrew Ilyas, Shibani Santurkar, Dimitris Tsipras, Logan Engstrom, Brandon Tran, and Aleksander Madry. Adversarial examples are not bugs, they are features. Advances in neural information processing systems, 32, 2019. Tooba Imtiaz, Morgan Kohler, Jared Miller, Zifeng Wang, Mario Sznaier, Octavia Camps, and Jennifer Dy. Saif: Sparse adversarial and interpretable attack framework. ar Xiv preprint ar Xiv:2212.07495, 2022. Ehsan Kazemi, Thomas Kerdreux, and Liqiang Wang. Minimally distorted structured adversarial attacks. International Journal of Computer Vision, 131(1):160 176, 2023. Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. University of Toronto, 2009. Yann Le Cun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. Huan Li and Zhouchen Lin. Accelerated proximal gradient methods for nonconvex programming. In Advances in neural information processing systems, pp. 379 387, 2015. Yiming Li, Baoyuan Wu, Yan Feng, Yanbo Fan, Yong Jiang, Zhifeng Li, and Shutao Xia. Toward adversarial robustness via semi-supervised robust training. ar Xiv preprint ar Xiv:2003.06974, 2020. Yuezun Li, Xin Yang, Baoyuan Wu, and Siwei Lyu. Hiding faces in plain sight: Disrupting ai face synthesis with adversarial perturbations. ar Xiv preprint ar Xiv:1906.09288, 2019. Aishan Liu, Xianglong Liu, Jiaxin Fan, Yuqing Ma, Anlan Zhang, Huiyuan Xie, and Dacheng Tao. Perceptual-sensitive gan for generating adversarial patches. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 1028 1035, 2019. Aishan Liu, Tairan Huang, Xianglong Liu, Yitao Xu, Yuqing Ma, Xinyun Chen, Stephen J Maybank, and Dacheng Tao. Spatiotemporal attacks for embodied agents. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XVII 16, pp. 122 138. Springer, 2020a. Aishan Liu, Jiakai Wang, Xianglong Liu, Bowen Cao, Chongzhi Zhang, and Hang Yu. Bias-based universal adversarial patch attack for automatic check-out. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XIII 16, pp. 395 410. Springer, 2020b. Published as a conference paper at ICLR 2025 Ze Liu, Han Hu, Yutong Lin, Zhuliang Yao, Zhenda Xie, Yixuan Wei, Jia Ning, Yue Cao, Zheng Zhang, Li Dong, et al. Swin transformer v2: Scaling up capacity and resolution. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 12009 12019, 2022. Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. In Proceedings of the International Conference on Learning Representations (ICLR), 2018. Apostolos Modas, Seyed-Mohsen Moosavi-Dezfooli, and Pascal Frossard. Sparsefool: a few pixels make a big difference. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9087 9096, 2019. Nina Narodytska and Shiva Prasad Kasiviswanathan. Simple black-box adversarial perturbations for deep networks. ar Xiv preprint ar Xiv:1612.06299, 2016. Yurii Nesterov. A method for solving the convex programming problem with convergence rate o(1/kˆ2). Proceedings of the USSR Academy of Sciences, 269, 1983. Nicolas Papernot, Patrick Mc Daniel, and Ian Goodfellow. Transferability in machine learning: from phenomena to black-box attacks using adversarial samples. ar Xiv preprint ar Xiv:1605.07277, 2016a. Nicolas Papernot, Patrick Mc Daniel, Somesh Jha, Matt Fredrikson, Z Berkay Celik, and Ananthram Swami. The limitations of deep learning in adversarial settings. In 2016 IEEE European symposium on security and privacy (Euro S&P), pp. 372 387. IEEE, 2016b. Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition, 2015. David Stutz, Matthias Hein, and Bernt Schiele. Disentangling adversarial robustness and generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6976 6987, 2019. Jiawei Su, Danilo Vasconcellos Vargas, and Kouichi Sakurai. One pixel attack for fooling deep neural networks. IEEE Transactions on Evolutionary Computation, 23(5):828 841, 2019. Hao Wang, Xiangyu Yang, and Xin Deng. A hybrid first-order method for nonconvex ℓp-ball constrained optimization. ar Xiv preprint ar Xiv:2104.04400, 2021. Kaidi Xu, Sijia Liu, Pu Zhao, Pin-Yu Chen, Huan Zhang, Quanfu Fan, Deniz Erdogmus, Yanzhi Wang, and Xue Lin. Structured adversarial attack: Towards general implementation and better interpretability. In International Conference on Learning Representations, 2019a. Yan Xu, Baoyuan Wu, Fumin Shen, Yanbo Fan, Yong Zhang, Heng Tao Shen, and Wei Liu. Exact adversarial attack to image captioning via structured output learning with latent variables. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4135 4144, 2019b. Sergey Zagoruyko and Nikos Komodakis. Wide residual networks. ar Xiv preprint ar Xiv:1605.07146, 2016. Jingfeng Zhang, Xilie Xu, Bo Han, Gang Niu, Lizhen Cui, Masashi Sugiyama, and Mohan Kankanhalli. Attacks which do not kill training make adversarial learning stronger. In International conference on machine learning, pp. 11278 11287. PMLR, 2020. Bolei Zhou, Aditya Khosla, Agata Lapedriza, Aude Oliva, and Antonio Torralba. Learning deep features for discriminative localization. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2921 2929, 2016. Hang Zhou, Dongdong Chen, Jing Liao, Kejiang Chen, Xiaoyi Dong, Kunlin Liu, Weiming Zhang, Gang Hua, and Nenghai Yu. Lg-gan: Label guided adversarial network for flexible targeted attack of point cloud based deep networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10356 10365, 2020. Published as a conference paper at ICLR 2025 Mingkang Zhu, Tianlong Chen, and Zhangyang Wang. Sparse and imperceptible adversarial attack via a homotopy algorithm. In International Conference on Machine Learning, pp. 12868 12877. PMLR, 2021. A UNTARGETED ADVERSARIAL ATTACK FORMULATION We can easily modify Eq. (1) to generate untargeted adversarial attacks by maximizing the loss L with respect to the true label l max w RM N C L(x + w, l) λD(w), (18) where λ > 0 is a regularization parameter. B GAUSSIAN BLUR KERNEL Given m {0, 1}M N and the square Gaussian blur kernel K Rn n, the convolution matrix m K [0, 1]M N is defined via l= ˆn Kk+ˆn+1,l+ˆn+1 mi+k,j+l, where ˆn = n 2 , is the floor function, and n is an odd number. In other words, Gaussian Blur kernel is a weighted mean of the surrounding pixels that gives more weight to the pixel near the current pixel. Setting M = m K and computing the matrix M RM N as in Eq. (9) via M ij = M ij + 1, if M ij = 0, q, else, (19) for 0 < q 1, means we increase the tradeoff parameters in Eq. (10) for pixels with a distance greater than ˆn = n 2 from any perturbed pixels. C EQUIVALENCE OF PROJECTED NAG IN ALGORITHM 2 TO NAG OF UNCONSTRAINED PROBLEMS Here we show that the projected NAG in Algorithm 2 has the same properties as NAG solving an unconstrained problem (Eq. (25)), and thus converges as such. Proof. Consider d = MNC, where our image x [Imin, Imax]d is vectorized. Let I denote the set that encompasses indices corresponding to entries with λ(ˆk) i,j,c λ(0) i,j,c after ˆk iterations. Also, let m := |I| < d. With these considerations, we can formulate the optimization problem in Eq. (11), which arises following iteration ˆk, as min w Rd L(x + w, t) + µ w 2 s.t. Aw = 0, (20) where A {0, 1}m d has rows (0, ..., 0, | {z } i 1 times 1, 0, ..., 0 | {z } d i times ), i I. (21) Published as a conference paper at ICLR 2025 We can eliminate the equality constraints using the nullspace method. Let H Rd d be an orthogonal matrix and R Rm m an upper triangular matrix, obtained by QR-decomposition of A , i.e., Further let H = (Y, Z) , where Y Rm d contains the first m rows of H and Z Rd {d m}. Because an orthogonal matrix H possesses full rank, both Y and Z also exhibit full rank. Hence we can uniquely write any w Rd as w = Y wy + Zwz = H wy wz with wy Rm, wz Rd m. In particular, for any w ker A we have 0 = Aw = AH wy wz = (R , 0) wy wz = R wy. (24) Given Eqs. (21) and (22), we can establish that both A and R possess rank m. Consequently, we can represent any w ker A as Zwz, where wz Rd m. Denoting the wz simply by z, we can formulate an unconstrained problem equivalent to Eq. (20) min z Rd m L(x + Zz, t) + µ Zz 2. (25) Setting f(w) = L(x + w, t) + µ w 2 and F(z) = f(Zz) we get from the update step of NAG wk+1 = Zzk+1 = Z (1 αk) z(k) σ F(z(k)) + αk z(k 1) σ F(z(k 1)) = (1 αk) w(k) σZZ f(w(k)) + αk w(k 1) σZZ f(w(k 1)) = (1 αk) w(k) σPV ( f(w(k))) + αk w(k 1) σPV ( f(w(k 1))) = PV (1 αk) w(k) σ f(w(k)) + αk w(k 1) σ f(w(k 1)) , where the last equality holds since w(k), w(k 1) V . Thus, the projected NAG in Algorithm 2 shares the properties of NAG solving the unconstrained problem in Eq. (25), ensuring its convergence. D EXPLANATION OF ASM, IS, AND CAM EVALUATION METRICS We provide further details on the ASM, IS, and CAM metrics as outlined in Sec. 3.2 of the paper. Consider a vectorized image x [Imin, Imax]d with true label l and target label t. Let Z(x) represent the logits of a classifier. The adversarial saliency map (ASM) (Papernot et al., 2016b; Xu et al., 2019a) is defined as follows [ASM(x, l, t)]i = Z(x)t S = i {1, ..., d} Z(x)t xi 0 or Z(x)l xi 0 . (27) The metric ASM(x, l, t) Rd 0 provides a measure of importance for each pixel. It is worth noting that a higher ASM value signifies greater pixel significance. We compute a binary mask B(x, l, t) {0, 1}d by [B(x, l, t)]i = 1, if [ASM(x, l, t)]i > ν, 0, otherwise, (28) where ν is some percentile of the entries of ASM(x, l, t). Given an adversarial perturbation w Rd, we can now compute the interpretability score (IS) as IS(w, x, l, t) = B(x, l, t) w 2 Published as a conference paper at ICLR 2025 Note that when IS(w, x, l, t) approaches 1, the perturbation primarily targets pixels crucial for the class prediction of the model. Conversely, IS scores nearing zero do not lend themselves to meaningful interpretation based on ASM scores. Let C be a convolutional neural network classifier and fk[i, j] be the activation of the unit k at the coordinates (i, j) in the last convolutional layer of C evaluated at x. Then, using global average pooling after the last convolutional layer, the input to the softmax corresponding to label l is X i,j fk[i, j] = X k wl kfk[i, j], where wl k are the weights corresponding to the label l for unit k. Since wl k indicate the importance of P i,j fk[i, j] for class l, the class activation map (CAM) (Zhou et al., 2016) is defined by [CAMl]i,j = X k wl kfk[i, j], and directly indicates the importance of activation at (i, j) for the classification of x as class l. To make the comparison of CAMl and x easier, the resulting class activation map is upscaled to the size of x using bicubic interpolation. D.1 HYPERPARAMETERS OF THE GSE ALGORITHM To find q, σ, µ, and ˆk, we run a grid search where ACP + ANC is the objective and ASR = 1.0 the constraint. Specifically, for CIFAR-10, we set q = 0.25, σ = 0.005, µ = 1, and ˆk = 30, while for Image Net, we use q = 0.9, σ = 0.05, µ = 0.1, and ˆk = 50. Due to the significant efficiency of our attack compared to other group-wise sparse attacks, we can easily run a grid search to find appropriate hyperparameters when working with different datasets. The grid search code used to find the hyperparameters for CIFAR-10 and Image Net is attached in https://github.com/wagnermoritz/GSE. E ADDITIONAL EXPERIMENTS E.1 EMPIRICAL PERFORMANCE In this section, we present untargeted and targeted attack results on the CIFAR-10 dataset (Krizhevsky et al., 2009) using a Wide Res Net classifier (Zagoruyko & Komodakis, 2016) and Image Net dataset (Deng et al., 2009) using a VGG19 classifier (Simonyan & Zisserman, 2015). Additionally, we include results on the NIPS2017 dataset (available at https://www.kaggle.com/competitions/ nips-2017-defense-against-adversarial-attack/data), comprising 1k images of the same dimensionality (299 299 3) as the Image Net dataset, on which we test the attacks against a VGG19 (Simonyan & Zisserman, 2015) and a Res Net50 (He et al., 2016) classifier. The results in Tab. 6 and Tab. 7 reinforce the findings of Sec. 3.4.1, highlighting GSE s ability to produce SOTA group-wise sparse adversarial attacks. Table 6: Untargeted attacks on a Wide Res Net classifier for CIFAR-10, VGG19 classifier for Image Net, and VGG19 and Res Net50 classifiers for NIPS2017. Tested on 10k images from the CIFAR-10 dataset, 10k images from the Image Net dataset, and 1k images from the NIPS2017 dataset. Attack ASR ACP ANC ℓ2 d2,0 CIFAR-10 Wide Res Net GSE (Ours) 100% 73.1 1.53 0.71 229 Str Attack 100% 78.9 4.76 1.01 289 FWnucl 97.6% 303 1.87 2.28 438 SAPF 100% 95.1 3.96 0.51 292 Image Net VGG19 GSE (Ours) 100% 813.9 5.43 1.63 1855 Str Attack 100% 3035 8.57 1.78 6130 FWnucl 85.8% 5964 2.46 2.73 7431 NIPS2017 VGG19 GSE (Ours) 100% 973 6.65 1.49 2254 Str Attack 100% 3589 10.1 1.56 6487 FWnucl 90.1% 6099 2.35 2.71 7701 NIPS2017 Res Net50 GSE (Ours) 100% 1332 9.81 1.46 3036 Str Attack 100% 8211 17.1 2.79 13177 FWnucl 48.2% 14571 5.01 1.77 17086 Published as a conference paper at ICLR 2025 Table 7: Targeted attacks performed a Wide Res Net classifier for CIFAR-10, VGG19 classifier for Image Net, and VGG19 and Res Net50 classifiers for NIPS2017. Tested on 1k images from each dataset, 9 target labels for CIFAR-10 and 10 target labels for Image Net and NIPS2017. Best case Average case Worst case Attack ASR ACP ANC ℓ2 d2,0 ASR ACP ANC ℓ2 d2,0 ASR ACP ANC ℓ2 d2,0 CIFAR-10 Wide Res Net GSE (Ours) 100% 49.8 1.04 0.50 141 94.8% 112 1.45 0.81 238 85.2% 234 2.71 1.25 388 Str Attack 100% 61.7 2.32 0.82 244 98.8% 196 6.31 1.64 431 93.1% 398 11.3 3.01 579 FWnucl 100% 210 1.12 1.56 324 82.1% 269 2.11 2.06 390 44.2% 365 3.56 3.58 472 SAPF 100% 85.3 1.97 0.45 223 94.8% 91.9 3.61 0.91 278 75.2% 101 6.25 1.36 371 Image Net VGG19 GSE (Ours) 100% 1867 6.15 2.54 3818 100% 6580 14.5 3.28 10204 100% 15526 24.8 4.39 20846 Str Attack 100% 4303 6.11 2.88 7013 100% 11568 16.7 3.59 17564 100% 21552 29.8 5.16 30451 FWnucl 56.2% 4612 1.51 2.90 6973 18.3% 9134 2.53 3.65 12050 0.0% N/A N/A N/A N/A NIPS2017 VGG19 GSE (Ours) 100% 2155 5.96 2.78 4617 100% 6329 15.4 3.44 11304 100% 15541 25.2 4.19 21581 Str Attack 100% 4197 6.55 2.45 6847 100% 11326 18.5 3.66 17332 100% 21018 34.1 5.38 30103 FWnucl 52.6% 4543 1.82 2.63 5783 14.1% 7517 2.92 3.81 11007 0.0% N/A N/A N/A N/A NIPS2017 Res Net50 GSE (Ours) 100% 3505 6.31 2.51 5698 100% 9127 16.2 2.87 15704 100% 20247 27.5 3.12 26574 Str Attack 100% 6344 6.92 2.54 9229 100% 15278 18.6 4.05 21090 100% 27922 33.7 6.51 35927 FWnucl 30.2% 9812 2.49 2.62 12845 11.8% 11512 8.02 3.69 19493 0.0% N/A N/A N/A N/A 50 60 70 80 90 PGD0 Sparse-RS Figure 5: IS vs. percentile ν for targeted versions of GSE vs. four other attacks. Evaluated on an Image Net VGG19 classifier. Tested on 1k images and 10 target labels for Image Net. E.2 EXPLAINABILITY Fig. 5 reinforces the findings of Sec. 3.4.2, demonstrating GSE s consistently higher IS scores compared to both SOTA group-wise sparse and SOTA sparse attacks, while using a VGG19 classifier (Simonyan & Zisserman, 2015) on Image Net. This underscores that GSE-generated perturbations are more focused on the most salient regions of the image. E.3 TRANSFERABILITY An intriguing application is evaluating the transferability performance (Papernot et al., 2016a) of various group-wise sparse attacks in the black-box setting, assessing whether an attack maintains a high ASR when targeting a different model. From Tab. 8, our GSE attack demonstrates transferability on par with SOTA group-wise sparse methods, outperforming Str Attack in the lower-triangular case and FWnucl in the diagonal case, while achieving a strong balance between in-model and out-model performance. The additional model, Swin-V2-B, is sourced from (Liu et al., 2022). Table 8: Transferability of targeted attacks on Image Net. Attack Attacked Model Res Net50 Swin-V2-B VGG19 GSE Res Net50 100% 5.15% 15.4% Swin-V2-B 8.61% 99.8% 18.3% VGG19 5.01% 4.26% 100% Str Attack Res Net50 100% 5.27% 18.1% Swin-V2-B 8.72% 100% 19.9% VGG19 4.95% 4.12% 100% FWnucl Res Net50 65.3% 5.30% 20.0% Swin-V2-B 10.2% 54.9% 19.3% VGG19 9.62% 6.05% 97.5% Published as a conference paper at ICLR 2025 Table 9: Ablation study on ˆk for GSE. Tested on with 1,000 samples from the Image Net dataset and a Res Net50 classifier. All values of ˆk lead to an attack success rate of 100%. ˆk 10 20 30 40 50 60 70 80 90 100 ACP 1260 1342 1437 1399 1558 1519 1462 1450 1462 1433 ANC 16.3 13.4 10.5 9.69 8.11 6.73 5.90 5.14 4.92 4.36 d2,0 3529 3559 3442 3387 3356 3265 3048 2814 2773 2656 ℓ2 1.44 1.42 1.41 1.39 1.43 1.41 1.42 1.40 1.43 1.44 E.4 ABLATION STUDY ON ˆk We conduct an ablation study on ˆk to evaluate the impact of Step 1 on GSE attack performance, as shown in Tab. 9. We see that ˆk is inversely correlated with the number of clusters (ANC) and group sparsity (d2,0), as anticipated. No significant correlations with other metrics were observed.