# boosting_adversarial_training_with_hypersphere_embedding__f00a5eeb.pdf Boosting Adversarial Training with Hypersphere Embedding Tianyu Pang , Xiao Yang , Yinpeng Dong, Kun Xu, Jun Zhu, Hang Su Dept. of Comp. Sci. & Tech., Institute for AI, BNRist Center Tsinghua-Bosch Joint ML Center, THBI Lab, Tsinghua University, Beijing, China {pty17, yangxiao19, dyp17}@mails.tsinghua.edu.cn kunxu.thu@gmail.com, {suhangss, dcszj}@mail.tsinghua.edu.cn Adversarial training (AT) is one of the most effective defenses against adversarial attacks for deep learning models. In this work, we advocate incorporating the hypersphere embedding (HE) mechanism into the AT procedure by regularizing the features onto compact manifolds, which constitutes a lightweight yet effective module to blend in the strength of representation learning. Our extensive analyses reveal that AT and HE are well coupled to benefit the robustness of the adversarially trained models from several aspects. We validate the effectiveness and adaptability of HE by embedding it into the popular AT frameworks including PGD-AT, ALP, and TRADES, as well as the Free AT and Fast AT strategies. In the experiments, we evaluate our methods under a wide range of adversarial attacks on the CIFAR-10 and Image Net datasets, which verifies that integrating HE can consistently enhance the model robustness for each AT framework with little extra computation. 1 Introduction The adversarial vulnerability of deep learning models has been widely recognized in recent years [4, 22, 60]. To mitigate this potential threat, a number of defenses have been proposed, but most of them are ultimately defeated by the attacks adapted to the specific details of the defenses [2, 8]. Among the existing defenses, adversarial training (AT) is a general strategy achieving the state-of-the-art robustness under different settings [53, 63, 71, 73, 74, 76, 83]. Various efforts have been devoted to improving AT from different aspects, including accelerating the training procedure [56, 57, 70, 79] and exploiting extra labeled and unlabeled training data [1, 10, 25, 78], which are conducive in the cases with limited computational resources or additional data accessibility. In the meanwhile, another research route focuses on boosting the adversarially trained models via imposing more direct supervision to regularize the learned representations. Along this line, recent progress shows that encoding triplet-wise metric learning or maximizing the optimal transport (OT) distance of data batch in AT is effective to leverage the inter-sample interactions, which can promote the learning of robust classifiers [34, 41, 45, 80]. However, optimization on the sampled triplets or the OT distance is usually of high computational cost, while the sampling process in metric learning could also introduce extra class biases on unbalanced data [48, 55]. In this work, we provide a lightweight yet competent module to tackle several defects in the learning dynamics of existing AT frameworks, and facilitate the adversarially trained networks learning more robust features. Methodologically, we augment the AT frameworks by integrating the hypersphere embedding (HE) mechanism, which normalizes the features in the penultimate layer and the weights in the softmax layer with an additive angular margin. Except for the generic benefits of HE on Equal contribution. Corresponding author. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. Table 1: Formulations of the AT frameworks without ( ) or with ( ) HE. The notations are defined in Sec. 2.2. We substitute the adversarial attacks in ALP with untargeted PGD as suggested [18]. Strategy HE Training objective LT Adversarial objective LA PGD-AT LCE(f(x ), y) LCE(f(x ), y) CE( ef(x ), y) LCE( ef(x ), y) ALP α LCE(f(x), y)+(1 α) LCE(f(x ), y)+λ W (z z ) 2 LCE(f(x ), y) CE( ef(x), y)+(1 α) Lm CE( ef(x ), y)+λ f W (ez ez ) 2 LCE( ef(x ), y) TRADES LCE(f(x), y) + λ LCE(f(x ), f(x)) LCE(f(x ), f(x)) CE( ef(x), y) + λ LCE( ef(x ), ef(x)) LCE( ef(x ), ef(x)) learning angularly discriminative representations [37, 39, 67, 72], we contribute to the extensive analyses (detailed in Sec. 3) showing that the encoded HE mechanism naturally adapts to AT. To intuitively explain the main insights, we take a binary classification task as an example, where the cross-entropy (CE) objective equals to maximizing L(x) = (W0 W1) z = W01 z cos(θ) on an input x with label y = 0. (i) If x is correctly classified, there is L(x) > 0, and adversaries aim to craft x such that L(x ) < 0. Since W01 and z are always positive, they cannot alter the sign of L. Thus feature normalization (FN) and weight normalization (WN) encourage the adversaries to attack the crucial component cos(θ), which results in more efficient perturbations when crafting adversarial examples in AT; (ii) In a data batch, points with larger z will dominate (vicious circle on increasing z ), which makes the model ignore the critical component cos(θ). FN alleviates this problem by encouraging the model to devote more efforts on learning hard examples, and well-learned hard examples will dynamically have smaller weights during training since cos(θ) is bounded; This can promote the worst-case performance under adversarial attacks; (iii) When there are much more samples of label 0, the CE objective will tend to have W0 W1 to minimize the loss. WN can relieve this trend and encourage W0 and W1 to diversify in directions. This mechanism alleviates the unbalanced label distributions caused by the untargeted or multi-targeted attacks applied in AT [23, 40], where the resulted adversarial labels depend on the semantic similarity among classes; (iv) The angular margin (AM) induces a larger inter-class variance and margin under the angular metric to further improve model robustness, which plays a similar role as the margin in SVM. Our method is concise and easy to implement. To validate the effectiveness, we consider three typical AT frameworks to incorporate with HE, namely, PGD-AT [40], ALP [29], and TRADES [81], as summarized in Table 1. We further verify the generality of our method by evaluating the combination of HE with previous strategies on accelerating AT, e.g., Free AT [57] and Fast AT [70]. In Sec. 4, we empirically evaluate the defenses on CIFAR-10 [31] and Image Net [15] under several different adversarial attacks, including the commonly adopted PGD [40] and other strong ones like the feature attack [35], FAB [13], SPSA [65], and NES [27], etc. We also test on the CIFAR-10-C and Image Net C datasets with corrupted images to inspect the robustness under general transformations [24]. The results demonstrate that incorporating HE can consistently improve the performance of the models trained by each AT framework, while introducing little extra computation. 2 Methodology In this section, we define the notations, introduce the hypersphere embedding (HE) mechanism, and provide the formulations under the adversarial training (AT) frameworks. Due to the limited space, we extensively introduce the related work in Appendix B, including those on combining metric learning with AT [34, 41, 44, 46, 80] and further present their bottlenecks. 2.1 Notations For the classification task with L labels in [L] := {1, , L}, a deep neural network (DNN) can be generally denoted as the mapping function f(x) for the input x as f(x) = S(W z + b), (1) where z = z(x; ω) is the extracted feature with model parameters ω, the matrix W = (W1, , WL) and vector b are respectively the weight and bias in the softmax layer, and S(h) : RL RL is the softmax function. One common training objective for DNNs is the cross-entropy (CE) loss defined as LCE(f(x), y) = 1 y log f(x), (2) where 1y is the one-hot encoding of label y and the logarithm of a vector is taken element-wisely. In this paper, we use (u, v) to denote the angle between vectors u and v. 2.2 The AT frameworks with HE Adversarial training (AT) is one of the most effective and widely studied defense strategies against adversarial vulnerability [6, 33]. Most of the AT methods can be formulated as a two-stage framework: min ω,W E [LT(ω, W|x, x , y)] , where x = arg max x B(x) LA(x |x, y, ω, W). (3) Here E[ ] is the expectation w.r.t. the data distribution, B(x) is a set of allowed points around x, LT and LA are the training and adversarial objectives, respectively. Since the inner maximization and outer minimization problems are mutually coupled, they are iteratively executed in training until the model parameters ω and W converge [40]. To promote the performance of the adversarially trained models, recent work proposes to embed pair-wise or triplet-wise metric learning into AT [34, 41, 80], which facilitates the neural networks learning more robust representations. Although these methods are appealing, they could introduce high computational overhead [41], cause unexpected class biases [26], or be vulnerable under strong adversarial attacks [35]. In this paper, we address the above deficiencies by presenting a lightweight yet effective module that integrates the hypersphere embedding (HE) mechanism with an AT procedure. Though HE is not completely new, our analysis in Sec. 3 demonstrates that HE naturally adapts to the learning dynamics of AT and can induce several advantages special to the adversarial setting. Specifically, the HE mechanism involves three typical operations including feature normalization (FN), weight normalization (WN), and angular margins (AM), as described below. Note that in Eq. (1) there is W z = (W 1 z, , W L z), and l [L], the inner product W l z = Wl z cos (θl), where θl = (Wl, z).1 Then the WN and FN operations can be denoted as WN operation: f Wl = Wl Wl ; FN operation: ez = z z , (4) Let cos θ = (cos (θ1), , cos (θL)) and f W be the weight matrix after executing WN on each column vector Wl. Then, the output predictions of the DNNs with HE become ef(x) = S(f W ez) = S(cos θ), (5) where no bias vector b exists in ef(x) [38, 67]. In contrast, the AM operation is only performed in the training phase, where ef(x) is fed into the CE loss with a margin m [68], formulated as Lm CE( ef(x), y) = 1 y log S(s (cos θ m 1y)). (6) Here s > 0 is a hyperparameter to improve the numerical stability during training [67]. To highlight our main contributions in terms of methodology, we summarize the proposed formulas of AT in Table 1. We mainly consider three representative AT frameworks including PGD-AT [40], ALP [29], and TRADES [81]. The differences between our enhanced versions (with HE) from the original versions (without HE) are colorized. Note that we apply the HE mechanism both on the adversarial objective LA for constructing adversarial examples (the inner maximization problem), and the training objective LT for updating parameters (the outer minimization problem). 3 Analysis of the benefits In this section, we analyze the benefits induced by the mutual interaction between AT and HE under the ℓp-bounded threat model [9], where B(x) = {x | x x p ϵ} and ϵ is the maximal perturbation. Detailed proofs for the conclusions below can be found in Appendix A. 1We omit the subscript of ℓ2-norm without ambiguity. 3.1 Formalized first-order adversary Most of the adversarial attacks applied in AT belong to the family of first-order adversaries [58], due to the computational efficiency. We first define the vector function Up as Up(u) = arg max v p 1 u v, where u Up(u) = u q. (7) Here q is the dual norm of p with 1 q = 1 [5]. Specially, there are U2(u) = u u 2 and U (u) = sign(u). If u = x LA, then Up( x LA) is the direction of greatest increase of LA under the first-order Taylor s expansion [30] and the ℓp-norm constraint, as stated below: Lemma 1. (First-order adversary) Given the adversarial objective LA and the set B(x) = {x | x x p ϵ}, under the first-order Taylor s expansion, the solution for maxx B(x) LA(x ) is x = x + ϵUp( x LA(x)). Furthermore, there is LA(x ) = LA(x) + ϵ x LA(x) q. According to the one-step formula in Lemma 1, we can generalize to the multi-step generation process of first-order adversaries under the ℓp-bounded threat model. For example, in the t-th step of the iterative attack with step size η [32], the adversarial example x(t) is updated as x(t) = x(t 1) + ηUp( x LA(x(t 1))), (8) where the increment of the loss is LA = LA(x(t)) LA(x(t 1)) = η x LA(x(t 1)) q. 3.2 The inner maximization problem in AT As shown in Table 1, the adversarial objectives LA are usually the CE loss between the adversarial prediction f(x ) and the target prediction f(x) or 1y. Thus, to investigate the inner maximization problem of LA, we expand the gradient of CE loss w.r.t. x as below: Lemma 2. (The gradient of CE loss) Let Wij = Wi Wj be the residual vector between two weights, and z = z(x ; ω) be the mapped feature of the adversarial example x , then there is x LCE(f(x ), f(x)) = X i =j f(x)if(x )j x (W ij z ). (9) If f(x) = 1y is the one-hot label vector, we have x LCE(f(x ), y) = P l =y f(x )l x (W ylz ). Lemma 2 indicates that the gradient of CE loss can be decomposed into the linear combination of the gradients on the residual logits W ij z . Let y be the predicted label on the finally crafted adversarial example x , where y = y. Based on the empirical observations [21, 43], we are justified to assume that f(x)y is much larger than f(x)l for l = y, and f(x )y is much larger than f(x )l for l {y, y }. Then we can approximate the linear combination in Eq. (9) with the dominated term as x LCE(f(x ), f(x)) f(x)yf(x )y x (W yy z ), where Wyy = Wy Wy . (10) Let θ yy = (Wyy , z ), there is W yy z = Wyy z cos(θ yy ) and Wyy does not depend on x . Thus by substituting Eq. (10) into Eq. (8), the update direction of each attacking step becomes Up[ x LCE(f(x ), f(x))] Up[ x ( z cos(θ yy ))], (11) where the factor f(x)yf(x )y is eliminated according to the definition of Up[ ]. Note that Eq. (11) also holds when f(x) = 1y, and the resulted adversarial objective is analogous to the C&W attack [7]. 3.3 Benefits from feature normalization To investigate the effects of FN alone, we deactivate the WN operation in Eq. (5) and denote f(x) = S(W ez). (12) Then similar to Eq. (11), we can obtain the update direction of the attack with FN applied as Up[ x LCE(f(x ), f(x))] Up[ x (cos(θ yy ))]. (13) norm constraint norm constraint Contour lines of Adv. example updated by (this work) Adv. example updated by Decision boundary noneffective norm constraint Figure 1: Intuitive illustration in the input space. When applying FN in LA, the adversary can take more effective update steps to move x across the decision boundary defined by cos(θ yy ) = 0. More effective adversarial perturbations. In the AT procedure, we prefer to craft adversarial examples more efficiently to reduce the computational burden [36, 66, 70]. As shown in Fig. 1, to successfully fool the model to classify x into the label y , the adversary needs to craft iterative perturbations to move x across the decision boundary defined by cos(θ yy ) = 0. Under the first-order optimization, the most effective direction to reach the decision boundary is along x (cos(θ yy )), namely, the direction with the steepest descent of cos(θ yy ). In contrast, the direction of x z is nearly tangent to the contours of cos(θ yy ), especially in high-dimension spaces, which is noneffective as the adversarial perturbation. Actually from Eq. (1) we can observe that when there is no bias term in the softmax layer, changing the norm z will not affect the predicted labels at all. By comparing Eq. (11) and Eq. (13), we can find that applying FN in the adversarial objective LA exactly removes the noneffective component caused by z , and encourages the adversarial perturbations to be aligned with the effective direction x (cos(θ yy )) under the ℓp-norm constraint. This facilitate crafting adversarial examples with fewer iterations and improve the efficiency of the AT progress, as empirically verified in the left panel of Fig. 2. Besides, in Fig. 1 we also provide three instantiations of the ℓp-norm constraint. We can see that if we do not use FN, the impact of the noneffective component of x z could be magnified under, e.g., the ℓ -norm constraint, which could consequently require more iterative steps and degrade the training efficiency. Better learning on hard (adversarial) examples. As to the benefits of applying FN in the training objective LT, we formally show that FN can promote learning on hard examples, as empirically observed in the previous work [50]. In the adversarial setting, this property can promote the worstcase performance under potential adversarial threats. Specifically, the model parameters ω is updated towards ω LCE. When FN is not applied, we can use similar derivations as in Lemma 2 to obtain ω LCE(f(x), y) = X l =y f(x)l Wyl (cos(θyl) ω z + z ω cos(θyl)), (14) where Wyl = Wy Wl and θyl = (Wyl, z). According to Eq. (14), when we use a mini-batch of data to update ω, the inputs with small ω z or ω cos(θyl) contribute less in the direction of model updating, which are qualitatively regarded as hard examples [50]. This causes the training process to devote noneffective efforts to increasing z for easy examples and consequently overlook the hard ones, which leads to vicious circles and could degrade the model robustness against strong adversarial attacks [51, 59]. As shown in Fig. 3, the hard examples in AT are usually the crafted adversarial examples, which are those we actually expect the model to focus on in the AT procedure. In comparison, when FN is applied, there is ω ez = 0, then ω is updated towards ω LCE(f(x), y) = X f(x)l Wyl ω cos(θyl). (15) In this case, due to the bounded value range [ 1, 1] of cosine function, the easy examples will contribute less when they are well learned, i.e., have large cos(θyl), while the hard examples could later dominate the training. This causes a dynamic training procedure similar to curriculum learning [3]. 3.4 Benefits from weight normalization In the AT procedure, we usually apply untargeted attacks [18, 40]. Since we do not explicit assign targets, the resulted prediction labels and feature locations of the crafted adversarial examples will Table 2: Classification accuracy (%) on CIFAR-10 under the white-box threat model. The perturbation ϵ = 0.031, step size η = 0.003. We highlight the best-performance model under each attack. Defense Clean PGD-20 PGD-500 MIM-20 FGSM Deep Fool C&W Fea Att. FAB PGD-AT 86.75 53.97 51.63 55.08 59.70 57.26 84.00 52.38 51.23 PGD-AT+HE 86.19 59.36 57.59 60.19 63.77 61.56 84.07 52.88 54.45 ALP 87.18 52.29 50.13 53.35 58.99 59.40 84.96 49.55 50.54 ALP+HE 89.91 57.69 51.78 58.63 65.08 65.19 87.86 48.64 51.86 TRADES 84.62 56.48 54.84 57.14 61.02 60.70 81.13 55.09 53.58 TRADES+HE 84.88 62.02 60.75 62.71 65.69 60.48 81.44 58.13 53.50 Table 3: Validation of combining Fast AT and Free AT with HE and m-HE on CIFAR-10. We report the accuracy (%) on clean and PGD, as well as the total training time (min). Defense Epo. Clean PGD-50 Time Fast AT 30 83.80 46.40 11.38 Fast AT+HE 30 82.58 52.55 11.48 Fast AT+m-HE 30 83.14 53.49 11.49 Free AT 10 77.21 46.14 15.78 Free AT+HE 10 76.85 50.98 15.87 Free AT+m-HE 10 77.59 51.85 15.91 Table 4: Top-1 classification accuracy (%) on Image Net under the white-box threat model. Model Method Clean PGD-10 PGD-50 Res Net-50 Free AT 60.28 32.13 31.39 Free AT+HE 61.83 40.22 39.85 Res Net-152 Free AT 65.20 36.97 35.87 Free AT+HE 65.41 43.24 42.60 WRN-50-2 Free AT 64.18 36.24 35.38 Free AT+HE 65.28 43.83 43.47 WRN-101-2 Free AT 66.15 39.35 38.23 Free AT+HE 66.37 45.35 45.04 depend on the unbalanced semantic similarity among different classes. For example, the learned features of dogs and cats are usually closer than those between dogs and planes, so an untargeted adversary will prefer to fool the model to predict the cat label on a dog image, rather than the plane label [42]. To understand how the adversarial class biases affect training, assuming that we perform the gradient descent on a data batch D = {(xk, yk)}k [N]. Then we can derive that l [L], the softmax weight Wl is updated towards Wl LCE(D) = X xk D f(xk)l zk, (16) where Dl is the subset of D with true label l. We can see that the weight Wl will tend to have larger norm when there are more data or easy examples in class l, i.e., larger |Dl| or zk for xk Dl. Besides, if an input x in the batch is adversarial, then f(x)y is usually small and consequently z will have a large effect on the update of Wy. Since there is W yy z < 0, Wy will be updated towards Wy both in norm and direction, which causes repetitive oscillation during training. When applying WN, the update of Wl will only depend on the averaged feature direction within each class, which alleviates the noneffective oscillation on the weight norm and speed up training [52]. Besides, when FN and WN are both applied, the inner products W z in the softmax layer will become the angular metric cos θ, as shown in Eq. (5). Then we can naturally introduce AM to learn angularly more discriminative and robust features [17, 75]. 3.5 Modifications to better utilize strong adversaries In most of the AT procedures, the crafted adversarial examples will only be used once in a single training step to update the model parameters [40], which means hard adversarial examples may not have an chance to gradually dominate the training as introduced in Eq. (15). Since ω cos(θyy ) = sin(θyy ) ωθyy , the weak adversarial examples around the decision boundary with θyy 90 have higher weights sin(θyy ). This makes the model tend to overlook the strong adversarial examples with large θyy , which contain abundant information. To be better compatible with strong adversaries, an easy-to-implement way is to directly substitute S(cos θ) in Eq. (5) with S( θ), using the arccos operator. We name this form of embedding as modified HE (m-HE), as evaluated in Table 3. Table 5: Top-1 classification accuracy (%) on CIFAR-10-C and Image Net-C. The models are trained on the original datasets CIFAR-10 and Image Net, respectively. Here m CA refers to the mean accuracy averaged on different corruptions and severity. Full version of the table is in Appendix C.6. Defense m CA Blur Weather Digital Defocus Glass Motion Zoom Snow Frost Fog Bright Contra Elastic Pixel JPEG CIFAR-10-C PGD-AT 77.23 81.84 79.69 77.62 80.88 81.32 77.95 61.70 84.05 44.55 80.79 84.76 84.35 PGD-AT+HE 77.29 81.86 79.45 78.17 80.87 80.77 77.98 62.45 83.67 45.11 80.69 84.16 84.10 ALP 77.73 81.94 80.31 78.23 80.97 81.74 79.26 61.51 84.88 45.86 80.91 85.09 84.68 ALP+HE 80.55 80.87 85.23 81.26 84.43 85.14 83.89 68.83 88.33 50.74 84.44 87.44 87.28 TRADES 75.36 79.84 77.72 76.34 78.66 79.52 76.94 59.68 82.06 43.80 78.53 82.65 82.31 TRADES+HE 75.78 80.55 77.61 77.26 79.62 79.23 76.53 61.39 82.33 45.04 79.29 82.50 82.40 Image Net-C Free AT 28.22 19.15 26.63 25.75 28.25 23.03 23.47 3.71 45.18 5.40 41.76 48.78 52.55 Free AT+HE 30.04 21.16 29.28 28.08 30.76 26.62 28.35 5.34 49.88 7.03 44.72 51.17 55.05 1 2 3 4 5 6 7 Iteration Steps of PGD in Training Robust Accuracy (%) PGD-AT + HE PGD-AT Standard Free AT Free AT + HE Figure 2: Left. Accuracy (%) under PGD-20, where the models are trained by PGD-AT with different iteration steps on CIFAR-10. Right. Visualization of the adversarial perturbations on Image Net. 4 Experiments CIFAR-10 [31] setup. We apply the wide residual network WRN-34-10 as the model architecture [77]. For each AT framework, we set the maximal perturbation ϵ = 8/255, the perturbation step size η = 2/255, and the number of iterations K = 10. We apply the momentum SGD [49] optimizer with the initial learning rate of 0.1, and train for 100 epochs. The learning rate decays with a factor of 0.1 at 75 and 90 epochs, respectively. The mini-batch size is 128. Besides, we set the regularization parameter 1/λ as 6 for TRADES, and set the adversarial logit pairing weight as 0.5 for ALP [29, 81]. The scale s = 15 and the margin m = 0.2 in HE, where different s and m correspond to different trade-offs between the accuracy and robustness, as detailed in Appendix C.3. Image Net [15] setup. We apply the framework of free adversarial training (Free AT) in Shafahi et al. [57], which has a similar training objective as PGD-AT and can train a robust model using four GPU workers. We set the repeat times m = 4 in Free AT. The perturbation ϵ = 4/255 with the step size η = 1/255. We train the model for 90 epochs with the initial learning rate of 0.1, and the mini-batch size is 256. The scale s = 10 and the margin m = 0.2 in HE.2 4.1 Performance under white-box attacks On the CIFAR-10 dataset, we test the defenses under different attacks including FGSM [22], PGD [40], MIM [16], Deepfool [42], C&W (l version) [7], feature attack [35], and FAB [13]. We report the classification accuracy in Table 2 following the evaluation settings in Zhang et al. [81]. We denote the iteration steps behind the attacking method, e.g., 10-step PGD as PGD-10. To verify that our strategy is generally compatible with previous work on accelerating AT, we combine HE with the one-step based Free AT and fast adversarial training (Fast AT) frameworks [70]. We provide the accuracy and training time results in Table 3. We can see that the operations in HE increase negligible computation, even in the cases pursuing extremely fast training. Besides, we also evaluate embedding m-HE (introduced in Sec. 3.5) and find it more effective than HE when combining with PGD-AT, Free AT and Fast AT that exclusively train on adversarial examples. On the Image Net dataset, we follow the evaluation settings in Shafahi et al. [57] to test under PGD-10 and PGD-50, as shown in Table 4. 2Code is available at https://github.com/Shawn XYang/AT_HE. Table 6: Classification accuracy (%) on the clean test data, and under two benchmark attacks Ray S and Auto Attack. Method Architecture Clean Ray S AA PGD-AT+HE WRN-34-10 86.25 57.8 53.16 WRN-34-20 85.14 59.0 53.74 Table 7: Attacking standardly trained WRN-34-10 with or without FN. Attack FN Acc. (%) PGD-1 67.09 62.89 PGD-2 50.37 33.75 Table 8: Classification accuracy (%) under different black-box query-based attacks on CIFAR-10. Method Iterations PGD-AT PGD-AT + HE ALP ALP + HE TRADES TRADES + HE ZOO - 73.47 74.65 72.70 74.90 71.92 74.65 20 73.64 73.66 73.11 74.45 72.12 72.36 50 68.93 69.31 68.39 68.28 68.20 68.42 80 65.67 65.97 65.14 63.89 65.32 65.62 20 74.68 74.87 74.41 75.91 73.40 73.47 50 71.22 71.48 70.81 71.11 70.29 70.53 80 69.16 69.92 68.88 68.53 68.83 69.06 Ablation studies. To investigate the individual effects caused by the three components FN, WN, and AM in HE, we perform the ablation studies for PGD-AT on CIFAR-10, and attack the trained models with PGD-20. We get the clean and robust accuracy of 86.43% / 54.46% for PGD-AT+FN, and 87.28% / 53.93% for PGD-AT+WN. In contrast, when we apply both the FN and WN, we can get the result of 86.20% / 58.95%. For TRADES, applying appropriate AM can increase 2% robust accuracy compared to only applying FN and WN. Detailed results are included in the Appendix C.3. Adaptive attacks. A generic PGD attack apply the cross-entropy loss on the model prediction as the adversarial objective, as reported in Table 2. To exclude the potential effect of gradient obfuscation [2], we construct an adaptive version of the PGD attack against our methods, which uses the training loss in Eq. (6) with the scalar s and margin m as the adversarial objective. In this case, when we apply the adaptive PGD-20 / PGD-500 attack, we will get a accuracy of 55.25% / 52.54% for PGD-AT+HE, which is still higher than the accuracy of 53.97% / 51.63% for PGD-AT (quote from Table 2). Benchmark attacks. We evaluate our enhanced models under two stronger benchmark attacks including Ray S [11] and Auto Attack [14] on CIFAR-10. We train WRN models via PGD-AT+HE, with weight decay of 5 10 4 [47]. For Ray S, we evaluate on 1, 000 test samples due to the high computation. The results are shown in Table 6, where the trained WRN-34-20 model achieves the state-of-the-art performance (no additional data) according to the reported benchmarks. 4.2 Performance under black-box attacks Query-based Black-box Attacks. ZOO [12] proposes to estimate the gradient at each coordinate ei as ˆgi, with a small finite-difference σ. In experiments, we randomly select one sampled coordinate to perform one update with ˆgi, and adopt the C&W optimization mechanism based on the estimated gradient. We set σ as 10 4 and max queries as 20, 000. SPSA [65] and NES [27] can make a full gradient evaluation by drawing random samples and obtaining the corresponding loss values. NES randomly samples from a Gaussian distribution to acquire the direction vectors while SPSA samples from a Rademacher distribution. In experiments, we set the number of random samples q as 128 for every iteration and σ = 0.001. We show the robustness of different iterations against untargeted score-based ZOO, SPSA, and NES in Table 8, where details on these attacks are in Appendix B.2. We include detailed experiments on the transfer-based black-box attacks in Appendix C.4. As expected, these results show that embedding HE can generally provide promotion under the black-box threat models [9], including the transfer-based and the query-based black-box attacks. 4.3 Performance under general-purpose attacks It has been shown that the adversarially trained models could be vulnerable to rotations [19], image corruptions [20] or affine attacks [62]. Therefore, we evaluate on the benchmarks with distributional shifts: CIFAR-10-C and Image Net-C [24]. As shown in Table 5, we report the classification accuracy B1-L3-conv1 B2-L0-conv2 B3-L4-conv1 B3-L4-conv2 B3-L3-conv1 B2-L1-conv1 B1-L4-conv1 B1-L2-conv2 B1-L1-conv1 B1-L2-conv1 B2-L4-conv2 B3-L3-conv2 B2-L3-conv2 conv1 B2-L2-conv1 B1-L1-conv2 B1-L4-conv2 B2-L2-conv2 B3-L1-conv2 B2-L4-conv1 B3-L0-conv2 B3-L0-conv1 B3-L2-conv1 B1-L3-conv2 B1-L0-conv1 bn1 B3-L2-conv2 B2-L1-conv2 B2-L0-conv1 B2-L3-conv1 B3-L1-conv1 B1-L0-conv2 Gradient Norm Ratio PGD-AT + HE Figure 3: The ratios of E( ω L(x ) / ω L(x) ) w.r.t. different parameters ω, where x and x are the adversarial example and its clean counterpart. Higher ratio values indicate more attention on the adversarial examples. B refers to block, L refers to layer, conv refers to convolution. under each corruption averaged on five levels of severity, where the models are WRN-34-10 trained on CIFAR-10 and Res Net-50 trained on Image Net, respectively. Here we adopt accuracy as the metric to be consistent with other results, while the reported values can easily convert into the corruption error metric [24]. We can find that our methods lead to better robustness under a wide range of corruptions that are not seen in training, which prevents the models from overfitting to certain attacking patterns. 4.4 More empirical analyses As shown in the left panel of Fig. 2, we separately use PGD-1 to PGD-7 to generate adversarial examples in training, then we evaluate the trained models under the PGD-20 attack. Our method requires fewer iterations of PGD to achieve certain robust accuracy, e.g., applying 3-step PGD in PGDAT+HE is more robust than applying 7-step PGD in PGD-AT, which largely reduces the necessary computation [69]. To verify the mechanism in Fig. 1, we attack a standardly trained WRN-34-10 (no FN applied) model, applying PGD-1 and PGD-2 with or without FN in the adversarial objective. We report the accuracy in Table 7. As seen, the attacks are more efficient with FN, which suggest that the perturbations are crafted along more effective directions. Besides, previous studies observe that the adversarial examples against robust models exhibit salient data characteristics [28, 54, 61, 64, 82]. So we visualize the untargeted perturbations on Image Net, as shown in the right panel of Fig. 2. We can observe that the adversarial perturbations produced for our method have sharper profiles and more concrete details, which are better aligned with human perception. Finally in Fig. 3, we calculate the norm ratios of the loss gradient on the adversarial example to it on the clean example. The model is trained for 70 epochs on CIFAR-10 using PGD-AT. The results verify that our method can prompt the training procedure to assign larger gradients on the crafted adversarial examples, which would benefit robust learning. 5 Conclusion In this paper, we propose to embed the HE mechanism into AT, in order to enhance the robustness of the adversarially trained models. We analyze the intriguing benefits induced by the interaction between AT and HE from several aspects. It is worth clarifying that empirically our HE module has varying degrees of adaptability on combining different AT frameworks, depending on the specific training principles. Still, incorporating the HE mechanism is generally conducive to robust learning and compatible with previous strategies, with little extra computation and simple code implementation. Broader Impact When deploying machine learning methods into the practical systems, the adversarial vulnerability can cause a potential security risk, as well as the negative impact on the crisis of confidence by the public. To this end, this inherent defect raises the requirements for reliable, general, and lightweight strategies to enhance the model robustness against malicious, especially adversarial attacks. In this work, we provide a simple and efficient way to boost the robustness of the adversarially trained models, which contributes to the modules of constructing more reliable systems in different tasks. Acknowledgements This work was supported by the National Key Research and Development Program of China (No.2020AAA0104304), NSFC Projects (Nos. 61620106010, 62076147, U19B2034, U1811461), Beijing Academy of Artificial Intelligence (BAAI), Tsinghua-Huawei Joint Research Program, a grant from Tsinghua Institute for Guo Qiang, Tiangong Institute for Intelligent Computing, and the NVIDIA NVAIL Program with GPU/DGX Acceleration. [1] Jean-Baptiste Alayrac, Jonathan Uesato, Po-Sen Huang, Alhussein Fawzi, Robert Stanforth, and Pushmeet Kohli. Are labels required for improving adversarial robustness? In Advances in Neural Information Processing Systems (Neur IPS), pages 12192 12202, 2019. [2] Anish Athalye, Nicholas Carlini, and David Wagner. Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. In International Conference on Machine Learning (ICML), 2018. [3] Yoshua Bengio, Jérôme Louradour, Ronan Collobert, and Jason Weston. Curriculum learning. In International Conference on Machine Learning (ICML), pages 41 48. ACM, 2009. [4] Battista Biggio, Igino Corona, Davide Maiorca, Blaine Nelson, Nedim Šrndi c, Pavel Laskov, Giorgio Giacinto, and Fabio Roli. Evasion attacks against machine learning at test time. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 387 402. Springer, 2013. [5] Stephen Boyd and Lieven Vandenberghe. Convex optimization. Cambridge university press, 2004. [6] Wieland Brendel, Jonas Rauber, Alexey Kurakin, Nicolas Papernot, Behar Veliqi, Sharada P Mohanty, Florian Laurent, Marcel Salathé, Matthias Bethge, Yaodong Yu, et al. Adversarial vision challenge. In The Neur IPS 18 Competition, pages 129 153. Springer, 2020. [7] Nicholas Carlini and David Wagner. Towards evaluating the robustness of neural networks. In IEEE Symposium on Security and Privacy (S&P), 2017. [8] Nicholas Carlini and David Wagner. Adversarial examples are not easily detected: Bypassing ten detection methods. In ACM Workshop on Artificial Intelligence and Security (AISec), 2017. [9] Nicholas Carlini, Anish Athalye, Nicolas Papernot, Wieland Brendel, Jonas Rauber, Dimitris Tsipras, Ian Goodfellow, Aleksander Madry, and Alexey Kurakin. On evaluating adversarial robustness. ar Xiv preprint ar Xiv:1902.06705, 2019. [10] Yair Carmon, Aditi Raghunathan, Ludwig Schmidt, Percy Liang, and John C Duchi. Unlabeled data improves adversarial robustness. In Advances in Neural Information Processing Systems (Neur IPS), 2019. [11] Jinghui Chen and Quanquan Gu. Rays: A ray searching method for hard-label adversarial attack. ar Xiv preprint ar Xiv:2006.12792, 2020. [12] Pin-Yu Chen, Huan Zhang, Yash Sharma, Jinfeng Yi, and Cho-Jui Hsieh. Zoo: Zeroth order optimization based black-box attacks to deep neural networks without training substitute models. In ACM Workshop on Artificial Intelligence and Security (AISec). ACM, 2017. [13] Francesco Croce and Matthias Hein. Minimally distorted adversarial examples with a fast adaptive boundary attack. In International Conference on Machine Learning (ICML), 2020. [14] Francesco Croce and Matthias Hein. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In International Conference on Machine Learning (ICML), 2020. [15] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009. [16] Yinpeng Dong, Fangzhou Liao, Tianyu Pang, Hang Su, Jun Zhu, Xiaolin Hu, and Jianguo Li. Boosting adversarial attacks with momentum. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018. [17] Gamaleldin Elsayed, Dilip Krishnan, Hossein Mobahi, Kevin Regan, and Samy Bengio. Large margin deep networks for classification. In Advances in neural information processing systems (Neur IPS), pages 842 852, 2018. [18] Logan Engstrom, Andrew Ilyas, and Anish Athalye. Evaluating and understanding the robustness of adversarial logit pairing. ar Xiv preprint ar Xiv:1807.10272, 2018. [19] Logan Engstrom, Brandon Tran, Dimitris Tsipras, Ludwig Schmidt, and Aleksander Madry. A rotation and a translation suffice: Fooling cnns with simple transformations. In International Conference on Machine Learning (ICML), 2019. [20] Justin Gilmer, Nicolas Ford, Nicolas Carlini, and Ekin Cubuk. Adversarial examples are a natural consequence of test error in noise. In International Conference on Machine Learning (ICML), 2019. [21] Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Deep Learning. MIT Press, 2016. http://www.deeplearningbook.org. [22] Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining and harnessing adversarial examples. In International Conference on Learning Representations (ICLR), 2015. [23] Sven Gowal, Jonathan Uesato, Chongli Qin, Po-Sen Huang, Timothy Mann, and Pushmeet Kohli. An alternative surrogate loss for pgd-based adversarial testing. ar Xiv preprint ar Xiv:1910.09338, 2019. [24] Dan Hendrycks and Thomas Dietterich. Benchmarking neural network robustness to common corruptions and perturbations. In International Conference on Learning Representations (ICLR), 2019. [25] Dan Hendrycks, Kimin Lee, and Mantas Mazeika. Using pre-training can improve model robustness and uncertainty. In International Conference on Machine Learning (ICML), 2019. [26] Elad Hoffer and Nir Ailon. Deep metric learning using triplet network. In International Workshop on Similarity-Based Pattern Recognition, pages 84 92. Springer, 2015. [27] Andrew Ilyas, Logan Engstrom, Anish Athalye, and Jessy Lin. Black-box adversarial attacks with limited queries and information. In International Conference on Machine Learning (ICML), 2018. [28] Andrew Ilyas, Shibani Santurkar, Dimitris Tsipras, Logan Engstrom, Brandon Anish Athalye, Tran, and Aleksander Madry. Adversarial examples are not bugs, they are features. In Advances in Neural Information Processing Systems (Neur IPS), 2019. [29] Harini Kannan, Alexey Kurakin, and Ian Goodfellow. Adversarial logit pairing. ar Xiv preprint ar Xiv:1803.06373, 2018. [30] Konrad Königsberger. Analysis 2 springer verlag, 2004. [31] Alex Krizhevsky and Geoffrey Hinton. Learning multiple layers of features from tiny images. Technical report, Citeseer, 2009. [32] Alexey Kurakin, Ian Goodfellow, and Samy Bengio. Adversarial examples in the physical world. In The International Conference on Learning Representations (ICLR) Workshops, 2017. [33] Alexey Kurakin, Ian Goodfellow, Samy Bengio, Yinpeng Dong, Fangzhou Liao, Ming Liang, Tianyu Pang, Jun Zhu, Xiaolin Hu, Cihang Xie, et al. Adversarial attacks and defences competition. ar Xiv preprint ar Xiv:1804.00097, 2018. [34] Pengcheng Li, Jinfeng Yi, Bowen Zhou, and Lijun Zhang. Improving the robustness of deep neural networks via adversarial training with triplet loss. In International Joint Conference on Artificial Intelligence (IJCAI), 2019. [35] Daquan Lin. https://github.com/Line290/Feature Attack, 2019. [36] Guanxiong Liu, Issa Khalil, and Abdallah Khreishah. Using single-step adversarial training to defend iterative adversarial examples. ar Xiv preprint ar Xiv:2002.09632, 2020. [37] Weiyang Liu, Yandong Wen, Zhiding Yu, Ming Li, Bhiksha Raj, and Le Song. Sphereface: Deep hypersphere embedding for face recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 212 220, 2017. [38] Weiyang Liu, Yan-Ming Zhang, Xingguo Li, Zhiding Yu, Bo Dai, Tuo Zhao, and Le Song. Deep hyperspherical learning. In Advances in neural information processing systems (Neur IPS), pages 3950 3960, 2017. [39] Weiyang Liu, Zhen Liu, Zhehui Chen, Bo Dai, Tuo Zhao, and Le Song. Deep hyperspherical defense against adversarial perturbations. Submission to ICLR, 2018. [40] Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. In International Conference on Learning Representations (ICLR), 2018. [41] Chengzhi Mao, Ziyuan Zhong, Junfeng Yang, Carl Vondrick, and Baishakhi Ray. Metric learning for adversarial robustness. In Advances in Neural Information Processing Systems (Neur IPS), pages 478 489, 2019. [42] Seyed-Mohsen Moosavi-Dezfooli, Alhussein Fawzi, and Pascal Frossard. Deepfool: a simple and accurate method to fool deep neural networks. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2574 2582, 2016. [43] Tianyu Pang, Chao Du, Yinpeng Dong, and Jun Zhu. Towards robust detection of adversarial examples. In Advances in Neural Information Processing Systems (Neur IPS), pages 4579 4589, 2018. [44] Tianyu Pang, Chao Du, and Jun Zhu. Max-mahalanobis linear discriminant analysis networks. In International Conference on Machine Learning (ICML), 2018. [45] Tianyu Pang, Kun Xu, Chao Du, Ning Chen, and Jun Zhu. Improving adversarial robustness via promoting ensemble diversity. In International Conference on Machine Learning (ICML), 2019. [46] Tianyu Pang, Kun Xu, Yinpeng Dong, Chao Du, Ning Chen, and Jun Zhu. Rethinking softmax cross-entropy loss for adversarial robustness. In International Conference on Learning Representations (ICLR), 2020. [47] Tianyu Pang, Xiao Yang, Yinpeng Dong, Hang Su, and Jun Zhu. Bag of tricks for adversarial training. ar Xiv preprint ar Xiv:2010.00467, 2020. [48] Omkar M Parkhi, Andrea Vedaldi, Andrew Zisserman, et al. Deep face recognition. In The British Machine Vision Conference (BMVC), 2015. [49] Ning Qian. On the momentum term in gradient descent learning algorithms. Neural networks, 12(1):145 151, 1999. [50] Rajeev Ranjan, Carlos D Castillo, and Rama Chellappa. L2-constrained softmax loss for discriminative face verification. ar Xiv preprint ar Xiv:1703.09507, 2017. [51] Leslie Rice, Eric Wong, and J Zico Kolter. Overfitting in adversarially robust deep learning. ar Xiv preprint ar Xiv:2002.11569, 2020. [52] Tim Salimans and Durk P Kingma. Weight normalization: A simple reparameterization to accelerate training of deep neural networks. In Advances in Neural Information Processing Systems (Neur IPS), pages 901 909, 2016. [53] Hadi Salman, Jerry Li, Ilya Razenshteyn, Pengchuan Zhang, Huan Zhang, Sebastien Bubeck, and Greg Yang. Provably robust deep learning via adversarially trained smoothed classifiers. In Advances in Neural Information Processing Systems (Neur IPS), pages 11289 11300, 2019. [54] Shibani Santurkar, Andrew Ilyas, Dimitris Tsipras, Logan Engstrom, Brandon Tran, and Aleksander Madry. Image synthesis with a single (robust) classifier. In Advances in Neural Information Processing Systems (Neur IPS), pages 1260 1271, 2019. [55] Florian Schroff, Dmitry Kalenichenko, and James Philbin. Facenet: A unified embedding for face recognition and clustering. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 815 823, 2015. [56] Leo Schwinn and Björn Eskofier. Fast and stable adversarial training through noise injection. ar Xiv preprint ar Xiv:2002.10097, 2020. [57] Ali Shafahi, Mahyar Najibi, Amin Ghiasi, Zheng Xu, John Dickerson, Christoph Studer, Larry S Davis, Gavin Taylor, and Tom Goldstein. Adversarial training for free! In Advances in Neural Information Processing Systems (Neur IPS), 2019. [58] Carl-Johann Simon-Gabriel, Yann Ollivier, Leon Bottou, Bernhard Schölkopf, and David Lopez-Paz. First-order adversarial vulnerability of neural networks and input dimension. In International Conference on Machine Learning (ICML), 2019. [59] Chuanbiao Song, Kun He, Liwei Wang, and John E Hopcroft. Improving the generalization of adversarial training with domain adaptation. In International Conference on Learning Representations (ICLR), 2019. [60] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. Intriguing properties of neural networks. In International Conference on Learning Representations (ICLR), 2014. [61] Guanhong Tao, Shiqing Ma, Yingqi Liu, and Xiangyu Zhang. Attacks meet interpretability: Attribute-steered detection of adversarial samples. In Advances in Neural Information Processing Systems (Neur IPS), pages 7717 7728, 2018. [62] Florian Tramèr and Dan Boneh. Adversarial training and robustness for multiple perturbations. In Advances in Neural Information Processing Systems (Neur IPS), pages 5858 5868, 2019. [63] Florian Tramèr, Alexey Kurakin, Nicolas Papernot, Dan Boneh, and Patrick Mc Daniel. Ensemble adversarial training: Attacks and defenses. In International Conference on Learning Representations (ICLR), 2018. [64] Dimitris Tsipras, Shibani Santurkar, Logan Engstrom, Alexander Turner, and Aleksander Madry. Robustness may be at odds with accuracy. In International Conference on Learning Representations (ICLR), 2019. [65] Jonathan Uesato, Brendan O Donoghue, Aaron van den Oord, and Pushmeet Kohli. Adversarial risk and the dangers of evaluating against weak attacks. In International Conference on Machine Learning (ICML), 2018. [66] S Vivek B and R Venkatesh Babu. Single-step adversarial training with dropout scheduling. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. [67] Feng Wang, Xiang Xiang, Jian Cheng, and Alan Loddon Yuille. Normface: l 2 hypersphere embedding for face verification. In ACM International Conference on Multimedia (ACM MM), pages 1041 1049. ACM, 2017. [68] Hao Wang, Yitong Wang, Zheng Zhou, Xing Ji, Dihong Gong, Jingchao Zhou, Zhifeng Li, and Wei Liu. Cosface: Large margin cosine loss for deep face recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 5265 5274, 2018. [69] Yisen Wang, Xingjun Ma, James Bailey, Jinfeng Yi, Bowen Zhou, and Quanquan Gu. On the convergence and robustness of adversarial training. In International Conference on Machine Learning (ICML), pages 6586 6595, 2019. [70] Eric Wong, Leslie Rice, and J. Zico Kolter. Fast is better than free: Revisiting adversarial training. In International Conference on Learning Representations (ICLR), 2020. [71] Tong Wu, Liang Tong, and Yevgeniy Vorobeychik. Defending against physically realizable attacks on image classification. In International Conference on Learning Representations (ICLR), 2020. [72] Zhirong Wu, Yuanjun Xiong, Stella X Yu, and Dahua Lin. Unsupervised feature learning via non-parametric instance discrimination. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 3733 3742, 2018. [73] Cihang Xie, Yuxin Wu, Laurens van der Maaten, Alan Yuille, and Kaiming He. Feature denoising for improving adversarial robustness. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019. [74] Xiaogang Xu, Hengshuang Zhao, and Jiaya Jia. Dynamic divide-and-conquer adversarial training for robust semantic segmentation. ar Xiv preprint ar Xiv:2003.06555, 2020. [75] Ziang Yan, Yiwen Guo, and Changshui Zhang. Deep defense: Training dnns with improved adversarial robustness. In Advances in Neural Information Processing Systems (Neur IPS), pages 419 428, 2018. [76] Ziqing Yang, Yiming Cui, Wanxiang Che, Ting Liu, Shijin Wang, and Guoping Hu. Improving machine reading comprehension via adversarial training. ar Xiv preprint ar Xiv:1911.03614, 2019. [77] Sergey Zagoruyko and Nikos Komodakis. Wide residual networks. In The British Machine Vision Conference (BMVC), 2016. [78] Runtian Zhai, Tianle Cai, Di He, Chen Dan, Kun He, John Hopcroft, and Liwei Wang. Adversarially robust generalization just requires more unlabeled data. ar Xiv preprint ar Xiv:1906.00555, 2019. [79] Dinghuai Zhang, Tianyuan Zhang, Yiping Lu, Zhanxing Zhu, and Bin Dong. You only propagate once: Accelerating adversarial training via maximal principle. In Advances in Neural Information Processing Systems (Neur IPS), 2019. [80] Haichao Zhang and Jianyu Wang. Defense against adversarial attacks using feature scatteringbased adversarial training. In Advances in Neural Information Processing Systems (Neur IPS), pages 1829 1839, 2019. [81] Hongyang Zhang, Yaodong Yu, Jiantao Jiao, Eric P Xing, Laurent El Ghaoui, and Michael I Jordan. Theoretically principled trade-off between robustness and accuracy. In International Conference on Machine Learning (ICML), 2019. [82] Tianyuan Zhang and Zhanxing Zhu. Interpreting adversarially trained convolutional neural networks. In International Conference on Machine Learning (ICML), 2019. [83] Chen Zhu, Yu Cheng, Zhe Gan, Siqi Sun, Thomas Goldstein, and Jingjing Liu. Freelb: Enhanced adversarial training for language understanding. In International Conference on Learning Representations (ICLR), 2020.