# removing_batch_normalization_boosts_adversarial_training__128dfbf8.pdf Removing Batch Normalization Boosts Adversarial Training Haotao Wang 1 Aston Zhang 2 Shuai Zheng 2 Xingjian Shi 2 Mu Li 2 Zhangyang Wang 1 Adversarial training (AT) defends deep neural networks against adversarial attacks. One challenge that limits its practical application is the performance degradation on clean samples. A major bottleneck identified by previous works is the widely used batch normalization (BN), which struggles to model the different statistics of clean and adversarial training samples in AT. Although the dominant approach is to extend BN to capture this mixture of distribution, we propose to completely eliminate this bottleneck by removing all BN layers in AT. Our normalizer-free robust training (No Frost) method extends recent advances in normalizerfree networks to AT for its unexplored advantage on handling the mixture distribution challenge. We show that No Frost achieves adversarial robustness with only a minor sacrifice on clean sample accuracy. On Image Net with Res Net50, No Frost achieves 74.06% clean accuracy, which drops merely 2.00% from standard training. In contrast, BN-based AT obtains 59.28% clean accuracy, suffering a significant 16.78% drop from standard training. In addition, No Frost achieves a 23.56% adversarial robustness against PGD attack, which improves the 13.57% robustness in BN-based AT. We observe better model smoothness and larger decision margins from No Frost, which make the models less sensitive to input perturbations and thus more robust. Moreover, when incorporating more data augmentations into No Frost, it achieves comprehensive robustness against multiple distribution shifts. Code and pretrained models are public1. Work done during the first author s internship at Amazon Web Services. 1University of Texas at Austin, Austin, USA 2Amazon Web Services, Santa Clara, USA. Correspondence to: Haotao Wang , Aston Zhang , Zhangyang Wang . Proceedings of the 39 th International Conference on Machine Learning, Baltimore, Maryland, USA, PMLR 162, 2022. Copyright 2022 by the author(s). 1https://github.com/amazon-research/norma lizer-free-robust-training 1. Introduction Deep neural networks (DNNs) are vulnerable to adversarial attacks (Szegedy et al., 2013), which generate adversarial images by adding slight manipulations on the original images to falsify model predictions. One of the most effective methods to defend adversarial attacks is adversarial training (AT) (Goodfellow et al., 2015; Zhang et al., 2019). It jointly fits a model on clean (original) images and adversarial images to improve the model s adversarial robustness (i.e., the accuracy on adversarial images). An improved adversarial robustness often comes at the cost of reducing accuracy on clean samples (Madry et al., 2018; Tsipras et al., 2019; Ilyas et al., 2019). However, for many real-world applications, high clean accuracy is a basic requirement while adversarial robustness is a favorable bonus. It is hence desirable to sustain high clean accuracy while achieving high adversarial robustness. Previous works (Xie & Yuille, 2020; Xie et al., 2020) pointed out that the widely used batch normalization (BN) layer (Ioffe & Szegedy, 2015) contributes to the undesirable trade-off between clean accuracy and adversarial robustness. In AT, clean and adversarial images are drawn from two different distributions. It is challenging for BN to capture those two different normalization statistics. Xie & Yuille (2020) proposed mixture BN (MBN) strategy for AT, which routes clean and adversarial images through two separate BN paths. MBN is adopted as the default option for many follow-up works (Xie & Yuille, 2020; Xie et al., 2020; Merchant et al., 2020; Li et al., 2020; Wang et al., 2020b; 2021). But as pointed out by the authors, it faces practical limitations: In practice there is no oracle to tell which BN path to choose during inference for each test sample. We explore an alternative solution. Since BN has limited capacity to estimate normalization statistics of samples from heterogeneous distributions, can we better handle this mixture distribution challenge by removing all BN layers in AT? Replacing BN with other normalization layers that do not calculate statistics across samples, such as instance normalization (IN) (Ulyanov et al., 2016), brings a small amount of benefit, but the results are still unsatisfying as shown in our experiments (Table 3). Instead, we focus on recently proposed normalizer-free networks (Brock et al., 2021a;b). Although these networks were proposed to match Removing Batch Normalization Boosts Adversarial Training state-of-the-art accuracy on standard training with improved hardware performance and memory efficiency, we leverage them for their unexplored benefit to handle data from mixture distributions. To this end, we propose the normalizer-free robust training (No Frost) method to improve the trade-off between clean accuracy and adversarial robustness for AT. No Frost is based on NF-Res Net (Brock et al., 2021a), a Res Net (He et al., 2016) variant without normalization layers but achieving an comparable accuracy on Image Net (Deng et al., 2009). Experimental results show that No Frost achieves a better accuracy-robustness trade-off compared with previous stateof-the-art AT methods based on BN or MBN models. For example, on Image Net with the Res Net50 backbone, No Frost achieves 11.96% higher robustness against APGD-CE attack (Croce & Hein, 2020) and 0.42% higher clean accuracy simultaneously compared with MBNAT (Xie & Yuille, 2020). No Frost also achieves 12.15% higher clean accuracy and 7.25% higher robustness against APGD-CE attack simultaneously than TRADES-FAT (Zhang et al., 2020). To explain the effectiveness of No Frost, we demonstrate that No Frost has better model smoothness (Zhang et al., 2019) and larger decision margins (Yang et al., 2020). Moreover, we show that No Frost can be generalized towards comprehensive robustness against distribution shifts beyond adversarial samples. In particular, we jointly train models on images generated by adversarial attacks and two other robust data augmentation methods, namely Deep Augment (Hendrycks et al., 2021) and texture-debiased augmentation (TDA) (Hermann et al., 2020). This extended version of No Frost simultaneously achieves better or comparable adversarial robustness and accuracy on multiple out-of-distribution (OOD) benchmark datasets (i.e., OOD robustness), such as Image Net-C (Hendrycks & Dietterich, 2019), Imagenet-R (Hendrycks et al., 2021), and Image Net Sketch (Wang et al., 2019), compared with previous stateof-the-art robust learning methods including Deep Augment and texture-debiased augmentation. In summary, our contributions are as follows: 1. We propose No Frost to improve the clean accuracy and adversarial robustness trade-off in AT. Our approach is simple and straightforward: just removing all BN layers to address the mixture distribution challenge. 2. To the best of our knowledge, we for the first time apply normalizer-free models to AT. We demonstrate the unexplored advantage of normalizer-free models in handling data from mixture distributions. 3. We show that No Frost achieves substantially better accuracy-robustness trade-off on Image Net. Using NFRes Net50, No Frost achieves 74.06% clean accuracy, dropping merely 2.00% from standard training, plus 23.56% adversarial robustness against PGD attack. In comparison, BN-based AT obtains only 59.28% clean accuracy with 13.57% adversarial robustness using Res Net50. 4. We demonstrate that when combining adversarial samples with other data augmentation methods, No Frost can simultaneously achieve adversarial robustness and out-of-distribution robustness. 2. Preliminary Before diving deep into the hidden benefit on model robustness brought by normalizer-free training, we need to describe the concept of model robustness that includes both adversarial robustness and OOD robustness (Section 2.1). Then we will revisit the mixture distribution challenge (Section 2.2) and normalizer-free networks (Section 2.3). 2.1. Model Robustness Model robustness refers to a model s performance under various data distribution shifts. Here we review two distribution shifts, adversarial examples and out-of-distribution examples, which are related to our work, and methods for improving model robustness. Adversarial robustness refers to a model s performance on adversarial samples. These adversarial samples are modified from the original (clean) samples by adversarial attacks (Szegedy et al., 2013; Madry et al., 2018; Carlini & Wagner, 2017; Xiao et al., 2018; Croce & Hein, 2020) to falsify the model. To defend these attacks, one of the most effective defense methods is adversarial training (AT) (Goodfellow et al., 2015; Madry et al., 2018; Zhang et al., 2019; 2020; Xie & Yuille, 2020). AT trains a model on both clean and adversarial images. Given a pair of clean image x and its label y sampled from the data distribution D, AT learns a robust classifier fθ with parameters θ by min θ E (x,y) D (1 λ)L(fθ(x), y) + λL(fθ(x ), y), (1) where L( , ) is the cross-entropy loss function. The adversarial image x in Equation 1 is generated from x by PGD attack (Madry et al., 2018): x(0) = Random Sample(B(x, ϵ)), x(t+1) = ΠB(x,ϵ)(x(t) + α sign( x(t)L(fθ(x(t)), y))), where B(x, ϵ) is the ℓ ball with radius ϵ around x, the initialization x(0) is randomly sampled from B(x, ϵ), ΠB(x,ϵ) means the nearest projection to B(x, ϵ), T is the total number of iterations, and α is the step size. The hyper-parameter Removing Batch Normalization Boosts Adversarial Training λ controls the weight of the loss on adversarial samples. When λ = 0 we obtain standard training. When λ = 1 we obtain PGDAT (Madry et al., 2018). Some previous works set λ = 0.5 (Kurakin et al., 2018; Wang et al., 2020b) to trade off clean accuracy and adversarial robustness, which we denote as standard adversarial training (SAT). Other works improve the above training approach. For example, TRADES (Zhang et al., 2019) simultaneously optimizes classification error and model smoothness. FAT (Zhang et al., 2020) uses early-stopped PGD attack to generate friendly adversarial samples, which can also be combined with TRADES (termed as TRADES-FAT). Out-of-distribution robustness refers to the model s performance on out-of-distribution (OOD) examples. To evaluate OOD robustness, there exists multiple benchmark datasets. Image Net-C (Hendrycks & Dietterich, 2019) adds natural corruptions such as Gaussian noise and motion blur on the Image Net validation set. Image Net-Sketch (Wang et al., 2019) contains sketch-like images to evaluate cross-domain transferability. Image Net-A (Hendrycks et al., 2019) contains naturally occurring real-world images which falsify state-of-the-art image classifiers. Image Net-R (Hendrycks et al., 2021) contains renditions such as painting and sculpture. To improve OOD robustness, a popular approach is through data augmentations (Zhong et al., 2017; Cubuk et al., 2018; Geirhos et al., 2019; Yun et al., 2019; Wang et al., 2019; Hendrycks et al., 2021; Gong et al., 2021; Wang et al., 2021). For example, Deep Augment (Hendrycks et al., 2021) first adds random noise onto the weights of an imageto-image model (e.g., an image super-resolution model). It then feeds clean images to the noisy image-to-image model and uses the output images as augmented data. In this way, Deep Augment obtains diverse augmented images and thus achieves state-of-the-art robustness on Image Net-C and Image Net-R. Texture-debiased augmentation (TDA) (Hermann et al., 2020) stacks color distortion, less aggressive random crops, and other simple augmentations to debias a model towards textures and is shown to improve model generalizability (e.g., from Image Net to Image Net-Sketch). 2.2. The Mixture Distribution Challenge Xie & Yuille (2020) have shown that solving Equation 1 with traditional BN-based networks leads to unsatisfying tradeoff between clean accuracy and adversarial robustness. The underlying reason is that clean and adversarial images are sampled from different distributions. It is difficult for BN to estimate the correct normalization statistics of such mixture of distributions. We show the misalignment between clean and adversarial distributions in Figure 1.2 Specifically, we train two Res Net26 models on Image Net by only using clean 2Similar observations were first made in (Xie & Yuille, 2020). We show them here for a more self-contained introduction. Figure 1. The mixture distribution challenge in adversarial training. Specifically, we show the channel-wise BN statistics of the 15-th layer in Res Net26 models obtained by standard training and adversarial training, respectively. Each dot represents the running mean and variance of a channel in the BN layer. We can see that clean and adversarial training samples have different feature statistics, and thus are sampled from different underlying distributions. Similar observations were first made in (Xie & Yuille, 2020). images (i.e., λ = 0 in Equation 1) and adversarial images (i.e., λ = 1), respectively. We then plot the channel-wise BN statistics of the 15-th layer (other layers are similar) in both models. As we can see, the running means and variances of clean (the green dots) and adversarial images (the red dots) are significantly different. To solve this mixture distribution challenge, Xie & Yuille (2020) proposed MBNAT. It uses a mixture BN (MBN) strategy to disentangle clean and adversarial statistics. Specifically, it uses two parallel BNs in each normalization layer, denoted by BNc and BNa. During training, clean (adversarial) images are routed to BNc (BNa). As a result, BNc (BNa) only estimates the distribution of clean (adversarial) images to avoid modeling a mixture distribution. During inference, we should use BNc (BNa) for a clean (adversarial) test image. There is, however, no oracle to tell us whether a test image is clean or adversarial. As shown by Xie & Yuille (2020); Xie et al. (2020), it is difficult to choose which path in practise: If BNc is used, the model will have a good clean accuracy while sacrificing robustness, and vice versa (see Appendix A for details). 2.3. Normalizer-Free Networks Batch normalization (BN) is originally proposed as a regularization to enable stable training of DNNs (Ioffe & Szegedy, 2015), and is then adopted as a basic building block of DNNs. Research on normalizer-free networks aims to remove BN from DNNs for better hardware efficiency. The first attempt to train normalizer-free (NF) deep residual networks uses stable weight initialization methods (Zhang et al., Removing Batch Normalization Boosts Adversarial Training 2018; De & Smith, 2020; Bachlechner et al., 2020). For example, Skip Init initializes residual blocks in NF networks close to identity mappings, ensuring signal propagation and well-behaved gradients (De & Smith, 2020). Although these initialization methods enable stable training of NF deep residual networks, the obtained test accuracy is still lower than that of well-tuned normalized models. More recently, Brock et al. (2021a) first obtained NF networks with performance competitive with traditional BN-based Res Nets (He et al., 2016) and Efficient Nets (Tan & Le, 2019). The authors proposed scaled weight standardization which normalizes the weights in each layer to prevent mean shift in hidden activations and thus stabled the training. Given the weight matrix W of a convolutional or fully connected layer, the proposed scaled weight standardization takes the following form: ˆ Wi,j = γ Wi,j µi where µi and σi are the mean and standard deviation of the i-th row of W , γ is a fixed number, and N is the batch size. This constraint is imposed throughout training as a differentiable operation in the forward pass. Brock et al. (2021b) further proposed an adaptive gradient clipping method, which enables normalizer-free models to train with large batch sizes and strong data augmentations for better test accuracy. 3.1. No Frost for Adversarial Training We adopt a simple strategy to address the mixture distribution challenge in AT. Since this challenge arises from the limited capability of BN to simultaneously encode the heterogeneous distributions of clean and adversarial samples, we simply remove all BN layers in the model. Specifically, our normalizer-free robust training (No Frost) method, solves the AT problem in Equation 1 by using NF networks (Brock et al., 2021a) as the model fθ. We use PGD attack (Madry et al., 2018) to generate adversarial samples in No Frost, and set λ = 0.5 for simplicity3. Without BN layers, No Frost naturally overcomes the limitation of inference-time oracle BN selection in MBNAT described in Section 2.2. 3.2. No Frost for Comprehensive Robustness To achieve the more challenging comprehensive robustness (i.e., to be simultaneously robust against multiple adversarial attacks and naturally occurring distribution shifts), we further generalize No Frost by combining it with other robust data augmentation methods. Different robust data augmentations have different strength and weakness: the best method 3Searching for the optimal λ value may lead to better performance, which we leave for future work. against one type of distribution shift may not be the best against another. For example, AT uses adversarial attack as a data augmentation method that augments clean images by adding worst-case additive noises, and achieves state-ofthe-art robustness against adversarial attacks; however, AT cannot achieve state-of-the-art robustness against natural distribution shifts (e.g., in Image Net-R and Image Net-Sketch). Similarly, Deep Augment is the state-of-the-art data augmentation method against distribution shifts caused by different renditions (e.g., art, cartoons, and graffiti in Image Net-R), but has little benefit against adversarial attacks. As a result, a naive way to achieve comprehensive robustness against multiple types of distribution shifts is to combine multiple robust data augmentations during training. However, jointly training on multiple different data augmentations may cause an even harder mixture distribution challenge than adversarial training that uses only one augmentation (i.e., adversarial attack). In view of this, we extend No Frost by incorporating more data augmentation methods to achieve comprehensive robustness. By default, we add two new robust data augmentation methods, namely Deep Augment and texture-debiased augmentation (TDA), for the extended version of No Frost (denoted as No Frost ). Formally, the optimization problem of No Frost is: min θ E (x,y) D (L(fθ(x), y)+ L(fθ(ˆx), y) + L(fθ(x ), y))/3, (2) where ˆx is the augmented image generated from x using either Deep Augment or TDA (each with half probability), x is the adversarial image generated from x, and fθ is a normalizer-free network. Other notations have the same meaning as in Equation 1. 4. Experiments In this section, we first describe the general experimental settings (Section 4.1). Then we show that No Frost outperforms previous adversarial training methods (Section 4.2) and achieves nicer properties such as better model smoothness and larger decision margins (Section 4.3). Finally, we evaluate No Frost for comprehensive robustness against multiple distribution shifts (Section 4.4). 4.1. Experimental Settings Datasets, models, and metrics All methods are trained on the Image Net (Deng et al., 2009) dataset. We use Res Net26 and Res Net50 (He et al., 2016) backbones, with different normalization strategies: BN (Ioffe & Szegedy, 2015), MBN (Xie & Yuille, 2020), and NF (Brock et al., 2021a). We evaluate clean accuracy on the Image Net validation set, and use the accuracy on adversarial test images as a metric for adver- Removing Batch Normalization Boosts Adversarial Training (a) Res Net26 (b) Res Net50 Figure 2. Trade-off between robustness and accuracy on Res Net26 (sub-figure (a)) and Res Net50 (sub-figure (b)) trained by MBNAT and No Frost. Adversarial robustness is evaluated on PGD (the top-left panel in each sub-figure), APGD-CE (the top-right panel in each sub-figure), Auto Attack (AA, the bottom-left panel in each sub-figure) and targeted PGD (denoted as PGDT; the bottom-right panel in each sub-figure) attacks. γ is the weight for interpolation between the two BNs in MBNAT. sarial robustness. We generate adversarial images on the Image Net validation set using white-box attacks (PGD (Madry et al., 2018), APGD-CE (Croce & Hein, 2020), APGD-DLR (Croce & Hein, 2020), MIA (Dong et al., 2018), CW (Carlini & Wagner, 2017)4), black-box attacks (Ray S (Chen & Gu, 2020), and Square (Andriushchenko et al., 2020)) and the Auto Attack (AA) (Croce & Hein, 2020). We evaluate OOD robustness against naturally occurring distribution shifts by measuring accuracy on Image Net-C (Hendrycks & Dietterich, 2019), Image Net-R (Hendrycks et al., 2021), and Image Net-Sketch (Wang et al., 2019). General hyper-parameters For all experiments, we train on Image Net for 90 epochs. We use the SGD optimizer with momentum 0.9. Batch size is 256. Weight decay factor is 5 10 5. The initial learning rate is 0.1 and decays following a cosine annealing scheduler. All experiments are conducted with 8 NVIDIA V100 GPUs. Implementation of adversarial training We study adversarial robustness under perturbation magnitude ϵ = 8 (on scale of 0 to 255 in unsigned 8-bit pixels). We set the maximum PGD attack iteration number T = 10 during training for all adversarial training methods. For TRADES and TRADES-FAT, we set the loss trade-off hyper-parameter 4We use the ℓ version of CW attack following (Zhang et al., 2020). to 1 following the original TRADES paper (Zhang et al., 2019). For FAT and TRADES-FAT, we set the PGD earlystop iteration to 1 following the original FAT paper (Zhang et al., 2020). Details for robustness evaluation Following (Zhang et al., 2019; 2020), we set attack iterations to be 20 for all white-box attacks. For all black-box attacks, we allow 400 queries per-sample on all compared models. For Auto Attack, we use the fast version with APGD-CE and APGD-DLR attacks. For Ray S, we evaluate on a subset of Image Net validation set with 1000 images due to high computational costs. 4.2. Adversarial Training Results We first compare No Frost with MBNAT, the de facto solution for resolving the AT mixture distribution challenge (Xie & Yuille, 2020; Xie et al., 2020; Merchant et al., 2020; Li et al., 2020; Wang et al., 2020b; 2021; 2022). As discussed in Section 2.2, MBN requires an empirical weighting value γ to be set for interpolation between BNc and BNa during inference (see Appendix A for more details). The original MBNAT paper (Xie & Yuille, 2020) uses γ = 1 to pursue the best adversarial robustness. In another work by the same first author (Xie et al., 2020), γ = 0 is applied for the best clean accuracy. We uniformly sample γ from interval Removing Batch Normalization Boosts Adversarial Training Table 1. Adversarial robustness of Res Net26 under perturbation magnitude ϵ = 8. Classification accuracy on clean images and under different adversarial attacks are reported. The best and second to the best numbers are shown in bold and underlined, respectively. Method Clean White-box Attacks Black-box Attacks AA PGD APGD-CE APGD-DLR MIA CW Ray S Square ST 72.68 0.01 0.00 0.00 0.00 0.00 18.2 27.5 0.00 SAT 52.65 10.55 5.02 5.30 8.84 9.18 30.5 44.7 3.78 TRADES 39.64 9.94 6.24 4.02 8.33 6.37 20.7 32.8 3.54 FAT 58.72 6.97 2.35 2.68 6.59 6.37 33.6 50.8 1.70 TRADES-FAT 55.65 11.91 5.79 6.14 10.83 10.81 31.1 46.7 4.63 No Frost 70.13 12.24 6.34 6.60 21.83 10.18 34.5 48.3 5.04 Table 2. Adversarial robustness of Res Net50 under perturbation magnitude ϵ = 8. Classification accuracy on clean images and under different adversarial attacks are reported. The best and second to the best numbers are shown in bold and underlined, respectively. Method Clean White-box attacks Black-box attacks AA PGD APGD-CE APGD-DLR MIA CW Ray S Square ST 76.06 0.04 0.00 0.00 0.00 0.00 22.5 31.4 0.00 SAT 59.28 13.57 7.80 8.46 10.28 11.02 27.4 40.2 6.23 TRADES 49.25 14.80 9.20 8.19 12.97 11.80 32.6 39.5 6.66 FAT 58.94 12.45 5.48 7.16 12.56 12.24 35.9 51.4 4.73 TRADES-FAT 60.52 11.67 4.71 5.90 11.28 10.29 34.5 48.6 3.87 No Frost 74.06 22.45 11.96 13.37 36.11 19.17 36.1 43.1 9.36 [0, 1] to obtain the robustness-accuracy Parato frontier of MBNAT. Comparison results between No Frost and MBNAT on Res Net26 and Res Net50 are shown in Figure 2. A point closer to the top-right corner represents a more desired model with higher clean accuracy and adversarial robustness. For MBNAT models, as the value of γ increases from 0 to 1, the influence of BNa gradually outweighs that of BNc (see Appendix A for more details). As a result, the adversarial robustness increases while the clean accuracy sharply drops. In contrast, No Frost simultaneously achieves decent clean accuracy and adversarial robustness. In other words, No Frost achieves a much more desired trade-off between clean accuracy and adversarial robustness compared with MBNAT. For example, on Res Net26, No Frost achieves 70.13% accuracy and 6.34% robustness against the APGDCE attack. To achieve comparable accuracy, MBNAT needs to set γ = 0 which leads to 0 robustness against APGDCE attack (6.34% less than No Frost) and 69.71% accuracy (0.42% less than No Frost). On the other hand, to achieve comparable robustness with No Frost, MBNAT needs to set γ = 0.9 which leads to 57.08% accuracy (13.05% less than No Frost) and 6.27% robustness (0.07% less than No Frost). We further compare No Frost with other adversarial training methods, including TRADES, FAT, and TRADES-FAT, on Image Net. The results on Res Net26 and Res Net50 are Table 3. Standard adversarial training (SAT) with IN-based networks yields worse robustness than No Frost. Experiments conducted on Res Net26 with different normalizers. SAT w/ BN 52.65 10.55 SAT w/ IN 56.78 11.06 No Frost 70.13 12.24 shown in Table 1 and 2, respectively. No Frost achieves significantly higher accuracy on clean images and better or comparable robustness against different attacks, compared with all those adversarial training methods. For example, on Res Net26, No Frost outperforms TRADES-FAT by 14.48% on clean accuracy, and 0.55% against APGD-CE attack. Since the mixture of distribution challenge is mainly caused by the limited capability of single BN layers to encode the mixture distribution of clean and adversarial samples, another possible solution is to replace BN with instance-level normalization layers, such as instance normalization (IN) (Ulyanov et al., 2016). We denote the method of replacing BN with IN in SAT as SAT w/ IN . The results are shown in Table 3. Both SAT w/ IN and No Frost achieve better accuracy on clean images and robustness against PGD attack, compared with the naive BN counterpart (i.e., SAT). This is intuitive since both methods are reasonable solutions for Removing Batch Normalization Boosts Adversarial Training Table 4. Mean and standard deviation of No Frost using Res Net26 with three different random seeds. Clean PGD APGD-CE APGD-DLR AA No Frost 70.15 0.03 12.19 0.10 6.34 0.04 6.57 0.06 5.01 0.09 the mixture of distribution problem in adversarial training. However, No Frost achieves considerably better performance than SAT w/ IN, with 13.35% higher accuracy and 1.18% higher robustness against PGD attack. Stability analyses In Table 4, we show the stability analysis results on No Frost over the randomness in the algorithm (e.g., random initialization, random batch sampling). Specifically, we run No Frost on Res Net26 with three different random seeds, and report the mean (denoted as µ) and standard deviation (denoted as σ) of the testing results on those three models in the form of µ σ in Table 4. We report accuracy on both clean and adversarial images generated by different attacks. As we can see, No Frost has stable performance with small standard derivations on both clean accuracy and adversarial robustness. 4.3. No Frost Leads to More Robust Model Properties In this section, we show that No Frost models have stronger model smoothness, larger decision margins (Wang et al., 2020c; Kim et al., 2021), and boundary thickness (Yang et al., 2020). All these have been shown to benefit model robustness (Sanyal et al., 2020; Wang et al., 2020c; Yang et al., 2020). In the following, we provide definitions for these properties and empirically show how they are influenced by removing normalization layers in adversarial training. Decision margin: Following (Kim et al., 2021), we define M(x) = p(x)y maxi =y p(x)i as the decision margin for a sample pair (x, y), where p(x) is the softmax probability of sample x. M(x) < 0 indicates a wrong prediction on sample x. Boundary thickness: Following (Yang et al., 2020), the boundary thickness of date x is defined as T(x) = x x 2 R 1 0 I{α < gij(tx + (1 t)x ) < β}dt, where gij( ) = p( )i p( )j, i and j are the predicted labels of x and x respectively, and I{ } is the indicator function. It measures the distance between two level sets gij( ) = α and gij( ) = β along the adversarial direction. We set α = 0, β = 0.75 and solve x via a targeted 20-step PGD attack following the original paper (Yang et al., 2020). Model smoothness: Following (Zhang et al., 2019; Kim et al., 2021), we use the KL divergence D(x) = KL(p(x) p(x )) as a measurement for model smoothness for sample x, where x is an adversarial image generated from x. A smaller D(x) indicates a stronger local model smoothness at x. Table 5. Decision margin, boundary thickness, and model smoothness of adversarially trained (under ϵ = 8) Res Net26 models with different normalization strategies. The best and second-best values are bolded and underlined, respectively. Normalization strategy (Method) Decision margin M(x)( ) Boundary thickness T(x)( ) Model smoothness D(x)( ) BN (SAT) 0.3241 17.51 4.927 MBN (MBNAT) 0.3143 13.78 1.119 NF (No Frost) 0.4700 31.49 2.996 We measure the above metrics on the 500 validation images from the first 10 classes on Image Net. We report their mean values over the 500 images in Table 5 and also show the distribution of those metrics using histograms in Figure 6 (in Appendix B.1). Compared with SAT, No Frost leads to larger decision margins, thicker boundaries, and stronger model smoothness. All these three properties are beneficial for model robustness (Moosavi-Dezfooli et al., 2019; Sanyal et al., 2020; Wang et al., 2020c; Yang et al., 2020). Another interesting observation is that, compared with SAT, MBNAT improves model smoothness while leaving the decision margin and boundary thickness almost unchanged. In contrast, No Frost improves all three properties over SAT. This is consistent with the recent finding that different defense methods improve robustness through different underlying mechanisms (Kim et al., 2021). Our findings suggest that MBNAT improves model robustness mainly through improving model smoothness, while No Frost improves robustness by simultaneously improving all three properties. 4.4. Comprehensive Robustness Now we evaluate No Frost (subsection 3.2) for comprehensive robustness on three robustness benchmark datasets (Image Net-C, Image Net-R, and Image Net-Sketch), two adversarial attacks (PGD and Square), together with clean accuracy on the Image Net validation set. Since No Frost jointly fits clean, adversarial, Deep Augment, and TDA samples, we compare it with the four stand-alone methods: Standard training (training with only clean images), SAT (training with both clean and adversarial images), Deep Augment, and TDA. We also include the naive combination of the four methods (i.e., jointly training on clean, adversarial, Deep Augment, and TDA samples on a traditional BN network) as a baseline, which is denoted as Combine . All methods are trained using the same settings in Section 4.1. Results are shown in Figure 3 and 4. Notably, the naive combination performs the worst in most cases. This shows the inferent difficulty in fitting multiple heterogeneous augmentations under the traditional BN. In contrast, equipped with the new normalizer-free strategy, Removing Batch Normalization Boosts Adversarial Training Figure 3. Model performance (accuracy in percentage) on different benchmark datasets or adversarial attacks. All methods are trained on Image Net with Res Net26. No Frost successfully fits all data augmentations within a single model and achieves comprehensive robustness. On Res Net26, No Frost achieves the best robustness on all evaluated OOD benchmark datasets and adversarial attacks. On Res Net50, although No Frost achieves slightly worse (3.06% less) robustness on Image Net-C than Deep Augment, it outperforms all baseline methods on other OOD benchmark datasets and adversarial attacks. For example, No Frost achieves 16.30%, 8.20%, and 2.05% higher robustness than Deep Augment on PGD attack, Square attack, and Image Net-R, respectively. 5. Discussions Apart from this work and the MBN papers (Xie & Yuille, 2020; Xie et al., 2020), there are other related works studying how BN affects model robustness. Benz et al. (2021a); Schneider et al. (2020) proposed to improve model robustness against natural image corruptions (e.g., random Gaussian noise and motion blurring) by unsupervised model adaptation. Specifically, they replace the BN statistic calculated on clean training images with those on unlabeled corrupted images. Adv BN (Shu et al., 2021) added adversarial perturbations on the BN statistics to increase model robustness against unseen distribution shifts such as style variations and image corruptions. Galloway et al. (2019) and Benz et al. (2021b) observed that, in standard training, BN grants models with better clean accuracy but harms their adversarial robustness. In contrast to their work, our paper utilizes normalizer-free networks to solve the mixture distribution challenge and improve the trade-off between clean accuracy and adversarial robustness in adversarial training. More related works on machine learning robustness can be found in a recent survey paper (Mohseni et al., 2021). Figure 4. Model performance (accuracy in percentage) on different benchmark datasets or adversarial attacks. All methods are trained on Image Net with Res Net50. Our paper shows that removing BN can significantly boost adversarial training. Yet, some existing test-time adaptation methods utilize the existence of BN to improve model robustness (Wang et al., 2020a; Nandy et al., 2021; Awais et al., 2020; Benz et al., 2021a). Those methods are not directly applicable on normalizer-free networks, and thus No Frost cannot be directly combined with those existing test-time adaptation methods for potentially further improved robustness. It will be our future work to study how to efficiently enable test-time adaption upon No Frost, potentially by designing new test-time adaptation methods tailored for NF networks. On the other hand, No Frost can potentially benefit from the future improvements in both fields of normalizerfree networks and adversarial training. 6. Conclusion In this paper, we address the issue of significant degradation on clean accuracy in adversarial training. The proposed No Frost method removes all BNs in AT. No Frost achieves a significantly more favorable trade-off between clean accuracy and adversarial robustness compared with previous BN-based AT methods: It achieves decent adversarial robustness with only minor degradation on clean accuracy. It is further generalized to achieve the more challenging goal of comprehensive robustness. We hope this study could be a stepping stone towards the exploration of normalizerfree training in improving model robustness and other fields with the challenge of data heterogeneity, such as distributed learning and domain generalization. Acknowledgement Z.W. is supported by the U.S. Army Research Laboratory Cooperative Research Agreement W911NF17-2-0196 Removing Batch Normalization Boosts Adversarial Training (IOBT REIGN) and an Amazon Research Award. Andriushchenko, M., Croce, F., Flammarion, N., and Hein, M. Square attack: a query-efficient black-box adversarial attack via random search. In European Conference on Computer Vision (ECCV), pp. 484 501, 2020. Awais, M., Shamshad, F., and Bae, S.-H. Towards an adversarially robust normalization approach. ar Xiv preprint ar Xiv:2006.11007, 2020. Bachlechner, T., Majumder, B. P., Mao, H. H., Cottrell, G. W., and Mc Auley, J. Re Zero is all you need: Fast convergence at large depth. ar Xiv preprint ar Xiv:2003.04887, 2020. Benz, P., Zhang, C., Karjauv, A., and Kweon, I. S. Revisiting batch normalization for improving corruption robustness. In IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 494 503, 2021a. Benz, P., Zhang, C., and Kweon, I. S. Batch normalization increases adversarial vulnerability and decreases adversarial transferability: A non-robust feature perspective. In IEEE International Conference on Computer Vision (ICCV), pp. 7818 7827, 2021b. Brock, A., De, S., and Smith, S. L. Characterizing signal propagation to close the performance gap in unnormalized Res Nets. In International Conference on Learning Representations (ICLR), 2021a. Brock, A., De, S., Smith, S. L., and Simonyan, K. Highperformance large-scale image recognition without normalization. ar Xiv preprint ar Xiv:2102.06171, 2021b. Carlini, N. and Wagner, D. Towards evaluating the robustness of neural networks. In IEEE Symposium on Security and Privacy (SP), pp. 39 57, 2017. Chen, J. and Gu, Q. Ray S: A ray searching method for hard-label adversarial attack. In International Conference on Knowledge Discovery and Data Mining (KDD), 2020. Croce, F. and Hein, M. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In International Conference on Machine Learning (ICML), pp. 2206 2216, 2020. Cubuk, E. D., Zoph, B., Mane, D., Vasudevan, V., and Le, Q. V. Auto Augment: Learning augmentation policies from data. ar Xiv preprint ar Xiv:1805.09501, 2018. De, S. and Smith, S. Batch normalization biases residual blocks towards the identity function in deep networks. Advances in Neural Information Processing Systems (Neur IPS), 2020. Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. Image Net: A large-scale hierarchical image database. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 248 255, 2009. Dong, Y., Liao, F., Pang, T., Su, H., Zhu, J., Hu, X., and Li, J. Boosting adversarial attacks with momentum. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9185 9193, 2018. Galloway, A., Golubeva, A., Tanay, T., Moussa, M., and Taylor, G. W. Batch normalization is a cause of adversarial vulnerability. ar Xiv preprint ar Xiv:1905.02161, 2019. Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F. A., and Brendel, W. Imagenet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. In International Conference on Learning Representations (ICLR), 2019. Gong, C., Ren, T., Ye, M., and Liu, Q. Max Up: A simple way to improve generalization of neural network training. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. Goodfellow, I. J., Shlens, J., and Szegedy, C. Explaining and harnessing adversarial examples. In International Conference on Learning Representations (ICLR), 2015. He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770 778, 2016. Hendrycks, D. and Dietterich, T. Benchmarking neural network robustness to common corruptions and perturbations. In International Conference on Learning Representations (ICLR), 2019. Hendrycks, D., Zhao, K., Basart, S., Steinhardt, J., and Song, D. Natural adversarial examples. ar Xiv preprint ar Xiv:1907.07174, 2019. Hendrycks, D., Basart, S., Mu, N., Kadavath, S., Wang, F., Dorundo, E., Desai, R., Zhu, T., Parajuli, S., Guo, M., Song, D., Steinhardt, J., and Gilmer, J. The many faces of robustness: A critical analysis of out-of-distribution generalization. In IEEE International Conference on Computer Vision (ICCV), 2021. Hermann, K., Chen, T., and Kornblith, S. The origins and prevalence of texture bias in convolutional neural networks. Advances in Neural Information Processing Systems (Neur IPS), 33, 2020. Ilyas, A., Santurkar, S., Tsipras, D., Engstrom, L., Tran, B., and Madry, A. Adversarial examples are not bugs, Removing Batch Normalization Boosts Adversarial Training they are features. In Advances in Neural Information Processing Systems (Neur IPS), 2019. Ioffe, S. and Szegedy, C. Batch Normalization: Accelerating deep network training by reducing internal covariate shift. In International Conference on Machine Learning (ICML), pp. 448 456, 2015. Kim, H., Lee, W., Lee, S., and Lee, J. Bridged adversarial training. ar Xiv preprint ar Xiv:2108.11135, 2021. Kurakin, A., Goodfellow, I. J., and Bengio, S. Adversarial examples in the physical world. In Artificial Intelligence Safety and Security, pp. 99 112. 2018. Li, Y., Yu, Q., Tan, M., Mei, J., Tang, P., Shen, W., Yuille, A., and Xie, C. Shape-texture debiased neural network training. ar Xiv preprint ar Xiv:2010.05981, 2020. Madry, A., Makelov, A., Schmidt, L., Tsipras, D., and Vladu, A. Towards deep learning models resistant to adversarial attacks. In International Conference on Learning Representations (ICLR), 2018. Merchant, A., Zoph, B., and Cubuk, E. D. Does data augmentation benefit from split batchnorms. ar Xiv preprint ar Xiv:2010.07810, 2020. Mohseni, S., Wang, H., Yu, Z., Xiao, C., Wang, Z., and Yadawa, J. Practical machine learning safety: A survey and primer. ar Xiv preprint ar Xiv:2106.04823, 2021. Moosavi-Dezfooli, S.-M., Fawzi, A., Uesato, J., and Frossard, P. Robustness via curvature regularization, and vice versa. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9078 9086, 2019. Nandy, J., Saha, S., Hsu, W., Mong, L., and Zhu, X. X. Covariate shift adaptation for adversarially robust classifier. In International Conference on Learning Representations Workshop (ICLRW), 2021. Sanyal, A., Dokania, P. K., Kanade, V., and Torr, P. H. How benign is benign overfitting? ar Xiv preprint ar Xiv:2007.04028, 2020. Schneider, S., Rusak, E., Eck, L., Bringmann, O., Brendel, W., and Bethge, M. Improving robustness against common corruptions by covariate shift adaptation. In Advances in Neural Information Processing Systems (Neur IPS), 2020. Shafahi, A., Najibi, M., Ghiasi, A., Xu, Z., Dickerson, J., Studer, C., Davis, L. S., Taylor, G., and Goldstein, T. Adversarial training for free! In Advances in Neural Information Processing Systems (Neur IPS), pp. 3358 3369, 2019. Shu, M., Wu, Z., Goldblum, M., and Goldstein, T. Encoding robustness to image style via adversarial feature perturbations. In Advances in Neural Information Processing Systems (Neur IPS), 2021. Szegedy, C., Zaremba, W., Sutskever, I., Bruna, J., Erhan, D., Goodfellow, I., and Fergus, R. Intriguing properties of neural networks. In International Conference on Learning Representations (ICLR), 2013. Tan, M. and Le, Q. Efficient Net: Rethinking model scaling for convolutional neural networks. In International Conference on Machine Learning (ICML), pp. 6105 6114, 2019. Tsipras, D., Santurkar, S., Engstrom, L., Turner, A., and Madry, A. Robustness may be at odds with accuracy. In International Conference on Learning Representations (ICLR), 2019. Ulyanov, D., Vedaldi, A., and Lempitsky, V. Instance normalization: The missing ingredient for fast stylization. ar Xiv preprint ar Xiv:1607.08022, 2016. Wang, D., Shelhamer, E., Liu, S., Olshausen, B., and Darrell, T. Tent: Fully test-time adaptation by entropy minimization. In International Conference on Learning Representations (ICLR), 2020a. Wang, H., Ge, S., Lipton, Z., and Xing, E. P. Learning robust global representations by penalizing local predictive power. In Advances in Neural Information Processing Systems (Neur IPS), pp. 10506 10518, 2019. Wang, H., Chen, T., Gui, S., Hu, T., Liu, J., and Wang, Z. Once-for-all adversarial training: In-situ tradeoff between robustness and accuracy for free. Advances in Neural Information Processing Systems (Neur IPS), pp. 7449 7461, 2020b. Wang, H., Xiao, C., Kossaifi, J., Yu, Z., Anandkumar, A., and Wang, Z. Augmax: Adversarial composition of random augmentations for robust training. In Advances in Neural Information Processing Systems (Neur IPS), 2021. Wang, H., Zhang, A., Zhu, Y., Zheng, S., Li, M., Smola, A., and Wang, Z. Partial and asymmetric contrastive learning for out-of-distribution detection in long-tailed recognition. In International Conference on Machine Learning (ICML), 2022. Wang, Y., Zou, D., Yi, J., Bailey, J., Ma, X., and Gu, Q. Improving adversarial robustness requires revisiting misclassified examples. In International Conference on Learning Representations (ICLR), 2020c. Wong, E., Rice, L., and Kolter, J. Z. Fast is better than free: Revisiting adversarial training. In International Conference on Learning Representations, 2020. Removing Batch Normalization Boosts Adversarial Training Xiao, C., Zhu, J.-Y., Li, B., He, W., Liu, M., and Song, D. Spatially transformed adversarial examples. ar Xiv preprint ar Xiv:1801.02612, 2018. Xie, C. and Yuille, A. Intriguing properties of adversarial training. In International Conference on Learning Representations (ICLR), 2020. Xie, C., Tan, M., Gong, B., Wang, J., Yuille, A., and Le, Q. V. Adversarial examples improve image recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. Yang, Y., Khanna, R., Yu, Y., Gholami, A., Keutzer, K., Gonzalez, J. E., Ramchandran, K., and Mahoney, M. W. Boundary thickness and robustness in learning models. Advances in Neural Information Processing Systems (Neur IPS), 2020. Yun, S., Han, D., Oh, S. J., Chun, S., Choe, J., and Yoo, Y. Cut Mix: Regularization strategy to train strong classifiers with localizable features. In IEEE International Conference on Computer Vision (ICCV), pp. 6023 6032, 2019. Zhang, H., Dauphin, Y. N., and Ma, T. Fixup initialization: Residual learning without normalization. In International Conference on Learning Representations (ICLR), 2018. Zhang, H., Yu, Y., Jiao, J., Xing, E. P., Ghaoui, L. E., and Jordan, M. I. Theoretically principled trade-off between robustness and accuracy. In International Conference on Machine Learning (ICML), pp. 7472 7482, 2019. Zhang, J., Xu, X., Han, B., Niu, G., Cui, L., Sugiyama, M., and Kankanhalli, M. Attacks which do not kill training make adversarial learning stronger. In International Conference on Machine Learning (ICML), pp. 11278 11287, 2020. Zhong, Z., Zheng, L., Kang, G., Li, S., and Yang, Y. Random erasing data augmentation. ar Xiv preprint ar Xiv:1708.04896, 2017. Removing Batch Normalization Boosts Adversarial Training A. How to Interpolate between Two BN Branches in MBNAT We follow (Merchant et al., 2020) to interpolate between the two BN branches (i.e., the BNc branch and BNa branch) during test time for MBNAT. Specifically, given an input test image x, we first forward it through the BNc branch (i.e., the MBN network using BNc at each normalization layer) to get the output logits zc, and then through the BNa branch (i.e., the MBN network using BNa at each normalization layer) to get the output logits za. We then average zc and za with a weighting hyper-parameter γ, i.e., z = (1 γ)zc + γza. Finally, we use the averaged logits z as the input to the softmax function to get the final prediction probabilities. As a result, when the value of γ increases from 0 to 1, the influence of BNa gradually outweighs that of BNc. When γ is 0 or 1, the MBN model falls back to the simple cases with only one BN (BNc when γ = 0 or BNa when γ = 1) at each normalization layer. This is the default interpolation method we used in our paper, which is denoted as MBNAT (logits) or simply MBNAT when used as the default. Besides the one suggested in (Merchant et al., 2020) (i.e., MBNAT (logits)), we have also investigated other possible interpolation methods between BNc and BNa. For example, we can interpolate the outputs of BNc and BNa at each MBN layer. Specifically, if the input feature of an MBN layer is denoted as fi, then the output feature fo = (1 γ)BNc(fi) + γBNa(fi), where BNc( ) and BNa( ) are the batch normalization operations by BNc and BNa, respectively. We denote this method as MBNAT (all) . We can also conduct this output mixing on some selected MBN layers, while keeping the two parallel outputs in other MBN layers. For example, we can randomly select p% MBN layers for mixing. We denote this method as MBNAT (random p%) . The results in Figure 5 show that MBNAT (logits) achieves the best robustness-accuracy trade-off curve among all compared interpolation strategies, so we use it as our default interpolation strategy for MBNAT. B. More Experimental Results B.1. Histograms of Decision Margin, Boundary Thickness, and Model Smoothness We have numerically compared the average decision margins, boundary thickness and model smoothness of SAT, MBNAT and No Frost in Table 5 (Section 4.3). Here in Figure 6, we visualize the distributions of these metrics on different models using histograms. Figure 6 is simply another way to show the results in Table 5, but gives more detailed information through histogram visualization. Figure 5. Trade-off between robustness and accuracy of different interpolation strategies on MBNAT with Res Net26. B.2. Normalizer-free Networks with Standard Training is Not Robust to Adversarial Attacks Table 6. Normalizer-free networks trained using standard training (only on clean images) is not robust against adversarial attacks. All methods are trained on Image Net with (NF-)Res Net50. Method Clean PGD APGD-CE ST w/ Res Net50 76.06 0.04 0.00 ST w/ NF-Res Net50 75.02 0.00 0.00 No Frost 74.06 22.45 11.96 In this section, we show that normalizer-free networks does not naturally have satisfactory adversarial robustness after standard training (i.e., training only on clean images). Specifically, we compare the results of standard training on Res Net50 (denoted as ST w/ Res Net50), standard training on NF-Res Net50 (denoted as ST w/ NF-Res Net50), and adversarial training on NF-Res Net50 (No Frost) in Table 6. As we can see, standard training on both Res Net50 and NF-Res Net50 have almost zero adversarial robustness. This shows that the robustness of No Frost is not simply the result of more robust network structure, but the combination of the AT algorithm and the AT-friendly normalizer-free network structure. B.3. Robustness under Different Perturbation Magnitudes In the main text, we evaluated adversarial robustness using adversarial attacks with perturbation magnitude ϵ = 8. In this section, we evaluate model robustness under different adversarial perturbation magnitudes. Specifically, we compare the robustness of the models in Table 1 (which are trained with ϵ = 8) on targeted PGD (denoted as PGDT) attack with ϵ ranging from 8 to 16. As shown in Figure 7, the advantage of No Frost holds on multiple different perturbation magnitudes. Removing Batch Normalization Boosts Adversarial Training (a) Decision margin (b) Boundary thickness (c) Model smoothness Figure 6. Histograms of decision margin, boundary thickness, and model smoothness of adversarially trained Res Net26 with different normalization strategies. All metrics are measured on the 500 validation images from the first 10 classes of Image Net. (a) M(x) as a metric for decision margin: the larger the better. (b) T(x) as a metric for boundaries robustness: the larger the better. (c) D(x) as as metric for model smoothness: the smaller the better. Figure 7. Adversarial robustness under different perturbation magnitudes. All methods are trained on Image Net with Res Net26. B.4. Adversarial Training Results with Small Perturbation Magnitudes In the main text, we set perturbation magnitude ϵ = 8 in both adversarial training and evaluation. Some previous works, such as Fast AT (Wong et al., 2020) and Free AT (Shafahi et al., 2019) conducted adversarial training on Image Net using smaller perturbation magnitudes such as ϵ = 2, 4, and also evaluation adversarial robustness using the same ϵ values. In this section, we compare No Frost with Fast AT and Free AT using the small ϵ setting. We use PGD attack with 10 and 50 steps (denoted as PGD-10 and PGD-50 respectively) to evaluate adversarial robustness, following (Shafahi et al., 2019). The results are shown in Table 7. No Frost largely outperforms Fast AT and Free AT under the small ϵ setting. Table 7. Adversarial robustness of Res Net50 under perturbation magnitude ϵ = 2, 4. Classification accuracy on clean images and under different adversarial attacks are reported. The best and second to the best numbers are shown in bold and underlined, respectively. ϵ Method Clean PGD-10 PGD-50 Fast AT 60.90 44.27 44.20 Free AT 64.45 43.52 43.39 No Frost 69.87 48.60 48.23 Fast AT 55.45 32.10 31.67 Free AT 60.21 32.77 31.88 No Frost 66.05 36.14 36.05