# if_your_data_distribution_shifts_use_selflearning__3554f515.pdf Published in Transactions on Machine Learning Research (11/2022) If your data distribution shifts, use self-learning Evgenia Rusak evgenia.rusak@bethgelab.org University of Tübingen Steffen Schneider steffen.schneider@bethgelab.org University of Tübingen George Pachitariu george.pachitariu@tue.ai University of Tübingen Luisa Eck luisa.eck@physics.ox.ac.uk University of Oxford Peter Gehler pgehler@amazon.com Amazon Tübingen Oliver Bringmann oliver.bringmann@uni-tuebingen.de University of Tübingen Wieland Brendel wieland.brendel@tuebingen.mpg.de Max-Planck Institute for Intelligent Systems Tübingen Matthias Bethge matthias.bethge@bethgelab.org University of Tübingen Reviewed on Open Review: https: // openreview. net/ forum? id= XXXX We demonstrate that self-learning techniques like entropy minimization and pseudo-labeling are simple and effective at improving performance of a deployed computer vision model under systematic domain shifts. We conduct a wide range of large-scale experiments and show consistent improvements irrespective of the model architecture, the pre-training technique or the type of distribution shift. At the same time, self-learning is simple to use in practice because it does not require knowledge or access to the original training data or scheme, is robust to hyperparameter choices, is straight-forward to implement and requires only a few adaptation epochs. This makes self-learning techniques highly attractive for any practitioner who applies machine learning algorithms in the real world. We present state-of-the-art adaptation results on CIFAR10-C (8.5% error), Image Net-C (22.0% m CE), Image Net-R (17.4% error) and Image Net-A (14.8% error), theoretically study the dynamics of selfsupervised adaptation methods and propose a new classification dataset (Image Net-D) which is challenging even with adaptation. Equal contribution. Work initiated during an internship at Amazon Tübingen. Joint senior authors. Published in Transactions on Machine Learning Research (11/2022) Figure 1: Robustness and adaptation to new datasets has traditionally been achieved by robust pre-training (with hand-selected/data-driven augmentation strategies, or additional data), unsupervised domain adaptation (with access to unlabeled samples from the test set), or, more recently, self-supervised learning methods. We show that on top of these different pre-training tasks, it is always possible (irrespective of architecture, model size or pre-training algorithm) to further adapt models to the target domain with simple self-learning techniques. 1 Introduction Deep Neural Networks (DNNs) can reach human-level performance in complex cognitive tasks (Brown et al., 2020; He et al., 2016a; Berner et al., 2019) if the distribution of the test data is sufficiently similar to the training data. However, DNNs are known to struggle if the distribution of the test data is shifted relatively to the training data (Geirhos et al., 2018; Dodge & Karam, 2017). Two largely distinct communities aim to increase the performance of models under test-time distribution shifts: The robustness community generally considers Image Net-scale datasets and evaluates models in an ad-hoc scenario. Models are trained on a clean source dataset like Image Net (Deng et al., 2009), using heavy data augmentation (Hendrycks et al., 2020a; Rusak et al., 2020; Geirhos et al., 2019) and/or large-scale pre-training (Xie et al., 2020a; Mahajan et al., 2018). The trained models are not adapted in any way to test-time distribution shifts. This evaluation scenario is relevant for applications in which very different distribution shifts are encountered in an unpredictable order, and hence misses out on the gains of adaptation to unlabeled samples of the target distribution. The unsupervised domain adaptation (UDA) community often considers smaller-scale datasets and assumes that both the source and the (unlabeled) target dataset are known. Models are trained on both datasets, e.g., with an adversarial objective (Ganin et al., 2016; Tzeng et al., 2017; Hoffman et al., 2018), before evaluation on the target domain data. This evaluation scenario provides optimal conditions for adaptation, but the reliance on the source dataset makes UDA more computationally expensive, more impractical and prevents the use of pre-trained models for which the source dataset is unknown or simply too large. We refer the reader to Farahani et al. (2021) for a review of UDA. In this work, we consider the source-free domain adaptation setting, a middle ground between the classical ad-hoc robustness setting and UDA in which models can adapt to the target distribution but without using the source dataset (Kundu et al., 2020; Kim et al., 2021; Li et al., 2020; Liang et al., 2020). This evaluation scenario is interesting for many practitioners and applications as an extension of the ad-hoc robustness scenario. It evaluates the possible performance of a deployed model on a systematic, unseen distribution shift at inference time: an embedded computer vision system in an autonomous car should adapt to changes without being trained on all available training data; an image-based quality control software may not necessarily open-source the images it has been trained on, but still has to be adapted to the lighting conditions at the operation location; a computer vision system in a hospital should perform robustly when tested on a scanner different from the one used for producing the training images importantly, it might not be known at development time which scanner the vision system will be tested on, and it might be prohibited to share images from many hospitals to run UDA. Published in Transactions on Machine Learning Research (11/2022) Can self-learning methods like pseudo-labeling and entropy-minimization also be used in this source-free domain adaptation setting?1 To answer this question, we perform an extensive study of several self-learning variants, and find consistent and substantial gains in test-time performance across several robustness and out-of-domain benchmarks and a wide range of models and pre-training methods, including models trained with UDA methods that do not use self-learning, see Figure 1. We also find that self-learning outperforms state-of-the-art source-free domain adaptation methods, namely Test-Time Training which is based on a self-supervised auxiliary objective and continual training (Sun et al., 2020), test-time entropy minimization (Wang et al., 2021), Meta Test-Time Training (Bartler et al., 2022), and (gradient-free) Batch Norm adaptation (Schneider et al., 2020; Nado et al., 2020), and can be improved with additional techniques such as diversity regularization (Mummadi et al., 2021). We perform a large number of ablations to study important design choices for self-learning methods in source-free domain adaptation. Furthermore, we show that a variant of pseudo-labeling with a robust loss function consistently outperforms entropy minimization on Image Net-scale datasets. We begin by positioning our work in the existing literature ( 2) and proceed with an overview of various self-learning variants that have been applied over the past years, and propose a new technique for robust pseudo-labeling ( 3). We then outline a rigorous experimental protocol that aims to highlight the strengths (and shortcomings) of various self-learning methods. We test various model architectures, with different pre-training schemes covering the most important models in unsupervised domain adaptation, robustness, and large-scale pre-training ( 4). Using this protocol, we show the effectiveness of self-learning across architectures, models and pre-training schemes ( 5). We proceed with an in-depth analysis of self-learning, both empirical ( 6) and theoretical ( 7). Since the outlined results on Image Net-C (22.0% m CE), Image Net-R (17.4% error) and Image Net-A (14.8%) approach clean performance (11.6% error for our baseline), we propose Image Net-D as a new benchmark, which we analyse in 8. We conclude by proposing a set of best practices for evaluating test-time adaptation techniques in the future to ensure scientific rigor and to enable fair model and method comparisons ( 9). 2 Related Work Test-Time Adaptation The main question we ask in this work is whether self-learning methods such as entropy minimization and different variants of pseudo-labeling can improve the performance of models when adapted at test-time to data with a distribution shift relative to the training data. Our work is most similar to test-time entropy minimization (TENT; Wang et al., 2021) since entropy minimization is one of our studied test-time adaptation techniques. The conceptual difference between our experiments and Wang et al. (2021) is that Wang et al. (2021) compare TENT to UDA while we argue that UDA can be regarded as a pretraining step, and self-learning can be used on top of any checkpoint pretrained with UDA. Wang et al. (2021) study the effectiveness of entropy minimization across different models, datasets and tasks. We expand upon their experimental setup and show that entropy minimization is effective across a large range of model architectures (convolutional neural networks, Vision Transformers (Dosovitskiy et al., 2021; Caron et al., 2021)), sizes (Res Net50 to Efficient Net-L2 (Xie et al., 2020a; Tan & Le, 2019)), and are orthogonal to robustification (e.g., Deep Augment (Hendrycks et al., 2020a)) and other pre-training schemes. Finally, we perform a large hyperparameter study for test-time entropy minimization, and thereby further improve upon the results reported by Wang et al. (2021). We also compare our self-learning results to gradient-free adaptation of batch normalization (BN; Ioffe & Szegedy, 2015) statistics (BN adapt; Schneider et al., 2020) who proposed re-estimating the BN statistics on the shifted data distribution. In Test-Time Training (TTT; Sun et al., 2020), the softmax cross-entropy loss is combined with a rotation prediction task (Gidaris et al., 2018) during pretraining on the source dataset. At test-time, the model is fine-tuned on the unlabeled test data with the self-supervised task. Sun et al. (2020) report results when 1Self-learning was defined by Tsypkin (1968) to denote learning when there is no external indication concerning the correctness of the response of the automatic system to the presented patterns . We opted to use this term to name the superset of pseudo-labeling, entropy minimization and self-supervised learning variants to highlight the fact these methods do not require ground truth labels for adaptation. Published in Transactions on Machine Learning Research (11/2022) adapting to a single test example, and also for online adaptation where the model successively adapts to a stream of data, where the data can either come from the same or a gradually changing distribution. We added a detailed comparison to Sun et al. (2020) in Appendix C.6. Liu et al. (2021) (TTT+++) replace the rotation prediction task with Sim CLR (Chen et al., 2020a), and find this modification improves performance. Eastwood et al. (2022) propose Feature Restoration where the approximate feature distribution under the target data is realigned with the feature distribution under source data. Eastwood et al. (2022) report results on CIFAR10 (among other datasets) for a Res Net18 architecture which allows us to include their method as a baseline. In a concurrent publication, Niu et al. (2022) propose an anti-forgetting test-time adaptation method called EATA, which combine sample-efficient entropy minimization with anti-forgetting weight regularization. They report accuracy numbers on Image Net-C for a Res Net50 which we include as a baseline. Mummadi et al. (2021) introduce a novel loss to stabilize entropy minimization by replacing the entropy by a non-saturating surrogate and a diversity regularizer based on batch-wise entropy maximization that prevents convergence to trivial collapsed solutions. To partially undo the distribution shift at test time, they additionally propose to add an input transformation module to the model. Mummadi et al. (2021) report results on the highest severity of Image Net-C for a Res Net50 which we include as a comparison to our results. We note that Mummadi et al. (2021) constitute concurrent unpublished work. Meta Test-Time Training (MT3; Bartler et al., 2022) combine meta-learning, self-supervision and test-time training to adapt a model trained on clean CIFAR10 (Krizhevsky et al., 2009) to CIFAR10-C (Hendrycks & Dietterich, 2019). We added a detailed comparison to Bartler et al. (2022) in Appendix C.7. The following papers also consider the setting of test-time adaptation, but are not used as direct baselines in our work, because they study other datasets or tasks. Azimi et al. (2022) show performance improvements when using test-time adaptation on video data. They show adaptation results for the popular test-time adaptation techniques of BN adaptation (Schneider et al., 2020), Test-Time Training (Sun et al., 2020) and test-time entropy minimization (Wang et al., 2021). MEMO (Zhang et al., 2021) maximizes the prediction consistency of different augmented copies regarding a given test sample. We do not consider the single-sample adaptation setting, and comparing our self-learning techniques to MEMO is not a fair setting for MEMO; unsurprisingly, techniques benefiting from multiple samples such as TENT (Wang et al., 2021) outperform MEMO. Ada Contrast by Chen et al. (2022) combines pseudo-labeling with other techniques, such as self-supervised contrastive learning on the target domain, soft k-nearest neighbors voting to stabilize the pseudo-labels, as well as consistency and diversity regularization. A direct comparison with their results is difficult because they evaluate on VISDA-C and Domain Net, and so we would need to train our methods on either of the datasets and perform a full hyperparameter search for a fair comparison. We do have one point of comparison: in their paper, Chen et al. (2022) perform better than TENT (Wang et al., 2021). However, we note that Chen et al. (2022) used the default hyperparameters of TENT and did not perform hyperparameter tuning on the new dataset, thus, we would expect the performance of properly tuned TENT to be better than reported in the paper. We expect the additional changes of Ada Contrast to further improve upon our simple self-learning baselines. Our work is conceptually similar to virtual adversarial domain adaptation in the fine-tuning phase of DIRT-T (Shu et al., 2018). In contrast to DIRT-T, our objective is simpler and we scale the approach to considerably larger datasets on Image Net scale. Iwasawa & Matsuo (2021) propose a test-time adaptation algorithm for the task of domain generalization based on computing the distance of each test sample and pseudo-prototypes for each class. Kumar et al. (2020) study the setting of self-learning for gradual domain adaptation. They find that self-learning works better if the data distribution changes slowly. The gradual domain adaptation setting differs from ours: instead of a gradual shift over time, we focus on a fixed shift at test time. Self-learning for domain adaptation Xie et al. (2020b) introduce In-N-Out which uses auxiliary information to boost both inand out-of-distribution performance. Ada Match (Berthelot et al., 2021) builds upon Fix Match (Sohn et al., 2020) and can be used for the tasks of unsupervised domain adaptation, semi-supervised learning and semi-supervised domain adaptation as a general-purpose algorithm. Prabhu et al. (2021) propose SENTRY, an algorithm based on judging the predictive consistency of samples from the target domain under different image transformations. Zou et al. (2019) show that different types of confidence regularization can improve the performance of self-learning. A theoretically motivated framework Published in Transactions on Machine Learning Research (11/2022) for self-learning in domain adaptation based on consistency regularization has been proposed by Wei et al. (2020) and then extended by Cai et al. (2021). The main differences from these works to ours are that they 1) utilize both source and target data during training (i.e., the classical UDA setup) whereas we only require access to unlabeled target data (source-free setup), 2) train their models from scratch whereas we adapt pretrained checkpoints to the unlabeled target data, and 3) are oftentimes more complicated (also in terms of the number of hyperparameters) than our approach due to using more than one term in the objective function. We would like to highlight that utilizing source data should always result in better performance compared to not using source data. Our contribution is to show that self-learning can still be very beneficial with a small compute budget and no access to source data. Our setup targets deployed systems , e.g., a self-driving car or a detection algorithm in a production line which adapts to the distribution shift on-the-fly and cannot (or should not) be retrained from scratch for every new domain shift. Model selection Gulrajani & Lopez-Paz (2021) show that model selection for hyperparameter tuning is non-trivial for the task of domain generalization, and propose model selection criteria under which models should be selected for this task. Following their spirit, we identify a model selection criterion for test-time adaptation, and rigorously use it in all our experiments. We outperform state-of-the-art techniques which did not disclose their hyperparameter selection protocols. 3 Self-learning for Test-Time Adaptation Tsypkin (1968) defines self-learning as learning when there is no external indication concerning the correctness of the response of the automatic system to the presented patterns , and thus, we use this term as a superset of different variants of pseudo-labeling and entropy minimization to highlight that these methods can be used for adaptation to unlabeled data. Different versions of self-learning have been used in both unsupervised domain adaptation (French et al., 2018; Shu et al., 2018), self-supervised representation learning (Caron et al., 2021), and in semi-supervised learning (Xie et al., 2020a). In a typical self-learning setting, a teacher network f t trained on the source domain predicts labels on the target domain. Then, a student model f s is fine-tuned on the predicted labels. In the following, let f t(x) denote the logits for sample x and let pt(j|x) σj(f t(x)) denote the probability for class j obtained from a softmax function σj( ). Similarly, f s(x) and ps(j|x) denote the logits and probabilities for the student model f s. For all techniques, one can optionally only admit samples where the probability maxj pt(j|x) exceeds some threshold. We consider three popular variants of self-learning: Pseudo-labeling with hard or soft labels, as well as entropy minimization. Hard Pseudo-Labeling (Lee, 2013; Galstyan & Cohen, 2007). We generate labels using the teacher and train the student on pseudo-labels i using the softmax cross-entropy loss, ℓH(x) := log ps(i|x), i = argmaxj pt(j|x) (1) Usually, only samples with a confidence above a certain threshold are considered for training the student. We test several thresholds but note that thresholding means discarding a potentially large portion of the data which leads to a performance decrease in itself. The teacher is updated after each epoch. Soft Pseudo-Labeling (Lee, 2013; Galstyan & Cohen, 2007). In contrast to the hard pseudo-labeling variant, we here train the student on class probabilities predicted by the teacher, j pt(j|x) log ps(j|x). (2) Soft pseudo-labeling is typically not used in conjunction with thresholding, since it already incorporates the certainty of the model. The teacher is updated after each epoch. Entropy Minimization (ENT; Grandvalet & Bengio, 2004; Wang et al., 2021). This variant is similar to soft pseudo-labeling, but we no longer differentiate between a teacher and student network. It Published in Transactions on Machine Learning Research (11/2022) corresponds to an instantaneous update of the teacher. The training objective becomes j ps(j|x) log ps(j|x). (3) Intuitively, self-learning with entropy minimization leads to a sharpening of the output distribution for each sample, making the model more confident in its predictions. Robust Pseudo-Labeling (RPL). Virtually all introduced self-learning variants use the softmax crossentropy classification objective. However, the softmax cross-entropy loss has been shown to be sensitive to label noise (Zhang & Sabuncu, 2018; Zhang et al., 2017). In the setting of domain adaptation, inaccuracies in the teacher predictions and, thus, the labels for the student, are inescapable, with severe repercussions for training stability and hyperparameter sensitivity as we show in the results. As a straight-forward solution to this problem, we propose to replace the cross-entropy loss by a robust classification loss designed to withstand certain amounts of label noise (Ghosh et al., 2017; Song et al., 2020; Shu et al., 2020; Zhang & Sabuncu, 2018). A popular candidate is the Generalized Cross Entropy (GCE) loss which combines the noise-tolerant Mean Absolute Error (MAE) loss (Ghosh et al., 2017) with the CE loss. We only consider the hard labels and use the robust GCE loss as the training loss for the student, i = argmaxj pt(j|x), ℓGCE(x, i) := q 1(1 ps(i|x)q), (4) with q (0, 1]. For the limit case q 0, the GCE loss approaches the CE loss and for q = 1, the GCE loss is the MAE loss (Zhang & Sabuncu, 2018). We test updating the teacher both after every update step of the student (RPL) and once per epoch (RPLep). Adaptation parameters. Following Wang et al. (2021), we only adapt the affine scale and shift parameters γ and β following the batch normalization layers (Ioffe & Szegedy, 2015) in most of our experiments. We verify that this type of adaptation works better than full model adaptation for large models in an ablation study in Section 6. Additional regularization in self-learning Different regularization terms have been proposed as a means to stabilize entropy minimization. Niu et al. (2022) propose an anti-forgetting weight regularization term, Liang et al. (2020); Mummadi et al. (2021); Chen et al. (2022) add a diversity regularizer, and Chen et al. (2022) use an additional consistency regularizer. These methods show improved performance with these regularization terms over simple entropy minimization, but also introduce additional hyperparameters, the tuning of which significantly increases compute requirements. In this work, we do not experiment with additional regularization, as the main point of our analysis is to show that pure self-learning is effective at improving the performance over the unadapted model across model architectures/sizes and pre-training schemes. For practitioners, we note that regularization terms can further improve the performance if the new hyperparameters are tuned properly. 4 Experiment design Datasets. Image Net-C (IN-C; Hendrycks & Dietterich, 2019) contains corrupted versions of the 50 000 images in the Image Net validation set. There are fifteen test and four hold-out corruptions, and there are five severity levels for each corruption. The established metric to report model performance on IN-C is the mean Corruption Error (m CE) where the error is normalized by the Alex Net error, and averaged over all corruptions and severity levels, see Eq. 20, Appendix C.1. Image Net-R (IN-R; Hendrycks et al., 2020a) contains 30 000 images with artistic renditions of 200 classes of the Image Net dataset. Image Net-A (IN-A; Hendrycks et al., 2019) is composed of 7500 unmodified real-world images on which standard Image Net-trained Res Net50 (He et al., 2016b) models yield chance level performance. CIFAR10 (Krizhevsky et al., 2009) and STL10 (Coates et al., 2011) are small-scale image recognition datasets with 10 classes each, and training sets of 50 000/5000 images and test sets of 10 000/8000 images, respectively. The digit datasets MNIST (Deng, 2012) and MNIST-M (Ganin et al., 2016) both have 60 000 training and 10 000 test images. Published in Transactions on Machine Learning Research (11/2022) Table 1: Self-learning decreases the error on Image Net-scale robustness datasets. Robust pseudo-labeling generally outperforms entropy minimization. number of w/o adapt w/ adapt ( ) w/ adapt ( ) m CE [%] on IN-C test ( ) parameters RPL ENT Res Net50 vanilla (He et al., 2016b) 2.6 107 76.7 50.5 (-26.2) 51.6 (-25.1) Res Net50 DAug+AM (Hendrycks et al., 2020a) 2.6 107 53.6 41.7 (-11.9) 42.6 (-11.0) Dense Net161 vanilla (Huang et al., 2017) 2.8 107 66.4 47.0 (-19.4) 47.7 (-18.7) Res Ne Xt10132 8d vanilla (Xie et al., 2017) 8.8 107 66.6 43.2 (-23.4) 44.3 (-22.3) Res Ne Xt10132 8d DAug+AM (Hendrycks et al., 2020a) 8.8 107 44.5 34.8 (-9.7) 35.5 (-9.0) Res Ne Xt10132 8d IG-3.5B (Mahajan et al., 2018) 8.8 107 51.7 40.9 (-10.8) 40.8 (-10.9) Efficient Net-L2 Noisy Student (Xie et al., 2020a) 4.8 108 28.3 22.0 (-6.3) 23.0 (-5.3) top1 error [%] on IN-R ( ) Res Net50 vanilla (He et al., 2016b) 2.6 107 63.8 54.1 (-9.7) 56.1 (-7.7) Efficient Net-L2 Noisy Student (Xie et al., 2020a) 4.8 108 23.5 17.4 (-6.1) 19.7 (-3.8) top1 error [%] on Image Net-A ( ) Efficient Net-L2 Noisy Student (Xie et al., 2020a) 4.8 108 16.5 14.8 (-1.7) 15.5 (-1.0) Hyperparameters. The different self-learning variants have a range of hyperparameters such as the learning rate or the stopping criterion. Our goal is to give a realistic estimation on the performance to be expected in practice. To this end, we optimize hyperparameters for each variant of pseudo-labeling on a hold-out set of IN-C that contains four types of image corruptions ( speckle noise , Gaussian blur , saturate and spatter ) with five different strengths each, following the procedure suggested in Hendrycks & Dietterich (2019). We refer to the hold-out set of IN-C as our dev set. On the small-scale datasets, we use the hold-out set of CIFAR10-C for hyperparameter tuning. On all other datasets, we use the hyperparameters obtained on the hold-out sets of IN-C (for large-scale datasets) or CIFAR10-C (on small-scale datasets). Models for Image Net-scale datasets. We consider five popular model architectures: Res Net50 (He et al., 2016b), Dense Net161 (Huang et al., 2017), Res Ne Xt101 (Xie et al., 2017), Efficient Net-L2 (Tan & Le, 2019), and the Vision Transformer (Vi T; Dosovitskiy et al., 2021) (see Appendix B.1 for details on the used models). For Res Net50, Dense Net and Res Ne Xt101, we include a simple vanilla version trained on Image Net only. For Res Net50 and Res Ne Xt101, we additionally include a state-of-the-art robust version trained with Deep Augment and Augmix (DAug+AM; Hendrycks et al., 2020a)2. For the Res Ne Xt model, we also include a version that was trained on 3.5 billion weakly labeled images (IG-3.5B; Mahajan et al., 2018). For Efficient Net-L2 we select the current state of the art on IN-C which was trained on 300 million images from JFT-300M (Chollet, 2017; Hinton et al., 2014) using a noisy student-teacher protocol (Xie et al., 2020a). Finally, for the Vi T, we use the model pretrained with DINO (Caron et al., 2021). We validate the Image Net and IN-C performance of all considered models and match the originally reported scores (Schneider et al., 2020). For Efficient Net-L2, we match Image Net top-1 accuracy up to 0.1% points, and IN-C up to 0.6% points m CE. Models for CIFAR10/ MNIST-scale datasets. For CIFAR10-C experiments, we use three Wide Res Nets (WRN; Zagoruyko & Komodakis, 2016): the first one is trained on clean CIFAR10 and has a depth of 28 and a width of 10, the second one is trained with CIFAR10 with the Aug Mix protocol (Hendrycks et al., 2020b) and has a depth of 40 and a width of 2, and the third one has a depth of 26 layers, and is pre-trained on clean CIFAR10 using the default training code from https://github.com/kuangliu/pytorch-cifar. We used this code-base to also train the Res Net18 and the Res Net50 models on CIFAR10. The remaining small-scale models are trained with UDA methods. We propose to regard any UDA method which requires joint training with source and target data as a pre-training step, similar to regular pre-training on Image Net, and use self-learning on top of the final checkpoint. We consider two popular UDA methods: self-supervised domain adaptation (UDA-SS; Sun et al., 2019) and Domain-Adversarial Training of Neural Networks (DANN; Ganin et al., 2016). In UDA-SS, the authors seek to align the representations of both domains by performing an auxiliary self-supervised task on both domains simultaneously. In all UDA-SS experiments, we use a Wide Res Net with a depth of 26 and a width of 16. In DANN, the authors learn a domain-invariant embedding 2see leaderboard at github.com/hendrycks/robustness Published in Transactions on Machine Learning Research (11/2022) Table 2: Self-learning decreases the error on small-scale datasets, for models pre-trained using data augmentation and unsupervised domain adaptation. Entropy minimization outperforms robust pseudo-labeling. number of w/o adapt w/ adapt ( ) w/ adapt ( ) top1 error [%] on CIFAR10-C ( ) parameters RPL ENT WRN-28-10 vanilla (Zagoruyko & Komodakis, 2016) 3.6 107 26.5 13.7 (-12.8) 13.3 (-13.2) WRN-40-2 AM (Hendrycks et al., 2020b) 2.2 106 11.2 9.0 (-2.2) 8.5 (-2.7) WRN-26-1-GN (Bartler et al., 2022) 1.5 106 18.6 18.0 (-0.6) 18.4 (0.2) WRN-26-1-BN (Zagoruyko & Komodakis, 2016) 1.5 106 25.8 15.1 (-10.7) 13.1 (-12.7) WRN-26-16 vanilla (Zagoruyko & Komodakis, 2016) 9.3 107 24.2 11.8 (-12.4) 11.2 (-13.0) WRN-26-16 UDA-SS (Sun et al., 2019) 9.3 107 27.7 18.2 (-9.5) 16.7 (-11.0) WRN-26-16 DANN (Ganin et al., 2016) 9.3 107 29.7 28.6 (-1.1) 28.5 (-1.2) UDA CIFAR10 STL10, top1 error on target [%]( ) WRN-26-16 UDA-SS (Sun et al., 2019) 9.3 107 28.7 22.9 (-5.8) 21.8 (-6.9) WRN-26-16 DANN (Ganin et al., 2016) 9.3 107 25.0 24.0 (-1.0) 23.9 (-1.1) UDA MNIST MNIST-M, top1 error on target [%]( ) WRN-26-16 UDA-SS (Sun et al., 2019) 9.3 107 4.8 2.4 (-2.4) 2.0 (-2.8) WRN-26-2 DANN (Ganin et al., 2016) 1.5 106 11.4 5.2 (-6.2) 5.1 (-6.3) by optimizing a minimax objective. For all DANN experiments except for MNIST MNIST-M, we use the same WRN architecture as above. For the MNIST MNIST-M experiment, the training with the larger model diverged and we used a smaller Wide Res Net version with a width of 2. We note that DANN training involves optimizing a minimax objective and is generally harder to tune. 5 Self-learning universally improves models Self-learning is a powerful learning scheme, and in the following section we show that it allows to perform test-time adaptation on robustified models, models obtained with large-scale pre-training, as well as (already) domain adapted models across a wide range of datasets and distribution shifts. Our main results on large-scale and small-scale datasets are shown in Tables 1 and 2. These summary tables show final results, and all experiments use the hyperparameters we determined separately on the dev set. The self-learning loss function, i.e. softor hard-pseudo-labeling / entropy minimization / robust pseudo-labeling, is a hyperparameter itself, and thus, in Tables 1 and 2, we show the overall best results. Results for the other loss functions can be found in Section 6 and in Appendix C. Self-learning successfully adapts Image Net-scale models across different model architectures on IN-C, IN-A and IN-R (Table 1). We adapt the vanilla Res Net50, Res Ne Xt101 and Dense Net161 models to IN-C and decrease the m CE by over 19 percent points in all models. Further, self-learning works for models irrespective of their size: Self-learning substantially improves the performance of the Res Net50 and the Res Next101 trained with DAug+AM, on IN-C by 11.9 and 9.7 percent points, respectively. Finally, we further improve the current state of the art model on IN-C the Efficient Net-L2 Noisy Student model and report a new state-of-the-art result of 22% m CE (which corresponds to a top1 error of 17.1%) on this benchmark with test-time adaptation (compared to 28% m CE without adaptation). Self-learning is not limited to the distribution shifts in IN-C like compression artefacts or blur. On IN-R, a dataset with renditions, self-learning improves both the vanilla Res Net50 and the Efficient Net-L2 model, the latter of which improves from 23.5% to a new state of the art of 17.4% top-1 error. For a vanilla Res Net50, we improve the top-1 error from 63.8% (Hendrycks et al., 2020a) to 54.1%. On IN-A, adapting the Efficient Net-L2 model using self-learning decreases the top-1 error from 16.5% (Xie et al., 2020a) to 14.8% top-1 error, again constituting a new state of the art with test-time adaptation on this dataset. Self-learning can also be used in an online adaptation setting, where the model continually adapts to new samples on IN-C in Fig. 7(i) or IN-R Fig. 7(ii), Appendix C.10. Adapting a Res Net50 on IN-A with RPL increases the error from 0% to 0.13% (chance level: 0.1%). Thus, an unadapted Res Net50 has 0% accuracy on IN-A by design and this error is increased to chance-level with Published in Transactions on Machine Learning Research (11/2022) self-learning. Since all labels on Image Net-A are wrong by design, predicting wrong labels as pseudo-labels does not lead to improvements beyond restoring chance-level performance. The finding that self-learning can be effective across model architectures has also been made by Wang et al. (2021) who show improved adaptation performance on CIFAR100-C for architectures based on self-attention (Zhao et al., 2020) and equilibrium solving (Bai et al., 2020), and also by Mummadi et al. (2021) who showed adaptation results for TENT and TENT+ which combines entropy minimization with a diversity regularizer for a Dense Net121 (Huang et al., 2017), a Mobile Net V2 (Sandler et al., 2018), a Res Ne Xt50 (Xie et al., 2017), and a robust model trained with DAug+AM on IN-C and IN-R. The improvements of self-learning are very stable: In Table 30, Appendix C.9, we show the averaged results across three different seeds for a Res Net50 model adapted with ENT and RPL. The unbiased std is roughly two orders of magnitude lower than any improvement we report in our paper, showcasing the robustness of our results to random initialization. Self-learning improves robustified and domain adapted models on small-scale datasets (Table 2). We test common domain adaptation techniques like DANN (Ganin et al., 2016) and UDA-SS (Sun et al., 2019), and show that self-learning is effective at further tuning such models to the target domain. We suggest to view unsupervised source/target domain adaptation as a step comparable to pre-training under corruptions, rather than an adaptation technique specifically tuned to the target set indeed, we can achieve error rates using, e.g., DANN + target adaptation previously only possible with source/target based pseudo-labeling, across different common domain adaptation benchmarks. For the UDA-SS experiments, we additionally trained a vanilla version with the same architecture using the widely used training code for CIFAR10 at https://github.com/kuangliu/pytorch-cifar (1.9k forks, 4.8k stars), and find that the vanilla trained model performs better both with and without adaptation compared to the UDA-SS model. We think that the UDA-SS model would need hyperparameter tuning; we did not perform any tuning for this model, especially because the authors provided scripts with hyperparameters they found to be optimal for different setups. In addition, the clean accuracy of the vanilla model on CIFAR10 (96.5%) is much higher than the average clean accuracy of the UDA-SS model (82.6%), which may explain or imply generally higher robustness under distribution shift (Miller et al., 2021). The finding that self-learning is more effective in the vanilla model compared to the UDA-SS model points towards the hypothesis that the network weights of the vanilla model trained on the source distribution are sufficiently general and can be tuned successfully using only the affine BN statistics, while the weights of the UDA-SS model are already co-adapted to both the source and the target distribution, and thus, self-learning is less effective. Self-learning also decreases the error on CIFAR10-C of the Wide Res Net model trained with Aug Mix (AM, Hendrycks et al., 2020b) and reaches a new state of the art on CIFAR10-C of 8.5% top1 error with test-time adaptation. Self-learning also improves large pre-trained models (Table 3). Unlike Batch Norm adaptation (Schneider et al., 2020), we show that self-learning transfers well to models pre-trained on a large amount of unlabeled data: self-learning decreases the m CE on IN-C of the Res Ne Xt101 trained on 3.5 billion weakly labeled samples (IG-3.5B, Mahajan et al., 2018) from 51.7% to 40.9%. Self-learning outperforms other test-time adaptation techniques on IN-C (Tables 4 and 5). The main point of our paper is showing that self-learning is effective across datasets, model sizes and pretraining methods. Here, we analyze whether our simple techniques can compete with other state-of-the-art adaptation methods. Overall, we find that self-learning outperforms several state-of-the-art techniques, but underperforms in some cases, especially when self-learning is combined with other techniques. By rigorous and fair model selection, we are able to improve upon TENT (Wang et al., 2021), and find that RPL performs better than entropy minimization on IN-C. We also compare to BN adaptation (Schneider et al., 2020), and find that 1) self-learning further improves the performance upon BN adapt, and 2) self-learning improves performance of a model pretrained on a large amount of data (Mahajan et al., 2018) (Table 3) which is a setting where BN adapt failed. Published in Transactions on Machine Learning Research (11/2022) Table 3: Unlike batch norm adaptation, self-learning adapts large-scale models trained on external data. m CE, test [%] ( ) w/o adapt BN adapt RPL Res Ne Xt101 vanilla 66.6 56.8 43.2 Res Ne Xt101 IG-3.5B 51.7 51.8 40.9 Table 4: Self-learning outperforms other test-time-adaptation techniques on IN-C. m CE [%] on IN-C test ( ) w/o adapt BN Adapt TENT EATA(lifelong) RPL ENT Res Net50 76.7 62.2 53.5 51.2 50.5 51.6 Table 5: Self-learning can further be improved when combining it with other techniques. literature results our results w/o adapt w/ adapt w/o adapt w/ adapt (RPL) w/ adapt (ENT) top1 error [%] on IN-C test, sev. 5( ) TTT, Res Net18 (Sun et al., 2020) 86.6 66.3 85.4 61.9 62.8 SLR, Res Net50 (Mummadi et al., 2021) 82.0 (46.9) 82.0 54.6 54.7 top1 error [%] on CIFAR10-C ( ) MT3, WRN-26-1-GN (Bartler et al., 2022) 35.7 24.4 18.6 18.4 18.0 TTT+++, Res Net50 (Liu et al., 2021) 29.1 9.8 24.9 14.6 12.4 BUFR, Res Net18 (Eastwood et al., 2022) 42.4 10.6 25.5 13.4 12.9 Table 6: Vision Transformers can be adapted with self-learning. w/o adapt w/ adapt w/ adapt w/ adapt w/ adapt m CE on IN-C test [%] ( ) affine layers bottleneck layers lin. layers all weights Vi T-S/16 62.3 51.8 46.8 45.2 43.5 ENT and RPL outperform the recently published EATA method (Niu et al., 2022). In EATA, high entropy samples are excluded from optimization, and a regularization term is added to prevent the model from forgetting the source distribution. EATA requires tuning of two additional parameters: the threshold for high entropy samples to be discarded and the regularization trade-off parameter β. RPL, ENT and simple hard PL outperform TTT (Sun et al., 2020); in particular, note that TTT requires a special loss function at training time, while our approach is agnostic to the pre-training phase. A detailed comparison to TTT is included in Appendix C.6. Mummadi et al. (2021) (SLR) is unpublished work and performs better than ENT and RPL on the highest severity of IN-C. SLR is an extension of entropy minimization where the entropy minimization loss is replaced with a version to ensure non-vanishing gradients of high confidence samples, as well as a diversity regularizer; in addition, a trainable module is prepended to the network to partially undo the distribution shift. Mummadi et al. (2021) introduce the hyperparameters δ as the trade-off parameter between their two losses, as well as κ as the momentum in their diversity regularization term. The success of SLR over ENT and RPL shows the promise of extending self-learning methods by additional objectives, and corroborates our findings on the effectiveness of self-learning. We also compare our approach to Meta Test-Time Training (MT3, Bartler et al., 2022), which combines meta-learning, self-supervision and test-time training for test-time adaptation. We find that both ENT and RPL perform better than MT3: using the same architecture as Bartler et al. (2022), our best error on CIFAR10-C is 18.0% compared to their best result of 24.4%. When exchanging Group Norm layers (Wu & He, 2018) for BN layers, the error further reduces to 13.1% (Table 11). We thus find that self-learning is more effective when adapting affine BN parameters instead of GN layers, which is consistent with the findings in Schneider et al. (2020). We included a detailed comparison to Bartler et al. (2022) in Appendix C.7. TTT+++ (Liu et al., 2021) outperforms both ENT and RPL on CIFAR10-C. Since Liu et al. (2021) do not report results on IN-C, it is impossible to judge whether their gains would generalize, although they do report much better results compared to TENT on Visda-C, so TTT+++ might also be effective on IN-C. Similar to TTT, TTT+++ requires a special loss function during pretraining and thus, cannot be used as an out-of-the-box adaptation technique on top of any pretrained checkpoint. Published in Transactions on Machine Learning Research (11/2022) Table 7: RPL (ENT) performs better on IN-C (CIFAR10-C). m CE, IN-C dev err, C10-C Res Net50 Res Ne Xt-101 Eff Net-L2 WRN-40 ENT 50.0 0.04 43.0 22.2 8.5 RPL 48.9 0.02 42.0 21.3 9.0 Table 8: RPL performs best without a threshold. threshold 0.0 0.5 0.9 m CE on IN-C dev [%] no adapt 69.5 soft PL 60.1 hard PL 53.8 51.9 52.4 RPL 49.7 49.9 51.8 Bottom-Up Feature Restoration (BUFR; Eastwood et al., 2022) outperforms self-learning on CIFAR10-C. The authors note that BUFR is applicable to dataset shifts with measurement shifts which stem from measurement artefacts, but not applicable to more complicated shifts where learning new features would be necessary. Self-supervised methods based on self-learning allow out-of-the-box test-time adaptation (Table 6). The recently published DINO method (Caron et al., 2021) is another variant of self-supervised learning that has proven to be effective for unsupervised representation learning. At the core, the method uses soft pseudo-labeling. Here, we test whether a model trained with DINO on the source dataset can be test-time adapted on IN-C using DINO to further improve out-of-distribution performance. We highlight that we specifically test the self-supervised DINO objective for its practicality as a test-time adaptation method, and did not switch the DINO objective for ENT or RPL to do test-time adaptation. Since the used model is a vision transformer model, we test different choices of adaptation parameters and find considerable performance improvements in all cases, yielding an m CE of 43.5% m CE at a parameter count comparable to a Res Net50 model. For adapting the affine layers, we follow Houlsby et al. (2019). 6 Understanding test-time adaptation with self-learning In the following section, we show ablations and interesting insights of using self-learning for test-time adaptation. If not specified otherwise, all ablations are run on the hold-out corruptions of IN-C (our dev set) with a vanilla Res Net50. Robust pseudo-labeling outperforms entropy minimization on large-scale datasets while the reverse is true on small-scale datasets (Table 7). We find that robust pseudo-labeling consistently improves over entropy minimization on IN-C, while entropy minimization performs better on smaller scale data (CIFAR10, STL10, MNIST). The finding highlights the importance of testing both algorithms on new datasets. The improvement is typically on the order of one percent point. Robust pseudo-labeling allows usage of the full dataset without a threshold (Table 8). Classical hard labeling needs a confidence threshold (T) for best performance, thereby reducing the dataset size, while best performance for RPL is reached for full dataset training with a threshold T of 0.0. We corroborate the results of Wang et al. (2021) who showed that TENT outperforms standard hard labeling with a threshold on the highest severity of CIFAR10-C and CIFAR100-C. We show that this result transfers to IN-C, for a variety of thresholds and pseudo-labeling variants. Short update intervals are crucial for fast adaptation (Table 9). Having established that RPL generally performs better than softand hard-labeling, we vary the update interval for the teacher. We find that instant updates are most effective. In entropy minimization, the update interval is instant per default. Adaptation of only affine layers is important in CNNs (Table 10). On IN-C, adapting only the affine parameters after the normalization layers (i.e., the rescaling and shift parameters β and γ) works better on a Res Net50 architecture than adapting all parameters or only the last layer. We indicate the number of adapted parameters in brackets. Note that for Vision Transformers, full model adaptation works better than affine adaptation (see Table 6). We also noticed that on convolutional models with a smaller parameter count like Res Net18, full model adaptation is possible. Wang et al. (2021) also used the affine BN parameters for test-time adaptation with TENT, and report that last layer optimization can improve performance but degrades with further optimization. They suggest that full model optimization does not improve performance Published in Transactions on Machine Learning Research (11/2022) Table 9: RPL performs best with instantaneous updates (Res Net50). Update interval (RPL) w/o adapt none epoch instant m CE, IN-C dev [%] 69.5 54.0 49.7 49.2 Table 10: RPL performs best when affine BN parameters are adapted (Res Net50). Mechanism w/o adapt last layer full model affine m CE, IN-C dev [%] 69.5 60.2 51.5 48.9 adapted parameters 0 2M 22.6M 5.3k Table 11: Self-learning works better in models with BN layers compared to models with GN layers, although un-adapted models with GN are more stable under distribution shift compared to models with BN layers. number of w/o adapt w/ adapt ( ) w/ adapt ( ) top1 error [%] on CIFAR10-C ( ) parameters RPL ENT WRN-26-1-BN (Zagoruyko & Komodakis, 2016) 1.5 106 25.8 15.1 (-10.7) 13.1 (-12.7) WRN-26-1-GN (Bartler et al., 2022) 1.5 106 18.6 18.4 (-0.2) 18.0 (-0.6) m CE [%] on IN-C test ( ) Res Net50 Big Transfer (Kolesnikov et al., 2020) 2.6 107 55.0 54.4 (-0.6) 56.4 (+1.4) Table 12: Our expanded analysis confirms hyperparameter choices from the literature. Method short updates adapt affine params use BN instead of GN adapt to full dataset TENT (Wang et al., 2021) EATA (Niu et al., 2022) SRL (Mummadi et al., 2021) MEMO (Zhang et al., 2021) RPL/ENT(ours) at all. In contrast, we find gains with full model adaptation, but stronger gains with adaptation of only affine parameters. Affine BN parameters work better for test-time adaptation compared to GN parameters. (Table 11). Schneider et al. (2020) showed that models with batch normalization layers are less robust to distribution shift compared to models with group normalization (Wu & He, 2018) layers. However, after adapting BN statistics, the adapted model outperformed the non-adapted GN model. Here, we show that these results also hold for test-time adaptation when adapting a model with GN or BN layers. We show that a Wide Res Net-26-1 (WRN-26-1) vanilla model with BN layers pretrained on clean CIFAR10 has a much higher error on CIFAR10-C than the same model with GN layers, but it has a much lower error after adaptation. The full results for the WRN-26-1 model can be found in Appendix C.7. Further, we test the pretrained Big Transfer (Kolesnikov et al., 2020) models which have GN layers, and find only small improvements with RPL , and no improvements with ENT. There are no pretrained weights released for the Big Transfer models which have BN layers, thus, a comparison similar to the WRN-26-1 model is not possible. A more detailed discussion on our Big Transfer results as well as a hyperparameter selection study can be found in Appendix E.2. Hyperparameters obtained on corruption datasets transfer well to real world datasets. When evaluating models, we select the hyperparameters discussed above (the learning rate and the epoch used for early stopping are the most critical ones) on the dev set (full results in Appendix C.2). We note that this technique transfers well to IN-R and -A, highlighting the practical value of corruption robustness datasets for adapting models on real distribution shifts. Learning rate and number of training epochs are important hyperparameters We tune the learning rate as well as the number of training epochs for all models, except for the Efficient Net-L2 model where we only train for one epoch due to computational constraints. In Tables 18, 19, 20, and 21 in Appendix C.2, we show that both ENT and RPL collapse after a certain number of epochs, showing the inherent instability of pseudo-labeling. While Wang et al. (2021) trained their model only for one epoch with one learning rate, Published in Transactions on Machine Learning Research (11/2022) Table 13: Self-learning leads to an increased ECE compared to the unadapted model. adaptation IN-C full IN-C sev 1 IN-C sev 2 IN-C sev 3 IN-C sev 4 IN-C sev 5 w/o adapt 2.3 0.8 2.3 0.3 2.1 0.3 2.1 0.3 2.2 0.4 2.9 1.6 RPL 10.6 7.3 6.6 0.5 7.9 1.5 8.9 2.1 11.2 3.3 18.7 12.6 ENT 10.6 7.3 6.6 0.5 7.9 1.5 8.9 2.1 11.2 3.3 18.7 12.6 we rigorously perform a full hyperparameter selection on a hold-out set (our dev set), and use the optimal hyperparameters on all tested datasets. We believe that this kind of experimental rigor is essential in order to be able to properly compare methods. Our analysis confirms hyperparameter choices from the literature Having searched over a broad space of hyperparameters and self-learning algorithms, we identified ENT and RPL as the best performing variants across different model sizes, architectures and pretraining techniques. We summarize the most important hyperparameter choices in Table 12 and compare them to those used in the literature when applying self-learning for test-time adaptation. Wang et al. (2021) identified that on CIFAR-C, entropy minimization outperforms hard PL, and found that affine adaptation of BN parameters works better than full model adaptation, and adapted to the full dataset since TENT does not have a threshold in contrast to hard PL. EATA (Niu et al., 2022) and SRL (Mummadi et al., 2021) are extensions of TENT, and thus, followed their hyperparameter choices. EATA does not adapt to the full dataset as they do not adapt to very similar samples or samples with high entropy values. MEMO (Zhang et al., 2021) adapts to a single sample and adapts all model weights. It would be interesting to study whether the performance of MEMO can be improved when adapting only affine BN parameters. Pure self-learning hurts calibration, (Table 13) Eastwood et al. (2022) show that approaches based on entropy minimization and pseudo-labeling hurt calibration due to the objective of confidence maximization. We corroborate their results and report the Expected Calibration Error (ECE; Naeini et al., 2015) when adapting a vanilla Res Net50 model with RPL and ENT. We report the mean ECE (lower is better) and standard deviation across corruptions (and severities) below. We observe that both methods increase ECE compared to the unadapted model. The increase in ECE is higher for more severe corruptions which can be explained by a successively stronger distribution shift compared to the source dataset. Self-learning leads to slightly decreased accuracy on the source dataset (Table 14) To judge the forgetting effect on the source distribution, we calculated the accuracy on clean Image Net for our adapted ENT and RPL checkpoints. We note that success of self-learning can partially be attributed to correcting the BN statistics of the vanilla model with respect to the distribution shift (Schneider et al., 2020). Thus, we only wish to examine the effect of fine-tuning of the affine BN parameters to the target distribution with respect to the source distribution. Thus, we again correct the mismatched statistics to the source dataset when calculating the accuracy. We report the mean and standard deviation across corruptions (and severities) below. We observe that both ENT and RPL lead to a decrease in performance on the source dataset, an effect which has also been observed by Niu et al. (2022). The decrease in performance is higher for more severe corruptions which can be explained by a increasingly stronger distribution shift compared to the source dataset. We find that the effect of forgetting is less pronounced in RPL compared to ENT. Additional experiments and ablation studies, as well as detailed results for all models and datasets can be found in Appendix C. We discuss additional proof-of-concept implementations on the WILDS benchmark (Koh et al., 2021), Big Transfer (Bi T; Chen et al., 2020b) models and on self-learning based UDA models in Appendix E. On WILDS, self-learning is effective for the Camelyon17 task with a systematic shift between train, validation and test sets (each set is comprised of different hospitals), while self-learning fails to improve on tasks with mixed domains, such as on the Rx Rx1 and the FMo W tasks.These results support our claim that self-learning is effective, while showing the important limitation when applied to more diverse shifts. Published in Transactions on Machine Learning Research (11/2022) Table 14: Self-learning leads to slightly decreased accuracy on the source dataset (clean IN). model top1 accuracy on Image Net val [%] w/o adapt 74.2 RPL adapted to IN-C (avg over corruptions, sev. 1) 74.7 0.6 RPL adapted to IN-C (avg over corruptions, sev. 2) 73.9 0.9 RPL adapted to IN-C (avg over corruptions, sev. 3) 73.0 1.3 RPL adapted to IN-C (avg over corruptions, sev. 4) 71.8 1.8 RPL adapted to IN-C (avg over corruptions, sev. 5) 69.8 3.0 ENT adapted to IN-C (avg over corruptions, sev. 1) 74.3 0.6 ENT adapted to IN-C (avg over corruptions, sev. 2) 73.2 1.1 ENT adapted to IN-C (avg over corruptions, sev. 3) 72.3 1.7 ENT adapted to IN-C (avg over corruptions, sev. 4) 70.7 2.3 ENT adapted to IN-C (avg over corruptions, sev. 5) 68.0 3.9 7 A simple model of stability in self-learning We observed that different self-learning schemes are optimal for small-scale vs. large-scale datasets and varying amount of classes. We reconsider the used loss functions, and unify them into ( f(x), entropy minimization sg(f(x)), pseudo-labeling. where we introduced student and teacher temperature τs and τt as parameters in the softmax function and the stop gradient operation sg. To study the learning dynamics, we consider a linear student network f s = ws Rd and a linear teacher network f t = wt Rd which are trained on N data points {xi}N i=1 with a binary cross-entropy loss function L defined as i=1 ℓ(xi) = σt(x i wt) log σs(x i ws) + σt( x i wt) log σs( x i ws) , where σt(z) = 1 1 + e z/τt and σs(z) = 1 1 + e z/τs . With stop gradient, student and teacher evolve in time according to ws = ws L ws, wt , wt = α(ws wt), (7) where α is the learning rate of the teacher. Without stop gradient, student and teacher are set equal to each other (following the instant updates we found to perform empirically best), and they evolve as w = w L(w), where ws = wt = w. (8) We restrict the theoretical analysis to the time evolution of the components of ws,t in direction of two data points xk and xl, ys,t k x k ws,t and ys,t l x l ws,t. All other components ys,t i with i = k, l are neglected to reduce the dimensionality of the equation system. It turns out that the resulting model captures the neural network dynamics quite well despite the drastic simplification of taking only two data points into account (see Figure 2 for a comparison of the model vs. self-learning on the CIFAR-C dataset). We obtain the dynamics: with stop gradient: ys k = x k ws (ℓ(xk) + ℓ(xl)) , ys l = x l ws (ℓ(xk) + ℓ(xl)) , yt k = α(yt k ys k), yt l = α(yt l ys l ), without stop gradient: yk = x k w (ℓ(xk) + ℓ(xl)) , yl = x l w (ℓ(xk) + ℓ(xl)) . With this setup in place, we can derive Published in Transactions on Machine Learning Research (11/2022) 0% 50% 100% 2 1 0 1 log10 τt τt = 2τs τt = τs 1 0 1 log10 τt Figure 2: For the two point model, we show accuracy, and for the CIFAR10-C simulation, we show improvement (yellow) vs. degradation (purple) over the non-adapted baseline (BAS). An important convergence criterion for pseudo-labeling (top row) and entropy minimization (bottom row) is the ratio of student and teacher temperatures; it lies at τs = τt for PL, and 2τs = τt for ENT. Despite the simplicity of the two-point model, the general convergence regions transfer to CIFAR10-C. Proposition 7.1 (Collapse in the two-point model). The student and teacher networks ws and wt trained with stop gradient do not collapse to the trivial representation x : x ws = 0, x wt = 0 if τs > τt. The network w trained without stop gradient does not collapse if τs > τt/2. Proof. see A.1. We validate the proposition on a simulated two datapoint toy dataset, as well as on the CIFAR-C dataset (Figure 2). In general, the size and location of the region where collapse is observed in the simulated model also depends on the initial conditions, the learning rate and the optimization procedure. An in depth discussion, as well as additional simulations are given in Appendix A. Entropy minimization with standard temperatures (τs = τt = 1) and hard pseudo-labeling (τt 0) are hence stable. The two-point learning dynamics vanish for soft pseudo-labeling with τs = τt, suggesting that one would have to analyze a more complex model with more data points. While this does not directly imply that the learning is unstable at this point, we empirically observe that both entropy minimization and hard labeling outperform soft-labeling in practice. The finding aligns with empirical work: For instance, Caron et al. (2021) fixed τs and varied τt during training, and empirically found an upper bound for τt above which the training was no longer stable. It also aligns with our findings suggesting that hard-labeling tends to outperform soft-labeling approaches, and soft-labeling performs best when selecting lower teacher temperatures. In practice, the result suggests that student temperatures should always exceed the teacher temperatures for pseudo-labeling, and student temperatures should always exceed half the teacher temperature for entropy minimization, which narrows the search space for hyperparameter optimization considerably. 8 Adapting models on a wider range of distribution shifts reveals limitations of robustification and adaptation methods Robustness datasets on Image Net-scale have so far been limited to a few selected domains (image corruptions in IN-C, image renditions in IN-R, difficult images for Res Net50 classifiers in IN-A). In order to test our approach on a wider range of complex distribution shifts, we re-purpose the dataset from the Visual Domain Adaptation Challenge 2019 (Domain Net; Saenko et al., 2019) as an additional robustness benchmark. Published in Transactions on Machine Learning Research (11/2022) Table 15: Self-learning decreases the top1 error on IN-D domains with strong initial performance, but fails to improve performance on challenging domains. domain Real Painting Clipart Sketch Infograph Quickdraw Image Net adapt w/o w/ w/o w/ w/o w/ w/o w/ w/o w/ w/o w/ w/o model Eff Net-L2 Noisy Student 29.2 27.9 42.7 40.9 45.0 37.9 56.4 51.5 77.9 94.3 98.4 99.4 11.6 Res Net50 DAug+AM 39.2 36.5 58.7 53.4 68.4 57.0 75.2 61.3 88.1 83.2 98.2 99.1 23.3 Res Net50 vanilla 40.1 37.3 65.1 57.8 76.0 63.6 82.0 73.0 89.6 85.1 99.2 99.8 23.9 Creation of Image Net-D The original Domain Net dataset comes with six image styles: Clipart , Real , Infograph , Painting , Quickdraw and Sketch , and has 345 classes in total, out of which 164 overlap with Image Net. We map these 164 Domain Net classes to 463 Image Net classes, e.g., for an image from the bird class in Domain Net, we accept all 39 bird classes in Image Net as valid predictions. Image Net also has ambiguous classes, e.g., it has separate classes for cellular telephone and dial phone . For these cases, we accept both predictions as valid. In this sense, the mapping from Domain Net to Image Net is a one-to-many mapping. We refer to the smaller version of Domain Net that is now compatible with Image Net-trained models as Image Net-D (IN-D). The benefit of IN-D over Domain Net is this re-mapping to Image Net classes which allows robustness researchers to easily benchmark on this dataset, without the need of re-training a model (as is common in UDA). We show example images from IN-D in Table 15. The detailed evaluation protocol on IN-D, our label-mapping procedure from Domain Net to Image Net along with justifications for our design choices and additional analysis are outlined in Appendix D. The most similar robustness dataset to IN-D is IN-R which contains renditions of Image Net classes, such as art, cartoons, deviantart, graffiti, embroidery, graphics and others. The benefit of IN-D over IN-R is that in IN-D, the images are separated according to the domain allowing for studying of systematic domain shifts, while in IN-R, the different domains are not distinguished. Image Net-Sketch (Wang et al., 2019) is a dataset similar to the Sketch domain of IN-D. More robust models perform better on IN-D. To test whether self-learning is helpful for more complex distribution shifts, we adapt a vanilla Res Net50, several robust IN-C models and the Efficient Net-L2 Noisy Student model on IN-D. We use the same hyperparameters we obtained on IN-C dev for all our IN-D experiments3. We show our main results in Table 15. Comparing the performance of the vanilla Res Net50 model to its robust DAug+AM variant, we find that the DAug+AM model performs better on all domains, with the most significant gains on the Clipart , Painting and Sketch domains. We show detailed results for all domains and all tested models in Appendix D.3, along with results on IN-C and IN-R for comparison. We find that the best performing models on IN-D are also the strongest ones on IN-C and IN-R which indicates good generalization capabilities of the techniques combined for these models, given the large differences between the three considered datasets. The Spearman s rank correlation coefficient between IN-C and IN-D (averaged over all domains) is 0.54, and 0.73 between IN-R and IN-D. Thus, the errors on IN-R are strongly correlated to errors on IN-D which can be explained by the similarity of IN-D and IN-R. We show Spearman s rank correlation cofficients for the individual domains versus IN-C/IN-R in Fig. 9 in Appendix D.5, and find correlation values above 0.8 between IN-R and IN-D for all domains except for the Real domain where the coefficient is almost zero. Further, we find that even the best models perform 20 to 30 percentage points worse on IN-D compared to their performance on IN-C or IN-R, indicating that IN-D might be a more challenging benchmark. All models struggle with some domains of IN-D. The Efficient Net-L2 Noisy Student model obtains the best results on most domains. However, we note that the overall error rates are surprisingly high compared to the model s strong performance on the other considered datasets (IN-A: 14.8% top-1 error, IN-R: 17.4% top-1 3In regards to hyperparameter selection, we performed a control experiment where we selected hyperparameters with leave-one-out cross validation this selection scheme actually performed worse than IN-C parameter selection (see Appendix D.2). Published in Transactions on Machine Learning Research (11/2022) error, IN-C: 22.0% m CE). Even on the Real domain closest to clean Image Net where the Efficient Net-L2 model has a top-1 error of 11.6%, the model only reaches a top-1 error of 29.2%. Self-learning decreases the top-1 error on all domains except for Infograph and Quickdraw . We note that both domains have very high error rates from the beginning and thus hypothesize that the produced pseudo-labels are of low quality. Error analysis on IN-D. We investigate the errors a Res Net50 model makes on IN-D by analyzing the most frequently predicted classes for different domains to reveal systematic errors indicative of the encountered distribution shifts. We find most errors interpretable: the classifier assigns the label comic book to images from the Clipart or Painting domains, website to images from the Infograph domain, and envelope to images from the Sketch domain. Thus, the classifier predicts the domain rather than the class. We find no systematic errors on the Real domain which is expected since this domain should be similar to Image Net. Detailed results on the most frequently predicted classes for different domains can be found in Fig. 9, Appendix D.5. IN-D should be used as an additional robustness benchmark. While the error rates on IN-C, -R and -A are at a well-acceptable level for our largest Efficient Net-L2 model after adaptation, IN-D performance is consistently worse for all models. We propose to move from isolated benchmark settings like IN-R (single domain) to benchmarks more common in domain adaptation (like Domain Net) and make IN-D publicly available as an easy to use dataset for this purpose. 9 Best practices and evaluation in test-time adaptation Based on our results as well as our discussion on previous work, we arrive at several proposals on how test-time adaptation should be evaluated in future work to ensure scientific rigor: 1. Cross-validation: We propose using the hold-out set of IN-C for model selection of all relevant hyperparameters, and then using these hyperparameters for testing on different datasets. 2. Comparison to simple baselines: With proper hyperparameter tuning, very simple baselines can perform on par with sophisticated approaches. This insight is also discussed by Gulrajani & Lopez-Paz (2021) for the setting of domain generalization and by Rusak et al. (2020) for robustness to common corruptions. 3. Using more robust models: Test-time adaptation can further improve upon robust models, which were pre-trained with more data or with UDA, or using protocols to increase robustness. A test-time adaptation method will be much more relevant to practitioners if it can improve upon the most robust model they can find for their task. 4. Important hyperparameters: We identify several important hyperparameters which affect the final performance in a crucial way, and thus, should be tuned to ensure fair comparisons: Number of adaptation epochs and learning rate: The final performance crucially depends on both of these parameters for all models and all methods that we have studied. Adaptation parameters While adaptation of affine batch normalization parameters works best for adaptation of CNNs, full adaptation performs best for Vi Ts. Therefore, it is important to benchmark test-time adaptation for different model architectures and adaptation parameters. 10 Conclusion We evaluated and analysed how self-learning, an essential component in many unsupervised domain adaptation and self-supervised pre-training techniques, can be applied for adaptation to both small and large-scale image recognition problems common in robustness research. We demonstrated new state-of-the-art adaptation results with the Efficient Net-L2 model on the benchmarks Image Net-C, -R, and -A, and introduced a new benchmark dataset (Image Net-D) which remains challenging even after adaptation. Our theoretical analysis shows the influence of the temperature parameter in the self-learning loss function on the training stability and provides guidelines how to choose a suitable value. Based on our extensive experiments, we formulate best practices for future research on test-time adaptation. Across the large diversity of (systematic) distribution Published in Transactions on Machine Learning Research (11/2022) shifts, architectures and pre-training methods we tested in this paper, we found that self-learning almost universally improved test-time performance. An important limitation of current self-learning methods is the observed instability over longer adaptation time frames. While we mitigate this issue through model selection (and show its robustness across synthetic and natural distribution shifts), this might not universally hold across distribution shifts encountered in practice. Concurrent work, e.g. Niu et al. (2022) tackle this problem through modifications of self-learning algorithms, and we think this direction will be important to continue to explore in future work. That being said, we hope that our results encourage both researchers and practitioners to experiment with self-learning if their data distribution shifts. Reproducibility Statement We attempted to make our work as reproducible as possible: We mostly used pre-trained models which are publicly available and we denoted the URL addresses of all used checkpoints; for the checkpoints that were necessary to retrain, we report the Github directories with the source code and used an official or verified reference implementation when available. We report all used hyperparameters in the Appendix and released the code. Software and Data Code for reproducing results of this paper is available at https://github.com/ bethgelab/robustness. Author Contributions St S and ER conceived the project with suggestions from MB and WB. ER ran initial experiments. St S performed large scale experiments on Image Net-C,R,A with support from PG, ER and WB. ER performed domain adaptation and CIFAR experiments with suggestions from St S. ER implemented all baselines that have been included as comparisons in the manuscript. GP ran DINO adaptation, WILDS, and self-ensembling experiments with suggestions from ER & St S. ER developed the Image Net-D dataset with suggestions from St S and WB, and performed the large scale experiments on Image Net-D. LE and St S developed the theoretical model, ER contributed the CIFAR-C experiments. WB, ER and St S wrote the initial manuscript with help from all authors. ER wrote an extensive related work section. ER reviewed and edited the manuscript in the course of the author-reviewer discussion period. Acknowledgements We thank Roland S. Zimmermann, Yi Zhu, Raghavan Manmatha, Matthias Kümmerer, Matthias Tangemann, Bernhard Schölkopf, Justin Gilmer, Shubham Krishna, Julian Bitterwolf, Berkay Kicanaoglu and Mohammadreza Zolfaghari for helpful discussions on pseudo labeling and feedback on our paper draft. We thank Yasser Jadidi and Alex Smola for support in the setup of our compute infrastructure. We thank the anonymous reviewers from TMLR and our Action Editor for their constructive feedback. We thank the International Max Planck Research School for Intelligent Systems (IMPRS-IS) for supporting E.R. and St.S.; St.S. acknowledges his membership in the European Laboratory for Learning and Intelligent Systems (ELLIS) Ph D program. This work was supported by the German Federal Ministry of Education and Research (BMBF) through the Tübingen AI Center (FKZ: 01IS18039A), by the Deutsche Forschungsgemeinschaft (DFG) in the priority program 1835 under grant BR2321/5-2 and in the SFB 1233, Robust Vision: Inference Principles and Neural Mechanisms (TP3), project number: 276693517. WB acknowledges financial support via an Emmy Noether Grant funded by the German Research Foundation (DFG) under grant no. BR 6382/1-1 and via the Open Philantropy Foundation funded by the Good Ventures Foundation. WB is a member of the Machine Learning Cluster of Excellence, EXC number 2064/1 Project number 390727645. Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin, M., Ghemawat, S., Irving, G., Isard, M., et al. Tensorflow: A system for large-scale machine learning. In 12th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 16), pp. 265 283, 2016. Cited on p. 48. Azimi, F., Palacio, S., Raue, F., Hees, J., Bertinetto, L., and Dengel, A. Self-supervised test-time adaptation on video data. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 3439 3448, 2022. Cited on p. 4. Bai, S., Koltun, V., and Kolter, J. Z. Multiscale deep equilibrium models. Advances in Neural Information Processing Systems, 33:5238 5250, 2020. Cited on p. 9. Published in Transactions on Machine Learning Research (11/2022) Bartler, A., Bühler, A., Wiewel, F., Döbler, M., and Yang, B. Mt3: Meta test-time training for self-supervised test-time adaption. In International Conference on Artificial Intelligence and Statistics, pp. 3080 3090. PMLR, 2022. Cited on pp. 3, 4, 8, 10, 12, 36, and 37 Berner, C., Brockman, G., Chan, B., Cheung, V., Debiak, P., Dennison, C., Farhi, D., Fischer, Q., Hashme, S., Hesse, C., et al. Dota 2 with large scale deep reinforcement learning. Ar Xiv preprint, abs/1912.06680, 2019. URL https://arxiv.org/abs/1912.06680. Cited on p. 2. Berthelot, D., Roelofs, R., Sohn, K., Carlini, N., and Kurakin, A. Adamatch: A unified approach to semi-supervised learning and domain adaptation, 2021. Cited on p. 4. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., Mc Candlish, S., Radford, A., Sutskever, I., and Amodei, D. Language models are few-shot learners. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. URL https://proceedings.neurips.cc/paper/ 2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html. Cited on p. 2. Cai, T., Gao, R., Lee, J. D., and Lei, Q. A theory of label propagation for subpopulation shift. ar Xiv preprint ar Xiv:2102.11203, 2021. Cited on p. 5. Caron, M., Touvron, H., Misra, I., Jégou, H., Mairal, J., Bojanowski, P., and Joulin, A. Emerging properties in self-supervised vision transformers. Ar Xiv preprint, abs/2104.14294, 2021. URL https: //arxiv.org/abs/2104.14294. Cited on pp. 3, 5, 7, 11, 15, and 31 Chen, D., Wang, D., Darrell, T., and Ebrahimi, S. Contrastive test-time adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 295 305, 2022. Cited on pp. 4 Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp. 1597 1607. PMLR, 2020a. Cited on p. 4. Chen, T., Kornblith, S., Swersky, K., Norouzi, M., and Hinton, G. E. Big self-supervised models are strong semi-supervised learners. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020b. URL https://proceedings. neurips.cc/paper/2020/hash/fcbc95ccdd551da181207c0c1400c655-Abstract.html. Cited on p. 13. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 1800 1807. IEEE Computer Society, 2017. doi: 10.1109/CVPR.2017.195. URL https://doi.org/10. 1109/CVPR.2017.195. Cited on pp. 7 and 30 Coates, A., Ng, A., and Lee, H. An analysis of single-layer networks in unsupervised feature learning. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, 2011. Cited on p. 6. Croce, F., Andriushchenko, M., Sehwag, V., Debenedetti, E., Flammarion, N., Chiang, M., Mittal, P., and Hein, M. Robustbench: a standardized adversarial robustness benchmark. Ar Xiv preprint, abs/2010.09670, 2020. URL https://arxiv.org/abs/2010.09670. Cited on p. 31. Deng, J., Dong, W., Socher, R., Li, L., Li, K., and Li, F. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2009), 20-25 June 2009, Miami, Florida, USA, pp. 248 255. IEEE Computer Society, 2009. doi: 10.1109/CVPR.2009.5206848. URL https://doi.org/10.1109/CVPR.2009.5206848. Cited on p. 2. Published in Transactions on Machine Learning Research (11/2022) Deng, L. The mnist database of handwritten digit images for machine learning research. IEEE Signal Processing Magazine, 29(6):141 142, 2012. Cited on p. 6. Dodge, S. F. and Karam, L. J. A study and comparison of human and deep learning recognition performance under visual distortions. In International Conference on Computer Communications and Networks, ICCCN 2017, 2017. Cited on p. 2. Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the 36th International Conference on Machine Learning, 2021. Cited on pp. 3 and 7 Eastwood, C., Mason, I., Williams, C. K., and Schölkopf, B. Source-free adaptation to measurement shift via bottom-up feature restoration. In International Conference on Learning Representations (ICLR), 2022. Cited on pp. 4, 10, 11, and 13 Farahani, A., Voghoei, S., Rasheed, K., and Arabnia, H. R. A brief review of domain adaptation. Advances in data science and information engineering, pp. 877 894, 2021. Cited on p. 2. French, G., Mackiewicz, M., and Fisher, M. H. Self-ensembling for visual domain adaptation. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. Open Review.net, 2018. URL https://openreview.net/ forum?id=rkpo Tax A-. Cited on pp. 5, 46, and 48 Galstyan, A. and Cohen, P. R. Empirical comparison of hard and soft label propagation for relational classification. In 17th international conference on Inductive logic programming, 2007. Cited on p. 5. Ganin, Y., Ustinova, E., Ajakan, H., Germain, P., Larochelle, H., Laviolette, F., Marchand, M., and Lempitsky, V. Domain-adversarial training of neural networks. The journal of machine learning research, 17(1): 2096 2030, 2016. Cited on pp. 2, 6, 7, 8, and 9 Geirhos, R., Temme, C. R. M., Rauber, J., Schütt, H. H., Bethge, M., and Wichmann, F. A. Generalisation in humans and deep neural networks. In Bengio, S., Wallach, H. M., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS 2018, December 3-8, 2018, Montréal, Canada, pp. 7549 7561, 2018. URL https://proceedings.neurips.cc/paper/2018/hash/ 0937fb5864ed06ffb59ae5f9b5ed67a9-Abstract.html. Cited on p. 2. Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F. A., and Brendel, W. Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. Open Review.net, 2019. URL https://openreview.net/forum?id=Bygh9j09KX. Cited on pp. 2 and 40 Ghosh, A., Kumar, H., and Sastry, P. S. Robust loss functions under label noise for deep neural networks. In Singh, S. P. and Markovitch, S. (eds.), Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, February 4-9, 2017, San Francisco, California, USA, pp. 1919 1925. AAAI Press, 2017. URL http://aaai.org/ocs/index.php/AAAI/AAAI17/paper/view/14759. Cited on p. 6. Gidaris, S., Singh, P., and Komodakis, N. Unsupervised representation learning by predicting image rotations. ar Xiv preprint ar Xiv:1803.07728, 2018. Cited on p. 3. Grandvalet, Y. and Bengio, Y. Semi-supervised learning by entropy minimization. In Advances in Neural Information Processing Systems 17 [Neural Information Processing Systems, NIPS 2004, December 13-18, 2004, Vancouver, British Columbia, Canada], pp. 529 536, 2004. URL https://proceedings.neurips. cc/paper/2004/hash/96f2b50b5d3613adf9c27049b2a888c7-Abstract.html. Cited on p. 5. Gulrajani, I. and Lopez-Paz, D. In search of lost domain generalization. In Proceedings of the 36th International Conference on Machine Learning, 2021. Cited on pp. 5 and 17 Published in Transactions on Machine Learning Research (11/2022) He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 770 778. IEEE Computer Society, 2016a. doi: 10.1109/CVPR.2016.90. URL https: //doi.org/10.1109/CVPR.2016.90. Cited on p. 2. He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 770 778. IEEE Computer Society, 2016b. doi: 10.1109/CVPR.2016.90. URL https: //doi.org/10.1109/CVPR.2016.90. Cited on pp. 6, 7, and 31 Hendrycks, D. and Dietterich, T. G. Benchmarking neural network robustness to common corruptions and perturbations. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. Open Review.net, 2019. URL https://openreview.net/forum?id=HJz6ti Cq Ym. Cited on pp. 4, 6, 7, and 39 Hendrycks, D., Zhao, K., Basart, S., Steinhardt, J., and Song, D. Natural adversarial examples. Ar Xiv preprint, abs/1907.07174, 2019. URL https://arxiv.org/abs/1907.07174. Cited on p. 6. Hendrycks, D., Basart, S., Mu, N., Kadavath, S., Wang, F., Dorundo, E., Desai, R., Zhu, T., Parajuli, S., Guo, M., et al. The many faces of robustness: A critical analysis of out-of-distribution generalization. Ar Xiv preprint, abs/2006.16241, 2020a. URL https://arxiv.org/abs/2006.16241. Cited on pp. 2, 3, 6, 7, 8, 30, 31, and 40 Hendrycks, D., Mu, N., Cubuk, E. D., Zoph, B., Gilmer, J., and Lakshminarayanan, B. Augmix: A simple data processing method to improve robustness and uncertainty. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020b. URL https://openreview.net/forum?id=S1gmrx HFv B. Cited on pp. 7, 8, 9, 31, and 40 Hinton, G., Vinyals, O., and Dean, J. Distilling the knowledge in a neural network. In NIPS Deep Learning Workshop, 2014. Cited on pp. 7 and 30 Hoffman, J., Tzeng, E., Park, T., Zhu, J.-Y., Isola, P., Saenko, K., Efros, A., and Darrell, T. Cycada: Cycle-consistent adversarial domain adaptation. In International conference on machine learning, pp. 1989 1998. Pmlr, 2018. Cited on p. 2. Houlsby, N., Giurgiu, A., Jastrzebski, S., Morrone, B., De Laroussilhe, Q., Gesmundo, A., Attariyan, M., and Gelly, S. Parameter-efficient transfer learning for NLP. In Proceedings of the 36th International Conference on Machine Learning, 2019. Cited on p. 11. Huang, G., Liu, Z., van der Maaten, L., and Weinberger, K. Q. Densely connected convolutional networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 2261 2269. IEEE Computer Society, 2017. doi: 10.1109/CVPR.2017.243. URL https://doi.org/10.1109/CVPR.2017.243. Cited on pp. 7, 9, 31, and 45 Ioffe, S. and Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Bach, F. R. and Blei, D. M. (eds.), Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6-11 July 2015, volume 37 of JMLR Workshop and Conference Proceedings, pp. 448 456. JMLR.org, 2015. URL http://proceedings.mlr.press/v37/ioffe15.html. Cited on pp. 3 and 6 Iwasawa, Y. and Matsuo, Y. Test-time classifier adjustment module for model-agnostic domain generalization. Advances in Neural Information Processing Systems, 34:2427 2440, 2021. Cited on p. 4. Kim, Y., Cho, D., Han, K., Panda, P., and Hong, S. Domain adaptation without source data. IEEE Transactions on Artificial Intelligence, 2021. Cited on p. 2. Koh, P. W., Sagawa, S., Marklund, H., Xie, S. M., Zhang, M., Balsubramani, A., Hu, W., Yasunaga, M., Phillips, R. L., Gao, I., Lee, T., David, E., Stavness, I., Guo, W., Earnshaw, B. A., Haque, I. S., Beery, S., Leskovec, J., Kundaje, A., Pierson, E., Levine, S., Finn, C., and Liang, P. WILDS: A benchmark of Published in Transactions on Machine Learning Research (11/2022) in-the-wild distribution shifts. In International Conference on Machine Learning (ICML), 2021. Cited on pp. 13, 45, and 48 Kolesnikov, A., Beyer, L., Zhai, X., Puigcerver, J., Yung, J., Gelly, S., and Houlsby, N. Big transfer (bit): General visual representation learning. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part V 16, pp. 491 507. Springer, 2020. Cited on pp. 12 Krizhevsky, A., Hinton, G., et al. Learning multiple layers of features from tiny images. 2009. Cited on pp. 4 Kumar, A., Ma, T., and Liang, P. Understanding self-training for gradual domain adaptation. In International Conference on Machine Learning, pp. 5468 5479. PMLR, 2020. Cited on p. 4. Kundu, J. N., Venkat, N., Babu, R. V., et al. Universal source-free domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4544 4553, 2020. Cited on p. 2. Lee, D.-H. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In ICML Workshop : Challenges in Representation Learning (WREPL), 2013. Cited on p. 5. Li, R., Jiao, Q., Cao, W., Wong, H.-S., and Wu, S. Model adaptation: Unsupervised domain adaptation without source data. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020. Cited on p. 2. Liang, J., Hu, D., and Feng, J. Do we really need to access the source data? source hypothesis transfer for unsupervised domain adaptation. In International Conference on Machine Learning, 2020. Cited on pp. 2 and 6 Liu, Y., Kothari, P., van Delft, B. G., Bellot-Gurlet, B., Mordan, T., and Alahi, A. Ttt++: When does selfsupervised test-time training fail or thrive? In Thirty-Fifth Conference on Neural Information Processing Systems, 2021. Cited on pp. 4 and 10 Mahajan, D., Girshick, R., Ramanathan, V., He, K., Paluri, M., Li, Y., Bharambe, A., and van der Maaten, L. Exploring the limits of weakly supervised pretraining. In Proceedings of the European Conference on Computer Vision (ECCV), 2018. Cited on pp. 2, 7, 9, 30, and 31 Marcel, S. and Rodriguez, Y. Torchvision the machine-vision package of torch. In ACM International Conference on Multimedia, 2010. Cited on pp. 31 and 48 Merkel, D. Docker: Lightweight linux containers for consistent development and deployment. Linux J., 2014 (239), 2014. ISSN 1075-3583. Cited on p. 48. Miller, J. P., Taori, R., Raghunathan, A., Sagawa, S., Koh, P. W., Shankar, V., Liang, P., Carmon, Y., and Schmidt, L. Accuracy on the line: on the strong correlation between out-of-distribution and in-distribution generalization. In International Conference on Machine Learning, pp. 7721 7735. PMLR, 2021. Cited on p. 9. Mummadi, C. K., Hutmacher, R., Rambach, K., Levinkov, E., Brox, T., and Metzen, J. H. Test-time adaptation to distribution shift by confidence maximization and input transformation. ar Xiv preprint ar Xiv:2106.14999, 2021. Cited on pp. 3, 4, 6, 9, 10, 12, and 13 Nado, Z., Padhy, S., Sculley, D., D Amour, A., Lakshminarayanan, B., and Snoek, J. Evaluating predictiontime batch normalization for robustness under covariate shift. Ar Xiv preprint, abs/2006.10963, 2020. URL https://arxiv.org/abs/2006.10963. Cited on p. 3. Naeini, M. P., Cooper, G., and Hauskrecht, M. Obtaining well calibrated probabilities using bayesian binning. In Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015. Cited on p. 13. Published in Transactions on Machine Learning Research (11/2022) Niu, S., Wu, J., Zhang, Y., Chen, Y., Zheng, S., Zhao, P., and Tan, M. Efficient test-time model adaptation without forgetting. In Proceedings of the 39th International Conference on Machine Learning, Proceedings of Machine Learning Research. PMLR, 2022. Cited on pp. 4, 6, 10, 12, 13, and 18 Paszke, A., Gross, S., Chintala, S., Chanan, G., Yang, E., De Vito, Z., Lin, Z., Desmaison, A., Antiga, L., and Lerer, A. Automatic differentiation in Py Torch. In NIPS Autodiff Workshop, 2017. Cited on p. 48. Prabhu, V., Khare, S., Kartik, D., and Hoffman, J. Sentry: Selective entropy optimization via committee consistency for unsupervised domain adaptation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 8558 8567, 2021. Cited on p. 4. Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, pp. 8748 8763. PMLR, 2021. Cited on p. 39. Recht, B., Roelofs, R., Schmidt, L., and Shankar, V. Do imagenet classifiers generalize to imagenet? In International Conference on Machine Learning, pp. 5389 5400. PMLR, 2019. Cited on p. 39. Rusak, E., Schott, L., Zimmermann, R., Bitterwolf, J., Bringmann, O., Bethge, M., and Brendel, W. Increasing the robustness of dnns against image corruptions by playing the game of noise. Ar Xiv preprint, abs/2001.06057, 2020. URL https://arxiv.org/abs/2001.06057. Cited on pp. 2, 17, and 40 Saenko, K., Peng, X., Usman, B., Saito, K., and Hu, P. Visual Domain Adaptation Challenge (Vis DA-2019), 2019. URL http://ai.bu.edu/visda-2019/. Cited on p. 15. Sandler, M., Howard, A. G., Zhu, M., Zhmoginov, A., and Chen, L. Mobilenetv2: Inverted residuals and linear bottlenecks. In 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, June 18-22, 2018, pp. 4510 4520. IEEE Computer Society, 2018. doi: 10.1109/CVPR.2018.00474. URL http://openaccess.thecvf.com/content_cvpr_2018/html/Sandler_ Mobile Net V2_Inverted_Residuals_CVPR_2018_paper.html. Cited on p. 9. Schneider, S., Rusak, E., Eck, L., Bringmann, O., Brendel, W., and Bethge, M. Improving robustness against common corruptions by covariate shift adaptation. In Advances in neural information processing systems, 2020. Cited on pp. 3, 4, 7, 9, 10, 12, 13, 30, 34, and 36 Shu, J., Zhao, Q., Chen, K., Xu, Z., and Meng, D. Learning adaptive loss for robust learning with noisy labels. Ar Xiv preprint, abs/2002.06482, 2020. URL https://arxiv.org/abs/2002.06482. Cited on p. 6. Shu, R., Bui, H. H., Narui, H., and Ermon, S. A DIRT-T approach to unsupervised domain adaptation. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. Open Review.net, 2018. URL https://openreview.net/ forum?id=H1q-TM-AW. Cited on pp. 4, 5, and 47 Sohn, K., Berthelot, D., Li, C.-L., Zhang, Z., Carlini, N., Cubuk, E. D., Kurakin, A., Zhang, H., and Raffel, C. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In Neur IPS, 2020. Cited on p. 4. Song, H., Kim, M., Park, D., and Lee, J.-G. Learning from noisy labels with deep neural networks: A survey. Ar Xiv preprint, abs/2007.08199, 2020. URL https://arxiv.org/abs/2007.08199. Cited on p. 6. Sun, Y., Tzeng, E., Darrell, T., and Efros, A. A. Unsupervised domain adaptation through self-supervision. Ar Xiv preprint, abs/1909.11825, 2019. URL https://arxiv.org/abs/1909.11825. Cited on pp. 7, 8, and 9 Sun, Y., Wang, X., Liu, Z., Miller, J., Efros, A., and Hardt, M. Test-time training with self-supervision for generalization under distribution shifts. In International conference on machine learning, pp. 9229 9248. PMLR, 2020. Cited on pp. 3, 4, 10, 35, and 36 Published in Transactions on Machine Learning Research (11/2022) Tan, M. and Le, Q. V. Efficientnet: Rethinking model scaling for convolutional neural networks. In Chaudhuri, K. and Salakhutdinov, R. (eds.), Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pp. 6105 6114. PMLR, 2019. URL http://proceedings.mlr.press/v97/tan19a.html. Cited on pp. 3, 7, and 31 Tange, O. Gnu parallel - the command-line power tool. ;login: The USENIX Magazine, 36(1):42 47, 2011. URL http://www.gnu.org/s/parallel. Cited on p. 48. Taori, R., Dave, A., Shankar, V., Carlini, N., Recht, B., and Schmidt, L. Measuring robustness to natural distribution shifts in image classification. Advances in Neural Information Processing Systems, 33:18583 18599, 2020. Cited on p. 39. Tsypkin, Y. Self-learning what is it? IEEE Transactions on Automatic Control, 13(6):608 612, 1968. doi: 10.1109/TAC.1968.1099015. Cited on pp. 3 and 5 Tzeng, E., Hoffman, J., Saenko, K., and Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7167 7176, 2017. Cited on p. 2. Virtanen, P., Gommers, R., Oliphant, T. E., Haberland, M., Reddy, T., Cournapeau, D., Burovski, E., Peterson, P., Weckesser, W., Bright, J., van der Walt, S. J., Brett, M., Wilson, J., Jarrod Millman, K., Mayorov, N., Nelson, A. R. J., Jones, E., Kern, R., Larson, E., Carey, C., Polat, İ., Feng, Y., Moore, E. W., Vand er Plas, J., Laxalde, D., Perktold, J., Cimrman, R., Henriksen, I., Quintero, E. A., Harris, C. R., Archibald, A. M., Ribeiro, A. H., Pedregosa, F., van Mulbregt, P., and Contributors, S. . . Sci Py 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17:261 272, 2020. doi: https://doi.org/10.1038/s41592-019-0686-2. Cited on p. 48. Wang, D., Shelhamer, E., Liu, S., Olshausen, B., and Darrell, T. Fully test-time adaptation by entropy minimization. In International Conference on Learning Representations (ICLR), 2021. Cited on pp. 3, 4, 5, 6, 9, 11, 12, and 13 Wang, H., Ge, S., Lipton, Z., and Xing, E. P. Learning robust global representations by penalizing local predictive power. In Advances in Neural Information Processing Systems, pp. 10506 10518, 2019. Cited on p. 16. Wei, C., Shen, K., Chen, Y., and Ma, T. Theoretical analysis of self-training with deep networks on unlabeled data. In ICLR, 2020. Cited on p. 5. Wightman, R. Pytorch image models. https://github.com/rwightman/pytorch-image-models, 2019. Cited on pp. 46 and 48 Wu, Y. and He, K. Group normalization. In Proceedings of the European Conference on Computer Vision (ECCV), 2018. Cited on pp. 10 and 12 Xie, Q., Luong, M., Hovy, E. H., and Le, Q. V. Self-training with noisy student improves imagenet classification. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13-19, 2020, pp. 10684 10695. IEEE, 2020a. doi: 10.1109/CVPR42600.2020.01070. URL https://doi.org/10.1109/CVPR42600.2020.01070. Cited on pp. 2, 3, 5, 7, 8, 30, 31, and 34 Xie, S., Girshick, R. B., Dollár, P., Tu, Z., and He, K. Aggregated residual transformations for deep neural networks. In 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pp. 5987 5995. IEEE Computer Society, 2017. doi: 10.1109/CVPR.2017.634. URL https://doi.org/10.1109/CVPR.2017.634. Cited on pp. 7, 9, and 31 Xie, S. M., Kumar, A., Jones, R., Khani, F., Ma, T., and Liang, P. In-n-out: Pre-training and self-training using auxiliary information for out-of-distribution robustness. ar Xiv preprint ar Xiv:2012.04550, 2020b. Cited on p. 4. Published in Transactions on Machine Learning Research (11/2022) Zagoruyko, S. and Komodakis, N. Wide residual networks. In Wilson, R. C., Hancock, E. R., and Smith, W. A. P. (eds.), Proceedings of the British Machine Vision Conference 2016, BMVC 2016, York, UK, September 19-22, 2016. BMVA Press, 2016. URL http://www.bmva.org/bmvc/2016/papers/paper087/index.html. Cited on pp. 7, 8, and 12 Zhang, C., Bengio, S., Hardt, M., Recht, B., and Vinyals, O. Understanding deep learning requires rethinking generalization. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. Open Review.net, 2017. URL https: //openreview.net/forum?id=Sy8gd B9xx. Cited on p. 6. Zhang, M., Levine, S., and Finn, C. Memo: Test time robustness via adaptation and augmentation. ar Xiv preprint ar Xiv:2110.09506, 2021. Cited on pp. 4, 12, 13, and 38 Zhang, Z. and Sabuncu, M. R. Generalized cross entropy loss for training deep neural networks with noisy labels. In Bengio, S., Wallach, H. M., Larochelle, H., Grauman, K., Cesa-Bianchi, N., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS 2018, December 3-8, 2018, Montréal, Canada, pp. 8792 8802, 2018. URL https://proceedings.neurips.cc/paper/2018/hash/ f2925f97bc13ad2852a7a551802feea0-Abstract.html. Cited on p. 6. Zhao, H., Jia, J., and Koltun, V. Exploring self-attention for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10076 10085, 2020. Cited on p. 9. Zou, Y., Yu, Z., Liu, X., Kumar, B., and Wang, J. Confidence regularized self-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 5982 5991, 2019. Cited on p. 4. Published in Transactions on Machine Learning Research (11/2022) A A two-point model of self-learning A.1 Proof of Proposition 7.1 Learning dynamics with stop gradient. Computing the stop gradient evolution defined in equation 7 explicitly yields ws = ws L = 1 σt(x i wt)σs( x i ws) σt( x i wt)σs(x i ws) xi wt = α(ws wt) The second equality uses the well-known derivative of the sigmoid function, zσ(z) = σ(z)σ( z). The equation system of 2d nonlinear, coupled ODEs for ws Rd and wt Rd in equation 10 is analytically difficult to analyze. Instead of studying the ODEs directly, we act on them with the data points x k , k = 1, . . . , N, and investigate the dynamics of the components x k ws,t ys,t k : x i xk σt(yt i)σs( ys i ) σt( yt i)σs(ys i ) yt k = α(ys k yt k). The learning rate of each mode ys k is scaled by (x k xi) which is much larger for i = k than for i = k in high-dimensional spaces. In the two-point approximation, we consider only the two (in absolute value) largest terms i = k, l for a given k in the sum in equation 11. Any changes that ys,t k (t) and ys,t l (t) might induce in other modes ys,t i (t) are neglected, and so we are left with only four ODEs: τs xk 2 σt(yt k)σs( ys k) σt( yt k)σs(ys k) τs (x k xl) σt(yt l)σs( ys l ) σt( yt l)σs(ys l ) , τs xl 2 σt(yt l)σs( ys l ) σt( yt l)σs(ys l ) τs (x k xl) σt(yt k)σs( ys k) σt( yt k)σs(ys k) yt k =α(ys k yt k), yt l = α(ys l yt l). The fixed points of equation 12 satisfy ys k = ys l = yt k = yt l = 0. (13) For α > 0, requiring yt k = yt l = 0 implies that ys k = yt k and ys l = yt l. For τs = τt, the two remaining equations ys k = ys l = 0 vanish automatically so that there are no non-trivial two-point learning dynamics. For τs = τt, there is a fixed point at ys,t k = ys,t l = 0 since at this point, each bracket in equation 12 vanishes individually: σt(yk,l)σs( yk,l) σs( yk,l)σt(yk,l) yk,l=0 = 1 4 = 0. (14) At the fixed point ys,t k = ys,t l = 0, ws and wt are orthogonal to both xk and xl and hence classification fails. If this fixed point is stable, ws and wt will stay at the fixed point once they have reached it, i.e. the model collapses. The fixed point is stable when all eigenvalues of the Jacobian J of the ODE system equation 12 evaluated at ys,t k = ys,t l = 0 are negative. Two eigenvalues λ are always negative, whereas the two other Published in Transactions on Machine Learning Research (11/2022) eigenvalues λ are positive if τs > τt: J ys,t k =ys,t l =0 = (x k xl) 4τ 2 s xk 2 4τsτt (x k xl) 4τsτt (x k xl) 4τ 2 s xl 2 (x k xl) 4τsτt xl 2 4τsτt α 0 α 0 0 α 0 α λ = 1 16τ 2s A2 + 32ατ 2s B(τs/τt 1) < 0, λ = 1 16τ 2s A2 + 32ατ 2s B(τs/τt 1) > 0 if τs > τt, where A = 8ατ 2 s B, B = xk + xl + q ( xk xl )2 + 4(x k xl)2 To sum up, training with stop gradient and τs > τt avoids a collapse of the two-point model to the trivial representation ys,t k = ys,t l = 0 since the fixed point is not stable in this parameter regime. Learning dynamics without stop gradient Without stop gradient, we set wt = ws w which leads to an additional term in the gradient: w = w L = 1 σt(x i w)σs( x i w) σt( x i w)σs(x i w) xi i=1 σt(x i w)σt( x i w) log σs(x i w) log σs( x i w) | {z } =log ((1+eyi/τs)/(1+e yi/τs))=yi/τs As before, we focus on the evolution of the two components yk = w xk and yl = w xl. yk = xk 2 1 τs (σt(yk)σs( yk) σt( yk)σs(yk)) + 1 τt σt(yk)σt( yk)yk + (x k xl) 1 τs (σt(yl)σs( yl) σt( yl)σs(yl)) + 1 τsτt σt(yl)σt( yl)yl yl = xl 2 1 τs (σt(yl)σs( yl) σt( yl)σs(yl)) + 1 τt σt(yl)σt( yl)yl + (x k xl) 1 τs (σt(yk)σs( yk) σt( yk)σs(yk)) + 1 τsτt σt(yk)σt( yk)yk There is a fixed point at yk = yl = 0 where each bracket in equation 17 vanishes individually, 1 τs (σt(yk,l)σs( yk,l) σt( yk,l)σs(yk,l)) + 1 τsτt σt(yk,l)σt( yk,l)yk,l yk,l = 0. (18) The Jacobian of the ODE system in equation 17 and its eigenvalues evaluated at the fixed point are given by J yk=yl=0 = (x k xl) 4τs (x k xl) 4τs λ1,2 = 1 8τs xk 4 + xl 4 2 xk 2 xl 2 + 4(x k xl)2 | {z } xk 2+ xl 2 + xk 2 + xl 2 | {z } 0 with equality if xk= xl Hence the fixed point is unstable when τs > τt/2 and thus the model without stop gradient does not collapse onto yk = yl = 0 in this regime, concluding the proof. Published in Transactions on Machine Learning Research (11/2022) A.2 Simulation of the two-point model For visualization purposes in the main paper, we set ws = wt = [0.5, 0.5] and train the model using instant gradient updates on the dataset with points x1 = [1, 0] and x2 = [0, 1] using SGD with learning rate 0.1 and momentum 0.9. We varied student and teacher temperatures on a log-scale with 250 points from 10 3 to 10. Qualitatively similar results can be obtained without momentum training, at higher learning rates (most likely due to the implicit learning rate scaling introduced by the momentum term). Note that the temperature scales for observing the collapse effect depend on the learning rate, and the exact training strategy lower learning rates can empirically prevent the model from collapsing and shift the convergence region. The result in Figure 2 will hence depend on the exact choice of learning rate (which is currently not considered in our continuous time evolution theory), while the predicted region without collapse is robust to details of the optimization. To visualize the impact of different hyperparameters, we show variants of the two point model with different learning rates using gradient descent with (Figure 3) and without momentum (Figure 4), and with different start conditions (Figure 5), which all influence the regions where the model degrades, but not the stable regions predicted by our theory. Published in Transactions on Machine Learning Research (11/2022) τt = 2τs τt = τs 2 1 0 1 log10 τt 0% 50% 100% 2 1 0 1 log10 τt 2 1 0 1 log10 τt 2 1 0 1 log10 τt 2 1 0 1 log10 τt Figure 3: Entropy minimization (top) Training two point model with momentum 0.9 and different learning rates with initialization ws = wt = [0.5, 0.5] . τt = 2τs τt = τs 2 1 0 1 log10 τt 0% 50% 100% 2 1 0 1 log10 τt 2 1 0 1 log10 τt 2 1 0 1 log10 τt 2 1 0 1 log10 τt Figure 4: Training a two point model without momentum and different learning rates with initialization ws = wt = [0.5, 0.5] . Note that especially for lower learning rates, longer training would increase the size of the collapsed region. τt = 2τs τt = τs 2 1 0 1 log10 τt 0% 50% 100% 2 1 0 1 log10 τt 2 1 0 1 log10 τt 2 1 0 1 log10 τt 2 1 0 1 log10 τt Figure 5: Training a two point model with momentum 0.9 and different learning rates with initialization ws = wt = [0.6, 0.3] . Published in Transactions on Machine Learning Research (11/2022) B Additional information on used models B.1 Details on all hyperparameters we tested for different models For all models except Efficient Net-L2, we adapt the batch norm statistics to the test domains following (Schneider et al., 2020). We do not expect significant gains for combining Efficient Net-L2 with batch norm adaptation: as demonstrated in (Schneider et al., 2020), models trained with large amounts of weakly labeled data do not seem to benefit from batch norm adaptation. Res Net50 models (IN-C) We use a vanilla Res Net50 model and compare softand hard-labeling against entropy minimization and robust pseudo-labeling. To find optimal hyperparameters for all methods, we perform an extensive evaluation and test (i) three different adaptation mechanisms (ii) several learning rates 1.0 10 4, 1.0 10 3, 1.0 10 2 and 5.0 10 2, (iii) the number of training epochs and (iv) updating the teacher after each epoch or each iteration. For all experiments, we use a batch size of 128. The hyperparameter search is performed on IN-C dev. We then use the optimal hyperparameters to evaluate the methods on the IN-C test set. Res Ne Xt101 models The Res Ne Xt101 model is considerably larger than the Res Net50 model and we therefore limit the number of ablation studies we perform for this architecture. Besides a baseline, we include a state-of-the-art robust version trained with Deep Augment+Augmix (DAug+AM, Hendrycks et al., 2020a) and a version that was trained on 3.5 billion weakly labeled images (IG-3.5B, Mahajan et al., 2018). We only test the two leading methods on the Res Ne Xt101 models (ENT and RPL). We vary the learning rate in same interval as for the Res Net50 model but scale it down linearly to account for the smaller batch size of 32. We only train the affine batch normalization parameters because adapting only these parameters leads to the best results on Res Net50 and is much more resource efficient than adapting all model parameters. Again, the hyperparameter search is performed only on the development corruptions of IN-C. We then use the optimal hyperparameters to evaluate the methods on the IN-C test set. Efficient Net-L2 models The current state of the art on IN, IN-C, IN-R and IN-A is an Efficient Net-L2 trained on 300 million images from JFT-300M (Chollet, 2017; Hinton et al., 2014) using a noisy student-teacher protocol (Xie et al., 2020a). We adapt this model for only one epoch due to resource constraints. During the hyperparameter search, we only evaluate three corruptions on the IN-C development set4 and test the learning rates 4.6 10 2, 4.6 10 3, 4.6 10 4 and 4.6 10 5. We use the optimal hyperparameters to evaluate ENT and RPL on the full IN-C test set (with all severity levels). UDA-SS models We trained the models using the scripts from the official code base at github.com/ yueatsprograms/uda_release. We used the provided scripts for the cases: (a) source: CIFAR10, target: STL10 and (b) source: MNIST, target: MNIST-M. For the case (c) source: CIFAR10, target: CIFAR10-C, we used the hyperparameters from case (a) since this case seemed to be the closest match to the new setting. We think that the baseline performance of the UDA-SS models can be further improved with hyperparameter tuning. DANN models To train models with the DANN-method, we used the Py Torch implementation of this paper at https://github.com/fungtion/DANN_py3. The code base only provides scripts and hyperparameters for the case (b) source: MNIST, target: MNIST-M. For the cases (a) and (c), we used the same optimizer and trained the model for 100 epochs. We think that the baseline performance of the DANN models can be further improved with hyperparameter tuning. Preprocessing For IN, IN-R, IN-A and IN-D, we resize all images to 256 256 px and take the center 224 224 px crop. The IN-C images are already rescaled and cropped. We center and re-scale the color values with µRGB = [0.485, 0.456, 0.406] and σRGB = [0.229, 0.224, 0.225]. For the Efficient Net-L2, we follow the 4We compare the results of computing the dev set on the 1, 3 and 5 severities versus the 1, 2, 3, 4 and 5 severities on our Res Ne Xt101 model in the Supplementary material. Published in Transactions on Machine Learning Research (11/2022) procedure in Xie et al. (2020a) and rescale all inputs to a resolution of 507 507 px and then center-crop them to 475 475 px. B.2 Full list of used models Image Net scale models Image Net trained models (Res Net50, Dense Net161, Res Ne Xt) are taken directly from torchvision (Marcel & Rodriguez, 2010). The model variants trained with Deep Augment and Aug Mix augmentations (Hendrycks et al., 2020b;a) are taken from https://github.com/hendrycks/imagenet-r. The weakly-supervised Res Ne Xt101 model is taken from the Py Torch Hub. For Efficient Net (Tan & Le, 2019), we use the Py Torch re-implementation available at https://github.com/rwightman/gen-efficientnet-pytorch. This is a verified re-implementation of the original work by Xie et al. (2020a). We verify the performance on Image Net, yielding a 88.23% top-1 accuracy and 98.546% top-5 accuracy which is within 0.2% points of the originally reported result (Xie et al., 2020a). On Image Net-C, our reproduced baseline achieves 28.9% m CE vs. 28.3% m CE originally reported by Xie et al. (2020a). As noted in the re-implementation, this offset is possible due to minor differences in the pre-processing. It is possible that our adaptation results would improve further when applied on the original codebase by Xie et al.. Small scale models We train the UDA-SS models using the original code base at github.com/ yueatsprograms/uda_release, with the hyperparameters given in the provided bash scripts. For our DANN experiments, we use the Py Torch implementation at github.com/fungtion/DANN_py3. We use the hyperparameters in the provided bash scripts. The following Table 16 contains all models we evaluated on various datasets with references and links to the corresponding source code. Table 16: Model checkpoints used for our experiments. Model Source Wide Res Net(28,10) (Croce et al., 2020) https://github.com/Robust Bench/robustbench/tree/master/robustbench Wide Res Net(40,2)+Aug Mix (Croce et al., 2020) https://github.com/Robust Bench/robustbench/tree/master/robustbench Res Net50 (He et al., 2016b) https://github.com/pytorch/vision/tree/master/torchvision/models Res Ne Xt101, 32 8d (He et al., 2016b) https://github.com/pytorch/vision/tree/master/torchvision/models Dense Net (Huang et al., 2017) https://github.com/pytorch/vision/tree/master/torchvision/models Res Ne Xt101, 32 8d (Xie et al., 2017) https://pytorch.org/hub/facebookresearch_WSL-Images_resnext/ Res Net50+Deep Augment+Aug Mix (Hendrycks et al., 2020a) https://github.com/hendrycks/imagenet-r Res Next101 (Hendrycks et al., 2020a) https://github.com/hendrycks/imagenet-r Res Next101 32 8d IG-3.5B (Mahajan et al., 2018) https://github.com/facebookresearch/WSL-Images/blob/master/hubconf.py Noisy Student Efficient Net-L2 (Xie et al., 2020a) https://github.com/rwightman/gen-efficientnet-pytorch Vi T-S/16 (Caron et al., 2021) https://github.com/facebookresearch/dino Published in Transactions on Machine Learning Research (11/2022) C Detailed and additional Results on IN-C C.1 Definition of the mean Corruption Error (m CE) The established performance metric on IN-C is the mean Corruption Error (m CE), which is obtained by normalizing the model s top-1 errors with the top-1 errors of Alex Net across the C=15 test corruptions and S=5 severities: m CE(model) = 1 PS s=1 errmodel c,s PS s=1 err Alex Net c,s . (20) The Alex Net errors used for normalization are shown in Table 17. Category Corruption top1 error Noise Gaussian Noise 0.886428 Shot Noise 0.894468 Impulse Noise 0.922640 Defocus Blur 0.819880 Glass Blur 0.826268 Motion Blur 0.785948 Zoom Blur 0.798360 Snow 0.866816 Frost 0.826572 Fog 0.819324 Brightness 0.564592 Contrast 0.853204 Elastic Transform 0.646056 Pixelate 0.717840 JPEG Compression 0.606500 Hold-out Noise Speckle Noise 0.845388 Hold-out Digital Saturate 0.658248 Hold-out Blur Gaussian Blur 0.787108 Hold-out Weather Spatter 0.717512 Table 17: Alex Net top1 errors on Image Net-C C.2 Detailed results for tuning epochs and learning rates We tune the learning rate for all models and the number of training epochs for all models except the Efficient Net-L2. In this section, we present detailed results for tuning these hyperparameters for all considered models. The best hyperparameters that we found in this analysis, are summarized in Table 22. Table 18: m CE in % on the IN-C dev set for ENT and RPL for different numbers of training epochs when adapting the affine batch norm parameters of a Res Net50 model. criterion ENT RPL lr 10 4 10 3 10 2 10 4 10 3 10 2 0 60.2 60.2 60.2 60.2 60.2 60.2 1 54.3 50.0 72.5 57.4 51.1 52.5 2 52.4 50.9 96.5 55.8 49.6 57.4 3 51.5 51.0 112.9 54.6 49.2 64.2 4 51.0 52.4 124.1 53.7 49.0 71.0 5 50.7 53.5 131.2 52.9 48.9 76.3 Table 19: m CE ( ) in % on the IN-C dev set for different learning rates for Efficient Net L2. We favor q = 0.8 over q = 0.7 due to slightly improved robustness to changes in the learning rate in the worst case error setting. lr (4.6 ) base 1 10 3 1 10 4 1 10 5 1 10 6 ENT 25.5 87.8 25.3 22.2 24.1 RPLq=0.7 25.5 60.3 21.3 23.3 n/a RPLq=0.8 25.5 58.2 21.4 23.4 n/a Published in Transactions on Machine Learning Research (11/2022) Table 22: The best hyperparameters for all models that we found on IN-C. For all models, we fine-tune only the affine batch normalization parameters and use q = 0.8 for RPL. The small batchsize for the Efficient Net model is due to hardware limitations. number of Model Method Learning rate batch size epochs vanilla Res Net50 ENT 1 10 3 128 1 vanilla Res Net50 RPL 1 10 3 128 5 vanilla Res Ne Xt101 ENT 2.5 10 4 128 1 vanilla Res Ne Xt101 RPL 2.5 10 4 128 4 IG-3.5B Res Ne Xt101 ENT 2.5 10 4 128 4 IG-3.5B Res Ne Xt101 RPL 2.5 10 3 128 2 DAug+AM Res Ne Xt101 ENT 2.5 10 4 128 1 DAug+AM Res Ne Xt101 RPL 2.5 10 4 128 4 Efficient Net-L2 ENT 4.6 10 5 8 1 Efficient Net-L2 RPL 4.6 10 4 8 1 Table 20: m CE in % on IN-C dev for entropy minimization for different learning rates and training epochs for Res Ne Xt101. (div.=diverged) ENT Baseline IG-3.5B DAug+AM lr 2.5 1e-4 1e-3 5e-3 1e-4 1e-3 5e-3 1e-4 1e-3 5e-3 epoch base 53.6 53.6 53.6 47.4 47.4 47.4 37.4 37.4 37.4 1 43.0 92.2 div. 40.9 40.4 58.6 35.4 46.4 div. 2 44.8 118.4 div. 39.8 41.5 69.5 35.5 90.8 div. 3 45.4 131.9 div. 39.3 42.6 76.1 35.5 122.5 div. 4 46.7 div. div. 39.1 44.2 84.3 35.6 133.8 div. Table 21: m CE in % on IN-C dev for robust pseudolabeling for different learning rates and training epochs for Res Ne Xt101. (div.=diverged) RPL Baseline IG-3.5B DAug+AM lr 2.5 1e-4 1e-3 5e-3 1e-4 1e-3 5e-3 1e-4 1e-3 5e-3 epoch base 53.6 53.6 53.6 47.4 47.4 47.4 37.4 37.4 37.4 1 43.4 51.3 div. 45.0 39.9 43.6 35.3 35.1 79.1 2 42.3 63.2 div. 43.4 39.3 48.2 34.9 35.6 121.2 3 42.0 72.6 div. 42.4 39.4 52.9 34.7 40.1 133.5 4 42.0 72.6 div. 42.4 39.4 52.9 34.7 40.1 133.5 C.3 Detailed results for all IN-C corruptions We outline detailed results for all corruptions and models in Table 23. Performance across the severities in the dataset is depicted in Figure 6. All detailed results presented here are obtained by following the model selection protocol outlined in the main text. 1 2 3 4 5 Severity 1 2 3 4 5 Severity 1 2 3 4 5 Severity RNx101 IG-3.5B 1 2 3 4 5 Severity RNx101 Deep Aug+Augmix 1 2 3 4 5 Severity Noisy Student L2 ENT RPL Base Figure 6: Severity-wise mean corruption error (normalized using the average Alex Net baseline error for each corruption) for Res Net50 (RN50), Res Next101 (RNx101) variants and the Noisy Student L2 model. Especially for more robust models (Deep Augment+Augmix and Noisy Student L2), most gains are obtained across higher severities 4 and 5. For weaker models, the baseline variant (Base) is additionally substantially improved for smaller corruptions. Published in Transactions on Machine Learning Research (11/2022) Table 23: Detailed results for each corruption along with mean corruption error (m CE) as reported in Table 2 in the main paper. We show (unnormalized) top-1 error rate averaged across 15 test corruptions along with the mean corruption error (m CE: which is normalized). Hyperparameter selection for both ENT and RPL was carried out on the dev corruptions as outlined in the main text. Mismatch in baseline m CE for Efficient Net-L2 can be most likely attributed to pre-processing differences between the original tensorflow implementation Xie et al. (2020a) and the Py Torch reimplementation we employ. We start with slightly weaker baselines for Res Net50 and Res Next101 than Schneider et al. (2020): Res Net50 and Res Next101 results are slightly worse than previously reported results (typically 0.1% points) due to the smaller batch size of 128 and 32. Smaller batch sizes impact the quality of re-estimated batch norm statistics when computation is performed on the fly Schneider et al. (2020), which is of no concern here due to the large gains obtained by pseudo-labeling. Res Net50 Baseline (Schneider et al., 2020) 62.2 Baseline (ours) 57.2 59.5 60.0 61.4 62.3 51.3 49.5 54.6 54.1 39.3 29.1 46.7 41.4 38.2 41.8 62.8 ENT 45.5 45.5 46.8 48.4 48.7 40.0 40.3 42.0 46.6 33.2 28.1 42.4 35.2 32.2 35.1 51.6 RPL 44.2 44.4 45.5 47.0 47.4 38.8 39.2 40.7 46.2 32.5 27.7 42.7 34.6 31.6 34.4 50.5 Res Ne Xt101 Baseline Baseline (Schneider et al., 2020) 56.7 Baseline (ours) 52.8 54.1 54.0 55.4 56.8 46.7 46.6 48.5 49.4 36.6 25.4 42.8 37.8 32.5 36.7 56.8 ENT 40.5 39.5 41.4 41.6 43.0 34.1 34.5 35.0 39.4 28.5 24.0 33.8 30.3 27.2 30.5 44.3 RPL 39.4 38.9 39.8 40.3 41.0 33.4 33.8 34.6 38.7 28.0 23.7 31.4 29.8 26.8 30.0 43.2 Res Ne Xt101 IG-3.5B Baseline (Schneider et al., 2020) 51.6 Baseline (ours) 50.7 51.5 53.1 54.2 55.5 45.5 44.7 41.7 42.0 28.1 20.1 33.8 35.4 27.8 33.9 51.8 ENT 38.6 38.3 40.4 41.4 41.5 33.8 33.6 32.2 34.6 24.1 19.7 26.3 27.6 24.2 27.9 40.8 RPL 39.1 39.2 40.8 42.1 42.4 33.7 33.5 31.8 34.7 23.9 19.6 26.1 27.5 23.8 27.5 40.9 Res Ne Xt101 Deep Aug+Augmix Baseline (Schneider et al., 2020) 38.0 Baseline (ours) 30.0 30.0 30.2 32.9 35.5 28.9 31.9 33.3 32.8 29.5 22.6 28.4 31.2 23.0 26.5 38.1 ENT 28.7 28.5 29.0 29.8 30.9 26.9 28.0 29.3 30.5 26.2 23.2 26.3 28.5 23.7 26.0 35.5 RPL 28.1 27.8 28.3 29.1 30.1 26.3 27.4 28.8 29.8 25.9 22.7 25.6 27.9 23.2 25.4 34.8 Noisy Student L2 Baseline (Xie et al., 2020a) 28.3 Baseline (ours) 21.6 22.0 20.5 23.9 40.5 19.8 23.2 22.8 26.9 21.0 15.2 21.2 24.8 17.9 18.6 28.9 ENT 18.5 18.7 17.4 18.8 23.4 16.9 18.8 17.1 19.6 16.8 14.1 16.6 19.6 15.8 16.5 23.0 RPL 17.8 18.0 17.0 18.1 21.4 16.4 17.9 16.4 18.7 15.7 13.6 15.6 19.2 15.0 15.6 22.0 Published in Transactions on Machine Learning Research (11/2022) C.4 Detailed results for the CIFAR10-C and UDA adaptation Table 24: Detailed results for each corruption along with mean error on CIFAR10-C as reported in Table 2 in the main paper. WRN-28-10 vanilla Baseline 53.0 41.2 44.7 18.5 49.0 22.3 24.4 18.1 25.0 11.2 6.7 17.4 16.2 28.0 22.4 26.5 BN adapt 20.8 17.6 22.7 8.1 28.4 10.9 9.2 14.2 13.0 8.7 6.8 8.5 13.5 12.1 21.0 14.4 ENT 18.5 15.9 20.6 7.8 25.5 10.6 8.5 13.1 12.3 8.3 6.9 8.0 12.6 11.1 18.9 13.3 RPL 19.1 21.4 16.3 8.1 26.4 8.9 10.9 13.7 12.9 6.9 13.1 19.7 8.2 11.4 8.7 13.7 WRN-40-2 AM Baseline 19.1 14.0 13.3 6.3 17.1 7.9 7.0 10.4 10.6 8.5 5.9 9.7 9.2 16.8 11.9 11.2 BN adapt 14.1 11.9 13.9 7.2 17.6 8.7 7.9 10.8 10.6 9.0 6.8 9.0 10.9 10.1 14.0 10.8 ENT 10.8 9.1 10.9 6.0 13.4 7.2 6.3 8.4 7.8 7.1 5.7 7.1 9.2 7.4 11.2 8.5 RPL 11.5 11.6 9.6 6.2 14.2 6.5 7.4 8.8 8.2 6.0 9.5 11.9 7.9 8.0 7.6 9.0 WRN-26-16 UDA-SS Baseline 26.0 24.7 19.3 22.4 56.2 32.4 32.1 31.7 31.2 26.6 15.8 20.4 26.3 21.5 28.9 27.7 BN adapt 20.5 19.0 15.6 13.5 43.1 19.4 18.3 23.1 21.2 16.2 12.8 14.1 20.9 16.7 23.4 19.9 ENT 16.9 16.7 12.3 11.3 37.6 15.6 14.8 18.3 18.2 13.4 10.8 11.9 17.9 14.4 20.9 16.7 RPL 18.1 17.1 13.2 11.9 41.5 17.3 16.1 20.4 19.1 14.5 11.8 12.7 18.8 18.1 22.6 18.2 WRN-26-16 vanilla Baseline 50.8 46.9 39.3 15.9 44.2 20.8 18.8 14.9 17.8 4.8 13.6 20.4 19.0 25.0 10.0 24.2 BN adapt 18.6 20.4 15.6 6.1 24.9 7.4 8.3 11.6 10.8 5.3 11.4 18.9 6.8 9.9 6.8 12.2 ENT 16.7 18.4 13.9 5.7 23.0 6.8 7.9 10.8 9.9 5.0 10.8 17.3 6.4 9.1 6.4 11.2 RPL 16.1 18.0 13.6 6.6 23.2 8.7 9.3 11.9 11.1 6.4 11.7 17.0 7.4 9.0 7.2 11.8 Table 25: Detailed results for the UDA methods reported in Table 2 of the main paper. Baseline BN adapt RPL ENT UDA CIFAR10 STL10, top1 error on target [%]( ) WRN-26-16 UDA-SS 28.7 24.6 22.9 21.8 WRN-26-16 DANN 25.0 25.0 24.0 23.9 UDA MNIST MNIST-M, top1 error on target [%]( ) WRN-26-16 UDA-SS 4.8 3.9 2.4 2.0 WRN-26-2 DANN 11.4 6.2 5.2 5.1 C.5 Ablation over the hyperparameter q for RPL For RPL, we must choose the hyperparameter q. We performed an ablation study over q and show results in Table 26, demonstrating that RPL is robust to the choice of q, with slight preference to higher values. Note: In the initial parameter sweep for this paper, we only compared q = 0.7 and q = 0.8. Given the result in Table 26, it could be interesting to re-run the models in Table 1 of the main paper with q = 0.9, which could yield another (small) improvement in m CE. Table 26: Image Net-C dev set m CE in %, vanilla Res Net50, batch size 96. We report the best score across a maximum of six adaptation epochs. q 0.5 0.6 0.7 0.8 0.9 m CE (dev) 49.5 49.3 49.2 49.2 49.1 C.6 Self-learning outperforms Test-Time Training (Sun et al., 2020) Sun et al. (2020) use a Res Net18 for their experiments on Image Net and only evaluate their method on severity 5 of IN-C. To enable a fair comparison, we trained a Res Net18 with both hard labeling and RPL and compare the efficacy of both methods to Test-Time Training in Table 27. For both hard labeling and RPL, Published in Transactions on Machine Learning Research (11/2022) we use the hyperparameters we found for the vanilla Res Net50 model and thus, we expect even better results for hyperparameters tuned on the vanilla Res Net18 model and following our general hyperparameter search protocol. While all methods (self-learning and TTT) improve the performance over a simple vanilla Res Net18, we note that even the very simple baseline using hard labeling already outperfoms Test-Time Training; further gains are possible with RPL. The result highlights the importance of simple baselines (like self-learning) when proposing new domain adaptation schemes. It is likely that many established DA techniques more complex than the basic self-learning techniques considered in this work will even further improve over TTT and other adaptation approaches developed exclusively in robustness settings. We report better accuracy numbers achieved with self-learning compared to online TTT, but note that they adapt to a single test sample while we make use of batches of data. Due to the single-image approach, they utilize GN instead of BN, and this may explain the performance gap to a certain degree as we show that BN is more effective for test-time adaptation compared to GN, even though the un-adapted BN models are less robust compared to the un-adapted GN models, as also noted by Sun et al. (2020). Further, Sun et al. (2020) optimize all model parameters while we only optimize the affine BN parameters which works better. Table 27: Comparison of hard-pseudo labeling and robust pseudo-labeling to Test-Time Training Sun et al. (2020): Top-1 error for a Res Net18 and severity 5 for all corruptions. Simple hard pseudo-labeling already outperforms TTT, robust pseudo labeling over multiple epochs yields additional gains. vanilla Res Net18 98.8 98.2 99.0 88.6 91.3 88.8 82.4 89.1 83.5 85.7 48.7 96.6 83.2 76.9 70.4 85.4 Test-Time Training 73.7 71.4 73.1 76.3 93.4 71.3 66.6 64.4 81.3 52.4 41.7 64.7 55.7 52.2 55.7 66.3 hard PL, (1 epoch) 73.2 70.8 73.6 76.5 75.6 63.9 56.1 59.0 65.9 48.4 39.7 85.2 50.4 47.0 51.5 62.5 ENT(1 epoch) 72.8 69.8 73.2 77.2 75.7 63.1 55.5 58.0 68.1 48.0 39.8 92.7 49.6 46.4 51.3 62.8 RPL (4 epochs) 71.3 68.3 71.7 76.2 75.6 61.5 54.4 56.9 67.1 47.3 39.3 93.2 48.9 45.7 50.4 61.9 C.7 Comparison to Meta Test-Time Training (Bartler et al., 2022) Bartler et al. (2022) report an error of 24.4% as their top result on CIFAR10-C which is far worse than our top results of 8.5% with an Aug Mix trained model, or 13.3% with a vanilla trained model, but a different architecture: they used a WRN-26-1 while we report results with a WRN-40-2. Thus, to make the comparison more fair, we tested our approach on their model architecture. A direct comparison is not straight-forward since Bartler et al. (2022) trained their models using Keras while our code-base is in Py Torch. Therefore, we first trained a baseline model in Py Torch on clean CIFAR10 using their architecture with the standard and widely used CIFAR10 training code available at https://github.com/kuangliu/pytorch-cifar (1.9k forks, 4.8k stars); the baseline test accuracy on clean CIFAR10 using the architecture of Bartler et al. (2022) is at 94.3%. As a second step, we adapted the baseline model with ENT and RPL on CIFAR10-C. Please see Table 28 for our results. BN adaptation is not possible for the model used by Bartler et al. (2022) since they use Group Norm instead of Batch Norm. Due to the usage of Group Norm, it is not evident whether the affine Group Norm parameters should be adapted, or all model parameters. Thus, we tested both, and report the results for both adaptation mechanisms. We find that adaptation of affine GN layers works better than full model adaptation, consistent with our results for the adaptation of BN layers. As the gains due to ENT and RPL seem lower than for our WRN-40-2 architecture, we hypothesize that the issue lies in the GN layers as the presence of GN instead of BN layers is the main difference between our WRN-40-2 and the new WRN-26-1 model. Therefore, we trained another WRN-26-1 model with BN layers instead of GN, and adapted this model using BN, ENT and RPL. The baseline accuracy on clean CIFAR10 of WRN-26-1-BN is 95.04%. Indeed, we find that adapting the model with BN instead of GN layers leads to much larger gains, which is consistent with findings by Schneider et al. (2020) who found that BN adaptation outperforms non-adapted models trained with GN. Thus, we conclude Published in Transactions on Machine Learning Research (11/2022) that self-learning techniques work better with models which have BN layers and less well with models with GN layers. For both model types (with GN or with BN layers), ENT works better than RPL which is consistent with our other results on small-scale datasets. Overall, our best result for the model architecture used by Bartler et al. (2022) (after replacing GN layers with BN layers) et al. is 13.1% which is much lower than their best result of 24.4%. Even when using GN layers, our best top-1 error is 18.0% which is significantly lower than the best result of Bartler et al. (2022). Table 28: Detailed results for our comparison to MT3 (Bartler et al., 2022) WRN-26-1-GN vanilla Baseline (Bartler et al., 2022) 50.5 47.2 56.1 23.7 51.7 24.3 26.3 25.6 34.4 28.1 13.5 25.0 27.4 55.8 29.8 34.6 adapted (Bartler et al., 2022) 30.1 29.5 41.8 15.6 33.7 22.8 18.7 20.2 18.8 24.1 13.8 22.4 23.7 27.6 22.7 24.4 WRN-26-1-GN vanilla Baseline [ours] 39.1 32.0 30.2 10.9 31.9 13.5 13.3 14.1 16.7 10.6 7.3 9.4 14.1 16.1 20.6 18.6 ENT [GN layers, ours] 41.6 30.7 32.8 9.7 30.9 11.8 11.4 13.5 14.9 10.0 7.2 8.9 13.2 13.3 19.8 18.0 RPL [GN layers, ours] 39.3 31.6 30.6 10.6 31.6 13.0 12.6 14.0 16.1 10.4 7.3 9.2 13.9 15.4 20.4 18.4 ENT [full model, ours] 61.2 46.0 37.6 9.1 30.2 11.1 10.4 13.1 14.4 10.0 7.2 8.7 13.1 11.9 19.7 20.3 RPL [full model, ours] 54.1 37.3 33.0 10.2 31.5 11.9 12.4 13.6 15.3 7.3 13.6 20.4 9.2 13.6 10.2 19.6 WRN-26-1-BN vanilla Baseline [ours] 55.7 44.8 43.1 15.5 44.2 20.5 21.1 17.2 20.7 6.6 15.6 21.3 23.6 24.7 11.8 25.8 BN adapt [ours] 28.0 28.9 24.2 10.5 31.5 12.7 14.0 18.4 18.1 9.4 17.1 26.9 13.2 15.3 12.7 18.7 ENT [BN layers, ours] 17.9 19.6 14.9 8.1 23.2 9.1 10.4 13.1 13.2 7.2 13.2 18.0 9.9 10.5 8.9 13.1 RPL [BN layers, ours] 21.3 22.0 17.9 9.3 26.6 10.5 11.8 15.1 15.0 7.8 14.7 20.8 12.1 11.8 10.4 15.1 C.8 Effect of batch size and linear learning rate scaling How is self-learning performance affected by batch size constraints? We compare the effect of different batch sizes and linear learning rate scaling. In general, we found that affine adaptation experiments on Res Net50 scale can be run with batch size 128 on a Nvidia V100 GPU (16GB), while only batch size 96 experiments are possible on RTX 2080 GPUs. The results in Table 29 show that for a Res Net50 model, higher batch size yields a generally better performance. Table 29: Image Net-C dev set m CE for various batch sizes with linear learning rate scaling. All results are computed for a vanilla Res Net50 model using RPL with q = 0.8, reporting the best score across a maximium of six adaptation epochs. batch size 16 32 64 80 96 128 learning rate ( 10 3) 0.125 0.250 0.500 0.625 0.750 1 dev m CE 53.8 51.0 49.7 49.3 49.2 48.9 C.9 Performance over different seeds in a Res Net50 on Image Net-C To limit the amount of compute, we ran RPL and ENT for our vanilla Res Net50 model three times with the optimal hyperparameters. The averaged results, displayed as mean (unbiased std) are: Table 30: Image Net-C performance for three seeds on a Res Net50 for ENT and RPL. Res Net50 + self-learning m CE on IN-C dev [%] m CE on IN-C test [%] ENT 50.0 (0.04) 51.6 (0.04) RPL 48.9 (0.02) 50.5 (0.03) Published in Transactions on Machine Learning Research (11/2022) C.10 Self-learning as continuous test-test adaptation We test our method on continuous test-time adaptation where the model adapts to a continuous stream of data from the same domain. In Fig. 7, we display the error of the Noisy Student L2 model while it is being adapted to Image Net-C and Image Net-R. The model performance improves as the model sees more data from the new domain. We differentiate continuous test-time adaptation from the online test-time adaptation setting (Zhang et al., 2021) where the model is adapted to each test sample individually, and reset after each test sample. 0 1 2 3 4 5 Samples [ 104] (i) Image Net-C Baseline ENT RPL 0 1 2 3 Samples [ 104] Top-1 Error [%] (ii) Image Net-R Figure 7: Evolution of error during online adaptation for Efficient Net-L2. Published in Transactions on Machine Learning Research (11/2022) Table 31: Statistics of one-to-many mappings from IN-D to Image Net. Number of IN classes one IN-D class is mapped to 1 2 3 4 5 6 7 8 13 28 39 132 Frequency of these mappings 102 32 13 3 5 1 1 2 1 2 1 1 D Detailed and additional Results on IN-D D.1 Detailed protocol for label mapping from Domain Net to Image Net The mapping was first done by comparing the class labels in Domain Net and the synset labels on Image Net. Afterwards, the resulting label maps were cleaned manually, because simply comparing class label strings resulted in imperfect matches. For example, images of the class hot dog in Domain Net were mapped to the class dog in IN. Another issue is that IN synset labels of different animal species do not contain the animal name in the text label, e.g., the class orangutan, orang, orangutang, Pongo pygmaeus does not contain the word monkey and we had to add this class to the hierarchical class monkey manually. We verified the mappings by investigating the class-confusion matrix of the true Domain Net class and the predicted Image Net classes remapped to Domain Net on the Real domain, and checked that the predictions lay on the main diagonal, indicating that Image Net classes have not been forgotten. The statistics for the mappings are shown in Table 31. Most IN-D classes (102) are mapped to one single Image Net class. A few IN-D classes are mapped to more than 20 Image Net classes: the IN-D classes monkey and snake are mapped to 28 Image Net monkey and snake species classes, the IN-D class bird is mapped to 39 Image Net bird species classes, and the IN-D class dog is mapped to 132 Image Net dog breed classes. We have considered max-pooling predictions across all sub-classes according to the Image Net class hierarchy following the approach of Taori et al. (2020) for Youtube-BB and Image Net-Vid (Recht et al., 2019). However, Radford et al. (2021) note that the resulting mappings are sometimes much less than perfect , thus, we decided to clean the mappings ourselves to increase the mappings quality. The full dictionary of the mappings will be released alongside the code. D.2 Evaluation protocol on IN-D The domains in IN-D differ in terms of their difficulty for the studied models. Therefore, to calculate an aggregate score, we propose normalizing the error rates by the error achieved by Alex Net on the respective domains to calculate the mean error, following the approach in Hendrycks & Dietterich (2019) for IN-C. This way, we obtain the aggregate score mean Domain Error (m DE) by calculating the mean over different domains, DEf d = Ef d EAlex Net d , m DE = 1 d=1 Ef d , (21) where Ef d is the top-1 error of a classifier f on domain d. Leave-one-out-cross-validation For all IN-D results we report in this paper, we chose the hyperparameters on the IN-C dev set. We tried a different model selection scheme on IN-D as a control experiment with Leave one out cross-validation (L1out CV): with a round-robin procedure, we choose the hyperparameters for the test domain on all other domains. We select the same hyperparameters as when tuning on the dev set: For the Res Net50 model, we select over the number of training epochs (with a maximum of 7 training epochs) and search for the optimal learning rate in the set [0.01, 0.001, 0.0001]. For the Efficient Net-L2 model, we train only for one epoch as before and select the optimal learning rate in the set [4.6 10 3, 4.6 10 4, 4.6 10 5, 4.6 10 6]. This model selection leads to worse results both for the Res Net50 and the Efficient Net-L2 models, highlighting the robustness of our model selection process, see Table 32. Published in Transactions on Machine Learning Research (11/2022) Table 32: m DE in % on IN-D for different model selection strategies. model model selection L1out CV IN-C dev Res Net50 RPLq=0.8 81.3 76.1 Res Net50 ENT 82.4 77.3 Efficient Net-L2 ENT 69.2 66.8 Efficient Net-L2 RPLq=0.8 69.1 67.2 D.3 Detailed results for robust Res Net50 models on IN-D We show detailed results for all models on IN-D for vanilla evaluation (Table 33) BN adaptation (Table 34), RPLq=0.8 (Table 35) and ENT(Table 36). For RPLq=0.8 and ENT, we use the same hyperparameters that we chose on our IN-C dev set. This means we train the models for 5 epochs with RPLq=0.8 and for one epoch with ENT. We evaluate the pre-trained and public checkpoints of SIN (Geirhos et al., 2019), ANT (Rusak et al., 2020), ANT+SIN (Rusak et al., 2020), Aug Mix (Hendrycks et al., 2020b), Deep Augment (Hendrycks et al., 2020a) and Deep Aug+Augmix (Hendrycks et al., 2020a) in the following tables. Table 33: Top-1 error on IN-D in % as obtained by robust Res Net50 models. For reference, we also show the m CE on IN-C and the top-1 error on IN-R. See main test for model references. Model Clipart Infograph Painting Quickdraw Real Sketch m DE IN-C IN-R vanilla 76.0 89.6 65.1 99.2 40.1 82.0 88.2 76.7 63.9 SIN 71.3 88.6 62.6 97.5 40.6 77.0 85.6 69.3 58.5 ANT 73.4 88.9 63.3 99.2 39.9 80.8 86.9 62.4 61.0 ANT+SIN 68.4 88.6 60.6 95.5 40.8 70.3 83.1 60.7 53.7 Aug Mix 70.8 88.6 62.1 99.1 39.0 78.5 85.4 65.3 58.9 Deep Augment 72.0 88.8 61.4 98.9 39.4 78.5 85.6 60.4 57.8 Deep Aug+Augmix 68.4 88.1 58.7 98.2 39.2 75.2 83.4 53.6 53.2 Table 34: Top1 error on IN-D in % as obtained by state-of-the-art robust Res Net50 models and batch norm adaptation, with a batch size of 128. See main text for model references. Model Clipart Infograph Painting Quickdraw Real Sketch m DE vanilla 70.2 88.2 63.5 97.8 41.1 78.3 80.2 SIN 67.3 89.7 62.2 97.2 44.0 75.2 79.6 ANT 69.2 89.4 63.0 97.5 42.9 79.5 80.7 ANT+SIN 64.9 88.2 60.0 96.8 42.6 73.0 77.8 Aug Mix 66.9 88.1 61.2 97.1 40.4 75.0 78.4 Deep Augment 66.6 89.7 60.0 97.2 42.5 75.1 78.8 Deep Aug+Augmix 61.9 85.7 57.5 95.3 40.2 69.2 74.9 Table 35: Top-1 error on IN-D in % as obtained by state-of-the-art robust Res Net50 models and RPLq=0.8. See main text for model references. Model Clipart Infograph Painting Quickdraw Real Sketch m DE vanilla 63.6 85.1 57.8 99.8 37.3 73.0 76.1 SIN 60.8 86.4 56.0 99.0 37.8 67.0 76.8 ANT 63.4 86.3 57.7 99.2 37.7 71.0 78.1 ANT+SIN 61.5 86.4 56.8 97.0 39.0 67.1 76.1 Aug Mix 59.7 83.4 54.1 98.2 35.6 70.1 74.6 Deep Augment 58.1 84.6 53.3 99.0 36.2 64.2 74.8 Deep Aug+Augmix 57.0 83.2 53.4 99.1 36.5 61.3 72.6 Published in Transactions on Machine Learning Research (11/2022) Table 36: Top-1 error on IN-D in % as obtained by state-of-the-art robust Res Net50 models and ENT. See main text for references to the used models. Model Clipart Infograph Painting Quickdraw Real Sketch m DE vanilla 65.1 85.8 59.2 98.5 38.4 75.8 77.3 SIN 62.1 87.0 57.3 99.1 39.0 68.6 75.5 ANT 64.2 86.9 58.7 97.1 38.8 72.8 76.5 ANT+SIN 62.2 86.8 57.7 95.8 40.1 68.7 75.2 Aug Mix 60.2 84.6 55.8 97.6 36.8 72.0 74.4 Deep Augment 59.5 85.7 54.4 98.0 37.1 66.4 73.3 Deep Aug+Augmix 58.4 84.3 54.7 98.5 38.1 63.6 72.7 Table 37: m DE on IN-D in % as obtained by robust Res Net50 models with a baseline evaluation, batch norm adaptation, RPLq=0.8 and ENT. See main text for model references. m DE on IN-D ( ) Model Baseline BN adapt RPLq=0.8 ENT vanilla 88.2 80.2 76.1 77.3 SIN 85.6 79.6 76.8 75.5 ANT 86.9 80.7 78.1 76.5 ANT+SIN 83.1 77.8 76.1 75.2 Aug Mix 85.4 78.4 74.6 74.4 Deep Augment 85.6 78.8 74.8 73.3 Deep Augment+Augmix 83.4 74.9 72.6 72.7 The summary results for all models are shown in Table 37. We show the top-1 error for the different IN-D domains versus training epochs for a vanilla Res Net50 in Fig. 8. We indicate the epochs 1 and 5 at which we extract the errors with dashed black lines. Figure 8: Top-1 error for the different IN-D domains for a Res Net50 and training with RPLq=0.8 and ENT. We indicate the epochs at which we extract the test errors by the dashed black lines (epoch 1 for ENT and epoch 5 for RPLq=0.8). Published in Transactions on Machine Learning Research (11/2022) D.4 Detailed results for the Efficient Net-L2 Noisy Student model on IN-D We show the detailed results for the Efficient Net-L2 Noisy Student model on IN-D in Table 38. Table 38: Top-1 error ( ) on IN-D in % for Efficient Net-L2 Domain Baseline ENT RPL Clipart 45.0 39.8 37.9 Infograph 77.9 91.3 94.3 Painting 42.7 41.7 40.9 Quickdraw 98.4 99.4 99.4 Real 29.2 28.7 27.9 Sketch 56.4 48.0 51.5 m DE 67.2 66.8 67.2 D.5 Detailed results on the error analysis on IN-D Analysing frequently predicted classes We analyze the most frequently predicted classes on IN-D by a vanilla Res Net50 and show the results in Fig. 9. The colors of the bars indicate whether the predicted class is part of the IN-D dataset: blue indicates that the class appear in the IN-D dataset, while orange means that the class is not present in IN-D. We make several interesting observations: First, we find most errors interpretable: it makes sense that a Res Net50 assigns the label comic book to images from the Clipart or Painting domains, or website to images from the Infograph domain, or envelope to images from the Sketch domain. Second, on the hard domain Quickdraw , the Res Net50 mostly predicts non-sensical classes that are not in IN-D, mirroring its almost chance performance on this domain. Third, we find no systematic errors on the Real domain which is expected since this domain should be similar to IN. Analyzing the correlation between the performance on IN-C/IN-R and IN-D We show the Spearman s rank correlation coefficients for errors on Image Net-D correlated to errors on Image Net-R and Image Net-C for robust Res Net50 models in Fig. 9. For this correlation analysis, we take the error numbers from Table 33. We find the correlation to be high between most domains in IN-D and IN-R which is expected since the distribution shift between IN-R and IN is similar to the distribution shift between IN-D and Image Net. The only domain where the Spearman s rank correlation coefficient is higher for IN-C is the Real domain which can be explained with IN-C being closer to real-world data than IN-R. Thus, we find that the Spearman s rank correlation coefficient reflects the similarity between different datasets. Filtering predictions on IN-D that cannot be mapped to Image Net We perform a second analysis: We filter the predicted labels according to whether they can be mapped to IN-D and report the filtered top-1 errors as well as the percentage of filtered out inputs in Table 39. We note that for the domains infograph and quickdraw , the Res Net50 predicts labels that cannot be mapped to IN-D in over 70% of all cases, highlighting the hardness of these two domains. Filtering labels and predictions on IN that cannot be mapped to Image Net-D To test for possible class-bias effects, we test the performance of a Res Net50 model on Image Net classes that can be mapped to IN-D and report the results in Table 39. First, we map IN labels to IN-D to make the setting as similar as possible to our experiments on IN-D and report the top-1 error (12.1%). This error is significantly lower compared to the top-1 error a Res Net50 obtains following the standard evaluation protocol (23.9%). This can be explained by the simplification of the task: While in IN there are 39 bird classes, these are all mapped to the same hierarchical class in IN-D. Therefore, the classes in IN-D are more dissimilar from each other than in IN. Additionally, there are only 164 IN-D classes compared to the 1000 IN classes, raising the chance level prediction. Published in Transactions on Machine Learning Research (11/2022) Table 39: top-1 error on IN and different IN-D domains for different settings: left column: predicted labels that cannot be mapped to IN-D are filtered out, right column: percentage of filtered out labels. Dataset top-1 error on filtered labels in % percentage of rejected inputs IN val 13.4 52.7 IN-D real 17.2 27.6 IN-D clipart 59.0 59.0 IN-D infograph 59.3 74.6 IN-D painting 39.5 42.4 IN-D quickdraw 96.7 76.1 IN-D sketch 65.6 47.9 Figure 9: Systematic predictions of a vanilla Res Net50 on IN-D for different domains. The colors of the bars indicate whether the predicted class is part of the IN-D dataset: blue indicates that the class appear in the IN-D dataset, while orange means that the class is not present in IN-D. The errors on IN-R are strongly correlated to errors on IN-D for most domains, except for the Real domain. If we further only accept predictions that can be mapped to IN-D, the top-1 error is slightly increased to 13.4%. In total, about 52.7% of all images in the IN validation set cannot be mapped to IN-D. D.6 Top-1 error on IN-D for Alex Net We report the top-1 error numbers on different IN-D as achieved by Alex Net in Table 40. We used these numbers for normalization when calculating m DE. Published in Transactions on Machine Learning Research (11/2022) Table 40: top-1 error on IN-D by Alex Net which was used for normalization. Dataset top-1 error in % IN-D real 54.887 IN-D clipart 84.010 IN-D infograph 95.072 IN-D painting 79.080 IN-D quickdraw 99.745 IN-D sketch 91.189 Published in Transactions on Machine Learning Research (11/2022) E Additional experiments E.1 Beyond Image Net classes: Self-learning on WILDS The WILDS benchmark (Koh et al., 2021) is comprised of ten tasks to test domain generalization, subpopulation shift, and combinations thereof. In contrast to the setting considered here, many of the datasets in WILDS mix several 10s or 100s domains during test time. The Camelyon17 dataset in WILDS contains histopathological images, with the labels being binary indicators of whether the central 32 32 region contains any tumor tissue; the domain identifies the hospital that the patch was taken from. Camelyon17 contains three different test splits with different domains and varying difficulty levels. For evaluation, we took the pretrained checkpoint from worksheets.codalab.org/worksheets/ 0x00d14c55993548a1823a710642f6d608 (camelyon17_erm_densenet121_seed0) for a Dense Net121 model (Huang et al., 2017) and verified the reported baseline performance numbers. We adapt the models using ENT or RPL for a maximum of 10 epochs using learning rates {3 10 5, 3 10 4, . . . 3 10 1}. The best hyperparameter is selected according to OOD Validation accuracy. The Rx Rx1 dataset in WILDS contains RGB images of cells obtained by fluorescent microscopy, with the labels indicating which of the 1,139 genetic treatments (including no treatment) the cells received; the domain identifies the batch in which the imaging experiment was run. The Rx Rx1 dataset contains three test splits, however, unlike Camelyon17, in all of the splits the domains are mixed. For evaluation, we took the pretrained checkpoint from worksheets.codalab.org/bundles/0x7d33860545b64acca5047396d42c0ea0 for a Res Net50 model and verified the reported baseline performance numbers. We adapt the models using ENT or RPL for a maximum of 10 epochs using base learning rates {6.25 10 6, 6.25 10 5, . . . 6.25 10 2}, which are scaled to the admissible batch size for single GPU adaptation using linear scaling. The best hyperparameter is selected according to OOD Validation accuracy. Table 41: Self-learning can improve performance on WILDS if a systematic shift is present on Camelyon17, the ood validation and test sets are different hospitals, for example. On datasets like Rx Rx1 and FMo W, we do not see an improvement, most likely because the ood domains are shuffled, and a limited amount of images exist for each test domain. Top-1 accuracy [%] Validation (ID) Validation (OOD) Test (OOD) Camelyon17 Baseline 81.4 88.7 63.1 BN adapt 97.8 (+16.4) 90.9 (+2.2) 88.0 (+24.9) ENT 97.6 (+16.2) 92.7 (+4.0) 91.6 (+28.5) RPL 97.6 (+16.2) 93.0 (+4.3) 91.0 (+27.9) Rx Rx1 Baseline 35.9 19.1 29.7 BN adapt 35.0 (-0.9) 19.1 (0.0) 29.4 (-0.3) ENT 34.8 (-1.1) 19.2 (+0.1) 29.4 (-0.3) RPL 34.8 (-1.1) 19.2 (+0.1) 29.4 (-0.3) FMo W Baseline 60.5 59.2 52.9 BN adapt 59.9 (-0.6) 57.6 (-1.6) 51.8 (-1.1) ENT 59.9 (-0.6) 58.5 (-0.7) 52.2 (-0.7) RPL 59.8 (-0.7) 58.6 (-0.6) 52.1 (-0.8) The FMo W dataset in WILDS contains RGB satellite images, with the labels being one of 62 building or land use categories; the domain specifies the year in which the image was taken and its geographical region (Africa, the Americas, Oceania, Asia, or Europe). The FMo W dataset contains four test splits for different time periods, for which all regions are mixed together. For evaluation, we took the pretrained checkpoint from //worksheets.codalab.org/bundles/0x20182ee424504e4a916fe88c91afd5a2 for a Dense Net121 model and verified the reported baseline performance numbers. We adapt the models using ENT or RPL for a maximum Published in Transactions on Machine Learning Research (11/2022) of 10 epochs using learning rates {5.0 10 6, 5.0 10 5, . . . 5.0 10 2}. The best hyperparameter is selected according to OOD Validation accuracy. While we see improvements on Camelyon17, neither BN adaptation nor self-learning can improve performance on Rx Rx1 or FMo W. Initial experiments on Poverty Map and i Wilds Cam also do not show improvements with self-learning. We hypothesize that the reason lies in the mixing of the domains: Both BN adaptation and our self-learning methods work best on systematic domain shifts. These results support our claim that self-learning is effective, while showing the important limitation when applied to more diverse shifts. E.2 Small improvements on Big Transfer models with Group normalization layers We evaluated Big Transfer models (Kolesnikov et al., 2020) provided by the timm library (Wightman, 2019). A difference to the Res Net50, Res Ne Xt101 and Efficient Net models is the use of group normalization layers, which might influence the optimal method for adaptation for this evaluation, we followed our typical protocol as performed on Res Net50 models, and used affine adaptation. For affine adaptation, a distilled Big Transfer Res Net50 model improves from 49.6 % to 48.4 % m CE on the Image Net-C development set, and from 55.0 % to 54.4 % m CE on the Image Net-C test set when using RPL (q = 0.8) for adaptation, at learning rate 7.5 10 4 at batch size 96 after a single adaptation epoch. Entropy minimization did not further improve results on the Image Net-C test set. An ablation over learning rates and epochs on the dev set is shown in Table 42, the final results are summarized in Table 43. Table 42: m CE in % on the IN-C dev set for ENT and RPL for different numbers of training epochs when adapting the affine batch norm parameters of a Big Transfer Res Net50 model. criterion ENT RPL lr, 7.5 10 5 10 4 10 3 10 5 10 4 10 3 0 49.63 49.63 49.63 49.63 49.63 49.63 1 49.44 50.42 52.59 49.54 48.89 48.95 2 49.26 50.27 56.47 49.47 48.35 50.77 3 49.08 52.18 60.06 49.39 48.93 51.45 4 48.91 52.03 60.50 49.31 50.01 51.53 5 48.80 51.97 62.91 49.24 49.96 51.34 6 48.83 52.10 62.96 49.16 49.71 51.19 7 48.83 52.10 62.96 49.16 49.71 51.19 Table 43: m CE in % on the IN-C dev set for ENT and RPL for different numbers of training epochs when adapting the affine batch norm parameters of a Big Transfer Res Net50 model. dev m CE test m CE Baseline 49.63 55.03 ENT 48.80 56.36 RPL 48.35 54.41 E.3 Can Self-Learning improve over Self-Learning based UDA? An interesting question is whether test-time adaptation with self-learning can improve upon self-learning based UDA methods. To investigate this question, we build upon French et al. (2018) and their released code base at github.com/Britefury/self-ensemble-visual-domain-adapt. We trained the Baseline models from scratch using the provided shell scripts with the default hyperparameters and verified the reported performance. For adaptation, we tested BN adaptation, ENT, RPL, as well as continuing to train in exactly the setup of French et al. (2018), but without the supervised loss. For the different losses, we adapt the models for a maximum of 10 epochs using learning rates {1 10 5, 1 10 4, . . . , 1 10 1}. Note that for this experiment, in contrast to any other result in this paper, we purposefully do not perform proper hyperparameter selection based on a validation dataset instead we report the best accuracy across all tested epochs and learning rates to give an upper bound on the achievable performance for test-time adaptation. As highlighted in Table 44, none of the four tested variants is able to meaningfully improve over the baseline, corroborating our initial hypothesis that self-learning within a full UDA setting is the optimal strategy, if dataset size and compute permits. On the other hand, results like the teacher refinement step in DIRT-T Published in Transactions on Machine Learning Research (11/2022) (Shu et al., 2018) show that with additional modifications in the loss function, it might be possible to improve over standard UDA with additional adaptation at test time. Table 44: Test-time adaptation marginally improves over self-ensembling. Baseline BN adapt ENT RPL Self-ensembling loss MNIST SVHN MT+TF 33.88 34.44 34.87 35.09 33.27 MT+CT* 32.62 34.11 34.25 34.21 33.36 MT+CT+TF 41.59 41.93 41.95 41.95 42.70 MT+CT+TFA 30.55 32.53 32.54 32.55 30.84 SVHN-specific aug. 97.05 96.82 96.91 96.87 97.12 MNIST USPS MT+TF 98.01 97.91 97.96 97.91 98.16 MT+CT* 88.34 88.39 88.54 88.39 88.44 MT+CT+TF 98.36 98.41 98.41 98.41 98.50 MT+CT+TFA 98.45 98.45 98.45 98.45 98.61 SVHN MNIST MT+TF 98.49 98.47 98.49 98.47 99.40 MT+CT* 88.34 88.36 88.36 88.36 89.36 MT+CT+TF 99.51 99.49 99.5 99.49 99.57 MT+CT+TFA 99.56 99.57 99.57 99.57 99.58 SVHN-specific aug. 99.52 99.49 99.5 99.49 99.65 USPS MNIST MT+TF 92.79 92.62 92.62 92.66 93.08 MT+CT* 99.11 99.13 99.14 99.13 99.21 MT+CT+TF 99.41 99.42 99.45 99.42 99.52 MT+CT+TFA 99.48 99.54 99.57 99.54 99.54 Published in Transactions on Machine Learning Research (11/2022) F Software stack We use different open source software packages for our experiments, most notably Docker (Merkel, 2014), scipy and numpy (Virtanen et al., 2020), GNU parallel (Tange, 2011), Tensorflow (Abadi et al., 2016), Py Torch (Paszke et al., 2017), timm (Wightman, 2019), Self-ensembling for visual domain adaptation (French et al., 2018), the WILDS benchmark (Koh et al., 2021), and torchvision (Marcel & Rodriguez, 2010).