# seed_selfsupervised_distillation_for_visual_representation__c8e0960e.pdf Published as a conference paper at ICLR 2021 SEED: SELF-SUPERVISED DISTILLATION FOR VISUAL REPRESENTATION Zhiyuan Fang , Jianfeng Wang , Lijuan Wang , Lei Zhang , Yezhou Yang , Zicheng Liu Arizona State University, Microsoft Corporation {zy.fang, yz.yang}@asu.edu {jianfw, lijuanw, leizhang, zliu}@microsoft.com This paper is concerned with self-supervised learning for small models. The problem is motivated by our empirical studies that while the widely used contrastive self-supervised learning method has shown great progress on large model training, it does not work well for small models. To address this problem, we propose a new learning paradigm, named SElf-Sup Ervised Distillation (SEED), where we leverage a larger network (as Teacher) to transfer its representational knowledge into a smaller architecture (as Student) in a self-supervised fashion. Instead of directly learning from unlabeled data, we train a student encoder to mimic the similarity score distribution inferred by a teacher over a set of instances. We show that SEED dramatically boosts the performance of small networks on downstream tasks. Compared with self-supervised baselines, SEED improves the top-1 accuracy from 42.2% to 67.6% on Efficient Net-B0 and from 36.3% to 68.2% on Mobile Net V3-Large on the Image Net-1k dataset. 1 INTRODUCTION 5 10 15 20 25 Numbers of Parameters (Millions) Image Net Top-1 Accuracy (%) SEED Eff-B1 Res-18 Res-34 Figure 1: SEED vs. Mo Co-V2 (Chen et al., 2020c)) on Image Net-1K linear probe accuracy. The vertical axis is the top-1 accuracy and the horizontal axis is the number of learnable parameters for different network architectures. Directly applying self-supervised contrastive learning (Mo Co-V2) does not work well for smaller architectures, while our method (SEED) leads to dramatic performance boost. Details of the setting can be found in Section 4. The burgeoning studies and success on self-supervised learning (SSL) for visual representation are mainly marked by its extraordinary potency of learning from unlabeled data at scale. Accompanying with the SSL is its phenomenal benefit of obtaining task-agnostic representations while allowing the training to dispense with prohibitively expensive data labeling. Major ramifications of visual SSL include pretext tasks (Noroozi & Favaro, 2016; Zhang et al., 2016; Gidaris et al., 2018; Zhang et al., 2019; Feng et al., 2019), contrastive representation learning (Wu et al., 2018; He et al., 2020; Chen et al., 2020a), online/offline clustering (Yang et al., 2016; Caron et al., 2018; Li et al., 2020; Caron et al., 2020; Grill et al., 2020), etc. Among them, several recent works (He et al., 2020; Chen et al., 2020a; Caron et al., 2020) have achieved comparable or even better accuracy than the supervised pre-training when transferring to downstream tasks, e.g. semi-supervised classification, object detection. The aforementioned top-performing SSL algorithms all involve large networks (e.g., Res Net-50 (He et al., 2016) or larger), with, however, little attention on small networks. Empirically, we find that existing techniques like contrastive learning do not work well on small networks. For instance, the linear probe top-1 accuracy on Image Net using Mo Co-V2 (Chen et al., 2020c) is only 36.3% with Mobile Net V3-Large (see Figure 1), which is much lower compared with its supervised training accuracy Published as a conference paper at ICLR 2021 75.2% (Howard et al., 2019). For Efficient Net-B0, the accuracy is 42.2% compared with its supervised training accuracy 77.1% (Tan & Le, 2019). We conjecture that this is because smaller models with fewer parameters cannot effectively learn instance level discriminative representation with large amount of data. To address this challenge, we inject knowledge distillation (KD) (Buciluˇa et al., 2006; Hinton et al., 2015) into self-supervised learning and propose self-supervised distillation (dubbed as SEED) as a new learning paradigm. That is, train the larger, and distill to the smaller both in self-supervised manner. Instead of directly conducting self-supervised training on a smaller model, SEED first trains a large model (as the teacher) in a self-supervised way, and then distills the knowledge to the smaller model (as the student). Note that the conventional distillation is for supervised learning, while the distillation here is in the self-supervised setting without any labeled data. Supervised distillation can be formulated as training a student to mimic the probability mass function over classes predicted by a teacher model. In unsupervised knowledge distillation setting, however, the distribution over classes is not directly attainable. Therefore, we propose a simple yet effective self-supervised distillation method. Similar to (He et al., 2020; Wu et al., 2018), we maintain a queue of data samples. Given an instance, we first use the teacher network to obtain its similarity scores with all the data samples in the queue as well as the instance itself. Then the student encoder is trained to mimic the similarity score distribution inferred by the teacher over these data samples. The simplicity and flexibility that SEED brings are self-evident. 1) It does not require any clustering/prototypical computing procedure to retrieve the pseudo-labels or latent classes. 2) The teacher model can be pre-trained with any advanced SSL approach, e.g., Mo Co-V2 (Chen et al., 2020c), Sim CLR (Chen et al., 2020a), SWAV (Caron et al., 2020). 3) The knowledge can be distilled to any target small networks (either shallower, thinner, or totally different architectures). To demonstrate the effectiveness, we comprehensively evaluate the learned representations on series of downstream tasks, e.g., fully/semi-supervised classification, object detection, and also assess the transferability to other domains. For example, on Image Net-1k dataset, SEED improves the linear probe accuracy of Efficient Net-B0 from 42.2% to 67.6% (a gain over 25%), and Mobile Net-V3 from 36.3% to 68.2% (a gain over 31%) compared to Mo Co-V2 baselines, as shown in Figure 1 and Section 4. Our contributions can be summarized as follows: We are the first to address the problem of self-supervised visual representation learning for small models. We propose a self-supervised distillation (SEED) technique to transfer knowledge from a large model to a small model without any labeled data. With the proposed distillation technique (SEED), we significantly improve the state-of-theart SSL performance on small models. We exhaustively compare a variety of distillation strategies to show the validity of SEED under multiple settings. 2 RELATED WORK Among the recent literature in self-supervised learning, contrastive based approaches show prominent results on downstream tasks. Majority of the techniques along this direction are stemming from noise-contrastive estimation (Gutmann & Hyvärinen, 2010) where the latent distribution is estimated by contrasting with randomly or artificially generated noises. Oord et al. (2018) first proposed Info-NCE to learn image representations by predicting the future using an auto-regressive model for unsupervised learning. Follow-up works include improving the efficiency (Hénaff et al., 2019), and using multi-view as positive samples (Tian et al., 2019b). As these approaches can only have the access to limited negative instances, Wu et al. (2018) designed a memory-bank to store the previously seen random representations as negative samples, and treat each of them as independent categories (instance discrimination). However, this approach also comes with a deficiency that the previously stored vectors are inconsistent with the recently computed representations during the earlier stage of pre-training. Chen et al. (2020a) mitigate this issue by sampling negative samples from a large batch. Concurrently, He et al. (2020) improve the memory-bank based method and propose to use Published as a conference paper at ICLR 2021 the momentum updated encoder for the remission of representation inconsistency. Other techniques include Misra & Maaten (2020) that combines the pretext-invariant objective loss with contrastive learning, and Wang & Isola (2020) that decomposes contrastive loss into alignment and uniformity objectiveness. Knowledge distillation (Hinton et al., 2015) aims to transfer knowledge from a cumbersome model to a smaller one without losing too much generalization power, which is also well investigated in model compression (Buciluˇa et al., 2006). Instead of mimicking the teacher s output logit, attention transfer (Zagoruyko & Komodakis, 2016) formulates knowledge distillation on attention maps. Similarly, works in (Ahn et al., 2019; Yim et al., 2017; Koratana et al., 2019; Huang & Wang, 2017) have utilized different learning objectives including consistency on feature maps, consistency on probability mass function, and maximizing the mutual information. CRD (Tian et al., 2019a), which is derived from CMC (Tian et al., 2019b), optimizes the student network by a similar objective to Oord et al. (2018) using a derived lower bound on mutual information. However, the aforementioned efforts all focus on task-specific distillation (e.g., image classification) during the fine-tuning phase rather than a task-agnostic distillation in the pre-training phase for the representation learning. Several works on natural language pre-training proposed to leverage knowledge distillation for a smaller yet stronger small models. For instances, Distill Bert (Sanh et al., 2019), Tiny Bert (Jiao et al., 2019), and Mobile Bert (Sun et al., 2020), have used knowledge distillation for model compression and shown their validity on multiple downstream tasks. Similar works also emphasize the value of smaller and faster models for language representation learning by leveraging knowledge distillation (Turc et al., 2019; Sun et al., 2019). These works all demonstrate the effectiveness of knowledge distillation for language representation learning in small models, while are not extended to the pre-training for visual representations. Notably, a recent concurrent work Comp Ress (Abbasi Koohpayegani et al., 2020) also point out the importance to develop better SSL method for smaller models. SEED closely relates to the above techniques but aims to facilitate visual representation learning during pre-training phase using distillation technique for small models, which as far as we know has not yet been investigated. 3.1 PRELIMINARY ON KNOWLEDGE DISTILLATION Knowledge distillation (Hinton et al., 2015; Buciluˇa et al., 2006) is an effective technique to transfer knowledge from a strong teacher network to a target student network. The training task can be generalized as the following formulation: ˆθS = arg min θS i Lsup(xi, θS, yi) + Ldistill(xi, θS, θT ), (1) where xi is an image, yi is the corresponding annotation, θS is the parameter set for the student network, and θT is the set for the teacher network. The loss Lsup is the alignment error between the network prediction and the annotation. For example in image classification task (Mishra & Marr, 2017; Shen & Savvides, 2020; Polino et al., 2018; Cho & Hariharan, 2019), it is normally a cross entropy loss. For object detection (Liu et al., 2019; Chen et al., 2017), it includes bounding box regression as well. The loss of Ldistill is the mimic error of the student network towards a pre-trained teacher network. For example in (Hinton et al., 2015), the teacher signal comes from the softmax prediction of multiple large-scale networks and the loss is measured by the Kullback Leibler divergence. In Romero et al. (2014), the task is to align the intermediate feature map values and to minimize the squared l2 distance. The effectiveness has been well demonstrated in the supervised setting with labeled data, but remains unknown for the unsupervised setting, which is our focus. 3.2 SELF-SUPERVISED DISTILLATION FOR VISUAL REPRESENTATION Different from supervised distillation, SEED aims to transfer knowledge from a large model to a small model without requiring labeled data, so that the learned representations in small model can be used for downstream tasks. Inspired by contrastive SSL, we formulate a simple approach for the distillation on the basis of instance similarity distribution over a contrastive instance queue. Similar to He et al. (2020), we maintain an instance queue for storing data samples encoding output from the Published as a conference paper at ICLR 2021 Instance Queue: D Inner product Cross-Entropy 𝐱 Teacher Probability Pre-trained and Frozen 𝐃+ Student Probability Figure 2: Illustration of our self-supervised distillation pipeline. The teacher encoder is pre-trained by SSL and kept frozen during the distillation. The student encoder is trained by minimizing the cross entropy of probabilities from teacher & student for an augmented view of an image, computed over a dynamically maintained queue. teacher. Given a new sample, we compute its similarity scores with all the samples in the queue using both the teacher and the student models. We require that the similarity score distribution computed by the student matches with that computed by the teacher, which is formulated as minimizing the cross entropy between the student and the teacher s similarity score distributions (as illustrated in Figure 2). Specifically, for a randomly augmented view xi of an image, it is first mapped and normalized into feature vector representations z T i = f T θ (xi)/||f T θ (xi)||2, and z S i = f S θ (xi)/||f S θ (xi)||2, where z T i , z S i RD, and f T θ and f S θ denote the teacher and student encoders, respectively. Let D = [d1...d K] denote the instance queue where K is the queue length and dj is the feature vector obtained from the teacher encoder. Similar to the contrastive learning framework, D is progressively updated under the first-in first-out strategy as distillation proceeds. That is, we en-queue the visual features of the current batch inferred by the teacher and de-queue the earliest seen samples at the end of iteration. Note that the maintained samples in queue D are mostly random and irrelevant to the target instance xi. Minimizing the cross entropy between the similarity score distribution computed by the student and teacher based on D softly contrasts xi with randomly selected samples, without directly aligning with the teacher encoder. To address this problem, we add the teacher s embedding (z T i ) into the queue and form D+ = [d1...d K, d K+1] with d K+1 = z T i . Let p T (xi; θT ; D+) denote the similarity score between the extracted teacher feature z T i and dj s (j = 1, ..., K + 1) computed by the teacher model. p T (xi; θT ; D+) is defined as p T (xi; θT , D+) = [ p T 1 ... p T K+1] , p T j = exp(z T i dj/τ T ) P d D+ exp(z T i d/τ T ), (2) and τ T is a temperature parameter for the teacher. Note, we use ()T to represent the feature from the teacher network and use ( ) to represent the inner product between two features. Similarly let p S(xi; θS, D+) denote the similarity score computed by the student model, which is defined as p S(xi; θS, D+) = [ p S 1 ... p S K+1] , where p S j = exp(z S i dj/τ S) P d D+ exp(z S i d/τ S), (3) and τ S is a temperature parameter for the student. Our self-supervised distillation can be formulated as minimizing the cross entropy between the similarity scores of the teacher, p T (xi; θT , D+), and the student, p S(xi; θS, D+), over all the instances xi, that is, ˆθS = arg min θS i p T (xi; θT , D+) log p S(xi; θS, D+) = arg min θS j exp(z T i dj/τ T ) P d D+ exp(z T i d/τ T ) log exp(z S i dj/τ S) P d D+ exp(z S i d/τ S). Since the teacher network is pre-trained and frozen, the queued features are consistent during training w.r.t. the student network. The higher the value of p T j is, the larger weight will be laid on p S j . Due Published as a conference paper at ICLR 2021 to the l2 normalization, similarity score between z T i and d K+1 remains constant 1 before softmax normalization, which is the largest among p T j . Thus, the weight for p S K+1 is the largest and can be adjusted solely by tuning the value of τ T . By minimizing the loss, the feature of z S i can be aligned with z T i and meanwhile contrasts with other unrelated image features in D. We further discuss the relation of these two goals with our learning objective in Appendix A.5. Relations with Info-NCE loss. When τ T 0, the softmax function for p T smoothly approaches to a one-hot vector, where p T K+1 equals 1 and all others 0. In this extreme case, the loss becomes i log exp(z T i z S i /τ) P d D+ exp(z S i d/τ), (5) which is similar to the widely-used Info-NCE loss (Oord et al., 2018) in contrastive-based SSL (see discussion in Appendix A.6. 4 EXPERIMENT 4.1 PRE-TRAINING Self-Supervised Pre-training of Teacher Network. By default, we use Mo Co-V2 (Chen et al., 2020c) to pre-train the teacher network. Following (Chen et al., 2020a), we use Res Net as the network backbone with different depths/widths and append a multi-layer-perceptron (MLP) layer (two linear layers and one Re LU (Nair & Hinton, 2010) activation layer in between) at the end of the encoder after average pooling. The dimension of the last feature dimension is 128. All teacher networks are pre-trained for 200 epochs due to the computational limitation unless explicitly specified. As our distillation is independent with the teacher pre-training algorithm, we also show results with other self-supervised pre-trained models for teacher network, e.g., SWAV (Caron et al., 2020), Sim CLR (Chen et al., 2020a). Self-Supervised Distillation on Student Network. We choose multiple smaller networks with fewer learnable parameters as the student network: Mobile Net-v3-Large (Howard et al., 2017), Efficient Net-B0 (Tan & Le, 2019), and smaller Res Net with fewer layers (Res Net-18, 34). Similar to the pre-training for teacher network, we add one additional MLP layer on the basis of the student network. Our distillation is trained with a standard SGD optimizer with momentum 0.9 and a weight decay parameter of 1e-4 for 200 epochs. The initial learning rate is set as 0.03 and updated by a cosine decay scheduler (Nair & Hinton, 2010) with 5 warm-up epochs and batch size 256. In Eq. 4, the teacher temperature is set as τ T = 0.01 and the student temperature is τ S = 0.2. The queue size of K is 65,536. In the following subsections and appendix, we also show results with different hyper-parameter values, e.g., for τ T and K. 4.2 FINE-TUNING AND EVALUATION In order to validate the effectiveness of self-supervised distillation, we choose to assess the performance of representations of the student encoder on several downstream tasks. We first report its performances of linear evaluation and semi-supervised linear evaluation on the Image Net ILSVRC2012 (Deng et al., 2009) dataset. To measure the feature transferability brought by distillation, we also conduct evaluations on other tasks, which include object detection and segmentation on the VOC07 (Everingham et al.) and MS-COCO (Lin et al., 2014) datasets. At the end, we compare the transferability of the features learned by distillation with ordinary self-supervised contrastive learning on the tasks of linear classification on datasets from different domains. Linear and KNN Evaluation on Image Net. We conduct the supervised linear classification on Image Net-1K, which contains 1.3M images for training, and 50,000 images for validation, spanning 1,000 categories. Following previous works in (He et al., 2020; Chen et al., 2020a), we train a single linear layer classifier on top of the frozen network encoder after self-supervised pretraining/distillation. SGD optimizer is used to train the linear classifier for 100 epochs with weight decay to be 0. The initial learning rate is set as 30 and is then reduced by a factor of 10 at 60 and 80 epochs (similar as in Tian et al. (2019a)). Notably, when training the linear classifier for Mobile Net and Efficient Net, we reduce the initial learning rate to 3. The results are reported in terms of Top-1 Published as a conference paper at ICLR 2021 Table 1: Image Net-1k test accuracy (%) using KNN and linear classification for multiple students and Mo Cov2 pre-trained deeper teacher architectures. denotes Mo Co-V2 self-supervised learning baselines before distillation. * indicates using a deeper teacher encoder pre-trained by SWAV, where additional small-patches are also utilized during distillation and trained for 800 epochs. K denotes Top-1 accuracy using KNN. T-1 and T-5 denote Top-1 and Top-5 accuracy using linear evaluation. First column shows Top-1 Acc. of Teacher network. First row shows the supervised performances of student networks. T S T-1 Eff-b0 Eff-b1 Mob-v3 R-18 R-34 K T-1 T-5 K T-1 T-5 K T-1 T-5 K T-1 T-5 K T-1 T-5 Supervised Acc. 77.3 79.2 75.2 72.1 75.0 - 30.0 42.2 68.5 34.4 50.7 74.6 27.5 36.3 62.2 36.7 52.5 77.0 41.5 57.4 81.6 R-50 67.4 46.0 61.3 82.7 46.1 61.4 83.1 44.8 55.2 80.3 43.4 57.9 82.0 45.2 58.5 82.6 +16.0 +19.1 +14.2 +16.1 +10.7 +8.8 +17.3 +18.9 +18.1 +6.7 +5.1 +4.8 +3.7 +1.1 +1.0 R-101 70.3 50.1 63.0 83.8 50.3 63.4 84.6 48.8 59.9 83.5 48.6 58.9 82.5 50.5 61.6 84.9 +20.1 +20.8 +15.3 +15.9 +12.7 +10.0 +21 .3 +23.6 +21.3 +11.9 +6.4 +5.5 +9.0 +4.2 +3.3 R-152 74.2 50.7 65.3 86.0 52.4 67.3 86.9 49.5 61.4 84.6 49.1 59.5 83.3 51.4 62.7 85.8 +20.7 +23.1 +17.5 +18.0 +16.6 +12.3 +22.0 +25.1 +22.4 +12.4 +7.0 +6.3 +9.9 +5.3 +4.2 R50 2 77.3 57.4 67.6 87.4 60.3 68.0 87.6 55.9 68.2 88.2 55.3 63.0 84.9 58.2 65.7 86.8 +27.4 +25.4 +18.9 +25.9 +17.3 +13.0 +18.9 +31.9 +26.0 +18.6 +10.5 +7.9 +16.7 +8.3 +5.2 0 20 40 60 0 Top-1 Accuracy (%) Number of Parameters of the Teacher (Millions) Efficient Net-b0 1% 10% 100% 0 20 40 60 0 Mobile Net-v3-large 1% 10% 100% 0 20 40 60 0 1% 10% 100% Figure 3: Image Net-1k Top-1 accuracy for semi-supervised evaluations using 1% (red line), 10% (blue line) of the annotations for linear fine-tuning, in comparison with the fully supervised (green line) linear evaluation baseline for SEED. For the points whose Teacher s number of parameters is at 0, we show the semi-supervised linear evaluation results of Mo Co-V2 without any distillation. The Student models tend to perform better on the semi-supervised tasks after distillation from larger Teachers. and Top-5 accuracy. We also perform classification using K-Nearest Neighbors (KNN) based on the learned 128d vector from the last MLP layer. The sample is classified by taking the most frequent label of its K (K = 10) nearest neighbors. Table 1 shows the results with various teacher networks and student networks. We list the baseline of contrastive self-supervised pre-training using Mo Co-V2 (Chen et al., 2020c) in the first row for each student architecture. We can see clearly that smaller networks perform rather worse. For example, Mobile Net-V3 can only reach 36.3%. This aligns well with previous conclusions from (Chen et al., 2020a;b) that bigger models are desired to perform better in contrastive-based self-supervised pretraining. We conjecture that this is mainly caused by the inability of smaller network to discriminate instances in a large-scale dataset. The results also clearly demonstrate that the distillation from a larger network helps boosting the performances of small networks, and show obvious improvement. For instance, with Mo Co-V2 pre-trained Res Net-152 (for 400 epochs) as the teacher network, the Top-1 accuracy of Mobile Net-V3-Large can be significantly improved from 36.3% to 61.4%. Furthermore, we use Res Net-50 2 (provided in Caron et al. (2020)) as the teacher network and adopt the multi-crop trick (see A.2 for details). The accuracy can be further improved to 68.2% (last row of Table 1) for Mobile Net-V3-Large with 800 epochs of distillation. We note that the gain benefited from distillation becomes more distinct on smaller architectures and we further study the effect of various teacher models in ablations. Semi-Supervised Evaluation on Image Net. Following (Oord et al., 2018; Kornblith et al., 2019; Kolesnikov et al., 2019), we evaluate the representation on the semi-supervised task, where a fixed 1% or 10% subsets of Image Net training data (Chen et al., 2020a) are provided with the annotations. After the self-supervised learning with and without distillation, we also train a classifier on top of the representation. The results are shown in Figure 3, where the baseline without distillation is depicted Published as a conference paper at ICLR 2021 Table 2: Object detection and instance segmentation results using contrastive self-supervised learning and SEED distillation using Res Net-18 as backbone: bounding-box AP (APbb) and mask AP (APmk) evaluated on VOC07-val and COCO testing split. More results on different backbones can be found in the Appendix. Subscript in green represents improvement is larger than 0.3. S T VOC Obj. Det. COCO Obj. Det. COCO Inst. Segm. APbb APbb 50 APbb 75 APbb APbb 50 APbb 75 APmk APmk 50 APmk 75 46.1 74.5 48.6 35.0 53.9 37.7 31.0 51.1 33.1 R-50 46.1( 0.0) 74.8(+0.3) 49.1(+0.5) 35.3(+0.3) 54.2(+0.3) 37.8(+0.1) 31.1(+0.1) 51.1( 0.0) 33.2(+0.1) R-101 46.8(+0.7) 75.8(+1.3) 49.3(+0.7) 35.3(+0.3) 54.3(+0.4) 37.9(+0.2) 31.3(+0.3) 51.3(+0.2) 33.4(+0.3) R-152 46.8(+0.7) 75.9(+1.4) 50.2(+1.6) 35.4(+0.4) 54.4(+0.5) 38.0(+0.3) 31.3(+0.3) 51.4(+0.3) 33.4(+0.3) 0 20 40 60 82 Top-1 Accuracy (%) Number of Parameters of the Teacher (Millions) 65.0 CIFAR-100 0 20 40 60 45 Res Net-18 Efficient Net Figure 4: Image Net-1k Accuracy (%) of student network (Efficient Net-B0 and Res Net-18) transferred to other domains (CIFAR-10, CIFAR-100, SUN-397 datasets) with and without distillation from lager architectures (Res Net-50/101/152). when teacher parameters are 0. As we can see, the accuracy is also improved remarkably with SEED distillation, and a stronger teacher network with more parameters leads to a better performed student network. Transferring to Classification. To further study whether the improvement of the learned representations by distillation is confined to Image Net, we evaluate on additional classification datasets to study the generalization and transferability of the feature representation. We strictly follow the linear evaluation and fine-tuning settings from (Kornblith et al., 2019; Chen et al., 2020a; Grill et al., 2020), that a linear layer is trained on the basis of frozen features. We report Top-1 Accuracy of models before and after distillation from various architectures on CIFAR-10, CIFAR-100 (Krizhevsky et al., 2009), SUN-397 (Xiao et al., 2010) datasets (see Figure 4). More details regarding pre-processing and training can be found in A.1.2. Notably, we observe that our distillation surpasses contrastive self-supervised pre-training consistently on all benchmarks, verifying the effectiveness of SEED. This also proves the generalization ability of the learned representations from distillation to a wide range of data domain and different classes. Transferring to Detection and Segmentation. We conduct two downstream tasks here. The first is Faster R-CNN (Ren et al., 2015) model for object detection trained on VOC-07+12 train+val set and evaluated on VOC-07 test split. The second is Mask R-CNN (He et al., 2017) model for the object detection and instance segmentation on COCO 2017 dataset (Lin et al., 2014). The pre-trained model serves as the initial weight and following He et al. (2020), we fine-tune all the layers of the model. More experiment settings can be found in A.2. The results are illustrated in Table 2. As we can see, on VOC, the distilled pre-trained model achieves a large improvement. With Res Net-152 as the teacher network, the Resnet18-based Faster R-CNN model shows +0.7 point improvement on AP, +1.4 improvement on AP50 and +1.6 on AP75. On COCO, the improvement is relatively minor and the reason could be that COCO training set has 118k training images while VOC has only 16.5k training images. A larger training set with more fine-tuning iterations reduces the importance of the initial weights. 4.3 ABLATION STUDY We now explore the effects of distillation using different Teacher architectures, Teacher Pre-training algorithms, various distillation strategies and hyper-parameters. Published as a conference paper at ICLR 2021 Table 3: Image Net-1k Accuracy (%) of student network (Res Net-18) distilled from variants of selfsupervised Res Net-50. P-E/D-E represent the pretraining and distillation epochs. T./S.-Top represent testing accuracy of Teacher and Student. represents distillation using additional small patches. First row is the Res Net-18 SSL baseline using Mo Co-v2 trained for 200 epochs. Teacher P-E D-E T. Top-1 S. Top-1 S. Top-5 52.5 77.0 Mo Co 200 200 60.6 52.1 77.0 Sim CLR 200 200 65.6 57.5 81.7 Mo Co-v2 200 200 67.4 57.9 82.0 800 200 71.1 60.5 83.5 SWAV 800 100 75.3 61.1 83.8 800 200 75.3 61.7 84.2 800 400 75.3 62.0 84.4 SWAV 800 200 75.3 62.6 84.8 Efficient Net-b0 R50 R101 R152 Teacher: Image Net Top-1 Accuracy (%) R50 R101 R152 Figure 5: Accuracy (%) of student networks (Efficient Net-b0 and Res Net-18) on Image Net distilled from wider Mo Co-v2 pre-trained Res Net (Res Net-50/101/152 2). Different Teacher Networks. Figure 5 summarizes the accuracy of Res Net-18 and Efficient Net-B0 distilled from wider and deeper Res Net architectures. We see clear performance improvement as depth and width of teacher network increase: compared to Res Net-50, deeper (Res Net-101) and wider (Res Net-50 2) substantially improve the accuracy. However, further architectural enlargement has relatively limited effects, and we suspect the accuracy might be limited by the student network capacity in this case. Different Teacher Pre-training Algorithms. In Table 3, we show the Top-1 accuracy of Res Net-18 distilled from Res Net-50 with different pre-training algorithms, i.e., Mo Co-V1 (He et al., 2020), Mo Co-V2 (Chen et al., 2020c), Sim CLR (Chen et al., 2020a), and SWAV (Caron et al., 2020)). Notably, the aforementioned methods all unanimously adopt contrastive-based pre-training except SWAV, which is based upon online clustering. We find that our SEED is agnostic to pre-training approaches, making it easy to use any self-supervised models (including clustering-based approach like SWAV) in self-supervised distillation. In addition, we observe that more training epochs for both teacher SSL and distillation epochs can bring beneficial gain. Other Distillation Strategies. We explore several alternative distillation strategies. l2-Distance: where the l2-distance of teacher & student s embeddings are minimized, motivated by Romero et al. (2014). K-Means: we exploit K-Means clustering to assign a pseudo-label based on the teacher network s representation. Online Clustering: we continuously update the clustering centers during distillation for pseudo-label generation. Binary Contrastive Loss: we adopt an Info-NCE alike loss for contrastive distillation (Tian et al., 2019a). We provide details for other strategies in A.4. Table 4 shows the results for each method on Res Net-18 (student) distilled from Res Net-50. From the results, the simple l2-distance minimizing approach can achieve a decent accuracy, which demonstrates the effectiveness of applying the distillation idea to the self-supervised learning. Beyond that, we study the effect of the original SSL (Mo Co-V2) supervision as supplementary loss to SEED and find it does not bring additional benefits to distillation. We find close results from these two strategies (Top-1 linear Acc.), SEED achieves 57.9%, while SEED + Mo Co-V2 achieves 57.6%. This implies that the loss of SEED can to a large extent cover the original SSL loss, and it is not necessary to conduct SSL any further during distillation. Meanwhile, our proposed SEED outperforms these alternatives with highest accuracy, which shows the superiority of aligning the student towards the teacher and contrasting with the irrelevant samples. Other Hyper-Parameters. Table 5 summarizes the distillation performances on multiple datasets using different temperature τ T . We observe a better performance when decreasing τ T to 0.01 for Image Net-1k and CIFAR-10 dataset, and to 1e-3 for CIFAR-100 datasets. When τ is large, the softmax-normalized similarity score of p T j between z T i and instance dj in the queue D+ also becomes large, which means the student s feature should be less discriminative with the features of other images to some extent. When τ T is 0, the teacher model will generate a one-hot vector, which only treats z T i as a positive instance and all others in the queue as negative. Thus, the best τ is a trade-off depending on the data distribution. We further compare effect of different hyper-parameters in A.8. Published as a conference paper at ICLR 2021 Table 4: Top-1/5 accuracy of linear classification results on Image Net using different distillation strategies on Res Net-18 (student) and Res Net-50 (teacher) architectures. Method Top-1 Acc. Top-5 Acc. l2-Distance 55.3 80.3 K-Means 51.0 75.8 Online Clustering 56.4 81.2 Binary Contr. Loss 57.4 81.5 SEED + Mo Co-V2 57.6 81.8 SEED 57.9 82.0 Table 5: Effect of τ T for the distillation of Res Net-18 (student), Res Net-50 (teacher) on multiple datasets. τ T Image Net CIFAR-10 CIFAR-100 Top-1 Top-5 Top-1 Top-1 0.3 54.8 80.0 78.7 46.6 0.1 54.9 80.1 83.0 50.1 0.05 56.5 81.3 84.4 56.2 0.01 57.9 82.0 87.5 60.6 1e-3 57.6 81.8 86.9 60.8 5 CONCLUSIONS Self-Supervised Learning is acknowledged for its remarkable ability in learning from unlabeled, and large scale data. However, a critical impedance for the SSL pre-training on smaller architecture comes from its low capacity of discriminating enormous number of instances. Instead of directly learning from unlabeled data, we proposed SEED as a novel self-supervised learning paradigm, which learns representation by self-supervised distillation from a bigger SSL pre-trained model. We show in extensive experiments that SEED effectively addresses the weakness of self-supervised learning for small models and achieves state-of-the-art results on various benchmarks of small architectures. Soroush Abbasi Koohpayegani, Ajinkya Tejankar, and Hamed Pirsiavash. Compress: Self-supervised learning by compressing representations. Advances in Neural Information Processing Systems, 33, 2020. Sungsoo Ahn, Shell Xu Hu, Andreas Damianou, Neil D Lawrence, and Zhenwen Dai. Variational information distillation for knowledge transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9163 9171, 2019. Cristian Buciluˇa, Rich Caruana, and Alexandru Niculescu-Mizil. Model compression. In Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining, pp. 535 541, 2006. Mathilde Caron, Piotr Bojanowski, Armand Joulin, and Matthijs Douze. Deep clustering for unsupervised learning of visual features. In Proceedings of the European Conference on Computer Vision (ECCV), pp. 132 149, 2018. Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. Unsupervised learning of visual features by contrasting cluster assignments. ar Xiv preprint ar Xiv:2006.09882, 2020. Guobin Chen, Wongun Choi, Xiang Yu, Tony Han, and Manmohan Chandraker. Learning efficient object detection models with knowledge distillation. In Advances in Neural Information Processing Systems, pp. 742 751, 2017. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. ar Xiv preprint ar Xiv:2002.05709, 2020a. Ting Chen, Simon Kornblith, Kevin Swersky, Mohammad Norouzi, and Geoffrey Hinton. Big self-supervised models are strong semi-supervised learners. ar Xiv preprint ar Xiv:2006.10029, 2020b. Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved baselines with momentum contrastive learning. ar Xiv preprint ar Xiv:2003.04297, 2020c. Published as a conference paper at ICLR 2021 Jang Hyun Cho and Bharath Hariharan. On the efficacy of knowledge distillation. In Proceedings of the IEEE International Conference on Computer Vision, pp. 4794 4802, 2019. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248 255. Ieee, 2009. M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results. http://www.pascalnetwork.org/challenges/VOC/voc2007/workshop/index.html. Zeyu Feng, Chang Xu, and Dacheng Tao. Self-supervised representation learning by rotation feature decoupling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 10364 10374, 2019. Spyros Gidaris, Praveer Singh, and Nikos Komodakis. Unsupervised representation learning by predicting image rotations. ar Xiv preprint ar Xiv:1803.07728, 2018. Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre H Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent: A new approach to self-supervised learning. ar Xiv preprint ar Xiv:2006.07733, 2020. Yandong Guo, Lei Zhang, Yuxiao Hu, Xiaodong He, and Jianfeng Gao. Ms-celeb-1m: A dataset and benchmark for large-scale face recognition. In European conference on computer vision, pp. 87 102. Springer, 2016. Michael Gutmann and Aapo Hyvärinen. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 297 304, 2010. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In Proceedings of the IEEE international conference on computer vision, pp. 2961 2969, 2017. Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729 9738, 2020. Olivier J Hénaff, Aravind Srinivas, Jeffrey De Fauw, Ali Razavi, Carl Doersch, SM Eslami, and Aaron van den Oord. Data-efficient image recognition with contrastive predictive coding. ar Xiv preprint ar Xiv:1905.09272, 2019. Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. ar Xiv preprint ar Xiv:1503.02531, 2015. Andrew Howard, Mark Sandler, Grace Chu, Liang-Chieh Chen, Bo Chen, Mingxing Tan, Weijun Wang, Yukun Zhu, Ruoming Pang, Vijay Vasudevan, et al. Searching for mobilenetv3. In Proceedings of the IEEE International Conference on Computer Vision, pp. 1314 1324, 2019. Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. Mobilenets: Efficient convolutional neural networks for mobile vision applications. ar Xiv preprint ar Xiv:1704.04861, 2017. Zehao Huang and Naiyan Wang. Like what you like: Knowledge distill via neuron selectivity transfer. ar Xiv preprint ar Xiv:1707.01219, 2017. Xiaoqi Jiao, Yichun Yin, Lifeng Shang, Xin Jiang, Xiao Chen, Linlin Li, Fang Wang, and Qun Liu. Tinybert: Distilling bert for natural language understanding. ar Xiv preprint ar Xiv:1909.10351, 2019. Published as a conference paper at ICLR 2021 Alexander Kolesnikov, Xiaohua Zhai, and Lucas Beyer. Revisiting self-supervised visual representation learning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 1920 1929, 2019. Animesh Koratana, Daniel Kang, Peter Bailis, and Matei Zaharia. Lit: Learned intermediate representation training for model compression. In International Conference on Machine Learning, pp. 3509 3518, 2019. Simon Kornblith, Jonathon Shlens, and Quoc V Le. Do better imagenet models transfer better? In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2661 2671, 2019. Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009. Junnan Li, Pan Zhou, Caiming Xiong, Richard Socher, and Steven CH Hoi. Prototypical contrastive learning of unsupervised representations. ar Xiv preprint ar Xiv:2005.04966, 2020. Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European conference on computer vision, pp. 740 755. Springer, 2014. Yifan Liu, Ke Chen, Chris Liu, Zengchang Qin, Zhenbo Luo, and Jingdong Wang. Structured knowledge distillation for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2604 2613, 2019. Asit Mishra and Debbie Marr. Apprentice: Using knowledge distillation techniques to improve low-precision network accuracy. ar Xiv preprint ar Xiv:1711.05852, 2017. Ishan Misra and Laurens van der Maaten. Self-supervised learning of pretext-invariant representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6707 6717, 2020. Vinod Nair and Geoffrey E Hinton. Rectified linear units improve restricted boltzmann machines. In ICML, 2010. Mehdi Noroozi and Paolo Favaro. Unsupervised learning of visual representations by solving jigsaw puzzles. In European Conference on Computer Vision, pp. 69 84. Springer, 2016. Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. ar Xiv preprint ar Xiv:1807.03748, 2018. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. In Advances in neural information processing systems, pp. 8026 8037, 2019. Antonio Polino, Razvan Pascanu, and Dan Alistarh. Model compression via distillation and quantization. ar Xiv preprint ar Xiv:1802.05668, 2018. Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in neural information processing systems, pp. 91 99, 2015. Adriana Romero, Nicolas Ballas, Samira Ebrahimi Kahou, Antoine Chassang, Carlo Gatta, and Yoshua Bengio. Fitnets: Hints for thin deep nets. 2014. Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, and Liang-Chieh Chen. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4510 4520, 2018. Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. ar Xiv preprint ar Xiv:1910.01108, 2019. Zhiqiang Shen and Marios Savvides. Meal v2: Boosting vanilla resnet-50 to 80%+ top-1 accuracy on imagenet without tricks. ar Xiv preprint ar Xiv:2009.08453, 2020. Published as a conference paper at ICLR 2021 Siqi Sun, Yu Cheng, Zhe Gan, and Jingjing Liu. Patient knowledge distillation for bert model compression. ar Xiv preprint ar Xiv:1908.09355, 2019. Zhiqing Sun, Hongkun Yu, Xiaodan Song, Renjie Liu, Yiming Yang, and Denny Zhou. Mobilebert: a compact task-agnostic bert for resource-limited devices. ar Xiv preprint ar Xiv:2004.02984, 2020. Mingxing Tan and Quoc V Le. Efficientnet: Rethinking model scaling for convolutional neural networks. ar Xiv preprint ar Xiv:1905.11946, 2019. Yonglong Tian, Dilip Krishnan, and Phillip Isola. Contrastive representation distillation. In International Conference on Learning Representations, 2019a. Yonglong Tian, Dilip Krishnan, and Phillip Isola. Contrastive multiview coding. ar Xiv preprint ar Xiv:1906.05849, 2019b. Iulia Turc, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Well-read students learn better: On the importance of pre-training compact models. ar Xiv preprint ar Xiv:1908.08962, 2019. Tongzhou Wang and Phillip Isola. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. ar Xiv preprint ar Xiv:2005.10242, 2020. Yuxin Wu, Alexander Kirillov, Francisco Massa, Wan-Yen Lo, and Ross Girshick. Detectron2, 2019. Zhirong Wu, Yuanjun Xiong, Stella X Yu, and Dahua Lin. Unsupervised feature learning via nonparametric instance discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3733 3742, 2018. Jianxiong Xiao, James Hays, Krista A Ehinger, Aude Oliva, and Antonio Torralba. Sun database: Large-scale scene recognition from abbey to zoo. In 2010 IEEE computer society conference on computer vision and pattern recognition, pp. 3485 3492. IEEE, 2010. Jianwei Yang, Devi Parikh, and Dhruv Batra. Joint unsupervised learning of deep representations and image clusters. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5147 5156, 2016. Junho Yim, Donggyu Joo, Jihoon Bae, and Junmo Kim. A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4133 4141, 2017. Sergey Zagoruyko and Nikos Komodakis. Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. ar Xiv preprint ar Xiv:1612.03928, 2016. Xiaohua Zhai, Avital Oliver, Alexander Kolesnikov, and Lucas Beyer. S4l: Self-supervised semisupervised learning. In Proceedings of the IEEE international conference on computer vision, pp. 1476 1485, 2019. Liheng Zhang, Guo-Jun Qi, Liqiang Wang, and Jiebo Luo. Aet vs. aed: Unsupervised representation learning by auto-encoding transformations rather than data. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2547 2555, 2019. Richard Zhang, Phillip Isola, and Alexei A Efros. Colorful image colorization. In European conference on computer vision, pp. 649 666. Springer, 2016. Xiao Zhang, Zhiyuan Fang, Yandong Wen, Zhifeng Li, and Yu Qiao. Range loss for deep face recognition with long-tailed training data. In Proceedings of the IEEE International Conference on Computer Vision, pp. 5409 5418, 2017. Published as a conference paper at ICLR 2021 We discuss more details and different hyperparameters for SEED during distillation. A.1 PSEUDO-IMPLEMENTATIONS We provide pseudo-code of the SEED distillation in Py Torch Paszke et al. (2019) style: 1 Q: maintaining queue of previous representations: (N X D) 2 T: Cumbersome encoder as Teacher. 3 S: Target encoder as Student. 4 temp_T, temp_S: temperatures of the Teacher & Student. 7 # activate evaluation mode for Teacher to freeze BN and updation. 10 for images in enumerate(loader): # Enumerate single crop-view 12 # augment image to get one identical view 13 images = aug(images) 15 # Batch-size 16 B = images.shape[0] 18 # extract embedding from S: 1 X D 19 X_S = S(images) 20 X_S = torch.norm(X_S, p=2, dim=1) 22 # use the gradient-free mode 23 with torch.no_grad(): 24 X_T = T(image) # embedding from T: 1 X D 25 X_T = torch.norm(X_T, p=2, dim=1) 27 # insert the current batch embedding from T 28 enqueue(Q, X_T) 30 # probability scores distribution for T, S: B X (N + 1) 31 S_Dist = torch.einsum( bd, dn -> bn , [X_S], Q.t().clone().detach()) 32 T_Dist = torch.einsum( bd, dn -> bn , [X_T], Q.t().clone().detach()) 34 # Apply temperatures for soft-labels 35 S_Dist /= temp_S 36 T_Dist = Soft Max(T_Dist/temp_T, dim=1) 38 # loss computation, use log_softmax for stable computation 39 loss = -torch.mul(T_Dist, Log_Soft Max(S_Dist, dim=1)).sum()/B 41 # update the random sample queue 42 dequeue(Q, B) # pop-out earliest B instances 44 # SGD updation 45 loss.backward() 46 update(S.params) Published as a conference paper at ICLR 2021 A.1.1 DATA AUGMENTATIONS Both our teacher pre-training and distillation adopt the data augmentations as follows: Random Resized Crop: The image is randomly resized with a scale of {0.2, 1.0}, then cropped to the size of 224 224. Random Color Jittering: with brightness to be {0.4, 0.4, 0.4, 0.1} with probability at 0.8. Random Gray Scale transformation: with probability at 0.2. Random Gaussian Blur transformation: with σ = {0.1, 0.2} and probability at 0.5. Horizontal Flip: Horizontal flip is applied with probability at 0.5. A.1.2 PRE-TRAINING AND DISTILLATION ON MOBILENET AND EFFICIENTNET Mobile Net (Howard et al., 2017) and Efficient Net (Tan & Le, 2019) have been considered as the smaller counterparts with larger models, i.e., Res Net-50 (with supervised training, Efficient Net B0 hits 77.2% Top-1 Acc., and Mobile Net-V3-large reaches 72.2% on Image Net testing split). Nevertheless, un-matched performances are observed in the task of self-supervised contrastive pretraining: i.e., Self-Supervised Learning (Mo Co-V2) on Mobile Net-V3 only yields 36.3% Top-1 Acc. on Image Net. We conjecture that several reasons might lead to this dilemma: 1. The inability of models with less parameters for handling large volume of categories and data, which exists also in other domains, i.e., face recognition (Guo et al., 2016; Zhang et al., 2017). 2. Less possibility for optimum parameters to be chosen when transferring to downstream tasks: models with more parameters after pre-training might produce a plenty cornucopia of optimum parameters for fine-tuning. To narrow the dramatic performance gap between smaller architectures using contrastive SSL with the larger, we explore with architectural manipulations and training hyper-parameters. In specific, we find that by adding a deeper projection head largely improves the representation quality, a.k.a., better performances on linear evaluation. We experiment with adding one additional linear projection head on the top of convolutional backbones. Similarly, we also expand the MLP projection head on Efficient Net-b0. Though recent work shows that fine-tuning from a middle layer of the projection head can produce a largely different result (Chen et al., 2020b), we consistently just use the representations from convolutional trunk without adding extra layers during the phase of linear evaluation. As shown in Table 6, pre-training with a deeper projection head dramatically helps the improvement on linear evaluations, adding 17% Top-1. Acc. for Mobile-v3-large, and we report the improved baselines in the main paper (see the first row in Table 1 of the main paper). We keep most of the hyper-parameters as the distillation on Res Net except reducing the weight-decay of them to 1e-5, following (Tan & Le, 2019; Sandler et al., 2018). Table 6: Linear evaluations on Image Net of Efficient Net and Mobile Net pre-trained using Mo Co-v2. A deeper projection head largely boosts the linear evaluation performances on smaller architectures. Model Deeper MLPs Top-1 Acc. Top-5 Acc. Efficient Net-b0 39.1 64.6 Efficient Net-b0 42.2 68.5 Mobile-v3-large 19.0 41.3 Mobile-v3-large 36.3 62.2 A.2 ADDITIONAL DETAILS OF EVALUATIONS We list additional details regarding our evaluation experiments in this section. Image Net-1k Semi-Supervised Linear Evaluation. Following Zhai et al. (2019); Chen et al. (2020a), we train the FC layers on the basis of our student encoder after distillation using a fraction Published as a conference paper at ICLR 2021 Table 7: Before and after distillation Top-1/5 test accuracy (%) on Image Net of Efficient Net-b0 and Mobile Netlarge without deeper MLPs. Student Teacher Top-1 Top-5 Efficient Net-b0 39.1 64.6 Res Net-50 59.2 81.2 Res Net-101 62.8 84.7 Res Net-152 63.3 85.6 Mobile Net-v3 19.0 41.3 Res Net-50 50.9 77.7 Res Net-101 57.6 82.6 Res Net-152 58.3 82.9 Table 8: Image Net-1k test accuracy (%) under KNN and linear classification on Res Net-50 encoder with deeper, Mo Co-V2/SWAV pre-trained teacher architectures. denotes Mo Co-V2 self-supervised learning baselines before distillation. * indicates using a stronger teacher encoder pre-trained by SWAV with additional small-patches during distillation. Teac. Stud. Res Net-50 Epoch KNN Top-1 Top-5 200 46.1 67.4 87.8 Res Net-50 200 46.1 67.5 87.8 +0.0 +0.1 +0.0 Res Net-101 200 52.3 69.1 88.7 +6.2 +1.7 +0.9 Res Net-152 200 53.2 70.4 90.5 +7.1 +3.0 +2.7 Res Net-50 2 800 59.0 74.3 92.2 +12.9 +6.9 +4.4 of labeled Image Net-1k dataset (1% and 10%), and evaluate it on the whole test split. The fraction of labeled dataset is constructed in a class-balanced way, with roughly 12 and 128 images per class . We use SGD optimizer and set initial learning rate to be 30 with a multiplier = Batch Size/256 without weight decaying for 100 epochs. We use the step-wise scheduler for the learning rate updating with 5 warm-up epochs, and the learning rate is reduced by 10 at 60 and 80 epochs. On smaller architectures like Efficient Net and Mobile Net, we reduce the initial learning rate to 3. During training, the image is center-cropped to the size of 224 224 with just Random Horizontal Flip as the data augmentation. For testing, we first resize the image to 256 256 and use the center cropped 224 224 for pre-processing. In Table 8, we show the distillation results on a larger encoder (Res Net-550) when using different teacher networks. Transfer Learning. We test the transferability of the representations learned from self-supervised distillation by conducting the linear evaluations using offline features on several other datasets. Specifically, a single layer logistic classifier is trained following (Chen et al., 2020a; Grill et al., 2020) using SGD optimizer without weight decay and momentum parameter at 0.9. We use CIFAR-10, CIFAR-100 (Krizhevsky et al., 2009) and SUN-397 (Xiao et al., 2010) as our testing beds. CIFAR: As the size for CIFAR dataset is 32 32, we resize all images to 224 224 pixels along the shorter side using bicubic resampling method, followed by a center crop operation. We set the learning rate at 1e-3 constantly and train it for 120 epochs. The hyper-parameters are searched using 10 fold cross-validation on the train split and report its final top-1 accuracy on the test split. The full image ids for semi-supervised evaluation on Image Net-1k can be found at https://github. com/google-research/simclr/tree/master/imagenet_subsets. Published as a conference paper at ICLR 2021 Table 9: Object detection and instance segmentation fine-tuned on VOC07: bounding-box AP (APbb) and mask AP (APmk) evaluated on VOC07-val. The first row shows the baseline from Mo Co-v2 backbones without distillation. Student Teacher VOC Object Detection APbb APbb 50 APbb 75 53.6 79.1 58.7 Res Net-50 53.7 (+0.1) 79.4 (+0.3) 59.2 (+0.5) Res Net-101 54.1 (+0.5) 79.8 (+0.7) 59.1 (+0.4) Res Net-152 54.4 (+0.8) 80.1 (+1.0) 59.9 (+1.2) 57.0 82.4 63.6 Res Net-50 57.0 (+0.0) 82.4 (+0.0) 63.6 (+0.0) Res Net-101 57.1 (+0.1) 82.8 (+0.4) 63.8 (+0.2) Res Net-152 57.3 (+0.3) 82.8 (+0.4) 63.9 (+0.3) Table 10: Object detection and instance segmentation fine-tuned on COCO: bounding-box AP (APbb) and mask AP (APmk) evaluated on COCO-val2017. The first several rows show the baselines from unsupervised backbones without distillation. Student Teacher Object Detection Instance Segmentation APbb APbb 50 APbb 75 APmk APmk 50 APmk 75 38.1 56.8 40.7 33.0 53.2 35.3 Res Net50 38.4 (+0.3) 57.0 (+0.2) 41.0 (+0.3) 33.3 (+0.3) 53.6 (+0.4) 35.4 (+0.1) Res Net101 38.5 (+0.4) 57.3 (+0.5) 41.4 (+0.7) 33.6 (+0.6) 54.1 (+0.9) 35.6 (+0.3) Res Net152 38.4 (+0.3) 57.0 (+0.2) 41.0 (+0.3) 33.3 (+0.3) 53.7 (+0.5) 35.3 (+0.0) SUN-397: We further extend our transferring evaluation to the scene dataset SUN-397 for a more diverse testing. The official dataset specifies 10 different train/test splits, with each contains 50 images per category covering 397 different scenes. We follow (Chen et al., 2020a; Grill et al., 2020) and use the first train/test split. For the validation set, we randomly pick 10 images (yielding 20% of the dataset), with identical optimizer parameters as CIFAR. Object Detection and Instance Segmentation. As indicated by (He et al., 2020), features produced by self-supervised pre-training have divergent distributions in downstream tasks, thus resulting the supervised pre-training picked hyper-parameters not applicable. To relieve this, He et al. (2020) uses feature normalization during the fine-tuning phase and train the BN layers. Different from previous transferring and linear evaluations where we exploit only offline features, model for detection and segmentation is trained with all parameters tuned. For this reason, annotations on COCO for segmentation gives much higher influence for the backbone model than the VOC dataset (see Table 9), and gives an offset to the pre-training difference (see Table 10). Thus, this makes the performance boosting by pre-training less obvious, and leads to trivial AP differences before and after distillation. Object Detection on PASCAL VOC-07: We train a C4 (He et al., 2017) based Faster R-CNN (Ren et al., 2015) as the detector with different Res Net architectures (Res Net-18, Res Net-34 and Res Net-50) for evaluating the transferability of features for object detection tasks. We use Detectron2 (Wu et al., 2019) for the implementations. We train our detector for 48k iterations with a batch size of 32 (8 images per GPU). The base learning rate is set to 0.01 with 200 warm-up iterations. We set the scale of images for training as [400, 800] and 800 at inference. Object Detection and Segmentation on COCO: We use Mask R-CNN (He et al., 2017) with the C4 backbone for the object detection and instance segmentation task on COCO dataset, with 2 schedule. Similar to the VOC detection, we tune the BN layers and all parameters. The model is trained for 180k iterations with initial learning rate set to 0.02. We set the scale of images for training as [600, 800] and 800 at inference. Published as a conference paper at ICLR 2021 Table 11: Linear evaluations on Image Net of Res Net-18 after distillation from the SWAV pre-trained Res Net-50 using either single view, cross-views, or small patch views. Method Multi-View(s) Top-1 Acc. Top-5 Acc. Identical-View 1 224 61.7 84.2 Cross-Views 2 224 58.2 81.7 Multi-Crops + Cross-Views 1 224 + 6 96 96 61.9 84.4 Multi-Crops + Identical-View 1 224 + 6 96 96 62.6 84.8 (a) (b) (c) (d) Figure 6: We experiment with different strategies of using views during distillation, which include: (a). Identical view for distillation. (b). Cross view distillation. (c). Large-small cross view distillation. (d). Large-small identical view distillation. A.3 SINGLE CROP V.S. MULTI-CROPS VIEW(S) FOR DISTILLATION In contrary with most contrastive SSL methods where two different augmented views of an image are utilized as the positive samples (see Figure 6-a), SEED uses an identical view for each image (see Figure 6-b) during distillation and yields better performances, as is shown in Table. 11. In addition, we have also experimented with two strategies of using small patches. To be specific, we follow the set-up in SWAV (Caron et al., 2020), that 6 small patches of the size 96 96 are sampled at the scale of (0.05, 0.14). Then, we apply the same augmentations as introduced previously as data pre-processing. Figure. 6-c shows the way that is similar in SWAV for small-patch learning, where both large and 6 small patches are fed into the student encoder, with the learning target (z T ) to be the embedding of large view from the teacher encoder. Figure. 6-d is the strategy we use during distillation, that both views are fed into student and teacher to produce the embeddings for small-views (z S s , z T s ) and large views (z S l , z T l ). Based on that, the distillation is formulated separately on the small and large views. Notably, we maintain two independent queues for storing historical data samples for the large and small views. A.4 STRATEGIES FOR OTHER DISTILLATION METHODS We compare the effect of distillation using different strategies with SEED. l2-Distance: We train the student encoder by minimizing the squared l2-distance of representations from student (z S i ) and teacher (z T i ) for an identical view xi. K-Means: We experiment with the K-Means clustering method to retrieve pseudo class labels for distillation. Specifically, we first extract offline image features using the SSL pre-trained Teacher network without any image augmentations. Based on this, we conduct our K-Means clustering with 4k and 16k unique centroids. Then the final centroids are used to produce pseudo labels for unlabelled instances. With that, we carry out the distillation by training the model on a classification task using the produced labels as the ground-truth. To avoid trivial solutions that the majority of images are assigned to a few clusters, we sample images based on a uniform distribution over pseudo-labels as clustering proceeds. We observe very close results when adjusting numbers of centroids. Online-Clustering: With K-Means for pseudo-label generation training, it does not lead to satisfying results (51.0% on Res Net-18 with Res Net-50 as Teacher) as instances might have not been accurately categorized by limited frozen centroids. Similar to (Caron et al., 2018; Li et al., 2020), we resort to the in-batch and dynamical clustering to substitute the frozen K-Means method. We conduct Published as a conference paper at ICLR 2021 K-Means clustering within a batch and continuously update the centroid based on the teacher feature representation as distillation goes on. This alleviates the above problems and yields a substantial performance improvement on Res Net-18 to 56.4%. Binary Contrastive Loss: We resort to CRD (Tian et al., 2019a) and adopt an info-NCE loss-alike training objective in unsupervised distillation tasks. Specifically, we treat representation features from Teacher and Student for instance xi as positive pairs, and random instances from D as negative samples: ˆθS = arg min θS i log h(z S i , z T i ) + K [log h(1 h(z S i , d T j ))], (6) where d T j D, h( ) is any family of functions that satisfy h: {z, d} [0, 1], e.g., cosine similarity. A.5 DISCUSSIONS ON SEED Our proposed learning objective for SEED is composed of two goals, that is to align the encoding z S by the student model with z T produced by the teacher model; meanwhile, z S also softly contrasts with random samples maintained in the D. This can be formulated more directly as minimizing the l2 distance of z T , z S, together with the cross-entropy computed using D: λa z T i z S i 2 λb p T (xi; θT , D) log p S(xi; θS, D) λa z T i z S i λb exp(z T i dj/τ T ) P d D exp(z T i d/τ T ) log exp(z S i dj/τ S) P d D exp(z S i d/τ S) Directly optimizing Eq. 7 can lead to apparent difficulty in searching optimal hyper-parameters (λa, λb, τ T and τ S). Our proposed objective on D+ indeed is an approximated upper-bound of the above objectiveness however much simplified: i p T (xi; θT , D+) log p S(xi; θS, D+) j exp(z T i dj/τ T ) P d D+ exp(z T i d/τ T ) | {z } wi j log exp(z S i dj/τ S) P d D+ exp(z S i d/τ S), (8) where we let wi j denote the weighting term regulated under τ T . Since the (K + 1)th element in D+ is our supplemented vector z T i , the above objective can be expanded into: n wi K+1 z S i z T i /τ S + log X d D+ exp(z S i d/τ S) j=1 wi j z S i dj/τ S + log X d D+ exp(z S i d/τ S) o (9) Note that the LSE term in the first line is strictly non-negative as the range of inner product for z S and d lies between -1, +1 : LSE(D+, z S i ) log M exp( 1/τ S) = log M exp( 5) > 0, (10) where M denotes the cardinality of the maintained queue D+ and is set to 65,536 in our experiment with τ S = 0.2 constantly. Meanwhile, the LSE term in the second line satisfies the following inequality: LSE(D+, z S i ) LSE(D, z S i ). (11) Published as a conference paper at ICLR 2021 Thus, this demonstrates that the objective for SEED as Eq. 8 is equivalent to minimizing a weakened upper-bound of e.q. 7: i p T (xi; θT , D+) log p S(xi; θS, D+) n wi K+1 ( z S i z T i /τ S) + j=1 wi j z S i dj/τ S + log X d D exp(z S i d/τ S) o wi K+1 τ S z S i z T i p T (xi; θT , D) log p S(xi; θS, D) (12) This proves that our LSEED directly relates to a more intuitive distillation formulation as Eq. 7 (l2 + cross entropy loss), and it implicitly contains the objective of aligning and contrasting. However, our training objective is much simplified. During practice, we find by regulating τ T , both training losses produce equal results. A.6 DISCUSSION ON THE RELATIONSHIP OF SEED WITH INFO-NCE The objective of distillation can be considered as a soft version of Info-NCE (Oord et al., 2018), with the only difference to be that SEED learns from the negative samples with probabilities instead of treating them all strictly as negative samples. To be more specific, following Info-NCE, the hard style contrastive distillation can be expressed as aligning with representations from the Teacher encoder and contrasting with all random instances: ˆθS = arg min θS LNCE = arg min θS i log exp(z T i z S i /τ) P d D exp(z S i d/τ) (13) which can be further deduced with two sub-terms consisting of positive sample alignment and contrasting with negative instances: n z S i z T i /τ | {z } alignment d D exp(z S i d/τ) | {z } contrasting Similarly, the objective of SEED can be dissembled into the weighted form of alignment and contrasting terms: LSEED = 1 N M j exp(z T i dj/τ T ) P d D+ exp(z T i d/τ T ) log exp(z S i dj/τ S) P d D+ exp(z S i d)/τ S exp(z T i dj/τ T ) P d D+ exp(z T i d/τ T ) | {z } wi j ( z S i z T i /τ S | {z } alignment d D exp(z S i d/τ S) | {z } contrasting (15) where the normalization term can be considered as soft labels, Wi = wi 1 . . . wi K+1 , which can weight the above loss as: LSEED = 1 N M j wi j n z S i z T i /τ S + log X d D exp(z S i d/τ S)) o , (16) When tuning hyper-parameter τ T towards 0, Wi can be altered into the format of one-hot vector with wi K+1 = 1, which is then degraded to the case of contrastive distillation as in equation 14. In practice, the choice of an optimal τ T can be dataset-specific. We show that the higher τ T (with labels be more soft ) can actually yield better results on other datasets, e.g., CIFAR-10 (Krizhevsky et al., 2009). Published as a conference paper at ICLR 2021 A.7 COMPATIBILITY WITH SUPERVISED DISTILLATION SEED conducts self-supervised distillation at the pre-training phase for the representation learning. However, we verify that SEED is compatible with traditional supervised distillation that happened during fine-tuning phrase at downstream, and can even produce better results. We begin with the SSL pre-training on a larger architecture (Res Net-152) using Mo Co-V2 and train it for 200 epochs as the teacher network. As images in CIFAR-100 are in the size of 32 32, we modify the first conv layer in Res Net with kernel size = 3 and stride = 1. We then compare the Top-1 accuracy of a smaller Res Net-18 on CIFAR-100 when using different distillation strategies when all parameters are trainable. First, we use SEED to pre-train Res Net-18 with Res-152 as the teacher model, and then evaluate in on the test split of CIFAR-100 using linear fine-tuning task. As we keep all parameters trainable during the fine-tuning phase, distillation on the pre-training only yields a trivial boost: 75.4% v.s. 75.2%. Then, we adopt the traditional distillation method, e.g., (Hinton et al., 2015), to first fine-tune the Res Net-152 model, and then use its output class probability to facilitate the linear classification task on Res Net-18 in the fine-tuning phrase. This improves the linear classification accuracy on Res Net-18 to 76.0%. At the end, we initialize the Res Net-18 with our SEED pre-trained Res Net-18, and equip it with the supervised classification distillation during fine-tuning. With that, we find that the performance of Res Net-18 is further boosted to 78.1%. We can conclude that our SEED is compatible with traditional supervised distillation that mostly happened at downstream for specific tasks, e.g., classification, object detection. Table 12: CIFAR-100 Top-1 Accuracy(%) of Res Net-18 with (or without) distillation at different phase: selfsupervised pre-training stage, and supervised classification fine-tuning. All backbone parameters of Res Net-18 are trainable in experiments. Pre-training Distill. Fine-tuning Distill. Top-1 Acc 75.2 75.4 76.0 78.1 4 5 6 7 8 9 10 11 12 Log(Size of Data Sample Queue) Image Net Top-1 Accuracy (%) Figure 7: Linear evaluation accuracy (%) of distillation between Res Net-18 (as the Student) and Res Net-50 (as the Teacher) using different size of queue when LR=0.03 and weight decay=1e-6. Note the axis is the log( ) value of queue lengths. A.8 ADDITIONAL ABLATION STUDIES We study effects of different hyper-parameters to distillation using a Res Net-18 (as Student) and a SWAV pre-trained Res Net-50 (as Teacher) with small patch views. In specific, we list the Top-1 Acc. on validation split of Image Net-1k using different lengths of queue (K=128, 512, 1,024, 4,096, 8,192, 16,384, 65,536) in Figure. 7. With the increasing of random data samples, the distillation boosts the accuracy of learned representations, however within a limited range: +1.5 when the queue size is Published as a conference paper at ICLR 2021 Table 13: Linear evaluation accuracy (%) of distillation between Res Net-18 (as the Student) and Res Net-50 (as the Teacher) using different learning rates when the queue size is 65,536 and weight decay=1e-6. LR Top-1 Acc. Top-5 Acc. 1 58.9 83.1 0.1 62.9 85.3 0.03 63.3 85.4 0.01 62.6 85.0 Table 14: Linear evaluation accuracy (%) of distillation between Res Net-18 (as the Student) and Res Net-50 (as the Teacher) using different weight decays when the queue size is 65,536 and LR=0.03. WD Top-1 Acc. Top-5 Acc. 1e-2 11.8 27.7 1e-3 62.3 84.7 1e-4 61.9 84.4 1e-5 61.6 84.2 1e-6 63.3 85.4 65,536 compared with 256. Furthermore, Table. 13 and 14 summarize the linear evaluation accuracy under different learning rates and weight decays.