# selfsupervised_learning_via_maximum_entropy_coding__522f930e.pdf Self-Supervised Learning via Maximum Entropy Coding Xin Liu Zhongdao Wang Yali Li Shengjin Wang Beijing National Research Center for Information Science and Technology (BNRist) Department of Electronic Engineering, Tsinghua University {xinliu20, wcd17}@mails.tsinghua.edu.cn {liyali13, wgsgj}@tsinghua.edu.cn A mainstream type of current self-supervised learning methods pursues a generalpurpose representation that can be well transferred to downstream tasks, typically by optimizing on a given pretext task such as instance discrimination. In this work, we argue that existing pretext tasks inevitably introduce biases into the learned representation, which in turn leads to biased transfer performance on various downstream tasks. To cope with this issue, we propose Maximum Entropy Coding (MEC), a more principled objective that explicitly optimizes on the structure of the representation, so that the learned representation is less biased and thus generalizes better to unseen downstream tasks. Inspired by the principle of maximum entropy in information theory, we hypothesize that a generalizable representation should be the one that admits the maximum entropy among all plausible representations. To make the objective end-to-end trainable, we propose to leverage the minimal coding length in lossy data coding as a computationally tractable surrogate for the entropy, and further derive a scalable reformulation of the objective that allows fast computation. Extensive experiments demonstrate that MEC learns a more generalizable representation than previous methods based on specific pretext tasks. It achieves state-of-the-art performance consistently on various downstream tasks, including not only Image Net linear probe, but also semi-supervised classification, object detection, instance segmentation, and object tracking. Interestingly, we show that existing batch-wise and feature-wise self-supervised objectives could be seen equivalent to low-order approximations of MEC. Code and pre-trained models are available at https://github.com/xinliu20/MEC. 1 Introduction Self-supervised learning (SSL) aims to learn rich and meaningful representations without relying on human annotations. Pursuing general-purpose representations, SSL models are typically used as pre-trained weights for providing a good initialization to downstream tasks. In this sense, SSL [12, 31, 10, 29, 15, 14, 86, 30] has seen great progress in computer vision, and can achieve competitive or even better performance on various downstream tasks compared to its supervised counterparts. At the core of current SSL methods is the design of pretext tasks. A pretext task is a (usually hand-crafted) learning objective in which the supervision signals could be mined from the data itself, with the aim of applying the learned representation to other downstream tasks. Early attempts of pretext tasks typically discard a certain property of the input data and then force the model to predict the discarded property. For instance, one can convert an RGB image to a gray-scale one, and train the model to predict the original color [87, 88]; or apply random rotation to image patches and ask the Corresponding author 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Barlow Twins Figure 1: Comparison of transfer learning performance on five image-based tasks (top row) and five video-based tasks (bottom row). Along each axis we plot the performance ranking of the representation on a specific downstream task, so a polygon with a larger area means better generalization capacity on various downstream tasks, across the board. model to predict the rotation angles [27]. In both cases, to fulfill the pretext task, the model must learn a meaningful representation that can describe object texture, shape or even category, therefore the learned representation transfers well to downstream tasks related to these features. Despite successes of existing pretext tasks, we find that they inevitably introduce biases into the learned representation, which conflicts with the original aim of general-purpose . For example, representations trained with the most prevalent image-level task, instance discrimination [20, 80], are found biased to image-level tasks such as image classification, while by contrast degenerate in patchor pixel-level tasks like object detection and semantic segmentation [81]. Even being transferred to image-level tasks, such representations still suffer from domain gaps in cases of different data distributions [26], e.g., classification on unseen categories other than Image Net objects. In this work, we are curious of what makes for generalizable representations, and pursue an explicit optimization with a criterion that directly measures the structure of representations, with the aim of minimizing the biases brought by the pretext task. To this end, we propose Maximum Entropy Coding (MEC). Inspired by the principle of maximum entropy in information theory, the basic hypothesis in MEC is that a generalizable representation should be the one that admit the maximum entropy among all plausible representations. Accordingly, optimizing towards maximum entropy leads to representations with good generalization capacity. The main challenge confronting us is that it is difficult, and computationally expensive, if possible, to estimate the distribution of a given representation (usually a finite set of high-dimensional vectors), so in turn it is difficult to estimate the entropy. To cope with this issue, we replace the optimization objective from the originally defined entropy to a computationally tractable surrogate, i.e., the necessary number of bits needed to encode the representations via lossy data coding [16]. The log-determinant term costs the most computation in the coding length function. By leveraging Taylor expansion of matrix, we further approximate this term with polynomial functions, leading to significant speed up and thus makes large-scale pre-training possible. In contrast to previous SSL methods that mainly evaluate on Image Net classification, we experiment on a wide variety of vision tasks to show the good generalization capacity of MEC. The considered tasks span across not only image-level recognition on various data distributions, but also patchor pixel-level tasks like object detection, instance segmentation, and object tracking. We show MEC generalizes well consistently across all tasks considered, while representations learned with previous pretext tasks usually perform well on closely related downstream tasks but degenerate on less related ones (see Figure 1). Besides empirical results, we find interesting equivalence between low-order approximations of MEC and existing batch-wise (e.g., Sim Siam [14]) or feature-wise objectives (e.g., Barlow Twins [86]), which provides a new perspective for a unified understanding of prevalent SSL methods. In summary, the main contributions of our work are as follows: To learn representations generalizable to various downstream tasks, we introduce the principle of maximum entropy into self-supervised learning. Maximum Entropy Coding Batch-wise Objectives Feature-wise Objectives batch dimension feature dimension ip OJa E1Ev NYNg Os KGe C1j Tn DYTSXEUc No IBlfjv HFPp WKxu NXDh Po R7gk WMo K1se4cx26z MLQdp1Moui V3Ir QI3gy KFx/2ef L2ZVc7hc92Ny Zp RIUm HCv V8tx E+xm Wmh FOR3Y7VTBZIB7t GVQ4Igq P5t MPEJHxumi MJbm CY0m7u+ODEd KDa PAVEZY9V8Njb/y1qp Ds/8j Ik1VSQ6Udhyp GO0Xh91GWSEs2HBj CRz My KSB9LTLQ5km2O4M2v Aj1k5JXLnk3br Fy CVPl4QAO4Rg8OIUKXEMVak BAw AM8wb Olr Efrx Xqdluas Wc8+/JH1/g NAp JW() LSim Siam MEC = Tr(µλZ1Z> () LBarlow T wins Figure 2: Illustration of our MEC and its relation to batch-wise and feature-wise objectives. We propose Maximum Entropy Coding (MEC), which explicitly optimizes on representations based on the principle of maximum entropy, and leverages the minimal coding length in lossy data coding as a computationally tractable surrogate for the entropy. We reformulate the log-determinant term in coding length function into a scalable form, which makes large-scale pre-training possible, and unifies existing batch-wise and feature-wise objectives as low-order approximations of our method. We show MEC representation generalizes well on a wide variety of imageand video-based downstream tasks, achieving state-of-the-arts on most tasks considered. In this section, we start with illustrating the maximum entropy principle, and then introduce a computationally tractable surrogate of the information entropy for high-dimensional vectors. We then present a scalable reformulation of the proposed surrogate, which makes large-scale training possible. And we further incorporate the view consistency prior for maximum entropy coding. Finally, we demonstrate how the proposed method can unify existing batch-wise and feature-wise SSL objectives. Please refer to Appendix E for more details about the proofs in this section. 2.1 Maximum Entropy Coding The maximum entropy principle. The main purpose of this work is to improve the generalization capacity of self-supervised learning representations across unseen downstream tasks and data distributions, and reduce the biases brought by specifically designed pretext tasks as much as possible. This naturally raises a question, i.e., what makes for a generalizable representation? To answer this question, we are particularly inspired by the maximum entropy principle in information theory, which states that the probability distribution that best represents the current state of knowledge about a system is the one with largest entropy, given a testable information (such as accuracy) and in this way no additional bias or assumptions is introduced [38, 39, 59]. We therefore hypothesis that a generalizable representation is the one that has the maximum entropy among all plausible representations. Intuitively, if we are able to express the entropy in a closed form, the maximum entropy principle then can serve as an optimization objective and supervise the representation learning. Minimal coding length as a surrogate for entropy. Entropy is originally defined on probability distributions [63], i.e., H(z) R p(z) log p(z)dz, for continuous random variables. However, it is very difficult to estimate the true distributions p(z) of a representation [6, 54], from a finite set of high dimensional vectors Z = z1, z2, . . . , zm Rd m. A handy fact is that entropy is conceptually equivalent to the minimal number of bits required to encode the data losslessly, so the minimal lossless coding length could be used to represent the entropy. However, lossless coding of continuous random variables is infeasible in our case since it often requires an infinite number of bits, breaking the numerical stability. Instead, we exploit the coding length in lossy data coding [16] as a computationally tractable surrogate for the entropy of continuous random variables. Given a set of samples Z, the minimal number of bits needed to encode Z subject to a distortion ϵ is given by the Figure 4: Effects of the distortion measure ϵ on maximum entropy coding (MEC). (a): Encoding the representations is akin to packing ϵ-balls into the representation space. (b): T-SNE [71] visualization of the representations learned with large ϵ (left plot, ϵ2 d = 0.12) and small ϵ (right plot, ϵ2 d = 0.01). (c): Linear and k NN accuracy and the spectral norm w.r.t the degree of distortion ϵ. following coding length function [47, 72]: log det Im + d mϵ2 Z Z , (1) where Im denotes the identity matrix with dimension m, and ϵ is the upper bound of the expected decoding error between z Z and the decoded bz, i.e., E [ z bz 2] ϵ. We note that the computation of log-determinant of high dimensional matrix in Equation (1) is highly expensive and may cause numerically unstable results for ill-conditioned matrix, which inhibits its application to large-scale pre-training (e.g., over 1 million images). Therefore, a scalable and stable reformulation of Equation (1) is required. We first rewrite Equation (1) as L = µ log det Im + λZ Z , where µ = m+d 2 and λ = d mϵ2 . Utilizing the identical equation det(exp(A)) = exp(Tr(A)) [35], we obtain L = Tr µ log Im + λZ Z , where Tr stands for the trace of the matrix. Finally, we apply Taylor series expansion to expand the logarithm of the matrix and obtain 256 512 1024 2048 dim time (ms) (log-scale) [0.24%] [0.18%] [0.11%] approx. origin Figure 3: Comparison of running time and relative approximation error between Equation (1) (origin) and Equation (2) (approx.) for different number of samples in Z (dim). with convergence condition: λZ Z 2 < 1, and this can be achieved by adjusting the hyperparameter ϵ (detailed in Section 2.2). Compared to Equation (1), there are only matrix multiplication and addition in Equation (2), which significantly speeds up the computation process and avoids the numerical unstability. To verify this, in Figure 3, we make a comparison of running time and relative approximation error between Equation (1) and (2) (using the first four terms). The results show that our reformulation can approximate the original coding length function with negligible error (all errors are well below 0.5%), and accelerate the computation considerably (over 50 acceleration for all cases), thus making large-scale pre-training possible. 2.2 Combining with the View Consistency Prior It should be noted that the necessary premise of the maximum entropy principle is that testable information is given as prior. For example, the testable information could be the accuracy of a predictive model: the most generalizable model should be the one with maximum entropy, but it is only when a group of models reaches a given accuracy. Otherwise, simply optimizing towards maximum entropy will lead to trivial solutions such as uniform distributions. In Equation (1), the prior is not considered. To introduce a prior, we employ a common practice in SSL by augmenting Z into two different views Z1 and Z2. More specifically, given a set of images D, an image x is sampled uniformly from D, and two augmented views x1 and x2 are obtained from x via a distribution of data augmentations T . Then they are fed to an encoder f, consisting of a backbone and a projector network, which produces ℓ2-normalized embeddings of z1 and z2. For a batch of m images, we have Z1 = z1 1, z2 1, . . . , zm 1 and similarly for Z2, which are two observations of the same Z. MEC aims to minimize the following loss: LMEC = µ log det Im + λZ 1 Z2 Tr λZ 1 Z2 k ! where the same notations in Section 2.1 apply, and n is the order of Taylor expansion. Compared with Equation (1), the formulation in (3) considers not only maximizing entropy, but also the view consistency prior mined from the data itself, therefore learning meaningful representations. As noted in Section 2.1, the convergence condition of Taylor expansion requires C 2 < 1, where C = λZ 1 Z2 and λ = d mϵ2 = 1 mϵ2 d . We show such condition can be strictly satisfied by setting ϵ2 d > 1 because of the inequality C 2 p C 1 C < 1. In practice, we empirically find that the Taylor expansion converges over a wide range of ϵd (Figure 4(c)) with a linear warm-up. From the preliminary experiments on CIFAR-10 [41] (detailed in Appendix B), we also find that the distributions of representations show progressive finer granularity as ϵd decreases (Figure 4(b)). This can be interpreted by the practical meaning of the distortion ϵd (Figure 4(a)), i.e., a smaller ϵd encourages the representation space to be encoded in finer granularity (and hence more uniform). By contrast, a small ϵd might break the semantic structures of similar images (i.e., tolerance). Therefore, a good choice of ϵd is needed to compromise the uniformity and tolerance properties [75, 74] of representations, which shares the same role as the temperature [80] term in contrastive learning. An overview of MEC is illustrated in Figure 2 and a Py Torch-like pseudocode is provided in Appendix A. The algorithm describes the minimalist variant of MEC, which can be further improved by integrating momentum encoder and asymmetric networks (detailed in experiments). 2.3 A Unified View of Batch-wise and Feature-wise SSL Objectives Current SSL methods based on Siamese networks can be roughly divided into two categories: batchwise methods [12, 31, 13, 15, 14, 10] and feature-wise methods [86, 5, 24, 36]. The former aims to minimize the distance between augmented views of the same sample while maximizing the distance between different samples, which can be viewed as decorrelating the different features in a batch. The latter, in contrast, tries to decorrelate the different vector components in the representation. The relationship between them has not been fully understood. Our work builds bridges between these two types of methods through the following derivation: LMEC = µ log det Im + λZ 1 Z2 | {z } batch-wise = µ log det Id + λZ1Z 2 | {z } feature-wise which can be proved since Z 1 Z2 Rm m and Z1Z 2 Rd d have the same nonzero eigenvalues. In Figure 2, under the framework of MEC, we show the equivalence between batch-wise and featurewise methods using two examples, Sim Siam [14] and Barlow Twins [86]. By taking Taylor expansion (Equation (2)) of the left side of Equation (4) and before the trace operation, the diagonal elements of the leading term (i.e., µλZ 1 Z2) measure the similarity between the views of the same images in a batch, and the objective of Sim Siam [14] is equivalent to maximizing the trace of this term. Similarly, the leading term of the right side expansion models the correlation between dimensions of the feature, and the objective of Barlow Twins [86] is equivalent to the second-order expansion of LMEC. With the above derivation, our method naturally subsumes the two different kinds of objectives as its low-order expansions, and we show in experiments that better downstream task performance can be achieved with higher-order approximations. We further show in Appendix E that our MEC can also bridge other self-supervised objectives. And we hope the direct tying of a family of objectives to a very grounded mathematical concept can inspire more new methods. 3 Experiments We perform self-supervised pre-training using the proposed MEC on the training set of the Image Net ILSVRC-2012 dataset [17]. After pre-training, we conduct extensive experiments to examine Table 1: Linear evaluation. All methods are based on standard Res Net-50 [33] pre-trained with two 224 224 views on Image Net training dataset. We perform pre-training with four different length of epochs following [14]. Method Pre-training epochs 100 200 400 800 Sim CLR [12] 66.5 68.3 69.8 70.4 Mo Co v2 [13] 67.4 69.9 71.0 72.2 BYOL [29] 66.5 70.6 73.2 74.3 Sw AV [10] 66.5 69.1 70.7 71.8 Sim Siam [14] 68.1 70.0 70.8 71.3 Barlow [86] 67.3 70.2 71.8 73.0 MEC 70.6 71.9 73.5 74.5 Table 2: Semi-supervised classification. We finetune the pre-trained model using 1% and 10% training samples of Image Net following [12, 29], and the top-1 and top-5 accuracy on Image Net val dataset are reported. Method 1% 10% Top 1 Top 5 Top 1 Top 5 Supervised 25.4 48.4 56.4 80.4 Sim CLR [12] 48.3 75.5 65.6 87.8 BYOL [29] 53.2 78.4 68.8 89.0 Barlow [86] 55.0 79.2 69.7 89.3 DINO [11] 52.2 78.2 68.2 89.1 VICReg [5] 54.8 79.4 69.5 89.5 MEC 55.9 79.6 70.3 89.7 Table 3: Transfer learning on object detection and instance segmentation tasks. We fine-tune the pre-trained model end-to-end on target datasets and tasks, following the standard protocol described in [13, 14, 29, 86]. We use Faster R-CNN [61] for VOC detection tasks and Mask R-CNN [32] (1 schedule) for COCO detection and instance segmentation tasks. All Faster/Mask R-CNN models are with the C4-backbone [79]. Bold entries are within 0.2 below the best. VOC 07 detection VOC 07+12 detection COCO detection COCO instance seg. Pre-train AP50 AP AP75 AP50 AP AP75 AP50 AP AP75 APmask 50 APmask APmask 75 Scratch 35.9 16.8 13.0 60.2 33.8 33.1 44.0 26.4 27.8 46.9 29.3 30.8 Supervised 74.4 42.4 42.7 81.3 53.5 58.8 58.2 38.2 41.2 54.7 33.3 35.2 Sim CLR [12] 75.9 46.8 50.1 81.8 55.5 61.4 57.7 37.9 40.9 54.6 33.3 35.3 Mo Co v2 [13] 77.1 48.5 52.5 82.5 57.4 64.0 58.9 39.3 42.5 55.8 34.4 36.5 BYOL [29] 77.1 47.0 49.9 81.4 55.3 61.1 57.8 37.9 40.9 54.3 33.2 35.0 Sw AV [10] 75.5 46.5 49.6 82.6 56.1 62.7 58.6 38.4 41.3 55.2 33.8 35.9 Barlow [86] 75.7 47.2 50.3 82.6 56.8 63.4 59.0 39.2 42.5 56.0 34.3 36.5 Sim Siam [14] 77.3 48.5 52.5 82.4 57.0 63.7 59.3 39.2 42.1 56.0 34.4 36.7 MEC 77.4 48.3 52.3 82.8 57.5 64.5 59.8 39.8 43.2 56.3 34.7 36.8 the learned representations on various tasks. The tasks considered include four image-level tasks: Image Net linear probing, Image Net semi-supervised classification, object detection, and instance segmentation; and five video-based tasks: single object tracking (SOT) [78], video object segmentation (VOS) [56], multi-object detection (MOT) [48], multi-object detection and segmentation (MOTS) [73] and pose tracking (Pose Track) [1]. An overview comparison with existing SSL representations across different tasks is presented in Section 3.1. Moreover, in Section 3.4, we present extensive ablations studies to demonstrate the behaviors and properties of the proposed method. All the implementation details of the aforementioned experiments are deferred to Appendix C. 3.1 Overview Comparison We evaluate MEC together with multiple previous SSL representations on the four image-based tasks and the five video-based tasks, plot their performance ranking along a axis for a specific task and obtain two sets of radar-like charts in Figure 1. Intuitively, it can be seen that the larger a polygon in a radar chart is, the better the corresponding representation generalize across downstream tasks. MEC consistently outperforms previous representations on both imageand video-based tasks considered, suggesting its good generalization capacity to various downstream tasks and data distributions. Below we detail experiments on each task. And additional experiment results can be found in Appendix D, including pre-training with different strategies and datasets, transfer learning on fine-grained classification benchmarks. 3.2 Evaluation on Image Net Linear probing. We train a linear classifier on top of frozen representations of the pre-trained model on the Image Net training set, and report the top-1 accuracy on the Image Net validation set, which is a 0 25 50 75 100 125 150 175 200 epochs simsiam w/o mec simsiam w/ mec 0 25 50 75 100 125 150 175 200 epochs simclr w/o mec simclr w/ mec method MEC k NN linear Sim Siam [14] 83.5 86.1 84.9 87.3 Sim CLR [12] 84.6 85.9 87.0 88.2 Figure 5: Effects of MEC regulation on other SSL objectives on CIFAR-10 [41]. Left plot: accuracy of a k NN classifier as a monitor of the pre-training process. Right table: linear probing and k NN classification accuracy of the pre-trained models with v.s. without our MEC regulation. standard and important protocol in SSL [88, 12, 31, 14, 86]. We evaluate the learned representations of models with four different pre-training epochs following [14], and make comparisons in Table 1 with state-of-the-art methods, whose results are reported in [14] and better than their original papers. Even so, our method MEC achieves highest accuracy under all pre-training epochs among those methods. In particular, with only 200-epoch pre-training, MEC outperforms several 800-epoch pre-trained methods, including Sim CLR [12], Sw AV [10] and Sim Siam [14], which indicates the pretraining efficiency of our method. Additionally, by incorporating other pre-training strategies, such as multi-crop [10, 11], our method can attain even better results, which are detailed in Appendix D. Furthermore, for a fixed given downstream task (i.e., image classification), MEC also improves generalization across different data distributions, which can be concluded from the transfer learning experiments on 11 fine-grained classification datasets in Appendix D. Semi-supervised classification. We fine-tune the pre-trained model on a small subset of Image Net for classification task. Specifically, we adopt the same fixed splits of 1% and 10% of Image Net training set as in [12] and report both top-1 and top-5 accuracies in Table 2. The experiment results show that our method consistently outperforms previous approaches in both 1% and 10% settings. And we also note that all SSL methods surpass the supervised counterparts by a large margin (e.g., over 20% accuracy using 1% labeled data), which emphasizes the importance of leveraging large amount of unlabeled data. 3.3 Transfer learning Object detection and instance segmentation. We transfer the pre-trained model by end-to-end fine-tuning on object detection and instance segmentation tasks, following the common practice of previous methods [31, 13, 14, 29, 86]. We report the results in Table 3. Similar to the observations in semi-supervised classification, the SSL methods outperform or are on par with the supervised counterparts in all tasks. Our method achieves better performance than previous approaches in most of the tasks and metrics considered. In particular, MEC is significantly better than the supervised baseline (+1.6 AP), Sim CLR (+1.9 AP), and previous best method Mo Co v2 (+0.5 AP) on COCO detection task. These results validate that the model pre-trained by MEC can generalize very well across different image-based tasks and datasets. Video-based tasks. To further evaluate whether MEC s representation generalizes beyond imagebased tasks, we transfer the pre-trained model to a series of video tasks based on a recently proposed evaluation platform named Uni Track [76]. In contrast to above evaluation processes, Uni Track does not require any additional training and hence provides a more direct comparison between different representations. We report the results in Table 4 and compare different methods using radar plots in Figure 1. As noted in Uni Track, there is no significant correlation between linear probe accuracy and tracking performance. However, our method achieves strong performance across linear probing and all five video tracking tasks. Specifically, MEC achieves the highest rank on four our of five tasks, and ranks among the top 2 methods on seven metrics out of all nine metrics. These results substantiate that the learned representations of MEC are more general-purpose and versatile. We attribute this important property of MEC to the principle of maximum entropy, the basis of our method, which enables the model to make best use of data for unseen downstream tasks during pre-training. Table 4: Transfer learning on five video tracking tasks. All methods use the same Res Net-50 backbone [33] and are evaluated based on Uni Track [76]. For each cell, we report the results obtained from features at [layer3 / layer4], and the best performance between the two is bolded. We then use the best of the two to rank the top 2 models (underlined) in each column. Pre-train SOT [78] VOS [56] MOT [48] MOTS [73] Pose Track [1] AUCXCorr AUCDCF J -mean IDF1 HOTA IDF1 HOTA IDF1 IDs Rand. Init. 10.3 / 9.0 28.0 / 20.0 29.3 / 33.9 8.4 / 8.9 8.4 / 8.5 20.8 / 23.1 25.9 / 28.7 40.2 / 38.5 88792 / 90963 Supervised 58.6 / 49.5 62.0 / 53.9 62.3 / 57.9 75.6 / 73.2 63.3 / 61.8 68.4 / 69.4 70.2 / 71.0 73.7 / 73.3 6969 / 7103 Ins Dis [80] 47.6 / 47.3 61.8 / 51.1 62.6 / 60.1 66.7 / 73.9 57.9 / 61.9 68.4 / 68.0 69.6 / 70.3 72.4 / 73.9 7106 / 7015 Mo Co v1 [31] 50.9 / 47.9 62.2 / 53.7 61.5 / 57.9 69.2 / 74.1 59.4 / 61.9 70.6 / 69.3 71.6 / 70.9 72.8 / 73.9 6872 / 7092 Sim CLR [12] 47.3 / 51.9 61.3 / 50.7 60.5 / 56.5 66.9 / 75.6 57.7 / 63.2 65.8 / 67.6 67.7 / 69.5 72.3 / 73.5 7084 / 7367 Mo Co v2 [13] 53.7 / 47.2 61.5 / 53.3 61.2 / 54.0 72.0 / 74.9 61.2 / 62.8 67.5 / 67.3 69.6 / 69.6 73.0 / 73.7 6932 / 7702 Infomin [68] 48.5 / 46.8 61.2 / 51.9 58.4 / 51.1 66.7 / 73.4 57.6 / 61.9 66.7 / 66.3 68.5 / 68.8 72.5 / 74.0 7066 / 7901 Bar Low [86] 44.5 / 55.5 60.5 / 60.1 61.7 / 57.8 63.7 / 74.5 55.4 / 62.4 68.7 / 67.4 69.5 / 69.8 72.3 / 74.3 7131 / 7456 BYOL [29] 48.3 / 55.5 58.9 / 56.8 58.8 / 54.3 65.3 / 74.9 56.8 / 62.9 70.1 / 66.8 70.8 / 69.3 72.4 / 73.8 7213 / 8032 Sw AV [10] 49.2 / 52.4 61.5 / 59.4 59.4 / 57.0 65.6 / 74.4 56.9 / 62.3 68.8 / 67.0 69.9 /69.5 72.7 / 73.6 7025 / 7377 Pix Pro [82] 40.5 / 49.2 57.4 / 49.3 56.4 / 52.2 61.7 / 67.7 54.3 / 58.6 64.2 / 66.2 65.1 / 67.6 72.4 / 73.1 7163 / 6953 Det Co [81] 55.0 / 47.1 59.0 / 53.2 62.3 / 56.1 75.3 / 72.9 62.8 / 61.6 67.8 / 66.8 70.0 / 69.4 73.9 / 73.3 7357 / 8009 MEC 56.8 / 54.7 62.3 / 60.6 62.0 / 57.5 72.7 / 75.0 61.2 / 63.3 70.7 / 66.5 72.0 / 69.0 73.1 / 74.1 6665 / 7752 Table 5: MEC ablation experiments with Res Net-50 on Image Net. We report the results of linear probing. The default settings of these experiments are marked in gray . (a) Smaller epochs. epochs 30 50 70 100 acc. 62.1 67.5 69.2 70.6 (b) Smaller batch sizes. batch 128 256 512 1024 acc. 69.5 70.1 70.4 70.6 (c) The distortion ϵd. ϵ2 d 0.02 0.06 0.25 1.0 acc. 70.5 70.6 70.2 69.5 (d) The momentum coefficient. m 0 0.9 0.99 0.996 0.999 acc. 69.0 70.1 70.4 70.6 70.2 (e) The order of expansion. order 1 2 3 4 5 acc. 70.0 70.3 70.5 70.6 70.6 (f) The projector network. proj asym. sym. 4096 8192 acc. 70.6 70.1 70.7 71.2 3.4 Ablation studies Batch size and training epoch. The computation and memory resources required for training self-supervised models are demanding. Previous methods often need a large batch size (e.g., 4096 in Sim CLR [12]) and long training time (e.g., 1000 epochs in Barlow Twins [86]) to work well. We hence test the efficiency of our method by pre-training the model with smaller training epochs (Table 5a) and batch sizes (Table 5b). The results show that our method can effectively avoid those prohibitive requirements. In particular, Our 50-epoch pre-trained model has already outperformed those 100-epoch pre-trained models of other methods (e.g., Sim CLR [12], BYOL [29], Barlow Twins [86]). Moreover, MEC is more robust to different batch sizes, and even a batch size of 128 performs well (with 1.1% accuracy drop). These results are noticeably different from those methods based on the pretext task of instance discrimination, where a large number of negative samples is required (e.g., over 4% accuracy drop of Sim CLR with a batch size of 128). This behavior also liberates MEC from those complex designs, such as memory queue [13] and online clustering [10]. Siamese networks play a critical role in discriminative SSL, and different methods often rely on different architectures to prevent mode collapse. We investigate the effects of different designs of Siamese networks by varying the momentum coefficient (Table 5d) and the projector network (Table 5f). Compared to other methods [13, 14, 29, 86], MEC still works well when the Siamese networks are direct weight-sharing (m=0) and symmetric (sym.). In contrast, BYOL [29] requires the momentum encoder to prevent collapse (0.3% accuracy without it), and asymmetric architectures are essential for Sim Siam [14] to work. These results demonstrate that the objective of MEC naturally avoids representational collapse without relying on special design of Siamese networks. We also observe that the performance can be further improved with a larger projector network, consistent with the conclusions in previous methods [86, 5]. Taylor series approximation. The computation of log-determinant of a high-dimensional matrix can be significantly accelerated by exploiting Taylor series approximation. And we study two important hyper-parameters of the process, i.e., the order of Taylor expansion (Table 5e) and the distortion Table 6: Comparison of SSL methods with different architectures. (Conv Nets v.s. Transformer) All methods are pre-trained with two 224 224 crops on Image Net training set, and we report the results of linear probing on Image Net validation set. model MEC i BOT [90] MAE [30] DINO [11] Mo Co v3 [15] Sim CLR [12] BYOL [29] Sw AV [10] R-50, 800-ep 74.5 - - - 73.8 70.4 74.3 71.8 Vi T-S, 300-ep 73.4 73.2 68.2 70.9 72.5 69.0 71.0 67.1 Vi T-B, 300-ep 76.5 75.6 69.3 72.6 76.5 73.9 73.9 71.6 measure ϵd (Table 5c). We find that the accuracy increases with higher order of expansion, since it produces more accurate approximation. Besides the linear probing, we find the COCO detection performance can be improved by a large margin (+1.5 AP) when the order is increased from 1 to 4. We also note that the extra two orders (from 2 to 4) benefit less than lifting from the first order to the fourth order approximation. It is reasonable since the relative approximation error of second-order expansion is already quite decent, lower than 0.5% (Figure 6), almost the same as fourth-order expansion (Figure 3). As for the distortion hyper-parameter, on one hand, it should be large enough so that the Taylor expansion is convergent, and we empirically observe this condition is satisfied over a wide range of degree of distortion. On the other hand, a good choice of ϵd is also needed to improve the representation quality, because it compromises the uniformity and tolerance properties of representations as depicted in Figure 4 and discussed in Section 2.2. Generalization across different backbones and objectives. Given the recent progress in Vision Transformers (Vi Ts) [21], it is worthwhile to examine the generalization ability of our method on both Vi Ts and Conv Nets, though previous methods usually focus on only one architecture. To this end, we make a straightforward extension by replacing the Conv Net encoder with a Vi T encoder, following the practice of [15]. The comparisons are made in Table 5 with previous methods based on Siamese networks [12, 29, 10, 11, 31], and also recently proposed masked auto-encoding methods [30, 90] tailored for Transformers. And the results show that MEC performs equally well with both architectures. Furthermore, since MEC is designed for optimizing the representation quality, it should be complementary to other methods that focus on pretext tasks. To validate this hypothesis, we apply MEC as a regulation term on the representations of a contrastive method, Sim CLR [12], and a non-contrastive method, Sim Siam [14]. The results in Figure 5 demonstrate that MEC improves both linear and k NN accuracy by a large margin (e.g., 2.3% linear probing accuracy for Sim CLR) with negligible computation overhead ( 1.05 running time). 4 Related Work Maximum entropy principle states that the probability distribution which best represents the current state of knowledge is the one with largest entropy, in the context of precisely stated prior data. This principle originates in Laplace s "Principle of Insufficient Reason" and is popularized by Jaynes [38, 39]. It has a wide variety of applications in different fields [40], ranging from statistical physics [58, 51], economics [2, 62], to reinforcement learning [77, 49] and supervised learning [57, 64, 22]. To the best of our knowledge, the usage of this principle has not been explored in the context of SSL. We notice a closely related work MCR2 [85] exploits the same coding length function (Equation (1)), but the motivation behind is quite different: they do not originate from maximum entropy, but instead base their method on coding rate reduction. Furthermore, MCR2 experiments on classification and clustering tasks, and is difficult to generalize to modern SSL settings due to the expensive and unstable computation. In contrast, our method is tailored for SSL and we reformulate the coding length function into a scalable and stable form for large-scale pre-training. Another closely related concept in information theory is mutual information, which has been widely used in SSL [34, 53, 68, 67, 69] based on the infomax principle [45]. Our method can also be viewed as an application of the infomax principle that maximizes the representation information, but without using mutual information estimators, which might result in worse transfer performance [70]. Pretext tasks are one of the core components in SSL. They refer to a type of tasks that mine a certain structure of the data, and then force learners to predict such a structure from degenerated input data. Labels are mined from the data itself, therefore dispensing with the need of human annotations. In natural language processing, the most commonly used pretext task is masked language modeling (MLM) [18, 60], where a random portion of the input sequence is masked and the model learns to predict these masked content. In computer vision, however, there emerge a much wider variety of pretext tasks [87, 88, 52, 83, 20, 80, 55, 19, 30, 4, 42], since images contain much more information than languages so that there are much more intrinsic structures to be mined as free learning signals Solving pretext tasks requires the model to understand the visual concepts present in images and thus useful representations can be learned. In this work, however, we argue that those handcrafted pretext tasks, like handcrafted feature engineering methods, might introduce undesired bias into the representations. So we pursue a straightforward approach to optimize on representations. And considering that the ultimate goal of self-supervised learning is a general-purpose representation, we explicitly encourage the generalization ability on downstream tasks and minimize the bias in the formulation of the pretext task, by introducing the maximum entropy principle. Siamese representation learning uses two identical neural networks to learn representations by comparing entities, in supervised manner [8, 65, 7], or recently in self-supervised manner [12, 31, 13, 14, 29, 10, 11, 86, 5, 23, 43]. In Siamese SSL, the two branches are weight-sharing [12, 14], or one is a moving average of the other [31, 29]. One crucial problem that needs to be carefully addressed in Siamese SSL is mode collapse, where all data is mapped to the same representation. Contrastive methods, such as Sim CLR [12] and Mo Co [31], prevent the trivial solution by minimizing the distance between augmented views of the same images while maximizing the distance between different images. Clustering-based methods [9, 3, 10] enforces consistency between cluster assignments of the views from same images, where the cluster centers work as negative prototypes. Non-contrastive methods [14, 29] remove the use of negative pairs and Sim Siam [14] concludes that the stop-gradient operation is critical to preventing mode collapse. Our work also leverages the basic view consistency prior inherent in Siamese representation learning, which gives testable information for the application of maximum entropy principle. Compared to previous work, our method of maximum entropy coding naturally avoids the low-entropy mode collapse problem and can be easily implemented without requiring instance discrimination [20, 12], memory queue [80, 31, 13] or online clustering [9, 3, 10]. Furthermore, we find interesting equivalence between low-order approximations of MEC and existing batch-wise or feature-wise objectives, which provides a new perspective for a unified understanding of SSL methods. 5 Conclusion What makes for generalizable representations? In this work, we argue that the best generalizable representation is supposed to be the one that admits the maximum entropy. Based on this assumption, we propose MEC, a novel pretext task that aims explicitly to generalize well. MEC pursues a representation with maximum entropy in all eligible ones that satisfies the view consistency prior of Siamese representation learning. The minimal coding length in lossy data coding is leveraged as a computationally tractable surrogate for the entropy; and in order to facilitate large-scale pretraining, we reformulate the coding length function into a scalable form. With the reformulation, we further demonstrate that MEC bridges the batch-wise and feature-wise SSL objectives. Extensive experiments shows MEC generalizes well across various image/video tasks, various data distributions, and different network architectures. Furthermore, MEC can be readily plugged into existing SSL methods such as Sim CLR and Sim Siam, and brings consistent improvements upon these strong representations. We hope our exploration will inspire the community to rethink the possibility of optimizing generalization explicitly in self-supervised representation learning. 6 Funding Transparency Statement Acknowledgement: This work is supported by the state key development program in 14th Five-Year under Grant No. 2021YFF0602103, 2021YFF0602102, 2021QY1702. We also thank for the research fund under Grant No. 2019GQG0001 from the Institute for Guo Qiang, Tsinghua University. [1] M. Andriluka et al., Posetrack: A benchmark for human pose estimation and tracking, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 5167 5176. [2] C. Arndt, S. Robinson, and F. Tarp, Parameter estimation for a computable general equilibrium model: A maximum entropy approach, Economic Modelling, vol. 19, no. 3, pp. 375 398, 2002. [3] Y. M. Asano, C. Rupprecht, and A. Vedaldi, Self-labelling via simultaneous clustering and representation learning, ar Xiv preprint ar Xiv:1911.05371, 2019. [4] H. Bao, L. Dong, and F. Wei, Beit: Bert pre-training of image transformers, ar Xiv preprint ar Xiv:2106.08254, 2021. [5] A. Bardes, J. Ponce, and Y. Le Cun, Vicreg: Variance-invariance-covariance regularization for self-supervised learning, ar Xiv preprint ar Xiv:2105.04906, 2021. [6] J. Beirlant, E. J. Dudewicz, L. Györfi, E. C. Van der Meulen, et al., Nonparametric entropy estimation: An overview, International Journal of Mathematical and Statistical Sciences, vol. 6, no. 1, pp. 17 39, 1997. [7] L. Bertinetto, J. Valmadre, J. F. Henriques, A. Vedaldi, and P. H. Torr, Fully-convolutional siamese networks for object tracking, in European conference on computer vision, Springer, 2016, pp. 850 865. [8] J. Bromley, I. Guyon, Y. Le Cun, E. Säckinger, and R. Shah, Signature verification using a" siamese" time delay neural network, Advances in neural information processing systems, vol. 6, 1993. [9] M. Caron, P. Bojanowski, A. Joulin, and M. Douze, Deep clustering for unsupervised learning of visual features, in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 132 149. [10] M. Caron, I. Misra, J. Mairal, P. Goyal, P. Bojanowski, and A. Joulin, Unsupervised learning of visual features by contrasting cluster assignments, Advances in Neural Information Processing Systems, vol. 33, pp. 9912 9924, 2020. [11] M. Caron et al., Emerging properties in self-supervised vision transformers, in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 9650 9660. [12] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton, A simple framework for contrastive learning of visual representations, in International conference on machine learning, PMLR, 2020, pp. 1597 1607. [13] X. Chen, H. Fan, R. Girshick, and K. He, Improved baselines with momentum contrastive learning, ar Xiv preprint ar Xiv:2003.04297, 2020. [14] X. Chen and K. He, Exploring simple siamese representation learning, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 15 750 15 758. [15] X. Chen, S. Xie, and K. He, An empirical study of training self-supervised vision transformers, in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 9640 9649. [16] T. M. Cover, Elements of information theory. John Wiley & Sons, 1999. [17] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, Imagenet: A large-scale hierarchical image database, in 2009 IEEE conference on computer vision and pattern recognition, Ieee, 2009, pp. 248 255. [18] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, Bert: Pre-training of deep bidirectional transformers for language understanding, ar Xiv preprint ar Xiv:1810.04805, 2018. [19] C. Doersch, A. Gupta, and A. A. Efros, Unsupervised visual representation learning by context prediction, in Proceedings of the IEEE international conference on computer vision, 2015, pp. 1422 1430. [20] A. Dosovitskiy, J. T. Springenberg, M. Riedmiller, and T. Brox, Discriminative unsupervised feature learning with convolutional neural networks, Advances in neural information processing systems, vol. 27, 2014. [21] A. Dosovitskiy et al., An image is worth 16x16 words: Transformers for image recognition at scale, ar Xiv preprint ar Xiv:2010.11929, 2020. [22] A. Dubey, O. Gupta, R. Raskar, and N. Naik, Maximum-entropy fine grained classification, Advances in neural information processing systems, vol. 31, 2018. [23] D. Dwibedi, Y. Aytar, J. Tompson, P. Sermanet, and A. Zisserman, With a little help from my friends: Nearest-neighbor contrastive learning of visual representations, in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 9588 9597. [24] A. Ermolov, A. Siarohin, E. Sangineto, and N. Sebe, Whitening for self-supervised representation learning, in International Conference on Machine Learning, PMLR, 2021, pp. 3015 3024. [25] M. Everingham, L. Van Gool, C. K. Williams, J. Winn, and A. Zisserman, The pascal visual object classes (voc) challenge, International journal of computer vision, vol. 88, no. 2, pp. 303 338, 2010. [26] D. Fu et al., Unsupervised pre-training for person re-identification, CVPR, 2021. [27] S. Gidaris, P. Singh, and N. Komodakis, Unsupervised representation learning by predicting image rotations, ar Xiv preprint ar Xiv:1803.07728, 2018. [28] P. Goyal et al., Accurate, large minibatch sgd: Training imagenet in 1 hour, ar Xiv preprint ar Xiv:1706.02677, 2017. [29] J.-B. Grill et al., Bootstrap your own latent-a new approach to self-supervised learning, Advances in Neural Information Processing Systems, vol. 33, pp. 21 271 21 284, 2020. [30] K. He, X. Chen, S. Xie, Y. Li, P. Dollár, and R. Girshick, Masked autoencoders are scalable vision learners, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022. [31] K. He, H. Fan, Y. Wu, S. Xie, and R. Girshick, Momentum contrast for unsupervised visual representation learning, in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 9729 9738. [32] K. He, G. Gkioxari, P. Dollár, and R. Girshick, Mask r-cnn, in Proceedings of the IEEE international conference on computer vision, 2017, pp. 2961 2969. [33] K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770 778. [34] R. D. Hjelm et al., Learning deep representations by mutual information estimation and maximization, ar Xiv preprint ar Xiv:1808.06670, 2018. [35] R. A. Horn and C. R. Johnson, Matrix analysis. Cambridge university press, 2012. [36] T. Hua, W. Wang, Z. Xue, S. Ren, Y. Wang, and H. Zhao, On feature decorrelation in selfsupervised learning, in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 9598 9608. [37] S. Ioffe and C. Szegedy, Batch normalization: Accelerating deep network training by reducing internal covariate shift, in International conference on machine learning, PMLR, 2015, pp. 448 456. [38] E. T. Jaynes, Information theory and statistical mechanics, Physical review, vol. 106, no. 4, p. 620, 1957. [39] E. T. Jaynes, Information theory and statistical mechanics. ii, Physical review, vol. 108, no. 2, p. 171, 1957. [40] J. N. Kapur, Maximum-entropy models in science and engineering. John Wiley & Sons, 1989. [41] A. Krizhevsky, Learning multiple layers of features from tiny images, Master s thesis, University of Tront, 2009. [42] H. Lee, K. Lee, K. Lee, H. Lee, and J. Shin, Improving transferability of representations via augmentation-aware self-supervision, in Advances in Neural Information Processing Systems, A. Beygelzimer, Y. Dauphin, P. Liang, and J. W. Vaughan, Eds., 2021. [Online]. Available: https://openreview.net/forum?id=U34r Qjn Imp M. [43] K.-H. Lee, A. Arnab, S. Guadarrama, J. Canny, and I. Fischer, Compressive visual representations, Advances in Neural Information Processing Systems, vol. 34, pp. 19 538 19 552, 2021. [44] T.-Y. Lin et al., Microsoft coco: Common objects in context, in European conference on computer vision, Springer, 2014, pp. 740 755. [45] R. Linsker, Self-organization in a perceptual network, Computer, vol. 21, no. 3, pp. 105 117, 1988. [46] I. Loshchilov and F. Hutter, Sgdr: Stochastic gradient descent with warm restarts, ar Xiv preprint ar Xiv:1608.03983, 2016. [47] Y. Ma, H. Derksen, W. Hong, and J. Wright, Segmentation of multivariate mixed data via lossy data coding and compression, IEEE transactions on pattern analysis and machine intelligence, vol. 29, no. 9, pp. 1546 1562, 2007. [48] A. Milan, L. Leal-Taixé, I. Reid, S. Roth, and K. Schindler, Mot16: A benchmark for multiobject tracking, ar Xiv preprint ar Xiv:1603.00831, 2016. [49] V. Mnih et al., Asynchronous methods for deep reinforcement learning, in International conference on machine learning, PMLR, 2016, pp. 1928 1937. [50] V. Nair and G. E. Hinton, Rectified linear units improve restricted boltzmann machines, in Icml, 2010. [51] T. K. Nakamura, Statistical mechanics of a collisionless system based on the maximum entropy principle, The Astrophysical Journal, vol. 531, no. 2, p. 739, 2000. [52] M. Noroozi and P. Favaro, Unsupervised learning of visual representations by solving jigsaw puzzles, in European conference on computer vision, Springer, 2016, pp. 69 84. [53] A. v. d. Oord, Y. Li, and O. Vinyals, Representation learning with contrastive predictive coding, ar Xiv preprint ar Xiv:1807.03748, 2018. [54] L. Paninski, Estimation of entropy and mutual information, Neural computation, vol. 15, no. 6, pp. 1191 1253, 2003. [55] D. Pathak, P. Krahenbuhl, J. Donahue, T. Darrell, and A. A. Efros, Context encoders: Feature learning by inpainting, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 2536 2544. [56] F. Perazzi, J. Pont-Tuset, B. Mc Williams, L. Van Gool, M. Gross, and A. Sorkine-Hornung, A benchmark dataset and evaluation methodology for video object segmentation, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 724 732. [57] G. Pereyra, G. Tucker, J. Chorowski, Ł. Kaiser, and G. Hinton, Regularizing neural networks by penalizing confident output distributions, ar Xiv preprint ar Xiv:1701.06548, 2017. [58] S. Pressé, K. Ghosh, J. Lee, and K. A. Dill, Principles of maximum entropy and maximum caliber in statistical physics, Reviews of Modern Physics, vol. 85, no. 3, p. 1115, 2013. [59] Principle of Maximum Entropy, https://en.wikipedia.org/wiki/Principle_of_ maximum_entropy. [60] A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, Improving language understanding by generative pre-training, 2018. [61] S. Ren, K. He, R. Girshick, and J. Sun, Faster r-cnn: Towards real-time object detection with region proposal networks, Advances in neural information processing systems, vol. 28, pp. 91 99, 2015. [62] E. Scharfenaker and J. Yang, Maximum entropy economics, The European Physical Journal Special Topics, vol. 229, no. 9, pp. 1577 1590, 2020. [63] C. E. Shannon, A mathematical theory of communication, The Bell system technical journal, vol. 27, no. 3, pp. 379 423, 1948. [64] J. Shawe-Taylor and D. Hardoon, Pac-bayes analysis of maximum entropy classification, in Artificial Intelligence and Statistics, PMLR, 2009, pp. 480 487. [65] Y. Taigman, M. Yang, M. Ranzato, and L. Wolf, Deepface: Closing the gap to human-level performance in face verification, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2014, pp. 1701 1708. [66] C. Tao et al., Exploring the equivalence of siamese self-supervised learning via a unified gradient framework, ar Xiv preprint ar Xiv:2112.05141, 2021. [67] Y. Tian, D. Krishnan, and P. Isola, Contrastive multiview coding, in European conference on computer vision, Springer, 2020, pp. 776 794. [68] Y. Tian, C. Sun, B. Poole, D. Krishnan, C. Schmid, and P. Isola, What makes for good views for contrastive learning? ar Xiv preprint ar Xiv:2005.10243, 2020. [69] Y.-H. H. Tsai, M. Q. Ma, M. Yang, H. Zhao, L.-P. Morency, and R. Salakhutdinov, Self-supervised representation learning with relative predictive coding, ar Xiv preprint ar Xiv:2103.11275, 2021. [70] M. Tschannen, J. Djolonga, P. K. Rubenstein, S. Gelly, and M. Lucic, On mutual information maximization for representation learning, ar Xiv preprint ar Xiv:1907.13625, 2019. [71] L. Van der Maaten and G. Hinton, Visualizing data using t-sne., Journal of machine learning research, vol. 9, no. 11, 2008. [72] R. Vidal, Y. Ma, and S. Sastry, Generalized principal component analysis (gpca), IEEE transactions on pattern analysis and machine intelligence, vol. 27, no. 12, pp. 1945 1959, 2005. [73] P. Voigtlaender et al., Mots: Multi-object tracking and segmentation, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 7942 7951. [74] F. Wang and H. Liu, Understanding the behaviour of contrastive loss, in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 2495 2504. [75] T. Wang and P. Isola, Understanding contrastive representation learning through alignment and uniformity on the hypersphere, in International Conference on Machine Learning, PMLR, 2020, pp. 9929 9939. [76] Z. Wang, H. Zhao, Y.-L. Li, S. Wang, P. Torr, and L. Bertinetto, Do different tracking tasks require different appearance models? Advances in Neural Information Processing Systems, vol. 34, 2021. [77] R. J. Williams and J. Peng, Function optimization using connectionist reinforcement learning algorithms, Connection Science, vol. 3, no. 3, pp. 241 268, 1991. [78] Y. Wu, J. Lim, and M.-H. Yang, Online object tracking: A benchmark, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2013, pp. 2411 2418. [79] Y. Wu, A. Kirillov, F. Massa, W.-Y. Lo, and R. Girshick, Detectron2, https://github.com/ facebookresearch/detectron2, 2019. [80] Z. Wu, Y. Xiong, S. X. Yu, and D. Lin, Unsupervised feature learning via non-parametric instance discrimination, in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3733 3742. [81] E. Xie et al., Detco: Unsupervised contrastive learning for object detection, in ar Xiv:2102.04803, 2021. [82] Z. Xie, Y. Lin, Z. Zhang, Y. Cao, S. Lin, and H. Hu, Propagate yourself: Exploring pixel-level consistency for unsupervised visual representation learning, in CVPR, 2021. [83] D. Xu, J. Xiao, Z. Zhao, J. Shao, D. Xie, and Y. Zhuang, Self-supervised spatiotemporal learning via video clip order prediction, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 10 334 10 343. [84] Y. You, I. Gitman, and B. Ginsburg, Large batch training of convolutional networks, ar Xiv preprint ar Xiv:1708.03888, 2017. [85] Y. Yu, K. H. R. Chan, C. You, C. Song, and Y. Ma, Learning diverse and discriminative representations via the principle of maximal coding rate reduction, Advances in Neural Information Processing Systems, vol. 33, pp. 9422 9434, 2020. [86] J. Zbontar, L. Jing, I. Misra, Y. Le Cun, and S. Deny, Barlow twins: Self-supervised learning via redundancy reduction, in International Conference on Machine Learning, PMLR, 2021, pp. 12 310 12 320. [87] R. Zhang, P. Isola, and A. A. Efros, Colorful image colorization, in European conference on computer vision, Springer, 2016, pp. 649 666. [88] R. Zhang, P. Isola, and A. A. Efros, Split-brain autoencoders: Unsupervised learning by cross-channel prediction, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 1058 1067. [89] B. Zhou, A. Khosla, A. Lapedriza, A. Torralba, and A. Oliva, Places: An image database for deep scene understanding, ar Xiv preprint ar Xiv:1610.02055, 2016. [90] J. Zhou et al., Ibot: Image bert pre-training with online tokenizer, ar Xiv preprint ar Xiv:2111.07832, 2021. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] Please see Appendix F. (c) Did you discuss any potential negative societal impacts of your work? [Yes] Please see Appendix G. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] A pseudocode is provided in Appendix A and code will be released. All instructions are in Appendix C. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] All implementation details are described in Appendix C. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] We report the average performance over three seeds. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] Please see Appendix C. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [Yes] Please see Appendix H. (c) Did you include any new assets either in the supplemental or as a URL? [Yes] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating?[N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]