# cribo_selfsupervised_learning_via_crossimage_objectlevel_bootstrapping__b3ce2248.pdf Published as a conference paper at ICLR 2024 : SELF-SUPERVISED LEARNING VIA CROSSIMAGE OBJECT-LEVEL BOOTSTRAPPING Tim Lebailly1 Thomas Stegm uller2 Behzad Bozorgtabar2,3 Jean-Philippe Thiran2,3 Tinne Tuytelaars1 1KU Leuven 2EPFL 3CHUV 1{firstname}.{lastname}@esat.kuleuven.be 2{firstname}.{lastname}@epfl.ch Leveraging nearest neighbor retrieval for self-supervised representation learning has proven beneficial with object-centric images. However, this approach faces limitations when applied to scene-centric datasets, where multiple objects within an image are only implicitly captured in the global representation. Such global bootstrapping can lead to undesirable entanglement of object representations. Furthermore, even object-centric datasets stand to benefit from a finer-grained bootstrapping approach. In response to these challenges, we introduce a novel Cross Image Object-Level Bootstrapping method tailored to enhance dense visual representation learning. By employing object-level nearest neighbor bootstrapping throughout the training, Cr IBo emerges as a notably strong and adequate candidate for in-context learning, leveraging nearest neighbor retrieval at test time. Cr IBo shows state-of-the-art performance on the latter task while being highly competitive in more standard downstream segmentation tasks. Our code and pretrained models are publicly available at https://github.com/tileb1/Cr IBo. 1 INTRODUCTION Over the past few years, the field of artificial intelligence has experienced significant advancements, primarily driven by the democratization of deep learning techniques (Krizhevsky et al., 2012). Selfsupervised learning (SSL) stands out as one of the major factors contributing to the success of deep learning. Indeed, SSL has opened the door to the training of large foundation models pretrained on massive amounts of uncurated data. In natural language processing (NLP), this large-scale taskagnostic pretraining has been the key to unlocking general-purpose features that perform well on various downstream tasks. Conversely, in computer vision, task-specific models and finetuning remain the de facto choice, in part due to the lack of consensus around the pretraining pretext task, such as is the case in NLP with masked language modeling (MLM). Nonetheless, Balaˇzevi c et al. (2023) showed that under the auspices of their contextual pretraining, dense nearest neighbor retrieval performed comparably with task-specific models while being denotes equal contribution. Figure 1: Positioning of Cri BO in the landscape of self-supervised learning. a) Illustration of the cross-image self-supervision concept. c) Depiction of object-level self-supervision. b) Cr IBo benefits from both learning paradigms. Published as a conference paper at ICLR 2024 significantly more efficient. The contextualization mechanism ensures the alignment of the pretraining phase with dense nearest neighbor evaluation by leveraging within and across image attention. Nevertheless, this mechanism only partially fulfills the alignment objective as i) the cross-image attention mechanism occurs from local to global features and ii) consistency is only encouraged between contextualized image-level representations of views from the same image. To mitigate these limitations, we propose to explicitly enforce cross-image consistency between object-level representations as shown in Figure 1. Most recent self-supervised learning algorithms, e.g., Chen et al. (2020); He et al. (2020); Grill et al. (2020a); Caron et al. (2021), revolve around the objective of learning image-level representations invariant to semantic-preserving data augmentations. Cr IBo departs from this paradigm in two ways. First, we encourage consistency at the granularity of the object in the images, and second, this consistency is enforced between objects from different images. In doing so, our pretraining aligns well with the objective of obtaining general-purpose representations tailored for dense nearest neighbor retrieval. Furthermore, by operating at the object level, Cr IBo elegantly mitigates the pitfall of contextual bias Singh et al. (2020), and is therefore compatible with scene-centric images which constitute the bulk of web-scale datasets. Our contributions are as follows. 1) To the best of our knowledge, Cr IBo is the first end-to-end online SSL approach that explicitly enforces cross-image consistency at the object/object-part level. 2) The resulting representations excel in in-context scene understanding tasks (Balaˇzevi c et al., 2023) without compromising the performance on standard segmentation benchmarks. 3) Moreover, Cr IBo is compatible with scene-centric datasets, addressing a gap in the existing cross-image SSL literature. 2 RELATED WORKS Image-level self-supervision. Seminal work Sim CLR (Chen et al., 2020) has played a pivotal role in promoting the widespread adoption of cross-view consistency as a fundamental principle for visual representation learning. However, such pretext task admits trivial solutions and contemporary works have yet to agree on the ideal way to prevent them. One effective and intuitive approach to address this issue involves the use of negative samples (Chen et al., 2020; Hjelm et al., 2018), which effectively mitigates degeneracy but requires the use of large batchsizes. This can be alleviated with a memory bank (He et al., 2020). Self-distillation methods, on the other hand, do not use explicit negatives. They avoid trivial solutions by using asymmetry in the form of e.g., an additional predictor (Chen & He, 2020; Grill et al., 2020b) on one branch, using stop-gradients (Chen & He, 2020; Caron et al., 2021; Li et al., 2021; Grill et al., 2020b), a momentum encoder (Caron et al., 2021; Grill et al., 2020b), etc. On the other end of the spectrum, clustering-based methods avoid trivial solutions by regularizing the assignment of the samples across a set of clusters (Asano et al., 2019; Caron et al., 2018; 2020; 2021; Zhuang et al., 2019). Localized self-supervision. A finer-grained self-supervision can be obtained by leveraging the inherent spatial structure of images and by enforcing cross-view consistency at a localized level. This ensures that the resulting features are better aligned with dense downstream tasks. These methods can be broadly categorized based on the level of granularity at which they enforce similarity. One line of work involves enforcing similarity at the feature level, known as the local-level approach, where individual features (or local-representations) are directly contrasted (Wang et al., 2021; Liu et al., 2020; O Pinheiro et al., 2020; Xie et al., 2021b; Lebailly & Tuytelaars, 2022; Bardes et al., 2022). Another approach operates at a coarser level, promoting similarity between semantically coherent groups of features, referred to as the object-level approach (Cho et al., 2021; H enaff et al., 2021; 2022; Xie et al., 2021a; Seitzer et al., 2022; Stegm uller et al., 2023; Wen et al., 2022). Alternatively, there have been propositions of dense fine-tuning strategies (Hamilton et al., 2022; Ziegler & Asano, 2022; Yun et al., 2022; Wang et al., 2022; Zadaianchuk et al., 2022). These techniques are designed to enhance models initially trained with an image-level objective (Caron et al., 2021) with localized self-supervision. However, the former cannot be used as stand-alone pretraining methods. Another way of using the inherent spatial structure is by leveraging patch location prediction (Caron et al., 2022) or by asking the model to reconstruct the input image from a corrupted/masked version thereof (He et al., 2022). Published as a conference paper at ICLR 2024 Cross-image self-supervision. Moving beyond the conventional cross-view consistency inherent in numerous self-supervised methods, image-level bootstrapping harnesses nearest neighbors in the latent space for self-labeling and metric learning (Dwibedi et al., 2021; Koohpayegani et al., 2021; Azabou et al., 2021; Xie et al., 2022). Current bootstrapping methods for representation learning face two challenges. Firstly, the effectiveness of using distance metrics in the latent space as a proxy for semantic closeness is contingent on the quality of the learned representations, especially during the initial stage of pretraining when the encoder is randomly initialized. To address this limitation, adaptive bootstrapping methods have been proposed, aiming to avoid systematic nearest neighbor bootstrapping (Lebailly et al., 2023). Secondly, in the context of scene-centric images, image-level bootstrapping reveals a tendency to entangle object representations, which can have detrimental effects on dense downstream tasks. These issues motivate the introduction of the objectlevel cross-image self-supervised learning paradigm. To that end, ORL (Xie et al., 2021a) introduced a three-stage procedure to identify pairs of region-of-interests (Ro Is) and encourage cross-image consistency between cropped objects at the image-level. Recently, Balaˇzevi c et al. (2023) proposed to leverage attention within and across images. The latter contextualizes the local representations of an image with the global ones from other images to adapt for in-context learning of various scene understanding tasks at test time. 3.1 PRELIMINARIES Before diving into the method, we briefly introduce the terminology used in this paper. Dense representation. A representation that explicitly preserves the spatial dependencies of the input image is referred to as a dense representation. We denote such representations by z RH W d or RHW d where H and W are the height and width of the image or a downscaled version thereof and d refers to the dimension of the latent space. The dense representation produced by a convolutional neural network is a stack of output feature maps, whereas a vision transformer yields an ordered sequence of tokens. Local representation. Given a spatial location n out of the N = HW possible ones, the associated slice of the dense representation zn Rd with n {1, 2, , N} is referred to as a local representation. Object representation. The object representation ck Rd of the kth object within an image is derived by aggregating the local representations associated with the corresponding object. We use the term object loosely to refer to a semantically coherent region within the image. Global representation. The global representation, denoted as z Rd, is a global aggregation over the spatial dimensions of the dense representation. As an example, this corresponds to the [CLS] token in a Vi T (Dosovitskiy et al., 2020) or the result of a global average pooling layer in case of a CNN. Bootstrapping Here, the term bootstrapping1 refers to the utilization of the model s current knowledge to enhance its performance further. In the scope of this study, this materializes as leveraging distances in the latent space to find object-level nearest neighbors that can be used to form positive pairs for self-distillation. 3.2 OBJECT-LEVEL CROSS-IMAGE BOOTSTRAPPING (CRIBO) In Section 1, we discuss the need for generalist models capable of performing various downstream tasks without requiring any finetuning. A recent study (Balaˇzevi c et al., 2023) demonstrated that this could be achieved by leveraging nearest neighbor retrieval at test time and contextualization 1The term bootstrapping is employed in the colloquial sense, as opposed to the statistical sense. Published as a conference paper at ICLR 2024 joint-space clustering retrieved nearest neighbors object representations memory bank object-level cross-image loss Figure 2: High-level overview of cross-image object-level bootstrapping (Cr IBo). Given an encoder f and pair of augmented views x1 and x2, object representations ck i (depicted as colored object masks) from each view i are computed. Using a memory bank, the nearest neighbors of each object representation ck 1 from the first view are retrieved. A self-supervised consistency loss (depicted as colored arrows) is then enforced between ck 2 and its corresponding retrieved neighbor from the other view nn(ck 1). throughout the training phase. Aligning with this aim, we propose further reducing the gap between the training and test phases by explicitly enforcing consistency between object-level nearest neighbors. The existing SSL studies exploiting nearest neighbors are built solely around global representations and are rooted in the assumption of datasets being object-centric. Hence, a simplistic extension of such frameworks to our proposed scenario is impracticable. To accomplish our goal, the following steps are required: 1. Identify semantically coherent regions within the image to create object-level representations (Sec. 3.2.1). 2. Match pairs of object-level representations across images (Sec. 3.2.2). 3. Apply cross-image consistency between pairs of matching object-level representations (Sec. 3.2.3). A high-level overview of our proposed method Cr IBo can be found in Figure 2. 3.2.1 SEMANTICALLY COHERENT IMAGE REGIONS The first step towards cross-image object-level bootstrapping is to identify semantically coherent image regions within the images. We argue that it is beneficial to use pairs of augmented views for each input image and to have a cross-view correspondence between the objects in both views. Indeed, image-level SSL methods yield significantly worsened performance when training without photometric augmentations (Dwibedi et al., 2021). This is observed when positive pairs originate from the same image and to a lesser extent (but still significant) when they do not. This drop in performance highlights the need to shed away low-level similarities between positive pairs of images with data augmentations. To prevent this from occurring with nearest neighbor (NN) positives, Dwibedi et al. (2021) refrain from enforcing consistency between a retrieved NN and its query. At the image level, this can be easily implemented by leveraging the duplicity of the views. More specifically, given two augmented views x1 and x2 as well as their corresponding global-representation z1 and z2, a similarity constraint is enforced between z1 and nn( z2) where nn( ) denotes the nearest neighbor operator. Conversely, at the object level, this requires a cross-view correspondence between object representations from both views. To that end, we rely on an online joint-space clustering algorithm from Stegm uller et al. (2023). Joint-space clustering. The clustering algorithm takes as input the dense representations z1 RN d and z2 RN d associated with the two augmented views and concatenates them along the token axis to obtain the dense representation of the joint-view, zcat R2N d. The 2N tokens are subsequently partitioned into K clusters to produce semantically coherent and spatially compact clusters. A hyperparameter λpos modulates the importance of the spatiality against that of the semantic (see Tab. 4). The resulting assignment is encoded in a matrix Q R2N K, which can Published as a conference paper at ICLR 2024 Figure 3: Illustration of cycle-consistent matchings. Such matchings are invariant to data augmentations and reciprocal, loosely speaking. be split in two to obtain the view-wise assignments Q 1, Q 2 RN K. Here Q i (n, k) denotes a soft-assignment of token n to object-cluster k in view i. These can in turn be used to compute the K object representations ci Rd in each view i as follows: n [N] Q i (n, k) zn i , i {1, 2} (1) Note that the columns of Q i undergo l1 normalization, rendering the object representations as affine combinations of local representations. It can be observed that the cross-view correspondence is provided for free as the object representation ck 1 from the first view is matched to ck 2 in the second one as a consequence of the joint-space clustering. In practice, it can be the case that some objects are exclusive to one of the two views i.e., P n [N] Q i (n, k) = 0. In that case, the object k is simply discarded from the view where it is present. In Section 3.2.2, we show that such cross-view correspondence can be harnessed to improve the training stability through a bootstrapping criterion. 3.2.2 CROSS-IMAGE OBJECT MATCHINGS In Section 3.2.1, we detail the methodology for identifying objects within the views of a given image and establish the strategy for their alignment from one view to another. Subsequent to this step is the critical phase of aligning such objects across images. For reasons that will be delineated in the next paragraph, we rely on a pair of memory banks M1 and M2, which are populated with object representations extracted from the first and second views during previous iterations. As such, the memory banks Mi can be queried with an object representation ck i to retrieve its nearest neighbor: nn(ck i , Mi) arg max c Mi c ck i c 2 ck i 2 (2) Here, we consider nn(ck 1, M1) and ck 1 as cross-image object-level matches. A visualization of cross-image object-level matches can be found in Appendix C.5 (Fig. 4 and Fig. 5). Cycle-consistent matchings. At the beginning of the pretraining phase, when the encoder is randomly initialized, relying on Equation (2) to identify cross-image object-level matches is not well grounded. Indeed, this straightforward approach assumes that the representations of similar objects are close in the embedding space, an assumption that does not hold at the initialization. Therefore, enforcing consistency between such pairs of semantically distinct objects should be avoided. To that end, we design a semantic closeness condition, which is only fulfilled when the matches are established based on high-level features, thus enabling the discarding of invalid positive pairs that do not share a semantic similarity. At a high-level, we consider a pair of objects as a valid positive pair if they are the nearest neighbors of each other. In practice, the criterion is implemented as follows. Starting from an object representation ck 1 in the batch of first views, its nearest neighbor nn(ck 1, M1) in the queue is found using Equation (2). To ensure that the criterion is invariant to data augmentation, we then switch to the queue of second views via the jump( ) operator: jump(ck 1) ck 2, jump(ck 2) ck 1 (3) Published as a conference paper at ICLR 2024 From this step, we obtain jump(nn(ck 1, M1)) and retrieve its NN in the batch of second views nn(jump(nn(ck 1, M1)), B2). The final step consists of switching back to the batch of first views and verifying if the resulting object is the same as the one we started from. More formally: O(ck i ) true if ck i = jump(nn(jump(nn(ck i , Mi)), Bj)) false otherwise i, j {1, 2}, i = j (4) A depiction of cycle-consistent matchings is provided in Figure 3. 3.2.3 SELF-SUPERVISED TRAINING OBJECTIVES At this point, we have all the necessary components to apply three distinct types of self-supervision: cross-view object-level, cross-image object-level, and global cross-image. All levels of selfsupervision in our method Cr IBo are based on the self-distillation mechanism proposed by Caron et al. (2021), which employs a student-teacher pair of siamese networks fs and ft, respectively. Given two augmented views x1 and x2, global representations and object representations can be obtained for each view and each network within the siamese pair. Cross-View Object-Level Self-Supervision. The object representations from the teacher and student are fed to their respective head hs and ht to output probability mass functions whose sharpness can be modulated via temperature parameters (τs and τt). For the teacher, this translates to: pk i,t = softmax L ht(ck i,t)/τt , i {1, 2} (5) The student s projections are obtained analogously. The cross-view object-level loss is expressed as the averaged cross-entropy over all pairs of object representations and making the loss symmetric w.r.t. the student/teacher: H(pk 1,t, pk 2,s) + H(pk 2,t, pk 1,s) (6) where H(a, b) = PL l=1 al log(bl) denotes the cross-entropy. Cross-Image Object-Level Self-Supervision. Once all object representations in the batch have been retrieved, the cross-image object-level loss is enforced in a manner analogous to its cross-view counterpart. The nearest neighbors are first fed to the projection head as follows: pk i,t = softmax L ht(nn(ck i,t))/τt , i {1, 2} (7) Note that these projections are only defined for the teacher since the nn( ) operator is not differentiable. The formulation of the cross-image object-level loss aligns with that in Equation (6) up to the filtering of invalid positive pairs (see Sec. 3.2.2): k [K] 1{O(ck 1)}H( pk 1,t, p2,s) + 1 k [K] 1{O(ck 2)}H( pk 2,t, p1,s) (8) where 1{ } denotes the indicator function and where Z1 and Z2 are normalization constants i.e., P k [K] 1{O(ck 1)} and P k [K] 1{O(ck 2)}, respectively. Empirical evidence on the utility of the cycle consistency can be found in Appendix C.4. Global Cross-View Self-Supervision. Importantly, both the clustering and the bootstrapping steps are contingent on the quality of the representations. To avoid potential instabilities, we incorporate a global cross-view loss into our framework, offering meaningful self-supervision at any stage of the pretraining. This loss is analogous to the one from Equation (6) except that it is applied to the global representation (z1,s, z2,s, z1,t and z2,t), which are fed to a dedicated head h: pi,t = softmax ht( zi,t/τ t) , i {1, 2} (9) Published as a conference paper at ICLR 2024 Table 1: Dense nearest neighbor retrieval. The quality of the learned spatial features is probed with a k-NN classifier and using different ratios of training data. The depicted m Io U scores are derived from the validation sets of two scene-centric datasets. refers to our own reproduction from official Git Hub repositories. If not specified, publicly available checkpoints are used. denotes results taken from Balaˇzevi c et al. (2023). ADE20K Pascal VOC Method Backbone Params Dataset Epochs 1/128 1/64 1/8 1/1 1/128 1/64 1/8 1/1 Scene-centric Slot Con (Wen et al., 2022) Res Net-50 25M COCO 800 9.9 11.8 17.3 22.1 37.3 42.8 52.9 57.2 ORL (Xie et al., 2021a) Res Net-50 25M COCO 800 9.1 10.3 13.6 16.6 30.2 33.3 42.0 45.6 DINO (Caron et al., 2021) Vi T-S/16 21M COCO 300 6.1 6.9 9.7 13.0 16.2 18.4 25.5 31.9 MAE (He et al., 2022) Vi T-S/16 21M COCO 300 3.7 4.1 5.4 6.8 8.5 9.3 12.2 15.9 Cr OC (Stegm uller et al., 2023) Vi T-S/16 21M COCO 300 7.6 9.0 13.1 18.0 27.1 31.4 40.3 47.1 Cr IBo Vi T-S/16 21M COCO 300 10.9 12.8 18.4 23.4 39.1 44.0 52.8 58.1 Object-centric Slot Con (Wen et al., 2022) Res Net-50 25M IN1K 200 10.6 12.0 18.1 24.0 40.7 44.7 56.4 61.6 DINO (Caron et al., 2021) Vi T-S/16 21M IN1K 800 9.4 10.6 14.6 18.4 24.5 28.7 38.7 46.1 Cr OC (Stegm uller et al., 2023) Vi T-S/16 21M IN1K 300 7.8 9.6 15.2 20.6 30.7 37.7 54.8 64.2 Time T (Salehi et al., 2023) Vi T-S/16 21M IN1K+YTVOS 800+30 11.7 13.6 19.7 24.6 38.9 44.3 56.0 62.7 Cr IBo Vi T-S/16 21M IN1K 800 13.7 16.5 23.2 28.3 52.7 59.3 69.3 73.2 DINO (Caron et al., 2021) Vi T-B/16 85M IN1K 400 11.1 12.6 17.6 22.0 29.2 34.7 47.2 54.9 MAE (He et al., 2022) Vi T-B/16 85M IN1K 1600 2.7 3.0 4.0 5.3 6.0 6.5 8.9 13.8 LOCA (Caron et al., 2022) Vi T-B/16 85M IN1K 600 - - - 18.5 - - - 57.5 Hummingbird (Balaˇzevi c et al., 2023) Vi T-B/16 85M IN1K 300 11.7 15.1 - 28.3 50.5 57.2 - 70.5 Cr IBo Vi T-B/16 85M IN1K 400 13.2 16.5 23.6 30.0 50.5 60.3 70.8 74.9 where τ s and τ t are the corresponding temperature parameters for the L-dimensional output distribution. The global cross-view loss is defined as: Lg cv = H(p1,t, p2,s) + H(p2,t, p1,s) (10) where H(a, b) = PL l=1 al log(bl). Finally, we obtain the overall training objective of Cr IBo as a sum of its individual components: Ltot = Lg cv + Lo cv + Lo ci (11) An ablation study over the inclusion of the individual loss terms can be found in Table 5. As in Caron et al. (2021), the student s parameters are optimized to minimize the above loss whereas the parameters of the teacher are updated via an exponential moving average of the student s weights. 4 EXPERIMENTS We first verify that the proposed pretraining method aligns well with the objective of in-context learning via nearest neighbor retrieval. We then show that in doing so, we do not compromise the performance on standard evaluations. General details on the experimental setup can be found in Appendix A. 4.1 DENSE NEAREST NEIGHBOR RETRIEVAL Inspired by Balaˇzevi c et al. (2023), we probe the quality of the learned features with a dense nearest neighbor (NN) classification task. The k-NN classifier is fitted on the local representations of a uniformly sub-sampled set of training images and evaluated on all the patches from the validation set of images. The sub-sampling factor is either 1, 8, 64 or 128. The only preprocessing applied to the images is a down-sampling to the closest resolution which is a multiple of the patch size. Following that step, patch-level labels are obtained via majority voting. We report the m Io U scores on Pascal VOC 2012 (Everingham et al.) and ADE20K (Zhou et al., 2017). When the sub-sampling ratio is greater than 1, the reported m Io U scores result from averaging over 5 independent runs. The number of nearest neighbor patches is set to k = 50. In Table 1, we observe that using object-level nearest neighbor bootstrapping throughout training is an adequate pretext task for this evaluation, which is not a surprise due to the inherent similarities between the training and test phases. Overall, Cr IBo outperforms all methods, but Hummingbird (Balaˇzevi c et al., 2023), by a significant margin and particularly following the scene-centric Published as a conference paper at ICLR 2024 Table 2: Linear segmentation with frozen backbones. The linear decoder from Segmenter Strudel et al. (2021) is trained on the frozen spatial features obtained with various self-supervised learning methods. We report the m Io U scores achieved on the validation sets of 4 different datasets. refers to our own reproduction from official Git Hub repositories. If not specified, publicly available checkpoints are used. Method Backbone Params Dataset Epochs Pascal Context Pascal VOC COCO-Stuff 164K ADE20K Scene-centric DINO (Caron et al., 2021) Vi T-S/16 21M COCO 300 27.3 43.9 19.9 18.9 MAE (He et al., 2022) Vi T-S/16 21M COCO 300 18.1 25.6 11.3 11.0 Cr OC (Stegm uller et al., 2023) Vi T-S/16 21M COCO 300 30.0 51.7 22.6 20.5 Cr IBo Vi T-S/16 21M COCO 300 37.5 61.0 29.1 27.5 Cr IBo Vi T-S/8 21M COCO 300 40.6 63.9 31.2 28.3 Object-centric DINO (Caron et al., 2021) Vi T-S/16 21M IN1K 800 32.4 49.0 24.1 23.4 Cr OC (Stegm uller et al., 2023) Vi T-S/16 21M IN1K 300 37.7 68.1 30.2 27.6 Time T (Salehi et al., 2023) Vi T-S/16 21M IN1K+YTVOS 800+30 40.6 67.9 32.0 29.9 Cr IBo Vi T-S/16 21M IN1K 800 41.7 73.7 33.8 31.9 DINO (Caron et al., 2021) Vi T-B/16 85M IN1K 400 38.9 64.6 31.9 30.5 MAE (He et al., 2022) Vi T-B/16 85M IN1K 1600 27.9 40.9 14.4 17.5 Cr IBo Vi T-B/16 85M IN1K 400 42.9 74.9 36.0 34.7 Table 3: Finetuning evaluation with Segmenter. Backbones pre-trained with different selfsupervised learning methods are finetuned using Segmenter (Strudel et al., 2021). We report the m Io U scores achieved on the validation sets of 4 different datasets. refers to our own reproduction from official Git Hub repositories. If not specified, publicly available checkpoints are used. Method Backbone Params Dataset Epochs Pascal Context Pascal VOC COCO-Stuff 164K ADE20K Scene-centric DINO (Caron et al., 2021) Vi T-S/16 21M COCO 300 33.5 66.1 35.6 35.0 MAE (He et al., 2022) Vi T-S/16 21M COCO 300 32.5 61.6 35.4 35.4 Cr OC (Stegm uller et al., 2023) Vi T-S/16 21M COCO 300 38.7 70.5 38.0 37.9 Cr IBo Vi T-S/16 21M COCO 300 41.9 74.2 39.4 39.3 Object-centric DINO (Caron et al., 2021) Vi T-S/16 21M IN1K 800 46.0 80.3 43.2 43.3 Cr OC (Stegm uller et al., 2023) Vi T-S/16 21M IN1K 300 46.0 80.9 42.9 42.8 Time T (Salehi et al., 2023) Vi T-S/16 21M IN1K+YTVOS 800+30 47.4 80.4 43.1 43.5 Cr IBo Vi T-S/16 21M IN1K 800 49.3 82.3 43.9 45.2 DINO (Caron et al., 2021) Vi T-B/16 85M IN1K 400 45.8 82.2 44.4 45.0 MAE (He et al., 2022) Vi T-B/16 85M IN1K 1600 47.9 82.7 45.5 46.4 Cr IBo Vi T-B/16 85M IN1K 400 49.2 83.4 44.6 46.0 pretraining. It is worth noting that the comparison with Balaˇzevi c et al. (2023) is not exactly appleto-apple, as i) we do not use multiple augmentations per image and ii) we refrain from balancing the classes in the training set. These differences in the implementation of the evaluation reflect the divergence in our objectives: we place a higher emphasis on the assessment of the intrinsic quality of the learned representations rather than the absolute value of the reported results. 4.2 SEGMENTATION WITH A LINEAR HEAD The frozen local representations are fed to the linear decoder head from Segmenter (Strudel et al., 2021) using the MMSegmentation (Contributors, 2020) implementation. Herein, spatial tokens undergo a linear projection to the class space, followed by bilinear up-sampling to match the input image dimensions, enabling the application of pixel-wise cross-entropy loss. Performance metrics, in terms of m Io U scores, are reported on the same 4 datasets (see Appendix A.2 for more details). As reported in Table 2, Cr IBo exhibits commendable performance in linear segmentation, following both scene-centric and object-centric pretrainings. Among the methods using a Vi T-B/16, Cr IBo strongly outperforms other baselines. In the case of Vi T-S/16, Cr IBo also outperforms all baselines but with a smaller margin, due mostly to Time T (Salehi et al., 2023) being a particularly strong baseline. It is noteworthy that as opposed to prevailing studies, e.g., Stegm uller et al. (2023); Bardes et al. (2022), we do not concatenate spatial tokens from multiple layers of the Vi T. Indeed, this trick is typically used to compare on an equal footing with Res Net-50, which has a much larger output dimension. As we only use and compare with Vi Ts, we opt for evaluations incorporating as minimal parameters as possible, aspiring to present a more accurate reflection of the features intrinsic quality. Published as a conference paper at ICLR 2024 Table 4: Ablation study on the hyperparameters of Cr IBo. Here, a Vi T-S/16 undergoes pretraining on COCO for 300 epochs with various combinations of hyperparameter values. The resulting features are then compared using dense nearest neighbors retrieval. We denote the size of the queues Mi in terms of the number of images present by S. Unless explicitly indicated, the values of the hyperparameters are set to λpos = 2.0, S = 25k, and K = 64. Positional weighting (λpos) Memory bank size (S) Number of objects (K) Dataset Lo ci 0.0 0.1 1.0 2.0 4.0 8.0 1k 5k 25k 4 8 16 32 64 Pascal VOC 44.9 46.3 47.5 47.5 45.8 44.2 - - - 47.3 47.6 48.1 47.1 47.5 Pascal VOC 55.9 57.1 58.1 58.1 57.5 57.0 46.7 47.3 58.1 49.9 52.9 54.6 57.0 58.1 ADE20K 16.9 18.0 18.9 18.4 17.8 16.3 - - - 17.7 18.0 18.9 18.5 18.4 ADE20K 21.6 21.8 23.1 23.4 22.6 22.5 18.1 18.4 23.4 18.1 19.4 20.3 21.7 23.4 4.3 END-TO-END FINETUNING WITH SEGMENTER In a real-world scenario, self-supervised pretraining often precedes end-to-end finetuning on the target dataset with an application-specific decoder head. The ability of the different pretraining paradigms to cope with that setting is assessed with an end-to-end finetuning of the backbone topped with the transformer-based decoder from Segmenter (Strudel et al., 2021). Here, the backbone s spatial features are fed to a transformer encoder along with K learnable class tokens. The resulting class and spatial tokens are projected onto one another to obtain patch-level predictions, subsequently upsampled to the input image s size to enforce the pixel-wise cross-entropy loss. We report the m Io U scores achieved on the same 4 datasets. The insights gained from the preceding experiments (Tabs. 1 and 2) underscore that models pretrained with Cr IBo have excellent generalists properties, i.e., allowing them to perform various tasks without requiring finetuning. The results depicted in Table 3 indicate that, when required, these models can also evolve into remarkable specialists. All that aside, after fine-tuning, all models tend to perform comparably, highlighting that the finetuning may be a suboptimal regime to compare SSL pretraining methods. 4.4 ABLATIONS We report in Table 4 the results of a grid search over various hyperparameters performed on COCO. Interestingly, we observe that Cr IBo performs the best when operating in the overclustering regime, that is, with a K much larger than the number of objects present in the images, but only when bootstrapping is used. We postulate that this is due to the fact that overclustering offers the possibility to find better matches. To illustrate, consider the image in Figure 2 depicting a zebra, a giraffe, and an elephant. With K = 1, this would necessitate identifying another image displaying the identical trio of animals; which is rare. With a larger value of K, the animals are isolated, expanding the number of good candidates. In the overclustering regime, the clusters represent object parts, which further facilitates the matchings. Similarly, a larger queue also amplifies the number of suitable candidates, a trend reflected in the results until reaching a saturation point at a queue size of 25k. More ablations are available in Appendix C. 5 CONCLUSION We introduce Cr IBo, the first end-to-end online cross-image object-level bootstrapping selfsupervised learning method. Cr IBo capitalizes on the augmented sample diversity derived from image-level bootstrapping while harnessing the advantages of localized self-supervision inherent to object-level learning. We extensively evaluate Cr IBo across multiple dense downstream tasks showcasing excellent performance on in-context scene understanding tasks and highlighting the merits of the cross-image object-level learning paradigm. Limitations. This study is developed around the Vi T architecture as it has become quite ubiquitous. As such, and for the sake of coherence, we only compare with other Vi T-based methods. As we have to cope with a computational budget, we chose to invest it in producing a well-grounded and motivated approach with extensive analysis, rather than in exploring larger datasets (Thomee et al., 2016; Changpinyo et al., 2021) or models, e.g., Vi T-Large or Vi T-Huge. Published as a conference paper at ICLR 2024 Acknowledgement. This project is funded by the European Research Council (ERC) under the European Union s Horizon 2020 research and innovation program (Grant Agreement No. 101021347). This work is also partially funded by the Personalized Health and Related Technologies (PHRT), grant number 2021/344; as well as the Fonds Wetenschappelijk Onderzoek (FWO), project G0A4720N. We acknowledge both Euro CC Belgium and the Swiss National Supercomputing Centre (project ID 606) for awarding this project access to the LUMI supercomputer. Yuki Markus Asano, Christian Rupprecht, and Andrea Vedaldi. Self-labelling via simultaneous clustering and representation learning. ar Xiv preprint ar Xiv:1911.05371, 2019. Mehdi Azabou, Mohammad Gheshlaghi Azar, Ran Liu, Chi-Heng Lin, Erik C. Johnson, Kiran Bhaskaran-Nair, Max Dabagia, Keith B. Hengen, William R. Gray Roncal, Michal Valko, and Eva L. Dyer. Mine your own view: Self-supervised learning through across-sample prediction. Co RR, abs/2102.10106, 2021. URL https://arxiv.org/abs/2102.10106. Ivana Balaˇzevi c, David Steiner, Nikhil Parthasarathy, Relja Arandjelovi c, and Olivier J H enaff. Towards in-context scene understanding. ar Xiv preprint ar Xiv:2306.01667, 2023. Adrien Bardes, Jean Ponce, and Yann Le Cun. Vicregl: Self-supervised learning of local visual features. ar Xiv preprint ar Xiv:2210.01571, 2022. Holger Caesar, Jasper Uijlings, and Vittorio Ferrari. Coco-stuff: Thing and stuff classes in context. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1209 1218, 2018. Mathilde Caron, Piotr Bojanowski, Armand Joulin, and Matthijs Douze. Deep clustering for unsupervised learning of visual features. In Proceedings of the European conference on computer vision (ECCV), pp. 132 149, 2018. Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. Unsupervised learning of visual features by contrasting cluster assignments. In Thirty-fourth Conference on Neural Information Processing Systems (Neur IPS), 2020. Mathilde Caron, Hugo Touvron, Ishan Misra, Herv e J egou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging properties in self-supervised vision transformers. ar Xiv preprint ar Xiv:2104.14294, 2021. Mathilde Caron, Neil Houlsby, and Cordelia Schmid. Location-aware self-supervised transformers. ar Xiv preprint ar Xiv:2212.02400, 2022. Soravit Changpinyo, Piyush Sharma, Nan Ding, and Radu Soricut. Conceptual 12m: Pushing web-scale image-text pre-training to recognize long-tail visual concepts. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3558 3568, 2021. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp. 1597 1607. PMLR, 2020. Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. Co RR, abs/2011.10566, 2020. URL https://arxiv.org/abs/2011.10566. Jang Hyun Cho, Utkarsh Mall, Kavita Bala, and Bharath Hariharan. Picie: Unsupervised semantic segmentation using invariance and equivariance in clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16794 16804, 2021. MMSegmentation Contributors. MMSegmentation: Openmmlab semantic segmentation toolbox and benchmark. https://github.com/open-mmlab/mmsegmentation, 2020. Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. Advances in neural information processing systems, 26:2292 2300, 2013. Published as a conference paper at ICLR 2024 Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248 255, 2009. doi: 10.1109/CVPR.2009.5206848. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. Debidatta Dwibedi, Yusuf Aytar, Jonathan Tompson, Pierre Sermanet, and Andrew Zisserman. With a little help from my friends: Nearest-neighbor contrastive learning of visual representations. pp. 9568 9577. IEEE Computer Society, October 2021. ISBN 978-1-66542-8125. doi: 10.1109/ICCV48922.2021.00945. URL https://www.computer.org/csdl/ proceedings-article/iccv/2021/281200j568/1Bm JDKW7JXq. M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisserman. The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results. http://www.pascalnetwork.org/challenges/VOC/voc2012/workshop/index.html. Yuxin Fang, Bencheng Liao, Xinggang Wang, Jiemin Fang, Jiyang Qi, Rui Wu, Jianwei Niu, and Wenyu Liu. You only look at one sequence: Rethinking transformer in vision through object detection. Advances in Neural Information Processing Systems, 34:26183 26197, 2021. Jean-Bastien Grill, Florian Strub, Florent Altch e, Corentin Tallec, Pierre H Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent: A new approach to self-supervised learning. ar Xiv preprint ar Xiv:2006.07733, 2020a. Jean-Bastien Grill, Florian Strub, Florent Altch e, Corentin Tallec, Pierre H. Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, Bilal Piot, Koray Kavukcuoglu, R emi Munos, and Michal Valko. Bootstrap your own latent: A new approach to self-supervised learning, 2020b. Mark Hamilton, Zhoutong Zhang, Bharath Hariharan, Noah Snavely, and William T Freeman. Unsupervised semantic segmentation by distilling feature correspondences. ar Xiv preprint ar Xiv:2203.08414, 2022. Kaiming He, Georgia Gkioxari, Piotr Doll ar, and Ross Girshick. Mask r-cnn. In Proceedings of the IEEE international conference on computer vision, pp. 2961 2969, 2017. Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729 9738, 2020. Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Doll ar, and Ross Girshick. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 16000 16009, 2022. Olivier J H enaff, Skanda Koppula, Jean-Baptiste Alayrac, Aaron van den Oord, Oriol Vinyals, and Jo ao Carreira. Efficient visual pretraining with contrastive detection. ar Xiv preprint ar Xiv:2103.10957, 2021. Olivier J H enaff, Skanda Koppula, Evan Shelhamer, Daniel Zoran, Andrew Jaegle, Andrew Zisserman, Jo ao Carreira, and Relja Arandjelovi c. Object discovery and representation networks. ar Xiv preprint ar Xiv:2203.08777, 2022. R Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Phil Bachman, Adam Trischler, and Yoshua Bengio. Learning deep representations by mutual information estimation and maximization. ar Xiv preprint ar Xiv:1808.06670, 2018. Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. Published as a conference paper at ICLR 2024 Soroush Abbasi Koohpayegani, Ajinkya Tejankar, and Hamed Pirsiavash. Mean shift for selfsupervised learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 10326 10335, October 2021. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25, 2012. Tim Lebailly and Tinne Tuytelaars. Global-local self-distillation for visual representation learning. ar Xiv preprint ar Xiv:2207.14676, 2022. Tim Lebailly, Thomas Stegm uller, Behzad Bozorgtabar, Jean-Philippe Thiran, and Tinne Tuytelaars. Adaptive similarity bootstrapping for self-distillation. ar Xiv preprint ar Xiv:2303.13606, 2023. Chunyuan Li, Jianwei Yang, Pengchuan Zhang, Mei Gao, Bin Xiao, Xiyang Dai, Lu Yuan, and Jianfeng Gao. Efficient self-supervised vision transformers for representation learning. Co RR, abs/2106.09785, 2021. URL https://arxiv.org/abs/2106.09785. Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Doll ar, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In European conference on computer vision, pp. 740 755. Springer, 2014. Songtao Liu, Zeming Li, and Jian Sun. Self-emd: Self-supervised object detection without imagenet. ar Xiv preprint ar Xiv:2011.13677, 2020. Roozbeh Mottaghi, Xianjie Chen, Xiaobai Liu, Nam-Gyu Cho, Seong-Whan Lee, Sanja Fidler, Raquel Urtasun, and Alan Yuille. The role of context for object detection and semantic segmentation in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 891 898, 2014. Pedro O O Pinheiro, Amjad Almahairi, Ryan Benmalek, Florian Golemo, and Aaron C Courville. Unsupervised learning of dense visual representations. Advances in Neural Information Processing Systems, 33:4489 4500, 2020. Mohammadreza Salehi, Efstratios Gavves, Cees G. M. Snoek, and Yuki M. Asano. Time does tell: Self-supervised time-tuning of dense image representations. ICCV, 2023. Maximilian Seitzer, Max Horn, Andrii Zadaianchuk, Dominik Zietlow, Tianjun Xiao, Carl-Johann Simon-Gabriel, Tong He, Zheng Zhang, Bernhard Sch olkopf, Thomas Brox, et al. Bridging the gap to real-world object-centric learning. ar Xiv preprint ar Xiv:2209.14860, 2022. Krishna Kumar Singh, Dhruv Mahajan, Kristen Grauman, Yong Jae Lee, Matt Feiszli, and Deepti Ghadiyaram. Don t judge an object by its context: Learning to overcome contextual bias. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11070 11078, 2020. Thomas Stegm uller, Tim Lebailly, Behzad Bozorgtabar, Tinne Tuytelaars, and Jean-Philippe Thiran. Croc: Cross-view online clustering for dense visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7000 7009, June 2023. Robin Strudel, Ricardo Garcia, Ivan Laptev, and Cordelia Schmid. Segmenter: Transformer for semantic segmentation. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 7262 7272, 2021. Bart Thomee, David A Shamma, Gerald Friedland, Benjamin Elizalde, Karl Ni, Douglas Poland, Damian Borth, and Li-Jia Li. Yfcc100m: The new data in multimedia research. Communications of the ACM, 59(2):64 73, 2016. Feng Wang, Huiyu Wang, Chen Wei, Alan Yuille, and Wei Shen. Cp2: Copy-paste contrastive pretraining for semantic segmentation. ar Xiv preprint ar Xiv:2203.11709, 2022. Xinlong Wang, Rufeng Zhang, Chunhua Shen, Tao Kong, and Lei Li. Dense contrastive learning for self-supervised visual pre-training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3024 3033, 2021. Published as a conference paper at ICLR 2024 Xin Wen, Bingchen Zhao, Anlin Zheng, Xiangyu Zhang, and Xiaojuan Qi. Self-supervised visual representation learning with semantic grouping. ar Xiv preprint ar Xiv:2205.15288, 2022. Jiahao Xie, Xiaohang Zhan, Ziwei Liu, Yew Soon Ong, and Chen Change Loy. Unsupervised objectlevel representation learning from scene images. Advances in Neural Information Processing Systems, 34:28864 28876, 2021a. Jiahao Xie, Xiaohang Zhan, Ziwei Liu, Yew-Soon Ong, and Chen Change Loy. Delving into inter-image invariance for unsupervised visual representations. International Journal of Computer Vision, 130(12):2994 3013, September 2022. ISSN 1573-1405. doi: 10.1007/ s11263-022-01681-x. URL http://dx.doi.org/10.1007/s11263-022-01681-x. Zhenda Xie, Yutong Lin, Zheng Zhang, Yue Cao, Stephen Lin, and Han Hu. Propagate yourself: Exploring pixel-level consistency for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16684 16693, 2021b. Sukmin Yun, Hankook Lee, Jaehyung Kim, and Jinwoo Shin. Patch-level representation learning for self-supervised vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8354 8363, 2022. Andrii Zadaianchuk, Matthaeus Kleindessner, Yi Zhu, Francesco Locatello, and Thomas Brox. Unsupervised semantic segmentation with self-supervised object-centric representations. ar Xiv preprint ar Xiv:2207.05027, 2022. Bolei Zhou, Hang Zhao, Xavier Puig, Sanja Fidler, Adela Barriuso, and Antonio Torralba. Scene parsing through ade20k dataset. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 633 641, 2017. Chengxu Zhuang, Alex Lin Zhai, and Daniel Yamins. Local aggregation for unsupervised learning of visual embeddings. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019. Adrian Ziegler and Yuki M Asano. Self-supervised learning of object parts for semantic segmentation. ar Xiv preprint ar Xiv:2204.13101, 2022. Published as a conference paper at ICLR 2024 A EXPERIMENTAL SETUP A.1 PRETRAINING Pre-training datasets. Our pretraining datasets include COCO (Lin et al., 2014) and Image Net-1k (Deng et al., 2009). The former includes about 118k scence-centric images while the latter includes about 1.3M object-centric images. Both are ubiquitous in scene-centric self-supervised learning and object-centric self-supervised learning, respectively. Network architecture. We employ vision transformers as our backbone f. This selection aligns with its prevalent use in contemporary methodologies and allows for fair and easy comparisons with existing works. The configuration of the projection heads closely follows that of Caron et al. (2021). Both the image-level head h and object-level head h share their weights, except for the final linear layer. Both the image-level and the object-level output distribution have the same size i.e., L = L = 65, 536. Optimization. The Vi T-small (Vi T-S/16) is trained for 800 epochs, while the Vi T-base (Vi T-B/16) is trained for 400 epochs. The Vi T-Base is only trained on Image Net-1k, while the Vi T-Small is trained on both COCO and Image Net-1k. Pretrainings on COCO use a batchsize of 256 while pretrainings on Image Net-1k use a batchsize of 1024. Learning rate, weight-decay, and other optimization-related hyperparameters are exactly the same as in DINO (Caron et al., 2021). Hyperparameters. Results reported in tables using Vi T-S/16 (apart from the gridsearch) are based on the following hyperparameters: (λpos, S, K) = (1.0, 25k, 32) and (λpos, S, K) = (2.0, 25k, 64) for pretrainings on Image Net-1K and COCO, respectively. Results reported for Vi T-B/16 are based using the following hyperparameters: (λpos, S, K) = (1.0, 25k, 32). A.2 EVALUATION PROTOCOLS Linear segmentation. The frozen local representations are fed to the linear decoder head from Segmenter (Strudel et al., 2021), and we use the implementation available through MMSegmentation (Contributors, 2020). Herein, spatial tokens undergo a linear projection to the class space, followed by bilinear up-sampling to match the input image dimensions, enabling the application of pixel-wise cross-entropy loss. Performance metrics, in terms of m Io U scores, are reported on four different datasets: Pascal Context (Mottaghi et al., 2014), Pascal VOC 2012 (Everingham et al.), COCO-Stuff 164K (Caesar et al., 2018) and ADE20K (Zhou et al., 2017). Regardless of the dataset used, the crop size is always 512 512 pixels, and the dataset configurations are kept as provided by MMSegmentation (Contributors, 2020). For ADE20K and COCO-Stuff 164K, we use the 160k iterations schedule whereas for Pascal VOC 2012 and Pascal Context we use the 80k iterations schedule. The only change we make to the default schedules configurations is that we use the Adam optimizer (Kingma & Ba, 2014) instead of SGD. For each pretraining method and dataset, we run with four different learning rates (8e-4, 3e-4, 1e-4, and 8e-5) and report the highest m Io U score. Finetuning with Segmenter. We perform an end-to-end finetuning of the backbone topped with the transformer-based decoder from Segmenter (Strudel et al., 2021). Here, the backbone s spatial features are fed to a transformer encoder along with K learnable class tokens. The resulting class and spatial tokens are projected onto one another to obtain patch-level predictions and subsequently up-sampled to the input image s size to enforce the pixel-wise cross-entropy loss. We report the m Io U scores achieved on 4 different datasets: Pascal Context (Mottaghi et al., 2014), Pascal VOC 2012 (Everingham et al.), COCO-Stuff 164K (Caesar et al., 2018) and ADE20K (Zhou et al., 2017). Regardless of the dataset used, the crop size is always 512 512 pixels, and the dataset configurations are kept as provided by MMSegmentation (Contributors, 2020). For ADE20K and COCOStuff 164K, we use the 160k iterations schedule whereas for Pascal VOC 2012 and Pascal Context, Published as a conference paper at ICLR 2024 we use the 80k iterations schedule. Two changes are made w.r.t. the default schedules configurations. First, we use the Adam optimizer (Kingma & Ba, 2014) instead of SGD and second, we use eta min = 0.1 lr instead of eta min = 1e-4. For each pretraining method and dataset, we run with four different learning rates (8e-5, 3e-5, 1e-5, and 8e-6) and report the highest m Io U score. B CLUSTERING ALGORITHM We briefly detail the clustering algorithm, particularly the distinctions from its original implementation proposed in Stegm uller et al. (2023). The overall objective of the clustering step is to assign the spatial tokens from both views of a common image to K semantically coherent groups. Formally, given the dense representations z1 and z2 of two augmented views, the joint representation zcat R2N d is obtained by a concatenation along the dimension of the tokens. The kth centroid ck Rd is initialized by uniformly sampling without replacement one of the 2N tokens. The clustering procedure relies on the Sinkhorn-Knopp algorithm (Cuturi, 2013) to solve the optimal transportation problem of assigning 2N tokens to K clusters/centroids. The transportation cost between tokens and centroids T(sem) R2N K is given by the negative cosine distance: T(sem) n,k = < ck, zn cat > ck zn cat (12) The Sinkhorn-Knopp algorithm efficiently computes a solution (an assignment matrix Q ) to the following optimization problem: Q = argmin Q Q < Q, T(sem) > 1 which aims to minimize the inner product of the transportation plan and the transportation cost, subject to an entropy constraint on the assignments. The transportation polytope Q defines the set of valid assignments: Q = {Q R2N K + | Q1K = 1 2N 12N, Q 12N = 1 We update the centroids by pooling over each cluster: n [2N] Q (n, k) zn cat (15) At this point, we could stop, but we observed improved results when doing multiple iterations of the above steps, i.e., restarting the procedure from Equation (12) and the centroids found in Equation (15). Empirically, we found that 5 iterations is a good trade-off between speed and accuracy. In practice, we use the hard version of the assignments obtained by ensuring that each token is only assigned to a single cluster. After a final l1-normalization of the columns of the herein obtained matrix, the view-wise object representations are computed as in Equation (1). Importantly, the above-defined algorithm diverges from the one proposed in Cr OC (Stegm uller et al., 2023) in multiple ways. The main difference is that we do not rely on an iterative procedure to find the optimal number of centroids K, and as such, the number of clusters is fixed. Another discrepancy is that we use uniform distributions for both the sampling of the initial centroids and the marginals of the assignments (see Eqs. (13) and (14)). We observed that these modifications simplify the overall algorithm without incurring a drop in performance. In the following paragraph, we discuss our adjustments to the positional cues of the clustering algorithm. Positional cues An important property of the selected clustering algorithm is its ability to leverage positional cues to compensate for the lack of high-level semantics in the features at the beginning of the training. The spatial guidance is incorporated in the algorithm via an additive constraint on the transportation cost matrix: T(tot) = T(sem) + λpos T(pos) (16) Published as a conference paper at ICLR 2024 Table 5: Ablation study on the individual contribution of each loss. Vi T-S/16 are trained on COCO for 300 epochs and evaluated with downstream tasks of different levels of granularity. Patch-level NN Object Detection with YOLO-S Image-level NN Lg cv Lo cv Lo ci Pascal VOC (m Io U) ADE20k (m Io U) COCO (APbb) COCO (APbb 50) COCO (APbb 75) IN1K (top-1 Acc.) 31.6 12.8 24.0 40.6 23.9 34.3 47.1 18.3 27.1 44.2 27.6 35.8 58.1 23.4 30.2 48.4 31.0 38.2 Table 6: Object detection with YOLO-S. Backbones pre-trained with different self-supervised learning methods are finetuned with YOLO-S (He et al., 2017). We report the mean average precision for the predicted bounding boxes (APbb) for different Io U thresholds on COCO. Method Backbone Params Dataset Epochs APbb APbb 50 APbb 75 Object-centric DINO Caron et al. (2021) Vi T-S/16 21M IN1K 800 35.4 56.0 36.6 Cr OC Stegm uller et al. (2023) Vi T-S/16 21M IN1K 300 35.7 55.5 37.0 Time T Salehi et al. (2023) Vi T-S/16 21M IN1K+YTVOS 800+30 35.0 55.5 36.2 Ours Cr IBo Vi T-S/16 21M IN1K 800 35.7 55.6 37.4 Similar to Cr OC, we use the 2D coordinates E(cat) R2N 2 of every patch/token in the joint dense representation zcat w.r.t. the upper left corner of the image that generated the two views. Up to normalization constants, the distance between the ith token and the jth centroid is given by: T(pos) ij = arg min l Sj ||E(cat) i E(cat) l ||2 (17) where Sj is the set of tokens assigned to the jth centroid from the previous iteration. It is noteworthy that the positional cost differs from that of Cr OC, which sets the position of a given cluster as the average position of the tokens assigned to it. In our approach, the distance between a token and a given centroid depends on the query token. This mitigates the issue of concentric objects sharing the same position. In Table 4, we ablate over the hyperparameter λpos, which modulates the importance of the positional cost in the clustering algorithm. C ADDITIONAL EXPERIMENTS C.1 CONTRIBUTION OF INDIVIDUAL LOSS TERMS Consecutive evaluations showcase the contribution of each loss in Table 5 on 3 different downstream tasks. It follows that all losses contribute a significant improvement to the overall performance of Cr IBo. C.2 OBJECT DETECTION WITH YOLO-S We consider an object detection downstream task to shed a different light on the quality of the learned embeddings. We perform an end-to-end finetuning of the backbone using YOLO-S (Fang et al., 2021) on COCO Lin et al. (2014). We assess performance based on the standard Average Precision metrics, namely i.e., APbb 50, APbb 75 and APbb, computed from bounding boxes predicted on the validation set. We strictly follow the finetuning protocol from Fang et al. (2021), with the sole modification being an adjustment to a batchsize of 32, as opposed to the originally used batchsize of 8. Our choice of object detection method was guided by its simplicity and minimal requirement for additional parameters, aiming to capture the nuances of various pretraining methods. However, Table 6 reveals that YOLO-S (Fang et al., 2021) is relatively insensitive to the pretraining scheme and provides less information than anticipated. In hindsight, it might have been more judicious to start Published as a conference paper at ICLR 2024 Figure 4: Visualization of cross-image object-level matchings on COCO. For a given query view (considered as view 1 here), object representations are computed via clustering. For each objectrepresentation ck 1 (highlighted in unique colors), its nearest neighbor nn(ck 1, M1) in the memory bank M1 is visualized. In this visualization, K = 12. with the parameters specific to YOLO-S already trained. Indeed, this approach involves appending a hundred detection tokens to the image tokens before transmitting the resulting sequence through the transformer encoder. As such, earlier layers are at risk of receiving strong and biased gradients at the beginning of the finetuning phase, potentially overriding the properties learned during pretraining and explaining the observed results. On the other hand, the results in Table 6 are typically in the ballpark of those reported in the original paper (Fang et al., 2021). C.3 SEGMENTATION WITH A LINEAR HEAD AND CONCATENATION We replicate the same linear segmentation experiment from the main text (see Section 4.2), but we concatenate the spatial tokens from the last 4 layers of the Vi Ts in Table 7. This scenario yields improved overall performance, but it s not clear whether these improvements result from the fact that different layers of the Vi T encode different information or from the increased number of trainable parameters. C.4 CYCLE CONSISTENCY BRINGS STABILITY We evaluate the effect of the cycle consistency criterion by comparing the results of various pretraining settings with and without the bootstrapping condition. Training is performed on Image Net-1k for 300 epochs for every combination of the following hyperparameters: - K: {4, 8, 12}. - λpos: {1., 2., 4.}. Table 7: Linear segmentation with frozen backbones. The linear decoder from Segmenter (Strudel et al., 2021) is trained on the frozen and concatenated spatial features from the last 4 layers of Vi Ts pretrained with various self-supervised learning methods. We report the m Io U scores achieved on the validation sets of 4 different datasets. Publicly available checkpoints are used when possible, and our reproduced results are indicated with a symbol. Method Backbone Params Epochs Dataset Pascal Context Pascal VOC COCO-Stuff 164K ADE20K Scene-centric DINO (Caron et al., 2021) Vi T-S/16 21M 300 COCO 28.9 47.4 21.6 20.3 MAE (He et al., 2022) Vi T-S/16 21M 300 COCO 19.9 29.4 13.0 13.5 Cr OC (Stegm uller et al., 2023) Vi T-S/16 21M 300 COCO 32.2 54.7 25.9 24.6 Cr IBo Vi T-S/16 21M 300 COCO 37.9 62.4 30.3 29.9 Object-centric DINO (Caron et al., 2021) Vi T-S/16 21M 800 IN1K 41.3 69.0 33.0 31.0 Cr OC (Stegm uller et al., 2023) Vi T-S/16 21M 300 IN1K 39.9 71.4 33.4 31.3 Time T (Salehi et al., 2023) Vi T-S/16 21M 800+30 IN1K+YTVOS 41.9 71.9 34.5 32.3 Cr IBo Vi T-S/16 21M 800 IN1K 42.4 75.2 34.6 33.4 DINO (Caron et al., 2021) Vi T-B/16 85M 400 IN1K 43.1 74.2 29.9 34.5 MAE (He et al., 2022) Vi T-B/16 85M 1600 IN1K 32.7 53.0 19.7 23.5 Cr IBo Vi T-B/16 85M 400 IN1K 43.0 75.7 36.5 35.4 Published as a conference paper at ICLR 2024 Figure 5: Visualization of cross-image object-level matchings on Image Net-1k. For a given a query view (considered as view 1 here), object representations are computed via clustering. For each object-representation ck 1 (highlighted in unique colors), its nearest neighbor nn(ck 1, M1) in the memory bank M1 is visualized. In this visualization, K = 12. - S: {25k, 50k}. We then evaluate the resulting models with the dense nearest neighbor retrieval task (see Section 4.1). In Table 8, we report the aggregated m Io U results for two datasets (Pascal VOC 2012 (Everingham et al.) and ADE20K (Zhou et al., 2017)) and the two settings, i.e., with and without the cycle consistency condition. The aggregation is either the minimum, the maximum, or the average of all the results. As can be observed in Table 8, the proposed criterion does not bring improvements per se, but stabilizes the training procedure. Indeed, under the systematic bootstrapping regime, the performance can be unpredictable, as exemplified by the large variations of m Io U scores. This happens when nearest neighbors are determined based on low-level cues, a scenario that adaptive bootstrapping strives to prevent through the employment of cycle consistency. C.5 VISUALIZATION OF CROSS-IMAGE OBJECT-LEVEL MATCHINGS Figure 4 and Figure 5 show the cross-image object-level matchings during a pretraining on COCO and Image Net-1k, respectively. Even though Image Net-1k is an object-centric dataset, classes can always be decomposed into finer grained classes. For example, the first image with the bird has tree branches present, which Cr IBo does not fail to recognize. C.6 BOOTSTRAPPING RATIO The evolution of the bootstrapping ratio over the training epochs is illustrated in Figure 6. Using the notations from the paper, the bootstrapping ratio corresponds to approximately Z1+Z2 2K averaged over the whole dataset. Z1 and Z2 are the normalization constants from Equation (8). In practice, the number of object representations in an image is slightly smaller than K as we discard the ones that do not span both views. At the initial epoch, the encoders are randomly initialized, making it very hard for the object representations to satisfy the cycle consistency condition. However, as learning progresses, representations become more semantically driven, and more bootstrapping is allowed. It Table 8: Ablation study on the adaptive bootstrapping using the cycle consistency condition. Results are aggregated over multiple hyperparameters. Trained on Image Net-1k for 300 epochs and evaluated using dense nearest neighbor retrieval (Tab. 1). Pascal VOC ADE20k Bootstrap criterion min. m Io U max. m Io U avg. m Io U min. m Io U max. m Io U avg. m Io U None 49.1 71.2 68.8 16.4 27.2 25.7 Cycle-consistency 68.0 71.5 69.9 24.8 27.6 26.0 Published as a conference paper at ICLR 2024 0 50 100 150 200 250 300 Epoch Bootstrapping ratio [%] K=4 K=8 K=16 K=32 K=64 Figure 6: Bootstrapping ratio over the epochs. More bootstrapping is allowed as the training progresses, and as the representations become more semantically driven. Table 9: High-level profiling of Cr IBo. Experiments are run on a single node with 4x AMD MI250x (2 compute die per GPU i.e., worldsize = 8) with a memory usage of 43.5 GB per compute die. The backbone is a Vi T-B/16 and the batchsize is set to 128 per compute die i.e., 1024 in total and K = 12. Cr IBo specific operations are highlighted (operations specific to object-level nearest neighbors). Description Operation Absolute time per iteration [ms] Relative time [%] Forward pass ft( ) + fs( ) + ht( ) + hs( ) 453.4 41.7 Clustering Q = C( ) 82.2 7.6 NN retrieval nn( ) + update(Mi) 47.5 4.4 Object projections ht( ) + hs( ) 11.2 1.0 Weights update backprop. + EMA 493.1 45.3 total 1087.4 100 is worth noting that a higher K results in a lower bootstrapping ratio as both the batch and the queue are larger (in terms of object representations), which imposes a more stringent condition to meet. C.7 HIGH-LEVEL PROFILING A detailed timing analysis of Cr IBo is performed in Table 9. It is worth noting that Cri Bo-specific operations take less than 15% of the total time. C.8 RUNTIME COMPARISON A runtime comparison with other methods can be found in Table 10. Cr IBo is faster than DINO and slower than MAE. Note that the quantitative comparisons in our paper are done with MAE trained for 1600 epochs and DINO/Cr IBo trained for 400 epochs. This leads to a similar total computational budget. Table 10: Runtime comparison. Experiments are run on a single node with 4x AMD MI250x (2 compute die per GPU i.e. worldsize = 8). Method Time per epoch [minutes:seconds] Memory per compute die [GB] Batchsize MAE 04:45 36 4096 DINO 29:43 45 1024 Cr IBo (K = 12) 23:58 44 1024 Cr IBo (K = 32) 28:22 60 1024 Published as a conference paper at ICLR 2024 Table 11: We investigate the usage of supervision in Cr IBo. To get a sense of how supervision could help Cr IBo, we ablate over the usage of labels for the clustering step and for the bootstrapping criterion. The hyperparameters (where applicable) are set to (λpos, S, K) = (2.0, 25k, 64). Supervision Patch-level NN Clustering Bootstrapping Pascal VOC ADE20k 52.5 21.2 52.3 20.2 59.5 23.4 58.1 23.4 C.9 SUPERVISED ORACLES An ablation using different supervised oracles is shown in Table 11. The supervision takes place at two different levels: 1) the clustering and 2) the bootstrapping criterion. 1. Instead of using an unsupervised clustering algorithm for computing object representations, the semantic segmentation labels from COCO and COCO-Stuff are used. The masks are downsampled (16x) to match the dense representation output resolution. Given an image, the object representation associated with a given mask label is the average pooling of the patches assigned to the label. 2. Instead of using the cycle-consistency condition for discarding supposed invalid pairs of NNs, the downsampled (16x) semantic masks of COCO and COCO-Stuff are used. A pair of NN object representations is deemed valid if the union of all patch labels is the same in both object representations. Surprisingly, using unsupervised clustering results in better performance than using the labels from COCO. In hindsight, this confirms again that the overclustering regime is where Cr IBo is most at ease. Indeed, the granularity of the mask annotations in COCO is not particularly high. The results would have probably been different using finer-grained annotations. However, replacing the cycleconsistency condition with a supervised oracle yields modest improvements. C.10 CLUSTERING ALGORITHM The clustering algorithm plays a pivotal role in the overall performance of Cr IBo. We compare the clustering algorithm, detailed in Appendix B, with a well-established alternative, i.e., the K-Means algorithm. Consequently, we pretrain Vi T-S/16 for 300 epochs on COCO (Lin et al., 2014), using a memory bank of size S = 25k, various numbers of clusters K along with the K-Means algorithm. Table 12 depicts the performance achieved by each clustering algorithm on the dense nearest neighbor retrieval task (see Sec. 4.1). It can be observed that using the Sinkhorn-based clustering from Appendix B leads to an improvement over the K-Means algorithm. This difference is only partially explained by the lack of positional cues in K-Means as shown by the row λpos = 0.0. Indeed, the Sinkhorn-based method, deprived of positional cues, still yields better scores, confirming the utility of the non-standard clustering algorithm. Table 12: Ablation study on the clustering algorithm. K = 4 K = 8 K = 16 K = 32 K = 64 Clustering Pascal VOC ADE20K Pascal VOC ADE20K Pascal VOC ADE20K Pascal VOC ADE20K Pascal VOC ADE20K K-Means 42.8 12.5 47.6 13.8 49.1 14.3 52.0 15.3 53.0 16.0 Sinkhorn (λpos = 0.0) 47.2 16.7 50.1 17.7 52.4 18.9 53.8 20.1 55.9 21.6 Sinkhorn (λpos = 2.0) 49.9 18.1 52.9 19.4 54.6 20.3 57.0 21.7 58.1 23.4 Published as a conference paper at ICLR 2024 Pascal VOC 2012 (Everingham et al.) A dataset composed of a training set includes 10,582 images distributed across 21 classes, with one being a background class. The validation set incorporates 1,449 images. Pascal Context (Mottaghi et al., 2014) A scene-centric dataset comprises 4,998 training images spanning 60 semantic classes (including the background). The validation set contains 5,105 images. COCO-Stuff 164K (Caesar et al., 2018) A scene-understanding dataset, featuring labels spanning 91 stuff categories and 80 things categories. The training set is composed of 118K images, while the validation set contains 5K images. ADE20K (Zhou et al., 2017) A dataset that encompasses scenes featuring fine-grained labels across 150 distinct semantic categories, making it one of the most demanding available semantic segmentation datasets. The training set comprises 20,210 images, and the validation set consists of 2,000 images.