# selfsupervised_visual_representation_learning_with_semantic_grouping__405ce509.pdf Self-Supervised Visual Representation Learning with Semantic Grouping Xin Wen1 Bingchen Zhao2, 3 Anlin Zheng1, 4 Xiangyu Zhang4 Xiaojuan Qi1 1University of Hong Kong 2University of Edinburgh 3Lunar AI 4MEGVII Technology {wenxin, xjqi}@eee.hku.hk zhaobc.gm@gmail.com {zhenganlin, zhangxiangyu}@megvii.com In this paper, we tackle the problem of learning visual representations from unlabeled scene-centric data. Existing works have demonstrated the potential of utilizing the underlying complex structure within scene-centric data; still, they commonly rely on hand-crafted objectness priors or specialized pretext tasks to build a learning framework, which may harm generalizability. Instead, we propose contrastive learning from data-driven semantic slots, namely Slot Con, for joint semantic grouping and representation learning. The semantic grouping is performed by assigning pixels to a set of learnable prototypes, which can adapt to each sample by attentive pooling over the feature and form new slots. Based on the learned data-dependent slots, a contrastive objective is employed for representation learning, which enhances the discriminability of features, and conversely facilitates grouping semantically coherent pixels together. Compared with previous efforts, by simultaneously optimizing the two coupled objectives of semantic grouping and contrastive learning, our approach bypasses the disadvantages of hand-crafted priors and is able to learn object/group-level representations from scene-centric images. Experiments show our approach effectively decomposes complex scenes into semantic groups for feature learning and significantly benefits downstream tasks, including object detection, instance segmentation, and semantic segmentation. Code is available at: https://github.com/CVMI-Lab/Slot Con. 1 Introduction Existing self-supervised approaches have demonstrated that visual representations can be learned from unlabeled data by constructing pretexts such as transformation prediction [22], instance discrimination [70, 31], and masked image modeling [2, 30, 67], etc. Among them, approaches based on instance discrimination [28, 6, 7, 11], which treat each image as a single class and employ a contrastive learning objective for training, have attained remarkable success and are beneficial to boost performance on many downstream tasks. However, this success is largely built upon the well-curated object-centric dataset Image Net [15], which has a large gap with the real-world data for downstream applications, such as city scenes [13] or crowd scenes [46]. Directly applying the instance discrimination pretext to these real-world data by simply treating the scene as a whole overlooks its intrinsic structures (e.g., multiple objects and complex layouts) and thus will limit the potential of pre-training with scene-centric data [61]. This leads to our focus: learning visual representations from unlabeled scene-centric data. Recent efforts to address this problem can be coarsely categorized into two types of research. One stream extends the instance discrimination task to pixel level for dense representation learning [54, 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Pixel-level Image-level Object-level (Ours) Consistency Consistency Consistency Object-centric Scene-centric Semantic Grouping Figure 1: Left: Unlike object-centric images that highlight one iconic object, scene-centric images often contain multiple objects with complex layouts. This adds to the data diversity and increases the potential of image representations, yet it challenges previous learning paradigms that simply treat an image as a whole or individual pixels. Middle: Contrastive-learning objectives built upon different levels of image representations, in which object-level contrastive learning is viewed as the most suitable solution for scene-centric data. Right: We jointly learn a set of semantic prototypes in order to perform semantic grouping over the pixel-level representations and form object-centric slots. 66, 73], which shows strong performance in downstream dense prediction tasks. Yet, these methods still lack the ability to model object-level relationships presented in scene-centric data, which is crucial for learning representations. Although another stream of works attempts to perform objectlevel representation learning, most of them still heavily rely on domain-specific priors to discover objects, which include saliency estimators [62, 56], unsupervised object proposal algorithms [68, 72], hand-crafted segmentation algorithms [81, 34] or unsupervised clustering [35]. However, if the representation is supervised by hand-crafted objectness priors, it will be discouraged from learning objectness from the data itself and prone to mistakes from priors. Therefore, the capability and generalizability of the representation will be limited. In this work, we aim at a fully learnable and data-driven approach to enable learning representations from scene-centric data for enhanced effectiveness, transferability and generalizability. We propose contrastive learning from data-driven semantic slots, namely Slot Con, for joint semantic grouping and representation learning. Semantic grouping is formulated as a feature-space pixel-level deep clustering problem where the cluster centers are initialized as a set of learnable semantic prototypes shared by the dataset, and grouping is achieved by assigning pixels to clusters. The cluster centers can then adapt to each sample by softly assigning pixels into cluster centers and aggregating their features via attentive pooling to form new ones, also called slots. Further, upon the learned slots from two random views of one image, a contrastive objective, which attempts to pull positive slots (i.e., slots from the same prototype and sample) together and push away negative ones, is employed for representation learning. The optimized representations will enhance the discriminability of prototypes and slots, which conversely facilitates grouping semantically coherent pixels together. Compared with previous efforts, by simultaneously optimizing the two coupled objectives of semantic grouping and contrastive representation learning, our method bypasses the disadvantages of hand-crafted objectness priors1 and is able to discover object/group-level representations from scene-centric images. We extensively assess the representation learning ability of our model by conducting transfer learning evaluation on COCO [46] object detection, instance segmentation, and semantic segmentation on Cityscapes [13], PASCAL VOC [20], and ADE20K [83]. Our method shows strong results with both COCO pre-training and Image Net-1K pre-training, bridging the gap between scene-centric and object-centric pre-training. As a byproduct, our method also achieves notable performance in unsupervised segmentation, showing strong ability in semantic concept discovery. In summary, our main contributions in this paper are: 1) We show that the decomposition of natural scenes (semantic grouping) can be done in a learnable fashion and jointly optimized with the representations from scratch. 2) We demonstrate that semantic grouping can bring object-centric representation learning to large-scale real-world scenarios. 3) Combining semantic grouping and representation learning, we unleash the potential of scene-centric pre-training, largely close its gap with object-centric pre-training and achieve state-of-the-art results in various downstream tasks. 1Things can be different if the data is not reliable, see Section G in the appendix for details. 2 Related work Our work is in the domain of self-supervised visual representation learning, where the goal is to learn visual representations without human annotations. We briefly review relevant works below. Image-level self-supervised learning aims at learning visual representations by treating each image as one data sample. To this end, a series of pretext tasks are designed in which the labels are readily available without human annotations. Early explorations range from low-level pixel-wise reconstruction tasks that include denoising [64], inpainting [55], and cross-channel prediction [80] to higher-level instance discrimination [17], rotation prediction [22], context prediction [16], jigsaw puzzle [52], counting [53], and colorization [79]. Modern variants of instance discrimination [17] equipped with contrastive learning [60, 36] have shown strong potential in learning transferable visual representations [70, 8, 31, 58, 84, 82]. Other works differ in their learning objectives but still treat an image as a whole [28, 10, 78]. To further utilize the complex structure in natural images, some works exploit local crops [6, 71, 74, 61] while others either step to pixelor object-level, detailed as follows. Pixel-level contrastive learning extends the instance discrimination task from image-level feature vectors to feature maps [54, 47, 66, 73]. Their main differences lie in the way positive pixel-pairs are matched (spatial adjacency [54, 73], feature-space nearest-neighbor [66], sink-horn matching [47]), and the image-level baseline they build upon (Mo Co v2 [54, 66], BYOL [47, 73]). Their pixellevel objective naturally helps learn dense representations that are favorable for dense prediction downstream tasks but lacks the grasp of holistic semantics and commonly require an auxiliary image-level loss to attain stronger performance [66, 73]. Object-level contrastive learning first discovers the objects in images and applies the contrastive objective over them, achieving a good balance in fine-grained structure and holistic semantics, yielding strong empirical gains with both object-centric [34, 35] and scene-centric data [72]. The key issue lies in finding objects in an image without supervision. Current works, however, still heavily rely on heuristic strategies that include saliency estimators [62, 56], selective-search [68, 72], hand-crafted segmentation algorithms [81, 34], or k-means clustering [35]. In contrast, our semantic grouping method is fully learnable and end to end, ensuring transferability and simplicity. Unsupervised semantic segmentation is an emerging task that targets addressing semantic segmentation with only unlabeled images. The first attempt of IIC [40] maximizes the mutual information of augmented image patches, and later works [38, 62] rely on saliency estimators as a prior to bootstrap semantic pixel representations. Recently, Pi CIE [39] adopt pixel-level deep clustering [5] to cluster the pixels into semantic groups, which Seg Discover [37] further improves by adopting super-pixels. On the other hand, Leopart [85], STEGO [29], and Free SOLO [65] exploits pre-trained networks attention maps for objectness. Still, they commonly rely on a (self-supervised) pre-trained network for initialization, while our method is trained fully from scratch. Object-centric representation learning is viewed as an essential component of data-efficient, robust and interpretable machine learning algorithms [27]. Towards unsupervised object-centric representation learning, a series of works have been proposed [26, 3, 19, 48, 25, 24, 23]. Directly extracting objects from images is challenging due to the lack of supervision and thus prior works have long been restricted to synthetic data [42, 51, 1]. Recent works try to step toward real-world videos; however, they either adopt motion [75, 41] or depth [18] as a cue for objectness to solve this problem. Instead, our method shows that we can first explicitly optimize for desirable properties of clusters (that can describe an object) over the dataset, then retrieve the objects from an image with the learned prototypes. More importantly, it first shows the possibility of learning object-centric representations from large-scale unlabelled scene-centric natural images. 3.1 Semantic grouping with pixel-level deep clustering Given a dataset D of unlabeled images, we aim at learning a set of prototypes S that classifies each pixel into a meaningful group, such that pixels within the same group are semantic-consistent (have Dense Projection Inverse Augmentation (Ro IAlign) Encoder & Projector Inverse Augmentation (Ro IAlign + Flip) Assignment Slots Figure 2: Overview of our proposed framework. Based on a shared pixel embedding function, the model learns to classify pixels into groups according to their feature similarity in a pixel-level deep clustering fashion (Sec. 3.1); the model produces group-level feature vectors (slots) through attentive pooling over the feature maps, and further performs group-level contrastive learning (Sec. 3.2). We omit the symmetrized loss computed by swapping the two views for simplicity. (best viewed in color) similar feature representations), and pixels between different groups are semantic-incoherent. We find that this problem can be viewed as unsupervised semantic segmentation [22, 39], and solved with pixel-level deep clustering [5, 76, 6, 7]. Intuitively, a meaningful semantic grouping should be invariant to data augmentations. Thus, for different augmentations of the same image, we enforce the pixels that lie in the same location to have similar assignment scores w.r.t. the same set of cluster centers (prototypes). Besides consistent grouping, the groups should be different from each other to ensure that the learned representations are discriminative and avoid trivial solutions, e.g., identical features. Together with standard techniques used in self-supervised learning (e.g., non-linear projector and momentum teacher [28, 7], etc.), this leads to the following framework. Specifically, as illustrated in Figure 2, our method contains two neural networks: the student network is parameterized by θ and include an encoder fθ, a projector gθ and a set of K learnable prototypes Sθ = s1 θ, s2 θ, . . . , s K θ RK D; the teacher network holds the same architecture with the student, but uses a different set of weights ξ that updates as an exponential moving average of θ. Given an input image x, two random augmentations are applied to produce two augmented views v1 and v2. Each augmented view vl {v1, v2} is then encoded with an encoder f into a hidden feature map hl RH W D, and then transformed with a multilayer perceptron (MLP) g to get zl = g hl RH W D. We then compute the assignment Pl θ of the projections zl θ with the corresponding prototypes Sθ, and enforce it to match the assignment Ql ξ produced with another view vl by the teacher network. More precisely, with ℓ2-normalized projections z = z/ z and prototypes s = s/ s , Ql ξ = softmax K zl ξ S ξ c /τt , Pl θ = softmax K zl θ S θ /τs RH W K , (1) with τs, τt > 0 temperature parameters that control the sharpness of the output distribution of the two networks. The c can be omitted and will be explained later. Note that the scale or layout of the two feature maps zl θ and zl ξ can be inconsistent due to geometric image augmentations, e.g., random crop, scale or flip. So we perform the inverse augmentation process (including Ro IAlign [32] and an optional flipping, details provided in Section A.1 in the appendix) on the predicted assignments to align their spatial locations: e Ql ξ = invaug Ql ξ , e Pl θ = invaug Pl θ . The inverse augmentation is performed on the assignments Q rather than the projections z to keep the context information out of the overlapping area for slot generation, which will be detailed in Section 3.2. Based on the aligned assignments, we apply the cross-entropy loss LCE θ,ξ qξ, pθ = P k qk ξ log pk θ to enforce the consistency in assignment score between spatial-aligned pixels from different views. The cross-entropy loss is averaged over all spatial locations to produce the grouping loss: LGroup θ,ξ = 1 H W h LCE θ,ξ e Q2 ξ[i, j], e P1 θ [i, j] + LCE θ,ξ e Q1 ξ[i, j], e P2 θ [i, j] i . (2) Directly optimizing the above objective resembles an unsupervised variant of Mean Teacher [57], which collapses as shown in [28]. In order to avoid collapsing, we follow [7] to maintain a mean logit c RK and reduce it when producing the teacher assignments Qξ, as indicated in Eq. 1. The mean logit stores an exponential moving average of all the logits produced by the teacher network: c λcc + (1 λc) 1 B H W z(i) ξ [j, k] S ξ , (3) where B stands for the batch size. Intuitively, reducing the mean logit amplifies the difference in assignment between different pixels, preventing all pixels from being assigned to the same prototype. Besides that, the teacher temperature τt is smaller than the student temperature τs to produce a sharper target and avoid uniform assignments. Both operations help avoid collapse and force the network to learn a meaningful semantic grouping. Discussion with DINO. The resulting solution for semantic grouping may seem like a naive extension of DINO [7]. However, this is not the whole picture. DINO is an image-level representation learning approach that adopts a large number of prototypes (e.g., 65536), while our objective is built on pixel-level representations and is tailored to learn meaningful semantic groups. Intuitively, DINO captures scene semantics that can be more complex and thus requires more prototypes to represent all scene-level variations, while our method learns object-level semantics which can be composed to depict a complicated scene and thus only requires a small number of prototypes (e.g., 256 for COCO). Besides, the representation learning objective of DINO is built on prototypes that are shared by the whole dataset, while our method instead adapts prototypes for each image and performs contrastive learning over groups (detailed in Section 3.2). In summary, we present a novel view for the decoupling of pixel-level online clustering (i.e., semantic grouping) and object-level representation learning. 3.2 Group-level representation learning by contrasting slots Inspired by Slot Attention [48], we then reuse the assignments computed by the semantic grouping module in Eq. 1 to perform attentive pooling over the dense projections z to produce group-level feature vectors (rephrased as slots), as shown in Figure 2. Intuitively, as the softmax normalization applies to the slot dimension, the attention coefficients sum to one for each individual input feature vector. As a result, the soft assignments A of the dense projections z w.r.t. the corresponding prototypes S can also be viewed as the attention coefficients that describe how the prototypes decompose the dense projections into non-overlapping groups. This inspires us to decompose the dense projections with attentive pooling following [48]. Specifically, for a dense projection zl θ produced from view vl, we extract K slots: i,j Al θ[i, j] i,j Al θ[i, j] zl θ[i, j] RK D, Al θ = softmax K zl θ S θ /τt RH W K , (4) where denotes the Hadamard product, and a similar operation applies for the teacher network to produce Sl ξ . Since the initial slots are shared by the whole dataset, the corresponding semantic may be missing in a specific view vl, thus producing redundant slots. Therefore, we compute the following binary indicator 1l to mask out the slots that fail to occupy a dominating pixel: 1k,l θ = i,j such that argmax K Al θ [i, j] = k , (5) and 1l ξ is computed similarly. Following the literature in object-level SSL [62, 34, 72, 35], we then apply a contrastive learning objective to discriminate the slot that holds the same semantic across views from distracting slots with the Info NCE [60] loss: LInfo NCE θ,ξ Sl θ, Sl ξ = 1 k=1 log 1k,l θ 1k,l qθ sk,l θ sk,l k 1k,l θ 1k ,l ξ exp qθ sk,l θ sk ,l ξ /τc . (6) This objective helps maximize the similarity between different views of the same slot, while minimizing the similarity between slots from another view with different semantics and all slots from other images. Note that here an additional predictor qθ with the same architecture as the projector gθ is applied to the slots Sθ as empirically it yields stronger performance [11, 68, 35]. And the resulting slot-level contrastive loss also follows a symmetric design like Eq. 2: LSlot θ,ξ = LInfo NCE θ,ξ S1 θ, S2 ξ + LInfo NCE θ,ξ S2 θ, S1 ξ . (7) 3.3 The overall optimization objective We jointly optimize the semantic grouping objective (Eq. 2) and the group-level contrastive learning objective (Eq. 7), controlled with a balancing factor λg: LOverall θ,ξ = λg LGroup θ,ξ + (1 λg)LSlot θ,ξ . (8) At each training step, the student network is optimized with gradients from the overall loss function: θ optimizer θ, θLOverall θ,ξ , η , where η denotes the learning rate; and the teacher network updates as an exponential moving average of the student network: ξ λtξ + (1 λt)θ, with λt denoting the momentum value. After training, only the teacher encoder fξ is kept for downstream tasks. 4 Experiments 4.1 Implementation details Table 1: Details of the datasets used for pre-training. Dateset #Img. #Obj./Img. #Class Image Net-1K [15] 1.28M 1.7 1000 COCO [46] 118K 7.3 80 COCO+ [46] 241K N/A N/A Pre-training datasets. We pre-train our models on COCO train2017 [46] and Image Net-1K [15], respectively. COCO train2017 [46] contains 118K images of diverse scenes with objects of multiple scales, which is closer to real-world scenarios. In contrast, Image Net-1K is a curated object-centric dataset containing more than 1.28M images, which is better for evaluating a model s potential with large-scale data. Besides, we also explore the limit of scene-centric pre-training on COCO+, i.e., COCO train2017 set plus the unlabeled2017 set. See details in Table 1. Data augmentation. The image augmentation setting is the same as BYOL [28]: a 224 224-pixel random resized crop with a random horizontal flip, followed by a random color distortion, random grayscale conversion, random Gaussian blur, and solarization. The crop pairs without overlap are discarded during training. Network architecture. We adopt Res Net-50 [33] as the default encoder for fθ and fξ. The projector gθ, gξ and predictor qθ are MLPs whose architecture are identical to that in [7] with a hidden dimension of 4096 and an output dimension of 256. Optimization. We adopt the LARS optimizer [77] to pre-train the model, with a batch size of 512 across eight NVIDIA 2080 Ti GPUs. Following [73], we utilize the cosine learning rate decay schedule [50] with a base learning rate of 1.0, linearly scaled with the batch size (Learning Rate = 1.0 Batch Size/256), a weight decay of 10 5, and a warm-up period of 5 epochs. Following [72, 66, 73], the model is pre-trained for 800 epochs on COCO(+) and 100/200 epochs on Image Net, respectively. Following the common practice of [72, 73], the momentum value λt of the teacher model starts from 0.99 and is gradually increased to 1 following a cosine schedule. Synchronized batch normalization and automatic mixed precision are also enabled during training. Hyper-parameters. The temperature values τs and τt in the student and teacher model are set to 0.1 and 0.07, respectively. Besides, the center momentum λc is set to 0.9. The default number of prototypes K is set to 256 for COCO(+) and 2048 for Image Net, according to our empirical finding that setting the number of prototypes close to the number of human-annotated categories can help downstream performance. The temperature value τc for the contrastive loss is set to 0.2 following [11], and the default balancing ratio λg is set to 0.5. 4.2 Evaluation protocols Following the common practice of previous self-supervised works [66, 73, 72], we evaluate the representation ability of the pre-trained model by taking it as the backbone of downstream tasks. Specifically, we add a newly initialized task-specific head to the pre-trained model for different downstream tasks, i.e., object detection and instance segmentation on COCO [46], and semantic segmentation on PASCAL VOC [20], Cityscapes [13], and ADE20K [83]. Object detection and instance segmentation. We train a Mask R-CNN [32] model with R50FPN [45] backbone implemented in Detectron2 [69]. We fine-tune all layers end-to-end on COCO train2017 split with the standard 1 schedule and report AP, AP50, AP75 on the val2017 split. Following [66, 73, 72] we train with the standard 1 schedule with Sync BN. Semantic segmentation. The evaluation details of PASCAL VOC and Cityscapes strictly follow [31]. We take our network to initialize the backbone of a fully-convolutional network [49] and fine-tune all the layers end-to-end. For PASCAL VOC, we fine-tune the model on train_aug2012 set for 30k iterations and report the mean intersection over union (m Io U) on the val2012 set. For Cityscapes, we fine-tune on the train_fine set for 90k iterations and evaluate it on the val_fine set. For ADE20K, we follow the standard 80k iterations schedule of MMSegmentation [12]. Unsupervised semantic segmentation. We also evaluate the model s ability of discovering semantic groups in complex scenes, which is accomplished by performing unsupervised semantic segmentation on COCO-Stuff [4]. We follow the common practice in this field [40, 39, 37] to merge the labels into 27 categories (15 "stuff" categories and 12 "thing" categories), and evaluate with a subset created by [40]. We perform inference with resolution 320 and number of prototypes 27 following the standard practice. The predicted labels are matched with the ground truth through Hungarian matching [43], and evaluated on m Io U and pixel accuracy (p Acc). 4.3 Transfer learning results COCO pre-training. In Table 2, we show the main results with COCO pre-training. There have been steady improvements in object-level pre-training with COCO, in which the top performance methods are Det Con [44] and ORL [72], which still rely on objectness priors like selective-search [59] or hand-crafted segmentation algorithms [21], yet they still fail to beat the pixel-level state-of-the-art Table 2: Main transfer results with COCO pre-training. We report the results in COCO [46] object detection, COCO instance segmentation, and semantic segmentation in Cityscapes [13], PASCAL VOC [20] and ADE20K [83]. Compared with other image-, pixel-, and object-level self-supervised learning methods, our method shows consistent improvements over different tasks without leveraging multi-crop [6] and objectness priors. ( : re-impl. w/ official weights; : full re-impl.) Method Epochs Multi crop Obj. Prior COCO detection COCO segmentation Semantic seg. (m Io U) APb APb 50 APb 75 APm APm 50 APm 75 City. VOC ADE random init. - 32.8 50.9 35.3 29.9 47.9 32.0 65.3 39.5 29.4 Image-level approaches Mo Co v2 [9] 800 38.5 58.1 42.1 34.8 55.3 37.3 73.8 69.2 36.2 Revisit. [61] 800 40.1 60.2 43.6 36.3 57.3 38.9 75.3 70.6 37.0 Pixel-level approaches Self-EMD [47] 800 39.3 60.1 42.8 - - - - - - Dense CL [66] 800 39.6 59.3 43.3 35.7 56.5 38.4 75.8 71.6 37.1 Pix Pro [73] 800 40.5 60.5 44.0 36.6 57.8 39.0 75.2 72.0 38.3 Object / Group-level approaches Det Con [34] 1000 39.8 59.5 43.5 35.9 56.4 38.7 76.1 70.2 38.1 ORL [72] 800 40.3 60.2 44.4 36.3 57.3 38.9 75.6 70.9 36.7 Ours (Slot Con) 800 41.0 61.1 45.0 37.0 58.3 39.8 76.2 71.6 39.0 Pix Pro [73].2 Our method alleviates such limitations and significantly improves over current objectlevel methods in all tasks, achieving consistent improvement over the previous approaches and even several methods that were pre-trained on the larger dataset Image Net-1K (Table 4). It is also notable that our method can achieve a better performance on the largest and most challenging dataset for segmentation, ADE20K, adding to the significance of this work. COCO+ pre-training. In Table 3, we report the results with COCO+ pre-training. The COCO+ is COCO train2017 plus unlabeled2017 set, which roughly doubles the number of training images and greatly adds to the data diversity. Our method further sees a notable gain in all tasks with extended COCO+ data, and even shows comparable results with our best-performing model pre-trained on Image Net-1K (5 large of COCO+), showing the great potential of scene-centric pre-training. Table 3: Pushing the limit of scene-centric pre-training. Our method further sees a notable gain in all tasks with extended COCO+ data, showing the great potential of scene-centric pre-training. Method Dataset Epochs COCO detection COCO segmentation Semantic seg. (m Io U) APb APb 50 APb 75 APm APm 50 APm 75 City. VOC ADE Slot Con COCO 800 41.0 61.1 45.0 37.0 58.3 39.8 76.2 71.6 39.0 Slot Con Image Net 100 41.4 61.6 45.6 37.2 58.5 39.9 75.4 73.1 38.6 Slot Con Image Net 200 41.8 62.2 45.7 37.8 59.1 40.7 76.3 75.0 38.8 ORL [72] COCO+ 800 40.6 60.8 44.5 36.7 57.9 39.3 - - - Slot Con COCO+ 800 41.8 62.2 45.8 37.8 59.4 40.6 76.5 73.9 39.2 Image Net-1K pre-training. In Table 4, we also benchmark our method with Image Net-1K pretraining and show its compatibility with object-centric data. Without selective-search for object proposals and without pre-training and transferring the FPN head, our method still beats most of the current works and largely closes the gap with the detection-specialized method So Co. Table 4: Main transfer results with Image Net-1K pre-training. Our method is also compatible with object-centric data and shows consistent improvements over different tasks without using FPN [45] and objectness priors. ( : re-impl. w/ official weights; : full re-impl.) Method Epochs w/ FPN Obj. Prior COCO detection COCO segmentation Semantic seg. (m Io U) APb APb 50 APb 75 APm APm 50 APm 75 City. VOC ADE random init. - 32.8 50.9 35.3 29.9 47.9 32.0 65.3 39.5 29.4 supervised 100 39.7 59.5 43.3 35.9 56.6 38.6 74.6 74.4 37.9 Image-level approaches Mo Co v2 [9] 800 40.4 60.1 44.2 36.5 57.2 39.2 76.2 73.7 36.9 Det Co [71] 200 40.1 61.0 43.9 36.4 58.0 38.9 76.0 72.6 37.8 Ins Loc [74] 200 40.9 60.9 44.7 36.8 57.8 39.4 75.4 72.9 37.3 Pixel-level approaches Dense CL [66] 200 40.3 59.9 44.3 36.4 57.0 39.2 76.2 72.8 38.1 Pix Pro [73] 100 40.7 60.5 44.8 36.8 57.4 39.7 76.8 73.9 38.2 Object / Group-level approaches Det Con [34] 200 40.6 - - 36.4 - - 75.5 72.6 - So Co [68] 100 41.6 61.9 45.6 37.4 58.8 40.2 76.5 71.9 37.8 Ours (Slot Con) 100 41.4 61.6 45.6 37.2 58.5 39.9 75.4 73.1 38.6 Ours (Slot Con) 200 41.8 62.2 45.7 37.8 59.1 40.7 76.3 75.0 38.8 4.4 Unsupervised semantic segmentation results Given our approach s consistent improvement in representation learning, we further analysis of how well our semantic grouping component can parse scenes quantitatively and qualitatively. It should be 2Pix Pro aggregates the global context with self-attention [63], so each pixel can also be viewed as an object-level embedding. Table 5: Main results in COCO-Stuff unsupervised semantic segmentation. Method m Io U p Acc Mask Contrast [62] 8.86 23.03 Pi CIE + H. [39] 14.36 49.99 Seg Discover [37] 14.34 56.53 Ours (Slot Con) 18.26 42.36 Image Pi CIE Ours clarified that this is just to help understand the learned prototypes, and not to propose a new SOTA. Unlike current SSL approaches that exhaustively enumerate the massive object proposals and report the best score [7, 35], we follow the common practice of unsupervised semantic segmentation [40, 39] to match the predicted results with the ground-truth using the Hungarian algorithm [43], where each ground-truth label is assigned to a prototype mutual-exclusively. For fair comparisons, the model used for evaluation is trained with K = 27 to match the number of categories of COCO-Stuff. As shown in Table 5, our method can surpass the previous works Pi CIE [39] and Seg Discover [37] with 4 points higher m Io U. Meanwhile, the p Acc is lower since we train the model with a lower resolution (7 7 v.s. 80 80 feature map). Besides Table 5, we also depict the visualization results, in which our method distinguishes confusing objects apart (4th column) and successfully localizes small objects (5th column). Since we only need to separate pixels with different semantics within the same image, the errors in category prediction can be ignored. 4.5 Ablation study Table 6: Ablation studies with COCO 800 epochs pre-training. We show the APb on COCO objection detection and m Io U on Cityscapes, PASCAL VOC, and ADE20K semantic segmentation. The default options are marked with a gray background. (a) Number of prototypes K COCO City VOC ADE 128 40.7 76.4 71.9 38.5 256 41.0 76.2 71.6 39.0 512 40.9 75.6 71.6 38.9 1024 40.7 75.8 70.9 39.1 (b) Loss balancing λg COCO City VOC ADE 0.3 41.0 76.1 72.1 37.9 0.5 41.0 76.2 71.6 39.0 0.7 40.5 75.2 71.5 38.4 1.0 40.4 74.2 70.1 38.6 (c) Teacher temperature τt COCO City VOC ADE 0.04 40.4 75.5 70.2 37.9 0.07 41.0 76.2 71.6 39.0 Number of prototypes K. Table 6a ablates the number of prototypes, we observe that the most suitable K for COCO detection is 256, which is close to its real semantic class number 172 (thing + stuff) [4]. Besides, the performance on Cityscapes and PASCAL VOC have a consistent tendency to drop, while the performance on ADE20K is consistently good if K is big enough. We hypothesize that a suitable K can encourage learning data-specific semantic features, which are only helpful when the pre-training and downstream data are alike (from COCO to COCO); increasing K produces fine-grained features that may lack discriminability in semantics but hold better transferability to ADE20K that require fine-grained segmentation [14]. Loss balancing weight λg. Table 6b ablates the balancing between the semantic grouping loss and the group-level contrastive loss, where the best balance is achieved with both losses treated equally. It is notable that when λg = 1.0, only the semantic grouping loss is applied, and the performance drops considerably, indicating the importance of our group-level contrastive loss for learning good representations. Teacher temperature τt. Table 6c ablates the teacher model s temperature parameter, indicating that a softer teacher distribution with τt = 0.07 helps achieve better performance. 4.6 Probing the prototypes Finally, we analyze whether the prototypes learn semantic meanings by visualizing their nearest neighbors in COCO val2017 split. We first perform semantic grouping on each image to split them into non-overlapping groups (segments), then pool each group to a feature vector, and retrieve the Figure 3: Examples of visual concepts discovered by Slot Con from the COCO val2017 split. Each column shows the top 5 segments retrieved with the same prototype, marked with reddish masks or arrows. Our method can discover visual concepts across various scenarios and semantic granularities regardless of small object size and occlusion. (best viewed in color) top 5 nearest-neighbor segments for each prototype according to cosine similarity. As shown in Figure 3, the prototypes well bind to semantic meanings that cover a wide range of scenarios and semantic granularities from animals, foods, and sports, to furniture, buildings, etc., localizing them well regardless of small object size and occlusion; and notably, without any human annotation. 5 Discussion on the emergence of objectness It should be clarified that the discovered objectiveness is at the semantic level, and object instances with identical semantics can be indistinguishable. Concerning the emergence of objectness, we impose geometric-covariance and photometric-invariance as guiding cues which force the model to decompose a large complex dataset into a small number of clusters through optimizing feature space and cluster centers. The emergence of objects/parts that are compositional and thus occupy a reasonable number of prototypes, can be viewed as a natural consequence under such conditions. Concerning granularity, our intuition is that the prototype number and the dataset distribution generate a bottleneck for the granularity of groups. For example, in the COCO dataset, with 256 prototypes, the model finds that splitting animals into cats, dots, elephants, etc., is enough and won t further separate them. In contrast, for humans (the most occupying category of COCO), as shown in the appendix, the model discovers not only human parts (Figure 7) but also different types of activities (Figure 8), indicating that parts are more helpful in this scenario and deserve more prototypes. 6 Conclusion This work presents a unified framework for joint semantic grouping and representation learning from unlabeled scene-centric images. The semantic grouping is performed by assigning pixels to a set of learnable prototypes, which can adapt to each sample by attentive pooling over the feature map and form new slots. Based on the learned data-dependent slots, a contrastive objective is employed for representation learning, enhancing features discriminability and facilitating the grouping of semantically coherent pixels together. By simultaneously optimizing the two coupled objectives of semantic grouping and contrastive learning, the proposed approach bypasses the disadvantages of handcrafted priors and can learn object/group-level representations from scene-centric images. Experiments show the proposed approach effectively decomposes complex scenes into semantic groups for feature learning and significantly facilitates downstream tasks, including object detection, instance segmentation, and semantic segmentation. Acknowledgments This work has been supported by Hong Kong Research Grant Council - Early Career Scheme (Grant No. 27209621), HKU Startup Fund, and HKU Seed Fund for Basic Research. The authors acknowledge Smart More, Lunar AI, and MEGVII for partial computing support. [1] Federico Baldassarre and Hossein Azizpour. Towards Self-Supervised Learning of Global and Object Centric Representations. ar Xiv:2203.05997 [cs], March 2022. 3 [2] Hangbo Bao, Li Dong, and Furu Wei. BEi T: BERT Pre-Training of Image Transformers. ar Xiv:2106.08254 [cs], June 2021. 1 [3] Christopher P. Burgess, Loic Matthey, Nicholas Watters, Rishabh Kabra, Irina Higgins, Matt Botvinick, and Alexander Lerchner. MONet: Unsupervised Scene Decomposition and Representation. ar Xiv:1901.11390 [cs, stat], January 2019. 3 [4] Holger Caesar, Jasper Uijlings, and Vittorio Ferrari. COCO-Stuff: Thing and Stuff Classes in Context. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1209 1218, June 2018. 7, 9 [5] Mathilde Caron, Piotr Bojanowski, Armand Joulin, and Matthijs Douze. Deep Clustering for Unsupervised Learning of Visual Features. In Vittorio Ferrari, Martial Hebert, Cristian Sminchisescu, and Yair Weiss, editors, Computer Vision ECCV 2018, pages 139 156, Cham, 2018. Springer International Publishing. 3, 4 [6] Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. Unsupervised learning of visual features by contrasting cluster assignments. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 9912 9924. Curran Associates, Inc., 2020. 1, 3, 4, 7 [7] Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging Properties in Self-Supervised Vision Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9650 9660, 2021. 1, 4, 5, 6, 9 [8] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A Simple Framework for Contrastive Learning of Visual Representations. In Proceedings of the 37th International Conference on Machine Learning, pages 1597 1607. PMLR, November 2020. 3 [9] Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved Baselines with Momentum Contrastive Learning. ar Xiv:2003.04297 [cs], March 2020. 7, 8 [10] Xinlei Chen and Kaiming He. Exploring Simple Siamese Representation Learning. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 15745 15753, Nashville, TN, USA, June 2021. IEEE. 3 [11] Xinlei Chen, Saining Xie, and Kaiming He. An Empirical Study of Training Self-Supervised Vision Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 9640 9649, 2021. 1, 6 [12] MMSegmentation Contributors. MMSegmentation: Openmmlab semantic segmentation toolbox and benchmark. https://github.com/open-mmlab/mmsegmentation, 2020. 7 [13] Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. The Cityscapes Dataset for Semantic Urban Scene Understanding. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 3213 3223. IEEE Computer Society, June 2016. 1, 2, 7 [14] Quan Cui, Bingchen Zhao, Zhao-Min Chen, Borui Zhao, Renjie Song, Jiajun Liang, Boyan Zhou, and Osamu Yoshie. Discriminability-Transferability Trade-Off: An Information-Theoretic Perspective. ar Xiv:2203.03871 [cs], March 2022. 9 [15] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Image Net: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pages 248 255, June 2009. 1, 6 [16] Carl Doersch, Abhinav Gupta, and Alexei A. Efros. Unsupervised Visual Representation Learning by Context Prediction. In 2015 IEEE International Conference on Computer Vision (ICCV), pages 1422 1430, December 2015. 3 [17] Alexey Dosovitskiy, Philipp Fischer, Jost Tobias Springenberg, Martin Riedmiller, and Thomas Brox. Discriminative unsupervised feature learning with exemplar convolutional neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(9):1734 1747, 2015. 3 [18] Gamaleldin F Elsayed, Aravindh Mahendran, Sjoerd van Steenkiste, Klaus Greff, Michael C Mozer, and Thomas Kipf. Savi++: Towards end-to-end object-centric learning from real-world videos. ar Xiv preprint ar Xiv:2206.07764, 2022. 3 [19] Martin Engelcke, Adam R. Kosiorek, Oiwi Parker Jones, and Ingmar Posner. GENESIS: Generative Scene Inference and Sampling with Object-Centric Latent Representations. In International Conference on Learning Representations, September 2019. 3 [20] Mark Everingham, S. M. Ali Eslami, Luc Van Gool, Christopher K. I. Williams, John Winn, and Andrew Zisserman. The Pascal Visual Object Classes Challenge: A Retrospective. International Journal of Computer Vision, 111(1):98 136, January 2015. 2, 7 [21] Pedro F. Felzenszwalb and Daniel P. Huttenlocher. Efficient Graph-Based Image Segmentation. International Journal of Computer Vision, 59(2):167 181, September 2004. 7 [22] Spyros Gidaris, Praveer Singh, and Nikos Komodakis. Unsupervised Representation Learning by Predicting Image Rotations. In International Conference on Learning Representations, February 2018. 1, 3, 4 [23] Anirudh Goyal, Aniket Rajiv Didolkar, Nan Rosemary Ke, Charles Blundell, Philippe Beaudoin, Nicolas Heess, Michael Curtis Mozer, and Yoshua Bengio. Neural production systems. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, 2021. 3 [24] Anirudh Goyal, Alex Lamb, Phanideep Gampa, Philippe Beaudoin, Charles Blundell, Sergey Levine, Yoshua Bengio, and Michael Curtis Mozer. Factorizing declarative and procedural knowledge in structured, dynamical environments. In International Conference on Learning Representations, 2021. 3 [25] Anirudh Goyal, Alex Lamb, Jordan Hoffmann, Shagun Sodhani, Sergey Levine, Yoshua Bengio, and Bernhard Schölkopf. Recurrent independent mechanisms. In International Conference on Learning Representations, 2021. 3 [26] Klaus Greff, Raphaël Lopez Kaufman, Rishabh Kabra, Nick Watters, Christopher Burgess, Daniel Zoran, Loic Matthey, Matthew Botvinick, and Alexander Lerchner. Multi-Object Representation Learning with Iterative Variational Inference. In Proceedings of the 36th International Conference on Machine Learning, pages 2424 2433. PMLR, May 2019. 3 [27] Klaus Greff, Sjoerd van Steenkiste, and Jürgen Schmidhuber. On the Binding Problem in Artificial Neural Networks. ar Xiv:2012.05208 [cs], December 2020. 3 [28] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, Bilal Piot, koray kavukcuoglu, Remi Munos, and Michal Valko. Bootstrap your own latent - a new approach to selfsupervised learning. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 21271 21284. Curran Associates, Inc., 2020. 1, 3, 4, 5, 6 [29] Mark Hamilton, Zhoutong Zhang, Bharath Hariharan, Noah Snavely, and William T. Freeman. Unsupervised semantic segmentation by distilling feature correspondences. In International Conference on Learning Representations, 2022. 3 [30] Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked Autoencoders Are Scalable Vision Learners. ar Xiv:2111.06377 [cs], December 2021. 1 [31] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum Contrast for Unsupervised Visual Representation Learning. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9726 9735, June 2020. 1, 3, 7 [32] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask R-CNN. In 2017 IEEE International Conference on Computer Vision (ICCV), pages 2980 2988, October 2017. 4, 7 [33] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep Residual Learning for Image Recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770 778, June 2016. 6 [34] Olivier J. Hénaff, Skanda Koppula, Jean-Baptiste Alayrac, Aaron van den Oord, Oriol Vinyals, and João Carreira. Efficient Visual Pretraining With Contrastive Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 10086 10096, 2021. 2, 3, 5, 7, 8 [35] Olivier J. Hénaff, Skanda Koppula, Evan Shelhamer, Daniel Zoran, Andrew Jaegle, Andrew Zisserman, João Carreira, and Relja Arandjelovi c. Object discovery and representation networks. ar Xiv:2203.08777 [cs], March 2022. 2, 3, 5, 6, 9 [36] R. Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Phil Bachman, Adam Trischler, and Yoshua Bengio. Learning deep representations by mutual information estimation and maximization. In International Conference on Learning Representations, September 2018. 3 [37] Haiyang Huang, Zhi Chen, and Cynthia Rudin. Seg Discover: Visual Concept Discovery via Unsupervised Semantic Segmentation. ar Xiv:2204.10926 [cs], April 2022. 3, 7, 9 [38] Jyh-Jing Hwang, Stella X Yu, Jianbo Shi, Maxwell D Collins, Tien-Ju Yang, Xiao Zhang, and Liang-Chieh Chen. Segsort: Segmentation by discriminative sorting of segments. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7334 7344, 2019. 3 [39] Jang Hyun Cho, Utkarsh Mall, Kavita Bala, and Bharath Hariharan. Pi CIE: Unsupervised Semantic Segmentation using Invariance and Equivariance in Clustering. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 16789 16799, Nashville, TN, USA, June 2021. IEEE. 3, 4, 7, 9 [40] Xu Ji, João F Henriques, and Andrea Vedaldi. Invariant information clustering for unsupervised image classification and segmentation. In Proceedings of the IEEE International Conference on Computer Vision, pages 9865 9874, 2019. 3, 7, 9 [41] Thomas Kipf, Gamaleldin F. Elsayed, Aravindh Mahendran, Austin Stone, Sara Sabour, Georg Heigold, Rico Jonschkowski, Alexey Dosovitskiy, and Klaus Greff. Conditional Object-Centric Learning from Video. ar Xiv:2111.12594 [cs, stat], March 2022. 3 [42] Thomas Kipf, Elise van der Pol, and Max Welling. Contrastive Learning of Structured World Models. In International Conference on Learning Representations, September 2019. 3 [43] H. W. Kuhn. The Hungarian method for the assignment problem. Naval Research Logistics Quarterly, 2(1-2):83 97, 1955. 7, 9 [44] Chunyuan Li, Jianwei Yang, Pengchuan Zhang, Mei Gao, Bin Xiao, Xiyang Dai, Lu Yuan, and Jianfeng Gao. Efficient Self-supervised Vision Transformers for Representation Learning. ar Xiv:2106.09785 [cs], June 2021. 7 [45] Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature Pyramid Networks for Object Detection. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 936 944, July 2017. 7, 8 [46] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C. Lawrence Zitnick. Microsoft COCO: Common Objects in Context. In David Fleet, Tomas Pajdla, Bernt Schiele, and Tinne Tuytelaars, editors, Computer Vision ECCV 2014, pages 740 755, Cham, 2014. Springer International Publishing. 1, 2, 6, 7 [47] Songtao Liu, Zeming Li, and Jian Sun. Self-EMD: Self-Supervised Object Detection without Image Net. ar Xiv:2011.13677 [cs], March 2021. 3, 7 [48] Francesco Locatello, Dirk Weissenborn, Thomas Unterthiner, Aravindh Mahendran, Georg Heigold, Jakob Uszkoreit, Alexey Dosovitskiy, and Thomas Kipf. Object-centric learning with slot attention. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 11525 11538. Curran Associates, Inc., 2020. 3, 5 [49] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 3431 3440, June 2015. 7 [50] Ilya Loshchilov and Frank Hutter. SGDR: Stochastic Gradient Descent with Warm Restarts. ar Xiv:1608.03983 [cs, math], May 2017. 6 [51] Sindy Löwe, Klaus Greff, Rico Jonschkowski, Alexey Dosovitskiy, and Thomas Kipf. Learning Object Centric Video Models by Contrasting Sets. ar Xiv:2011.10287 [cs], November 2020. 3 [52] Mehdi Noroozi and Paolo Favaro. Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, Computer Vision ECCV 2016, pages 69 84, Cham, 2016. Springer International Publishing. 3 [53] Mehdi Noroozi, Hamed Pirsiavash, and Paolo Favaro. Representation Learning by Learning to Count. In 2017 IEEE International Conference on Computer Vision (ICCV), pages 5899 5907, October 2017. 3 [54] Pedro O O. Pinheiro, Amjad Almahairi, Ryan Benmalek, Florian Golemo, and Aaron C Courville. Unsupervised learning of dense visual representations. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 4489 4500. Curran Associates, Inc., 2020. 1, 3 [55] Deepak Pathak, Philipp Krähenbühl, Jeff Donahue, Trevor Darrell, and Alexei A. Efros. Context Encoders: Feature Learning by Inpainting. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2536 2544, June 2016. 3 [56] Ramprasaath R. Selvaraju, Karan Desai, Justin Johnson, and Nikhil Naik. CASTing Your Model: Learning to Localize Improves Self-Supervised Representations. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 11053 11062, June 2021. 2, 3 [57] Antti Tarvainen and Harri Valpola. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. 5 [58] Yonglong Tian, Chen Sun, Ben Poole, Dilip Krishnan, Cordelia Schmid, and Phillip Isola. What makes for good views for contrastive learning? In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 6827 6839. Curran Associates, Inc., 2020. 3 [59] J. R. R. Uijlings, K. E. A. van de Sande, T. Gevers, and A. W. M. Smeulders. Selective Search for Object Recognition. International Journal of Computer Vision, 104(2):154 171, September 2013. 7 [60] Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation Learning with Contrastive Predictive Coding. ar Xiv:1807.03748 [cs, stat], January 2019. 3, 5 [61] Wouter Van Gansbeke, Simon Vandenhende, Stamatios Georgoulis, and Luc V Gool. Revisiting contrastive methods for unsupervised learning of visual representations. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 16238 16250. Curran Associates, Inc., 2021. 1, 3, 7 [62] Wouter Van Gansbeke, Simon Vandenhende, Stamatios Georgoulis, and Luc Van Gool. Unsupervised Semantic Segmentation by Contrasting Object Mask Proposals. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 10052 10062, 2021. 2, 3, 5, 9 [63] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 17, pages 6000 6010, Red Hook, NY, USA, December 2017. Curran Associates Inc. 8 [64] Pascal Vincent, Hugo Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol. Extracting and composing robust features with denoising autoencoders. In Proceedings of the 25th International Conference on Machine Learning - ICML 08, pages 1096 1103, Helsinki, Finland, 2008. ACM Press. 3 [65] Xinlong Wang, Zhiding Yu, Shalini De Mello, Jan Kautz, Anima Anandkumar, Chunhua Shen, and Jose M Alvarez. Freesolo: Learning to segment objects without annotations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14176 14186, 2022. 3 [66] Xinlong Wang, Rufeng Zhang, Chunhua Shen, Tao Kong, and Lei Li. Dense Contrastive Learning for Self-Supervised Visual Pre-Training. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 3023 3032, Nashville, TN, USA, June 2021. IEEE. 2, 3, 6, 7, 8 [67] Chen Wei, Haoqi Fan, Saining Xie, Chao-Yuan Wu, Alan Yuille, and Christoph Feichtenhofer. Masked Feature Prediction for Self-Supervised Visual Pre-Training. ar Xiv:2112.09133 [cs], December 2021. 1 [68] Fangyun Wei, Yue Gao, Zhirong Wu, Han Hu, and Stephen Lin. Aligning Pretraining for Detection via Object-Level Contrastive Learning. In Advances in Neural Information Processing Systems, volume 34, pages 22682 22694. Curran Associates, Inc., 2021. 2, 3, 6, 8 [69] Yuxin Wu, Alexander Kirillov, Francisco Massa, Wan-Yen Lo, and Ross Girshick. Detectron2. https: //github.com/facebookresearch/detectron2, 2019. 7 [70] Zhirong Wu, Yuanjun Xiong, Stella X. Yu, and Dahua Lin. Unsupervised Feature Learning via Nonparametric Instance Discrimination. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3733 3742, June 2018. 1, 3 [71] Enze Xie, Jian Ding, Wenhai Wang, Xiaohang Zhan, Hang Xu, Peize Sun, Zhenguo Li, and Ping Luo. Det Co: Unsupervised Contrastive Learning for Object Detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 8392 8401, 2021. 3, 8 [72] Jiahao Xie, Xiaohang Zhan, Ziwei Liu, Yew Soon Ong, and Chen Change Loy. Unsupervised object-level representation learning from scene images. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 28864 28876. Curran Associates, Inc., 2021. 2, 3, 5, 6, 7, 8 [73] Zhenda Xie, Yutong Lin, Zheng Zhang, Yue Cao, Stephen Lin, and Han Hu. Propagate Yourself: Exploring Pixel-Level Consistency for Unsupervised Visual Representation Learning. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 16679 16688, Nashville, TN, USA, June 2021. IEEE. 2, 3, 6, 7, 8 [74] Ceyuan Yang, Zhirong Wu, Bolei Zhou, and Stephen Lin. Instance Localization for Self-supervised Detection Pretraining. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 3986 3995, June 2021. 3, 8 [75] Charig Yang, Hala Lamdouar, Erika Lu, Andrew Zisserman, and Weidi Xie. Self-Supervised Video Object Segmentation by Motion Grouping. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7177 7188, 2021. 3 [76] Asano Ym, Rupprecht C, and Vedaldi A. Self-labelling via simultaneous clustering and representation learning. In International Conference on Learning Representations, September 2019. 4 [77] Yang You, Igor Gitman, and Boris Ginsburg. Large Batch Training of Convolutional Networks. ar Xiv:1708.03888 [cs], September 2017. 6 [78] Jure Zbontar, Li Jing, Ishan Misra, Yann Le Cun, and Stephane Deny. Barlow Twins: Self-Supervised Learning via Redundancy Reduction. In Proceedings of the 38th International Conference on Machine Learning, pages 12310 12320. PMLR, July 2021. 3 [79] Richard Zhang, Phillip Isola, and Alexei A. Efros. Colorful Image Colorization. In Bastian Leibe, Jiri Matas, Nicu Sebe, and Max Welling, editors, Computer Vision ECCV 2016, pages 649 666, Cham, 2016. Springer International Publishing. 3 [80] Richard Zhang, Phillip Isola, and Alexei A. Efros. Split-Brain Autoencoders: Unsupervised Learning by Cross-Channel Prediction. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 645 654, July 2017. 3 [81] Xiao Zhang and Michael Maire. Self-supervised visual representation learning from hierarchical grouping. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 16579 16590. Curran Associates, Inc., 2020. 2, 3 [82] Bingchen Zhao and Xin Wen. Distilling Visual Priors from Self-Supervised Learning. In Adrien Bartoli and Andrea Fusiello, editors, Computer Vision ECCV 2020 Workshops, pages 422 429, Cham, 2020. Springer International Publishing. 3 [83] Bolei Zhou, Hang Zhao, Xavier Puig, Tete Xiao, Sanja Fidler, Adela Barriuso, and Antonio Torralba. Semantic Understanding of Scenes Through the ADE20K Dataset. International Journal of Computer Vision, 127(3):302 321, March 2019. 2, 7 [84] Rui Zhu, Bingchen Zhao, Jingen Liu, Zhenglong Sun, and Chang Wen Chen. Improving Contrastive Learning by Visualizing Feature Transformation. ar Xiv:2108.02982 [cs], August 2021. 3 [85] Adrian Ziegler and Yuki M. Asano. Self-Supervised Learning of Object Parts for Semantic Segmentation. ar Xiv:2204.13101 [cs], April 2022. 3 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] The limitations are discussed in Section G in the appendix. (c) Did you discuss any potential negative societal impacts of your work? [Yes] The potential negative societal impacts are discussed in Section G in the appendix. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] Link to the code is provided in the abstract. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [Yes] (c) Did you include any new assets either in the supplemental material or as a URL? [N/A] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]