# bootstrapping_semantic_segmentation_with_regional_contrast__2da899a0.pdf Published as a conference paper at ICLR 2022 BOOTSTRAPPING SEMANTIC SEGMENTATION WITH REGIONAL CONTRAST Shikun Liu1, Shuaifeng Zhi1, Edward Johns2, and Andrew J. Davison1 1Dyson Robotics Lab, Imperial College London 2Robot Learning Lab, Imperial College London shikun.liu17@imperial.ac.uk We present Re Co, a contrastive learning framework designed at a regional level to assist learning in semantic segmentation. Re Co performs pixel-level contrastive learning on a sparse set of hard negative pixels, with minimal additional memory footprint. Re Co is easy to implement, being built on top of off-the-shelf segmentation networks, and consistently improves performance, achieving more accurate segmentation boundaries and faster convergence. The strongest effect is in semisupervised learning with very few labels. With Re Co, we achieve high quality semantic segmentation model, requiring only 5 examples of each semantic class. 1 INTRODUCTION Semantic segmentation is an essential part of applications such as scene understanding and autonomous driving, whose goal is to assign a semantic label to each pixel in an image. Significant progress has been achieved by use of large datasets with high quality human annotations. However, labelling images with pixel-level accuracy is time consuming and expensive; for example, labelling a single image in City Scapes can take more than 90 minutes (Cordts et al., 2016). When deploying semantic segmentation models in practical applications where only limited labelled data are available, high quality ground-truth annotation is a significant bottleneck. To reduce the need for labelled data, there is a recent surge of interest in leveraging unlabelled data for semi-supervised learning. Previous methods include improving segmentation models via adversarial learning (Hung et al., 2019; Mittal et al., 2019) and self-training (Zou et al., 2019; 2018; Zhu et al., 2020). Others focus on designing advanced data augmentation strategies to generate pseudo image-annotation pairs from unlabelled images (Olsson et al., 2021; French et al., 2020). Bicycle Sheep Horse Query Class Sampling Distribution Person Sheep Horse Horse Person Bicycle Sheep Person Bicycle Pulling from Pushing to Class Mean Representation Representation Figure 1: Re Co pushes representations within a class closer to the class mean representation, whilst simultaneously pushing these representations away from negative representations sampled in different classes. The sampling distribution from negative classes is adaptive to each query class. In both semi-supervised and supervised learning, a segmentation model often predicts smooth label maps, because neighbouring pixels are usually of the same class, and rarer high-frequency regions are typically only found in object boundaries. This learning bias produces blurry contours and regularly mis-labels rare objects. After carefully examining the label predictions, we further observe that wrongly labelled pixels are typically confused with very few other classes; e.g. a pixel labelled as rider has a much higher chance of being wrongly classified as person, compared to train or bus. By understanding this class structure, learning can be actively focused on the challenging pixels to improve overall segmentation quality. Here we propose Re Co, a contrastive learning framework designed at a regional level. Specifically, Re Co is a new loss function which helps semantic segmentation not only to learn from local context Published as a conference paper at ICLR 2022 (neighbouring pixels), but also from global semantic class relationships across the entire dataset. Re Co performs contrastive learning on a pixel-level dense representation, as visualised in Fig. 1. For each semantic class in a mini-batch, Re Co samples a set of pixel-level representations (queries), and encourages them to be close to the class mean representation (positive keys), and simultaneously pushes them away from representations sampled from other classes (negative keys). For pixel-level contrastive learning with high-resolution images, it is impractical to sample all pixels. In Re Co, we actively sample a sparse set of queries and keys, consisting of less than 5% of all available pixels. We sample negative keys from a learned distribution based on the relative distance between the mean representation of each negative key and the query class. This distribution can be interpreted as a pairwise semantic class relationship, dynamically updated during training. We sample queries for those having a low prediction confidence. Active sampling helps Re Co to rapidly focus on the most confusing pixels for each semantic class, and requires minimal additional memory. Re Co enables a high-accuracy segmentation model to be trained with very few human annotations. We evaluate Re Co in a semi-supervised setting, with two different modes: i) Partial Dataset Full Labels a sparse subset of training images, where each image has full ground-truth labels, and the remaining images are unlabelled; ii) Partial Labels Full Dataset all images have some labels, but covering only a sparse subset of pixels within each image. In both settings, we show that Re Co can consistently improve performance across all methods and datasets. 2 RELATED WORK Semantic Segmentation One recent direction is in designing more effective deep convolutional neural networks. Fully convolutional networks (FCNs) (Long et al., 2015) are the foundation of modern segmentation network design. They were later improved with dilated/atrous convolutions with larger receptive fields, capturing more long range information (Chen et al., 2017; 2018). Alternative approaches include encoder-decoder architectures (Ronneberger et al., 2015; Kirillov et al., 2019), sometimes using skip connections (Ronneberger et al., 2015) to refine filtered details. A parallel direction is to improve optimisation strategies, by designing loss functions that better respect class imbalance (Lin et al., 2017) or using rendering strategy to refine uncertain pixels from high-frequency regions improving the label quality (Kirillov et al., 2020). Re Co is built upon this line of research, to improve segmentation by providing additional supervision on hard pixels. Semi-supervised Classification and Segmentation The goal of semi-supervised learning is to improve model performance by taking advantage of a large amount of unlabelled data during training. Here consistency regularisation and entropy minimisation are two common strategies. The intuition is that the network s output should be invariant to data perturbation and geometric transformation. Based on these strategies, many semi-supervised methods have been developed for image classification (Sohn et al., 2020; Tarvainen & Valpola, 2017; Berthelot et al., 2019; Kuo et al., 2020). However, for segmentation, generating effective pseudo-labels and well-designed data augmentation are non-trivial. Some solutions improved the quality of pseudo-labelling, using adversarial learning (Hung et al., 2019; Mittal et al., 2019) or enforcing consistency from different augmented images (French et al., 2020; Olsson et al., 2021). In this work, we show that we can improve the performance of current semi-supervised segmentation methods by jointly training with a suitable auxiliary task. Contrastive Learning Contrastive learning learns a similarity function to bring views of the same data closer in representation space, whilst pushing views of different data apart. Most recent contrastive frameworks learn similarity scores based on global representations of the views, parameterising data with a single vector (He et al., 2020; Chen et al., 2020; Khosla et al., 2020). Dense representations, on the other hand, rely on pixel-level representations and naturally provide additional supervision, capturing fine-grained pixel correspondence. Contrastive pre-training based on dense representations has recently been explored, and shows better performance in dense prediction tasks, such as object detection and keypoint detection (Wang et al., 2021b; O. Pinheiro et al., 2020). Contrastive Learning for Semantic Segmentation Contrastive learning has been recently studied to improve semantic segmentation, with a number of different design strategies. Zhang et al. (2021) and Zhao et al. (2021) both perform contrastive learning via pre-training, based on the generated auxiliary labels and ground-truth labels respectively, but at the cost of huge memory consump- Published as a conference paper at ICLR 2022 tion. In contrast, ours performs contrastive learning whilst requiring much less memory, via active sampling. In concurrent work, (Wang et al., 2021a; Alonso et al., 2021) also perform contrastive learning with active sampling. However, whilst both these methods are applied to a stored feature bank, ours focuses on sampling features on-the-fly. Active sampling in Alonso et al. (2021) is further based on learnable, class-specific attention modules, whilst ours only samples features based on relation graphs and prediction confidence, without introducing any additional computation overhead, which results in a simpler and much more memory-efficient implementation. 3 RECO REGIONAL CONTRAST 3.1 PIXEL-LEVEL CONTRASTIVE LEARNING Let (X, Y ) be a training dataset with training images x X and their corresponding C-class pixellevel segmentation labels y Y , where y can be either provided in the original dataset, or generated automatically as pseudo-labels. A segmentation network f is then optimised to learn a mapping fθ : X 7 Y , parameterised by network parameters θ. This segmentation network f can be decomposed into two parts: an encoder network: φ : X 7 Z, and a decoder classification head ψc : Z 7 Y . To perform pixel-level contrastive learning, we additionally attach a decoder representation head ψr on top of the encoder network φ, parallel to the classification head, mapping the encoded feature into a higher m-dimensional dense representation with the same spatial resolution as the input image: ψr : Z 7 R, R Rm. This representation head is only applied during training to guide the classifier using the Re Co loss as an auxiliary task, and is removed during inference. A pixel-level contrastive loss is a function which encourages queries rq to be similar to the positive key r+ k , and dissimilar to the negative keys r k . All queries and keys are sampled from the decoder representation head: rq, r+, k R. In Re Co, we use a pixel-level contrastive loss across all available semantic classes in each mini-batch, with the distance between keys and queries measured by their normalised dot product. The general formation of the Re Co loss Lreco is then defined as: rq Rcq log exp(rq rc,+ k /τ) exp(rq rc,+ k /τ) + P r k Rc k exp(rq r k /τ) , (1) for which C is a set containing all available classes in the current mini-batch, τ is the temperature control of the softness of the distribution, Rc q represents a query set containing all representations whose labels belong to class c, Rc k represents a negative key set containing all representations whose labels do not belong to class c, and rc,+ k represents the positive key which is the mean representation of class c. Suppose P is a set containing all pixel coordinates with the same resolution as R, these queries and keys are then defined as: [u,v] P 1(y[u,v] = c)r[u,v], Rc k = [ [u,v] P 1(y[u,v] = c)r[u,v], rc,+ k = 1 |Rcq| rq Rcq rq. (2) 3.2 ACTIVE HARD SAMPLING ON QUERIES AND KEYS Contrastive learning on all pixels in high-resolution images would be computationally expensive. Here, we introduce active hard sampling strategies to optimise only a sparse set of queries and keys. Active Key Sampling When classifying a pixel, a semantic network might be uncertain only over a very small number of candidates, among all available classes. The uncertainty from these candidates typically comes from a close spatial (e.g. rider and bicycle) or semantic (e.g. horse and cow) relationship. To reduce this uncertainty, we propose to sample negative keys non-uniformly, based on the relative distance between each negative key class and the query class. This involves building a pair-wise class relationship graph G, with G R|C| |C|, computed and dynamically updated for each mini-batch. This pair-wise relationship is measured by the normalised dot product between the mean representation from a pair of two classes and is defined as: G[p, q] = rp,+ k rq,+ k , p, q C, and p = q. (3) We further apply Soft Max to normalise these pair-wise relationships among all negative classes j for each query class c, which produces a probabilistic distribution: Published as a conference paper at ICLR 2022 exp(G[c, i])/ P j C,j =c exp(G[c, j]). We sample negative keys for each class i based on this distribution, to learn the corresponding query class c. This procedure allocates more samples to hard, confusing classes chosen specifically for each query class, helping the segmentation network to learn a more accurate decision boundary. Active Query Sampling Due to the natural class imbalance in semantic segmentation, it is easy to over-fit on common classes, such as the road and building classes in the City Scapes dataset, or the background class in the Pascal VOC dataset. These common classes contribute to the majority of pixel space in training images, and so randomly sampling queries will under-sample rare classes and provide minimal supervision to these classes. Therefore, we instead sample hard queries for those whose corresponding pixel prediction confidence is below a defined threshold. Accordingly, Re Co s loss would then guide the segmentation network by providing appropriate supervision on these less certain pixels. The easy and hard queries are defined as follows, and visualised in Fig. 2, Rc, easy q = [ rq Rcq 1(ˆyq > δs)rq, Rc, hard q = [ rq Rcq 1(ˆyq δs)rq, (4) where ˆyq is the predicted confidence of label c after the Soft Max operation corresponding to the same pixel location as rq, and δs is the user-defined confidence threshold. (a) Confidence Map (b) Easy Queries (c) Hard Queries Figure 2: Easy and hard queries (shown in white) determined from the predicted confidence map in the Cityscapes dataset. Here we set the confidence threshold δs = 0.97. 3.3 SEMI-SUPERVISED SEMANTIC SEGMENTATION WITH RECO Re Co can easily be added to modern semi-supervised segmentation methods without changing the training pipeline, with no additional cost at inference time. To incorporate Re Co, we simply add an additional representation head ψr as described in Section 3.1, and apply the Re Co loss (in Eq. 1) to this representation using the sampling strategy introduced in Section 3.2. We apply the Mean Teacher framework (Tarvainen & Valpola, 2017) following prior state-of-the-art semi-supervised segmentation methods (Olsson et al., 2021; Mittal et al., 2019). Instead of using the original segmentation network fθ (which we call the student model), we instead use fθ (which we call the teacher model) to generate pseudo-labels from unlabelled images, where θ is a moving average of the previous state of θ during training optimisation: θ t = λθ t 1 +(1 λ)θt, with a decay parameter λ = 0.99. This teacher model can be treated as a temporal ensemble of student models across training time t, resulting in more stable predictions for unlabelled images. The student model fθ is then used to train on the augmented unlabelled images, with pseudo-labels as the ground-truths. For all pixels with defined ground-truth labels, we apply the Re Co loss on dense representations corresponding to all valid pixels. For all pixels without such labels, we only sample pixels whose predicted pseudo-label confidence is greater than a threshold δw. This avoids sampling pixels which are likely to have incorrect pseudo-labels. We apply the Re Co loss to a combined set of labelled and unlabelled pixels. The overall training loss for semi-supervised segmentation is then the linear combination of supervised cross-entropy loss (on ground-truth labels), unsupervised cross-entropy loss (on pseudo-labels) and Re Co loss: Ltotal = Lsupervised + η Lunsupervised + Lreco, (5) where η is defined as the percentage of pixels whose predicted confidence are greater than δs, a scalar re-weighting the contribution for unsupervised loss, following prior methods (Olsson et al., 2021; Published as a conference paper at ICLR 2022 Labelled Images Augmented Images Student Network Predicted Features Ground-truth Labels Predicted Labels Supervised Loss Unlabelled Images Augmented Images Student Network Predicted Features Augmented Labels Predicted Labels Unsupervised Loss Generated Labels Teacher Network Feature Space Masked Features Figure 3: Visualisation of the Re Co framework applied to semi-supervised segmentation and trained with three losses. A supervised loss is computed based on labelled data with ground-truth annotations. An unsupervised loss is computed for unlabelled data with generated pseudo-labels. And finally a Re Co loss is computed based on pixel-level dense representation predicted from both labelled and unlabelled images. Mittal et al., 2019). This makes sure the segmentation network would not be dominated by gradients produced by uncertain pseudo-labels, which typically occur during the early stage of training. Fig. 3 shows a visualisation of the Re Co framework for semi-supervised segmentation. 4 EXPERIMENTS Semi-Supervised Segmentation Benchmark Redesign We propose two modes of semisupervised segmentation tasks, aiming at two different applications. i) Partial Dataset Full Labels: A small subset of the images is trained with complete ground-truth labels for each image, whilst the remaining training images are unlabelled. This is the de-facto standard of evaluating semi-supervised segmentation in prior works. ii) Partial Labels Full Dataset: All images are trained with partial labels, but only a small percentage of labels are provided for each class in each training image. We create the dataset by first randomly sampling a pixel for each class, and then continuously apply a [5 5] square kernel for dilation until we meet the percentage criteria. 1 Pixel 1% Labels 5% Labels 25% Labels Background Boat Dog Person Undefined Figure 4: Example of training labels for Pascal VOC dataset in Partial Labels Full Dataset setting. (1 Pixel is zoomed 5 times for better visualisation.) The Partial Dataset Full Label setting evaluates the ability to generalise semantic classes given a few examples with perfect boundary information. The Partial Label Full Dataset evaluates learning semantic class completion given many examples with no or minimal boundary information. Datasets We experiment on segmentation datasets: Cityscapes (Cordts et al., 2016) and Pascal VOC 2012 (Everingham et al., 2015) in both partial and full label setting. We also evaluate on a more difficult indoor scene dataset SUN RGB-D (Song et al., 2015) in the full label setting only, mainly due to the low quality annotations making it difficult for fair evaluation in the partial label setting. An example of the partially labelled Pascal VOC is shown in Fig. 4. Strong Baselines Prior semi-supervised segmentation methods are typically designed with different backbone architectures, and trained with different strategies, which makes it difficult to compare them fairly. In this work, we standardise the baselines and implement four strong semi-supervised segmentation methods ourselves: S4GAN (Mittal et al., 2019): an adversarial learning based semisupervised method; Cut Out (French et al., 2020), Cut Mix (French et al., 2020), Class Mix (Olsson Published as a conference paper at ICLR 2022 et al., 2021): three image augmentation strategies designed specifically for semi-supervised segmentation. Our implementations for all baselines obtain performance on par with, and most of the time surpassing, the performance reported in each original publication, giving us a set of strong baselines. All baselines and our method were implemented on Deep Lab V3+ (Chen et al., 2018) with Res Net-101 backbone (He et al., 2016), and all with the same optimisation strategies. Detailed hyper-parameters used for each dataset are provided in the Appendix A. 4.1 RESULTS ON PASCAL VOC, CITYSCAPES, SUN RGB-D (FULL LABELS) First, we compared our results to semi-supervised baselines in a full label setting. We applied Re Co on top of Class Mix, which consistently outperformed other semi-supervised baselines. Table 1 shows the mean Io U validation performance on three datasets over three individual runs (different labelled and unlabelled data splits). The number of labelled images shown in the three columns for each dataset, are chosen such that the least-appeared classes have appeared in 5, 15 and 50 images respectively. In the fewest-label setting for each dataset, applying Re Co with Class Mix can improve results by a significant margin, with up to 1.5 4.5% absolute improvement. Pascal VOC City Scapes SUN RGB-D Method 60 labels 200 labels 600 labels 20 labels 50 labels 150 labels 50 labels 150 labels 500 labels Supervised 37.79 53.87 64.04 38.12 45.42 54.93 19.79 28.78 37.73 S4GAN (Mittal et al., 2019) 47.95 61.25 66.21 37.65 47.08 56.46 20.53 29.79 38.08 Cut Out (French et al., 2020) 52.96 63.57 69.85 42.52 50.15 59.42 25.94 34.45 41.25 Cut Mix (French et al., 2020) 53.71 66.95 72.42 44.02 54.72 62.24 27.60 37.55 42.69 Class Mix (Olsson et al., 2021) 49.06 67.95 72.50 45.61 55.56 63.94 28.42 37.55 42.46 Re Co + Class Mix 53.31 69.81 72.75 49.86 57.69 65.04 29.65 39.14 44.55 Table 1: mean Io U validation performance for Pascal VOC, City Scapes and SUN RGB-D datasets. We report the mean over three independent runs for all methods. Pascal VOC 1/16 [92] 1/8 [183] 1/4 [366] 1/2 [732] Adv Sem Seg (Hung et al., 2019) 39.69 47.58 59.97 65.27 Mean Teacher (Tarvainen & Valpola, 2017) 48.70 55.81 63.01 69.16 CCT (Ouali et al., 2020) 33.10 47.60 58.80 62.10 GCT (Ke et al., 2020) 46.04 54.98 64.71 70.67 VAT (Miyato et al., 2018) 36.92 49.35 56.88 63.34 Cut Mix (French et al., 2020) 55.58 63.20 68.36 69.87 Pseudo Seg (Zou et al., 2021) 57.60 65.50 69.14 72.41 Re Co + Class Mix 64.78 72.02 73.14 74.69 Table 2: mean Io U validation performance for Pascal VOC with data partition and training strategy proposed in Pseudo Seg (Zou et al., 2021). The percentage and the number of labelled data used are listed in the first row. To further justify the effectiveness of Re Co, we also include results on existing benchmarks, to compare with other semi-supervised methods in Table 2. Here, all baselines were re-implemented and reported in the Pseudo Seg setting (Zou et al., 2021), where the labelled images are sampled from the original PASCAL dataset, with a total of 1.4k images. In both benchmarks, Re Co shows state-of-the-art performance, and specifically is able to reach Pseudo Seg s performance, whilst requiring only half the labelled data. Additional results are further shown in Appendix C. In Fig. 5, we present qualitative results from the semi-supervised setup with the fewest labels: 20 labels for City Scapes and 50 labels SUN RGB-D datasets. The 60 labelled Pascal VOC is further shown in Appendix D. In Fig. 5, we can see the advantage of Re Co, where the edges and boundaries of small objects are clearly more pronounced such as in the person and bicycle classes in City Scapes, and the lamp and pillow classes in SUN RGB-D. Interestingly, we found that in SUN RGB-D, though all methods may confuse ambiguous class pairs such as table and desk or window and curtain, Re Co still produces consistently sharp and accurate object boundaries compared to the Supervised and Class Mix baselines where labels are noisy near object boundaries. 4.2 RESULTS ON PASCAL VOC AND CITYSCAPES (PARTIAL LABELS) In the partial label setting, we evaluated on the City Scapes and Pascal VOC datasets. Table 3 compared Re Co to the two best semi-supervised baselines and a supervised baseline. Again, we see Re Co can improve performance in all cases when applied on top of Class Mix, with around 1 3% Published as a conference paper at ICLR 2022 Ground Truth Supervised (20 Labels) Class Mix (20 Labels) Re Co + Class Mix (20 Labels) Road Sidewalk Building Wall Fence Pole Traffic Light Traffic Sign Vegetation Terrain Sky Person Rider Car Truck Bus Train Motorcycle Bicycle Undefined Ground Truth Supervised (50 Labels) Class Mix (50 Labels) Re Co + Class Mix (50 Labels) Wall Floor Cabinet Bed Chair Sofa Table Door Window Bookshelf Picture Counter Blinds Desk Shelves Curtain Dresser Pillow Mirror Floor Mat Clothes Ceiling Books Fridge TV Paper Towel Bath Curtain Box Whiteboard Person Nightstand Toilet Sink Lamp Bathtub Bag Undefined Figure 5: Visualisation of City Scapes and SUN RGB-D validation set trained on 20 and 50 labelled images respectively. Interesting regions are shown in white arrows. Pascal VOC (Partial) Method 1 pixel 1% labels 5% labels 25% labels Supervised 60.33 66.17 69.16 73.75 Cut Mix 63.50 70.83 73.04 75.64 Class Mix 63.69 71.04 72.90 75.79 Re Co + Class Mix 66.11 72.67 74.09 75.96 City Scapes (Partial) Method 1 pixel 1% labels 5% labels 25% labels Supervised 44.08 52.89 56.65 63.43 Cut Mix 46.91 54.90 59.69 65.61 Class Mix 47.42 56.68 60.96 66.46 Re Co + Class Mix 49.66 58.97 62.32 66.92 Table 3: mean Io U validation performance for Pascal VOC and Cityscapes datasets trained on 1, 1%, 5% and 25% labelled pixels per class per image. We report the mean over three independent runs for all methods. absolute improvement. We observe less relative performance improvement than in the full label setting; very sparse ground-truth annotations could confuse Re Co, resulting in inaccurate supervision. We show qualitative results on Pascal VOC dataset trained on 1 labelled pixel per class per image in Fig. 6. As in the full label setting, we see smoother and more accurate boundary predictions from Re Co. More visualisations from City Scapes are shown in Appendix E. 4.3 ABLATIVE ANALYSIS Next we present an ablative analysis on 20 labelled City Scapes images to understand the behaviour of Re Co with respect to hyper-parameters. We use our default experimental setting from Section 4.1, using Re Co with Class Mix. Additional ablations are further shown in Appendix B. Number of Queries and Keys We first evaluate the performance by varying the number of queries and keys used in Re Co framework, whilst fixing all other hyper-parameters. In Fig. 7a and 7b, we can observe that performance is better when sampling more queries and keys, but after a certain Published as a conference paper at ICLR 2022 Background Aeroplane Bicycle Bird Boat Bottle Bus Car Cat Chair Cow Table Dog Horse Motorbike Person Plant Sheep Sofa Train Monitor Undefined Figure 6: Visualisation of Pascal VOC validation set with Class Mix (left) vs. with Re Co (right) trained on 1 labelled pixel per class per image. Interesting regions are shown in white arrows. point, the improvements become marginal. Notably, even in our smallest option of having 32 queries per class in a mini-batch consisting of less than 0.5% among all available pixel space this can still improve performance by a non-trivial margin. Compared to a concurrent work (Zhang et al., 2021) which requires 10k queries and 40k keys in each training iteration, Re Co can be optimised with 50 more efficiency in terms of memory footprint. Ratio of Unlabelled Data We evaluate how Re Co can generalise across different levels of unlabelled data. In Fig. 7c, we show that by only training on 10% unlabelled data in the original setting, we can already surpass the Class Mix baseline. This shows that Re Co can achieve strong generalisation not only in label efficiency but also in data efficiency. Choice of Semi-Supervised Method Finally, we show that Re Co is robust to the choice of different semi-supervised methods. In Fig. 7d, we can see that Re Co obtains a better performance from a variety choice of semi-supervised baselines. Effect of Active Sampling In Fig. 7e, we see that randomly sampling queries and keys gives much less improvement compared to active sampling in our default setting. Particularly, hard query sampling has a dominant effect on generalisation: if we instead only sample from easy queries, Re Co only marginally improves on the baseline. This further verifies that most queries are redundant. 32 64 128 256 44 49.55 49.83 (a) Number of Queries 64 128 256 512 44 (b) Number of Keys 2% 5% 10% 100% 40 (c) Ratio of Unlabelled Data Cut Out Cut Mix Class Mix 40 (d) Semi-Supervised Method Random Query Random Key Active Query Random Key Easy Query Active Key 46.56 46.38 (e) Effect of Sampling Figure 7: mean Io U validation performance on 20 labelled City Scapes dataset based on different choices of hyper-parameters. Grey: Class Mix (if not labelled otherwise) in our default setting. Light Blue: Re Co + Class Mix (if not labelled otherwise) in a different hyper-parameter setting. Dark Blue: Re Co + Class Mix in our default setting. 5 VISUALISATIONS AND INTERPRETABILITY OF CLASS RELATIONSHIPS In this section, we visualise the pair-wise semantic class relation graph defined in Eq. 3, additionally supported by a semantic class dendrogram using the off-the-shelf hierarchical clustering algorithm in Sci Py (Virtanen et al., 2020) for better visualisation. The features for each semantic class used in both visualisations are averaged across all available pixel embeddings in each class from the validation set. In all visualisations, we compared features learned with Re Co built on top of supervised Published as a conference paper at ICLR 2022 learning compared to a standard supervised learning method trained on all data, representing the semantic class relationships of the full dataset. Using the same definitions in Section 3.1, we first choose such pixel embedding to be the embedding Z predicted from the encoder network φ in both supervised learning and with Re Co. We also show the visualisation for embedding R which is the actual representation we used for Re Co loss and active sampling. pole traffic light traffic sign train motorcycle road sidewalk building pole traffic light traffic sign vegetation bus train motorcycle traffic light traffic sign (a) Embed. Z (Supervised) pole traffic light traffic sign train motorcycle road sidewalk building pole traffic light traffic sign vegetation bus train motorcycle traffic light traffic sign (b) Embed. Z (Re Co + Supervised) pole traffic light traffic sign train motorcycle road sidewalk building pole traffic light traffic sign vegetation bus train motorcycle sky traffic light traffic sign (c) Embed. R (Re Co + Supervised) Figure 8: Visualisation of semantic class relation graph (top) and its corresponding semantic class dendrogram (bottom) on City Scapes dataset. Brighter colour represents closer (more confused) relationship. Best viewed in zoom. In Fig. 8, we present the semantic class relationship and dendrogram for the City Scapes dataset by embedding R and Z with and without Re Co. We can clearly see that Re Co helps disentangle features compared to supervised learning where many pairs of semantic classes are similar. In addition, we find that the dendrogram generated by Re Co based on embedding Z is more structured, showing a clear and interpretable semantic tree by grouping semantically similar classes together: for example, all large transportation classes car, truck, bus and train are under the same parent branch. In addition, we find that nearly all classes based on embedding R are perfectly disentangled, except for bus and train, suggesting the City Scapes dataset might not have sufficient bus and train examples to learn a distinctive representation for these two classes. fridge whiteboard dresser night stand curtain shower curtain floor floor mat 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Figure 9: Visualisation of semantic class dendrogram based on embedding R on SUN RGB-D dataset using Re Co + Supervised method. Best viewed in zoom. The pair-wise relation graph helps us to understand the distribution of semantic classes in each dataset, and clarifies the pattern of incorrect predictions from the trained semantic network. We additionally provide a dendrogram based on embedding R for the SUN RGBD dataset, clearly showing ambiguous class pairs, such as night stand and dresser; table and desk; floor and floormat, consistent with our results shown in Fig. 5. Complete visualisations of these semantic class relationships are shown in Appendix F. 6 CONCLUSION In this work, we have presented Re Co, a new pixel-level contrastive framework with active sampling, designed specifically for semantic segmentation. Re Co can improve performance in semantic segmentation methods with minimal additional memory footprint. In particular, Re Co has shown its strongest effect in semi-supervised learning with very few labels, where we improved on the stateof-the-art by a large margin. In further work, we aim to design effective contrastive frameworks for video representation learning. Published as a conference paper at ICLR 2022 REPRODUCIBILITY All of the information for reproducibility is shown in Appendix A. Code is available at https: //github.com/lorenmt/reco. ACKNOWLEDGEMENT This work has been supported by Dyson Technology Ltd. We thank Zhe Lin for the initial discussion and Zhengyang Feng for his help on the evaluation metric design. I nigo Alonso, Alberto Sabater, David Ferstl, Luis Montesano, and Ana C Murillo. Semi-supervised semantic segmentation with pixel-level contrastive learning from a class-wise memory bank. In Proceedings of the International Conference on Computer Vision (ICCV), 2021. David Berthelot, Nicholas Carlini, Ian Goodfellow, Nicolas Papernot, Avital Oliver, and Colin A Raffel. Mixmatch: A holistic approach to semi-supervised learning. In Advances in Neural Information Processing Systems (Neur IPS), 2019. Liang-Chieh Chen, George Papandreou, Iasonas Kokkinos, Kevin Murphy, and Alan L Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 2017. Liang-Chieh Chen, Yukun Zhu, George Papandreou, Florian Schroff, and Hartwig Adam. Encoderdecoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), 2018. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning (ICML), 2020. Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. Mark Everingham, SM Ali Eslami, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. The pascal visual object classes challenge: A retrospective. International Journal of Computer Vision (IJCV), 2015. Zhengyang Feng, Qianyu Zhou, Qiqi Gu, Xin Tan, Guangliang Cheng, Xuequan Lu, Jianping Shi, and Lizhuang Ma. Dmt: Dynamic mutual training for semi-supervised learning. ar Xiv preprint ar Xiv:2004.08514, 2020. Geoff French, Timo Aila, Samuli Laine, Michal Mackiewicz, and Graham Finlayson. Semisupervised semantic segmentation needs strong, high-dimensional perturbations. In Proceedings of the British Machine Vision Conference (BMVC), 2020. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. Wei Chih Hung, Yi Hsuan Tsai, Yan Ting Liou, Yen Yu Lin, and Ming Hsuan Yang. Adversarial learning for semi-supervised semantic segmentation. In Proceedings of the British Machine Vision Conference (BMVC), 2019. Published as a conference paper at ICLR 2022 Zhanghan Ke, Di Qiu, Kaican Li, Qiong Yan, and Rynson W.H. Lau. Guided collaborative training for pixel-wise semi-supervised learning. In Proceedings of the European Conference on Computer Vision (ECCV), 2020. Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. Supervised contrastive learning. Advances in Neural Information Processing Systems (Neur IPS), 2020. Alexander Kirillov, Ross Girshick, Kaiming He, and Piotr Doll ar. Panoptic feature pyramid networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Alexander Kirillov, Yuxin Wu, Kaiming He, and Ross Girshick. Pointrend: Image segmentation as rendering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. Chia-Wen Kuo, Chih-Yao Ma, Jia-Bin Huang, and Zsolt Kira. Featmatch: Feature-based augmentation for semi-supervised learning. In Proceedings of the European Conference on Computer Vision (ECCV), 2020. Xin Lai, Zhuotao Tian, Li Jiang, Shu Liu, Hengshuang Zhao, Liwei Wang, and Jiaya Jia. Semisupervised semantic segmentation with directional context-aware consistency. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021. Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Doll ar. Focal loss for dense object detection. In Proceedings of the International Conference on Computer Vision (ICCV), 2017. Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015. Robert Mendel, Luis Antonio de Souza, David Rauber, Jo ao Paulo Papa, and Christoph Palm. Semisupervised segmentation based on error-correcting supervision. In Proceedings of the European Conference on Computer Vision (ECCV), 2020. Sudhanshu Mittal, Maxim Tatarchenko, and Thomas Brox. Semi-supervised semantic segmentation with high-and low-level consistency. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 2019. Takeru Miyato, Shin-ichi Maeda, Masanori Koyama, and Shin Ishii. Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 2018. Pedro O O. Pinheiro, Amjad Almahairi, Ryan Benmalek, Florian Golemo, and Aaron C Courville. Unsupervised learning of dense visual representations. In Advances in Neural Information Processing Systems (Neur IPS), 2020. Viktor Olsson, Wilhelm Tranheden, Juliano Pinto, and Lennart Svensson. Classmix: Segmentationbased data augmentation for semi-supervised learning. In IEEE Winter Conference on Applications of Computer Vision (WACV), 2021. Yassine Ouali, C eline Hudelot, and Myriam Tami. Semi-supervised semantic segmentation with cross-consistency training. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020. Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), 2015. Kihyuk Sohn, David Berthelot, Nicholas Carlini, Zizhao Zhang, Han Zhang, Colin A Raffel, Ekin Dogus Cubuk, Alexey Kurakin, and Chun-Liang Li. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In Advances in Neural Information Processing Systems (Neur IPS), 2020. Published as a conference paper at ICLR 2022 Shuran Song, Samuel P Lichtenberg, and Jianxiong Xiao. Sun rgb-d: A rgb-d scene understanding benchmark suite. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015. Antti Tarvainen and Harri Valpola. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Advances in Neural Information Processing Systems (Neur IPS), 2017. Pauli Virtanen, Ralf Gommers, Travis E Oliphant, Matt Haberland, Tyler Reddy, David Cournapeau, Evgeni Burovski, Pearu Peterson, Warren Weckesser, Jonathan Bright, et al. Scipy 1.0: fundamental algorithms for scientific computing in python. Nature methods, 2020. Wenguan Wang, Tianfei Zhou, Fisher Yu, Jifeng Dai, Ender Konukoglu, and Luc Van Gool. Exploring cross-image pixel contrast for semantic segmentation. In Proceedings of the International Conference on Computer Vision (ICCV), 2021a. Xinlong Wang, Rufeng Zhang, Chunhua Shen, Tao Kong, and Lei Li. Dense contrastive learning for self-supervised visual pre-training. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2021b. Feihu Zhang, Philip Torr, Rene Ranftl, and Stephan R Richter. Looking beyond single images for contrastive semantic segmentation learning. In Advances in Neural Information Processing Systems (Neur IPS), 2021. Xiangyun Zhao, Raviteja Vemulapalli, Philip Andrew Mansfield, Boqing Gong, Bradley Green, Lior Shapira, and Ying Wu. Contrastive learning for label efficient semantic segmentation. In Proceedings of the International Conference on Computer Vision (ICCV), 2021. Yi Zhu, Zhongyue Zhang, Chongruo Wu, Zhi Zhang, Tong He, Hang Zhang, R Manmatha, Mu Li, and Alexander Smola. Improving semantic segmentation via self-training. ar Xiv preprint ar Xiv:2004.14960, 2020. Yang Zou, Zhiding Yu, BVK Kumar, and Jinsong Wang. Domain adaptation for semantic segmentation via class-balanced self-training. ar Xiv preprint ar Xiv:1810.07911, 2018. Yang Zou, Zhiding Yu, Xiaofeng Liu, BVK Kumar, and Jinsong Wang. Confidence regularized self-training. In Proceedings of the International Conference on Computer Vision (ICCV), 2019. Yuliang Zou, Zizhao Zhang, Han Zhang, Chun-Liang Li, Xiao Bian, Jia-Bin Huang, and Tomas Pfister. Pseudoseg: Designing pseudo labels for semantic segmentation. In Proceedings of the International Conference on Learning Representations (ICLR), 2021. Published as a conference paper at ICLR 2022 A IMPLEMENTATION DETAILS We trained all methods with SGD optimiser with learning rate 2.5 10 3, momentum 0.9, and weight decay 5 10 4. We adopted the polynomial annealing policy to schedule the learning rate, which is multiplied by (1 iter total iter)power with power = 0.9, and trained for 40k iterations for all datasets. Code is attached in the supplementary material. For City Scapes, we first downsampled all images in the dataset to half resolution [512 1024] prior to use. We extracted [512 512] random crops and used a batch size of 2 during training. For Pascal VOC, we extracted [321 321] random crops, applied a random scale between [0.5, 1.5], and used a batch size of 10 during training. For SUN RGB-D, we first rescaled all images to [384 512] resolution, extracted [321 321] random crops, applied a random scale between [0.5, 1.5], and used a batch size of 5. We additionally re-organised the original training and validation split in SUN RGB-D dataset from 5285 and 5050 to 9860 and 475 samples respectively, to increase the amount of training data which we think is more appropriate for semi-supervised task. All datasets were additionally augmented with Gaussian blur, colour jittering, and random horizontal flip. The pre-processing for City Scapes and Pascal VOC are consistent with the prior work (Olsson et al., 2021). In Table 2, we extracted [513 513] random crops and applied a random scale between [0.5, 2.0], following Pseudo Seg s training setup (Zou et al., 2021). In our Re Co framework, we sampled 256 query samples and 512 key samples and used temperature τ = 0.5 for each mini-batch, which we found to work well in all datasets. The dimensionality for pixel-level representation was set to m = 256. The confidence thresholds were set to δw = 0.7 and δs = 0.97. B ADDITIONAL ABLATIVE STUDIES City Scapes 20 Labels 50 Labels Supervised 38.10 47.10 Semi-Supervised 28.59 43.74 Semi-Supervised + Re Co 29.16 46.96 Class Mix 45.61 55.56 Class Mix + Reco 49.86 57.69 Table 4: mean Io U validation performance for 20 and 50 labelled City Scapes data for supervised method (top) and semi-supervised methods (bottom). Re Co Only Results Table 4 shows Re Co designed with and without data augmentation, trained on 20 and 50 labelled City Scapes dataset. We observe that using pure semisupervised learning with additional unlabelled data will lead to worse performance compared to supervised learning without such labelled data. This shows data augmentation strategies designed for semi-supervised segmentation are the key component to make best use of the unlabelled data. Although the vanilla Re Co still performs better compared to standard semi-supervised learning, the active sampling of Re Co based on incorrect pseudo-labels leads to marginal improvement compared to a pure data augmentation method like Class Mix. Therefore, Re Co performs better as an auxiliary framework combined with a strong semi-supervised method. Compared to Feature Bank Methods We have experimented with Re Co with a stored feature bank framework similar to the design in the concurrent works (Alonso et al., 2021; Wang et al., 2021a). We found that just by replacing our batch-wise sampling method with a feature bank sampling method will achieve a similar performance (49.34 m Io U) compared to our original design (49.86 m Io U) on 20 labelled City Scapes, but with a slower training speed. This verifies our assumption that batch-wise sampling is an accurate approximation of class distribution over the entire dataset. Published as a conference paper at ICLR 2022 C RESULTS ON SEMI-SUPERVISED SEGMENTATION BENCHMARKS Here, we present quantitative results for other semi-supervised semantic segmentation benchmarks in City Scapes and Pascal VOC datasets. Note that, this benchmark is much less challenging compared to our proposed benchmark in Section 4.1, evaluated with significantly less number of labelled images. Since some methods applied with different backbones and training strategies, we compared each result with respect to its performance gap compared to its corresponding fully supervised result, as shown in brackets, to ensure fairness following Feng et al. (2020). In Table 5, we show results for Re Co applied on top of Class Mix, and trained with both Deep Labv2 (Chen et al., 2017) and Deep Labv3+ (Chen et al., 2018). We can observe that Re Co achieved the best performances in most cases in both datasets, showing its robustness to different backbone architectures and number of labelled training images. Pascal VOC Backbone 1/106 [100] 1/50 [212] 1/20 [529] 1/8 [1323] Full [10582] Adv Sem Seg (Hung et al., 2019) Deep Labv2 - 57.20(17.70) 64.70(10.20) 69.50(5.40) 74.90 S4GAN (Mittal et al., 2019) Deep Labv2 - 63.30(12.30) 67.20(8.40) 71.40(4.20) 75.60 Cut Mix (French et al., 2020) Deep Labv2 53.79(18.71) 64.81(7.73) 66.48(6.06) 67.60(4.94) 72.54 Class Mix (Olsson et al., 2021) Deep Labv2 54.18(19.95) 66.15(7.98) 67.77(6.36) 71.00(3.13) 74.13 CCT (Ouali et al., 2020) PSPNet - - - 70.45(4.80) 75.25 CAC (Lai et al., 2021) PSPNet - - - 72.50(3.90) 76.40 GCT (Ke et al., 2020) Deep Labv2 - - - 70.57(3.49) 74.06 DMT (Feng et al., 2020) Deep Labv2 63.04(11.71) 67.15(7.60) 69.92(4.83) 72.70(2.05) 74.75 Re Co + Class Mix Deep Labv2 63.16(11.20) 66.41(7.95) 68.85(5.51) 71.00(3.36) 74.36 Re Co + Class Mix Deep Labv3+ 63.60(14.15) 72.14(5.61) 73.66(4.09) 74.62(3.13) 77.75 City Scapes Backbone 1/30 [100] 1/8 [372] 1/4 [744] 1/2 [1488] Full [2975] Adv Sem Seg (Hung et al., 2019) Deep Labv2 - 58.80(7.60) 62.30(4.10) 65.70(0.70) 66.40 S4GAN (Mittal et al., 2019) Deep Labv2 - 59.30(6.50) 61.90(3.90) - 65.80 Cut Mix (French et al., 2020) Deep Labv2 51.20(16.33) 60.34(7.19) 63.87(3.66) - 67.53 Class Mix (Olsson et al., 2021) Deep Labv2 54.07(12.12) 61.35(4.84) 63.63(2.56) 66.29( 0.10) 66.19 DMT (Feng et al., 2020) Deep Labv2 54.81(13.36) 63.03(5.13) - - 68.16 ECS (Mendel et al., 2020) Deep Labv3+ - 67.38(7.38) 70.70(4.06) 72.89(1.87) 74.76 Re Co + Class Mix Deep Labv2 56.53(12.07) 64.94(3.66) 67.53(1.07) 68.69( 0.09) 68.60 Re Co + Class Mix Deep Labv3+ 60.28(10.20) 66.44(4.04) 68.50(1.98) 70.63( 0.15) 70.48 Table 5: mean Io U validation performance in semi-supervsed Pascal VOC and City Scapes datasets. We list the percentage along with the number of labelled images at the top row. The first and second best performances in each data partition setting are coloured in red and orange respectively. trained images in doubled resolution. All results were taken from the corresponding publications. Published as a conference paper at ICLR 2022 D VISUALISATION ON PASCAL VOC (TRAINED WITH 60 LABELLED IMAGES) In the full label setting, the baselines Supervised and Class Mix are very prone to completely misclassifying rare objects such as boat, bottle and table, while our method can predict these rare classes accurately. Ground Truth Supervised (60 Labels) Class Mix (60 Labels) Re Co + Class Mix (60 Labels) Background Aeroplane Bicycle Bird Boat Bottle Bus Car Cat Chair Cow Table Dog Horse Motorbike Person Plant Sheep Sofa Train Monitor Invalid Published as a conference paper at ICLR 2022 E VISUALISATION ON CITYSCAPES (TRAINED WITH 1% LABELLED PIXEL) In the partial label setting, the performance improvements are less pronounced compared to the full label setting in City Scapes dataset. The improvements typically come from the more accurate predictions in small object boundaries such as in traffic light and traffic sign. Learning semantics with partial labels with minimal boundary information remains an open research question and still has huge scope for improvements. Ground Truth Supervised (1% Labels) Class Mix (1% Labels) Re Co + Class Mix (1% Labels) Road Sidewalk Building Wall Fence Pole Traffic Light Traffic Sign Vegetation Terrain Sky Person Rider Car Truck Bus Train Motorcycle Bicycle Invalid Published as a conference paper at ICLR 2022 F VISUALISATION ON SEMANTIC CLASS RELATIONSHIP FROM PASCAL VOC (TOP) AND SUN RGB-D (BOTTOM) We show features learned by Re Co are more disentangled compared to the Supervised baseline in all datasets, which helps the segmentation model to learn a better decision boundary. Brighter colour represents closer (more confused) relationship. Best viewed in zoom. cow dining table potted plant dining table potted plant cow dining table potted plant dining table potted plant cow dining table potted plant dining table potted plant potted plant dining table (a) Embed. Z (Supervised) potted plant dining table sofa aeroplane (b) Embed. Z (Re Co + Supervised) dining table bird tvmonitor potted plant 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 (c) Embed. R (Re Co + Supervised) wall floor cabinetbed chair sofa table door window bookshelf picture counter blinds desk shelves curtain dresser pillow mirror floor mat clothes ceiling books fridgetv paper towel shower curtainbox whiteboard person night stand toilet sink lamp bathtubbag wall floor cabinet sofa table door window bookshelf picture counter desk shelves curtain dresser pillow mirror floor mat clothes ceiling books fridge tv paper towel shower curtain box whiteboard person night stand sink lamp bathtub wall floor cabinetbed chair sofa table door window bookshelf picture counter blinds desk shelves curtain dresser pillow mirror floor mat clothes ceiling books fridgetv paper towel shower curtainbox whiteboard person night stand toilet sink lamp bathtubbag wall floor cabinet sofa table door window bookshelf picture counter desk shelves curtain dresser pillow mirror floor mat clothes ceiling books fridge tv paper towel shower curtain box whiteboard person night stand sink lamp bathtub wall floor cabinetbed chair sofa table door window bookshelf picture counter blinds desk shelves curtain dresser pillow mirror floor mat clothes ceiling books fridgetv paper towel shower curtainbox whiteboard person night stand toilet sink lamp bathtubbag wall floor cabinet sofa table door window bookshelf picture counter desk shelves curtain dresser pillow mirror floor mat clothes ceiling books fridge tv paper towel shower curtain box whiteboard person night stand sink lamp bathtub floor floor mat shower curtain curtain whiteboard bag bookshelf paper night stand (a) Embed. Z (Supervised) chair floor mat dresser night stand paper bookshelf sink shower curtain wall whiteboard (b) Embed. Z (Re Co + Supervised) tv fridge whiteboard dresser night stand curtain shower curtain floor floor mat 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 (c) Embed. R (Re Co + Supervised)