# selfsupervised_learning_with_kernel_dependence_maximization__47082a64.pdf Self-Supervised Learning with Kernel Dependence Maximization Yazhe Li Deep Mind and Gatsby Unit, UCL yazhe@google.com Roman Pogodin Gatsby Unit, UCL roman.pogodin.17@ucl.ac.uk Danica J. Sutherland UBC and Amii dsuth@cs.ubc.ca Arthur Gretton Gatsby Unit, UCL arthur.gretton@gmail.com We approach self-supervised learning of image representations from a statistical dependence perspective, proposing Self-Supervised Learning with the Hilbert Schmidt Independence Criterion (SSL-HSIC). SSL-HSIC maximizes dependence between representations of transformations of an image and the image identity, while minimizing the kernelized variance of those representations. This framework yields a new understanding of Info NCE, a variational lower bound on the mutual information (MI) between different transformations. While the MI itself is known to have pathologies which can result in learning meaningless representations, its bound is much better behaved: we show that it implicitly approximates SSLHSIC (with a slightly different regularizer). Our approach also gives us insight into BYOL, a negative-free SSL method, since SSL-HSIC similarly learns local neighborhoods of samples. SSL-HSIC allows us to directly optimize statistical dependence in time linear in the batch size, without restrictive data assumptions or indirect mutual information estimators. Trained with or without a target network, SSL-HSIC matches the current state-of-the-art for standard linear evaluation on Image Net [1], semi-supervised learning and transfer to other classification and vision tasks such as semantic segmentation, depth estimation and object recognition. Code is available at https://github.com/deepmind/ssl_hsic. 1 Introduction Learning general-purpose visual representations without human supervision is a long-standing goal of machine learning. Specifically, we wish to find a feature extractor that captures the image semantics of a large unlabeled collection of images, so that e.g. various image understanding tasks can be achieved with simple linear models. One approach takes the latent representation of a likelihood-based generative model [2 8]; such models, though, solve a harder problem than necessary since semantic features need not capture low-level details of the input. Another option is to train a self-supervised model for a pretext task, such as predicting the position of image patches, identifying rotations, or image inpainting [9 14]. Designing good pretext tasks, however, is a subtle art, with little theoretical guidance available. Recently, a class of models based on contrastive learning [15 22] has seen substantial success: dataset images are cropped, rotated, color shifted, etc. into several views, and features are then trained to pull together representations of the positive pairs of views of the same These authors contributed equally. Work done in part while at the Toyota Technological Institute at Chicago. 35th Conference on Neural Information Processing Systems (Neur IPS 2021). 24M 94M 250M 375M Number of parameters Image Net top-1 accuracy (%) Sim CLR (2x) Sim CLR (4x) (Res Net-200-2x) Barlow Twins SSL-HSIC (ours) (Res Net-50)(Res Net-50-2x) (Res Net-50-4x) Figure 1: Top-1 accuracies with linear evaluation for different Res Net architecture and methods: supervised (as in [25]), SSL-HSIC (with a target network; ours), BYOL [25], Sw AV [18], Sim CLR [19], Mo Co v2 [20] and Barlow Twins [22]. Figure 2: Statistical dependence view of contrastive learning: representations of transformed images should highly depend on image identity. Measuring dependence with HSIC, this pushes different images representation distributions apart (black arrows) and pulls representations of the same image together (colored shapes). source image, and push apart those of negative pairs (from different images). These methods are either understood from an information theoretic perspective as estimating the mutual information between the positives [15], or explained as aligning features subject to a uniformity constraint [23]. Another line of research [24, 25] attempts to learn representation without the negative pairs, but requires either a target network or stop-gradient operation to avoid collapsing. We examine the contrastive framework from a statistical dependence point of view: feature representations for a given transformed image should be highly dependent on the image identity (Figure 2). To measure dependence, we turn to the Hilbert-Schmidt Independence Criterion (HSIC) [26], and propose a new loss for self-supervised learning which we call SSL-HSIC. Our loss is inspired by HSIC Bottleneck [27, 28], an alternative to Information Bottleneck [29], where we use the image identity as the label, but change the regularization term. Through the dependence maximization perspective, we present a unified view of various selfsupervised losses. Previous work [30] has shown that the success of Info NCE cannot be solely attributed to properties of mutual information, in particular because mutual information (unlike kernel measures of dependence) has no notion of geometry in feature space: for instance, all invertible encoders achieve maximal mutual information, but they can output dramatically different representations with very different downstream performance [30]. Variational bounds on mutual information do impart notions of locality that allow them to succeed in practice, departing from the mutual information quantity that they try to estimate. We prove that Info NCE, a popular such bound, in fact approximates SSL-HSIC with a variance-based regularization. Thus, Info NCE can be thought of as working because it implicitly estimates a kernel-based notion of dependence. We additionally show SSL-HSIC is related to metric learning, where the features learn to align to the structure induced by the self-supervised labels. This perspective is closely related to the objective of BYOL [25], and can explain properties such as alignment and uniformity [23] observed in contrastive learning. Our perspective brings additional advantages, in computation and in simplicity of the algorithm, compared with existing approaches. Unlike the indirect variational bounds on mutual information [15, 31, 32], SSL-HSIC can be directly estimated from mini-batches of data. Unlike negative-free methods, the SSL-HSIC loss itself penalizes trivial solutions, so techniques such as target networks are not needed for reasonable outcomes. Using a target network does improve the performance of our method, however, suggesting target networks have other advantages that are not yet well understood. Finally, we employ random Fourier features [33] in our implementation, resulting in cost linear in batch size. Our main contributions are as follows: We introduce SSL-HSIC, a principled self-supervised loss using kernel dependence maximization. We present a unified view of contrastive learning through dependence maximization, by establishing relationships between SSL-HSIC, Info NCE, and metric learning. Our method achieves top-1 accuracy of 74.8% and top-5 accuracy of 92.2% with linear evaluations (see Figure 1 for a comparison with other methods), top-1 accuracy of 80.2% and Top-5 accuracy of 94.7% with fine-tuning, and competitive performance on a diverse set of downstream tasks. 2 Background 2.1 Self-supervised learning Recent developments in self-supervised learning, such as contrastive learning, try to ensure that features of two random views of an image are more associated with each other than with random views of other images. Typically, this is done through some variant of a classification loss, with one positive pair and many negatives. Other methods can learn solely from positive pairs, however. There have been many variations of this general framework in the past few years. Van den Oord et al. [15] first formulated the Info NCE loss, which estimates a lower bound of the mutual information between the feature and the context. Sim CLR [19, 34] carefully investigates the contribution of different data augmentations, and scales up the training batch size to include more negative examples. Mo Co [17] increases the number of negative examples by using a memory bank. BYOL [25] learns solely on positive image pairs, training so that representations of one view match that of the other under a moving average of the featurizer. Instead of the moving average, Sim Siam [24] suggests a stop-gradient on one of the encoders is enough to prevent BYOL from finding trivial solutions. Sw AV [18] clusters the representation online, and uses distance from the cluster centers rather than computing pairwise distances of the data. Barlow Twins [22] uses an objective related to the cross-correlation matrix of the two views, motivated by redundancy reduction. It is perhaps the most related to our work in the literature (and their covariance matrix can be connected to HSIC [35]), but our method measures dependency more directly. While Barlow Twins decorrelates components of final representations, we maximize the dependence between the image s abstract identity and its transformations. On the theory side, Info NCE is proposed as a variational bound on Mutual Information between the representation of two views of the same image [15, 32]. Tschannen et al. [30] observe that Info NCE performance cannot be explained solely by the properties of the mutual information, however, but is influenced more by other factors, such as the formulation of the estimator and the architecture of the feature extractor. Essentially, representations with the same MI can have drastically different representational qualities. better for SSL 1 MI(z 2 ,z )= log 2 for all cases Figure 3: Three distributions of positive examples for two classes (green and purple) that have the same mutual information, but drastically different quality for downstream learners. To see this, consider a problem with two inputs, A and B (Figure 3, green and purple), and a one-dimensional featurizer, parameterized by the integer M, which maps A to Uniform({0, 2, . . . , 2M}) and B to Uniform({1, 3, . . . , 2M + 1}). When M = 0, the inputs are encoded into linearly separable features A = 0 and B = 1 (Figure 3, bottom). Otherwise when M > 0, they are interspersed like ABABABAB a representation which is much harder to work with for downstream learners. Nevertheless, the mutual information between the features of any two augmentations of the same input (a positive pair) is independent of M, that is H[Z1] H[Z1|Z2] = log 2 for any M. Note that Info NCE would strongly prefer M = 0, indeed behaving very differently from MI. Later theories suggest that contrastive losses balance alignment of individual features and uniformity of the feature distribution [23], or in general alignment and some loss-defined distribution [36]. We propose to interpret the contrastive loss through the lens of statistical dependence, and relate it to metric learning, which naturally leads to alignment and uniformity. 2.2 Hilbert-Schmidt Independence Criterion (HSIC) The Hilbert-Schmidt Independence Criterion (HSIC) [26] is a kernel-based measure of dependence between probability distributions. Like mutual information, for a wide range of kernels HSIC(X, Y ) = 0 if and only if X and Y are independent [37], and large values of the measure correspond to more dependence. Unlike mutual information, HSIC incorporates a notion of ge- ometry (via the kernel choice), and is both statistically and computationally easy to estimate. It has been used in a variety of applications, particularly for independence testing [38], but it has also been maximized in applications such as feature selection [39], clustering [40, 41], active learning [42], and as a classification loss called HSIC Bottleneck [27, 28] (similar ideas were expressed in [43, 44]). HSIC measures the dependence between two random variables by first taking a nonlinear feature transformation of each, say φ : X F and ψ : Y G (with F and G reproducing kernel Hilbert spaces, RKHSes1), and then evaluating the norm of the cross-covariance between those features: HSIC(X, Y ) = E[φ(X) ψ(Y ) ] E[φ(X)] E[ψ(Y )] 2 Here HS is the Hilbert-Schmidt norm, which in finite dimensions is the usual Frobenius norm. HSIC measures the scale of the correlation in these nonlinear features, which allows it to identify nonlinear dependencies between X and Y with appropriate features φ and ψ. Inner products in an RKHS are by definition kernel functions: k(x, x ) = φ(x), φ(x ) F and l(y, y ) = ψ(y), ψ(y ) G. Let (X , Y ), (X , Y ) be independent copies of (X, Y ); this gives HSIC(X, Y ) = E [k(X, X )l(Y, Y )] 2 E [k(X, X )l(Y, Y )] + E [k(X, X )] E [l(Y, Y )] . (2) HSIC is also straightforward to estimate: given i.i.d. samples {(x1, y1), . . . , (x N, y N)} drawn i.i.d. from the joint distribution of (X, Y ), Gretton et al. [26] propose an estimator \ HSIC(X, Y ) = 1 (N 1)2 Tr(KHLH) , (3) where Kij = k(xi, xj) and Lij = l(yi, yj) are the kernel matrices, and H = I 1 N 11 is called the centering matrix. This estimator has an O(1/N) bias, which is not a concern for our uses; however, an unbiased estimator with the same computational cost is available [39]. 3 Self-supervised learning with Kernel Dependence Maximization Our method builds on the self-supervised learning framework used by most of the recent selfsupervised learning approaches [16 19, 22, 24, 25]. For a dataset with N points xi, each point goes through a random transformation tp(xi) (e.g. random crop), and then forms a feature representation zp i = fθ(tp(xi)) with an encoder network fθ. We associate each image xi with its identity yi, which works as a one-hot encoded label: yi RN and (yi)d = 1 iff d = i (and zero otherwise). To match the transformations and image identities, we maximize the dependence between zi and yi such that zi is predictive of its original image. To build representations suitable for downstream tasks, we also need to penalize high-variance representations. These ideas come together in our HSIC-based objective for self-supervised learning, which we term SSL-HSIC: LSSL HSIC(θ) = HSIC (Z, Y ) + γ p HSIC (Z, Z) . (4) Unlike contrastive losses, which make the zp i from the same xi closer and those from different xj more distant, we propose an alternative way to match different transformations of the same image with its abstract identity (e.g. position in the dataset). Our objective also resembles the HSIC bottleneck for supervised learning [27] (in particular, the version of [28]), but ours uses a square root for HSIC(Z, Z). The square root makes the two terms on the same scale: HSIC(Z, Y ) is effectively a dot product, and p HSIC(Z, Z) a norm, so that e.g. scaling the kernel by a constant does not change the relative amount of regularization;2 this also gives better performance in practice. Due to the one-hot encoded labels, we can re-write HSIC (Z, Y ) as (see Appendix A) HSIC(Z, Y ) Ez1,z2 pos [k(z1, z2)] Ez1Ez2 [k(z1, z2)] , (5) 1In a slight abuse of notation, we use φ(x)ψ(y) for the tensor product φ(x) ψ(y) F G. 2Other prior work on maximizing HSIC [41, 45] used HSIC(Z, Y )/ p HSIC(Z, Z) HSIC(Y, Y ), or equivalently [46] the distance correlation [47]; the kernel-target alignment [48, 49] is also closely related. Here, the overall scale of either kernel does not change the objective. Our HSIC(Y, Y ) is constant (hence absorbed in γ), and we found an additive penalty to be more stable in optimization than dividing the estimators. where the first expectation is over the distribution of positive pairs (those from the same source image), and the second one is a sum over all image pairs, including their transformations. The first term in (5) pushes representations belonging to the same image identity together, while the second term keeps mean representations for each identity apart (as in Figure 2). The scaling of HSIC(Z, Y ) depends on the choice of the kernel over Y , and is irrelevant to the optimization. This form also reveals three key theoretical results. Section 3.1 shows that Info NCE is better understood as an HSIC-based loss than a mutual information between views. Section 3.2 reveals that the dependence maximization in HSIC(Z, Y ) can also be viewed as a form of distance metric learning, where the cluster structure is defined by the labels. Finally, HSIC(Z, Y ) is proportional to the average kernel-based distance between the distribution of views for each source image (the maximum mean discrepancy, MMD; see Appendix B.2). 3.1 Connection to Info NCE In this section we show the connection between Info NCE and our loss; see Appendix B for full details. We first write the latter in its infinite sample size limit (see [23] for a derivation) as LInfo NCE(θ) = Ez1,z2 pos [k(z1, z2)] + Ez1 log Ez2 [exp (k(z1, z2))] , (6) where the last two expectations are taken over all points, and the first is over the distribution of positive pairs. The kernel k(z1, z2) was originally formulated as a scoring function in a form of a dot product [15], and then a scaled cosine similarity [19]. Both functions are valid kernels. Now assume that k(z1, z2) doesn t deviate much from Ez2 [k(z1, z2)], Taylor-expand the exponent in (6) around Ez2 [k(z1, z2)], then expand log(1 + Ez2(. . . )) Ez2(. . . ). We obtain an HSIC(Z, Y )- based objective: LInfo NCE(θ) Ez1,z2 pos [k(z1, z2)] + Ez1Ez2 [k(z1, z2)] | {z } HSIC(Z,Y ) 2 Ez1 [Varz2 [k(z1, z2)]] | {z } variance penalty Since the scaling of HSIC(Z, Y ) is irrelevant to the optimization, we assume scaling to replace with =. In the small variance regime, we can show that for the right γ, HSIC (Z, Y ) + γ HSIC (Z, Z) LInfo NCE(θ) + o(variance) . (8) For HSIC(Z, Z) 1, we also have that HSIC (Z, Y ) + γ HSIC (Z, Z) LSSL HSIC(θ) (9) due to the square root. Info NCE and SSL-HSIC in general don t quite bound each other due to discrepancy in the variance terms, but in practice the difference is small. Why should we prefer the HSIC interpretation of Info NCE? Initially, Info NCE was suggested as a variational approximation to the mutual information between two views [15]. It has been observed, however, that using tighter estimators of mutual information leads to worse performance [30]. It is also simple to construct examples where Info NCE finds different representations while the underlying MI remains constant [30]. Alternative theories suggest that Info NCE balances alignment of positive examples and uniformity of the overall feature representation [23], or that (under strong assumptions) it can identify the latent structure in a hypothesized data-generating process, akin to nonlinear ICA [50]. Our view is consistent with these theories, but doesn t put restrictive assumptions on the input data or learned representations. In Section 5 (summarized in Table 8a), we show that our interpretation gives rise to a better objective in practice. 3.2 Connection to metric learning Our SSL-HSIC objective is closely related to kernel alignment [48], especially centered kernel alignment [49]. As a kernel method for distance metric learning, kernel alignment measures the agreement between a kernel function and a target function. Intuitively, the self-supervised labels Y imply a cluster structure, and HSIC(Z, Y ) estimates the degree of agreement between the learned features and this cluster structure in the kernel space. This relationship with clustering is also established in [40, 41, 45], where labels are learned rather than features. The clustering perspective is more evident when we assume linear kernels over both Z and Y , and Z is unit length and centered:3 HSIC(Z, Y ) 1 M Tr(Y Z ZY ) + Tr(Z Z) NM p=1 zp i 2 2 NM = p=1 zp i zi 2 2 NM, (10) with M the number of augmentations per image and zi = P p zp i /M the average feature vector of the augmented views of xi. We emphasize, though, that (10) assumes centered, normalized data with linear kernels; the right-hand side of (10) could be optimized by setting all zp i to the same vector for each i, but this does not actually optimize HSIC(Z, Y ). Equation (10) shows that we recover the spectral formulation [51] and sum-of-squares loss used in the k-means clustering algorithm from the kernel objective. Moreover, the self-supervised label imposes that the features from transformations of the same image are gathered in the same cluster. Equation (10) also allows us to connect SSL-HSIC to non-contrastive objectives such as BYOL, although the connection is subtle because of its use of predictor and target networks. If each image is augmented with two views, we can compute (10) using zi (z1 i + z2 i )/2, so the clustering loss becomes P i ||z1 i z2 i ||2 2. This is exactly the BYOL objective, only that z2 i in BYOL comes from a target network. The assumption of centered and normalized features for (10) is important in the case of BYOL: without it, BYOL can find trivial solutions where all the features are collapsed to the same feature vector far away from the origin. The target network is used to prevent the collapse. SSL-HSIC, on the other hand, rules out such a solution by building the centering into the loss function, and therefore can be trained successfully without a target network or stop gradient operation. 3.3 Estimator of SSL-HSIC To use SSL-HSIC, we need to correctly and efficiently estimate (4). Both points are non-trivial: the self-supervised framework implies non-i.i.d. batches (due to positive examples), while the estimator in (3) assumes i.i.d. data; moreover, the time to compute (3) is quadratic in the batch size. First, for HSIC(Z, Z) we use the biased estimator in (3). Although the i.i.d. estimator (3) results in an O(1/B) bias for B original images in the batch size (see Appendix A), the batch size B is large in our case and therefore the bias is negligible. For HSIC(Z, Y ) the situation is more delicate: the i.i.d. estimator needs re-scaling, and its bias depends on the number of positive examples M, which is typically very small (usually 2). We propose the following estimator: \ HSIC(Z, Y ) = l ipl k(zp i , zl i) 1 B2M 2 X ijpl k(zp i , zl j) 1 M 1 where i and j index original images, and p and l their random transformations; k is the kernel used for latent Z, l is the kernel used for the labels, and l = l(i, i) l(i, j) (l for same labels minus l for different labels). Note that due to the one-hot structure of self-supervised labels Y , the standard (i.i.d.-based) estimator would miss the 1/N scaling and the M 1 correction (the latter is important in practice, as we usually have M = 2). See Appendix A for the derivations. For convenience, we assume l = N (any scaling of l can be subsumed by γ), and optimize b LSSL HSIC(θ) = \ HSIC(Z, Y ) + γ q \ HSIC(Z, Z) . (12) The computational complexity of the proposed estimators is O(B2M 2) for each mini-batch of size B with M augmentations. We can reduce the complexity to O(BM) by using random Fourier features (RFF) [33], which approximate the kernel k(z1, z2) with a carefully chosen random D-dimensional approximation R(z1) R(z2) for R(z) : RDz RD, such that k(z1, z2) = E R(z1) R(z2) . Fourier frequencies are sampled independently for the two kernels on Z in HSIC(Z, Z) at each training step. We leave the details on how to construct R(z) for the kernels we use to Appendix C. 3Centered Z is a valid assumption for BYOL, as the target network keeps representations of views with different image identities away from each other. For high-dimensional unit vectors, this can easily lead to orthogonal representations. We also observe centered representations empirically: see Appendix B.3. 4 Experiments In this section, we present our experimental setup, where we assess the performance of the representation learned with SSL-HSIC both with and without a target network. First, we train a model with a standard Res Net-50 backbone using SSL-HSIC as objective on the training set of Image Net ILSVRC-2012 [1]. For evaluation, we retain the backbone as a feature extractor for downstream tasks. We evaluate the representation on various downstream tasks including classification, object segmentation, object detection and depth estimation. 1 k(z 2 ,z ) Figure 4: Architecture and SSL-HSIC objective. A self-supervised label y an indicator of the image identity is associated with an image x. Image transformation functions t are sampled and applied to the original image, resulting in views t1(x) and t2(x). Features z1 and z2 are obtained after passing the augmented views through encoder (f), projector (g), and possibly predictor (q) networks, while label y is retained. Kernel matrices, K for the latents and L for the labels, are computed on the mini-batch of data; SSL-HSIC is estimated with K and L as in (12). The blue boxes reflect two potential options: when using a target network, ξ is a moving average of θ, and a predictor network q is added; without the target network, q is removed and ξ is simply equal to θ. 4.1 Implementation Architecture Figure 4 illustrates the architecture we used for SSL-HSIC in this section. To facilitate comparison between different methods, our encoder fθ uses the standard Res Net-50 backbone without the final classification layer. The output of the encoder is a 2048-dimension embedding vector, which is the representation used for downstream tasks. As in BYOL [25], our projector g and predictor q networks are 2-layer MLPs with 4096 hidden dimensions and 256 output dimensions. The outputs of the networks are batch-normalized and rescaled to unit norm before computing the loss. We use an inverse multiquadric kernel (IMQ) for the latent representation (approximated with 512 random Fourier features that are resampled at each step; see Appendix C for details) and a linear kernel for labels. γ in (4) is set to 3. When training without a target network, unlike Sim Siam [24], we do not stop gradients for either branch. If the target network is used, its weights are an exponential moving average of the online network weights. We employ the same schedule as BYOL [25], τ = 1 0.01 (cos(πt/T) + 1)/2 with t the current step and T the total training steps. Image augmentation Our method uses the same data augmentation scheme as BYOL (see Appendix D.1). Briefly, we first draw a random patch from the original image and resize it to 224 224. Then, we apply a random horizontal flip, followed by color jittering, consisting of a random sequence of brightness, contrast, saturation, hue adjustments, and an optional grayscale conversion. Finally Gaussian blur and solarization are applied, and the view is normalized with Image Net statistics. Optimization We train the model with a batch size of 4096 on 128 Cloud TPU v4 cores. Again, following [19, 25], we use the LARS optimizer [52] with a cosine decay learning rate schedule over 1000 epochs. The base learning rate to all of our experiments is 0.4 and it is scaled linearly [53] with the batch size lr = 0.4 batch_size/256. All experiments use weight decay of 10 6. Learning kernel parameters We use a linear kernel for labels, since the type of kernel only scales (12). Our inverse multiquadric kernel for the latent Z has an additional kernel scale parameter. We optimize this along with all other parameters, but regularize it to maximize the entropy of the distribution kσ(s), where sij = zi zj 2; this amounts to maximizing log k σ(s) 2 (Appendix D.1.2). 4.2 Evaluation Results Linear evaluation on Image Net Learned features are evaluated with the standard linear evaluation protocol commonly used in evaluating self-supervised learning methods [15 22, 25]. Table 1 reports the top-1 and top-5 accuracies obtained with SSL-HSIC on Image Net validation set, and compares to previous self-supervised learning methods. Without a target network, our method reaches 72.2% top-1 and 90.7% top-5 accuracies. Unlike BYOL, the SSL-HSIC objective prevents the network from finding trivial solutions as explained in Section 3.2. Adding the target network, our method outperforms most previous methods, achieving top-1 accuracy of 74.8% and top-5 accuracy of 92.2%. The fact that we see performance gains from adopting a target network suggests that its effect is not yet well understood, although note discussion in [25] which points to its stabilizing effect. Table 1: Linear evaluation on the Image Net validation set. Top-1(%) Top-5(%) Supervised [54] 75.9 92.8 Sim CLR [19] 69.3 89.0 Mo Co v2 [20] 71.1 90.1 BYOL [25] 74.3 91.6 Sw AV [18] 75.3 - Barlow Twins [22] 73.2 91.0 SSL-HSIC (w/o target) 72.2 90.7 SSL-HSIC (w/ target) 74.8 92.2 Table 2: Fine-tuning on 1%, 10% and 100% of the Image Net training set and evaluating on the validation set. Top-1(%) Top-5(%) 1% 10% 100% 1% 10% 100% Supervised [54] 25.4 56.4 75.9 48.4 80.4 92.8 Sim CLR [19] 48.3 65.6 76.0 75.5 87.8 93.1 BYOL [25] 53.2 68.8 77.7 78.4 89.0 93.9 Sw AV [18] 53.9 70.2 - 78.5 89.9 - Barlow Twins [22] 55.0 69.7 - 79.2 89.3 - SSL-HSIC (w/o target) 45.3 65.5 76.4 72.7 87.5 93.2 SSL-HSIC (w/ target) 52.1 67.9 77.2 77.7 88.6 93.6 Semi-supervised learning on Image Net We fine-tune the network pretrained with SSL-HSIC on 1%, 10% and 100% of Image Net, using the same Image Net splits as Sim CLR [19]. Table 2 summarizes the semi-supervised learning performance. Our method, with or without a target network, has competitive performance in both data regimes. The target network has the most impact on the small-data regime, with 1% labels. Table 3: Comparison of transfer learning performance on 12 image datasets. Supervised-IN is trained on Image Net with supervised pretrainining. Random init trains on individual dataset with randomly initialized weights. MPCA refers to mean per-class accuracy; AP50 is average precision at Io U=0.5. Dataset Birdsnap Caltech101 Cifar10 Cifar100 DTD Aircraft Food Flowers Pets Cars SUN397 VOC2007 Metric Top-1 MPCA Top-1 Top-1 Top-1 MPCA Top-1 MPCA MPCA Top-1 Top-1 AP50 Supervised-IN [19] 53.7 94.5 93.6 78.3 74.9 61.0 72.3 94.7 91.5 67.8 61.9 82.8 Sim CLR [19] 37.4 90.3 90.6 71.6 74.5 50.3 68.4 90.3 83.6 50.3 58.8 80.5 BYOL [25] 57.2 94.2 91.3 78.4 75.5 60.6 75.3 96.1 90.4 66.7 62.2 82.5 SSL-HSIC (w/o target) 50.6 92.3 91.5 75.9 75.3 57.9 73.6 95.0 88.2 59.3 61.0 81.4 SSL-HSIC (w/ target) 57.8 93.5 92.3 77.0 76.2 58.5 75.6 95.4 91.2 62.6 61.8 83.3 Fine-tuned: Supervised-IN [19] 75.8 93.3 97.5 86.4 74.6 86.0 88.3 97.6 92.1 92.1 94.3 85.0 Random init [19] 76.1 72.6 95.9 80.2 64.8 85.9 86.9 92.0 81.5 91.4 53.6 67.3 Sim CLR [19] 75.9 92.1 97.7 85.9 73.2 88.1 88.2 97.0 89.2 91.3 63.5 84.1 BYOL [25] 76.3 93.8 97.8 86.1 76.2 88.1 88.5 97.0 91.7 91.6 63.7 85.4 SSL-HSIC (w/o target) 73.1 91.5 97.4 85.3 75.3 87.1 87.5 96.4 90.6 91.6 62.2 84.1 SSL-HSIC (w/ target) 74.9 93.8 97.8 84.7 75.4 88.9 87.7 97.3 91.7 91.8 61.7 84.1 Transfer to other classification tasks To investigate the generality of the representation learned with SSL-HSIC, we evaluate the transfer performance for classification on 12 natural image datasets [55 64] using the same procedure as [19, 25, 65]. Table 3 shows the top-1 accuracy of the linear evaluation and fine-tuning performance on the test set. SSL-HSIC gets state-of-the-art performance on 3 of the classification tasks and reaches strong performance on others for this benchmark, indicating the learned representations are robust for transfer learning. Transfer to other vision tasks To test the ability of transferring to tasks other than classification, we fine-tune the network on semantic segmentation, depth estimation and object detection tasks. We use Pascal VOC2012 dataset [58] for semantic segmentation, NYU v2 dataset [66] for depth estimation and COCO [67] for object detection. Object detection outputs either bounding box or object segmentation (instance segmentation). Details of the evaluations setup is in Appendix D.2. Table 4 and Table 5 shows that SSL-HSIC achieves competitive performance on all three vision tasks. Table 4: Fine-tuning performance on semantic segmentation and depth estimation. Mean Intersection over Union (m Io U) is reported for semantic segmentation. Relative error (rel), root mean squared error (rms), and the percent of pixels (pct) where the error is below 1.25n thresholds are reported for depth estimation. VOC2012 NYU v2 Method m Io U pct.< 1.25 pct.< 1.252 pct.< 1.253 rms rel Supervised-IN 74.4 81.1 95.3 98.8 0.573 0.127 Sim CLR 75.2 83.3 96.5 99.1 0.557 0.134 BYOL 76.3 84.6 96.7 99.1 0.541 0.129 SSL-HSIC(w/o target) 74.9 84.1 96.7 99.2 0.539 0.130 SSL-HSIC(w/ target) 76.0 83.8 96.8 99.1 0.548 0.130 Table 5: Fine-tuning performance on COCO object detection tasks. Precision, averaged over 10 Io U (Intersection over Union) thresholds, is reported for both bounding box and object segmentation. Method APbb APmk Supervised 39.6 35.6 Sim CLR 39.7 35.8 Mo Co v2 40.1 36.3 BYOL 41.6 37.2 Sw AV 41.6 37.8 SSL-HSIC(w/o target) 40.5 36.3 SSL-HSIC(w/ target) 41.3 36.8 5 Ablation Studies We present ablation studies to gain more intuition on SSL-HSIC. Here, we use a Res Net-50 backbone trained for 100 epochs on Image Net, and evaluate with the linear protocol unless specified. Res Net architectures In this ablation, we investigate the performance of SSL-HSIC with wider and deeper Res Net architecture. Figure 1 and Table 6 show our main results. The performance of SSL-HSIC gets better with larger networks. We used the supervised baseline from [25] which our training framework is based on ([19] reports lower performance). The performance gap between SSL-HSIC and the supervised baseline diminishes with larger architectures. In addition, Table 7 presents the semi-supervise learning results with subsets 1%, 10% and 100% of the Image Net data. Table 6: Top-1 and top-5 accuracies for different Res Net architectures using linear evaluation protocol. SSL-HSIC BYOL[25] Sup.[25] Res Net Top1 Top5 Top1 Top5 Top1 Top5 50 (1x) 74.8 92.2 74.3 91.6 76.4 92.9 50 (2x) 77.9 94.0 77.4 93.6 79.9 95.0 50 (4x) 79.1 94.5 78.6 94.2 80.7 95.3 200 (2x) 79.6 94.8 79.6 94.9 80.1 95.2 Table 7: Top-1 and top-5 accuracies for different Res Net architectures using semisupervised fine-tuning. Res Net 1% 10% 100% 1% 10% 100% 50 (1x) 52.1 67.9 77.2 77.7 88.6 93.6 50 (2x) 61.2 72.6 79.3 83.8 91.2 94.7 50 (4x) 67.0 75.4 79.7 87.4 92.5 94.8 200(2x) 69.0 76.3 80.5 88.3 92.9 95.2 Regularization term We compare performance of Info NCE with SSL-HSIC in Table 8a since they can be seen as approximating the same HSIC(Z, Y ) objective but with different forms of regularization. We reproduce the Info NCE result in our codebase, using the same architecture and data augmentiation as for SSL-HSIC. Trained for 100 epochs (without a target network), Info NCE achieves 66.0% top-1 and 86.9% top-5 accuracies, which is better than the result reported in [19]. For comparison, SSL-HSIC reaches 66.7% top-1 and 87.6% top-5 accuracies. This suggests that the regularization employed by SSL-HSIC is more effective. Kernel type We investigate the effect of using different a kernel on latents Z. Training without a target network or random Fourier feature approximation, the top-1 accuracies for linear, Gaussian, and inverse multiquadric (IMQ) kernels are 65.27%, 66.67% and 66.72% respectively. Non-linear kernels indeed improve the performance; Gaussian and IMQ kernels reach very similar performance for 100 epochs. We choose IMQ kernel for longer runs, because its heavy-tail property can capture more signal when points are far apart. Number of RFF Features Table 8b shows the performance of SSL-HSIC with different numbers of Fourier features. The RFF approximation has a minor impact on the overall performance, as long as we resample them; fixed sets of features performed poorly. Our main result picked 512 features, for substantial computational savings with minor loss in accuracy. Batch size Similar to most of the self-supervised learning methods [19, 25], SSL-HSIC benefits from using a larger batch size during training. However, the drop of performance from using smaller batch size is not as pronounced as it is in Sim CLR[19] as shown in Table 8c. Table 8: Linear evaluation results when varying different hyperparameters. (a) Regularization Top-1 Top-5 SSL-HSIC 66.7 87.6 Info NCE 66.0 86.9 (b) # Fourier features # RFFs Top-1(%) 64 66.0 128 66.2 256 66.2 512 66.4 1024 66.5 2048 66.5 No Approx. 66.7 (c) Batch size Batch Size SSL-HSIC Sim CLR 256 63.7 57.5 512 65.6 60.7 1024 66.7 62.8 2048 67.1 64.0 4096 66.7 64.6 (d) Projector/predictor size Output Dim Top-1(%) 64 65.4 128 66.0 256 66.4 512 66.6 1024 66.6 Projector and predictor output size Table 8d shows the performance when using different output dimension for the projector/predictor networks. The performance saturates at 512 dimensions. 6 Conclusions We introduced SSL-HSIC, a loss function for self-supervised representation learning based on kernel dependence maximization. We provided a unified view on various self-supervised learning losses: we proved that Info NCE, a lower bound of mutual information, actually approximates SSL-HSIC with a variance-based regularization, and we can also interpret SSL-HSIC as metric learning where the cluster structure is imposed by the self-supervised label, of which the BYOL objective is a special case. We showed that training with SSL-HSIC achieves performance on par with the state-of-the-art on the standard self-supervised benchmarks. Although using the image identity as self-supervised label provides a good inductive bias, it might not be wholly satisfactory; we expect that some images pairs are in fact more similar than others, based e.g. on their Image Net class label. It will be interesting to explore methods that combine label structure discovery with representation learning (as in Sw AV [18]). In this paper, we only explored learning image representations, but in future work SSL-HSIC can be extended to learning structure for Y as well, building on existing work [41, 45]. Broader impact Our work concentrates on providing a more theoretically grounded and interpretable loss function for self-supervised learning. A better understanding of self-supervised learning, especially through more interpretable learning dynamics, is likely to lead to better and more explicit control over societal biases of these algorithms. SSL-HSIC yields an alternative, clearer understanding of existing selfsupervised methods. As such, it is unlikely that our method introduces further biases than those already present for self-supervised learning. The broader impacts of the self-supervised learning framework is an area that has not been studied by the AI ethics community, but we think it calls for closer inspection. An important concern for fairness of ML algorithms is dataset bias. Image Net is known for a number of problems such as offensive annotations, non-visual concepts and lack of diversity, in particular for underrepresented groups. Existing works and remedies typically focus on label bias. Since SSL doesn t use labels, however, the type and degree of bias could be very different from that of supervised learning. To mitigate the risk of dataset bias, one could employ dataset re-balancing to correct sampling bias [68] or completely exclude human images from the dataset while achieving the same performance [69]. A new topic to investigate for self-supervised learning is how the bias/unbiased representation could be transferred to downstream tasks. We are not aware of any work in this direction. Another area of concern is security and robustness. Compared to supervised learning, self-supervised learning typically involves more intensive data augmentation such as color jittering, brightness adjustment, etc. There is some initial evidence suggesting self-supervised learning improves model robustness [70]. However, since data augmentation can either be beneficial [71] or detrimental [72] depending on the type of adversarial attacks, more studies are needed to assess its role for self-supervised learning. Acknowledgments and Disclosure of Funding The authors would like to thank Olivier J. Hénaff for valuable feedback on the manuscript and the help for evaluating object detection task. We thank Aaron Van den Oord and Oriol Vinyals for providing valuable feedback on the manuscript. We are grateful to Yonglong Tian, Ting Chen and the BYOL authors for the help with reproducing baselines and evaluating downstream tasks. This work was supported by Deep Mind, the Gatsby Charitable Foundation, the Wellcome Trust, NSERC, and the Canada CIFAR AI Chairs program. [1] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei. Image Net large scale visual recognition challenge. International Journal of Computer Vision 115.3 (2015), pp. 211 252. [2] G. E. Hinton and R. R. Salakhutdinov. Reducing the dimensionality of data with neural networks. Science 313.5786 (2006), pp. 504 507. [3] P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, P.-A. Manzagol, and L. Bottou. Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion. Journal of Machine Learning Research 11.12 (2010). [4] A. Coates and A. Y. Ng. Learning feature representations with k-means. Neural networks: Tricks of the trade. Springer, 2012, pp. 561 580. [5] I. Higgins, L. Matthey, A. Pal, C. Burgess, X. Glorot, M. Botvinick, S. Mohamed, and A. Lerchner. beta-VAE: Learning basic visual concepts with a constrained variational framework. ICLR. 2016. [6] A. van den Oord, O. Vinyals, and K. Kavukcuoglu. Neural Discrete Representation Learning. Neur IPS. 2017. ar Xiv: 1711.00937. [7] W. Grathwohl, R. T. Q. Chen, J. Bettencourt, and D. Duvenaud. Scalable Reversible Generative Models with Free-form Continuous Dynamics. ICLR. 2019. ar Xiv: 1810.01367. [8] T. B. Brown, B. Mann, N. Ryder, M. Subbiah, J. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam, G. Sastry, A. Askell, S. Agarwal, A. Herbert-Voss, G. Krueger, T. Henighan, R. Child, A. Ramesh, D. M. Ziegler, J. Wu, C. Winter, C. Hesse, M. Chen, E. Sigler, M. Litwin, S. Gray, B. Chess, J. Clark, C. Berner, S. Mc Candlish, A. Radford, I. Sutskever, and D. Amodei. Language Models are Few-Shot Learners. Neur IPS. 2020. ar Xiv: 2005.14165. [9] C. Doersch, A. Gupta, and A. A. Efros. Unsupervised Visual Representation Learning by Context Prediction. ICCV. 2015. ar Xiv: 1505.05192. [10] G. Larsson, M. Maire, and G. Shakhnarovich. Learning representations for automatic colorization. ECCV. Springer. 2016, pp. 577 593. [11] M. Noroozi and P. Favaro. Unsupervised learning of visual representations by solving jigsaw puzzles. ECCV. Springer. 2016, pp. 69 84. [12] D. Pathak, P. Krahenbuhl, J. Donahue, T. Darrell, and A. A. Efros. Context encoders: Feature learning by inpainting. CVPR. 2016, pp. 2536 2544. [13] R. Zhang, P. Isola, and A. A. Efros. Colorful image colorization. ECCV. Springer. 2016, pp. 649 666. [14] S. Gidaris, P. Singh, and N. Komodakis. Unsupervised representation learning by predicting image rotations. ICLR. 2018. ar Xiv: 1803.07728. [15] A. van den Oord, Y. Li, and O. Vinyals. Representation learning with contrastive predictive coding. 2018. ar Xiv: 1807.03748. [16] P. Bachman, R. D. Hjelm, and W. Buchwalter. Learning representations by maximizing mutual information across views. Neur IPS. 2019. ar Xiv: 1906.00910. [17] K. He, H. Fan, Y. Wu, S. Xie, and R. B. Girshick. Momentum Contrast for Unsupervised Visual Representation Learning. CVPR. 2019. ar Xiv: 1911.05722. [18] M. Caron, I. Misra, J. Mairal, P. Goyal, P. Bojanowski, and A. Joulin. Unsupervised learning of visual features by contrasting cluster assignments. Neur IPS. 2020. ar Xiv: 2006.09882. [19] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton. A Simple Framework for Contrastive Learning of Visual Representations. ICML. 2020. ar Xiv: 2002.05709. [20] X. Chen, H. Fan, R. B. Girshick, and K. He. Improved Baselines with Momentum Contrastive Learning. 2020. ar Xiv: 2003.04297. [21] O. Henaff. Data-efficient image recognition with contrastive predictive coding. ICML. 2020. ar Xiv: 1905.09272. [22] J. Zbontar, L. Jing, I. Misra, Y. Le Cun, and S. Deny. Barlow Twins: Self-Supervised Learning via Redundancy Reduction. 2021. ar Xiv: 2103.03230. [23] T. Wang and P. Isola. Understanding Contrastive Representation Learning through Alignment and Uniformity on the Hypersphere. 2020. ar Xiv: 2005.10242. [24] X. Chen and K. He. Exploring Simple Siamese Representation Learning. 2020. ar Xiv: 2011.10566. [25] J.-B. Grill, F. Strub, F. Altché, C. Tallec, P. H. Richemond, E. Buchatskaya, C. Doersch, B. A. Pires, Z. D. Guo, M. G. Azar, B. Piot, K. Kavukcuoglu, R. Munos, and M. Valko. Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning. Neur IPS. 2020. ar Xiv: 2006.07733. [26] A. Gretton, O. Bousquet, A. Smola, and B. Schölkopf. Measuring statistical dependence with Hilbert-Schmidt norms. ALT. Springer. 2005, pp. 63 77. [27] K. W.-D. Ma, J. P. Lewis, and W. B. Kleijn. The HSIC Bottleneck: Deep Learning without Back-Propagation. AAAI. 2020. [28] R. Pogodin and P. Latham. Kernelized information bottleneck leads to biologically plausible 3-factor Hebbian learning in deep networks. Neur IPS (2020). [29] N. Tishby and N. Zaslavsky. Deep Learning and the Information Bottleneck Principle. IEEE Information Theory Workshop. 2015. ar Xiv: 1503.02406. [30] M. Tschannen, J. Djolonga, P. K. Rubenstein, S. Gelly, and M. Lucic. On mutual information maximization for representation learning. ICLR. 2020. ar Xiv: 1907.13625. [31] I. Belghazi, S. Rajeswar, A. Baratin, R. D. Hjelm, and A. C. Courville. MINE: Mutual Information Neural Estimation. ICML. 2018. ar Xiv: 1801.04062. [32] B. Poole, S. Ozair, A. van den Oord, A. A. Alemi, and G. Tucker. On Variational Bounds of Mutual Information. ICML. 2019. ar Xiv: 1905.06922. [33] A. Rahimi and B. Recht. Random Features for Large-Scale Kernel Machines. Neur IPS. 2007. [34] T. Chen, S. Kornblith, K. Swersky, M. Norouzi, and G. Hinton. Big Self-Supervised Models are Strong Semi-Supervised Learners. Neur IPS. 2020. ar Xiv: 2006.10029. [35] Y.-H. H. Tsai, S. Bai, L.-P. Morency, and R. Salakhutdinov. A note on connecting barlow twins with negative-sample-free contrastive learning. ar Xiv preprint ar Xiv:2104.13712 (2021). [36] T. Chen and L. Li. Intriguing Properties of Contrastive Losses. 2020. ar Xiv: 2011.02803. [37] Z. Szabó and B. K. Sriperumbudur. Characteristic and Universal Tensor Product Kernels. Journal of Machine Learning Research 18.233 (2018), pp. 1 29. [38] A. Gretton, K. Fukumizu, C. H. Teo, L. Song, B. Schölkopf, and A. J. Smola. A kernel statistical test of independence. Neur IPS. 2007. [39] L. Song, A. Smola, A. Gretton, J. Bedo, and K. Borgwardt. Feature selection via dependence maximization. Journal of Machine Learning Research 13.5 (2012). [40] L. Song, A. Smola, A. Gretton, and K. M. Borgwardt. A Dependence Maximization View of Clustering. ICML. 2007. [41] M. B. Blaschko and A. Gretton. Learning Taxonomies by Dependence Maximization. Neur IPS. 2009. [42] S. Jain, G. Liu, and D. Gifford. Information Condensing Active Learning. 2020. ar Xiv: 2002.07916. [43] D. Wu, Y. Zhao, Y.-H. H. Tsai, M. Yamada, and R. Salakhutdinov. Dependency Bottleneck in Auto-encoding Architectures: an Empirical Study. 2018. ar Xiv: 1802.05408. [44] A. Nøkland and L. H. Eidnes. Training neural networks with local error signals. ICML. 2019. ar Xiv: 1901.06656. [45] M. B. Blaschko, W. Zaremba, and A. Gretton. Taxonomic Prediction with Tree-Structured Covariances. Machine Learning and Knowledge Discovery in Databases. Springer Berlin Heidelberg, 2013, pp. 304 319. [46] D. Sejdinovic, B. Sriperumbudur, A. Gretton, and K. Fukumizu. Equivalence of distancebased and RKHS-based statistics in hypothesis testing. The Annals of Statistics 41.5 (2013), pp. 2263 2291. [47] G. J. Székely, M. L. Rizzo, and N. K. Bakirov. Measuring and testing dependence by correlation of distances. The Annals of Statistics 35.6 (2007), pp. 2769 2794. [48] N. Cristianini, J. Shawe-Taylor, A. Elisseeff, and J. Kandola. On Kernel-Target Alignment. Neur IPS. Vol. 14. 2002. [49] C. Cortes, M. Mohri, and A. Rostamizadeh. Algorithms for learning kernels based on centered alignment. The Journal of Machine Learning Research 13.1 (2012), pp. 795 828. [50] R. S. Zimmermann, Y. Sharma, S. Schneider, M. Bethge, and W. Brendel. Contrastive Learning Inverts the Data Generating Process. 2021. ar Xiv: 2102.08850. [51] H. Zha, X. He, C. Ding, M. Gu, and H. D. Simon. Spectral relaxation for k-means clustering. Neur IPS. 2001, pp. 1057 1064. [52] Y. You, I. Gitman, and B. Ginsburg. Scaling SGD Batch Size to 32K for Image Net Training. 2017. ar Xiv: 1708.03888. [53] P. Goyal, P. Dollár, R. B. Girshick, P. Noordhuis, L. Wesolowski, A. Kyrola, A. Tulloch, Y. Jia, and K. He. Accurate, Large Minibatch SGD: Training Image Net in 1 Hour. 2017. ar Xiv: 1706.02677. [54] X. Zhai, A. Oliver, A. Kolesnikov, and L. Beyer. S4L: Self-Supervised Semi-Supervised Learning. ICCV. 2019. ar Xiv: 1905.03670. [55] L. Fei-Fei, R. Fergus, and P. Perona. Learning Generative Visual Models from Few Training Examples: An Incremental Bayesian Approach Tested on 101 Object Categories. CVPR Workshop. 2004. [56] M.-E. Nilsback and A. Zisserman. Automated flower classification over a large number of classes. 2008 Sixth Indian Conference on Computer Vision, Graphics & Image Processing. IEEE. 2008, pp. 722 729. [57] A. Krizhevsky. Learning multiple layers of features from tiny images. 2009. URL: https: //www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf. [58] M. Everingham, L. Van Gool, C. K. Williams, J. Winn, and A. Zisserman. The Pascal visual object classes (VOC) challenge. IJCV 88.2 (2010), pp. 303 338. [59] J. Xiao, J. Hays, K. A. Ehinger, A. Oliva, and A. Torralba. SUN database: Large-scale scene recognition from abbey to zoo. CVPR. 2010. [60] O. M. Parkhi, A. Vedaldi, A. Zisserman, and C. V. Jawahar. Cats and Dogs. CVPR. 2012. [61] M. Cimpoi, S. Maji, I. Kokkinos, S. Mohamed, and A. Vedaldi. Describing Textures in the Wild. CVPR. 2013. ar Xiv: 1311.3618. [62] J. Krause, J. Deng, M. Stark, and L. Fei-fei. Collecting a large-scale dataset of fine-grained cars. The Second Workshop on Fine-Grained Visual Categorization. 2013. [63] L. Bossard, M. Guillaumin, and L. Van Gool. Food-101 mining discriminative components with random forests. ECCV. Springer. 2014, pp. 446 461. [64] S. Maji, E. Rahtu, J. Kannala, M. B. Blaschko, and A. Vedaldi. Fine-Grained Visual Classification of Aircraft. ar Xiv: 1306.5151. [65] S. Kornblith, J. Shlens, and Q. V. Le. Do Better Image Net Models Transfer Better? CVPR. 2019. ar Xiv: 1805.08974. [66] N. Silberman, D. Hoiem, P. Kohli, and R. Fergus. Indoor Segmentation and Support Inference from RGBD Images. ECCV. 2012. [67] T.-Y. Lin, M. Maire, S. J. Belongie, L. D. Bourdev, R. B. Girshick, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick. Microsoft COCO: Common Objects in Context. ECCV. 2014. ar Xiv: 1405.0312. [68] K. Yang, K. Qinami, L. Fei-Fei, J. Deng, and O. Russakovsky. Towards fairer datasets: Filtering and balancing the distribution of the people subtree in the imagenet hierarchy. Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency. 2020, pp. 547 558. [69] Y. Asano, C. Rupprecht, A. Zisserman, and A. Vedaldi. PASS: An Image Net replacement for self-supervised pretraining without humans (2021). [70] D. Hendrycks, M. Mazeika, S. Kadavath, and D. Song. Using self-supervised learning can improve model robustness and uncertainty. ar Xiv preprint ar Xiv:1906.12340 (2019). [71] A. Sablayrolles, M. Douze, C. Schmid, Y. Ollivier, and H. Jégou. White-box vs black-box: Bayes optimal strategies for membership inference. International Conference on Machine Learning. PMLR. 2019, pp. 5558 5567. [72] D. Yu, H. Zhang, W. Chen, J. Yin, and T.-Y. Liu. How Does Data Augmentation Affect Privacy in Machine Learning? Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 35. 12. 2021, pp. 10746 10753. [73] L. Grafakos. Classical Fourier Analysis. Vol. 2. Springer, 2008. [74] R. Piessens. The Hankel transform. The Transforms and Applications Handbook 2.9 (2000). [75] F. Johansson et al. mpmath: a Python library for arbitrary-precision floating-point arithmetic (version 0.18). http://mpmath.org/. 2013. [76] J. Long, E. Shelhamer, and T. Darrell. Fully Convolutional Networks for Semantic Segmentation. CVPR. 2015. ar Xiv: 1411.4038. [77] O. J. Hénaff, S. Koppula, J.-B. Alayrac, A. van den Oord, O. Vinyals, and J. Carreira. Efficient Visual Pretraining with Contrastive Detection. 2021. ar Xiv: 2103.10957. [78] K. He, G. Gkioxari, P. Dollár, and R. B. Girshick. Mask R-CNN. ICCV. 2017. ar Xiv: 1703.06870. [79] T.-Y. Lin, P. Dollár, R. B. Girshick, K. He, B. Hariharan, and S. J. Belongie. Feature Pyramid Networks for Object Detection. CVPR. 2017. ar Xiv: 1612.03144.