# selfsupervised_representation_learning_with_meta_comprehensive_regularization__ac7174bc.pdf Self-Supervised Representation Learning with Meta Comprehensive Regularization Huijie Guo1,*, Ying Ba4,5,*, Jie Hu6, Lingyu Si2,3, Wenwen Qiang2,3, , Lei Shi1, 1Beihang University 2Institute of Software Chinese Academy of Sciences 3University of Chinese Academy of Sciences 4Gaoling School of Artificial Intelligence, Renmin University of China 5Beijing Key Laboratory of Big Data Management and Analysis Methods 6Meituan {guo hj, leishi}@buaa.edu.cn, yingba@ruc.edu.cn, hujie@ios.ac.cn, {lingyu, qiangwenwen}@iscas.ac.cn Self-Supervised Learning (SSL) methods harness the concept of semantic invariance by utilizing data augmentation strategies to produce similar representations for different deformations of the same input. Essentially, the model captures the shared information among multiple augmented views of samples, while disregarding the non-shared information that may be beneficial for downstream tasks. To address this issue, we introduce a module called Comp Mod with Meta Comprehensive Regularization (MCR), embedded into existing selfsupervised frameworks, to make the learned representations more comprehensive. Specifically, we update our proposed model through a bi-level optimization mechanism, enabling it to capture comprehensive features. Additionally, guided by the constrained extraction of features using maximum entropy coding, the self-supervised learning model learns more comprehensive features on top of learning consistent features. In addition, we provide theoretical support for our proposed method from information theory and causal counterfactual perspective. Experimental results show that our method achieves significant improvement in classification, object detection and instance segmentation tasks on multiple benchmark datasets. Introduction Deep learning models have exhibited remarkable capabilities, leading to the widespread adoption of machine learning across diverse fields. Despite the impressive performance of supervised learning methods, their heavy reliance on labeled data for model training poses limitations on their generalization ability and scalability. To address this challenge, Self-Supervised Learning (SSL) has emerged as a promising paradigm that bridges the gap between supervised and unsupervised learning by generating supervised signals directly from the samples without the need for manual annotation. Currently, SSL has achieved remarkable results in computer vision (Tian, Krishnan, and Isola 2020; Chen, Xie, and He *These authors contributed equally. Corresponding author. Co-corresponding author. Copyright 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Figure 1: Loss of task-related information caused by data augmentation in SSL methods. (a), the positive sample pair (xa, xb) can be obtained from the input x by Random Cropping and Cutout. (b) formally presents the semantics related to label in different augmented views, where h( ) represents the amount of attributes related to the label in sample. 2021; Caron et al. 2020) and natural language processing (Baevski et al. 2020; Akbari et al. 2021; Zhou et al. 2020). The general framework of self-supervised representation learning consists of two key components: data augmentation and loss function, which try to learn invariance to the transformation generated by data augmentation on the same sample while maintain discrimination to different samples. In practice, data augmentation generates two augmented views of the same image by applying random strategies, such as Cutout (De Vries and Taylor 2017), Coloring (Zhang, Isola, and Efros 2016), Random Cropping (Takahashi, Matsubara, and Uehara 2019), etc. Several studies (Zheng et al. 2021; Shorten and Khoshgoftaar 2019; Zhang and Ma 2022; Tian et al. 2020) also have suggested that not all data augmentations are beneficial for downstream tasks. For instance, rotation invariance may help some flower categories but harm animal recognition (Xiao et al. 2020). Similarly, color invariance may have opposite effects on animal and flower classification tasks. Therefore, recent works have proposed adaptive augmentation strategies to adapt to different data and task environments (Li et al. 2022a; Yang et al. 2022). Data augmentation strategies are widely used in SSL to create positive pairs of images that share the same label. However, these strategies may not preserve all the semantic information that is relevant to the label in the augmented views. For example, suppose an image s label is bird and The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) it only refers to the foreground object, not the background. Figure 1(a) shows two views of the same image x created by Random Cropping and Cutout, denoted as x1 and x2. Note that x1 contains the bird s beak while x2 does not, and x2 contains the bird s wings while x1 does not. A common assumption in SSL is that the semantic content of an image should be invariant to the applied transformations. However, this assumption can be broken by the transformation methods and may not hold for all label-related attributes, such as the bird s beak and wings. Figure 1(b) illustrates a function h( ) that measures the amount of label-related attributes in an image. The representations learned by SSL methods are based on the shared information between different augmented views, such as the Intersection over Union (Io U). However, this shared information may not capture the entire foreground of the input, and some label-related attributes may be dropped in the model training process. The more label-related information is preserved in the training process, the better the model can learn. Therefore, models trained using traditional SSL methods may exhibit subpar performance in downstream tasks due to the loss of labelrelated information during the training process. To address the aforementioned issue, we propose utilizing a more comprehensive representation to guide the training of SSL model, enabling the model to focus on non-shared semantic information that might be beneficial for downstream tasks, thereby enhancing model s generalization capability. We propose a plug-and-play module called Comp Mod with Meta Comprehensive Regularization to guide the learning of SSL methods by obtaining comprehensive features. Specifically, we employ semantic complementarity to fuse augmented features in a low-dimensional space, utilizing a bilevel optimization mechanism to obtain comprehensive representation that guide the learning of SSL methods. Our contributions are the following: From the information theory, we analyze that data augmentation in SSL may lead to the lack of task-related information, which in turn reduces the generalization ability of the model. We design a plug-and-play module, called Comp Mod, to induce existing SSL methods to learn comprehensive feature representations. Comp Mod ensures comprehensive feature exploration through a bi-level optimization mechanism and constrained extraction of features with maximum entropy coding, guaranteeing complete mining of feature completeness. A causal counterfactual analysis provides theoretical support for our proposed method. Empirical evaluations of the proposed method substantiate its superior performance in classification, object detection and instance segmentation tasks. Related Work Recently, various frameworks have emerged for selfsupervised representation learning, which can be broadly classified into two types (Garrido et al. 2022; Balestriero and Le Cun 2022): sample-based and dimension-based contrastive learning methods. Sample-based contrastive methods learn visual representations by constructing pairs of samples and applying contrastive loss function. These methods encourage the embeddings of augmented views of the same image to be close to each other, while simultaneously pushing away the embeddings of different images. Some notable methods, such as Sim CLR (Chen et al. 2020), utilize Info NCE as the loss function and rely on the quality and quantity of negative samples. However, these methods also necessitate greater computational resources. Mo Co (He et al. 2020) tackles this issue by constructing a dynamic dictionary bank that expands the pool of available negative samples. On the other hand, some studies have investigated whether SSL can still work without negative samples. BYOL (Grill et al. 2020) and Sim Siam (Chen and He 2021) utilize a distillation-like mechanism to learn representations by computing the similarity between positives, without the need for negative samples. Dimension-based contrastive methods learn visual representations by optimizing the information content of the learned representations and reducing feature redundancy. Barlow Twins (Zbontar et al. 2021) endeavors to make the normalized cross-correlation matrix of the augmented embeddings close to the identity matrix. The loss function of VICReg (Bardes, Ponce, and Lecun 2022) consists of three items: invariance, variance and covariance regularization item. TCR (Li et al. 2022b) employs the Maximum Coding Rate Reduction (MCR2) objective to learn feature subspaces that are both informative and discriminative. Liu (Liu et al. 2022) proposed using maximum entropy coding for contrastive learning, based on the principle of maximum entropy in information theory, and established a connection between sample-based and dimension-based SSL. These works mentioned above are based on the invariance of semantic among augmented views, while ignoring the partial loss of label-related information in each view after augmentation, leading to imperfect consistency in semantic information across views. By leveraging the comprehensive information between views, our work allows the feature extractor to gather more abundant information, thereby inducing the learned sample representations to be more generalizable. Our proposed Meta Comprehensive Regularization can be integrated into existing SSL framework. Methodology Figure 2 shows the overview of our proposed method. We design a new module, Comp Mod, to improve existing selfsupervised method. Next, we first theoretically analyze the lack of partial semantic information caused by data augmentation is not conducive to downstream tasks in SSL from an information-theoretic perspective, and then introduce our proposed method and the training process of the model. Contrastive Learning Let D = {xi}n i=1 denote the unlabeled training set, where xi is an input image. Two augmented views x1 i and x2 i of the sample xi are generated by different augmentation strategies t1 and t2 sampled from a augmentation distribution A. The augmented views are fed into a shared encoder fθ to obtain their representations h1 i = fθ(x1 i ) and The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Figure 2: Illustration of self-supervised representation learning framework with Meta Comprehensive Regularization. h2 i = fθ(x2 i ), which are then mapped via a projector gϕ onto the embedding space, z1 i = gϕ(h1 i ) and z2 i = gϕ(h2 i ). We denote the embedding matrix of augmentation view 1 as Z1 = [z1 1, ..., z1 i , ..., z1 n]T Rn d, where d is the dimension of the embedding space, so does matrix Z2. Represented by Sim CLR, the objective function of the Contrastive Learning (CL) employs the Noise Contrastive Estimation (NCE) loss (Gutmann and Hyv arinen 2010): Lssl = Ez1 i ,z2 i [ log e s(z1 i ,z2 i )/τ e s(z1 i ,z2 i )/τ + P zj e s(z1 i ,zj )/τ ] (1) where s( , ) denotes the cosine similarity and τ represents the temperature hyper-parameter, zj is the negative sample, zj Z1 Z2/{z1 i , z2 i }, Z1 and Z2 represent the sets of augmented views in the embedding space, respectively. Analysis Based on Information Theory We assume that the original input images inherently encompass all relevant task-related information, e.g., I(x; T) = H(T), where x D is a random variable, I denotes the mutual information, H represents the information entropy, and T refers to a random variable for the downstream task. As evident from Figure 1, data augmentation on the input sample results in loss of task-relevant information within the data. Consequently, we deduce: I(x1; T), I(x2; T) H(T), where (x1, x2) {(x1 i , x2 i )}n i=1. Also, we can obtain: I(x1; x2; T) H(T). A general explanation for CL is to maximize the mutual information between two augmented views(Wang et al. 2022): max f,g I(z1; z2) (2) where z1 and z2 are random variables, (z1, z2) (Z1, Z2). Applying Data Processing Inequality (Klir and Wierman 1999) to the Markov chain x x1(x2) z1(z2), we have: H(x) I(x1; x2) I(z1; z2) I(x; T) I(x1; x2; T) I(z1; z2; T) (3) Based on Eq. 2 and Eq. 3, we can draw the conclusion that due to the disruption caused by data augmentation to the semantic information of input samples, contrastive learning is constrained to extract only a subset of task-related information. In order to elucidate this conclusion, we begin by providing a definition for comprehensive representation, followed by deriving the following theorem. Definition 1. (Comprehensive representation) For a random variable z defined in encoder space. z is a comprehensive representation if and only if I(z; T) = H(T). Theorem 1. (Task-Relevant information in representations) In contrastive learning, given a random variable x representing the original sample space, two random variables x1 and x2 characterizing the sample space after augmentation, and two random variable z1 and z2 denoting the augmented samples within the feature space, we have: H(T) {I(x1; T), I(x2; T)} I(x1; x2; T) H(T) {I(z1; T), I(z2; T)} I(z1; z2; T) (4) From Theorem 1, we deduce the following: (1) Data augmentation leads to a reduction in the amount of task-related information present within the data. (2) The task-related information contained in the commonalities between z1 and z2 is individually less than the task-related information within z1 and z2. Obviously, if the representation of the augmented view is close to the comprehensive representation, it is more beneficial to the downstream tasks. Next, we propose to use Meta Comprehensive Regularization to force the augmented representation to be a comprehensive representation. Meta Comprehensive Regularization To address the issue of semantic loss resulting from data augmentation, we propose a new module called Comp Mod, which learns a more comprehensive representation and helps facilitate model learning, as shown in Figure 2. This module can be directly incorporated into the traditional SSL framework and can complement existing SSL methods. Then, we introduce the module Comp Mod. Different augmentations of the same sample are typically derived through various data augmentation techniques, empirically implying that each augmentation results in distinct The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) semantic information loss. As a result, a method that yields a comprehensive feature representation without semantic information loss, distinct from the original input data representation, involves integrating features from different perspectives of the same sample. Thus, a more comprehensive representation of xi can be obtained by the following formula: ˆhi := h1 i h2 i (5) where represents the fusion strategy, such as the concatenation of vectors: ˆhi = h1 i h2 i R2d , where d is the output dimension of a backbone network. ˆhi fuses the semantic information from all augmented views, so as to solve the problem of partial semantic missing. Next, a projection head gξ parameterized by ξ maps ˆhi onto the same embedding space as z1 i and z2 i . And then, we obtain the so-called more comprehensive embedding of sample xi, denoted as ˆzi = gξ(ˆhi) Rd. The comprehensive embedding matrix is defined as ˆZ = [ˆz1, ..., ˆzi, ..., ˆzn]T Rn d. Simple fusion alone does not guarantee that the learned features encompass all semantics. Inspired by the maximum entropy principle in information theory, a generalizable representation should be the one with the maximum entropy among all possible representations, corresponding to the maximization of the semantic information associated with the true label. Here, we use the code length of the lossy data coding (Cover 1999) to calculate the entropy of the embedding matrix ˆZ, which apply the Taylor series expansion: Lcomp( ˆZ) = Tr(µ k (λ ˆZ ˆZT )k) (6) where k is the order of Taylor expansion, µ and λ are hyperparameters. Therefore, to further ensure the comprehensiveness of the obtained ˆZ, we propose that the obtained ˆZ should minimize Lcomp( ˆZ). Next, we use the comprehensive representation to guide the learning of the backbone network. To enhance the semantic richness of the representation in each augmented view, we constrain the information contained in the augmented embeddings Z1, Z2 to equal the information contained in the comprehensive embedding ˆZ. We propose to minimize the following loss: Lmcr(Z1, Z2) = k (λ ˆZZT t )k) (7) where Zt is the embedding matrix of augmented view, t = 1, 2. As we can see, when ˆZ is predetermined and carries maximal information content, in order to minimize Eq. 7, it is necessary for Zt to be equal to ˆZ. Thus, minimizing Eq. 7 can be considered as a conduit for transferring comprehensive information from ˆZ to both Z1 and Z2, enabling them to compensate for the semantic loss incurred by data augmentation. Consequently, while extracting consistent semantic information from Z1 and Z2, minimizing Eq. 7 facilitates the extraction of comprehensive semantic information. Algorithm 1: The main algorithm Input: Training set D; Batch Size n; Encoder function fθ; Projection Head gϕ; Multi-layer Network gξ. Parameter: Regularization Parameter: λ1, λ2 Output: The optimal encoder: fθ 1: for sample batch X from D do 2: # generate two augmented views 3: x1 i , x2 i = t1(xi), t2(xi), t1, t2 A, xi X 4: # obtain the augmented embeddings 5: z1 i = gϕ(fθ(x1 i )) 6: z2 i = gϕ(fθ(x2 i )) 7: # obtain the more comprehensive embedding 8: ˆzi = gξ(ˆhi), where ˆhi = h1 i h2 i 9: # Under the fixed ξ, update {θξ, ϕξ} 10: {θξ, ϕξ} = {θ, ϕ} r θ,ϕ(Lssl + λ1Lmcr) 11: # Under the fixed {θξ, ϕξ}, update ξ 12: ξ = ξ r ξ(Lssl(θξ, ϕξ) + λ2Lcomp) 13: end for Model Objective Finally, we present the objective during the training phase, which can be divided into two steps. The first step is to learn fθ and gϕ that can extract feature representation. The second step is to learn gξ that can obtain comprehensive representation by a bi-level optimization mechanism. The training process is shown in Algorithm 1. Specifically, in the first step of each epoch, we fix gξ and update fθ and gϕ through the following formulation: {θ, ϕ} = {θ, ϕ} r θ,ϕ(Lssl + λ1Lmcr) (8) where r is the learning rate and λ1 is the hyperparameter. The purpose of learning Eq. 8 is to extract consistency semantic information between Z1 and Z2. However, Lssl + λ1Lmcr in Eq. 8 enables both Z1 and Z2 to simultaneously possess comprehensive semantic information. Thus, learning Eq. 8 can result in the consistency information between Z1 and Z2 being comprehensive information. In the second step of each epoch, we fix fθ and gϕ and update gξ through the following formulation: ξ = ξ r ξ(Lssl(θξ, ϕξ) + λ2Lcomp) s.t.{θξ, ϕξ} = {θ, ϕ} r θ,ϕ(Lssl + λ1Lmcr) (9) where Lssl(θξ, ϕξ) represents that the loss Lssl is calculated based on fθξ and gϕξ, and λ2 is the hyperparameter. It is important to note that during the computation of Lssl, gξ is not involved, hence direct differentiation of Lssl with respect to ξ is not possible. However, when computing matrix Lssl(θξ, ϕξ), as indicated by Eq. 9, θξ and ϕξ can be treated as functions of ξ, allowing for direct differentiation of Lssl(θξ, ϕξ) with respect to ξ. Simultaneously, optimizing ξLssl(θξ, ϕξ) can be conceptualized as follows: by manipulating ξ to induce changes in Lssl(θξ, ϕξ), constrained by the conditions outlined in Eq. 9, where Lssl(θξ, ϕξ) is consistently optimized under these ξ-induced circumstances. Subsequently, among all optimal states of Lssl(θξ, ϕξ), the objective is to identify a configuration that minimize the magnitude of ξLssl(θξ, ϕξ). So, optimizing The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Figure 3: Structural causal model of latent variables. We assume that part of the semantic missing in the augmented view x1 and x2 compared to the original view x. These exclusive semantic c is affected by the semantic c and the applied augmentation strategy t. ξLssl(θξ, ϕξ) can be simply regarded as the mechanism for modulating ξ to reduce its magnitude, thereby enhancing the similarity between matrices Z1 and Z2. As widely acknowledged, it is only when matrices Z1 and Z2 encapsulate a greater abundance of analogous semantic information that matrices Z1 and Z2 can exhibit increased similarity. Therefore, optimizing ξLssl(θξ, ϕξ) can compel gξ to extract a greater amount of semantic information. When coupled with the influence of Lcomp, this process enables Eq. 9 to coerce gξ into capturing comprehensive semantic information. Causal Interpretation We use a Structural Causal Model (SCM) (Pearl and Mackenzie 2018) to describe the causal relationship between variables in self-supervised learning, whereby data augmentation can be viewed as a counterfactual. We define c to be the semantic of original sample, c is the invariant semantics across augmented views, c is the exclusive semantics in the augmented view. We assume that the samples are only generated by their semantic information (or label) and exogenous variables, as shown in Fig.3. The augmentation strategy t is randomly sampled from the strategy set T . We can formulate the relation between these variables: c := fc(uc), c := f c(c, t), c := f c(c, t), x := f(c, ux); where uc, ux is independent exogenous variable, and f c, f c are deterministic functions. Given a factual observation x = f(c , ux ). If a new augmentation strategy t is applied, it is equivalent to an intervention mechanism affecting f c, f c: do( c := f c(c , t ), c := f c(c , t )) Using the modified SCM by fixing all exogenous variables, the relation between the variables in the counterfactual sample x = f(c , ux ) and its augmented views can be reformulated as: c := c , c := f c(c , t ), c := f c(c , t ); Thus, data augmentation in self-supervised learning can be viewed as counterfactual. Then, we can obtain: Theorem 2. Assume that the data generating process is consistent with the above description. Let f : x z be any smooth function that can minimize the following objective: E(z1 i ,z2 i ) (Z1,Z2) h z1 i z2 i 2 2 i Hmin(zt i) (10) where Hmin( ) denotes the minimum entropy of all augmented embeddings, zt i = {z1 i , z2 i }. In this way, the learned encoder fθ can capture all semantic information related to the original sample. The above theorem states that when we constrain the embeddings to have the maximum entropy, the SSL model can constrain the learned representation to contain more semantics. Note that the Eq. 10 is similar to Eq. 9. Therefore, this theorem also provides a theoretical basis for the proposed bi-level learning method. Experimental Results In this section, we first evaluate it on classification task using linear evaluation and semi-supervised learning settings. Then, we validate our method on object detection and instance segmentation tasks in computer vision. Experimental Setting Datasets. For the classification task, we evaluate our proposed method on the following six image datasets, including CIFAR-10 and CIFAR-100 dataset (Krizhevsky 2009), STL-10 dataset (Coates, Ng, and Lee 2011), Tiny Image Net dataset (Le and Yang 2015), Image Net-100 dataset (Russakovsky et al. 2015), and Image Net dataset (Russakovsky et al. 2015). For transfer learning, we validate our method by the performance on the object detection and semantic segmentation tasks on COCO (Lin et al. 2014) dataset. Default Setting. Each input sample generates two corresponding positive samples in the experiment. The image augmentation strategies comprise the following image transformations: random cropping, resizing, horizontal flipping, color jittering, converting to grayscale and gaussian blurring. Detailed experimental settings for different downstream tasks can be found in Appendix B. In the experiment, we use Resnet18 or Resnet50 as our base encoder network, along with a 3-layer MLP projection head to project the representation to a embedding space. Comp Mod Details. The Comp Mod consists of a multilayer linear network, which is set to 2d d d, where d is the output dimension of the backbone network fθ. Downstream Tasks Self-supervised Learning We conduct a classification task to test our proposed method. For comparison, we take Sim CLR (Chen et al. 2020), Barlow Twins (Zbontar et al. 2021), BYOL (Grill et al. 2020), Sim Siam (Chen and He 2021), W-MES (Ermolov et al. 2021), Sw AV (Caron et al. 2020), Mo Co (He et al. 2020), CMC (Tian, Krishnan, and Isola 2020), SSL-HSIC (Li et al. 2021) and VICReg (Bardes, Ponce, and Lecun 2022) as baselines. We validate the proposed method with the results of a linear classifier and a 5-nearest neighbor classifier. Table 1 shows the The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Methods CIFAR-10 CIFAR-100 STL-10 Tiny Image Net linear 5-nn linear 5-nn linear 5-nn linear 5-nn Sim CLR 91.80 88.42 66.83 56.56 90.51 85.68 48.82 32.86 Barlow Twins 90.88 88.78 66.67 56.39 90.71 85.31 49.74 33.61 BYOL 91.73 89.45 66.60 56.82 91.99 88.64 51.00 36.24 Sim Siam 91.51 89.31 66.73 56.87 91.92 88.54 50.92 35.98 W-MSE 91.99 89.87 67.64 56.45 91.75 88.59 49.22 35.44 Sw AV 90.17 86.45 65.23 54.77 89.12 84.12 47.13 31.07 SSL-HSIC 91.95 89.91 67.22 57.01 92.06 88.87 51.42 36.03 VICReg 91.08 88.93 67.15 56.47 91.11 86.24 50.17 34.24 Sim CLR+ 93.91 91.21 68.91 58.22 92.77 87.98 51.01 35.23 BYOL+ 93.96 91.53 68.74 58.01 94.95 89.88 53.51 37.95 Barlow Twins+ 92.54 90.75 68.29 57.84 93.12 89.61 51.72 35.21 Table 1: Classification accuracy on small and medium datasets. Top 1 accuracy(%) of linear classifier and a 5-nearest neighbors classifier for different datasets with a Res Net-18. Best results are in bold. Methods Image Net-100 Image Net top-1 top-5 top-1 top-5 Sim CLR 70.15 89.75 69.32 89.15 Mo Co 72.81 91.64 71.13 - CMC 73.58 92.06 66.21 87.03 BYOL 74.89 92.83 74.31 91.62 Sw AV 75.77 92.86 75.30 - DCL 74.60 92.08 - - RELIC - - 74.81 92.23 SSL-HSIC - - 72.13 90.33 ICL-MSR 72.08 91.60 70.73 90.43 Barlow Twins 72.88 90.99 73.22 91.01 Sim CLR+ 72.21 91.23 71.89 91.52 BYOL+ 76.95 93.94 75.11 93.55 Barlow Twins+ 76.88 94.11 75.62 92.13 Table 2: Evaluation on Image Net-100 and Image Net datasets. The representations are obtained with a Res Net18 with our method on top 1 accuracy(%) of linear classifier and a 5-nn classifier. Best results are in bold. performance of different SSL methods, where method+ denotes our proposed method. The results show that our proposed method improves the classification performance, in which Sim CLR+ and BYOL+ improve by more than 2% on CIFAR10 and CIFAR100 dataset, while BYOL+ improves by about 2.5% on Tiny Image Net dataset. Furthermore, we test our method for classification on two larger datasets, Image Net-100 and Image Net. For comparison, we also add several other methods including Mo Co (He et al. 2020), CMC (Tian, Krishnan, and Isola 2020), ICLMSR (Qiang et al. 2022) and RELIC (Mitrovic et al. 2020). The results in Table 2 demonstrate that our method still improves over the baseline, e.g., Barlow Twins+ achieves 4% performance improvement on Image Net-100, Sim CLR+ and Barlow Twins+ achieve more than 2% on Image Net. Semi-supervised Learning The detailed experimental setup follows the most common evaluation protocol for semi-supervised learning, as in Appendix B. Table 3 reports Methods Epochs 1% 10% top-1 top-5 top-1 top-5 Sim CLR 1000 48.3 75.5 65.6 87.8 BYOL 1000 53.2 78.4 68.8 89.0 Sw AV 1000 53.9 78.5 70.2 89.9 Barlow Twins 1000 55.0 79.2 69.7 89.3 Sim CLR+ 1000 49.1 75.8 65.8 88.0 BYOL+ 1000 54.6 78.9 69.2 89.3 Barlow Twins+ 1000 56.1 79.8 70.2 89.9 Table 3: Semi-supervised classification. We finetune the pretrained model using 1% and 10% training samples of Image Net following (Zbontar et al. 2021), and the top-1 and top-5 under linear evaluation are reported. the classification results on Image Net compared with existing methods using two pre-trained models. From the results, Barlow Twins+ is 1.1% better than Barlow Twins, and BYOL+ increases by about 1.4% at the 1% subset setting. Transfer Learning We evaluate our method for the localization based tasks of object detection and instance segmentation on COCO (Lin et al. 2014) datasets. Image Net supervised pre-training is often used as initialization for finetuning downstream tasks. Several different self-supervised methods are used for performance comparison. We report the results of our proposed method compared with baselines in Table 5, showing that the proposed method brings performance improvements on different downstream tasks. Ablation Experiments Parametric Sensitivity In this section, we conduct an experimental investigation of the model trade-off parameters. Specifically, we vary λ1 and λ1 in the range of [0.001, 0.01, 0.1, 1], and record the classification accuracy of our method using a Res Net-18 on CIFAR-10 dataset with the Sim CLR+ method. The results in Table 6 indicates that our method has minimal variation in accuracy, indicating that hyperparameter tuning is easy in practice. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) ID Data augmentations Methods horizontal flip rotate random crop random grey color jitter Sim CLR+ BYOL+ Barlow Twins+ 1 93.65 92.64 91.75 2 92.31 92.78 92.09 3 92.78 93.16 92.15 4 93.36 92.99 91.95 5 93.47 92.74 92.23 6 93.72 92.89 91.89 7 93.75 93.78 92.48 8 93.91 93.96 92.54 Table 4: Comparison of different data augmentations by using a Res Net-18 on CIFAR-10 dataset. Methods Object Det. Instance Seg. AP AP50 AP75 AP AP50 AP75 Supervised 38.2 58.2 41.2 33.3 54.7 35.2 Sim CLR 37.9 57.7 40.9 33.2 54.6 35.3 Sw AV 37.6 57.6 40.2 33.0 54.2 35.1 BYOL 37.9 57.8 40.9 33.1 54.3 35.0 Sim Siam 37.9 57.5 40.9 33.3 54.2 35.2 Barlow Twins 39.2 59.0 42.5 34.2 56.0 36.5 Sim CLR+ 38.1 58.1 41.0 33.7 54.6 35.1 BYOL+ 39.7 59.1 42.9 35.4 56.1 36.2 Barlow Twins+ 39.1 59.3 43.1 35.2 56.2 36.9 Table 5: Transfer learning. We pre-train the network on Image Net dataset. Then, we learn representation on the object detection and instance segmentation tasks on COCO dataset using Mask P-CNN. Evaluation is on AP, AP50 and AP75. Analysis of Data Augmentation we compare the linear classification accuracy under different augmentation strategies on CIFAR-10 dataset as shown in Table 4. As can be seen, there is no significant difference in classification accuracy, indicating that our method can be applied to different augmentation strategies. Fusion Strategy and Optimization In this section, we first investigate the impact of different fusion strategies. Mixup (Verma et al. 2021) can be used to fuse features in representation space. We can obtain the more comprehensive representation using the following formula: ˆhi = α h1 i + (1 α) h2 i (11) where α is a coefficient sampled from a uniform distribution, α U(0, 1). By adjusting α, we can control the semantic information to be biased towards h1 i or h2 i . Another strategy is to achieve semantic fusion in the embedding space: zi := z1 i z2 i = z1 i z2 i (12) Then zi is mapped onto the embedding space to obtain ˆzi via a projection head gζ composed of a multi-layer linear network (2d d d): ˆzi = gζ( zi). Additionally, our proposed method utilizes a bi-level optimization mechanism for model optimization during training. λ2 λ1 0.001 0.01 0.1 1 0.001 91.79 92.13 92.77 91.75 0.01 92.56 91.89 93.91 91.11 0.1 93.35 92.23 92.35 90.78 1 93.03 91.26 92.77 90.08 Table 6: Parametric analysis of λ1 and λ2. Mixup M(h) M(z) α 0.1 0.3 0.5 0.7 0.9 - - Acc. 91.03 91.33 91.87 91.34 91.56 93.96 92.58 No bi-level 92.05 Table 7: Analysis of Fusion Strategy and Optimization. Experimental results are based on the classification with BYOL+. M(h) means fusion in representation space, while, M(z) means fusion in embedding space In this section, we also analyze the impact of not using this strategy in experiments. Table 7 shows the results of BYOL+ utilizing different feature fusion strategies and optimization on the classification task on CIFAR-10 dataset. Experimental results demonstrate that the fusion strategy and optimization we adopt achieve the best results. In this paper, we find that data augmentation in SSL may lead to the lack of task-related information from information theory,resulting in a reduction of the model s performance in downstream tasks. To this end, we design a novel module Comp Mod with Meta Comprehensive Regularization as a complement to existing SSL frameworks. Comp Mod exploits a bi-level optimization mechanism and constraint based on maximum entropy coding to enable more information to be discovered, thereby enhancing the generalization of the learned model. Moreover, a causal interpretation provide theoretical support for the proposed method. Finally, the performance of various downstream tasks validates the effectiveness of our proposed method. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Acknowledgements This work was supported by China Postdoctoral Science Foundation under Grant 2023M743639, National Key R&D Program of China (2021YFB3500700), NSFC Grant 62172026, National Social Science Fund of China 22&ZD153, the Fundamental Research Funds for the Central Universities and SKLSDE. References Akbari, H.; Yuan, L.; Qian, R.; Chuang, W.-H.; Chang, S.-F.; Cui, Y.; and Gong, B. 2021. Vatt: Transformers for multimodal self-supervised learning from raw video, audio and text. Advances in Neural Information Processing Systems, 34: 24206 24221. Baevski, A.; Zhou, Y.; Mohamed, A.; and Auli, M. 2020. wav2vec 2.0: A framework for self-supervised learning of speech representations. Advances in neural information processing systems, 33: 12449 12460. Balestriero, R.; and Le Cun, Y. 2022. Contrastive and noncontrastive self-supervised learning recover global and local spectral embedding methods. Advances in Neural Information Processing Systems, 35: 26671 26685. Bardes, A.; Ponce, J.; and Lecun, Y. 2022. VICReg: Variance-Invariance-Covariance Regularization For Self Supervised Learning. In ICLR 2022-International Conference on Learning Representations. Caron, M.; Misra, I.; Mairal, J.; Goyal, P.; Bojanowski, P.; and Joulin, A. 2020. Unsupervised learning of visual features by contrasting cluster assignments. Advances in Neural Information Processing Systems, 33: 9912 9924. Chen, T.; Kornblith, S.; Norouzi, M.; and Hinton, G. 2020. A simple framework for contrastive learning of visual representations. In International conference on machine learning, 1597 1607. PMLR. Chen, X.; and He, K. 2021. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 15750 15758. Chen, X.; Xie, S.; and He, K. 2021. An empirical study of training self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 9640 9649. Coates, A.; Ng, A.; and Lee, H. 2011. An analysis of singlelayer networks in unsupervised feature learning. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, 215 223. JMLR Workshop and Conference Proceedings. Cover, T. M. 1999. Elements of information theory. John Wiley & Sons. De Vries, T.; and Taylor, G. W. 2017. Improved regularization of convolutional neural networks with cutout. ar Xiv preprint ar Xiv:1708.04552. Ermolov, A.; Siarohin, A.; Sangineto, E.; and Sebe, N. 2021. Whitening for self-supervised representation learning. In International Conference on Machine Learning, 3015 3024. PMLR. Garrido, Q.; Chen, Y.; Bardes, A.; Najman, L.; and Lecun, Y. 2022. On the duality between contrastive and non-contrastive self-supervised learning. ar Xiv preprint ar Xiv:2206.02574. Grill, J.-B.; Strub, F.; Altch e, F.; Tallec, C.; Richemond, P.; Buchatskaya, E.; Doersch, C.; Avila Pires, B.; Guo, Z.; Gheshlaghi Azar, M.; et al. 2020. Bootstrap your own latenta new approach to self-supervised learning. Advances in neural information processing systems, 33: 21271 21284. Gutmann, M.; and Hyv arinen, A. 2010. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, 297 304. JMLR Workshop and Conference Proceedings. He, K.; Fan, H.; Wu, Y.; Xie, S.; and Girshick, R. 2020. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 9729 9738. Klir, G.; and Wierman, M. 1999. Uncertainty-based information: elements of generalized information theory, volume 15. Springer Science & Business Media. Krizhevsky, A. 2009. Learning Multiple Layers of Features from Tiny Images. Master s thesis, University of Tront. Le, Y.; and Yang, X. 2015. Tiny imagenet visual recognition challenge. CS 231N, 7(7): 3. Li, J.; Qiang, W.; Zheng, C.; Su, B.; and Xiong, H. 2022a. Metaug: Contrastive learning via meta feature augmentation. In International Conference on Machine Learning, 12964 12978. PMLR. Li, Y.; Pogodin, R.; Sutherland, D. J.; and Gretton, A. 2021. Self-supervised learning with kernel dependence maximization. Advances in Neural Information Processing Systems, 34: 15543 15556. Li, Z.; Chen, Y.; Le Cun, Y.; and Sommer, F. T. 2022b. Neural manifold clustering and embedding. ar Xiv preprint ar Xiv:2201.10000. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Doll ar, P.; and Zitnick, C. L. 2014. Microsoft coco: Common objects in context. In Computer Vision ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, 740 755. Springer. Liu, X.; Wang, Z.; Li, Y.-L.; and Wang, S. 2022. Selfsupervised learning via maximum entropy coding. Advances in Neural Information Processing Systems, 35: 34091 34105. Mitrovic, J.; Mc Williams, B.; Walker, J.; Buesing, L.; and Blundell, C. 2020. Representation learning via invariant causal mechanisms. ar Xiv preprint ar Xiv:2010.07922. Pearl, J.; and Mackenzie, D. 2018. The book of why: the new science of cause and effect. Basic books. Qiang, W.; Li, J.; Zheng, C.; Su, B.; and Xiong, H. 2022. Interventional Contrastive Learning with Meta Semantic Regularizer. In International Conference on Machine Learning, 18018 18030. PMLR. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. 2015. Imagenet large scale visual recognition challenge. International journal of computer vision, 115(3): 211 252. Shorten, C.; and Khoshgoftaar, T. M. 2019. A survey on image data augmentation for deep learning. Journal of big data, 6(1): 1 48. Takahashi, R.; Matsubara, T.; and Uehara, K. 2019. Data augmentation using random image cropping and patching for deep CNNs. IEEE Transactions on Circuits and Systems for Video Technology, 30(9): 2917 2931. Tian, Y.; Krishnan, D.; and Isola, P. 2020. Contrastive multiview coding. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XI 16, 776 794. Springer. Tian, Y.; Sun, C.; Poole, B.; Krishnan, D.; Schmid, C.; and Isola, P. 2020. What makes for good views for contrastive learning? Advances in neural information processing systems, 33: 6827 6839. Verma, V.; Luong, T.; Kawaguchi, K.; Pham, H.; and Le, Q. 2021. Towards domain-agnostic contrastive learning. In International Conference on Machine Learning, 10530 10541. PMLR. Wang, H.; Guo, X.; Deng, Z.-H.; and Lu, Y. 2022. Rethinking minimal sufficient representation in contrastive learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 16041 16050. Xiao, T.; Wang, X.; Efros, A. A.; and Darrell, T. 2020. What Should Not Be Contrastive in Contrastive Learning. In International Conference on Learning Representations. Yang, K.; Zhou, T.; Tian, X.; and Tao, D. 2022. Identitydisentangled adversarial augmentation for self-supervised learning. In International Conference on Machine Learning, 25364 25381. PMLR. Zbontar, J.; Jing, L.; Misra, I.; Le Cun, Y.; and Deny, S. 2021. Barlow twins: Self-supervised learning via redundancy reduction. In International Conference on Machine Learning, 12310 12320. PMLR. Zhang, J.; and Ma, K. 2022. Rethinking the augmentation module in contrastive learning: Learning hierarchical augmentation invariance with expanded views. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 16650 16659. Zhang, R.; Isola, P.; and Efros, A. A. 2016. Colorful image colorization. In Computer Vision ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 1114, 2016, Proceedings, Part III 14, 649 666. Springer. Zheng, M.; You, S.; Wang, F.; Qian, C.; Zhang, C.; Wang, X.; and Xu, C. 2021. Re SSL: Relational Self-Supervised Learning with Weak Augmentation. Advances in Neural Information Processing Systems, 34. Zhou, W.; Lee, D.-H.; Selvam, R. K.; Lee, S.; and Ren, X. 2020. Pre-training Text-to-Text Transformers for Conceptcentric Common Sense. In International Conference on Learning Representations. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)