# metaug_contrastive_learning_via_meta_feature_augmentation__2f249fa7.pdf Met Aug: Contrastive Learning via Meta Feature Augmentation Jiangmeng Li * 1 2 3 Wenwen Qiang * 1 2 3 Changwen Zheng 1 3 Bing Su 4 5 Hui Xiong 6 7 What matters for contrastive learning? We argue that contrastive learning heavily relies on informative features, or hard (positive or negative) features. Early works include more informative features by applying complex data augmentations and large batch size or memory bank, and recent works design elaborate sampling approaches to explore informative features. The key challenge toward exploring such features is that the source multi-view data is generated by applying random data augmentations, making it infeasible to always add useful information in the augmented data. Consequently, the informativeness of features learned from such augmented data is limited. In response, we propose to directly augment the features in latent space, thereby learning discriminative representations without a large amount of input data. We perform a meta learning technique to build the augmentation generator that updates its network parameters by considering the performance of the encoder. However, insufficient input data may lead the encoder to learn collapsed features and therefore malfunction the augmentation generator. A new margin-injected regularization is further added in the objective function to avoid the encoder learning a degenerate mapping. To contrast all features in one gradient back-propagation step, we adopt the proposed optimization-driven unified contrastive loss *Equal contribution 1Science & Technology on Integrated Information System Laboratory, Institute of Software Chinese Academy of Sciences, Beijing, China 2University of Chinese Academy of Sciences, Beijing, China 3Southern Marine Science and Engineering Guangdong Laboratory (Guangzhou), Guangdong, China. 4Gaoling School of Artificial Intelligence, Renmin University of China, Beijing, China 5Beijing Key Laboratory of Big Data Management and Analysis Methods, Beijing, China 6Thrust of Artificial Intelligence, The Hong Kong University of Science and Technology (Guangzhou), Guangzhou, China 7Department of Computer Science Engineering, The Hong Kong University of Science and Technology, Hong Kong SAR, China. Correspondence to: Bing Su . Proceedings of the 39 th International Conference on Machine Learning, Baltimore, Maryland, USA, PMLR 162, 2022. Copyright 2022 by the author(s). instead of the conventional contrastive loss. Empirically, our method achieves state-of-the-art results on several benchmark datasets. 1. Introduction Contrastive learning methods have achieved empirical success in computer vision (Chopra et al., 2005; Hadsell et al., 2006). Under the setting of self-supervised learning (SSL), recent researches demonstrate the superiority of contrastive methods (Hjelm et al., 2018; Tian et al., 2019; Chuang et al., 2020; Robinson et al., 2020). Typically, these approaches learn features by contrasting different views (e.g., different random data augmentations) of the image in hidden space. We recap the preliminaries of the conventional contrastive learning paradigm: every two views of the same image are considered to be a positive pair, and every two views of the different images are considered to be a negative pair; the contrastive loss (Gutmann & Hyvärinen, 2010; Oord et al., 2018) guides the learned features to bring positive pairs together and push negative pairs farther apart. However, this learning paradigm suffers from the need for a large number of pairs to contrast, e.g., large batch size or memory bank size, because many pairs are not informative to the model, i.e., positive pairs are pretty close and negative pairs are already very apart in hidden space. These pairs have few contributions to the optimization. Contrastive methods need numerous pairs and expect to collect informative ones, and therefore complex data augmentations (e.g., jittering, random cropping, separating color channels, etc.) (Bachman et al., 2019; Chen et al., 2020) and large-scale memory banks (Tian et al., 2019; He et al., 2020) are effective in improving the performance of contrastive models on downstream tasks. The success of the recent works depends on the elaborate selection of informative negative pairs (Chuang et al., 2020; Robinson et al., 2020). These methods focus on designing sampling strategies to assign larger weights to informative pairs, which rely on enough and informative positive pairs and do not need large amounts of negative pairs. When the number of pairs to contrast is limited, the contrastive loss may cause conventional contrastive learning approaches to learn collapsed features (Zbontar et al., 2021; Grill et al., Met Aug: Contrastive Learning via Meta Feature Augmentation 2020), e.g., outputting the same feature vector for all images. Nowadays, many researchers have noticed potential environmental problems brought by training deep learning models (Xu et al., 2021), for instance, (Strubell et al., 2019) reports a remarkable example that the carbon dioxide emissions generated by training a Transformer (Vaswani et al., 2017) is equivalent to 200 round trips between San Francisco and New York by plane. Therefore, we motivate our method to perform an efficient self-supervised contrastive approach to learn anti-collapse and discriminative features based on a restricted amount of images in a training epoch (e.g., small batch size) and plain neural networks with limited parameters. Much research effort has been devoted to strong augmentations on data, but the informativeness of the features learned from the augmented data is hard to exactly measure, since the data is fed into mapping-agnostic deep neural networks to generate the features. Instead, we directly tackle augmentations on features and show that appropriate feature augmentations can sharply improve the optimization. To this end, we propose Meta Feature Augmentation (Met Aug), which learns view-specific encoders (with projection heads) and auxiliary meta feature augmentation generators (MAGs) by margin-injected meta feature augmentation and optimization-driven unified contrast. Suppose the input data has M views, and the multi-view data is fed into the encoder to generate the latent features. We initialize M neural networks as MAGs for views, which are used to augment the features of each view. We contrast all original and augmented features for bi-optimization training. Through such a learning paradigm, Met Aug can improve the performance of self-supervised contrastive learning. To learn anti-collapse and discriminative features from a restricted amount of images, Met Aug relies on two key ingredients: 1) margin-injected meta feature augmentation, where MAGs use the performance of the encoder in one iteration to improve the view-specific feature augmentations for the next iteration. In this way, MAGs promote the encoder to efficiently explore the discriminative information of the input. For the original features and the augmented features generated by MAGs, we inject a margin Rσ between the similarities of them, which avoids the instance-level feature collapse; 2) optimization-driven unified contrast, which contrasts all features in one gradient back-propagation step. Such proposed contrast can also amplify the impact of the instance similarity that deviates far from the optimum and weaken the impact of the instance similarity that is close to the optimum. We conduct head-to-head comparisons on various benchmark datasets, which prove the effectiveness of margin-injected meta feature augmentation and optimization-driven unified contrast. Contributions: We propose margin-injected meta feature augmenta- tion, which directly augments the latent features to generate informative and anti-collapse features. Benefiting from such features, encoders can efficiently capture discriminative information. We propose optimization-driven unified contrast to include all available features in one step of backpropagation and weight the similarities of paired features by measuring their contributions to optimization. Empirically, Met Aug improves the downstream task performance on different benchmark datasets. 2. Related works Self-supervised learning. Under the setting of unsupervised learning, SSL methods have achieved impressive success, which constructs auxiliary tasks to learn discriminative information from the unlabeled inputs. Deep Info Max (Hjelm et al., 2018) explores to maximize the mutual information between an input and the output of a deep neural network encoder by different mutual information estimations. CPC (Oord et al., 2018) proposes to adopt noise-contrastive estimation (NCE) (Gutmann & Hyvärinen, 2010) as the contrastive loss to train the model to measure the mutual information of multiple views deduced by the Kullback Leibler divergence (Goldberger et al., 2003). CMC (Tian et al., 2019) and AMDIM (Bachman et al., 2019) employ contrastive learning on the multi-view data. Sw AV (Caron et al., 2020) compares the cluster assignments under different views instead of directly comparing features by using more views (e.g., six views). Sim CLR (Chen et al., 2020) and Mo Co (He et al., 2020) use large batch or memory bank to enlarge the amount of available negative features to learn good representations. Instead of exploring informative features by adopting various data augmentations and enlarging the number of features, our method focuses on straightforwardly generating informative features to contrast. Recent works explore imposing stronger constraints on the conventional contrastive learning paradigm or propose alternative loss functions (instead of contrastive loss). Debiased CL(Chuang et al., 2020) and Hard CL (Robinson et al., 2020) consider to directly collect informative features to contrast by designing sampling strategies, which are inspired by positive-unlabeled learning methods (Elkan & Noto, 2008; du Plessis et al., 2014). Motivated by (Sridharan & Kakade, 2008), (Tsai et al., 2020) proposes an information theoretical framework for SSL, which, guided by the theory, uses information bottleneck to restrict the learned features and maintain the sufficient self-supervision. BYOL (Grill et al., 2020), W-MSE (Ermolov et al., 2020), and Barlow Twins (Zbontar et al., 2021) present a crucial issue that insufficient self-supervision (e.g., not enough negative features) may lead to the feature collapse in hidden space. To tackle the Met Aug: Contrastive Learning via Meta Feature Augmentation Figure 1. Met Aug s architecture. Dashed blue box represents the data encoding process, and dashed red box represents the meta feature augmentation. In training, we first fix aωj and aωj , and then train fθj( ), gϑj( ), fθj ( ) and gϑj ( ) by using LMet Aug. Next, we fix the encoders and projection heads, and train aωj and aωj in a meta updating manner. The networks are iteratively trained until convergence. mentioned issue, we propose a new margin-injected regularization in meta feature augmentation to avoid generating degenerate features. DACL (Verma et al., 2021) proposes a new data augmentation that applies to domain-agnostic problems. Loo C (Xiao et al., 2021) learns to capture varying and invariant factors for visual representations by constructing separate embedding spaces for each augmentation. These methods explore informative features from the perspective of data augmentation, while our straightforward idea behind our method is to augment features in the latent space. Meta learning. The objective of meta learning is to automatically learn the learning algorithm. Early works (a et al., 2002; Bengio et al., 2002; Schmidhuber, 2014) aim to guide the model (e.g., neural network) to learn prior knowledge about how to learn new knowledge, so that the model can efficiently learn new knowledge, e.g., the model can be quickly fine-tuned to specific downstream tasks with few training steps and achieve good performance. Recently, researchers explored to use meta learning to find optimal hyper-parameters (Li et al., 2017) and appropriately initialize a neural network for few-shot learning (Finn et al., 2017; Snell et al., 2017; Vinyals et al., 2016). Recent approaches (Chen et al., 2016; Jaderberg et al., 2016; Ma et al., 2018; Liu et al., 2019) have focused on learning optimizers or generating a gradient-driven loss for deep neural networks in the field of NLP, computer vision, etc. Our goal is to learn representations that capture information shared between multiple different views by performing selfsupervised contrastive learning. Formally, we suppose the input multi-view dataset as X = {X1, X2, ..., XN}, where N denotes the number of samples. Xi represents a collection of M views of the sample, where i {1, ..., N}. For each sample Xi, we denote xi as a random variable representing views following xi P(Xi), and xj i denotes the j-th view of the i-th sample, where j {1, ..., M}. 3.1. Contrastive learning preliminary We recap the preliminaries of contrastive learning(Tian et al., 2019; Chen et al., 2020): the foundational idea behind contrastive learning is to learn an embedding that maximizes agreement between the views of the same sample and separates the views of different samples in latent space. Given a multi-view dataset X, we treat pairs of the views of the same sample {xj i, xj i }, where j, j {1, ..., M}, as positives, versus pairs of the views of the different samples {xj i, xj i }, where i = i , as negatives. To impose contrastive learning, we feed the input xj i into a view-specific encoder fθj( ) to learn a representation hj i, and hj i is mapped into a feature zj i by a projection head gϑj( ), where θj and ϑj are the network parameters of fθj( ) and gϑj( ), respectively. A discriminating function d( ) is adopted to measure the similarity of {zj i , zj i }, where i = i . The encoder fθj( ) and projection head gϑj( ) are trained by using a contrastive loss (Oord et al., 2018), which is formulated as follows: log d ({z+}) d ({z+}) + K P k=1 d ({z }k) where XS = {{z+}, {z }1, {z }2, ..., , {z }k} is a set of pairs randomly sampled from X, which includes a positive {z+} and K negatives {z }k, k {1, ..., K}, because contrastive loss can only use one positive in an iteration. In test, the projection head gϑj( ) is discarded, and the representation hj i is directly used for downstream tasks. 3.2. Margin-injected meta feature augmentation Recent contrastive methods rely on complex data augmentations to increase the informativeness of views. Yet this lack of guidance approach leads to the demand for a large number of training data (e.g., large batch size and memory bank). We propose a meta feature augmentation method, Met Aug: Contrastive Learning via Meta Feature Augmentation 1.0 0 0.1 0.2 0.3 0.4 0 (b) w/ margin-injected regularization 0.5 0.6 0.7 0.8 1.0 0 0.1 0.2 0.3 0.4 0 (a) w/o margin-injected regularization 0.5 0.6 0.7 0.8 1.0 0 0.1 0.2 0.3 0.4 0 (c) w/o margin-injected regularization 0.5 0.6 0.7 0.8 1.0 0 0.1 0.2 0.3 0.4 0 (d) w/ margin-injected regularization 0.5 0.6 0.7 0.8 Figure 2. Similarity histograms obtained by our method (with or without margin-injected regularization) on CIFAR-10. (a) and (b) demonstrate the summarized similarity of positives (i.e., {ˆz+}) that include original features and augmented features. (c) and (d) demonstrate the statistical results of the original features learned by our model. Blue histograms represent the similarity between the features of the same image s views, and red histograms represent the similarity between the features of the different images views. which creates informative augmented features by updating parameters of its own network according to the performance (gradient) of the encoder (see Appendix A.3 for our rethinking of augmented features). A visualization of the overall Met Aug architecture is shown in Figure 1. To this end, we build a group of MAGs aω( ) = {aω1( ), ..., aωM ( )} for all M views, where ω = {ω1, ..., ωM}. To be simplified, we define fθ and gϑ as the groups of view-specific encoders and projection heads, respectively, i.e., θ = {θ1, ..., θM} and ϑ = {ϑ1, ..., ϑM}. In training, the encoders fθ( ) and the projection heads gϑ( ) are trained alongside the MAG aω( ) (with the network parameters ω). Following the protocol of meta learning (Finn et al., 2017; Liu et al., 2019), we firstly train fθ( ) and gϑ( ) under the learning paradigm of self-supervised contrastive learning. Then, aω( ) is updated by computing its gradients with respect to the performance of fθ( ) and gϑ( ). Here, we measure the performance of fθ( ) and gϑ( ) by the gradients of them when the corresponding contrastive loss is back-propagated. Concretely, all of fθ( ), gϑ( ), and aω( ) are iteratively trained until convergence. Specifically, we first update network parameters θ and ϑ of the encoders and projection heads by adopting the conventional contrastive loss. Then, we train the MAG aω( ) in a meta learning manner. We encourage the augmented features to be informative, and the encoders fθ( ) can better explore the discriminative information by jointly using the original and augmented features to contrast. Hence, the performance of the encoders would be promoted on the same training data. To update network parameters ω of the aω( ), we formalize the meta updating objective as follows: arg min ω L n g ϑ f θ e X , aω g ϑ f θ e X o (2) where e X represents a minibatch sampled from the training dataset X, n g ϑ f θ e X , aω g ϑ f θ e X o denotes a set including both original features and meta augmented features. ϑ and θ represent the parameter sets of the encoders and projection heads, respectively, which are computed by the updating of one gradient back-propagation: θ = θ ℓ θL n gϑ fθ e X , aω gϑ fθ e X o ϑ = ϑ ℓ ϑL n gϑ fθ e X , aω gϑ fθ e X o (3) where ℓis the learning rate shared between θ and ϑ. The idea behind the meta updating objective is that we perform the second-derivative technique (Finn et al., 2017; Zhang et al., 2018; Liu et al., 2019) to train aω( ). Specifically, a derivative over the derivative (Hessian matrix) of the combination {θ, ϑ} is used to update ω, where {θ, ϑ} is a parameter set conjoining θ and ϑ. We compute the derivative with respect to ω by using a retained computational graph of {θ, ϑ}. However, in practice, we find a critical issue: when the original features are not informative enough, large gradients are difficult to generate by contrasting the uninformative features, the MAGs aω( ) are inclined to create collapsed augmented features, e.g., the augmented features and the original features are very similar. We consider the reason for the feature collapse is that small gradient changes of the encoders alongside projection heads gϑ(fθ( )) lead to the update step-size of aω( ) to become extensively small, which leaves the optimization of aω( ) to fall into a local optimum. The augmented features are such that without any extra useful information. To tackle this issue, we further inject a margin to encourage aω( ) to generate more complex and informative augmented features, which can be considered as a regularization term in the meta updating objective. See Figure 2(a) for the details of the augmented feature collapse issue, and we observe that, without margin-injected regularization, MAGs tend to generate collapsed features that are very similar with the original features. Formally, we formulate the approach to generate margins for aω( ) by σ+ = min h min d({z+}k+) , max d({z }k ) i σ = max h min d({z+}k+) , max d({z }k ) i (4) where d({z+}k+) is a set of the outputs (similarities) Met Aug: Contrastive Learning via Meta Feature Augmentation of positives computed by the discriminating function d( ), and k+ {1, ..., K+} where K+ represents the number of positives in a minibatch. n d({z }k ) o is a set of the discriminating outputs of negatives, and k {1, ..., K } where K represents the number of negatives. Note that only original features are used in Equation 4. We call the formulated margin generation approach as "Large", and we also propose two more approaches, called "Medium" and "Small". In Appendix A.2, we conduct comparisons to evaluate the effects of the three margin generation approaches. We inject the margins between the augmented features and original features by adding a regularization term in the meta updating objective, and the regularization is defined as: h d({ˆz+}ˆk+) σ+i h σ d({ˆz }ˆk ) i where {ˆz+}ˆk+ denotes a positive that includes one original feature and one augmented feature, and ˆK+ denotes the number of such positives. {ˆz }ˆk represents likewise one of ˆK negatives, each of which includes one original feature and one augmented feature. [ ]+ denotes the cut-offat-zero function, which is defined as [a]+ = max(a, 0). We then integrate such regularization to the updating of ω by ω ω ℓ ωL n g ϑ f θ e X , aω g ϑ f θ e X o + α Rσ where ℓ represents the learning rate of ω, and α is a hyperparameter balancing the impact of the loss of margininjected regularization term. Rσ restricts MAGs to generate informative features that are more different with the original features (see Figure 2(b)). In practice, Figure 2(c) and (d) show that the features learned by our method (with margin-injected regularization) are more concentrated, e.g., the features of the same image are more similar and the gap between the features of the different images are enlarged, which proves informative augmented features can further lead the encoders to learn non-collapsed (scattered) features. 3.3. Optimization-driven unified contrast We propose to jointly contrast all features (including the original features and the meta augmented features) in one gradient back-propagation step. Motivated by (Schroff et al., 2015), we introduce the following optimization-driven unified loss function to replace the conventional contrastive Algorithm 1 Met Aug Input: Multi-view dataset X with M views of each sample, minibatch size n, and hyperparameters α, β, δ. Initialize The neural network parameters: θ and ϑ for view-specific encoders fθ( ) and projection heads gϑ( ), ω for MAGs, i.e., aω( ). The learning rates: ℓand ℓ . repeat for t-th training iteration do Iteratively sample minibatch e X = {Xi}tn i=(t 1)n. # regular contrastive training step θ θ ℓ θLMet Aug(fθ, gϑ, aω, e X) ϑ ϑ ℓ ϑLMet Aug(fθ, gϑ, aω, e X) end for for t-th training iteration do Iteratively sample minibatch e X = {Xi}tn i=(t 1)n. # compute fast weights # retain computational graph θ = θ ℓ θLMet Aug(fθ, gϑ, aω, e X) ϑ = ϑ ℓ ϑLMet Aug(fθ, gϑ, aω, e X) # meta training step using second derivative ω ω ℓ ω LMet Aug(f θ, g ϑ, aω, e X) + α Rσ end for until θ, ϑ, and ω converge. loss as follows: k =1 d({z }k ) k+=1 d({z+}k+) + λ + (7) where [ ]+ ensures that L 0 is always held. Note that all original features and augmented features are involved. λ is a margin between the summarized instance similarities to enhance the capability of the similarity separation. However, we find that the difference between PK k =1 d({z }k ) and PK+ k+=1 d({z+}k+) is not the larger the better. Excessive increases of the difference may undermine the convergence in optimization. We thereby wish to adopt a margin λ that leads to preferable convergence. We reform the loss in Equation 7 by adding a temperature coefficient β as follows: exp β d({z }k ) d({z+}k+) + λ ) (8) when β + , Equation 8 is Equation 7. Inspired by (Sun et al., 2020), we use weighting factors Γ and Γ+ to modulate the impacts of d({z }k ) and d({z+}k+). Such approach aims to give greater weight to the similarity Met Aug: Contrastive Learning via Meta Feature Augmentation Table 1. Comparison of different methods on classification accuracy (top 1). We use conv and fc as backbones in the experiments. denotes that the methods have reduced learnable parameters (See Appedix B.1). Model Tiny Image Net STL-10 CIFAR10 CIFAR100 conv fc conv fc conv fc conv fc Fully supervised 36.60 68.70 75.39 42.27 Bi GAN 24.38 20.21 71.53 67.18 62.57 62.74 37.59 33.34 NAT 13.70 11.62 64.32 61.43 56.19 51.29 29.18 24.57 DIM 33.54 36.88 72.86 70.85 73.25 73.62 48.13 45.92 Split Brain 32.95 33.24 71.55 63.05 77.56 76.80 51.74 47.02 Sw AV 39.56 0.2 38.87 0.3 70.32 0.4 71.40 0.3 68.32 0.2 65.20 0.3 44.37 0.3 40.85 0.3 Sim CLR 36.24 0.2 39.83 0.1 75.57 0.3 77.15 0.3 80.58 0.2 80.07 0.2 50.03 0.2 49.82 0.3 CMC 41.58 0.1 40.11 0.2 83.03 85.06 81.31 0.2 83.28 0.2 58.13 0.2 56.72 0.3 Mo Co 35.90 0.2 41.37 0.2 77.50 0.2 79.73 0.3 76.37 0.3 79.30 0.2 51.04 0.2 52.31 0.2 BYOL 41.59 0.2 41.90 0.1 81.73 0.3 81.57 0.2 77.18 0.2 80.01 0.2 53.64 0.2 53.78 0.2 Barlow Twins 39.81 0.3 40.34 0.2 80.97 0.3 81.43 0.3 76.63 0.3 78.49 0.2 52.80 0.2 52.95 0.2 DACL 40.61 0.2 41.26 0.1 80.34 0.2 80.01 0.3 81.92 0.2 80.87 0.2 52.66 0.2 52.08 0.3 Loo C 42.04 0.1 41.93 0.2 81.92 0.2 82.60 0.2 83.79 0.2 82.05 0.2 54.25 0.2 54.09 0.2 Sim CLR + Debiased 38.79 0.2 40.26 0.2 77.09 0.3 78.39 0.2 80.89 0.2 80.93 0.2 51.38 0.2 51.09 0.2 Sim CLR + Hard 40.05 0.3 41.23 0.2 79.86 0.2 80.20 0.2 82.13 0.2 82.76 0.1 52.69 0.2 53.13 0.2 CMC + Debiased 41.64 0.2 41.36 0.1 83.79 0.3 84.20 0.2 82.17 0.2 83.72 0.2 58.48 0.2 57.16 0.2 CMC + Hard 42.89 0.2 42.01 0.2 83.16 0.3 85.15 0.2 83.04 0.2 86.22 0.2 58.97 0.3 59.13 0.2 Met Aug (only OUCL) 42.02 0.1 42.14 0.2 84.09 0.2 84.72 0.3 85.98 0.2 87.13 0.2 59.21 0.2 58.73 0.2 Met Aug 44.51 0.2 45.36 0.2 85.41 0.3 85.62 0.2 87.87 0.2 88.12 0.2 59.97 0.3 61.06 0.2 that deviates from the optimum and smaller weight to the similarity that has the close proximity with the optimum. Γ = [d({z }k ) O ]+ and Γ+ = [O+ d({z+}k+)]+, where O and O+ represents the expected optimums of d({z }k ) and d({z+}k+). Note that, we further propose a variation of Γ including Γ and Γ+, and the comparisons of them are demonstrated in Section 4.4. γ+ and γ is used to replace λ and add Γ and Γ+ in Equation 8: k+=1 exp β Γ (d({z }k ) γ ) Γ+(d({z+}k+) γ+) ) We limit d({z }k ) and d({z+}k+) in the range of [0, 1] by normalizing the features in {z }k and {z }k , such that theoretically, the optimum of d({z }k ) is 0, the optimum of d({z+}k+) is 1. The positive of d({z }k ) O and O+ d({z+}k+) can easily be guaranteed. To cut the number of hyperparameters, we reform Equation 9 into β (d({z+}k+) 1) 2 + (d({z }k )) 2 2γ2 ) (10) which is derived by setting O+ = 1 + γ, O = γ, γ+ = 1 γ, and γ = γ. 3.4. Model objective Concretely, we adopt margin-injected meta feature augmentation in the contrastive learning paradigm to achieve desired discriminative multi-view representations, and the proposed LMet Aug is incorporated to replace the conventional contrastive loss L. The final model objective is defined as: LMet Aug = Lori OUCL + δ Laug OUCL (11) where Lori OUCL represents the loss NOT including the meta augmented features, Laug OUCL represents the loss including such features, and δ is a coefficient that controls the balance between them (we perform parameter comparisons in Appendix A.4). It is worthy to note that the margin-injected regularization Rσ is only used in meta training the MAGs, i.e., aω( ), while in regular training of encoders and projection heads, Rσ is discarded. Rσ restricts the augmented features to be informative so that such features can lead the encoder to efficiently and effectively learn discriminative representations. The training process is detailed by Algorithm 1. 4. Experiments We benchmark our Met Aug on five established datasets: Tiny Image Net (Krizhevsky et al., 2009), STL-10 (Coates et al., 2011), CIFAR10 (Krizhevsky et al., 2009), CIFAR100 (Krizhevsky et al., 2009), and Image Net (Jia et al., 2009). The compared benchmark methods include: Bi GAN (Donahue et al., 2016), NAT (Bojanowski & Joulin, 2017), DIM (Hjelm et al., 2018), Split Brain (Zhang et al., 2017), CPC Met Aug: Contrastive Learning via Meta Feature Augmentation Table 2. Performance (accuracy) on the CIFAR10 and STL-10 datasets with Res Net-50 (He et al., 2016). Model CIFAR10 STL-10 Average Sw AV 83.15 82.93 83.04 Sim CLR 84.63 83.75 84.19 CMC 86.10 86.83 86.47 BYOL 87.14 87.56 87.35 Barlow Twins 85.84 86.02 85.93 DACL 86.93 88.11 87.52 Loo C 87.80 88.62 88.21 Sw AV + Hard 83.99 84.51 84.25 Sim CLR + Hard 86.91 85.48 86.20 CMC + Hard 88.25 87.79 88.02 Met Aug (only OUCL) 88.79 88.31 88.55 Met Aug 91.09 90.26 90.68 (Hénaff et al., 2019), Sw AV (Caron et al., 2020), Sim CLR (Chen et al., 2020), CMC (Tian et al., 2019), Mo Co (He et al., 2020), Sim Siam (Chen & He, 2020), Info Min Aug. (Tian et al., 2020), BYOL (Grill et al., 2020), Barlow Twins (Zbontar et al., 2021), DACL (Verma et al., 2021) Loo C (Xiao et al., 2021), Debiased (Chuang et al., 2020), Hard (Robinson et al., 2020), and NNCLR (Dwibedi et al., 2021). 4.1. Efficiently performing Met Aug Implementations. To efficiently perform CL within a restricted amount of the inputs in training, we uniformly set the batch size as 64 (see Appendix A.1 for the comparisons under different setting of batch size). For the experiments with conv and fc as the backbone networks, we adopt a network with the 5 convolutional layers in Alex Net (Krizhevsky et al., 2012) as conv and a network with further 2 fully connected layers as fc. Inspired by the backbone splitting setting of Split Brain (Zhang et al., 2017), we evenly split the Alex Net into sub-networks across the channel dimension, and each sub-network is the view-specific encoder (see Appendix B for the detailed implementation). For the experiments with Res Net-50, we directly change the encoder network to Res Net-50. All backbone encoders are not pre-trained. Met Aug (only OUCL) is the ablation variant without margin-injected meta feature augmentation. Given an RGB image, we convert it to the Lab image color space and split it into L and ab channels. During contrastive learning, RGB, L, and ab are used as three views of the image. Before feeding the views into our model, we simply adopt the same data augmentations in CMC (Tian et al., 2019). Especially, the major contribution of DACL is the proposed data augmentation (i.e., mixup-noise) so that we particularly add mixup data augmentation for DACL. In training, a memory bank (Wu et al., 2018) is adopted to facilitate calculations. We retrieve 4096 past features from the memory bank to derive negatives. The learning rates and weight decay rates are uniform over comparisons. Table 3. Linear evaluation results on Image Net. We follow the setting of (Tian et al., 2019; Dwibedi et al., 2021) to compare with other benchmark SSL methods with conv and Res Net-50. Met Aug denotes Met Aug further leveraging the advances of the positive re-sampling technique (Dwibedi et al., 2021). Note that the batch size adopted by the compared methods are inconsistent in the comparisons using Res Net-50, i.e., CMC adopts 128 yet others adopt 4096. Therefore, we report the comparisons of Met Aug using Res Net-50 with the batch size of 128 in Appendix A.5. Model conv Res Net-50 top 1 top 5 Fully supervised 50.5 - - Split Brain 32.8 - - CPC v2 - 63.8 85.3 Sw AV 38.0 0.3 71.8 - Sim CLR 37.7 0.2 71.7 - CMC 42.6 - - Mo Co 39.4 0.2 71.1 - Sim Siam - 71.3 - Info Min Aug. - 73.0 91.1 BYOL 41.1 0.2 74.3 91.6 Barlow Twins 39.6 0.2 - - NNCLR - 75.4 92.3 DACL 41.8 0.2 - - Loo C 43.2 0.2 - - Sim CLR + Debiased 38.9 0.3 - - Sim CLR + Hard 41.5 0.2 - - Met Aug 45.1 0.2 - - Met Aug - 76.0 93.2 Comparison on downstream tasks. We collect the results of 20 trials for comparisons. The average result of the last 20 epochs is used as the final result of each trial, and the 95% confidence intervals are also reported, while the results without 95% confidence intervals are quoted from the published papers. We compare Met Aug against a fully-supervised method (similar to Alex Net (Krizhevsky et al., 2012)) and the state-of-the-art unsupervised methods. Table 1 shows the comparisons on four benchmark datasets. The last two rows of tables represent the results of our methods. As demonstrated in tables, Met Aug beats the best prior methods on all datasets. Even compared with the fully-supervised method trained end-to-end (without fine-tuning) for the architecture presented, the proposed method has a significant improvement on most downstream tasks, which demonstrates that Met Aug can better model discriminative information when supervision is insufficient (e.g., the training data is limited). The ablation model (i.e., Met Aug (only OUCL)) outperforms most unsupervised methods but falls short of the performance of Met Aug. Thus, the ablation study proves the effectiveness of our proposed margin-injected meta feature augmentation and optimization-driven unified contrast. DACL and Loo C propose to enhance contrastive learning from the perspective of data augmentation, while Met Aug Met Aug: Contrastive Learning via Meta Feature Augmentation Table 4. Comparison of applying benchmark SSL methods with different data augmentations by using the fc backbone on CIFAR10. Data augmentations Methods horizontal rotate random random color mixup DACL Loo C Met Aug flip crop grey jitter 1 - 80.73 87.05 2 - 81.16 87.53 3 - 80.70 86.81 4 - 81.64 87.79 5 - 82.05 88.12 6 - 82.16 88.01 7 80.87 82.21 88.22 8 82.09 83.17 88.65 improves contrastive learning from the perspective of feature augmentation. The idea behind our method is simple but effective, since contrastive learning works directly on features, and the augmented images need one step of encoding to become features. The experimental results support that Met Aug achieves better performance on benchmarks. Performing Met Aug on Res Net. We perform classification comparisons on the CIFAR10 and STL-10 by using Res Net50. Table 2 shows that Met Aug and the ablation variant outperform the compared methods, which indicates that Met Aug has strong adaptability to different encoders. 4.2. Benchmarking Met Aug on Image Net Implementation. To comprehensively understand the performance of our proposed Met Aug, we conduct comparisons on Image Net and make fair comparisons with benchmark methods. The backbone encoder is conv or Res Net-50, and the results are demonstrated in Table 3. Met Aug is a decoupled approach so that we can introduce Met Aug in the learning paradigm of state-of-the-art to improve the performance, e.g, for the experiments using conv or Res Net-50, we perform Met Aug in CMC or NNCLR, respectively. Results. As shown in Table 3, we find that Met Aug can effectively promote the performance of benchmark methods in the comparisons using both conv and Res Net-50. The results support that our proposed meta feature augmentation can enable different encoders to model discriminative information even in the large-scale dataset. 4.3. Is Met Aug robust for data augmentation? To illustrate the impacts of different data augmentations, we conducted multiple comparisons on CIFAR10 shown in Table 4. Note that horizontal flip and rotate are similar, and we use them together in the 1-th comparison. In the 5-th comparison, we take the same data augmentations as the setting of comparisons in Section 4.1. The data augmentations adopted in the 6-th comparison are as same as the setting of Met Aug (only OUCL) w/ Γ Met Aug (only OUCL) w/o Γ Accuracy(%) Figure 3. The impact of different ϕdec on the performance of our method using Γ. Comparisons are conducted on Tiny Image Net with conv as the encoders. Loo C (Xiao et al., 2021). Additionally, mixup is proposed by DACL (Verma et al., 2021). We observe from Table 4 that Met Aug outperforms the compared methods in all comparisons. It is worth noting that even using weak data augmentation degenerates the performance of our method as well as benchmark methods, but the performance degeneration of our method is minimal compared to others, e.g., from 8-th and 1-th comparison, we find that the gap of Met Aug is 1.60%, while that of Loo C is 2.44%. The results support that Met Aug is robust for various data augmentations. 4.4. Do the variant of Γ promote Met Aug? In practice, we find that the introduction of the weighting factor Γ cannot directly improve our proposed method. Our conjuncture lies in that Γ may cause the loss to converge excessively fast, which leaves the network parameters at a local minimum. Therefore, we propose a variant to replace Γ in Equation 9, i.e., Γ = Γ ϕdec where ϕdec is a linear attenuation coefficient to linearly attenuate the impact of Γ so that the difference between the current value and the optimum becomes smaller. We use Met Aug (only OUCL) to demonstrate the effectiveness of the proposed variant. The results are shown in Figure 3. We observe that the performance of our method get peak value when ϕdec is 6, which manifests that introducing a certain linear attenuation to Γ can promote Met Aug. Met Aug: Contrastive Learning via Meta Feature Augmentation 5. Conclusion We conclude that exploring informative features is the key to contrastive learning. Different from the conventional contrastive methods that collect enough informative features to learn a good representation by enlarging the batch or memory bank, we motivate Met Aug to learn a discriminative representation from a restricted amount of images. Our method proposes margin-injected meta feature augmentation to straightforwardly augment features to be informative and avoid learning degenerate features. To efficiently make use of all available features, Met Aug further proposes optimization-driven unified contrast. Experimental evaluations demonstrate that Met Aug achieves the state-of-the-art. Acknowledgements The authors would like to thank the anonymous reviewers for their valuable comments. This work is supported in part by National Natural Science Foundation of China No. 61976206 and No. 61832017, Key Special Project for Introduced Talents Team of Southern Marine Science and Engineering Guangdong Laboratory (Guangzhou) No. GML2019ZD0603, National Key Research and Development Program of China No. 2019YFB1405100, Beijing Outstanding Young Scientist Program NO. BJJWZYJH012019100020098, Beijing Academy of Artificial Intelligence (BAAI), the Fundamental Research Funds for the Central Universities, the Research Funds of Renmin University of China 21XNLG05, and Public Computing Cloud, Renmin University of China. This work is also supported in part by Intelligent Social Governance Platform, Major Innovation & Planning Interdisciplinary Platform for the Double-First Class Initiative, Renmin University of China, and Public Policy and Decision-making Research Lab of Renmin University of China. a, S., Bengio, S., Bengio, Y., Cloutier, J., and Gecsei, J. On the optimization of a synaptic learning rule. 2002. Arora, S., Khandeparkar, H., Khodak, M., Plevrakis, O., and Saunshi, N. A theoretical analysis of contrastive unsupervised representation learning. 2019. Bachman, P., Hjelm, R. D., and Buchwalter, W. Learning representations by maximizing mutual information across views. In Neur IPS 2019, 2019. Bengio, Y., Bengio, S., and Cloutier, J. Learning a synaptic learning rule. In Ijcnn-91-seattle International Joint Conference on Neural Networks, 2002. Bojanowski, P. and Joulin, A. Unsupervised learning by predicting noise. ar Xiv preprint ar Xiv:1704.05310, 2017. Caron, M., Misra, I., Mairal, J., Goyal, P., Bojanowski, P., and Joulin, A. Unsupervised learning of visual features by contrasting cluster assignments. 2020. Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. A simple framework for contrastive learning of visual representations. ar Xiv preprint ar Xiv:2002.05709, 2020. Chen, X. and He, K. Exploring simple siamese representation learning. 2020. Chen, Y., Hoffman, M. W., Colmenarejo, S. G., Denil, M., Lillicrap, T. P., Botvinick, M., and Freitas, N. D. Learning to learn without gradient descent by gradient descent. 2016. Chopra, S., HAd Sell, R., and Lecun, Y. Learning a similarity metric discriminatively, with application to face verification. In 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 05), 2005. Chuang, C. Y., Robinson, J., Lin, Y. C., Torralba, A., and Jegelka, S. Debiased contrastive learning. 2020. Coates, A., Ng, A., and Lee, H. An analysis of singlelayer networks in unsupervised feature learning. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, 2011. Cubuk, E. D., Zoph, B., Shlens, J., and Le, Q. V. Randaugment: Practical data augmentation with no separate search. Co RR, abs/1909.13719, 2019. URL http: //arxiv.org/abs/1909.13719. Donahue, J., Krähenbühl, P., and Darrell, T. Adversarial feature learning. ar Xiv preprint ar Xiv:1605.09782, 2016. du Plessis, M. C., Niu, G., and Sugiyama, M. Analysis of learning from positive and unlabeled data. In Advances in Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems 2014, December 8-13 2014, Montreal, Quebec, Canada, 2014. Dwibedi, D., Aytar, Y., Tompson, J., Sermanet, P., and Zisserman, A. With a little help from my friends: Nearestneighbor contrastive learning of visual representations. 2021. Elkan, C. and Noto, K. Learning classifiers from only positive and unlabeled data. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, Nevada, USA, August 24-27, 2008. ACM, 2008. Met Aug: Contrastive Learning via Meta Feature Augmentation Ermolov, A., Siarohin, A., Sangineto, E., and Sebe, N. Whitening for self-supervised representation learning. 2020. Finn, C., Abbeel, P., and Levine, S. Model-agnostic metalearning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, Proceedings of Machine Learning Research, pp. 1126 1135. PMLR, 2017. Goldberger, J., Gordon, S., and Greenspan, H. An efficient image similarity measure based on approximations of kl-divergence between two gaussian mixtures. In IEEE International Conference on Computer Vision, 2003. Grill, J. B., Strub, F., Altché, F., Tallec, C., Richemond, P. H., Buchatskaya, E., Doersch, C., Pires, B. A., Guo, Z. D., and Azar, M. G. Bootstrap your own latent: A new approach to self-supervised learning. 2020. Gutmann, M. and Hyvärinen, A. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. 2010. Hadsell, R., Chopra, S., and Lecun, Y. Dimensionality reduction by learning an invariant mapping. In 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 06), 2006. He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, 2016. He, K., Fan, H., Wu, Y., Xie, S., and Girshick, R. Momentum contrast for unsupervised visual representation learning. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020. Hjelm, R. D., Fedorov, A., Lavoie-Marchildon, S., Grewal, K., Bachman, P., Trischler, A., and Bengio, Y. Learning deep representations by mutual information estimation and maximization. ar Xiv preprint ar Xiv:1808.06670, 2018. Hénaff, O., Srinivas, A., Fauw, J. D., Razavi, A., Doersch, C., Eslami, S., and Oord, A. Data-efficient image recognition with contrastive predictive coding. 2019. Jaderberg, M., Czarnecki, W. M., Osindero, S., Vinyals, O., Graves, A., and Kavukcuoglu, K. Decoupled neural interfaces using synthetic gradients. 2016. Jia, D., Wei, D., Socher, R., Li, L. J., Kai, L., and Li, F. F. Imagenet: A large-scale hierarchical image database. Proc of IEEE Computer Vision and Pattern Recognition, 2009. Krizhevsky, A., Hinton, G., et al. Learning multiple layers of features from tiny images. 2009. Krizhevsky, A., Sutskever, I., and Hinton, G. Imagenet classification with deep convolutional neural networks. In NIPS, 2012. Li, Z., Zhou, F., Fei, C., and Hang, L. Meta-sgd: Learning to learn quickly for few-shot learning. 2017. Liu, S., Davison, A. J., and Johns, E. Self-supervised generalisation with meta auxiliary learning. 2019. Ma, C., Shen, C., Dick, A., Wu, Q., Wang, P., Hengel, A., and Reid, I. Visual question answering with memoryaugmented networks. IEEE, 2018. Oord, A. v. d., Li, Y., and Vinyals, O. Representation learning with contrastive predictive coding. ar Xiv preprint ar Xiv:1807.03748, 2018. Robinson, J., Chuang, C. Y., Sra, S., and Jegelka, S. Contrastive learning with hard negative samples. 2020. Schmidhuber, J. Learning complex, extended sequences using the principle of history compression. Neural Computation, 4(2):234 242, 2014. Schroff, F., Kalenichenko, D., and Philbin, J. Facenet: A unified embedding for face recognition and clustering. 2015. Snell, J., Swersky, K., and Zemel, R. S. Prototypical networks for few-shot learning. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, 2017. Sridharan, K. and Kakade, S. M. An information theoretic framework for multi-view learning. Conference on Learning Theory, 2008. Strubell, E., Ganesh, A., and Mccallum, A. Energy and policy considerations for deep learning in nlp. 2019. Sun, Y., Cheng, C., Zhang, Y., Zhang, C., Zheng, L., Wang, Z., and Wei, Y. Circle loss: A unified perspective of pair similarity optimization. 2020. Tian, Y., Krishnan, D., and Isola, P. Contrastive multiview coding. ar Xiv preprint ar Xiv:1906.05849, 2019. Tian, Y., Sun, C., Poole, B., Krishnan, D., Schmid, C., and Isola, P. What makes for good views for contrastive learning. 2020. Tsai, Y., Wu, Y., Salakhutdinov, R., and Morency, L. P. Self-supervised learning from a multi-view perspective. 2020. Met Aug: Contrastive Learning via Meta Feature Augmentation Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., and Polosukhin, I. Attention is all you need. 2017. Verma, V., Luong, T., Kawaguchi, K., Pham, H., and Le, Q. V. Towards domain-agnostic contrastive learning. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, Proceedings of Machine Learning Research. PMLR, 2021. Vinyals, O., Blundell, C., Lillicrap, T., Kavukcuoglu, K., and Wierstra, D. Matching networks for one shot learning. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona, Spain, 2016. Wu, Z., Xiong, Y., Yu, S. X., and Lin, D. Unsupervised feature learning via non-parametric instance discrimination. 2018. Xiao, T., Wang, X., Efros, A. A., and Darrell, T. What should not be contrastive in contrastive learning. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. Open Review.net, 2021. Xu, J., Zhou, W., Fu, Z., Zhou, H., and Li, L. A survey on green deep learning. Co RR, 2021. Zbontar, J., Li, J., Misra, I., Lecun, Y., and Deny, S. Barlow twins: Self-supervised learning via redundancy reduction. 2021. Zhang, R., Isola, P., and Efros, A. A. Split-brain autoencoders: Unsupervised learning by cross-channel prediction. 2017. Zhang, Y., Tang, H., and Jia, K. Fine-grained visual categorization using meta-learning optimization with sample selection of auxiliary data. In Computer Vision - ECCV 2018 - 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part VIII, Lecture Notes in Computer Science, pp. 241 256. Springer, 2018. Met Aug: Contrastive Learning via Meta Feature Augmentation Figure 4. Comparison of different methods on classification accuracy (top1) under various settings of batch size. We conducted experiments on CIFAR10 with conv encoder. A. Appendix - Extended comparisons In this section, we provide several experimental analyses about the advantages of our proposed method. The experiments to find appropriate hyperparameters are conducted as well, and in detail, we conduct comparisons of using different hyperparameters on the validation set of corresponding benchmark datasets. A.1. Can Met Aug perform consistently under different settings of batch size? As the results shown in Table 1, 2, and 3, we observe that Met Aug achieves our expectation that learning anticollapse and discriminative representations from a restricted amount of images in a training step (i.e., the batch size is limited). However, we conduct further experiments to explore whether Met Aug has consistent performance under settings of larger batch sizes. From Figure 4, we observe that with the increase of batch size, each compared method achieves better performance on the downstream task. We conjecture that with the enlarging of batch size, the number of available features in a training step is increased, so that models may explore more informative features to promote the performance of contrastive learning. Yet, comparing our method with the benchmark methods, we find that the gap between the performance of Met Aug (only OUCL) and the compared methods becomes smaller. We extend the mentioned conjecture: as more informative features can be explored by all methods in a training step, OUCL s advantage becomes less significant. OUCL aims to include all available features to efficiently train the model and avoid the optimization fall into a local optimum, and the increase of batch size, which means sufficient self-supervision, can naturally promote the efficiency of optimization and avoid the fall into a local optimum. 60.77% 61.06% 60.59% 60.40% 59.74% 60.28% 59.80% 59.63% Figure 5. Heatmap of injected margin variant comparisons. Table 5. Performance (accuracy) of Met Aug with or without the augmented features on CIFAR10 with conv encoder. Model δ w/ augmented w/o augmented features features Sim CLR - 80.58 DACL - 81.92 Loo C - 83.79 CMC + Hard - 83.04 10 1 85.85 85.48 10 2 85.91 85.99 10 3 86.57 86.65 10 4 87.42 87.41 10 5 87.72 87.87 10 6 87.26 87.47 10 7 86.90 87.19 10 8 86.12 86.35 Yet the advance of OUCL is always maintained, which is supported by the comparison. Only Loo C s performance can gradually catch up with the performance of Met Aug (only OUCL). We research the setting of Loo C, and find that Loo C leverages more than one (e.g., three) contrastive loss in a training step, which allows Loo C to train the model multiple times. We observe that, even in the large batch size, Met Aug can still improve the state-of-the-art methods by a significant margin. Concretely, Met Aug maintains its superiority over compared method under different settings of batch size. A.2. Variants of the injected margin We denote min+ as min({d({z+}k+)}) and max as max({d({z }k )}). For "Medium", both σ+ and σ equal to mean[min+, max ]. σ+ = max[min+, max ] and σ = min[min+, max ] in "Small". In Figure 5, we conduct comparisons on CIFAR100 with fc. We observe that, whether our method uses Lcontrast or the proposed LOUCL, all three variants can improve Met Aug, and our method with "Large" achieve the best performance. The experiments further prove the effectiveness of the two key ingredients of Met Aug. A.3. Understanding of the augmented features To understand the augmented features, we conduct a comparison of Met Aug by adopting the augmented features in Met Aug: Contrastive Learning via Meta Feature Augmentation (a) Fix α, and study on β and δ (b) Fix β, and study on α and δ (c) Fix δ, and study on α and β Figure 6. Impacts of the hyperparameters α, β, and δ of our proposed method. We conducted comparisons based on Met Aug on CIFAR10 with fc encoder. To measure the influences, we iteratively fixed one parameter and then study on the others by selecting them in the ranges, respectively. The range of Accuracy(%) 0.30 0.25 0.20 0.15 0.35 0.40 0.45 0.50 41.46 41.55 42.06 42.14 Figure 7. Impacts of the hyperparameter γ of our proposed method. We conducted comparisons based on Met Aug (onlu OUCL) on Tiny Image Net with fc encoder. To measure the impact of γ, we iteratively select γ and observe the accuracy of the method. the test or not. As shown in Table 5, the results of Met Aug using the augmented features in the test are listed in the w/ augmented features column, and the results of Met Aug NOT using the augmented features in the test are listed in the w/o augmented features column (which is the regular approach in the test). We select δ from the range of {10 1, 10 2, 10 3, 10 4, 10 5, 10 6, 10 7, 10 8} to generate different augmented features. Specifically, the approach of adopting the augmented features in the test is that we use MAGs to generate augmented features and such features are treated as the same as the original features, i.e., augmented features are regarded as additions of original features. Note that, in regular test (i.e., w/ augmented features), we use the representation hj i and discard the projection head gϑj( ) to feed into the classifier, while, in the test of using augmented features, we have to use the feature zj i generated by the projection head gϑj( ) to feed into the classifier, because MAGs work on the feature zj i . From Table 5, we observe that, generally, Met Aug w/o aug- mented features beats Met Aug w/ augmented features. The reasons behind such phenomenon are: 1) the augmented features are generated to lead the encoders to learn discriminative representations (e.g., hj i), which indicates that the augmented features contribute to the improvement of the encoders, but this does not mean that the augmented features are discriminative for downstream tasks; 2) in the test of using augmented features, we do not discard the projection head gϑj( ), and recent works prove that the approach of using a projection head in training and discarding such head in the test can significantly improve the performance of the model on downstream tasks (Chen et al., 2020; He et al., 2020). Out of the understanding of the experimental results, we think that the augmented features contain useful information that can improve the encoder, but such information may not be discriminative to downstream tasks. A.4. Synthetic comparison of hyperparameters To intuitively understand the impacts of hyperparameters, we conduct comparisons by using various combinations of them for the proposed Met Aug. Specifically, α controls the impact of the proposed margin-injected regularization term. The hyperparameter β is proposed as a temperature coefficient in OUCL. γ is a specific parameter to replace the hyperparameters in OUCL such that the number of hyperparameters can be reduced. δ balances the impact of OUCL that uses augmented features and OUCL that does not use these features. As demonstrated in Figure 7, we first solely study γ s impact on Met Aug, because γ is only used in OUCL function, and in practice, we find that, compared with other hyperparameters, γ has less impact on our method. We conduct experiments on Tiny Image Net with fc encoder and select γ from the corresponding range for Met Aug (only OUCL) to clarify its impact, and the results indicate that appropriate selected γ can indeed promote the performance of our Met Aug: Contrastive Learning via Meta Feature Augmentation Table 6. Comparisons on Image Net with different view settings using Res Net-50 as the encoder. denotes our reimplementations. Rand Augment is proposed by (Cubuk et al., 2019). View Method Res Net-50 setting top 1 top 5 CMC 64.0 85.5 CMC 63.3 84.8 Met Aug (only OUCL) 63.9 85.7 Met Aug (only MAG) 64.4 86.0 Met Aug 65.1 86.2 CMC 64.8 86.1 CMC 64.5 86.2 CMC + Rand Augment 66.2 87.0 CMC + Rand Augment 65.7 86.8 Met Aug (only OUCL) 65.0 86.3 Met Aug (only MAG) 65.9 86.8 Met Aug 66.4 87.1 Met Aug + Rand Augment 66.7 87.5 method, but the differences between the impacts of different γ are limited. Then, we fix γ = 0.40 and study on the impacts of other hyperparameters. As the results are shown in Figure 6, the plots further elaborate our parameter studies results with Met Aug on the CIFAR10 benchmark dataset with fc encoder. To explore the influence of β and δ, we first fixed α = 10 13, and then we selected β from the range of {2, 22, 23, 24, 25, 26, 27, 28} and δ from the range of {10 1, 10 2, 10 3, 10 4, 10 5, 10 6, 10 7, 10 8}. Following the same experimental principle as above, we selected α from the range of {10 3, 10 5, 10 7, 10 9, 10 11, 10 13, 10 15, 10 17}. See Figure 6(a), (b), and (c) for the details of the comparisons. In general, good classification performance highly depends on the β and δ terms. Also, α is an intensely necessary supplement for adapting the interval between similarities of augmented features and original features, which avoids to learn degenerate representations. We also find that the potential to improve the learned representations grows with the adjustment of term β, e.g., the initial loss becomes relatively large. A.5. Discussion of the comparisons using Res Net-50 on Image Net In Table 3, we do not report the experimental results of CMC using Res Net-50, because the batch sizes adopted by them are inconsistent, i.e., CMC adopts 128 yet other methods adopt 4096. Therefore, the performance of CMC is not competitive compared to other compared methods, including Met Aug . Our proposed Met Aug can be treated as a feature augmentation approach so that Met Aug can be embedded into each self-supervised learning architecture. The original Met Aug is implemented based on CMC, so for a fair comparison, Met Aug adopts the same batch size as CMC, while Met Aug is implemented based on NNCLR (Dwibedi et al., 2021) with the batch size of 4096. Therefore, Met Augast apparently outperforms Met Aug, so we only report the best variant of our method, i.e., Met Augast. Whereas, we have comprehensively evaluated whether our method can really improve the performance of CMC on Image Net, which is shown in Table 6. Met Aug can be treated as a component to improve baseline methods, which is implemented based on CMC . We observe that Met Aug beats CMC and even achieves better performance than CMC + Rand Augment, which proves that our method can not only improve CMC but also beat Rand Augment. We further employ Rand Augment in our method. The results show that such a variant achieves the best performance, which is consistent with 4.3, i.e., stronger data augmentation improves the performance of our method. B. Appendix - Implementation In this paper, we introduce a novel self-supervised representation learning approach, i.e., Meta Feature Augmentation (Met Aug), of which Figure 1 depicts the overview framework. The following subsections provide the design details of Met Aug. B.1. Network architecture In the experiments, neural network classification methods (i.e., conv and fc) are adopted as the backbone networks, and the classifiers (i.e., the linear networks) on the representations extracted from the encoders are performed on downstream classification tasks. According to the principle of building the encoders, the Alex Net is split across the channel dimension with the idea that split-Alex Net can also perform well in learning representations between views, which only has the halved learnable parameters (Zhang et al., 2017). We build the Alex Net with 5 convolutional layers, 2 linear layers, and a fully connected layer followed by a l2 normalization function. Then the split-Alex Nets (i.e., the sub-networks) are regarded as the encoders. In experiments, we use conv and fc, which use the corresponding layers of Alex Net. Note that we split Alex Net across channels for RGB, L, and ab views. in the test, we concatenate representations layer-wise from the encoders into one to achieve the final representations of the inputs. We develop the classifier by leveraging a linear network followed by a softmax output function. Following the proposed experimental setting of the previous literature (Oord et al., 2018; Hjelm et al., 2018; Arora et al., 2019; Tian et al., 2019), we evaluate the quality of the learned representations by freezing the weights of backbone encoders and training the linear classifier in the test. Met Aug: Contrastive Learning via Meta Feature Augmentation B.2. Algorithm description Met Aug is an end-to-end representation learning method: we iteratively train the encoders and MAGs by backpropagating LMet Aug, and the training process is based on Adam gradient optimization. The proposed Met Aug is a generalized approach, which can be used for various downstream tasks, e.g., classification, clustering, regression, etc. We can straightforwardly train the encoders, pretrained by Met Aug, on downstream tasks. Here, we provide a pseudo-code for Met Aug training described in the style of Py Torch, which is without the inclusion of the detailed matrix processing or helper utility functions & codes that are irrelevant to the algorithm: