# localizing_memorization_in_ssl_vision_encoders__9d000299.pdf Localizing Memorization in SSL Vision Encoders Wenhao Wang1, Adam Dziedzic1, Michael Backes1, Franziska Boenisch1 1CISPA, Helmholtz Center for Information Security Recent work on studying memorization in self-supervised learning (SSL) suggests that even though SSL encoders are trained on millions of images, they still memorize individual data points. While effort has been put into characterizing the memorized data and linking encoder memorization to downstream utility, little is known about where the memorization happens inside SSL encoders. To close this gap, we propose two metrics for localizing memorization in SSL encoders on a per-layer (Layer Mem) and per-unit basis (Unit Mem). Our localization methods are independent of the downstream task, do not require any label information, and can be performed in a forward pass. By localizing memorization in various encoder architectures (convolutional and transformer-based) trained on diverse datasets with contrastive and non-contrastive SSL frameworks, we find that (1) while SSL memorization increases with layer depth, highly memorizing units are distributed across the entire encoder, (2) a significant fraction of units in SSL encoders experiences surprisingly high memorization of individual data points, which is in contrast to models trained under supervision, (3) atypical (or outlier) data points cause much higher layer and unit memorization than standard data points, and (4) in vision transformers, most memorization happens in the fully-connected layers. Finally, we show that localizing memorization in SSL has the potential to improve fine-tuning and to inform pruning strategies. 1 Introduction Self-supervised learning (SSL) ([16, 17, 13, 4, 30, 27, 29]) enables pre-training large encoders on unlabeled data to generate feature representations for a multitude of downstream tasks. Recently, it was found that, even though their training datasets are large, SSL encoders still memorize individual data points ([36, 47]). While prior work characterizes the memorized data and studies the effect of memorization to improve downstream generalization ([47]), little is known about where in SSL encoders memorization happens. The few works on localizing memorization are usually confined to supervised learning (SL) ([3, 45, 35]), or operate in the language domain ([55, 37, 7, 44, 14]). In particular, most results are coarse-grained and localize memorization on a per-layer basis [3, 45] and/or require labels [3, 35]. To close the gap, we propose two novel metrics for localizing memorization in SSL encoders in the vision domain. Our Layer Mem localizes memorization of the training data within the SSL encoders on a layer-level. For a more fine-grained localization, we turn to memorization in individual units (i.e., neurons in fully-connected layers or channels in convolutional layers). We propose Unit Mem which measures memorization of individual training data points through the units sensitivity to these points. Both our metrics can be computed independently of a downstream task, in a forward pass without gradient calculation, and without labels, which makes them computationally efficient and well-suited for the large SSL encoders pretrained on unlabeled data. By performing a systematic study on localizing memorization with our two metrics on various encoder architectures (convolutional Correspondence to boenisch@cispa.de Accepted for publication at: 38th Conference on Neural Information Processing Systems (Neur IPS 2024). and transformer-based) trained on diverse vision datasets with contrastive and non-contrastive SSL frameworks, we make the following key discoveries: Memorization happens through the entire SSL encoder. By analyzing our Layer Mem scores between subsequent layers, we find that the highest memorizing layers in SSL are not necessarily the last ones, which is in line with findings recently reported for SL [35]. While there is a tendency that higher per-layer memorization can be observed in deeper layers, similar to SL [3, 45], our analysis of memorization on a per-unit level highlights that highly memorizing units are distributed across the entire SSL encoder, and can also be found in the first layers. Units in SSL encoders experience high memorization. By analyzing SSL encoders with our Unit Mem metric, we find that a significant fraction of their units experiences high memorization of individual training data points. This stands in contrast with models trained using SL for which we observe high class memorization, measured as the unit s sensitivity to any particular class. While these results are in line with the two learning paradigms objectives where SL optimizes to separate different classes whereas SSL optimizes foremost for instance discrimination [48], it is a novel discovery that this yields significantly different memorization patterns between SL and SSL down to the level of individual units. Atypical data points cause higher memorization in layers and units. While prior work has shown that SSL encoders overall memorize atypical data points more than standard data points [47], our study reveals that the effect is constant throughout all encoder layers. Hence, there are no particular layers responsible for memorizing atypical data points, similarly as observed in SL [35]. Yet, memorization of atypical data points can be attributed on a unit-level where we observe that the highest memorizing units align with the highest memorized (atypical) data points and that overall atypical data points cause higher unit memorization than standard data points. Memorization in vision transformers happens mainly in the fully-connected layers. The memorization of transformers [46] was primarily investigated in the language domain [26, 43]. However, the understanding in the vision domain is lacking, and due to the difference in input and output tokens (language transformers operate on discrete tokens while vision transformers operate on continuous ones), the methods for analysis and the findings are not easily transferable. Yet, with our methods to localize memorization, we are the first to show that the same trend holds in vision transformers that was previously reported for language transformers, namely that memorization happens in the fully-connected layers. Finally, we investigate future applications that could benefit from localizing memorization and identify more efficient fine-tuning and memorization-informed pruning strategies as promising directions. In summary, we make the following contributions: We propose Layer Mem and Unit Mem, the first practical metrics to localize memorization in SSL encoders on a per-layer basis and down to the granularity of individual units. We perform an extensive experimental evaluation to localize memorization in various encoder architectures trained on diverse vision datasets with different SSL frameworks. Through our metrics, we gain new insights into the memorization patterns of SSL encoders and can compare them to the ones of SL models. We show that the localization of memorization can yield practical benefits for encoder fine-tuning and pruning. 2 Related Work SSL. SSL relies on large amounts of unlabeled data to train encoder models that return representations for a multitude of downstream tasks [6]. Especially in the vision domain, a wide range of SSL frameworks have recently been introduced [16, 17, 13, 4, 30, 27, 29]. Some of them rely on contrastive loss functions [16, 29, 27] whereas others train with non-contrastive objective functions [41, 17, 13, 30]. Memorization in SL. Memorization was extensively studied in SL [52, 2, 15]. In particular, it was shown that it can have a detrimental effect on data privacy, since it enables data extraction attacks [9, 10, 11]. At the same time, memorization seems to be required for generalization, in particular for long-tailed data distributions [23, 24]. It was also shown that harder or more atypical data points [2, 43] experience higher memorization. While all these works focus on studying memorization from the data perspective and concerning its impact on the learning algorithm, they do not consider where memorization happens. Memorization in SSL. Even though SSL rapidly grew in popularity during recent years, work on studying memorization in SSL is limited. Meehan et al. [36] proposed to quantify Déjà Vu memorization of SSL encoders with respect to particular data points by comparing the representations of these data points with the representations of a labeled public dataset. Data points whose k nearest public neighbors in the representation space are highly consistent in labels are considered to be memorized. Since SSL is aimed to train without labels, this approach is limited in practical applicability. More recently, Wang et al. [47] proposed SSLMem, a definition of memorization for SSL based on the leave-one-out definition from SL [23, 24]. Instead of relying on labels, this definition captures memorization through representation alignment, i.e., measuring the distance between representations of a data point s multiple augmentations. Since both works rely on the output representations to quantify memorization, neither of them is suitable for performing fine-grained localization of memorization. Yet, we use the setup of SSLMem as a building block to design our Layer Mem metric which localizes memorization per layer. Localizing Memorization. In SL, most work focuses on localizing memorization on a per-layer basis and suggests that memorization happens in the deeper layers [3, 45]. By analyzing which neurons have the biggest impact on predicting the correct label of a data point, Maini et al. [35] were able to study memorization on a per-unit granularity. They do so by zero-ing out random units until a label flip occurs. Their findings suggest that only a few units are responsible for memorizing outlier data points. Yet, due to the absence of labels in SSL, this approach is inapplicable to our work. In the language domain, a significant line of work aims at localizing where semantic facts are stored within large language transformers [55, 37, 7, 44]. Chang et al. [14] even proposed benchmarks for localization methods in the language domain. In the injection benchmark (INJ Benchmark), they fine-tune a small number of neurons and then assess whether the localization method detects the memorization in these neurons. The deletion benchmark (DEL Benchmark) first performs localization, followed by the deletion of the responsible neurons, and a final assessment of the performance drop on the data points detected as memorized in the identified neurons. Since in SSL, performance drop cannot be measured directly due to the absence of a downstream task, the deletion approach is not applicable. Instead, we verify our Unit Mem metric in a similar vein to the INJ Benchmark by fine-tuning a single unit on a data point and localizing memorization as we describe in Section 5.2. Studying Individual Units in ML Models. Early work in SL [22] already suggested that units at different model layers fulfill different functions: while units in lower layers are responsible for extracting general features, units in higher layers towards the model output are responsible for very specific features [51]. In particular, it was found that units represent different concepts required for the primary task [5], where some units focus on single concepts whilst others are responsible for multiple concepts [38, 54]. While these differences have been identified between the units of models trained with SL, we perform a corresponding investigation in the SSL domain through the lens of localizing memorization. 3 Background and Setup SSL and Notation. We consider an SSL training framework M. The encoder f : Rn Rs is pre-trained, or in short trained, on the unlabeled dataset D to output representations of dimensionality s. Throughout the training, as the encoder improves, its alignment loss LA(f, x) = d(f(x ), f(x )) between the representations of two random augmentations x , x of any training data point x decreases with respect to a distance metric d (e.g., Euclidean distance). This effect has also been observed in non-contrastive SSL frameworks [53]. We denote by f l, l [1, . . . L] the lth layer of encoder f. Data points from the test set D are denoted as x. Memorized Data. Prior work in the SL domain usually generates outliers for measuring memorization by flipping the labels of training data points [23, 24, 35]. This turns these points into outliers that experience a higher level of memorization and leave the strongest possible signal in the model. Yet, such an approach is not suitable in SSL where labels are unavailable. Therefore, we rely on the SSLMem metric proposed by [47] to identify the most (least) memorized data points for a given encoder. The findings based on the SSLMem metric indicate that the most memorized data points correspond to atypical and outlier samples. SSLMem for Quantifying Memorization. SSLMem quantifies the memorization of individual data points by SSL encoders. It is, to the best of our knowledge, the only existing method for quantifying memorization in SSL without reliance on downstream labels. SSLMem for a training data point x is defined as SSLf(x) = E f M(D) E x ,x Aug(x)d (f(x ), f(x )) ; SSLg(x) = E g M(D\x) E x ,x Aug(x)d (g(x ), g(x )) SSLMemf,g(x) = SSLg(x) SSLf(x) (1) where f and g are two classes of SSL encoders whose training dataset D differs in data point x. x and x denote two augmentations randomly drawn from the augmentation set Aug that is used during training and d is a distance metric, here ℓ2-distance. Experimental Setup. We localize memorization in encoders trained with different SSL frameworks on five common vision datasets, namely CIFAR10, CIFAR100, SVHN, STL10, and Image Net. We leverage different model architectures from the Res Net family, including Res Net9, Res Net18, Res Net34, and Res Net50. We also analyze Vision Transformers (Vi Ts) using their Tiny and Base versions. Results are reported over three independent trials. To identify the most memorized training data points, we rely on the SSLMem metric and follow the setup from [47]. More details on the experimental setup can be found in Appendix B.2 For the readers convenience, we include a glossary with short explanations for all concepts and background relevant to this work in Appendix A. 4 Layer-Level Localization of Memorization In order to localize memorization on a per-layer granularity, we propose a new Layer Mem metric which relies on the SSLMem metric, as a building block. Since the SSLMem as defined in Equation (1) is not normalized, we introduce the following normalization to the range [0, 1] SSLMem f,g(x) = SSLg(x) SSLf(x) SSLf(x) + SSLg(x) (2) such that values close to 0 denote no memorization while 1 denotes the highest memorization. This makes the score more interpretable. While SSLMem returns a memorization score per data point for a given encoder, Layer Mem returns a memorization score per encoder layer l, measured on a (sub)set D = {x1, ..., x|D |} D of training data D. Similar to SSLMem, Layer Mem makes use of a second encoder g as a reference to detect memorization as Layer Mem D (l) = 1 |D | i=1 SSLMem f l,gl(xi). (3) f l, gl denote the output of encoders f and g after layer l, respectively. Intuitively, our Layer Mem metric measures the average per-layer memorization over training data points xi D . As our Layer Mem build on SSLMem , it also inherits the above normalization. Since Equation (3) operates on different layers outputs which in turn depend on all previous layers, Layer Mem risks to report accumulated memorization up to layer l. Therefore, we also define Layer Mem D (l) for all layers l > 1 as Layer Mem D (l) = Layer Mem D (l) Layer Mem D (l 1). (4) This reports the increase in memorization of layer l with respect to the previous layer l 1. 2Our code is attached as supplementary material. 4.1 Experimental Results and Observations We present our core results and provide additional ablations on our Layer Mem in Appendix C.3. Table 1: Layer-based Memorization Scores. Res N denotes a residual connection that comes from the previous N-th convolutional layer. Layer Layer Mem LM Layer Mem Top50 LM Top50 Layer Mem Least50 1 0.091 - 0.144 - 0.003 2 0.123 0.032 0.225 0.081 0.012 3 0.154 0.031 0.308 0.083 0.022 4 0.183 0.029 0.402 0.094 0.031 Res2 0.185 0.002 0.403 0.001 0.041 5 0.212 0.027 0.479 0.076 0.051 6 0.246 0.034 0.599 0.120 0.061 7 0.276 0.030 0.697 0.098 0.071 8 0.308 0.032 0.817 0.120 0.073 Res6 0.311 0.003 0.817 0 0.086 Memorization Increases but not Monotonically. We report the Layer Mem scores in Table 1 for the Res Net9-based SSL encoder trained with Sim CLR on CIFAR10 (further per-layer breakdown and scores for Res Net18, Res Net34, and Res Net50 are presented in Table 15, Table 16, and Table 17 in Appendix C.3). We report Layer Mem across the 100 randomly chosen training data points, their Layer Mem (denoted as LM), followed by Layer Mem for only the Top 50 memorized data points, their Layer Mem (denoted as LM Top50), and Layer Mem for only the Least 50 memorized data points. The results show that our Layer Mem indeed increases with layer depth in SSL, similar to the trend observed for SL [45], i.e., deeper layers experience higher memorization than early layers. However, our Layer Mem presents the memorization from a more accurate perspective, where we discard the accumulated memorization from previous layers, including the residual connections. Layer Mem indicates that the memorization increases in all the layers but is not monotonic. We also study the differences in localization of the memorization for most memorized (outliers and atypical examples) vs. least memorized data points (inliers), shown as columns Layer Mem Top50 and Layer Mem Least50 in Table 1, respectively. While we observe that the absolute memorization for the most memorized data points is significantly higher than for the least memorized data points, they both follow the same trend of increasing memorization in deeper layers. The Layer Mem for the most memorized points (denoted as LM Top50 in Table 1) indicates that, following the overall trend, high memorization of the atypical samples is also spread over the entire encoder and not confined to particular layers. Table 2: Memorization in Vi T occurs primarily in the deeper blocks and more in the fully connected than attention layers. Vi T Block Layer Mem Layer Mem Attention Layer 2 0.028 0.008 6 0.114 0.009 12 0.281 0.010 Fully-Connected Layer 2 0.039 0.011 6 0.129 0.015 12 0.303 0.022 Memorization in Vision Transformers. The memorization of Transformers [46] was, so far, primarily investigated in the language domain [26, 43], however, its understanding in the vision domain is lacking. The fully-connected layers in language transformers were shown to act as key-value memories. Still, findings from language transformers cannot be easily transferred to vision transformers (Vi Ts) [20]: while language transformers operate on the level of discrete and interpretable input and output tokens, Vi Ts operate on continuous input image patches and output representations. Through the analysis of our newly proposed metric for memorization in SSL, in Table 2 (Vi T-Tiny trained on CIFAR10 using MAE [30]), we are the first to show that memorization in Vi Ts occurs more in deeper blocks and that within the blocks, fully-connected layers memorize more than attention layers. We present the full set of results for Layer Mem and Layer Mem over all blocks in Table 10. Memorization in Different SSL Frameworks. We also study the differences in memorization behavior between different SSL frameworks. Therefore, we compare the Layer Mem score between corresponding layers of a Res Net50 trained on Image Net with Sim CLR [16] and DINO [13], and of a Vi T-Base encoder trained on Image Net with DINO and MAE [30]. We ensure by early stopping that the resulting linear probing accuracies of the encoder pairs are similar for better comparability of their memorization. The Image Net downstream task performance within both encoder pairs is 66.12% for Sim CLR and 68.44% for DINO; and 60.43% for MAE and 60.17% for DINO. Our results in Table 4 show that encoders with the same architecture trained with different SSL frameworks experience a similar memorization pattern, namely that memorization occurs primarily in the deeper blocks/layers. In Figure 12 in Appendix C.1, we additionally show that memorization patterns between different Table 3: Consistency in 100 most memorized samples according to Layer Mem. We report the pairwise overlap between the most memorized samples and the consistency in ranking of most memorized samples using the statistical Kendall s Tau test (τ-statistic, p-value). While we observe high overlap and statistical similarity within adjacent layers, especially towards the end of the network, there is low similarity and overlap between early and late layers. Layers Overlap % τ, p Layers Overlap % τ, p Layers Overlap % τ, p Layers Overlap % τ, p 1 2 79 0.607, 4.07e-29 1 3 52 0.505, 9.08e-11 1 4 47 0.412, 6.07e-7 1 5 24 0.240, 1.18e-2 1 6 19 0.181, 6.01e-2 1 7 18 0.167, 744e-2 1 8 16 0.104, 1.27e-1 2 3 70 0.562, 8.48e-19 2 4 64 0.544, 3.95e-16 2 5 36 0.288, 2.08e-4 2 6 30 0.249, 3.96e-3 2 7 28 0.241, 9.96e-3 2 8 27 0.247, 5.53e-3 3 4 82 0.665, 4.41e-42 3 5 51 0.512, 6.67e-11 3 6 42 0.356, 8.31e-5 3 7 39 0.319, 8.31e-5 3 8 37 0.310, 1.09e-4 4 5 68 0.557, 1.61e-18 4 6 54 0.509, 6.31e-11 4 7 48 0.412, 6.67e-7 4 8 45 0.396, 2.50e-6 5 6 72 0.559, 4.11e-18 5 7 61 0.531, 4.19e-14 5 8 58 0.527, 1.08e-14 6 7 84 0.657, 1.47e-42 6 8 79 0.644, 4.17e-37 7 8 94 0.837, 9.71e-76 SSL frameworks are similar down to the individual unit level, i.e., the number of highly memorizing units and the magnitude of memorization are roughly the same. We present the full results for all Res Net50 layers and all Vi T blocks in Table 29 and Table 31, respectively. Variability and Consistency of Memorization cross Different Layers. We use Layer Mem to analyze the variability and consistency between the samples memorized by different layers in a Res Net9 vision encoder trained with CIFAR10 dataset. The results are shown in Table 3 and Figure 13 in appendix C.5. The overlap within the 100 most memorized samples between adjacent layers is usually high but decreases the further the layers are separated. Our statistical analysis to compare the similarity of the orderings within different layers most memorized samples using the Kendall s rank correlation coefficient shows that while for closer layers, we manage to reject the null hypothesis ( no correlation ) with high statistical confidence (low p-value) which is not the case for further away layers. 4.2 Verification of Layer-Based Memorization Table 4: The layer-based memorization is similar across encoders trained with different frameworks. LM=Layer Mem, LM= Layer Mem. Res Net50 Layer Sim CLR DINO Number LM LM LM LM 2 0.040 0.003 0.041 0.002 27 0.161 0.005 0.165 0.006 49 0.302 0.008 0.311 0.007 Vi T-Base Block MAE DINO Number LM LM LM LM 2 0.037 0.010 0.036 0.011 6 0.120 0.015 0.116 0.014 12 0.274 0.019 0.271 0.019 To analyze whether our Layer Mem metric and its variant indeed localize memorization correctly, we first replace different layers of an encoder and then compute linear probing accuracy on various downstream tasks. Since prior work shows that memorization in SSL is required for downstream generalization [47], we expect the highest performance drop when replacing the layers identified as most memorizing. Our results in Appendix C.7 verify this intuition. They show that by replacing the most memorizing layers of an encoder trained on a dataset A, e.g., CIFAR10, with the equivalent layers of another dataset B, e.g., STL10, the linear probing accuracy drop for CIFAR10 is significantly larger than when when replacing random or least memorizing layers. Surprisingly, at the same time, the replacement of the most memorizing layers from the CIFAR10 trained encoder with STL10 layers also causes the highest increase in STL10 linear probing accuracy (again in comparison to replacing random or least memorizing layers). See a full set of results for replacing any combination of 1, 2, and 3 layers in Table 30, Table 32, and Table 33, respectively. These results suggest that we might be able to improve standard encoder fine-tuning by localizing the most memorizing layers and fine-tuning these instead of the last layer(s) currently the standard practice for fine-tuning in SSL. We verify this assumption in Table 5 and show that fine-tuning the most memorizing layers indeed yields the highest downstream performance on the fine-tuning dataset. This shows that localizing memorization might have practical application for more efficient fine-tuning in the future. 5 Unit-Level Localization of Memorization Experiments from the previous section highlight that we are able to localize the memorization of data points in particular layers of the SSL encoders. This raises the even more fundamental question on Table 5: Fine-tuning most memorizing layers. We train a Res Net9 encoder with Sim CLR on CIFAR10 and fine-tune different (combinations of) layers on the STL10 dataset, resized to 32x32x3. We train a linear layer trained on top of the encoder (HEAD) and report STL10 test accuracy after fine-tuning. Fine-tuning the most memorizing layer(s), in contrast to the last layer(s), yields higher fine-tuning results. Fine-tuned Layers Accuracy (%) None (HEAD) 48.6% 1.12% 6 (highest Layer Mem) + HEAD 53.0% 0.86% 8 (last layer, highest Layer Mem) + HEAD 52.7% 0.97% 6,8 + HEAD 56.7% 0.84% 7,8 + HEAD 55.3% 0.77% 4,6,8 (highest Layer Mem) + HEAD 57.9% 0.79% 6,7,8 + HEAD 56.5% 0.95% whether it is possible to trace down SSL memorization to a unit-level. To answer this question, we design Unit Mem, a new metric to localize memorization in individual units of SSL encoders. We use the term unit to refer to both an activation map from a convolutional layer (single-layer output channel) or an individual neuron within a fully connected layer. Our Unit Mem metric quantifies for every unit u in the SSL encoder how much u is sensitive to, i.e., memorizes, any particular training data point. Therefore, Unit Mem relates the maximum unit activation that occurs for a data point xk in the training data (sub)set D D with the mean unit activation on all other data points in D \ {xk}. The design of Unit Mem is inspired by the class selectivity metric defined for SL by [39]. Class selectivity was derived from selectivity indices commonly used in neuroscience [19, 8, 25] and quantifies a unit s discriminability between different classes. It was used as an indicator of good generalization in SL. We provide more background in Appendix D.1. To leverage ideas from class selectivity for identifying memorization, we integrate three fundamental changes in our metric in comparison to the class selectivity metric. While class selectivity is calculated on classes of the test set and relies on class labels, our Unit Mem is (1) label-agnostic and (2) computed on individual data points from the training dataset to determine their memorization. Additionally, to account for SSL s strong reliance on augmentations, (3) we calculate Unit Mem over the expectation on the augmentation set used during training. Research from the privacy community [49, 34] suggests that those augmentations leave a stronger signal in ML models than the original data point, i.e., relying on the unaugmented point alone might under-report memorization. We verify this effect in Figure 5 in Appendix C.1. We note that through these fundamental changes Unit Mem is able to measure memorization of individual data points within a class rather than to solely distinguish between classes or concepts like the original class selectivity. We provide further insights into this difference and perform experimental verification which highlights that Unit Mem captures individual data points memorization rather than capturing classes or concepts in Appendix C.2. To formalize our Unit Mem, we first define the mean activation µ of unit u on a training point x as µu(x) = E x Aug(x)activationu(x ), (5) where the activation for convolutions feature maps is averaged across all elements of the feature map and for fully connected layers is an output from a single neuron (which is averaged across all patches of x in Vi Ts). Further, for the unit u, we compute the maximum mean activation µmax,u across all instances from D , where N = |D |, as µmax,u = max({µu(xi)}N i=1). (6) Let k be the index of the maximum mean activation µu(xk), i.e., the argmax. Then, we calculate the corresponding mean activity µ max across all the remaining N 1 instances from D as µ max,u = mean({µu(xi)}N i=1,i =k). (7) Finally, we define the Unit Mem of unit u as Unit Mem D (u) = µmax,u µ max,u µmax,u + µ max,u . (8) (a) Different datasets. (b) Unit Mem vs SSLMem. (c) Unit Mem with DP. Figure 1: Insights into Unit Mem. We train a Res Net9 encoder with Sim CLR: (a) Different datasets, including SVHN, CIFAR10, and STL10. We report the Unit Mem of the last convolutional layer (conv4_2); (b) Comparing alignment between SSLMem and Unit Mem on CIFAR10. Data points with higher general memorization (SSLMem) tend to experience higher Unit Mem; (c) Using different strengths of privacy protection according to DP during training on CIFAR10 and Vit-Base The value of the Unit Mem metric is bounded between 0 and 1, where 0 indicates that the unit is equally activated by all training data points, while value 1 denotes exclusive memorization, where only a single data point triggers the activation, while all other points leave the unit inactive. 5.1 Experimental Results and Observations We present our core results and provide detailed additional ablations on our Unit Mem Appendix C.1. Highly Memorizing Units Occur over Entire Encoder. Our analysis highlights that over all encoder architectures and SSL training frameworks, highly memorizing units are spread over the entire encoder. While, on average, earlier layers exhibit lower Unit Mem than deeper layers, even the first layer contains highly memorizing units as shown in Figure 3 (first row). Figure 1a shows that this trend holds over different datasets. Yet, the SVHN dataset, which is visually less complex than the CIFAR10 or STL10 dataset, has the lowest number of highly memorizing units. This observation motivates us to study the relationship between the highest memorized (atypical or hard to learn) data points and the highest memorizing units. Most Memorized Samples and Units Align. To draw a connection between data points and unit memorization, we analyze which data points are responsible for the highest µmax scores. This corresponds to a data point which causes the highest activations of a unit, while other points activate the unit only to a small degree or not at all. We show the results in Figure 1b (also in Table 12 as well as in Figure 7 in Appendix C.1). For each unit u in the last convolutional layer of the Res Net9 trained on CIFAR10, we measure its Unit Mem score, then we identify which data point is responsible for the unit s µmax,u, and finally measure this point s SSLMem score. We plot the Unit Mem and SSLMem scores for each unit and its corresponding point. Our results highlight that the data points that experience the highest memorization according to the SSLMem score are also the ones memorized in the most memorizing units. Given the strong memorization in individual units, we next look into two methods to reduce it and analyze their impact. Differential Privacy reduces Unit Memorization. The gold standard to guarantee privacy in ML is Differential Privacy (DP) [21]. DP formalizes that any training data point should only have a negligible influence on the final trained ML model. To implement this, individual data points gradients during training are clipped to a pre-defined norm, and controlled amounts of noise are added [1]. This limits the influence that each training data point can have on the final model. Building on the DP framework for SSL encoders [50], we train a Vi T-Tiny using MAE on CIFAR10 with three different privacy levels in DP usually indicated with ε. We train non-private (ε = ), little private (ε = 20), and highly private (ε = 8) encoders and apply our Unit Mem to detect and localize memorization. Our results in Figure 1c highlight that while with increasing privacy levels, the average Unit Mem decreases, there are still individual units that experience high memorization. Data Point vs Class Memorization. Since stronger training augmentations yield higher class clustering [31] (i.e., the fact that data points from the same downstream class are close to each other in representation space but distant to data points from other classes), we also analyze how the SSL encoders differ from the standard class discriminators, namely SL trained models. Therefore, we Figure 3: Significantly more (less) units memorize data points rather than classes in SSL (SL). We measure the Class Mem vs Unit Mem for 10000 samples from CIFAR100, with 100 random samples per class. Each i-th column represents the i-th convolutional layer in Res Net9, with 8 convolution layers, where the 1st row is for SSL while the 2nd row for SL. The red diagonal line denotes y = x. go beyond our previous experiments that measure memorization of units with respect to individual data points and additionally study unit memorization at a class-granularity. Therefore, we adjust the class selectivity metric from [39] to perform on the training dataset rather than on the test data set as the original class selectivity. To avoid confusion between the two versions, refer to our adapted metric as Class Mem (see Appendix D.2 for an explicit definition). Equipped with Unit Mem and Class Mem, we study the behavior of SSL encoders and compare between SSL and SL. For our comparison, we train an encoder with Sim CLR and a model with SL using the standard cross entropy loss, both on the CIFAR100 dataset using Res Net9. We remove the classification layer from the SL trained model to obtain the same architecture as for the encoder trained with Sim CLR. Figure 2: Unit Mem and Class Mem for SL and SSL. For comparability, we early stop the SL training once it reaches a comparable performance to the linear probing accuracy on CIFAR100 obtained by the SSL encoder. Our results in Figure 2 show that overall, in SSL throughout all layers, average memorization of individual data points is higher than class memorization, whereas in SL, in deeper layers, the class memorization increases significantly. We hypothesize that this effect is due to earlier layers in SL learning more general features which are independent of the class whereas later layers learn features that are highly class dependent. For SSL, such a difference over the network does not seem to exist; both scores increase slightly, however, probably due to the SSL learning paradigm, memorization of individual data points remains higher. To better understand the memorization of units on the micro level, we investigate further the individual units in each layer. In Figure 3, we plot the Class Mem vs Unit Mem for each unit and in each of the eight encoder layers of Res Net9. Most units for SSL (row 1) constantly exhibit higher Unit Mem than Class Mem, i.e., they cluster under the diagonal line, which suggests that most units memorize individual data points across the whole network. Contrary, the initial layers for the model trained with SL have a slight tendency to memorize data points over entire classes whereas in later layers, this trend drastically reverses and most units memorize classes. In Appendix C.4, we investigate how this different memorization behavior between SL and SSL affects downstream generalization. 5.2 Verification of Unit-Based Memorization We verify the unit-based localization of memorization with our Unit Mem by deliberately inserting memorization of particular units and checking if our Unit Mem correctly detects it. Therefore, we first train a Sim Siam-based [17] Res Net18 encoder trained on the CIFAR10 dataset. We select Sim Siam over Sim CLR for this experiment since Sim CLR, as a contrastive SSL framework, cannot train on a single data point. Then, using Layer Mem, we identity the last convolutional layer in Res Net18 (i.e., layer 4.1.conv2) as the layer with the highest Layer Mem and Layer Mem memorization. We select the unit from the layer with the highest µmax and also pick a unit with no activation (µmax = 0) for some test data points. Then, we fine-tune these units using a single test data point and report the change in Unit Mem for the chosen units in Table 13. The results show that our Unit Mem correctly detects the increase in memorization in both units. Additionally, we analyze the impact of zero-ing out the most or least memorizing vs. random units. Again with the argument that memorization is required for downstream generalization in SSL [47], we expect the highest performance drop when Table 6: Removing the least/most memorized units according to Unit Mem preserves most/least linear probing performance. We prune according to units with highest or lowest Unit Mem either per layer or for the entire network (total). We also present baselines where we prune randomly selected units. The standard deviation for this baseline is reported over 10 independent trials where different random units were pruned. We train the Res Net9 encoder using CIFAR10 and compute the Unit Mem score using 5000 data points from the train set. Pruning % of Selected Downstream Accuracy (%) Strategy Units CIFAR10 SVHN STL10 No Pruning - 70.44 78.22 69.12 Top Unit Mem per layer 10 53.04 63.84 50.94 Random per layer 10 58.09 1.76 67.04 2.44 55.71 2.18 Low Unit Mem per layer 10 62.58 72.26 59.26 Top Unit Mem per layer 20 48.30 55.88 43.18 Random per layer 20 51.34 1.21 58.01 1.34 46.74 0.97 Low Unit Mem per layer 20 54.84 62.60 50.02 Top Unit Mem total 10 49.16 61.28 47.30 Random total 10 56.77 2.09 67.09 1.56 53.89 2.33 Low Unit Mem total 10 62.62 72.28 59.30 zero-ing out the most memorizing units. Our results in Table 6 in and in Appendix C.1 confirm this hypothesis and show that removing the most memorizing units yields the highest loss in linear probing accuracy on various downstream tasks while pruning the least memorized units preserves better downstream performance than removing random units. These results suggest that future work may benefit from using our Unit Mem metric for finding which units within a network can be pruned while preserving high performance. 6 Conclusions We propose the first practical metrics for localizing memorization within SSL encoders on a per-layer and per-unit level. By analyzing different SSL architectures, frameworks, and datasets using our metrics, we find that while memorization in SSL increases in deeper layers, a significant fraction of highly memorizing units can be encountered over the entire encoder. Our results also show that SSL encoders significantly differ from SL trained models in their memorization patterns, with the former constantly memorizing data points and the latter increasingly memorizing classes. Finally, using our metrics for localizing memorization presents itself as an interesting direction towards more efficient encoder fine-tuning and pruning. Acknowledgments The project on which this paper is based was funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation), Project number 550224287. Additional funding came from the Initiative and Networking Fund of the Helmholtz Association in the framework of the Helmholtz AI project call under the name PAFMIM , funding number ZT-I-PF-5-227. Responsibility for the content of this publication lies with the authors. [1] Martin Abadi, Andy Chu, Ian Goodfellow, H Brendan Mc Mahan, Ilya Mironov, Kunal Talwar, and Li Zhang. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC conference on computer and communications security, pages 308 318, 2016. [2] Devansh Arpit, Stanisław Jastrzebski, Nicolas Ballas, David Krueger, Emmanuel Bengio, Maxinder S Kanwal, Tegan Maharaj, Asja Fischer, Aaron Courville, Yoshua Bengio, et al. A closer look at memorization in deep networks. In International conference on machine learning, pages 233 242. PMLR, 2017. [3] Robert Baldock, Hartmut Maennel, and Behnam Neyshabur. Deep learning through the lens of example difficulty. Advances in Neural Information Processing Systems, 34:10876 10889, 2021. [4] Adrien Bardes, Jean Ponce, and Yann Le Cun. VICReg: Variance-invariance-covariance regularization for self-supervised learning. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=xm6YD62D1Ub. [5] David Bau, Bolei Zhou, Aditya Khosla, Aude Oliva, and Antonio Torralba. Network dissection: Quantifying interpretability of deep visual representations. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 6541 6549, 2017. [6] Yoshua Bengio, Aaron Courville, and Pascal Vincent. Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8): 1798 1828, 2013. doi: 10.1109/TPAMI.2013.50. [7] Steven Bills, Nick Cammarata, Dan Mossing, Henk Tillman, Leo Gao, Gabriel Goh, Ilya Sutskever, Jan Leike, Jeff Wu, and William Saunders. Language models can explain neurons in language models. URL https://openaipublic. blob. core. windows. net/neuronexplainer/paper/index. html.(Date accessed: 14.05. 2023), 2023. [8] K. H. Britten, M. N. Shalden, W. T. Newsome, and J. A. Movshon. The analysis of visual motion: A comparison of neuronal and psychophysical performance. Journal of Neuroscience, 12:4745 4765, 1992. [9] Nicholas Carlini, Chang Liu, Úlfar Erlingsson, Jernej Kos, and Dawn Song. The secret sharer: Evaluating and testing unintended memorization in neural networks. In 28th USENIX Security Symposium (USENIX Security 19), pages 267 284, 2019. [10] Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, et al. Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21), pages 2633 2650, 2021. [11] Nicholas Carlini, Matthew Jagielski, Chiyuan Zhang, Nicolas Papernot, Andreas Terzis, and Florian Tramer. The privacy onion effect: Memorization is relative. Advances in Neural Information Processing Systems, 35:13263 13276, 2022. [12] Nicolas Carlini, Jamie Hayes, Milad Nasr, Matthew Jagielski, Vikash Sehwag, Florian Tramer, Borja Balle, Daphne Ippolito, and Eric Wallace. Extracting training data from diffusion models. In 32nd USENIX Security Symposium (USENIX Security 23), pages 5253 5270, 2023. [13] Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pages 9650 9660, 2021. [14] Ting-Yun Chang, Jesse Thomason, and Robin Jia. Do localization methods actually localize memorized data in llms? ar Xiv preprint ar Xiv:2311.09060, 2023. [15] Satrajit Chatterjee. Learning and memorization. In International conference on machine learning, pages 755 763. PMLR, 2018. [16] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597 1607. PMLR, 2020. [17] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 15750 15758, 2021. [18] Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, pages 215 223. JMLR Workshop and Conference Proceedings, 2011. [19] Russell L. de Valois, E. William Yund, and Norva Hepler. The orientation and direction selectivity of cells in macaque visual cortex. Vision Research, 22:531 544, 1982. URL https://api.semanticscholar.org/Corpus ID:33506510. [20] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=Yicb Fd NTTy. [21] Cynthia Dwork. Differential privacy. In International colloquium on automata, languages, and programming, pages 1 12. Springer, 2006. [22] Dumitru Erhan, Yoshua Bengio, Aaron Courville, and Pascal Vincent. Visualizing higher-layer features of a deep network. 2009. [23] Vitaly Feldman. Does learning require memorization? a short tale about a long tail. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pages 954 959, 2020. [24] Vitaly Feldman and Chiyuan Zhang. What neural networks memorize and why: Discovering the long tail via influence estimation. Advances in Neural Information Processing Systems, 33: 2881 2891, 2020. [25] David Freedman and John Assad. Experience-dependent representation of visual categories in parietal cortex. Nature, 443:85 8, 10 2006. doi: 10.1038/nature05078. [26] Mor Geva, Roei Schuster, Jonathan Berant, and Omer Levy. Transformer feed-forward layers are key-value memories. In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wentau Yih, editors, Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 5484 5495, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.emnlp-main.446. URL https://aclanthology.org/2021.emnlp-main.446. [27] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, 33:21271 21284, 2020. [28] Kaiming He, X. Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770 778, 2015. URL https://api.semanticscholar.org/Corpus ID:206594692. [29] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 9729 9738, 2020. [30] Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Dollár, and Ross Girshick. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16000 16009, 2022. [31] Weiran Huang, Mingyang Yi, Xuyang Zhao, and Zihao Jiang. Towards the generalization of contrastive self-supervised learning. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=XDJwu EYHhme. [32] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009. [33] Hongbin Liu, Jinyuan Jia, Wenjie Qu, and Neil Zhenqiang Gong. Encodermi: Membership inference against pre-trained encoders in contrastive learning. In Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security, pages 2081 2095, 2021. [34] Hongbin Liu, Jinyuan Jia, Wenjie Qu, and Neil Zhenqiang Gong. Encodermi: Membership inference against pre-trained encoders in contrastive learning. In Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security, pages 2081 2095, 2021. [35] Pratyush Maini, Michael C Mozer, Hanie Sedghi, Zachary C Lipton, J Zico Kolter, and Chiyuan Zhang. Can neural network memorization be localized? ar Xiv preprint ar Xiv:2307.09542, 2023. [36] Casey Meehan, Florian Bordes, Pascal Vincent, Kamalika Chaudhuri, and Chuan Guo. Do ssl models have déjà vu? a case of unintended memorization in self-supervised learning. ar Xiv e-prints, pages ar Xiv 2304, 2023. [37] Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. Locating and editing factual associations in gpt. Advances in Neural Information Processing Systems, 35:17359 17372, 2022. [38] Ari S Morcos, David GT Barrett, Neil C Rabinowitz, and Matthew Botvinick. On the importance of single directions for generalization. In International Conference on Learning Representations, 2018. [39] Ari S. Morcos, David G.T. Barrett, Neil C. Rabinowitz, and Matthew Botvinick. On the importance of single directions for generalization. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=r1iu Qjx CZ. [40] Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y Ng. Reading digits in natural images with unsupervised feature learning. 2011. [41] Ashwini Pokle, Jinjin Tian, Yuchen Li, and Andrej Risteski. Contrasting the landscape of contrastive and non-contrastive learning. In Gustau Camps-Valls, Francisco J. R. Ruiz, and Isabel Valera, editors, Proceedings of The 25th International Conference on Artificial Intelligence and Statistics, volume 151 of Proceedings of Machine Learning Research, pages 8592 8618. PMLR, 28 30 Mar 2022. URL https://proceedings.mlr.press/v151/pokle22a.html. [42] Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, et al. Imagenet large scale visual recognition challenge. International journal of computer vision, 115:211 252, 2015. [43] Ildus Sadrtdinov, Nadezhda Chirkova, and Ekaterina Lobacheva. On the memorization properties of contrastive learning. ar Xiv preprint ar Xiv:2107.10143, 2021. [44] Chandan Singh, Aliyah R Hsu, Richard Antonello, Shailee Jain, Alexander G Huth, Bin Yu, and Jianfeng Gao. Explaining black box text modules in natural language with language models. ar Xiv preprint ar Xiv:2305.09863, 2023. [45] Cory Stephenson, Abhinav Ganesh, Yue Hui, Hanlin Tang, Sue Yeon Chung, et al. On the geometry of generalization and memorization in deep neural networks. In International Conference on Learning Representations, 2020. [46] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention is all you need. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL https://proceedings.neurips.cc/paper_files/paper/2017/file/ 3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf. [47] Wenhao Wang, Muhammad Ahmad Kaleem, Adam Dziedzic, Michael Backes, Nicolas Papernot, and Franziska Boenisch. Memorization in self-supervised learning improves downstream generalization. In The Twelfth International Conference on Learning Representations (ICLR), 2024. [48] Yifei Wang, Qi Zhang, Yisen Wang, Jiansheng Yang, and Zhouchen Lin. Chaos is a ladder: A new theoretical understanding of contrastive learning via augmentation overlap. In International Conference on Learning Representations, 2021. [49] Da Yu, Huishuai Zhang, Wei Chen, Jian Yin, and Tie-Yan Liu. How does data augmentation affect privacy in machine learning? In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 10746 10753, 2021. [50] Yaodong Yu, Maziar Sanjabi, Yi Ma, Kamalika Chaudhuri, and Chuan Guo. Vip: A differentially private foundation model for computer vision. ar Xiv preprint ar Xiv:2306.08842, 2023. [51] Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In Computer Vision ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part I 13, pages 818 833. Springer, 2014. [52] Chiyuan Zhang, Samy Bengio, Moritz Hardt, Benjamin Recht, and Oriol Vinyals. Understanding deep learning requires rethinking generalization. In International Conference on Learning Representations, 2016. [53] Qi Zhang, Yifei Wang, and Yisen Wang. How mask matters: Towards theoretical understandings of masked autoencoders. Advances in Neural Information Processing Systems, 35:27127 27139, 2022. [54] Bolei Zhou, Yiyou Sun, David Bau, and Antonio Torralba. Revisiting the importance of individual units in cnns via ablation. ar Xiv preprint ar Xiv:1806.02891, 2018. [55] Chen Zhu, Ankit Singh Rawat, Manzil Zaheer, Srinadh Bhojanapalli, Daliang Li, Felix Yu, and Sanjiv Kumar. Modifying memories in transformer models. ar Xiv preprint ar Xiv:2012.00363, 2020. For the reader s convenience, we provide a glossary with all important terms and concepts related to our work in Table 7. Table 7: Glossary. We present a concise overview on the concepts relevant to this work. Concept Explanation Atypical examples Data points that are uncommon in the data distribution and different in terms of their features. Examples: Figure 1 from [47]. Sometimes also called outliers . Class Selectivity A metric proposed by [39] which quantifies a unit s discriminability between different classes, measured on the test data. Class Mem Our adaptation of Class Selectivity measured on the training data. Downstream Generalization Expresses how well an encoder is suited to solve some downstream tasks. For classification, it is often measured by linear probing, i.e., training an additional classification layer on top of the encoder output. Layer Mem Our proposed metric to quantify memorization of any layer in the SSL encoder. Memorization A phenomenon where a machine learning model stores detailed information on its training data. Memorized Data Point A data point that experiences high memorization by a machine learning model. Memorization Pattern A general trend in the low-level memorization of an SSL encoder, i.e., in which layers or units do memorization localize. Unit Term used to refer to an individual neuron in fully connected layers or a channel in convolutional layers. Unit Mem Our proposed metric to quantify the memorization of any unit in the SSL encoder. B Experimental Setup Datasets. We base our experiments on Image Net ILSVRC-2012 [42], CIFAR10 [32], CIFAR100 [32], SVHN [40], and STL10 [18]. Models. We use the Res Net family of models [28], including Res Net9, Res Net18, Res Net34, and Res Net50. In Table 8, we present the detailed architecture of the Res Net9 model. Table 8: Architecture of Res Net9. In the Number of Units column, we present the number of activation maps (corresponding to individual filters in the filter bank). Conv-Layer ID Layer Name Number of Units Number of Parameters 1 Conv1 32 896 - BN1 32 64 - Max Pool1 32 0 2 Conv2-0 64 18496 - BN2-0 64 128 - Max Pool2-0 64 0 3 Conv2-1 64 36928 - BN2-1 64 128 - Max Pool2-1 64 0 4 Conv2-2 64 36928 - BN2-2 64 128 - Max Pool2-2 64 0 5 Conv3 128 73856 - BN3 128 256 - Max Pool3 128 0 6 Conv4-0 256 295168 - BN4-0 256 512 - Max Pool4-0 256 0 7 Conv4-1 256 590080 - BN4-1 256 512 - Max Pool4-1 256 0 8 Conv4-2 256 590080 - BN4-2 256 512 - Max Pool4-2 256 0 SSL Frameworks. We base our experimentation on four state-of-art SSL encoders: MAE [30], Sim CLR [16], DINO [13], and Sim Siam [17]. Training Setup. Our experimental setup for training the encoders mainly follows [47] and we rely on their naming conventions and refer to the data points that are used to train encoder f, but not reference encoder g as candidate data points. In total, we use 50000 data points as training samples for CIFAR10, SVHN, and STL10 and 100000 for Image Net with 5000 candidate data points per dataset. The encoders evaluated in the paper are trained with batch size 1024, and trained 600 epochs for CIFAR10, SVHN, and STL10, and 300 epochs for Image Net. We set the batch size to 1024 for all our experiments and train for 600 epochs on CIFAR10, SVHN, and STL10, and for 300 epochs on Image Net. As a distance metric to measure representation alignment, we use the ℓ2 distance. We repeat all experiments with three independent seeds and report average and standard deviation. For reproducibility, we detail our full setup in Table 9 with the standard parameters that are used throughout the paper if not explicitly specified otherwise. Training Augmentations. We generate augmentations at random from the following augmentation sets (p indicates augmentation probability): SL, standard, (referred to as weak augmentations): Color Jitter(0.9-0.9-0.9-0.5, p=0.4), Random Horizontal Flip(p=0.5), Random Grayscale(p=0.1), Random Resized Crop(size=32) SSL, standard, (referred to as normal augmentations): Color Jitter(0.8-0.8-0.8-0.2, p=0.8), Random Horizontal Flip(p=1.0), Random Grayscale(p=0.2), Random Resized Crop(size=32) SSL, stronger, (referred to as strong augmentations): Color Jitter(0.8-0.8-0.8-0.2, p=0.9), Random Horizontal Flip(p=1.0), Random Grayscale(p=0.5), Random Resized Crop(size=32), Random Vertical Flip(p=1.0) SSL (independent): Gaussian Blur(kernel_size=(4, 4), sigma=(0.1, 5.0), p=0.8), Random Invert(p=0.2), Random Resized Crop(size=32), Random Vertical Flip(p=1.0) Masking (MAE): 75% random masking Details on Computing Unit Mem. Relying on insights from [33], we calculate Equation (5) for the activations within Unit Mem over ten augmentations since this has shown to yield a strong signal on the augmented data point. For convolutional feature maps, the activation of the unit is calculated as the average of all elements in the feature map. In Vi Ts, where we measure activation over fully-connected layers, we compute the activation per neuron and average across all patches of a given input. For example, for Vi T Tiny encoder pretrained on CIFAR10, the input image of resolution 32x32 is patchified into 64 patches, each of size 4x4. Then, each patch is represented by a 192 dimensional embedding. The classification (CLS) embedding is prepended to the remaining 64 embeddings. Overall, we obtain 65 patches. The last fully connected layer has 192 neurons. For each neuron, we average its activations across the 65 patches. In the case of Vi T Base, we have 768-dimensional embeddings and 197 patches for the input image of resolution 224x224. Details on Fine-Tuning with one Test Data Point We provide the exact details of our experiments to verify our Unit Mem through deliberate insertion of memorization in Section 5.2. We train a Sim Siam-based [17] Res Net18 encoder on the CIFAR10 dataset and use Layer Mem to identity layer 4.1.conv2, i.e., the last convolutional layer in Res Net18, as the layer with highest accumulated memorization. We select the unit from the layer with the highest µmax and also pick a unit with no activation (µmax = 0) for some test data points. Then, for compatability with pytorch which does not support individual unit training, we lock all parameters except for the targeted layer and train the model with a single sample from the testing dataset. We choose the sample that achieves the highest activation µmax on the unit with the highest Unit Mem. We save the checkpoints after each epoch and test the µmax for the selected two units. Our results in Table 13 show that the value of µmax for the selected data point increases in both units and the data point remains the one responsible for the µmax. Details on Hardware resources usage We finish all our experiments on two devices: a cloud server with four A100 GPUs and a local workstation with Intel 13700k CPU, Nvidia 4090 graphics card and 64GB of RAM Table 9: Training Setup for SSL Frameworks and Hyperparameters. Two numbers denote Image Net / Others. Model Training Linear Probing MAE Sim CLR DINO Sim Siam MAE Sim CLR DINO Sim Siam Training Epochs 300 / 600 300 / 600 300 / 600 - / 200 45 / 90 45 / 90 45 / 90 - / 30 Warm-up Epochs 30 / 60 30 / 60 30 / 60 - / 24 5 / 10 5 / 10 5 / 10 - / 3 Batch Size 2048 4096 1024 128 4096 4096 4096 256 Optimizer Adam W LARS Adam W SGD LARS LARS LARS SGD Learning rate 1.2e-3 4.8 2e-3 2.5e-2 1.6 4.8 1.6 5e-2 Learning rate Schedule Cos. Decay Cos. Decay Cos. Decay Cos. Decay Cos. Decay Cos. Decay Cos. Decay Cos. Decay C Additional Experiments C.1 Additional Insights into Unit Mem Unit Mem increases over training. First, we assess how Unit Mem evolves over training of the SSL encoder. Therefore, we train a Res Net9 encoder using Sim CLR on the CIFAR10 dataset for 800 epochs, using 120 warm-up epochs. Every five epochs, we measure the Unit Mem. Our results in Figure 4 depict the average Unit Mem of the Res Net9 s last convolutional layer. Figure 4: Average Unit Mem of layer 8 over training. We observe that the unit memorization monotonically increases throughout training and that the increase is particularly high during the first epochs. After the warm-up, we observe that the increase in unit memorization stagnates until the level of memorization on the unit level converges. The same trend can be observed over all layers which indicates that SSL encoders increase unit memorization throughout training. Measuring Unit Mem without using augmentations leads to an under-reporting of memorization. Figure 5: Unit Mem w & w/o augmentations. To assess the impact on using augmentation to implement Equation (5) for the calculation of our Unit Mem has an impact on the reported results, we train two Res Net9 models on the CIFAR100 dataset, one using Sim CLR, the other one using standard SL with cross entropy loss. During training we rely on the standard augmentations for SL and SSL reported above. To measure memorization, we once use ten augmentations from the training augmentation set, and no augmentations otherwise and report the results in Figure 5. We find that while the trend of the reported memorization is equal in both settings, the Unit Mem measured without augmentations remains constantly lower than when measured with augmentations. This suggests that when measuring Unit Mem, it is important to use augmentations to avoid under-reporting of the memorization. Figure 6: Size of D . The number of data points used to measure Unit Mem does not have a significant impact on the reported memorization. Using the same Res Net9, trained with Sim CLR on CIFAR100, we assess whether the number of data points that we use to calculate Unit Mem (the size of D ) has an impact on the reported memorization. Then, we measure Unit Mem using 100 random data point chosen one from each class in CIFAR100, 100 purely randomly chosen data points, and randomly chosen CIFAR100 data points. We present our findings in Figure 6. Our results highlight that all the lines are within each other s standard deviation, indicating that there is no significant difference in the reported Unit Mem, dependent on the make up of the dataset D . Most memorized data points align with the most memorizing units. We train a Res Net9 on CIFAR10 using Sim CLR and measure Unit Mem for the 300 most and 300 least memorized data points identified using SSLMem by [47]. The measurement of the two sets (most vs least memorized data points) is performed independently. Our results in Figure 7 show that the Unit Mem calculated on the most memorized data points is significantly higher than on the least memorized data points (we verify the significance with a statistical t-test in Table 11. Figure 7: Least vs most memorized data points. While also some of the least memorized data points lead to a high activation of the units, highest activation (on average and in particular) can be observed for the most memorized data points. This underlines the trend observed in Table 12 which shows that highly memorized data points align with the highly memorizing units. Computing Unit Mem based on the median yields similar results to using the mean. Our Unit Mem metric is inspired by the class selectivity defined for SL by [39] which quantifies a unit s discriminability between different classes, see Appendix D.1. Yet, we calculate the µ max,u in Equation (7) using the median on the other individual training data points activations while Class Selectivity computes their equivalent of µ max,u using the mean on all other test classes activations. (b) Mean Figure 8: Mean vs Median. We show in Figure 8 over the 300 most and least memorized data points for a Res Net9 trained with Sim CLR on CIFAR10 that using the median for Unit Mem yields very similar results to using the mean. For SSL, the concrete augmentation set has no strong impact when measuring Unit Mem. We additionally set out to study the impact of the augmentation set used to calculate Unit Mem. Therefore, we calculate Unit Mem on the Res Net9 trained on CIFAR10 using Sim CLR using different augmentations sets. For SSL, we measure once with the standard training augmentations ("Normal"), with an independent set of augmentations of similar strength ("Independent"), with a weaker augmentation set for which we rely on the augmentations used to train the SL model ("Weak"), and an independent very strong set of augmentations modeled after MAE training and using a masking of 75% of the input image ("Masking"). Our results in Figure 9 depict the Unit Mem over the last convolutional layer of the Res Net9 encoders. Figure 9: Different augmentation sets. They highlight that the weak and independent augmentations report extremely similar Unit Mem to the original set of training augmentations used. For SL, the impact of using different augmentations during training and measuring Unit Mem is more expressed. We also measured for a weak augmentation set ("Normal"), an independent weak set ("Independent"), strong augentations for which we relied on the standard SSL augmentations ("Strong"), and the 75% masking ("Masking"). We observe that using the augmentations from training to calculate Unit Mem yields the highest localization of memorization. Figure 10: Different augmentation sets. Stronger augmentations reduce memorization. We also analyze how the training augmentation strength can impact the final encoder s Unit Mem. We use Color Jitter, Horizontal Flip, Random Grayscale, and Random Resized Crop as augmentations. Their strength is determined by the probability of applying them and their level of distortion. In Appendix B, we present the exact parameters specified for each of them under different strengths. Our results in Figure 10 suggest that stronger augmentations yield lower perunit memorization. These findings are in line with prior theoretical work on SSL [48] highlighting that SSL performs foremost the task of instance discrimination (i.e., differentiating between individual images), but achieves clustering according to classes due to the augmentations: with stronger augmentations, multiple data points augmented views will look extremely similar (e.g., the wheels of two different images of cars), such that they eventually activate the same unit. Thereby, this unit memorizes individual data points less while units trained with weaker augmentations depend on and memorize individual data points more. Note that we do not observe a strong dependency of our reported Unit Mem on the concrete augmentation set used to calculate the metric (see Equation (5)) as we show in Figure 9 in Appendix C.1. Yet, using the original set of training augmentations, as we do for our experiments, yields the strongest signal. Figure 11: Different weight decay Stronger weight decay reduces memorization. To analyze how training weight decay affects the final encoder s Unit Mem, we train a Res Net9 using Sim CLR on CIFAR10 with three different levels of weight decay. Our results in Figure 11 show that stronger weight decay yields lower memorization, yet also decreases linear probing accuracy. Different SSL frameworks yield similar memorization pattern. We compare the Unit Mem score between corresponding layers of a Res Net50 pre-trained on Image Net with Sim CLR and DINO, as well as for Vi T-Base encoders pre-trained on Image Net with DINO and MAE. We ensure that the number of epochs, batch sizes, training dataset sizes, and the resulting linear probing accuracies of the encoders are similar for direct comparability. Our results in Figure 12 depict the Unit Mem of the last convolutional layer of the Res Net50, and the final block s fully-connected layer in the Vi T. The plot indicates that the different SSL frameworks applied to the same architecture with the same dataset yield similar memorization pattern. Table 10: The memorization in Vi T occurs primarily in the deeper blocks and more in the fully-connected than attention layers. We present the results for Vi T Tiny pre-trained on CIFAR10 using MAE. Layer Mem AT T N = Layer Mem AT T N Res Block F C N 1, Layer Mem F C N = Layer Mem F C N Res Block AT T N , Block Mem N = Res Block F C N Res Block F C N 1. Vi T Block Attention Layer Fully-Connected Layer Number Layer Mem Layer Mem Res Block Layer Mem Layer Mem Res Block Block Mem 1 0.006 - 0.007 0.020 - 0.022 - 2 0.028 0.006 0.028 0.039 0.011 0.040 0.018 3 0.046 0.006 0.047 0.060 0.013 0.061 0.021 4 0.067 0.006 0.067 0.083 0.017 0.085 0.024 5 0.092 0.007 0.091 0.105 0.014 0.106 0.021 6 0.114 0.008 0.114 0.129 0.015 0.131 0.025 7 0.140 0.009 0.139 0.155 0.016 0.156 0.025 8 0.164 0.008 0.164 0.182 0.018 0.182 0.026 9 0.191 0.009 0.190 0.210 0.020 0.211 0.029 10 0.220 0.009 0.220 0.240 0.020 0.241 0.030 11 0.249 0.008 0.249 0.271 0.022 0.271 0.030 12 0.280 0.009 0.280 0.303 0.023 0.304 0.033 (a) Sim CLR vs. DINO (b) DINO vs. MAE Figure 12: Different SSL frameworks. Additional Verification of Unit Mem. We present the additional verification of the Unit Mem metric in Table 14. Therein, we perform two additional experiments to the verification presented in Section 5.2. First, we finetune the most memorizing unit and the inactive unit with 300 (instead of 1) data points from the test set (a). We observe that the data points that experienced the highest memorization for the selected unit remains the highest memorized of the 300 data points. Additionally, it experiences the highest memorization in the unit that used to be inactive. Second, we fine-tune the most memorizing unit and the inactive unit with the most memorized data point, but with a batch-size of 300 were we duplicate the data point 300 times (b). We observe that Table 11: Most vs Least Memorized Data Points. We train a Res Net9 using Sim CLR on CIFAR10 follwoing the setup by [47]. We then take the 50 most and 50 least memorized data points according to SSLMem and calculate the Unit Mem over for the two sets of points. In the table, we report the average per-layer Unit Mem of the two sets independently. We also perform a statistical t-test to find whether the Unit Mem scores differ among most and least memorized data points. With p << 0.05, we are able to reject the null-hypothesis and find that the memorization according to Unit Mem differs significantly between the most and least memorized data points. Layer mean Unit Mem mean Unit Mem t-test Name most memorized (10% units) least memorized (10% units) p-value conv1 0.507 0.235 65.89/0.00 conv2-0 0.501 0.231 66.25/0.00 conv2-1 0.503 0.233 65.94/0.00 conv2-2 0.512 0.240 65.12/0.00 conv3 0.509 0.242 64.13/0.00 conv4-0 0.514 0.246 63.67/0.00 conv4-1 0.515 0.245 64.09/0.00 conv4-2 0.522 0.248 64.18/0.00 Table 12: Highly memorized data points align with most memorizing units. We select 10% of the most memorizing units according to Unit Mem in the last layer (conv-4-2) of the Res Net9 encoder pre-trained on CIFAR10. The 1st row represents the number of times a given data point was responsible for µmax, the 2nd row counts for how many daat points this applies. The last column shows that the highest memorized sample (SSLMem of 0.891) is responsible for the µmax in the largest number of units (5). Metric Used Average SSLMem Score Frequency 0.694 0.813 0.833 0.857 0.891 # of times Responsible for µmax 1 2 3 4 5 # of Samples 10 2 1 1 1 the effect of the fine-tuning on this point s memorization is far more expressed than when fine-tuning with 300 different data points. Additional Verification of Unit Mem. In Table 13, we prune, i.e., zero out neurons according to their level of memorization. Our results indicate that by pruning the most memorizing neurons, we cause the highest drop in downstream performance. C.2 Unit Mem Measures Memorization of Individual Data Points To highlight that Unit Mem reports memorization of individual data points rather than the a unit s ability to recognize class-wide concepts, we designed an additional experiment. For the experiment, we rely on the class concept of "wheel" as an example. In the STL10 dataset, three classes have a concept wheel, namely Truck, Plane, Car. If Unit Mem was to report simply a unit s sensitivity to concepts of different classes (rather than individual data points), we would see a drop in Unit Mem as we increase the percentage of data points with the concept wheel in the batch used to compute the metric. This is because then all data points should equally activate the unit, resulting in low memorization according to Equation (8). We perform the experiment in Table 34 with 1000 data points chosen from different classes, namely 1) all classes (here 30% of the data points have wheels), 2) the classes Truck, Plane, Car (close to 100% of the samples now have the concept of wheels), and 3) purely the class car (close to 100% wheels). In 2) and 3), close to 100% of the samples now have the concept of wheels. Thus, if the units were responsible for the concept wheel, they would have a very high activation over all samples and the reported Unit Mem should be very low. However, in our results, we see that we do have units with very high Unit Mem. These can, in turn not be the units for the class-concept wheel, but must be units that focus on individual characteristics of the individual training images. This means that there must be unique features from the individual images that are still memorized that go beyond the concepts that are the same within a class. Table 13: Verification of the Unit Mem metric for memorization in individual units. The SSL model based on Sim Siam with Res Net18 architecture and trained on CIFAR10 is fine-tuned on a single data point. We select two units with the highest and lowest Unit Mem scores. The data point used for fine-tuning achieves µmax in both units. The Unit Mem score increases only for the two selected units while it remains unchanged for the remaining units. Targeted Number of Fine-Tuning Epochs Unit 0 10 50 200 500 1000 Highest Unit Mem 0.754 0.761 0.792 0.814 0.824 0.826 Lowest Unit Mem 0 0 0 0.0008 0.0021 0.0109 Table 14: The µmax and µmin after fine-tuning for different number of epochs. (a) The µmax and µmin after fine-tuning for different numbers of epochs. This is fine-tuned with 300 data samples from the test dataset. The samples were not seen during the initial training of the encoder, thus only a single filter is affected by them. trained nodes original 10 epoch 50 epoch 200 epoch 500 epoch 1000 epoch Most Mem filter 77th 0.754 0.766 0.809 0.819 0.826 0.828 Least Mem filter 459th 0 0 0.011 0.038 0.046 0.051 (b) The µmax and µmin after fine-tuning for different number of epochs This is fine-tuned with only highest µmax samples while 300 duplication from training datasets. trained nodes original 10 epoch 50 epoch 200 epoch 500 epoch 1000 epoch Most Mem filter 77th 0.754 0.798 0.846 0.857 0.861 0.862 Least Mem filter 459th 0 0.039 0.065 0.079 0.081 0.081 C.3 Additional Insights into Layer Mem Layer Mem is not sensitive to the size and composition of the batch. In our ablation study in , we show that Layer Mem is not sensitive to the size and composition of batch it is computed on. The results can be found in Table 19, where report the Layer Mem measured for different number of candidate data points. We pre-trained a Res Net9 using Sim CLR on CIFAR10 and determined Layer Mem on batches of different sizes. For each batch size, we use 3 independent seeds (i.e., different batch compositions) and report the average Layer Mem score and its standard deviation. The results show that the reported Layer Mem score is, indeed, similar across all setups. This indicates Layer Mem s insensitivity to the choice of the batch used to compute it. Full Results with Memorization Scores over all Layers. We present the Layer Mem score for Res Net18 in Table 15, Res Net34 in Table 16, and Res Net50 in Table 17, all trained on CIFAR10 and using the Sim CLR framework. We show the further breakdown of the memorization within the layers in Table 18. We observe that the batch normalization layers (denoted as BN) together with the Max Pool layers have a negligible impact on memorization and most of the memorization in each layer is due to the convolutional operations. This is due to the much larger number of parameters in the convolutional filters than in the batch normalization layers and no additional parameters in the Max Pool layers, as shown in Table 8. However, the memorization reported per convolutional layer is not correlated with the number of parameters of the layer. For instance, our Layer Mem reports the highest memorization for the 6-th layer, while layers 7 and 8 have each twice as many parameters, see Table 1. Table 15: Full results Res Net18. We depict our Layer Mem of the final trained model (at the end of training with CIFAR10, Res Net18 with Sim CLR). Layer Layer Mem conv1 0.074 0.010 max pool 0.092 0.007 conv2-1 0.101 0.012 conv2-2 0.110 0.006 conv2-3 0.123 0.013 conv2-4 0.134 0.010 conv3-1 0.146 0.008 conv3-2 0.155 0.013 conv3-3 0.166 0.011 conv3-4 0.183 0.007 conv4-1 0.193 0.006 conv4-2 0.206 0.009 conv4-3 0.220 0.011 conv4-4 0.239 0.010 conv5-1 0.246 0.014 conv5-2 0.257 0.007 conv5-3 0.272 0.011 conv5-4 0.295 0.012 averge-pool 0.266 0.009 fully-connected 0.224 0.010 softmax 0.207 0.007 Table 16: Full results Res Net34. We depict our Layer Mem of the final trained model (at the end of training with CIFAR10, Res Net34 with Sim CLR). Layer Layer Mem conv1 0.037 0.008 max pool 0.069 0.013 conv2-1 0.078 0.008 conv2-2 0.083 0.007 conv2-3 0.091 0.012 conv2-4 0.096 0.009 conv2-5 0.107 0.016 conv2-6 0.115 0.010 conv3-1 0.124 0.011 conv3-2 0.128 0.013 conv3-3 0.131 0.007 conv3-4 0.138 0.008 conv3-5 0.143 0.013 conv3-6 0.149 0.015 conv3-7 0.157 0.013 conv3-8 0.166 0.009 conv4-1 0.172 0.006 conv4-2 0.178 0.010 conv4-3 0.181 0.012 conv4-4 0.186 0.008 conv4-5 0.194 0.013 conv4-6 0.201 0.007 conv4-7 0.205 0.009 conv4-8 0.211 0.011 conv4-9 0.218 0.006 conv4-10 0.227 0.012 conv4-11 0.235 0.010 conv4-12 0.246 0.007 conv5-1 0.257 0.011 conv5-2 0.264 0.014 conv5-3 0.273 0.008 conv5-4 0.285 0.012 conv5-5 0.299 0.011 conv5-6 0.313 0.015 averge-pool 0.297 0.009 fully-connected 0.241 0.013 softmax 0.233 0.006 Ablation on Layer Mem s sensitivity. We perform an additional ablation to show that Layer Mem is not sensitive to the number of samples in the batch used to compute it or the composition of the batch, i.e., which samples are chosen. Our results in Table 19 highlight that over different batches with 100, 500, 1000, and 5000 samples, the observed Layer Mem scores are alike. This indicates Layer Mem s insensitivity to the choice of the batch used to compute it. Table 17: Full results Res Net50. We depict our Layer Mem of the final trained model (at the end of training with CIFAR10, Res Net50 with Sim CLR). Layer Layer Mem conv1 0.046 0.006 max pool 0.066 0.012 conv2-1 0.071 0.008 conv2-2 0.068 0.013 conv2-3 0.073 0.012 conv2-4 0.079 0.015 conv2-5 0.082 0.014 conv2-6 0.083 0.010 conv2-7 0.088 0.007 conv2-8 0.094 0.011 conv2-9 0.103 0.014 conv3-1 0.109 0.010 conv3-2 0.112 0.012 conv3-3 0.118 0.009 conv3-4 0.123 0.007 conv3-5 0.127 0.010 conv3-6 0.133 0.011 conv3-7 0.136 0.013 conv3-8 0.140 0.008 conv3-9 0.144 0.005 conv3-10 0.149 0.008 conv3-11 0.156 0.011 conv3-12 0.165 0.007 conv4-1 0.168 0.012 conv4-2 0.175 0.010 conv4-3 0.181 0.006 conv4-4 0.187 0.009 conv4-5 0.192 0.008 conv4-6 0.198 0.014 conv4-7 0.204 0.010 conv4-8 0.211 0.008 conv4-9 0.217 0.011 conv4-10 0.225 0.005 conv4-11 0.231 0.015 conv4-12 0.235 0.011 conv4-13 0.241 0.012 conv4-14 0.248 0.009 conv4-15 0.253 0.011 conv4-16 0.262 0.016 conv4-17 0.268 0.012 conv4-18 0.279 0.011 conv5-1 0.292 0.008 conv5-2 0.293 0.005 conv5-3 0.298 0.012 conv5-4 0.308 0.010 conv5-5 0.316 0.014 conv5-6 0.315 0.012 conv5-7 0.326 0.007 conv5-8 0.327 0.011 conv5-9 0.335 0.013 averge-pool 0.328 0.007 fully-connected 0.266 0.014 softmax 0.245 0.010 C.4 Memorization in SL vs. SSL We conducted an additional experiment where we trained a Res Net9 on CIFAR100 with SSL (Sim CLR) and SL (cross-entropy loss). For the SL model, we remove the classification layer to turn it into an encoder. Then, we report linear probing accuracies on multiple downstream tasks in Table 20. Our results highlight that the SL pretrained encoders exhibit a significantly higher downstream accuracy on their pretraining dataset than the SSL encoder. We assume that this is because of the class memorization. In contrast, the SL pretrained encoders perform significantly worse on other datasets than the SSL pretrained encoders since they might overfit the representations to their classes rather than provide more general (instance-based) representations as the SSL encoders. Additionally, we note that prior work has shown that the MAE encoder provides the highest performance when a few last layers are fine-tuned. The results in the original MAE paper in Figure 9 [30] indicate that fine-tuning a few last layers/blocks (e.g., 4 or 6 blocks out of 24 in Vi T-Large) can achieve accuracy close to full fine-tuning (when all 24 blocks are fine-tuned). This is in line with our observation that the difference between Unit Mem and Class Mem is the highest in the few last layers/blocks. Thus, fine-tuning only these last layers/blocks suffices for good downstream performance. Table 18: Layer-based Memorization Scores. We present the layer-wise memorization of an SSL encoder pretrained on CIFAR10 using Res Net9 with Sim CLR. The 1st column represents the IDs of convolutional layers and the 2nd column shows the name of the layers. Residual N denotes that the residual connection comes from the previous N-th convolutional layer. We report Layer Mem across the 100 randomly chosen training data points, their Layer Mem (denoted as LM), followed by Layer Mem for only the Top 50 memorized data points, their Layer Mem (denoted as Top50), and Layer Mem for only the Least 50 memorized data points. The projection head layer (denoted as head) is used only for training. ID Name Layer Mem LM Layer Mem Top50 Top50 Layer Mem Least50 1 Conv1 0.091 - 0.144 - 0.003 - BN1 0.091 0.000 0.144 0 0.004 - Max Pool 0.097 0.006 0.158 0.014 0.004 2 Conv2-0 0.123 0.026 0.225 0.067 0.012 - BN2-0 0.124 0.001 0.225 0 0.012 - Max Pool 0.128 0.004 0.236 0.011 0.013 3 Conv2-1 0.154 0.026 0.308 0.072 0.022 4 Conv2-2 0.183 0.029 0.402 0.094 0.031 - Residual2 0.185 0.002 0.403 0.01 0.041 5 Conv3 0.212 0.027 0.479 0.076 0.051 - BN3 0.211 -0.001 0.480 0.001 0.051 - Max Pool 0.215 0.004 0.486 0.006 0.050 6 Conv4-0 0.246 0.031 0.599 0.113 0.061 - BN4-0 0.244 -0.002 0.600 0.001 0.060 - Max Pool 0.247 0.003 0.603 0.003 0.061 7 Conv4-1 0.276 0.029 0.697 0.094 0.071 8 Conv4-2 0.308 0.032 0.817 0.120 0.073 - Residual6 0.311 0.003 0.817 0 0.086 - head 0.319 0.008 0.819 0.002 0.097 - Max Pool 0.318 -0.001 0.819 0 0.096 - FC 0.192 -0.126 0.409 -0.410 0.071 Table 19: Layer Mem is not sensitive to the number of samples used for its calculation. We pre-train a Res Net9 using Sim CLR on CIFAR10 and determined Layer Mem on batches of different sizes. For each batch size, we use three independent seeds (i.e., different batch compositions) and report the average Layer Mem score and its standard deviation. The results show that the reported Layer Mem score is, indeed, similar across all setups. This indicates Layer Mem s insensitivity to the choice of the batch used to compute it. Layer 100 samples 500 samples 1000 samples 5000 samples 1 0.092 8e-4 0.093 7e-4 0.089 7e-4 0.091 8e-4 2 0.122 9e-4 0.124 1e-3 0.122 8e-4 0.121 7e-4 3 0.150 1e-3 0.154 9e-4 0.151 8e-4 0.152 6e-4 4 0.181 1e-3 0.182 8e-4 0.180 8e-4 0.181 8e-4 Res2 0.184 9e-4 0.185 8e-4 0.183 9e-4 0.184 6e-4 5 0.213 8e-4 0.213 8e-4 0.212 7e-4 0.212 8e-4 6 0.246 1e-3 0.249 7e-4 0.247 8e-4 0.245 9e-4 7 0.277 7e-4 0.281 9e-4 0.277 8e-4 0.276 4e-4 8 0.309 9e-4 0.314 8e-4 0.310 7e-4 0.307 7e-4 Res6 0.310 1e-3 0.316 1e-3 0.313 8e-4 0.309 9e-4 C.5 Visualization for Variability and Consistency of Memorization cross Different Layers. We present the top 10 most memorized samples of each layer for the Res Net9 vision encoder trained with the CIFAR10 dataset in Figure 13. The results show that the overlap within the top 10 most memorized samples between adjacent layers is usually high but decreases the further the layers are separated. This aligns with the results of overlap rate and Kendall s Tau test reported in Table 3. C.6 Layer-based Memorization Across Different SSL Frameworks and Datasets We present the full results for the Table 4, which show that the layer-based memorization is similar across encoders trained with different SSL frameworks. The results for the Res Net50 architecture trained with Sim CLR and DINO using the Image Net dataset are presented in Table 29, and the results for the Vi T-Base architecture trained with MAE and DINO using the Image Net dataset are shown in Table 31. Table 20: Comparing the impact of memorization on downstream generalization between SSL and SL. We train a Res Net9 on CIFAR100 with SSL (pretrained on CIFAR100 using Sim CLR and SL (cross-entropy loss, trained until convergence). For the SL model, we remove the classification layer to turn it into an encoder. Then, we report linear probing accuracies on multiple downstream tasks in Encoder CIFAR100 CIFAR10 STL10 SVHN SSL 65.4% 0.98% 57.6% 0.87% 48.7% 0.98% 59.2% 0.76% SL (trained until convergence on CIFAR100, last layer removed) 66.1% 1.12% 56.7% 0.83% 46.1% 1.04% 58.6% 0.82% Figure 13: The most memorized samples per layer according to Layer Mem. C.7 Verification of Layer-Based Memorization To analyze whether our Layer Mem metric and its variant indeed localize memorization correctly, we first replace different layers of an encoder and then compute linear probing accuracy on various downstream tasks. Since prior work shows that memorization in SSL is required for downstream generalization [47], we expect the highest performance drop when replacing the layers identified as most memorizing. We verify this hypothesis and train a Res Net9 encoder f1 on the CIFAR10 dataset and compute the Layer Mem and Layer Mem scores per layer. Then, we select the three most memorized, random, and least memorized layers and replace them with the corresponding layers from a Res Net9 trained on STL10 (f2). Our results in Table 21 show that the highest linear probing accuracy drop on the CIFAR10 test set for f1 is caused by replacing the three most memorized layers from f1 according to the Layer Mem score. The second biggest drop is observed when replacing according to Layer Mem, highlighting that indeed our Layer Mem metric and its variant identify the most crucial layers in SSL encoders for memorization. Surprisingly, the replacement of the layers in f1 with the corresponding layers from f2 causes the biggest simultaneous increase in the downstream accuracy for the STL10 dataset. We observe the same trends when f2 is trained on SVHN (Table 22b), for replacing single layers in Res Net9 (Appendix C.7), and replacing whole blocks in Res Net50 (Appendix C.7) instead of only individual layers as we present in Appendix C.7. The above analysis verifies that the Layer Mem score and its variant identify the most crucial layers in SSL encoders. They further strengthen the claims that memorization is required for generalization [23, 24, 47]. Replacing layers in Res Net9 for SVHN. In Table 22b, we show the effect of replacing the most and least vs random layers of a CIFAR10 trained Res Net9 on the downstream performance. We replace the layers with the corresponding ones from a Res Net9 encoder trained on SVHN. Table 21: Replacing the most/least memorized layers according to Layer Mem causes the most/least changes in downstream performance. We study the effect of replacing layers of the Res Net9 encoder trained on CIFAR10 with layers from another Res Net9 encoder trained on STL10 and report the linear probing accuracy on the CIFAR10 and STL10 test sets. Results for the impact of replacing any combination of 1, 2, and 3 layers on downstream accuracy are shown in Appendix C.12. Replacement Criteria Replaced Layer(s) CIFAR10 STL10 None (Baseline) None 69.08% 1.05% 17.81% 0.92% Most Memorized Layer Mem 4 6 8 36.59% 1.13% 32.33% 0.88% Most Memorized Layer Mem 6 7 8 39.07% 1.05% 29.82% 0.91% Random 4 5 7 43.22% 1.08% 25.89% 0.93% Least Memorized Layer Mem 2 3 4 49.95% 1.21% 24.71% 0.99% Least Memorized Layer Mem 2 3 5 59.14% 0.91% 23.10% 1.06% Table 22: Evaluating the effect of replacing layers of the Res Net9 encoder pre-trained on CIFAR10 with layers from Res Net9 pre-trained on STL10. We report the linear probing accuracy of Res Net9 with the replaced layers and tested on the CIFAR10, STL10 test sets. (a) CIFAR10 & STL10 Replaced Criterium Replaced Layer(s) CIFAR10 STL10 None (Baseline) None 69.08% 1.05% 17.81% 0.92% Most Memorized (delta) 4 6 8 36.59% 1.13% 32.33% 0.88% Most Memorized (absolute) 6 7 8 39.07% 1.05% 29.82% 0.91% Random 4 5 7 43.22% 1.08% 25.89% 0.93% Least Memorized (delta) 2 3 5 59.14% 0.91% 23.10% 1.06% Least Memorized (absolute) 2 3 4 49.95% 1.21% 24.71% 0.99% (b) CIFAR10 & SVHN Replaced Criterium Replaced Layer(s) CIFAR10 SVHN None (Baseline) None 69.08% 1.05% 19.33% 0.65% Most Memorized (delta) 4 6 8 33.07% 1.51% 34.05% 1.01% Most Memorized (absolute) 6 7 8 34.97% 0.84% 31.87% 1.21% Random 4 5 7 39.28% 0.74% 26.04% 0.82% Least Memorized (delta) 2 3 5 52.81% 1.03% 21.05% 0.89% Least Memorized (absolute) 2 3 4 45.39% 1.10% 24.66% 0.57% Replacing blocks in Res Net50 trained on CIFAR10 with Sim CLR. We present the results for replacing blocks in Res Net50 trained on CIFAR10 using Sim CLR in Appendix C.7. Statistics of Batch-Norm layer for different datasets. Batch-norm layers between different datasets might have different statistics. This could impact the downstream performance. To investigate the changes, we measured the cosine similarity between the weights and biases of the batch-norm layers for two encoders (trained on CIFAR10 and STL10, respectively). The results in Table 25 show a high per-layer cosine similarity (average over all layers=0.823). This suggests that the statistics are similar, hence, no adjustment is required. We hypothesize that the similarity stems from the fact that the data distributions are similar and that we normalize both input datasets according to the Image Net normalization parameters. C.8 Layer Mem with Different Distance Metrics In addition to the ℓ2 distance, we also used 3 other distance metrics (ℓ1, cosine similarity, and angular distance) to evaluate the stability of Layer Mem. Our results in Table 26 highlight that 1) the memorization scores are very similar, independent of the choice of the distance metric, and 2) the most memorizing layers according to and Layer Mem are the same over all metrics. This suggests that our findings are independent of the choice of distance metric. Table 23: Evaluating the effect of replacing layers of the Res Net9 encoder pre-trained on CIFAR10 with layers from Res Net9 pre-trained on STL10. We report the linear probing accuracy of Res Net9 with the replaced layers and tested on the CIFAR10, STL10 test sets. (a) CIFAR10 & STL10 Replaced Criterium Replaced Layer(s) CIFAR10 STL10 None (Baseline) None 69.08% 1.05% 17.81% 0.92% Most Memorized (delta) 6 59.84% 1.20% 21.98% 0.41% Most Memorized (absolute) 8 60.02% 0.94% 21.67% 0.72% Random 5 62.98% 0.57% 20.44% 0.85% Least Memorized (delta) 2 65.52% 0.74% 18.94% 0.63% Least Memorized (absolute) 3 64.21% 1.08% 18.89% 0.81% (b) CIFAR10 & SVHN Replaced Criterium Replaced Layer(s) CIFAR10 SVHN None 69.08% 1.05% 19.33% 0.65% Most Memorized (delta) 6 59.22% 0.97% 22.47% 0.57% Most Memorized (absolute) 8 59.69% 1.04% 21.60% 0.92% Random 5 61.07% 1.12% 21.09% 0.69% Least Memorized (delta) 2 62.93% 1.08% 20.44% 0.71% Least Memorized (absolute) 3 62.35% 0.81% 20.18% 0.98% Table 24: Evaluating the effect of replacing blocks of Res Net50 pre-trained on CIFAR10 with blocks from Res Net50 pre-trained on STL10 and SVHN. The accuracy in the table is the linear probing accuracy of Res Net50 on CIFAR10. We replace block 3 of conv layers, which was selected according to the biggest Layer Mem between two layers (not the absolute value of the Layer Mem score of the layers). (a) CIFAR10 & STL10 Replaced Criterium Replaced Layer(s) CIFAR10 STL10 None / 77.12% 1.42% 18.22% 0.88% Most Memorized C4_B6 C3_B4 C2_B3 43.66% 1.20% 25.78% 0.95% Random C3_B2 C4_B4 C5_B2 51.09% 1.01% 22.55% 1.17% Least Memorized C2_B1 C2_B2 C3_B3 57.41% 0.74% 20.10% 1.11% (b) CIFAR10 & SVHN Replaced Criterium Replaced Layer(s) CIFAR10 SVHN None / 77.12% 1.42% 28.44% 1.23% Most Memorized C4_B6 C3_B4 C2_B3 35.21% 0.94% 38.11% 1.08% Random C3_B2 C4_B4 C5_B2 44.19% 0.97% 32.44% 1.25% Least Memorized C2_B1 C2_B2 C3_B3 49.06% 1.31% 29.98% 0.85% C.9 Layer Mem with Different Augmentation Strength The results, reported in Table 27 highlight that stronger training augmentations reduce Layer Mem. C.10 Layer Mem with Different Initialization of Trainable parameters We performed an additional experiment where we trained encoders f and g independently with a different random seed (yielding f and g ) to study how random initialization of trainable parameters can affect the memorization of final vision encoder.The results are reported in Table 28. We compared the overlap in most memorized samples between encoder f (from the paper) and f . The results (Table 4, attached PDF) show that overlap is overall high (min. 69% in layer 2) and increases in the later layers (max. 90%, final layer). Table 25: Cosine similarities between batch-norm layer outputs for Res Net9 trained on CIFAR10 and STL10. We normalize the training data according to the Image Net parameters and train the encoders using Sim CLR. We calculate the cosine similarity over the weights (γ) and the bias (β) of the respective encoders trained batch-norm layers. Layer 1 2 3 4 5 6 7 8 Cosine Similarity 0.797 0.844 0.823 0.811 0.779 0.805 0.847 0.881 Table 26: Layer Mem (LM) and -LM under different distance metrics. We report for ℓ1, ℓ2 (see original submission), cosine similarity (Cos. Sim), and angular distance (Ang. Dist). The results highlight that our memorization measure is independent of the underlying metric. (Res Net9, CIFAR10, Sim CLR). ℓ1 ℓ2 Cos. Sim. Ang. Dist. Layer LM LM LM LM LM LM LM LM 1 0.099 - 0.091 - 0.104 - 0.096 - 2 0.128 0.029 0.123 0.032 0.134 0.030 0.128 0.032 3 0.159 0.031 0.154 0.031 0.163 0.029 0.160 0.032 4 0.187 0.028 0.183 0.029 0.190 0.027 0.191 0.031 Res2 0.192 0.005 0.185 0.002 0.193 0.003 0.193 0.002 5 0.221 0.029 0.212 0.027 0.220 0.027 0.222 0.029 6 0.256 0.035 0.246 0.034 0.256 0.036 0.259 0.037 7 0.289 0.033 0.276 0.030 0.288 0.032 0.293 0.034 8 0.325 0.036 0.308 0.032 0.321 0.033 0.328 0.035 Res6 0.329 0.004 0.311 0.003 0.323 0.002 0.329 0.001 C.11 Impact of Layer Replacement on Layer Memorization According to the definition of SSLMem Equation (2), we let the representations of a given input data point x pass through the same (replaced) layer in both f and g. We show the Layer Mem and Layer Mem scores after replacing a single layer in the Res Net9 encoder pre-trained using Sim CLR on the CIFAR10 dataset in Table 23 and Table 30. The Layer Mem score of the replaced layer always drops as expected since this layer does not memorize any original training data points. The decrease in Layer Mem between the initial and replaced layers is smaller in the earlier layers (e.g., 1st layer) as compared to the later layers (e.g., 6th layer). This might be because, in general, the earlier layers from different models might be more similar as they are responsible for extracting general features instead of specific ones for a given dataset. The most important take-away from these experiments is that the Layer Mem is not affected significantly and its values show the same trends after the layer replacement. C.12 Layer Replacement for Single, Two, and Three Layers at a Time We perform the experiment with the replacement of 1 layer in Table 30, 2 layers Table 32, and 3 layers Table 33. The following results confirm our results from Table 21 in the main paper. When only a single layer is replaced, then the 6th (not the last layer) is the most important one. This layer had the highest Layer Mem score. Note that the replacement of the 6th layer causes the highest drop in accuracy on the original CIFAR10 dataset and the highest gain in accuracy on STL10. Next, when two layers are replaced, then layers 6th and 8th play the most important roles, where Table 27: Layer Mem (LM) and -LM under different augmentation sets used during training. We use the augmentations defined in Appendix B during training and metric calculation. The results show that stronger augmentations reduce memorization. (Res Net9, CIFAR10, Sim CLR). weak normal strong Layer LM LM LM LM LM LM 1 0.092 - 0.091 - 0.089 - 2 0.123 0.031 0.123 0.032 0.120 0.031 3 0.154 0.031 0.154 0.031 0.150 0.030 4 0.184 0.030 0.183 0.029 0.178 0.028 Res2 0.187 0.003 0.185 0.002 0.181 0.003 5 0.215 0.028 0.212 0.027 0.208 0.027 6 0.249 0.034 0.246 0.034 0.241 0.033 7 0.280 0.031 0.276 0.030 0.269 0.028 8 0.313 0.033 0.308 0.032 0.300 0.031 Res6 0.315 0.002 0.311 0.003 0.302 0.002 Table 28: Overlap in 100 most memorized samples according to Layer Mem between 2 different encoders. We train encoders with different seeds and report the per-layer overlap in their most memorized samples. We observe an overall high overlap, especially in the last layer. Layer 1 2 3 4 5 6 7 8 Overlap % 73 69 75 89 85 88 86 90 their replacement with layers from the encoder trained on STL10 causes the highest drop on CIFAR10 and the highest performance increase on STL10. This is contrary to the common intuition, which would suggest the replacement of the last two layers instead. D Additional Setup D.1 Class Selectivity We denote the class selectivity metric as Class Selectivity. It was proposed by [39] to quantify a unit s discriminability between different classes and described more in the main part of the paper in Section 5. We derive the basic metric in more detail here. To compute the Class Selectivity metric per unit u, first the class-conditional mean activity is calculated for the test dataset D. We denote each test data point as xi D. We assume M classes CM j=1, each with its corresponding test data points xc Cj, where c = 1, 2, ..., |Cj|. We define the mean activation µ of unit u for class Cj as µu(Cj) = 1 |Cj| xc Cj activationu( xc), (9) where the activation for convolutional feature maps is averaged across all elements of the feature map. Further, for the unit u, we compute the maximum mean activation µmax,u across all classes C, where M = |C|, as µmax,u = max({ µu( xi)}M i=1). (10) Let p be the index location of the maximum mean activation µu( xp), i.e., the argmax. Then, we calculate the corresponding mean activity µ max,u across all the remaining M 1 classes as µ max,u = 1 M 1 j=1,j =p µu(Cj). (11) Finally, the class selectivity is then calculated as follows Class Selectivity(u) = µmax,u µ max,u µmax,u + µ max,u , (12) where µmax,u represents the highest class-conditional mean activity and µ max,u denotes the mean activity across all other classes (for unit u and computed on the test dataset D). D.2 Class Memorization We use a similar definition as Class Selectivity for Class Mem, which measures how much a given unit is responsible for the memorization of a class. While Class Selectivity is calculated on the test set, we compute Class Mem on the training dataset. To compute the Class Mem metric per unit u, first, the class-conditional mean activity is calculated for the training dataset D . We denote each train data point as xi D . We assume M classes CM j=1, each with its corresponding train data points xc Cj, where c = 1, 2, ..., |Cj|. We define the mean activation µ of unit u for class Cj as µu(Cj) = 1 |Cj| xc Cj activationu(xc), (13) where the activation for convolutional feature maps is averaged across all elements of the feature map. Further, for the unit u, we compute the maximum mean activation µmax,u across all classes C, where M = |C|, as µmax,u = max({ µu(xi)}M i=1). (14) Let p be the index location of the maximum mean activation µu( xp), i.e., the argmax. Then, we calculate the corresponding mean activity µ max,u across all the remaining M 1 classes as µ max,u = 1 M 1 j=1,j =p µu(Cj). (15) Finally, the class Memorization is then calculated as follows Class Mem(u) = µmax,u µ max,u µmax,u + µ max,u , (16) where µmax,u represents the highest class-conditional mean activity and µ max,u denotes the mean activity across all other classes (for unit u and computed on the train dataset D ). E Extended Related Work Localizing Memorization on the Level of Individual Units. In Section 5, we considered memorization from the perspective of individual units and identified that pruning the least/most memorized units according to Unit Mem preserves the least/most performance (as shown in Table Table 6). The work by Maini et al. [35] characterized individual examples as mislabeled based on the low number of channels or filters that need to be zeroed out to flip the prediction. They observe that significantly more neurons need to be zeroed out to flip clean examples compared to mislabeled ones. A similar experiment in the SSL domain could potentially reveal a similar trend, where noisy examples are harder to learn and primarily influence a small number of units. However, SSL encoders do not have discrete output changes from zeroing out individual units. One could pre-train the encoder and add linear probing, but this would require labels for the SSL training set, making it inapplicable. Even with labeled data and fine-tuning, identifying noisy SSL examples based on the SSLMem score may not match mislabeled examples in SL. The lack of a discrete oracle and the potential mismatch between noisy SSL and mislabeled SL examples makes it difficult to identify individual units responsible for predictions of selected examples using prior methods. F Impact & Limitations The fact that memorization can enable privacy attacks, such as data extraction [9, 10, 12], has been established in prior work. Yet, this paper advances the field of machine learning towards a novel fundamental understanding on where in SSL encoders memorization happens, and how memorization differs between standard SL models and SSL encoders. Our insights hold the potential to yield societal benefits in the form of the design of novel methods to reduce memorization, improve fine-tuning, and yield better model pruning algorithms. G Additional Results Table 29: All-layer memorization. We train the Res Net50 encoder using Sim CLR and DINO SSL frameworks on the Image Net dataset. We report the full results with the Layer Mem and Layer Mem scores for each layer. Res Net50 Layer Sim CLR DINO Layer Mem Layer Mem Layer Mem Layer Mem conv1 0.038 0.001 - 0.040 0.002 - max pool 0.039 0.002 0.001 0.040 0.002 0.000 conv2-1 0.041 0.002 0.002 0.043 0.001 0.003 conv2-2 0.044 0.002 0.003 0.045 0.001 0.002 conv2-3 0.048 0.001 0.004 0.048 0.002 0.003 conv2-4 0.052 0.002 0.004 0.052 0.001 0.004 conv2-5 0.055 0.001 0.003 0.056 0.001 0.004 conv2-6 0.059 0.002 0.004 0.060 0.001 0.004 conv2-7 0.063 0.001 0.004 0.065 0.001 0.005 conv2-8 0.068 0.001 0.005 0.069 0.002 0.004 conv2-9 0.072 0.002 0.004 0.073 0.001 0.004 conv3-1 0.077 0.002 0.005 0.078 0.002 0.005 conv3-2 0.081 0.003 0.004 0.083 0.001 0.005 conv3-3 0.086 0.002 0.005 0.088 0.002 0.005 conv3-4 0.092 0.001 0.006 0.094 0.001 0.006 conv3-5 0.097 0.001 0.005 0.099 0.002 0.005 conv3-6 0.103 0.002 0.006 0.104 0.001 0.005 conv3-7 0.108 0.002 0.005 0.110 0.001 0.006 conv3-8 0.112 0.002 0.004 0.115 0.002 0.005 conv3-9 0.117 0.001 0.005 0.120 0.001 0.005 conv3-10 0.123 0.002 0.006 0.126 0.002 0.006 conv3-11 0.128 0.001 0.005 0.131 0.003 0.005 conv3-12 0.134 0.002 0.006 0.136 0.002 0.005 conv4-1 0.139 0.002 0.005 0.142 0.002 0.006 conv4-2 0.145 0.002 0.006 0.148 0.003 0.006 conv4-3 0.150 0.003 0.005 0.153 0.002 0.005 conv4-4 0.156 0.003 0.006 0.159 0.003 0.006 conv4-5 0.161 0.002 0.005 0.164 0.003 0.005 conv4-6 0.166 0.003 0.005 0.169 0.004 0.005 conv4-7 0.172 0.004 0.006 0.175 0.002 0.006 conv4-8 0.178 0.003 0.006 0.181 0.003 0.006 conv4-9 0.183 0.002 0.005 0.186 0.002 0.005 conv4-10 0.189 0.003 0.006 0.192 0.003 0.006 conv4-11 0.194 0.004 0.005 0.198 0.004 0.006 conv4-12 0.200 0.003 0.006 0.203 0.005 0.005 conv4-13 0.207 0.006 0.007 0.210 0.003 0.007 conv4-14 0.212 0.002 0.005 0.216 0.004 0.006 conv4-15 0.218 0.003 0.006 0.221 0.003 0.005 conv4-16 0.224 0.004 0.006 0.228 0.005 0.007 conv4-17 0.229 0.003 0.005 0.234 0.003 0.006 conv4-18 0.235 0.005 0.006 0.240 0.003 0.006 conv5-1 0.242 0.003 0.007 0.247 0.004 0.007 conv5-2 0.249 0.003 0.007 0.254 0.005 0.007 conv5-3 0.257 0.002 0.008 0.261 0.004 0.007 conv5-4 0.265 0.002 0.008 0.269 0.005 0.008 conv5-5 0.272 0.004 0.007 0.276 0.004 0.007 conv5-6 0.279 0.003 0.007 0.283 0.003 0.007 conv5-7 0.287 0.003 0.008 0.292 0.006 0.009 conv5-8 0.295 0.003 0.008 0.300 0.004 0.008 conv5-9 0.304 0.005 0.009 0.309 0.004 0.009 Table 30: Replace a single layer. We follow the settings from the Table 21 (same encoder) and replace a single layer at a time. Replaced Layers CIFAR10 STL10 None 69.37% 1.07% 20.09% 0.64% 1 63.07% 0.93% 18.79% 0.62% 2 65.19% 0.87% 19.01% 0.77% 3 64.47% 1.15% 18.99% 0.81% 4 60.29% 0.74% 20.44% 0.66% 5 62.74% 0.82% 19.93% 0.59% 6 59.91% 1.09% 21.92% 0.67% 7 60.77% 0.75% 20.97% 0.52% 8 60.04% 0.90% 21.71% 0.58% max(=layer X) 65.19% 0.87%(2) 21.92% 0.67%(6) min(=layer X) 59.91% 1.09%(6) 18.79% 0.62%(2) Table 31: All-layer memorization. We train the Vi T-Base encoder using MAE and DINO SSL frameworks on the Image Net dataset. We report the full results with the Layer Mem and Layer Mem scores for each block. Vi T-Base MAE DINO Block Number Layer Mem Layer Mem Layer Mem Layer Mem 1 0.019 0.001 - 0.019 0.001 - 2 0.037 0.001 0.011 0.036 0.002 0.012 3 0.055 0.002 0.013 0.056 0.003 0.012 4 0.075 0.002 0.013 0.077 0.002 0.014 5 0.095 0.002 0.012 0.096 0.004 0.013 6 0.118 0.004 0.016 0.119 0.006 0.015 7 0.139 0.003 0.015 0.142 0.004 0.014 8 0.163 0.005 0.018 0.168 0.005 0.017 9 0.188 0.004 0.017 0.193 0.003 0.018 10 0.215 0.006 0.018 0.219 0.006 0.019 11 0.243 0.005 0.021 0.247 0.005 0.020 12 0.271 0.003 0.020 0.275 0.004 0.019 Table 32: Replace two layers. We follow the setting from the Table 21 (same encoder) and replace two layers at a time. Replaced Layers CIFAR10 STL10 None 69.37% 1.07% 18.44% 0.64% 1 2 53.44% 0.90% 20.31% 0.51% 1 3 52.51% 0.83% 20.80% 0.60% 1 4 50.77% 0.78% 22.16% 0.66% 1 5 50.98% 0.97% 22.02% 0.59% 1 6 46.09% 0.77% 21.46% 0.69% 1 7 49.59% 0.89% 24.87% 0.66% 1 8 45.96% 0.94% 21.66% 0.71% 2 3 55.44% 0.73% 19.31% 0.72% 2 4 53.61% 0.92% 24.18% 0.68% 2 5 53.34% 1.06% 20.39% 0.57% 2 6 48.59% 0.81% 23.32% 0.70% 2 7 51.07% 1.13% 21.97% 0.52% 2 8 50.15% 0.82% 22.57% 0.61% 3 4 52.99% 1.01% 21.09% 0.73% 3 5 52.67% 0.90% 21.00% 0.81% 3 6 48.22% 0.79% 23.48% 0.62% 3 7 50.81% 0.86% 22.09% 0.70% 3 8 49.07% 0.92% 23.19% 0.67% 4 5 50.49% 0.96% 22.41% 0.55% 4 6 44.88% 0.91% 23.61% 0.49% 4 7 46.38% 1.13% 24.04% 0.62% 4 8 45.09% 0.75% 24.11% 0.66% 5 6 46.02% 1.07% 24.29% 0.81% 5 7 49.21% 1.00% 22.99% 0.80% 5 8 45.71% 0.94% 24.33% 0.73% 6 7 44.76% 0.88% 24.90% 0.54% 6 8 44.13% 1.01% 25.08% 0.64% 7 8 44.98% 0.94% 24.91% 0.81% max (=layer X) 55.44% 0.73% (2 3) 25.08% 0.64% (6 8) min (=layer X) 44.13% 1.01% (6 8) 19.31% 0.72% (2 3) Table 33: Replace three layers. We follow the settings from the Table 21 (same encoder) and replace three layers at a time. Replaced Layers CIFAR10 STL10 None 69.37% 1.07% 18.44% 0.64% 1 2 3 49.77% 0.66% 21.95% 0.42% 1 2 4 48.33% 0.65% 23.28% 0.77% 1 2 5 49.51% 0.70% 22.18% 0.57% 1 2 6 45.31% 0.59% 26.41% 0.53% 1 2 7 48.99% 0.84% 22.54% 0.57% 1 2 8 46.29% 0.48% 25.74% 0.62% 1 3 4 47.66% 0.52% 24.37% 0.42% 1 3 5 49.04% 0.65% 22.20% 0.66% 1 3 6 45.19% 0.72% 26.57% 0.61% 1 3 7 48.19% 0.58% 23.76% 0.82% 1 3 8 46.20% 0.83% 25.67% 0.60% 1 4 5 47.35% 0.60% 24.99% 0.58% 1 4 6 43.89% 0.70% 29.57% 0.67% 1 4 7 44.53% 0.66% 27.55% 0.71% 1 4 8 43.94% 0.78% 29.26% 0.52% 1 5 6 44.23% 0.70% 27.89% 0.52% 1 5 7 48.03% 0.55% 23.30% 0.63% 1 5 8 44.71% 0.71% 26.99% 0.58% 1 6 7 43.30% 0.44% 29.81% 0.57% 1 6 8 41.72% 0.70% 30.71% 0.67% 1 7 8 44.59% 0.83% 28.06% 0.47% 2 3 4 48.89% 0.38% 22.76% 0.38% 2 3 5 50.48% 0.67% 21.71% 0.60% 2 3 6 46.98% 0.57% 25.68% 0.46% 2 3 7 49.81% 0.62% 22.31% 0.49% 2 3 8 48.07% 0.93% 23.74% 0.70% 2 4 5 48.55% 0.79% 23.90% 0.82% 2 4 6 44.99% 0.58% 27.88% 0.57% 2 4 7 47.78% 0.68% 24.87% 0.75% 2 4 8 45.41% 0.86% 26.98% 0.51% 2 5 6 45.91% 0.44% 26.47% 0.60% 2 5 7 48.37% 0.55% 22.90% 0.50% 2 5 8 47.18% 0.52% 25.57% 0.81% 2 6 7 45.62% 0.69% 26.78% 0.48% 2 6 8 42.99% 0.63% 29.41% 0.43% 2 7 8 46.89% 0.93% 25.77% 0.63% 3 4 5 47.90% 0.56% 24.74% 0.48% 3 4 6 43.32% 0.58% 28.73% 0.60% 3 4 7 45.80% 0.57% 26.59% 0.65% 3 4 8 44.49% 0.55% 28.36% 0.46% 3 5 6 45.49% 0.71% 26.89% 0.90% 3 5 7 48.14% 0.73% 23.66% 0.58% 3 5 8 46.61% 0.69% 25.90% 0.39% 3 6 7 44.01% 0.72% 28.20% 0.58% 3 6 8 42.00% 0.65% 29.93% 0.33% 3 7 8 45.21% 0.43% 27.26% 0.41% 4 5 6 41.84% 0.76% 30.20% 0.53% 4 5 7 44.20% 0.72% 27.64% 0.48% 4 5 8 42.31% 0.82% 29.81% 0.33% 4 6 7 40.02% 0.71% 32.55% 0.58% 4 6 8 37.66% 0.49% 31.94% 0.38% 4 7 8 40.96% 0.62% 30.78% 0.56% 5 6 7 42.77% 0.60% 29.66% 0.69% 5 6 8 40.55% 0.68% 31.02% 0.47% 5 7 8 43.79% 0.91% 28.54% 0.55% 6 7 8 38.95% 0.57% 31.59% 0.66% max (=layer X) 50.48% 0.67% (2 3 5) 31.94% 0.38% (4 6 8) min (=layer X) 37.66% 0.49% (4 6 8) 21.71% 0.60% (2 3 5) Table 34: Unit Mem distinguishes between individual examples within a class. We use 1000 samples for each experiment to compute the Unit Mem score. All: denotes all classes, TPC: stands for the 3 following classes Truck, Plance, and Car classes, while Car: is simply the car class. Layer All All All TPC TPC TPC Car Car Car Number min max avg min max avg min max avg Layer 1 0 0 0.845 0.014 0.366 0.011 0.007 1e-4 0.801 0.018 0.357 0.009 0.011 1e-4 0.829 0.015 0.360 0.010 Layer 2 0.006 9e-5 0.832 0.016 0.352 0.010 0 0 0.789 0.015 0.350 0.013 0.009 8e-5 0.810 0.011 0.351 0.013 Layer 3 0 0 0.841 0.017 0.363 0.008 0 0 0.800 0.014 0.355 0.010 0.010 9e-5 0.825 0.009 0.356 0.008 Layer 4 0 0 0.871 0.012 0.377 0.009 0.004 1e-4 0.833 0.019 0.373 0.012 0.015 2e-4 0.844 0.014 0.371 0.009 Layer 5 0.010 2e-4 0.859 0.016 0.381 0.008 0.016 3e-4 0.810 0.013 0.375 0.011 0.013 1e-4 0.837 0.012 0.380 0.010 Layer 6 0.020 4e-4 0.905 0.018 0.403 0.011 0.022 3e-4 0.868 0.019 0.381 0.008 0.030 5e-4 0.879 0.014 0.394 0.007 Layer 7 0.019 3e-4 0.894 0.013 0.398 0.009 0.021 3e-4 0.859 0.014 0.380 0.013 0.019 3e-4 0.861 0.017 0.387 0.011 Layer 8 0.017 2e-4 0.905 0.013 0.409 0.010 0.25 4e-4 0.863 0.017 0.385 0.010 0.024 4e-4 0.870 0.013 0.397 0.012 Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: We introduce our main contributions and key findings from line 5 to line 19 in the abstract and line 73 to line 80 in Section 1 Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: We discuss our Limitations in Appendix F Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speechto-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: All related results are clearly stated and referenced in either the main paper or the appendix. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and cross-referenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: The detailed experimental setup is introduced in Appendix B and related source code is uploaded to open-review. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: All related source code is uploaded to open-review and experiments are conducted on open source datasets. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips. cc/public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https://nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: The detailed experimental setup is introduced in Appendix B. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: For the average results of all multiple experiments in the paper, we report the standard deviation in the tables and draw the error bar used to represent the standard deviation in the figures. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: The hardware usage is introduced in Appendix B Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/ Ethics Guidelines? Answer: [Yes] Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: In this work, we introduce two metrics for locating memorization in SSL vision encoders, as well as some key findings that result from experimenting with our metrics. No potential direct social impact is expected. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: All datasets used to train models for this paper are safe public datasets, including Image Net ILSVRC-2012 [42], CIFAR10 [32], CIFAR100 [32], SVHN [40], and STL10 [18] (This is also introduced in Appendix B). Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: All models and datasets usage is detailed introduced in Appendix B. We totally abbey the terms of use of these public sources. All other code works are done by ourselves with no potential risks of licenses. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/ datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [Yes] Justification: We provide all our code for metrics mentioned in our paper to open review and write a detailed Readme file for the usage of our code. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.