# implicit_variance_regularization_in_noncontrastive_ssl__0715e0bc.pdf Implicit variance regularization in non-contrastive SSL Manu Srinath Halvagal1,2, Axel Laborieux1, Friedemann Zenke1,2 {firstname.lastname}@fmi.ch 1 Friedrich Miescher Institute for Biomedical Research, Basel, Switzerland 2 Faculty of Science, University of Basel, Basel, Switzerland These authors contributed equally. Non-contrastive self-supervised learning (SSL) methods like BYOL and Sim Siam rely on asymmetric predictor networks to avoid representational collapse without negative samples. Yet, how predictor networks facilitate stable learning is not fully understood. While previous theoretical analyses assumed Euclidean losses, most practical implementations rely on cosine similarity. To gain further theoretical insight into non-contrastive SSL, we analytically study learning dynamics in conjunction with Euclidean and cosine similarity in the eigenspace of closed-form linear predictor networks. We show that both avoid collapse through implicit variance regularization albeit through different dynamical mechanisms. Moreover, we find that the eigenvalues act as effective learning rate multipliers and propose a family of isotropic loss functions (Iso Loss) that equalize convergence rates across eigenmodes. Empirically, Iso Loss speeds up the initial learning dynamics and increases robustness, thereby allowing us to dispense with the exponential moving average (EMA) target network typically used with non-contrastive methods. Our analysis sheds light on the variance regularization mechanisms of non-contrastive SSL and lays the theoretical grounds for crafting novel loss functions that shape the learning dynamics of the predictor s spectrum. 1 Introduction SSL has emerged as a powerful method to learn useful representations from vast quantities of unlabeled data [1 8]. In SSL, the network s objective is to pull together its outputs for two differently augmented versions of the same input, so that they learn representations that are predictive across randomized transformations [9]. To avoid the trivial solution whereby the network output becomes constant, also called representational collapse, SSL methods use either a contrastive objective to push apart representations of unrelated images [2, 3, 10 13] or other non-contrastive strategies. Non-contrastive methods comprise explicit variance regularization techniques [6, 7, 14], whitening approaches [15, 16], and asymmetric losses as in Bootstrap Your Own Latent (BYOL) [1] and Sim Siam [5]. Asymmetric losses break symmetry between the two branches by passing one of the representations through a predictor network and stopping gradients from flowing through the other target branch. How this architectural modification prevents representational collapse is not obvious and has been the focus of several theoretical [17 21] and empirical studies [22 24]. A significant advance was provided by Tian et al. [17] who showed that linear predictors align with the correlation matrix of the embeddings, and proposed the closed-form predictor Direct Pred based on this insight. However, previous analyses assumed a Euclidean loss at the output [17 19, 21] except [20], whereas practical implementations typically use the cosine loss [1, 5] which yields superior performance on downstream tasks. This difference raises the question whether analysis based on the Euclidean loss provides an accurate account of the learning dynamics under the cosine loss. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). In this work, we provide a comparative analysis of the learning dynamics for the Euclidean and cosine-based asymmetric losses in the eigenspace of the closed-form predictor Direct Pred. Our analysis shows how both losses implicitly regularize the variance of the representations, revealing a connection between asymmetric losses and explicit variance regularization in VICReg [7]. Yet, the learning dynamics induced by the two losses are markedly different. While the learning dynamics of different eigenmodes decouple in the Euclidean case, dynamics remain coupled for the cosine loss. Moreover, our analysis shows that for both losses, the predictor s eigenvalues act as learning rate multipliers, thereby slowing down learning for modes with small eigenvalues. Based on our analysis, we craft an isotropic loss function (Iso Loss) for each case that resolves this problem and speeds up the initial learning dynamics. Furthermore, Iso Loss works without an EMA target network possibly because it boosts small eigenvalues, the purported role of the EMA in Direct Pred [17]. In summary, our main contributions are the following: We analyze the SSL dynamics in the eigenspace of closed-form linear predictors for asymmetric Euclidean and cosine losses and show that both perform implicit variance regularization, but with markedly different learning dynamics. Our analysis shows that predictor eigenvalues act as learning rate multipliers which slows down learning for small eigenvalues. We propose isotropic loss functions for both cases that equalize the dynamics across eigenmodes and improve robustness, thereby allowing to learn without an EMA target network. 2 Eigenspace analysis of the learning dynamics To gain a better analytic understanding of the SSL dynamics underlying non-contrastive methods such as BYOL and Sim Siam [1, 5], we analyze them in the predictor s eigenspace. Specifically we proceed in three steps. First, building on Direct Pred, we invoke the neural tangent kernel (NTK) to derive simple dynamic expressions of the predictor s eigenmodes for Euclidean and cosine loss. This formulation uncovers the implicit variance regularization mechanisms that prevent representational collapse. Using the eigenspace framework, we illustrate how removing the predictor or the stopgradient results in collapse or run-away dynamics. Finally, we find that predictor eigenvalues act as learning rate multipliers for their associated mode, thereby slowing down learning for small eigenvalues. We derive a modified isotropic loss function (Iso Loss) that provides more equalized learning dynamics across modes, which showcases how our analytic insights help to design novel loss functions that actively shape the predictor spectrum. However, before we start our analysis, we will briefly review Direct Pred [17] and the NTK [25], a powerful theoretical tool linking representational changes and parameter updates. We will rely on both concepts for our analysis. 2.1 Background and problem setup We begin by reviewing Direct Pred [17] and defining our notation. In the following, we consider a Siamese neural network z = f (x; θ) with output z RM, input x RN, and parameters θ. We further assume a linear predictor network WP RM M and use the same parameters for the online and target branches as in Sim Siam [5]. We denote pairs of representations as z(1), z(2) corresponding to pairs of inputs x(1), x(2) related through augmentation and implicitly assume that all losses are averaged over many augmented pairs. The asymmetric loss function (Fig. 1a), introduced in BYOL [1], is then given by: L = d WPz(1), SG(z(2)) , where SG denotes the stop-gradient operation, and d is either the Euclidean distance metric d(a, b) = 1 2 a b 2 or the cosine distance metric d(a, b) = a b a b . We refer to the corresponding loss functions as Leuc and Lcos respectively. Direct Pred. Tian et al. [17] showed that a linear predictor in the BYOL setting aligns during learning with the correlation matrix of representations Cz := Ex zz , where the expectation is taken over the data distribution. Since the correlation matrix is a real symmetric matrix, one can diagonalize it over R: Cz = UDCU , where U is an orthogonal matrix whose columns are the eigenvectors of Cz and DC is the real-valued diagonal matrix of the eigenvalues sm with m [1, M]. Stop-grad x Predictor Online net Target net L = d WPz(1), SG(z(2)) b Euclidean distance Cosine distance Figure 1: (a) Schematic of a Siamese network with a predictor network and a stop-gradient on the target network branch. The target network can be a copy (Sim Siam [5]) or a moving average (BYOL [1]) of the online network. In either case, the target network is not optimized with gradient descent. (b) Visualization of learning dynamics under the Euclidean distance metric showing learning update directions along two eigenmodes, with the light cloud representing the distribution of the representations z, the darker cloud representing the predictor outputs WPz, and the dotted circle indicates the steady state λ1,2 = 1, reached during learning. All eigenvalues converge to one. (c) Same as (b), but for the cosine distance. The dotted line indicates the steady state λ1 = λ2. Given this eigendecomposition, the authors proposed Direct Pred, in which the predictor is not learned via gradient descent but directly set to: WP = fα (Cz) = UDα CU , (1) where α is a positive constant exponent applied element-wise to DC. The eigenvalues λm of the predictor matrix WP are then λm = sα m. We use D to denote the diagonal matrix containing the eigenvalues λm. While Direct Pred used α = 0.5, the follow-up study Direct Copy [18] showed that α = 1 is also effective while avoiding the expensive diagonalization step. While Tian et al. [17] based their analysis on the Euclidean loss Leuc, most practical models, including Tian et al. s large-scale experiments, relied on the cosine similarity loss Lcos. This discrepancy raises the question to what extent setting the predictor to the above expression is justified for the cosine loss. Empirically, we find that a trainable linear predictor does align its eigenspace with that of the representation correlation matrix also for the cosine loss (see Fig. 4 in Appendix A). Neural tangent kernel (NTK). The NTK is a powerful analytical tool characterizing the learning dynamics of neural networks [25, 26]. Here, we recall the definition of the empirical NTK [26] corresponding to a single instantiation of the network s parameters θ. If |D| denotes the size of the training dataset, L : RM R an arbitrary loss function, X, the training data concatenated into one vector of size N|D|, and Z = z(X), the concatenated output of size M|D|, then the empirical NTK is the (M|D| M|D|)-sized matrix: Θt(X, X) = θZ θZ , and the continuous-time gradient-descent dynamics [26] of the representations z are given by: dt = ηΘt(x, X) ZL . (2) In other words, the empirical NTK links the representational dynamics dz dt under gradient descent on the parameters θ, and the representational gradient ZL. 2.2 Implicit variance regularization in non-contrastive SSL As a starting point for our analysis, we first express the relevant loss functions in the eigenbasis of the predictor network. We do this using a closed-form linear predictor as prescribed by Direct Pred. In the following, we use ˆz = U z to denote the representation expressed in the eigenbasis. Lemma 1. (Euclidean and cosine loss in the predictor eigenspace) Let WP be a linear predictor set according to Direct Pred with eigenvalues λm, and ˆz the representations expressed in the predictor s eigenbasis. Then the asymmetric Euclidean loss Leuc and cosine loss Lcos can be expressed as: m |λmˆz(1) m SG(ˆz(2) m )|2 , (3) λmˆz(1) m SG(ˆz(2) m ) Dˆz(1) SG(ˆz(2)) . (4) for which we defer the simple proof to Appendix B. Rewriting the losses in the eigenbasis makes it clear that the asymmetric loss with Direct Pred can be viewed as an implicit loss function in the predictor s eigenspace, where the variance of each mode naturally appears through the λm terms. In the following analysis, we will show how the learning dynamics implicitly regularize these variances λm. From Eq. (3) we directly see that Leuc is a sum of M terms, one for each eigenmode, which decouples the learning dynamics, a fact first noted by Tian et al. [17]. In contrast, the form of Lcos yields coupled dynamics due to the Dˆz(1) = q P k(λkˆz(1) k )2 term in the denominator. This coupling arises from the normalization of the representation vectors to the unit hypersphere when calculating the cosine distance. The normalization effectively removes one degree of freedom and, in the process, adds a dependence between all the representation dimensions (Fig. 1b and 1c). To get an analytic handle on the evolution of the eigen-representations ˆz as the encoder learns, we first note that if training were to update the representations directly, instead of indirectly through updating the weights θ, they would evolve along the following representational gradients : ˆz(1)Leuc = Dˆz(1) ˆz(2) D , (5) ˆz(1)Lcos = Dˆz(2) Dˆz(1) ˆz(2) + (Dˆz(1)) ˆz(2) Dˆz(1) 3 ˆz(2) D2ˆz(1) . (6) In practice, however, representations of different samples do not evolve independently along these gradients, but influence each other through parameter changes in θ. This interdependence of representations and parameters are captured by the empirical NTK Θt(X, X) (cf. Eq. (2)). Because the NTK is positive semi-definite, loosely speaking, gradient descent on the parameters changes representations in the direction of the above representational gradients. To see this link more formally, we express the NTK in the eigenbasis as ˆΘt(X, X) = θ ˆZ θ ˆZ where ˆZ = ˆzt(X) = U zt(X). Since we are concerned with the learning dynamics in this rotated basis, we will rewrite Eq. (2) for continuous-time gradient descent for a generic loss function L as: dt = η ˆΘt(x, X) ˆ ZL . (7) Note, that structurally these dynamics are the same as the embedding space dynamics in Eq. (2) but merely expressed in the predictor eigenbasis (see Lemma 2 in Appendix B for a derivation). Although ˆΘt changes over time and is generally intractable in finite-width networks, it is positive semidefinite. This property guarantees that the cosine angle between the representational training dynamics under the parameter-space optimization of a neural network d dt ˆZ ˆΘt ˆ ZL and the dynamics that would result from optimizing the representations d dt ˆZ ˆ ZL is non-negative: * = η D ˆ ZL, ˆΘt ˆ ZL E 0. In other words, the representational updates due to network training lie within a 180-degree cone of the dynamics prescribed by Eqs. (5) and (6). This guarantee makes it possible to draw qualitative conclusions about asymptotic collective behavior, e.g., whether a network is bound to collapse or not, from analyzing the more tractable dynamics that follow the representational gradients d dt ˆZ ˆ ZL of the transformed BYOL/Sim Siam loss. For ease of analysis, we now consider linear networks with Gaussian i.i.d inputs, an important limiting case amenable for theoretical analysis [27]. In this setting the empirical NTK becomes the identity and the simplified representational dynamics are exact, allowing us to fully characterize the representational dynamics for Leuc and Lcos in the following two theorems. In the proofs for these theorems, we show that the assumption of Gaussian inputs can be relaxed further. Theorem 1. (Representational dynamics under Leuc) For a linear network with i.i.d Gaussian inputs learning with Leuc, the representational dynamics of each mode m independently follow the gradient of the loss ˆz Leuc. More specifically, the dynamics uncouple and follow M independent differential equations: dˆz(1) m dt = η Leuc ˆz(1) m (t) = ηλm ˆz(2) m λmˆz(1) m , (8) which, after taking the expectation over augmentations yields the dynamics: dt = ηλm (1 λm) ˆzm . (9) We provide the proof in Appendix B and appreciate that d dt ˆzm has the same sign as ˆzm whenever λm < 1 and the opposite sign whenever λm > 1. These dynamics are convergent and approach an eigenvalue λm of one, thereby preventing collapse of mode m. Since the eigenmodes are orthogonal and uncorrelated, and the condition simultaneously holds for all modes, this ultimately prevents both representational and dimensional collapse [28]. Since the eigenvalues also correspond to the variance of the representations, the underlying mechanism constitutes an implicit form of variance regularization. Finally, we note that the above decoupling of the dynamics for the Euclidean loss has been described previously in Tian et al. [17]. Nevertheless, the representational dynamics are different for the commonly used cosine loss Lcos. Theorem 2. (Representational dynamics under Lcos) For a linear network with i.i.d Gaussian inputs trained with Lcos, the dynamics follow a system of M coupled differential equations: dˆz(1) m dt = η λm Dˆz(1) 3 ˆz(2) k =m λk λk(ˆz(1) k )2ˆz(2) m λmˆz(1) m ˆz(1) k ˆz(2) k , (10) and reach a regime in which the eigenvalues are comparable in magnitude. In this regime, the expected update over augmentations is well approximated by: dt ηλm E ˆz2 m Dˆz 3 k =m λk (λk λm) , (11) where we have assumed averages over augmentations. See Appendix B for the proof. Theorem 2 states that d dt ˆzm has the same or different sign as ˆzm depending on the sign of the aggregate sum P k =m λk(λk λm). This relation suggests that a steady state is only reached through mutual agreement when the non-zero eigenvalues are all equal. In contrast to the Euclidean case, there is no pre-specified target value (see Fig. 5 in Appendix A). Thus, the cosine loss also induces implicit variance regularization, but through a markedly different mechanism in which eigenmodes cooperate. 2.3 Stop-grad and predictor network are essential for implicit variance regularization. We now extend our analysis to explain the known failure modes due to ablating the predictor or the stop-gradient for each distance metric. When we omit the stop-grad operator from Leuc, we have: Leuc no SG = 1 2 WPz(1) z(2) 2 dˆzm dt = η (1 λm)2 ˆzm , (12) dt ˆzm and ˆzm always have opposite signs (see Appendix C for the derivation). This drives the representations toward zero with exponentially decaying eigenvalues, causing the notorious representational collapse [5]. Omitting the stop-grad operator from Lcos yields a nontrivial expression for the dynamics causing the largest eigenmode to diverge (see Appendix C) . Interestingly, this is different from the collapse to zero inferred for the Euclidean distance. Similarly, when removing the predictor network in the Euclidean loss case, the dynamics read: Leuc no Pred = 1 2 z(1) SG(z(2)) 2 dˆzm dt = 0 , (13) meaning that no learning updates occur. When the predictor is removed in the cosine loss case, the dynamics are: Lcos no Pred = z(1) SG(z(2)) z(1) SG(z(2)) dˆzm E ˆz2 k ˆz 3 As we show in Appendix C, these dynamics also avoid collapse. However, the effective learning rates become impractically small without the eigenvalue factors from Eq. (11). We summarized the predicted dynamics of all settings in Table 1. Thus, our analysis provides mechanistic explanations for why stop-grad and predictor networks are required for avoiding collapse in non-contrastive SSL. Table 1: Summary of eigendynamics as predicted by our analysis for linear networks. Loss dˆzm/dt Predicted dynamics Leuc λm(1 λm) λs converge to 1, large ones faster. Leuc no SG (1 λm)2 All λs collapse. Leuc no Pred 0 No learning updates. Leuc iso (1 λm) λs converge to 1 at homogeneous rates. k =m λk(λk λm) λs converge to equal values. Lcos no SG Appendix C All λs diverge. Lcos no Pred Appendix C λs converge to equal values at low rates. k =m λk(λk λm) λs converge to equal values at homogeneous rates. 2.4 Isotropic losses that equalize convergence across eigenmodes In Eqs. (9) and (11) the eigenvalues appear as multiplicative learning rate modifiers in front of the difference terms that determine the fixed point. Hence, modes with larger eigenvalues converge faster than modes with smaller eigenvalues, reminiscent of previous theoretical work on supervised learning [27]. We hypothesized that the anisotropy in learning dynamics could lead to slow convergence for small eigenvalue modes or instability for large eigenvalues. To alleviate this issue, we designed alternative isotropic loss functions that equalize relaxation dynamics for all eigenmodes by exploiting the stop-grad function. Put simply, this involves taking the dynamics from Eqs. (8) and (10), removing the leading λm term, and deriving the loss function that would result in the desired dynamics. One such isotropic Iso Loss function for the Euclidean distance is: Leuc iso = 1 2 z(1) SG(z(2) + z(1) WPz(1)) 2. (15) We note that this Iso Loss has the same numerical value as Leuc, but the gradient flow is modified by placing the prediction inside the stop-grad and also adding and subtracting z(1) inside and outside of the stop-grad. The associated idealized learning dynamics in our analytic framework are given by: dˆzm dt = η (1 λm) ˆzm, (16) where the λm factor (cf. Eq. (9)) disappeared (Table 1). Similarly, for the cosine distance, Lcos iso = (z(1)) SG z(2) WPz(1) z(2) 2SG (WPz(1)) z(2) WPz(1) 3 z(2) W 1/2 P z(1) 2 (17) is one possible Iso Loss, in which W 1/2 P = UD1/2U with the square-root applied element-wise to the diagonal matrix D. While this Iso Loss does not preserve numerical equality with the original loss Lcos, it achieves the desired effect of removing the leading λm learning-rate modifier (cf. Table 1). a Leuc no SG 0 1000 Iteration b Leuc no Pred 0 100000 Iteration 0 1000 Iteration 0 1000 Iteration 0 1000 Iteration Lcos no Pred 0 100000 Iteration 0 1000 Iteration 0 1000 Iteration Figure 2: Evolution of representations (top) and eigenvalues (below) of WP throughout training with different loss functions. The representational trajectories correspond to training with M = 2 for visualization and the points signify the final network outputs. The eigenvalues were computed with dimensions N = 15 and M = 10. (a) Omitting the stop-grad leads to representational collapse in the Euclidean case (top), and diverging eigenvalues for the cosine case (bottom). (b) No learning occurs without the predictor with the Euclidean distance, but learning does occur with the cosine distance, although at low rates. Note the change in scale of the time-axis. (c) Optimizing the BYOL/Sim Siam loss leads to isotropic representations under both distance metrics. (d) Optimizing Iso Loss has the same effect, but with uniform convergence dynamics for all eigenvalues for both distance metrics. 3 Numerical experiments To validate our theoretical findings (cf. Table 1), we first simulated a small linear Siamese neural network as shown in Fig.1a, for which Theorems 1 and 2 hold exactly. We fed the network with independent standard Gaussian inputs, and generated pairs of augmentations using isotropic Gaussian perturbations of standard deviation σ = 0.1. We then trained the linear encoder with each configuration described above. Training the network with Leuc no SG resulted in collapse with exponentially decaying eigenvalues, whereas Lcos no SG succumbed to diverging eigenvalues as predicted (Fig. 2a). Training without the predictor caused vanishing updates for Leuc no Pred and slow learning for Lcos no Pred, in line with our analysis (Fig. 2b). Optimizing Leuc, the representations become increasingly isotropic with all the eigenvalues λm converging to one (Fig. 2c, top), whereas optimizing Lcos also resulted in the eigenvalues converging to the same value, but different from one (Fig. 2c, bottom). The anisotropy in the dynamics of different eigenmodes noted above is particularly striking in the case of the Euclidean distance (Fig. 2c). Training with Leuc iso and Lcos iso resulted in similar convergence properties as their non-isotropic counterparts, but the eigenmodes converged at more homogeneous rates (Fig. 2d). Finally, we confirmed that these findings were qualitatively similar in the corresponding nonlinear networks with Re LU activation (see Fig. 6 in Appendix A). Thus, our theoretical findings hold up in simple Siamese networks. Leuc Leuc iso Leuc BYOL 0 500 1000 Epoch 0 500 1000 Epoch 0 500 1000 Epoch Lcos Lcos iso Lcos BYOL 0 500 Epoch Figure 3: Learning dynamics for a Res Net-18 network trained with different loss functions. (a) Evolution of the eigenvalues of the representation correlation matrix during training for closed-form predictors as prescribed by Direct Pred (left) and Iso Loss (center). Right: Standard BYOL with the nonlinear trainable predictor. For clarity, we plot only one in ten eigenvalues. Both Leuc and Leuc iso drive the eigenvalues to converge quickly and remain constant thereafter with relatively small fluctuations (note the logarithmic scale). BYOL results in the eigenvalues being spread across a large range of magnitudes. (b) Linear readout validation accuracy for Leuc and Leuc iso during the first 500 training epochs. Iso Loss accelerates the initial learning dynamics as predicted by the theory. (c) Same as in (a) but for the cosine distance. Lcos recruits few large eigenvalues, but drives them gradually to the same magnitude, whereas Lcos iso quickly recruits all eigenvalues and causing them to converge to an isotropic solution. In contrast, BYOL recruits eigenvalues in a step-wise manner. (d) Same as (b) but for the cosine distance. 3.1 Theory qualitatively captures dynamics in nonlinear networks and real-world datasets. To investigate how well our theoretical analysis holds up in non-toy settings, we performed several self-supervised learning experiments on CIFAR-10, CIFAR-100 [29], STL-10 [30], and Tiny Image Net [31]. We based our implementation1 on the Solo-learn library [32], and used a Res Net-18 backbone [33] as the encoder and the cosine loss, unless mentioned otherwise (see Appendix D for details). As baselines for comparison, we trained the same backbone using BYOL with the nonlinear predictor and Direct Pred with the closed-form linear predictor. We recorded the online readout accuracy of a linear classifier trained on frozen features following standard practice, evaluated either on the held-out validation or test set where available. We found that the eigenvalue dynamics of the representational correlation matrix in the Res Net-18 closely mirrored the analytical predictions for the closed-form predictor. For Euclidean distances (Fig. 3a), the eigenvalues for Direct Pred and Iso Loss converged to a small range of values around one. However, the dynamics for BYOL with a learnable nonlinear predictor deviated significantly with the eigenvalues distributed over a larger range. Consistent with our analysis, Iso Loss had faster initial dynamics for the eigenvalues which also resulted in a faster initial improvements in model performance (Fig. 3b). The faster learning with Iso Loss was even more evident for the cosine distance (Fig. 3c). Surprisingly, BYOL, which uses a nonlinear predictor also closely matched the predicted dynamics in the case of the cosine distance. Furthermore, the dynamics showed a stepwise learning phenomenon wherein eigenvalues are progressively recruited one-by-one, consistent with recent findings for other SSL methods [34]. Finally, Iso Loss exhibited faster initial learning (Fig. 3d), in agreement with our theoretical analysis. Thus, our theoretical analysis accurately predicts key properties of the eigenvalue dynamics in nonlinear networks trained on real-world datasets. 1Code is available at https://github.com/fmi-basel/implicit-var-reg 3.2 Iso Loss promotes eigenvalue recruitment and works without an EMA target network. To further investigate the impact of Iso Loss on learning, we first verified that it does not have any adverse effects on downstream classification performance. We found that Iso Loss matched or outperformed Direct Pred on all benchmarks (Table 2) when trained with an EMA target network as used in the original studies. Yet, it performed slightly worse than BYOL, which uses a nonlinear predictor and an EMA target network. Because EMA target networks are thought to amplify small eigenvalues [17], we speculated that Iso Loss may work without it. We repeated training for the closed-form predictor losses without EMA to test this idea. We found that Lcos iso was indeed robust to EMA removal. However, it caused a slight drop in performance (Table 2) and a notable reduction in the recruitment of small eigenvalues (see Fig. 7 in Appendix A). In contrast, optimizing the standard BYOL/Sim Siam loss Lcos with the symmetric linear predictor was unstable, as reported previously [17]. Finally, we confirmed the above findings also hold for α = 1 (cf. Eq. (1)) as prescribed by Direct Copy [18] (see Table 3 in Appendix A). Thus, Iso Loss allows training without an EMA target network. Table 2: Linear readout validation accuracy in % stddev over five random seeds. The denotes crashed runs, known to occur with symmetric predictors like Direct Pred [17]. Starred values were taken from the Solo-learn library [32]. Model EMA CIFAR-10 CIFAR-100 STL-10 Tiny Image Net BYOL Yes 92.6 70.5 91.7 0.1 38.3 1.5 Sim Siam No 90.7 0.2 66.3 0.4 87.5 0.7 39.8 0.6 Direct Pred (α = 0.5) Yes 92.0 0.2 66.6 0.5 88.8 0.3 40.1 0.5 No 12.1 1.3 1.6 0.6 10.4 0.1 1.3 0.2 Iso Loss (ours) Yes 91.5 0.2 69.0 0.2 89.0 0.3 44.8 0.4 No 91.5 0.2 64.3 0.3 87.4 0.1 40.4 0.4 The above result suggests that Iso Loss promotes the recruitment of small eigenvalues in closed-form predictors. Another factor that has been implicated in suppressing recruitment is weight decay [18]. To probe how weight decay and Iso Loss affect small eigenvalue recruitment, we repeated the above simulations with EMA and different amounts of weight decay. Indeed, we observed less eigenvalue recruitment with increasing weight decay for Direct Pred (Appendix A, Fig. 8a), but not for Iso Loss (Fig. 8b). However, for Iso Loss larger weight decay resulted in lower magnitudes of all eigenvalues. Hence, Iso Loss reduces the impact of weight decay on eigenvalue recruitment. 4 Discussion We provided a comprehensive analysis of the SSL representational dynamics in the eigenspace of closed-form linear predictor networks (i.e., Direct Pred and Direct Copy) for both the Euclidean loss and the more commonly used cosine similarity. Our analysis revealed how asymmetric losses prevent representational and dimensional collapse through implicit variance regularization along orthogonal eigenmodes, thereby formally linking predictor-based SSL with explicit variance regularization approaches [6, 7, 14]. Our work provides a theory framework which further complements the growing body of work linking contrastive and non-contrastive SSL [24, 35 38]. We empirically validated the key predictions of our analysis in linear and nonlinear network models on several datasets, including CIFAR-10/100, STL-10, and Tiny Image Net. Moreover, we found that the eigenvalues of the predictor network act as learning rate multipliers, causing anisotropic learning dynamics. We derived Euclidean and cosine Iso Losses, which counteract this anisotropy and enable closed-form linear predictor methods to work without an EMA target network, thereby further consolidating its presumed role in boosting small eigenvalues [17]. To our knowledge, this is the first work to comprehensively characterize asymmetric SSL learning dynamics for the cosine distance metric widely used in practice. However, our analysis rests on several assumptions. First, the analytic link through the NTK between gradient descent on parameters and the representational changes is an approximation in nonlinear networks. Moreover, we assumed Gaussian i.i.d inputs for proving Theorems 1 and 2. Although these assumptions generally do not hold in nonlinear networks, our analysis qualitatively captures their overall learning behavior and predicts how networks respond to changes in the stop-grad placement. In summary, we have provided a simple theoretical explanation of how asymmetric loss configurations prevent representational collapse in SSL and elucidate their inherent dependence on the placement of the stop-grad operation. We further demonstrated how the eigenspace framework allows crafting new loss functions with a distinct impact on the SSL learning dynamics. We provided one specific example of such loss functions, Iso Loss, which equalizes the learning dynamics in the predictor s eigenspace, resulting in faster initial learning and improved stability. In contrast to Direct Pred, Iso Loss learns stably without an EMA target network. Our work thus lays out an effective framework for analyzing and developing new SSL loss functions. Acknowledgments and Disclosure of Funding This project was supported by the Swiss National Science Foundation [grant numbers PCEFP3_202981 and TMPFP3_210282], by EU s Horizon Europe Research and Innovation Programme (grant agreement number 101070374) funded through SERI (ref 1131-52302), and the Novartis Research Foundation. The authors declare no competing interests. [1] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, 33:21271 21284, 2020. [2] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597 1607. PMLR, 2020. [3] Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. Unsupervised learning of visual features by contrasting cluster assignments. Advances in Neural Information Processing Systems, 33:9912 9924, 2020. [4] Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, pages 9650 9660, 2021. [5] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15750 15758, 2021. [6] Jure Zbontar, Li Jing, Ishan Misra, Yann Le Cun, and Stéphane Deny. Barlow twins: Selfsupervised learning via redundancy reduction. In International Conference on Machine Learning, pages 12310 12320. PMLR, 2021. [7] Adrien Bardes, Jean Ponce, and Yann Le Cun. Vicreg: Variance-invariance-covariance regularization for self-supervised learning. ar Xiv preprint ar Xiv:2105.04906, 2021. [8] Mahmoud Assran, Mathilde Caron, Ishan Misra, Piotr Bojanowski, Florian Bordes, Pascal Vincent, Armand Joulin, Mike Rabbat, and Nicolas Ballas. Masked siamese networks for label-efficient learning. In Computer Vision ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23 27, 2022, Proceedings, Part XXXI, pages 456 473. Springer, 2022. [9] Yonglong Tian, Chen Sun, Ben Poole, Dilip Krishnan, Cordelia Schmid, and Phillip Isola. What makes for good views for contrastive learning? Advances in neural information processing systems, 33:6827 6839, 2020. [10] Jane Bromley, Isabelle Guyon, Yann Le Cun, Eduard Säckinger, and Roopak Shah. Signature verification using a" siamese" time delay neural network. Advances in neural information processing systems, 6, 1993. [11] Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. ar Xiv preprint ar Xiv:1807.03748, 2018. [12] Tongzhou Wang and Phillip Isola. Understanding contrastive representation learning through alignment and uniformity on the hypersphere. In International Conference on Machine Learning, pages 9929 9939. PMLR, 2020. [13] Ting Chen, Calvin Luo, and Lala Li. Intriguing properties of contrastive losses. Advances in Neural Information Processing Systems, 34:11834 11845, 2021. [14] Manu Srinath Halvagal and Friedemann Zenke. The combination of Hebbian and predictive plasticity learns invariant object representations in deep sensory networks. bio Rxiv, accepte, 2022. doi: 10.1101/2022.03.17.484712. URL https://www.biorxiv.org/content/10. 1101/2022.03.17.484712v2. [15] Aleksandr Ermolov, Aliaksandr Siarohin, Enver Sangineto, and Nicu Sebe. Whitening for self-supervised representation learning. In International Conference on Machine Learning, pages 3015 3024. PMLR, 2021. [16] Xi Weng, Lei Huang, Lei Zhao, Rao Anwer, Salman H Khan, and Fahad Shahbaz Khan. An investigation into whitening loss for self-supervised learning. Advances in Neural Information Processing Systems, 35:29748 29760, 2022. [17] Yuandong Tian, Xinlei Chen, and Surya Ganguli. Understanding self-supervised learning dynamics without contrastive pairs. In International Conference on Machine Learning, pages 10268 10278. PMLR, 2021. [18] Xiang Wang, Xinlei Chen, Simon S Du, and Yuandong Tian. Towards demystifying representation learning with non-contrastive self-supervision. ar Xiv preprint ar Xiv:2110.04947, 2021. [19] Kang-Jun Liu, Masanori Suganuma, and Takayuki Okatani. Bridging the gap from asymmetry tricks to decorrelation principles in non-contrastive self-supervised learning. Advances in Neural Information Processing Systems, 35:19824 19835, 2022. [20] Chenxin Tao, Honghui Wang, Xizhou Zhu, Jiahua Dong, Shiji Song, Gao Huang, and Jifeng Dai. Exploring the equivalence of siamese self-supervised learning via a unified gradient framework. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14431 14440, 2022. [21] Pierre H Richemond, Allison Tam, Yunhao Tang, Florian Strub, Bilal Piot, and Felix Hill. The edge of orthogonality: A simple view of what makes byol tick. ar Xiv preprint ar Xiv:2302.04817, 2023. [22] Pierre H Richemond, Jean-Bastien Grill, Florent Altché, Corentin Tallec, Florian Strub, Andrew Brock, Samuel Smith, Soham De, Razvan Pascanu, Bilal Piot, et al. Byol works even without batch statistics. ar Xiv preprint ar Xiv:2010.10241, 2020. [23] Xiao Wang, Haoqi Fan, Yuandong Tian, Daisuke Kihara, and Xinlei Chen. On the importance of asymmetry for siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16570 16579, 2022. [24] Chaoning Zhang, Kang Zhang, Chenshuang Zhang, Trung X Pham, Chang D Yoo, and In So Kweon. How does simsiam avoid collapse without negative samples? a unified understanding with self-supervised contrastive learning. ar Xiv preprint ar Xiv:2203.16262, 2022. [25] Arthur Jacot, Franck Gabriel, and Clément Hongler. Neural tangent kernel: Convergence and generalization in neural networks. Advances in neural information processing systems, 31, 2018. [26] Jaehoon Lee, Lechao Xiao, Samuel Schoenholz, Yasaman Bahri, Roman Novak, Jascha Sohl-Dickstein, and Jeffrey Pennington. Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent. Advances in Neural Information Processing Systems, 32:8572 8583, 2019. URL https://proceedings.neurips.cc/paper/2019/hash/ 0d1a9651497a38d8b1c3871c84528bd4-Abstract.html. [27] Andrew M. Saxe, James L. Mc Clelland, and Surya Ganguli. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. ar Xiv:1312.6120 [cond-mat, q-bio, stat], December 2013. URL http://arxiv.org/abs/1312.6120. [28] Li Jing, Pascal Vincent, Yann Le Cun, and Yuandong Tian. Understanding dimensional collapse in contrastive self-supervised learning. ar Xiv preprint ar Xiv:2110.09348, 2021. [29] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images, 2009. [30] Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, pages 215 223. JMLR Workshop and Conference Proceedings, 2011. [31] Ya Le and Xuan Yang. Tiny imagenet visual recognition challenge. CS 231N, 7(7):3, 2015. [32] Victor Guilherme Turrisi da Costa, Enrico Fini, Moin Nabi, Nicu Sebe, and Elisa Ricci. sololearn: A library of self-supervised methods for visual representation learning. Journal of Machine Learning Research, 23(56):1 6, 2022. URL http://jmlr.org/papers/v23/21-1155. html. [33] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. [34] James B Simon, Maksis Knutins, Liu Ziyin, Daniel Geisz, Abraham J Fetterman, and Joshua Albrecht. On the stepwise nature of self-supervised learning. ar Xiv preprint ar Xiv:2303.15438, 2023. [35] Quentin Garrido, Yubei Chen, Adrien Bardes, Laurent Najman, and Yann Lecun. On the duality between contrastive and non-contrastive self-supervised learning. ar Xiv preprint ar Xiv:2206.02574, 2022. [36] Randall Balestriero and Yann Le Cun. Contrastive and non-contrastive self-supervised learning recover global and local spectral embedding methods. ar Xiv preprint ar Xiv:2205.11508, 2022. [37] Yann Dubois, Stefano Ermon, Tatsunori B Hashimoto, and Percy S Liang. Improving selfsupervised learning by characterizing idealized representations. Advances in Neural Information Processing Systems, 35:11279 11296, 2022. [38] Ashwini Pokle, Jinjin Tian, Yuchen Li, and Andrej Risteski. Contrasting the landscape of contrastive and non-contrastive learning. ar Xiv preprint ar Xiv:2203.15702, 2022. A Supplementary Material 0 100 200 Sorted index 0 100 200 Sorted index Lcos Leuc b Figure 4: Eigenspace alignment between Cz = Ex zz and a linear predictor WP trained with gradient descent. Following [1], we measure eigenspace alignment as the cosine between ui and WPui for every eigenvector ui of Cz. (a) Measured alignment for every eigenvector of Cz ordered by sorted eigenvalue indices over training epochs, when the network is trained with Leuc. (b) Same as (a) but for training with Lcos. 0 1000 Steps 0 1000 Steps Figure 5: Comparison between the dynamics under Euclidean and cosine asymmetric losses for different initializations in a network with M = 2 output neurons. (a) Observed dynamics of the eigenvalues in the two-neuron toy network under three different initializations. Both eigenvalues always converge to 1 regardless of the initialization. (b) Same as (a), but for the cosine distance. Under different initializations, the two eigenvalues converge to arbitrary, but equal, values. Table 3: Linear readout validation accuracy in % stddev over five random seeds. The denotes crashed runs, known to occur with symmetric predictors like Direct Pred [1]. Model α EMA CIFAR-10 CIFAR-100 STL-10 Tiny Image Net Direct Copy 1 Yes 91.3 0.2 68.7 0.3 89.3 0.2 45.3 0.3 No 12.1 0.6 1.5 0.5 10.3 0.1 0.6 0.1 Direct Pred 0.5 Yes 92.0 0.2 66.6 0.5 88.8 0.3 40.1 0.5 No 12.1 1.3 1.6 0.6 10.4 0.1 1.3 0.2 Iso Loss (ours) 1 Yes 91.5 0.2 69.3 0.2 89.1 0.3 45.6 0.9 No 91.4 0.2 63.0 0.3 86.9 0.4 38.5 0.5 0.5 Yes 91.5 0.2 69.0 0.2 89.0 0.3 44.8 0.4 No 91.5 0.2 64.3 0.3 87.4 0.1 40.4 0.4 a Leuc no SG 0 1000 Iteration b Leuc no Pred 0 100000 Iteration 0 1000 Iteration 0 1000 Iteration 0 1000 Iteration 0 100000 Iteration 0 1000 Iteration 0 1000 Iteration Figure 6: Same as Fig. 2 but with a Re LU nonlinearity on the embeddings. We observe learning dynamics qualitatively similar to the linear network. 0 500 1000 Epoch 0 500 1000 Epoch Figure 7: Eigenvalue dynamics for learning under Iso Loss (Lcos iso ) with and without the EMA target network. Removing the EMA results in markedly different dynamics with fewer eigenmodes recruited during training. wd = 4 10 4 wd = 10 5 wd = 10 6 a 0 500 1000 Epoch 0 500 1000 Epoch 0 500 1000 Epoch Figure 8: Effect of weight decay on eigenvalue recruitment for Direct Pred and Iso Loss. (a) Evolution of eigenvalues during learning under Direct Pred (Lcos) with EMA and different amounts of weight decay. Decreasing weight decay correlates with the number of eigenvalues recruited during learning. (b) Same as (a) but for Iso Loss (Lcos iso ). All eigenvalues are recruited independent of the strength of weight decay. However, the magnitude of the eigenvalues inversely correlates with the magnitude of weight-decay. Lemma 1. (Euclidean and cosine loss in the predictor eigenspace) Let WP be a linear predictor set according to Direct Pred with eigenvalues λm, and ˆz the representations expressed in the predictor s eigenbasis. Then the asymmetric Euclidean loss Leuc and cosine loss Lcos can be expressed as: m |λmˆz(1) m SG(ˆz(2) m )|2 , (3) λmˆz(1) m SG(ˆz(2) m ) Dˆz(1) SG(ˆz(2)) . (4) Proof. Under Direct Pred, the predictor is a symmetric matrix with eigendecomposition WP = UDU . Since U is an orthogonal matrix, we also have UU = I so that we can simplify the losses as follows: 2 WPz(1) SG(z(2)) 2 2 UDU z(1) SG(UU z(2)) 2 2 Dˆz(1) SG(ˆz(2)) 2 m |λmˆz(1) m SG(ˆz(2) m )|2 WPz(1) SG(z(2)) WPz(1) SG(z(2)) = (z(1)) UDU SG(z(2)) UDU z(1) SG(UU z(2)) = (ˆz(1)) D SG(ˆz(2)) Dˆz(1) SG(ˆz(2)) λmˆz(1) m SG(ˆz(2) m ) Dˆz(1) SG(ˆz(2)) , where we used the fact that U is orthogonal and therefore does not change the Euclidean norm. ˆz = U z is the representation rotated into the eigenbasis. Lemma 2. (Learning dynamics in a rotated basis) Assuming that a given loss L is optimized by gradient descent on the parameters of a neural network with network outputs z, a given orthogonal transformation ˆz = U z and learning rate η, then the rotated representations ˆz evolve according to the dynamics: dˆz dt = η ˆΘt(x, X) ˆ ZL , where ˆΘt(X, X) = θ ˆZ θ ˆZ is the empirical NTK expressed in the rotated basis. Proof. Let θ be the parameters of the neural network. Then we obtain the representational dynamics using the chain rule in the continuous-time gradient-flow setting [2]: dt = θˆz dθ dt = θˆz ( η θL) = θˆz η θ ˆZ ˆ ZL = η ˆΘt(x, X) ˆ ZL . The above is a reiteration of the derivation of Eq. (2) given by Lee et al. [2], with an additional orthogonal transformation on the network outputs. We proceed by proving the following Lemma which we will use in our proofs of Theorems 1 and 2. Lemma 3. The NTK for a linear network is invariant under orthogonal transformations of the network output. Proof. We first note that for a linear network, the parameters θ are just the feedforward weights W. Therefore, for any orthogonal transformation U of the network output: ˆz = U f(x) = U Wx θˆz = W ˆz = W U Wx = x U , (18) where is the Kronecker product resulting from the fact that every input vector component appears in the update once for each output component. We now study ˆΘt(X, X), the transformed empirical NTK (cf. Lemma 2). The (M M) diagonal blocks in the full (M|D| M|D|) empirical NTK ˆΘt(X, X) correspond to single samples and the off-diagonal blocks are cross-terms between samples, where |D| denotes the size of the training dataset and M the dimension of the outputs. We can develop a generic expression for each (M M) block ˆΘt(xi, xj) corresponding to the interactions between samples i and j as: ˆΘt(xi, xj) = W ˆzi W ˆz j = x i U x j U = x i U (xj U) = x i xj U U = x i xj IM = x i xj IM. (19) where we have used the fact that (A B) = A B and (A B)(C D) = AC BD. Here, IM is the identity matrix of size M. Noting that Eq. (19) is unchanged when U is just the identity matrix completes the proof. Theorem 1. (Representational dynamics under Leuc) For a linear network with i.i.d Gaussian inputs learning with Leuc, the representational dynamics of each mode m independently follow the gradient of the loss ˆz Leuc. More specifically, the dynamics uncouple and follow M independent differential equations: dˆz(1) m dt = η Leuc ˆz(1) m (t) = ηλm ˆz(2) m λmˆz(1) m , (8) which, after taking the expectation over augmentations yields the dynamics: dt = ηλm (1 λm) ˆzm . (9) Proof. For a linear network with weights W RM N, we have from Lemma 3 that the empirical NTK ˆΘ(X, X) in the orthogonal eigenbasis is equal to the empirical NTK Θ(X, X) in the original basis. Furthermore from the proof for the lemma (see Eq. (19) above), each (M M) block of the full (M|D| M|D|) empirical NTK is given by: ˆΘt(xi, xj) = x i xj IM. (20) where IM RM M is the identity. Eq. (20) gives the total effective interaction between the samples i and j from the dataset. For high-dimensional inputs x drawn from an i.i.d standard Gaussian distribution, we have x i xj δij by the central limit theorem. Therefore, in the special case of a linear network with Gaussian i.i.d inputs, the representational dynamics (Lemma 2) simplify as follows: dˆz(1) i dt = η ˆΘt(xi, X) ˆ ZL = η ˆΘt(xi, xi) ˆzi L η X j =i ˆΘt(xi, xj) ˆzj L = η x i xi ˆzi L η X x i xj ˆzj L = η ˆzi L . (21) While the assumption of Gaussian i.i.d inputs is quite restrictive, we offer a generalizing interpretation here. Specifically, the above argument also holds when the inputs x are not all mutually orthogonal, but fall into P orthogonal clusters in the input dataset. Then, we would have x i xj δpi=pj where pi is the label of the cluster corresponding to sample i. If Pi is the number of all the samples with the same label pi, then Eq. (21) would simply be scaled to give dˆz(1) i dt = ηPi ˆzi L. For brevity, we proceed with the simplest case Eq. (21) in which every input is orthogonal. For Leuc, the representational gradient ˆzi L is then given by: ˆzi Leuc = Dtˆz(1) i ˆz(2) i Dt Noting that Dt is just a diagonal matrix containing the eigenvalues λm and dropping the sample subscript i for notational ease, we obtain for the m-th component of ˆzi Leuc: ˆzm = λm(λmˆz(1) m ˆz(2) m ) . Substituting this result in Eq. (21) gives us Eq. (8), the expression we were looking for. Finally, introducing ˆzm E[ˆz(1) m ] = E[ˆz(2) m ] as the expectation over augmentations, we find that each eigenmode evolves independently in expectation value as: " dˆz(1) m dt dt = ηλm E[ˆz(2) m ] λm E[ˆz(1) m ] = ηλm (1 λm) ˆzm . Theorem 2. (Representational dynamics under Lcos) For a linear network with i.i.d Gaussian inputs trained with Lcos, the dynamics follow a system of M coupled differential equations: dˆz(1) m dt = η λm Dˆz(1) 3 ˆz(2) k =m λk λk(ˆz(1) k )2ˆz(2) m λmˆz(1) m ˆz(1) k ˆz(2) k , (10) and reach a regime in which the eigenvalues are comparable in magnitude. In this regime, the expected update over augmentations is well approximated by: dt ηλm E ˆz2 m Dˆz 3 k =m λk (λk λm) , (11) Proof. We can retrace the steps from the proof for Theorem 1 until Eq. (21): dˆz(1) i dt = η ˆzi L . ˆzi L is a vector of dimension M. Ignoring the sample subscript i for simplicity, and focusing on the m-th component of ˆzi L, we get: λmˆz(1) m SG(ˆz(2) m ) Dˆz(1) SG(ˆz(2)) ˆz(1) m = λmˆz(2) m Dˆz(1) ˆz(2) + P k λkˆz(1) k ˆz(2) k Dˆz(1) 3 ˆz(2) λ2 mˆz(1) m = λm Dˆz(1) 3 ˆz(2) Dˆz(1) 2ˆz(2) m λmˆz(1) m k λkˆz(1) k ˆz(2) k = λm Dˆz(1) 3 ˆz(2) k λ2 k(ˆz(1) k )2 ! ˆz(2) m λmˆz(1) m k λkˆz(1) k ˆz(2) k = λm Dˆz(1) 3 ˆz(2) k =m λk λk(ˆz(1) k )2ˆz(2) m λmˆz(1) m ˆz(1) k ˆz(2) k dˆz(1) m dt = η L = ηλm Dˆz(1) 3 ˆz(2) k =m λk λk(ˆz(1) k )2ˆz(2) m λmˆz(1) m ˆz(1) k ˆz(2) k , proving Eq. (10). Assuming sufficiently small augmentations, ˆz(1) k and ˆz(2) k carry the same sign, and the net sign of both terms inside the parenthesis is fully determined by γm sign(ˆz(1) m ). Hence, we may write: dˆz(1) m dt = ηλmγm Dˆz(1) 3 ˆz(2) λ2 k(ˆz(1) k )2|ˆz(2) m | λmλk|ˆz(1) m ||ˆz(1) k ||ˆz(2) k | . It is useful to separate out γm in this manner because every other term in the expression is now non-negative. Then sign(γm dˆz dt ) = sign(ˆzm dˆzm dt ) tells us whether ˆzm tends to increase or decrease in magnitude, as we have argued in the main text. Asymptotic analysis. To get a handle on how the different eigenvalues influence each other, we consider two important limiting cases. First, we consider the asymptotic regime dominated by one eigenvalue, and show that it tends towards a more symmetric solution in which the gap between different eigenvalues decreases. Second, we derive asymptotic expressions for the near-uniform regime in which all eigenvalues are comparable in size and show that this solution tends toward the uniform solution (cf. Eq. (11)). To facilitate our analysis, we define each mode s relative contribution χm |ˆzm| ˆz and evaluate Eq. (10) taking the expectation value over augmentations: " dˆz(1) m dt " (ˆz(1) k )2ˆz(2) m Dˆz(1) 3 ˆz(2) " ˆz(1) m ˆz(1) k ˆz(2) k Dˆz(1) 3 ˆz(2) λ2 k E ˆz2 k Dˆz 3 λmλk E ˆzmˆzk λ2 k E χ2 k ˆz 2 E [χm] λmλk E χmχk ˆz 2 In the second equality we used the fact that the expectation value taken over augmentations is conditioned on the input sample, which makes them conditionally independent. One dominant eigenvalue. First, we consider the low-rank regime in which one eigenvalue dominates. Without loss of generality, we assume λ1 λk k = 1. We then have: χ1 1 χk ϵ (0 < ϵ 1) k = 1 Plugging these values into Eq. (22) gives the following dynamics for the dominant eigenmode: λ2 k E ϵ2 ˆz 2 E [1] λ1λk E ϵ ˆz 2 λ2 kϵ2 E ˆz 2 E [1] λ1λkϵ2 E ˆz 2 = ηλ1γ1ϵ2E ˆz 2 k =1 λk (λk λ1) . These updates are always opposite in sign to the representation component, which corresponds to decaying dynamics for the leading eigenmode because γ1 dˆz1 For all other modes we have: λ2 kϵ2 E ˆz 2 E [ϵ] λmλkϵ2 E ˆz 2 λ2 1 E ˆz 2 E [ϵ] λmλ1ϵ E ˆz 2 = ηλmγmϵ E ˆz 2 λ1(λ1 λm) + ϵ2 X k {m,1} λk(λk λm) ηλmγmλ1ϵ E ˆz 2 so that γm dˆzm dt > 0, i.e, the updates have the same sign as the representation component, which corresponds to growth dynamics. In other words: The dominant eigenvalue pulls all the other eigenvalues up, a form of implicit cooperation between the eigenmodes. We also note that the non-dominant eigenmodes increase at a rate proportional to ϵ, whereas the dominant eigenmode decreases at a slower rate proportional to ϵ2. Thus, for sensible initializations with at least one large and many small eigenvalues, the modes will tend toward an equilibrium at some non-zero intermediate value, without a dominant mode. Next we study this other limiting case in which all eigenvalues are of similar size. Near-uniform regime. To study the dynamics in a near-uniform regime, we note that all χm are of order O(1) in ˆzm, whereas the eigenvalues λm are of order O(ˆz2 m). In this setting, the effect of the eigenvalue terms λm on the dynamics is stronger than the χm terms which are bounded between 0 and 1. With a sufficiently high-dimensional representation, all χm terms will be centered around 1/ M. Based on these observations, we may make the simplifying assumption that the contributions are all approximately equal, i.e, χi = χ for all i. Substituting this value in Eq (22) gives: dt = ηλmγm E χ2 ˆz 2 k =m λk (λk λm) . (23) Finally, substituting for χ, which by assumption are all approximately equal: χ χm = |ˆzm| and absorbing back the sign from γm, we obtain the approximate dynamics in Eq (11): dt ηλm E ˆz2 m Dˆz 3 k =m λk (λk λm) C Derivation of idealized learning dynamics for different loss variations C.1 Removing the stop-grad from the Euclidean loss Leuc Omitting the stop-grad operator from Leuc gives: Leuc no SG = 1 2 WPz(1) z(2) 2 m |λmˆz(1) m ˆz(2) m |2 . Tracing the steps to prove Theorem 1 and assuming Gaussian i.i.d inputs for a linear network, we write: Leuc no SG ˆzm = Leuc no SG ˆz(1) m + Leuc no SG ˆz(2) m = λmˆz(1) m ˆz(2) m λm λmˆz(1) m ˆz(2) m = λmˆz(1) m ˆz(2) m (λm 1) dt = ηE Leuc no SG ˆzm = η λm E[ˆz(1) m ] E[ˆz(2) m ] (λm 1) = η (1 λm)2 ˆzm , which results in decaying representations and thus collapse. C.2 Removing the stop-grad from the cosine loss Lcos Following the same arguments as above, omitting the stop-grad operator from Lcos gives: Lcos no SG = WPz(1) z(2) WPz(1) z(2) Lcos no SG ˆzm = λm Dˆz(1) 3 ˆz(2) λ2 k(ˆz(1) k )2ˆz(2) m λmλkˆz(1) m ˆz(1) k ˆz(2) k + λ2 kλm(ˆz(1) k )3 λkˆz(2) m ˆz(2) k ˆz(1) k + λm Dˆz(1) 3 ˆz(2) λ3 m(ˆz(1) m )3 λm(ˆz(2) m )2ˆz(1) m , so that, when taking the expectation value over augmentations, the dynamics follow: dt = ηE Lcos no SG ˆzm λk E χ2 k ˆz 2 E [χm] λm E χmχk ˆz 2 λmλk E χ3 k ˆz 3 E χk ˆz Dˆz 3 E [χmχk ˆz ] λ2 m E χ3 m ˆz 3 E χm ˆz Dˆz 3 E χ2 m ˆz . In the asymptotic regime with dominant eigenvalue λ1, we get the dynamics: dt = ηλ1γ1 X λ1λmϵ3E ˆz 3 λ2 1 E ˆz 3 ηλ4 1γ1 E ˆz 3 dt = ηλmγm X λmλkϵ3E ˆz 3 λ2 mϵ3E ˆz 3 ηλ2 mγmλ2 1 E ˆz 3 Thus, all eigenmodes diverge because γm dˆzm Similarly, we find divergent dynamics when starting in the near-uniform regime: dt = ηλmγm E χ2 ˆz 2 k =m λk (λk λm) + ηλ2 mγm E χ3 ˆz 3 ηλmγm E χ ˆz ηλ2 mγm E χ3 ˆz 3 selecting the terms with the highest power in the eigenvalues. Thus, omission of stop-grad precludes successful representation learning for both the Euclidean and the cosine loss, but due to different mechanisms. Euclidean loss yields collapse, whereas the cosine loss succumbs to run-away activity. C.3 Removing the predictor from the Euclidean loss Leuc To analyze the representational dynamics in the absence of the predictor network, we consider Leuc no Pred: Leuc no Pred = 1 2 z(1) SG(z(2)) 2 m |ˆz(1) m SG(ˆz(2) m )|2 . The dynamics resulting from this loss function are a special case of the dynamics derived in Theorem 1 with all the eigenvalues equal to one (λk = 1). In particular Eq. (8) becomes: dˆz(1) m dt = η Leuc no Pred ˆz(1) m (t) = η ˆz(2) m ˆz(1) m , which evaluates to 0 under expectation over augmentations. Hence there is no learning without the predictor. C.4 Removing the predictor from the cosine loss Lcos Similarly, when we remove the predictor from Lcos it yields: Lcos no Pred = ˆz(1) m SG(ˆz(2) m ) ˆz(1) SG(ˆz(2)) , so that Eq. (10) becomes: dˆz(1) m dt = η ˆz(1) 3 ˆz(2) (ˆz(1) k )2ˆz(2) m ˆz(1) m ˆz(1) k ˆz(2) k . (24) Here, the near-uniform approximation (Eq. (11)) of ignoring the differences in χ between different eigenmodes is not valid. This is because the λ-terms are no longer present, and the effects of the χ-terms on the dynamics cannot be treated as negligible. In particular, setting WP = I in the dynamics derived in Theorem 2 would yield zero dynamics. However taking the expectation of Eq. (24) over augmentations yields the non-zero dynamics: dˆzm E ˆz2 k ˆz 3 which consists of terms of order O 1 ˆzm and O(1) in ˆzm, hinting at slower dynamics compared to Eq. (10). To study these granular effects, we would need to explicitly model the effect of the augmentations, for which we do not have a good statistical model. In lieu of deriving these dynamics analytically, we make an observation which restricts the possible dynamical behavior. Specifically, the sum of the eigenvalues remains constant throughout training. To show this, we begin by writing out the expression for the derivative over time of the sum of all the eigenvalues: d dt d dt Edata ˆz2 m We can derive the term inside the expectation by adding up the dynamics given by Eq. (25) for all the different eigenmodes: ˆz(1) m dˆz(1) m dt = η ˆz(1) 3 ˆz(2) (ˆz(1) k )2ˆz(1) m ˆz(2) m (ˆz(1) m )2ˆz(1) k ˆz(2) k m ˆz(1) m dˆz(1) m dt = η ˆz(1) 3 ˆz(2) (ˆz(1) k )2ˆz(1) m ˆz(2) m (ˆz(1) m )2ˆz(1) k ˆz(2) k m ˆz(1) m dˆz(1) m dt proving that d m λm = 0, i.e, the sum of the eigenvalues is conserved. This precludes collapsing dynamics where all eigenvalues go to zero as well as diverging dynamics where at least one eigenvalue diverges. C.5 Isotropic losses for equalized convergence rates In Expressions (9) and (11) we see that the overall learning dynamics have a quadratic dependence on the eigenvalues with a root near collapsed solutions, which causes these modes to learn slower. We reasoned that this anisotropy could be detrimental for learning. To address this issue, we sought to derive alternative loss functions that encourage isotropic learning dynamics for all modes. C.5.1 Euclidean Iso Loss. We start by deriving an Iso Loss function for the Euclidean case Leuc. To avoid the unwanted quadratic dependence, we first note that we would like to arrive at the following expression for the dynamics: dt = η (1 λm) ˆzm . By recalling the Euclidean loss and corresponding dynamics: m |λmˆz(1) m SG(ˆz(2) m )|2 dˆzm dt = ηλm (1 λm) ˆzm , we note that the leading λm term has no influence on the overall sign of the dynamics, and is introduced by the second step in the chain rule: ˆz(1) m = (λmˆz(1) ˆz(2)) ˆz(1) m (λmˆz(1) ˆz(2)) . Based on this realization we see that this second step needs to be modified. To that end, we start with the desired derivative: Leuc iso ˆz(1) m = (λmˆz(1) ˆz(2)) ˆz(1) m (ˆz(1) ˆz(2)) , and see that several loss functions are possible. The one we have reported in Eq. (15) we derived by applying an appropriate stop-grad while integrating: Leuc iso ˆz(1) m = (ˆz(1) m + λmˆz(1) ˆz(2) ˆz(1) m ) ˆz(1) m (ˆz(1) ˆz(2)) . Leuc iso = 1 m |ˆz(1) m SG(ˆz(2) m + ˆz(1) m λmˆz(1) m )|2 Another alternative loss with the same desired isotropic learning dynamics, but using a different placement of the stop-gradient operators, is given by: m SG λmˆz(1) m ˆz(2) m ˆz(1) m SG(ˆz(2) m ) C.5.2 Cosine Similarity Iso Loss. Since most practical SSL approaches rely on cosine similarity, which suffers from a similar anisotropy of the learning dynamics, we sought to find Iso Losses in this setting. With the same goal as above, we would like to arrive at the dynamics: dt = η ˆz(2) m Dˆz(1) ˆz(2) η P k λkˆz(1) k ˆz(2) k Dˆz(1) 3 ˆz(2) λmˆz(1) m starting from the cosine loss and corresponding dynamics: λmˆz(1) m SG(ˆz(2) m ) Dˆz(1) SG(ˆz(2)) (26) dt = η λmˆz(2) m Dˆz(1) ˆz(2) η P k λkˆz(1) k ˆz(2) k Dˆz(1) 3 ˆz(2) λ2 mˆz(1) m . (27) The Iso Loss in this case can be derived by noting how λm arises in each of the two terms in Eq. (27), and engineering an alternative loss function corresponding to each term separately. In the first term, λm arises from the partial derivative of the numerator λmˆz(1) m SG(ˆz(2) m ) in the original loss (Eq. (26)). This can be remediated by using ˆz(1) m SG(ˆz(2) m ) as the numerator instead. In the second term in Eq. (27), λ2 m arises from the partial derivative of Dˆz(1) = q P k(λkˆz(1) m )2 in the denominator. We can reduce λ2 m to λm by instead taking the partial derivative of D1/2ˆz(1) = q P k(λ1/2 k ˆz(1) m )2. Putting these insights together, we arrive at the desired partial derivative: Lcos iso ˆz(1) m = 1 Dˆz(1) ˆz(2) ˆz(1) m ˆz(2) m ˆz(1) m + P k λkˆz(1) k ˆz(2) k Dˆz(1) 3 ˆz(2) 1 2 λm(ˆz(1) m )2 Dˆz(1) ˆz(2) (ˆz(1)) ˆz(2) ˆz(1) m + P k λkˆz(1) k ˆz(2) k Dˆz(1) 3 ˆz(2) 1 2 D1/2ˆz(1) 2 and the integrated Iso Loss in eigenspace: Lcos iso = (ˆz(1)) SG Dˆz(1) ˆz(2) (Dˆz(1)) ˆz(2) Dˆz(1) 3 ˆz(2) D1/2ˆz(1) 2 . Rotating all terms back to the original space gives the desired Iso Loss for cosine similarity as reported (Eq. (17)): Liso = (z(1)) SG z(2) WPz(1) z(2) 2SG (WPz(1)) z(2) WPz(1) 3 z(2) W 1/2 P z(1) 2 . D Experimental methods Self-supervised pretraining. We used the CIFAR-10, CIFAR-100 [3], STL-10 [4], and Tiny Image Net [5] datasets for self-supervised learning with a Res Net-18 [6] encoder and the Sim CLR set of transformations [7]. We also adopted several modifications of Res Net-18 which have been proposed to deal with the low resolution of the images in these datasets [7]. The Res Net modifications comprise using 3 3 convolutional kernels instead of 7 7 kernels and skipping the first max-pooling operation. The Solo-learn library [8] also provides specialized sets of augmentations that work well for these datasets, which we adopted as well. The configurations we used for each dataset are summarized in Table 4. We used Batch Norm in the backbone and the projector multi-layer perceptron (MLP) in the hidden layer for all methods. For BYOL, we included Batch Norm also in the hidden layer of the predictor MLP. We used a projection dimension of 256 for the projection MLP using one hidden layer with 4096 units, and the same architecture for the nonlinear predictor for the BYOL baseline. For networks using EMA target networks, we used the LARS optimizer with learning rate 1.0 whereas for networks without the EMA, we used stochastic gradient descent with momentum 0.9 and learning rate 0.1. Furthermore, we used a warmup period of 10 epochs for the learning rate followed by a cosine decay schedule and a batch size of 256. We also used a weight decay 4 10 4 for the closed-form predictor models and 10 5 for the nonlinear predictor models. For evaluation, we removed the projection MLP and used the embeddings at the pooled output of the Res Net convolutional layers following standard practice. For the EMA, we started with τbase = 0.99 and increased τEMA to 1 with a cosine schedule exactly following the configuration reported in [9]. For Direct Pred, we used α = 0.5 and τ = 0.998 for the moving average estimate of the correlation matrix updated at every step. CIFAR-10 CIFAR-100 STL-10 Tiny Image Net Resolution 32 32 32 32 96 96 64 64 Kernel size 3 3 3 3 7 7 3 3 First max-pool No No Yes Yes Blur No No Yes No Linear evaluation protocol. We reported the held-out classification accuracy on the test sets for CIFAR-10/100 and STL-10, and the validation set for Tiny Image Net, after online training of the gradient-isolated linear classifier on each labeled example in the training set during pretraining. Compute resources All simulations were run on an in-house cluster consisting of 5 nodes with 4 V100 NVIDIA GPUs each, one node with 4 A100 NVIDIA GPUs, and one node with 8 A40 NVIDIA GPUs. Runs on CIFAR-10/100 took about 8 hours each, and the STL-10 and Tiny Imagenet runs took about 24 hours each. [1] Yuandong Tian, Xinlei Chen, and Surya Ganguli. Understanding self-supervised learning dynamics without contrastive pairs. In International Conference on Machine Learning, pages 10268 10278. PMLR, 2021. [2] Jaehoon Lee, Lechao Xiao, Samuel Schoenholz, Yasaman Bahri, Roman Novak, Jascha Sohl-Dickstein, and Jeffrey Pennington. Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent. Advances in Neural Information Processing Systems, 32:8572 8583, 2019. URL https://proceedings.neurips.cc/paper/2019/hash/ 0d1a9651497a38d8b1c3871c84528bd4-Abstract.html. [3] Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images, 2009. [4] Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, pages 215 223. JMLR Workshop and Conference Proceedings, 2011. [5] Ya Le and Xuan Yang. Tiny imagenet visual recognition challenge. CS 231N, 7(7):3, 2015. [6] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. [7] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pages 1597 1607. PMLR, 2020. [8] Victor Guilherme Turrisi da Costa, Enrico Fini, Moin Nabi, Nicu Sebe, and Elisa Ricci. sololearn: A library of self-supervised methods for visual representation learning. Journal of Machine Learning Research, 23(56):1 6, 2022. URL http://jmlr.org/papers/v23/21-1155.html. [9] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, 33:21271 21284, 2020.