# unifying_selfsupervised_clustering_and_energybased_models__3872bdb3.pdf Published in Transactions on Machine Learning Research (08/2025) Unifying Self-Supervised Clustering and Energy-Based Models Emanuele Sansone emanuele.sansone@kuleuven.be Department of Electrical Engineering (ESAT) KU Leuven Robin Manhaeve robin.manhaeve@kuleuven.be Department of Computer Science KU Leuven Reviewed on Open Review: https: // openreview. net/ forum? id= NW0u Ke6IZa Self-supervised learning excels at learning representations from large amounts of data. At the same time, generative models offer the complementary property of learning information about the underlying data generation process. In this study, we aim at establishing a principled connection between these two paradigms and highlight the benefits of their complementarity. In particular, we perform an analysis of self-supervised learning objectives, elucidating the underlying probabilistic graphical models and presenting a standardized methodology for their derivation from first principles. The analysis suggests a natural means of integrating self-supervised learning with likelihood-based generative models. We instantiate this concept within the realm of cluster-based self-supervised learning and energy models, introducing a lower bound proven to reliably penalize the most important failure modes and unlocking full unification. Our theoretical findings are substantiated through experiments on synthetic and real-world data, including SVHN, CIFAR10, and CIFAR100, demonstrating that our objective function allows to jointly train a backbone network in a discriminative and generative fashion, consequently outperforming existing self-supervised learning strategies in terms of clustering, generation and out-of-distribution detection performance by a wide margin. We also demonstrate that the solution can be integrated into a neuro-symbolic framework to tackle a simple yet non-trivial instantiation of the symbol grounding problem. The code is publicly available at https://github.com/emsansone/GEDI.git. 1 Introduction Self-supervised learning (SSL) has achieved remarkable results in recent years thanks to its ability to learn high-quality representations from large amounts of unlabeled data (Balestriero et al., 2023). At the same time, generative models have provided valuable insights into the unknown generative processes underlying data. The synergy between these two distinct areas of machine learning holds great potential, for example by leveraging knowledge of the underlying generative process to learn more robust representations, or by using learned representations to synthesize new data. However, a principled theory and methodology that bridges self-supervised learning and generative modeling is missing. In this work, we take a significant step toward closing this gap by demonstrating, for the first time, the feasibility of learning a self-supervised clustering model in a generative manner. To achieve this, we address three key challenges: formulation, integration, and unification. The first challenge arises from the diverse range of SSL objectives proposed in recent years and the absence of a common perspective. We tackle this by providing a probabilistic interpretation of these objectives through their underlying probabilistic graphical models. This perspective reveals a common methodology Published in Transactions on Machine Learning Research (08/2025) for deriving existing objectives from first principles. The resulting probabilistic framework naturally paves the way to connect SSL with likelihood-based generative models, thereby addressing the second challenge of integration. The final challenge, unification, involves reconciling the variety of architectural heuristics commonly used to prevent trivial representations - heuristics that often hamper generative capabilities. We overcome this by introducing a novel objective, the GEnerative DIscriminative lower bound (GEDI). We prove that GEDI is guaranteed to avoid the triad of failure modes (Sansone, 2023), including representation collapse, cluster collapse, and the problem of label inconsistency with data augmentations. These guarantees eliminate the need for common architectural heuristics and enable training a neural network in both a generative (akin to energy-based models) and a discriminative manner (similar to SSL clustering). The proposed unified view ultimately leads to improved performance in terms of clustering, generation, and out-of-distribution detection compared to standard cluster-based SSL methods. We substantiate our theoretical findings through experiments conducted on both toy and real-world datasets. Specifically, our results demonstrate that GEDI can achieve a significant improvement in terms of clustering performance compared to state-of-the-art baselines on SVHN, CIFAR-10 and CIFAR-100. Additionally, in the context of generation performance, GEDI can effectively compete with existing energy-based solutions, whereas traditional SSL approaches fall short. Most importantly, we highlight that the generative nature of GEDI is a crucial aspect that enhances the model s ability to detect out-of-distribution data, surpassing purely discriminative baselines in this regard. Finally, we show that GEDI can be easily integrated into a neuro-symbolic framework like Deep Prob Log (Manhaeve et al., 2018) and leverage its clustering nature to learn higher quality symbolic representations when performing symbol grounding (Harnad, 1990; Barsalou, 1999; Manhaeve et al., 2018; 2021; Sansone & Manhaeve, 2022; 2023a;b; Marconato et al., 2023). The article is structured as follows: In 2, we provide a probabilistic interpretation of three classes of SSL approaches, namely contrastive, negative-free and cluster-based methods. In 3, we showcase the integration of SSL with likelihood-based generative models. In 4, we take a step further and provide an instantiation of the general framework, thus providing the first principled objective to unify cluster-based and energy-based models and that is guaranteed to avoid the triad of failure modes. In 5, we review related work on SSL and in 6 we discuss the experimental analysis. Finally, in 7, we conclude by discussing future research directions for SSL. 2 A Probabilistic Formulation of SSL Objectives Figure 1: Different function blocks and variables used throughout the analysis of existing SSL approaches. Let us introduce the random quantities used throughout this section to analyze self-supervised learning approaches. We use x Ω, where Ωis a compact subset of Rd, to identify observed data drawn independently from an unknown distribution p(x) and x Ωto identify observed data drawn independently from a stochastic augmentation strategy T (x |x). We use ξ Rh, and z Rl, to identify latent representations in a deep network architecture, see Fig. 1) for a visualization. Importantly, the first latent representation is obtained through an encoding function enc : Ω Rh, whereas the second representation is obtained through the composition of enc and a projector function proj : Rh Rl, namely g : Ω Rl. Finally, we introduce a categorical variable y to identify the final output of the deep network architecture. We focus our analysis on three main classes of self-supervised learning approaches: 1) contrastive, 2) negativefree and 3) cluster-based methods.1 For each, we construct the underlying probabilistic graphical model (cf. Figure 2). This allows us to derive their corresponding objective functions from first principles and 1We focus on purely encoder-based techniques and leave the extension to predictive models (e.g. masked autoencoders) for future work. Published in Transactions on Machine Learning Research (08/2025) (a) Contrastive (CT) (b) Negative-Free (NF) (c) Cluster-Based (CB) Figure 2: Probabilistic graphical models for the different classes of self-supervised learning approaches. White and grey nodes represent hidden and observed vectors/variables, respectively. Solid arrows define the generative process, whereas blue dashed arrows identify auxiliary posterior densities/distributions. Index i {1, . . . , n} is used to identify training samples and their corresponding representations. highlight that all objectives share a common structure. Further on, we will use these insights to integrate self-supervised learning with generative models. 2.1 Contrastive SSL We demonstrate that contrastive self-supervised learning can be modeled as an instance classification problem, and reveal that the learning objective can be decomposed into two main terms: an entropy and a discriminative term. For the simplicity of the exposition, we focus on one of the main contrastive objectives, Info NCE (den Oord et al., 2018). Same conclusions can be obtained for other objectives such as CPC (O. Henaff, 2020), Sim CLR (Chen et al., 2020), Proto CPC (Lee, 2022), KSCL (Xu et al., 2022) to name a few, as demonstrated in appendices E to H. In the following, we introduce the ground truth and model predictive distributions and demonstrate how contrastive learning encourages alignment between the two. Definition 1 (Constrastive SSL ground truth joint distribution). p(x1:n, y1:n) j=1 p(xj)δ(yj j) with yj {1, . . . , n} and δ a delta function. Definition 2 (Constrastive SSL model distribution). p(x1:n, y1:n; Θ) i=1 p(xi)p(yi|xi; Θ) p(yi|xi; Θ) esim(g(xi;θ),g(xyi;θ))/τ Pn k=1 esim(g(xi;θ),g(xk;θ))/τ with sim : Rl Rl R a similarity function, τ > 0 a temperature parameter and Θ = {θ, {xi}n i } a set of parameters. The ground truth distribution defines an underlying instance classifier, where each input xi is associated to a unique natural number identifier through the δ function. The following Lemma provides an alternative yet equivalent interpretation of contrastive learning (the proof is given in Appendix B). Lemma 1. Given Definitions 1 and 2, maximizing the Info NCE objective is equivalent to maximize the following log-likelihood lower bound: Ep(x1:n,y1:n){log p(x1:n, y1:n)} Hp(x1:n) | {z } Neg. entropy term i=1 log esim(g(xi;θ),g(xi;θ))/τ Pn k=1 esim(g(xi;θ),g(xk;θ))/τ | {z } discriminative SSL term LCT (Θ) Published in Transactions on Machine Learning Research (08/2025) Moreover, the maximization of this lower bound is equivalent to solve an instance classification problem. From the above Lemma, we observe that the contrastive learning objective involves a discriminative and a negative entropy term. Only the discriminative SSL term is usually optimized, due to the fact that the entropy terms does not depend on Θ. 2.2 Negative-Free SSL We demonstrate that negative-free SSL enforces two important properties, namely the decorrelation of features in z and the invariance w.r.t. data augmentation over ξ. Furthermore, we reveal that negative-free SSL can be decomposed into two main parts, similarly to contrastive SSL. We focus the analysis on a recent negative-free criterion, namely Cor Info Max (Ozsoy et al., 2022). Similar conclusions can be derived for other negative-free approaches, including Barlow Twins (Zbontar et al., 2021), Vic Reg (Bardes et al., 2022a;b) and W-MSE (Ermolov et al., 2021) (please refer to Appendix I for further details). In the following, we introduce the ground truth and model distributions and demonstrate how negative-free SSL encourages alignment of their covariance matrices. In effect, this drives the sample covariance matrix toward a diagonal form, thereby decorrelating the features. Definition 3 (Negative-free SSL ground-truth distribution). p(z, ξ1:n, x1:n, x 1:n) p(z) i=1 p(xi|z)p(ξi)p(x i|xi, ξi) with Gaussian priors p(z) = N(z|0, I), p(ξi) = N(ξi|0, γ 1I), and assume the following conditional independencies p(x i|xi, ξi) = T (x i|xi) and p(xi|z) = p(xi). Definition 4 (Negative-free SSL auxiliary model distributions). q(ξi|xi, x i; Θ) N(ξi|enc(xi; θ) enc(x i; θ), I) q(z|x1:n; Θ) N(z|0, Σ) with Σ Pn i=1(g(xi; θ) g)(g(xi; θ) g)T + βI, β > 0 chosen to ensure the positive definiteness of Σ and g = 1/n Pn i=1 g(xi; θ) and Θ = {θ}. We can state the following Lemma with proof provided in Appendix C. Lemma 2. Given Definitions 3 and 4, maximizing the Cor Info Max objective (Ozsoy et al., 2022) is equivalent to maximize the following log-likelihood lower bound: Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} Hp(x1:n) | {z } Neg. entropy term Ep(x1:n){KL(q(z|x1:n; Θ) p(z))} | {z } discriminative SSL term LNF (Θ) i=1 Ep(xi)T (x i|xi){KL(q(ξi|xi, x i; Θ) p(ξi))} | {z } Continuation of LNF (Θ) where the first Kullback-Leibler term KL(q(z|x1:n; Θ) p(z)) KL(N(z|0, Σ) N(z|0, I)), the second term KL(q(ξi|xi, x i; Θ) p(ξi)) KL(N(ξi|enc(xi; θ) enc(x i; θ), I) N(ξi|0, γ 1I)) and const being a constant for the optimization over Θ. In other words, the discriminative terms promote two properties. Indeed, the first term standardizes the sample covariance of the latent representation, thus decorrelating its feature elements, and the the second term promote the invariance to data augmentations. These two properties are common among other negativefree methods, as shown in Appendix I. Notably, the negative-entropy term in the log-likelihood lower bound is not influenced by the optimization. Published in Transactions on Machine Learning Research (08/2025) 2.3 Cluster-Based SSL We demonstrate that cluster-based SSL admits a probabilistic interpretation based on the graphical model in Fig. 2(c), thus revealing that the objective can be decomposed in a negative entropy and a discriminative term. The analysis focusses on a recent approach Sw AV (Caron et al., 2020), but is more generally applicable to other cluster-based approaches, including Deep Cluster (Caron et al., 2018) and Se LA (Asano et al., 2020) to name a few. In the following, we introduce the ground truth and model predictive distributions and demonstrate how contrastive learning encourages alignment between the two. Definition 5 (Cluster-Based SSL ground-truth joint distribution). p(x1:n, x 1:n, y1:n) i=1 p(xi)T (x i|xi)p(yi|xi; Θ) with yi {1, . . . , c} being a categorical variable to identify one of c clusters, p(yi|xi; Θ) = e UT :yi G:i/τ P y e UT :y G:i/τ , where U Rh c is the matrix2 of cluster centers, G = [g(x1; θ), . . . , g(xn; θ)] Rh n is a representation matrix and Θ = {θ, U} is the set of parameters. Definition 6 (Cluster-Based SSL auxiliary clustering distribution). q(y1:n|x 1:n) i=1 q(yi|x i) We can state the following Lemma (proof is provided in Appendix D) Lemma 3. Given Definitions 5 and 6, maximizing the Sw AV objective (Caron et al., 2020) is equivalent to maximize the following log-likelihood lower bound: Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} Hp(x1:n) | {z } Neg. entropy term i=1 Ep(xi)T (x i|xi) n Eq(yi|x i) log p(yi|xi; Θ) + Hq(yi|x i) o | {z } discriminative SSL term LCB(Θ) An important aspect of cluster-based SSL is that the categorical variables y1:n are unobserved. A naive maximization of LCB(Θ) can lead to obtain trivial solutions like the one corresponding to uniformative predictions, namely q(yi|x i) = pγ(yi|xi) = Uniform({1, . . . , c}) for all i = 1, . . . , n. Fortunately, the problem can be avoided and solved exactly using an alternating optimization strategy based on the Sinkhorn-Knopp algorithm, where the alternation occurs between the optimization over the auxiliary distribution and Θ. This is indeed the procedure used in several cluster-based SSL approaches, like Sw AV (Caron et al., 2020), Deep Cluster (Caron et al., 2018) and Se LA (Asano et al., 2020) to name a few. We will see later on that we can obtain a different lower bound for cluster-based SSL which doesn t require asymmetries, in the form of alternating optimization, stopping gradients or momentum encoders. 3 Integrating SSL and Likelihood-Based Generative Models In all three classes of SSL approaches (cf. Lemmas 3, 6 and 9), the expected data log-likelihood can be lower bounded by the sum of two contribution terms, namely a negative entropy Hp(x1:n) and a discriminative log-likelihood term, chosen from LCT (Θ), LCB(Θ) and LNF (Θ). A connection to generative models emerges by additionally lower bounding the negative entropy term, namely: Hp(x1:n) = Ep(x1:n){log p(x1:n)} 2We use subscripts to select rows and columns. For instance, U:y identify y th column of matrix U. Published in Transactions on Machine Learning Research (08/2025) i=1 Ep(xi){log p(xi)} Ep(xi){log pΨ(xi)} + KL(p(xi) pΨ(xi)) i=1 Ep(xi){log pΨ(xi)} | {z } CE(p, pΨ) where pΨ(x) is a generative model parameterized by Ψ. Notably, the relation in (1) can be substituted in any of the objectives previously derived for the different SSL classes, thus allowing to integrate generative and SSL models into a single Bayesian framework. It is important to mention that much can be gained by synergistically optimizing the resulting hybrid objectives. For instance, the recent work EBCLR (Kim & Ye, 2022) proposes a specific instantiation of this general idea, by combining energy-based models with contrastive SSL approaches and showcasing the improved discriminative performance. EBCLR can be obtained by observing that the conditional density in Definition 1 can be decomposed into joint and marginal densities (similarly to what is done in (Grathwohl et al., 2020) for a supervised classifier):3 p(y|x; Θ) esim(g(x),g(xy))/τ Pn k=1 esim(g(x),g(xk))/τ p(y, x; Θ) = esim(g(x),g(xy))/τ p(x; Θ) = Pn k=1 esim(g(x),g(xk))/τ Γ(Θ) = e E(x;Θ) where E(x, Θ) = log Pn j=1 esim(g(xℓj ),g(x))/τ defines the energy score of the marginal density. By choosing pΨ(x) = p(x; Θ) and sim(z, z ) = z z 2 in Eq. (2), one recovers the exact formulation of EBCLR (Kim & Ye, 2022). Moreover, this integrated framework is not bound to the specific family of energy-based models, rather one could in principle specify any likelihood-based model (e.g. VAE, diffusion model, normalizing flow) in Eq. (1), consequently giving rise to a wide spectrum of possible hybrid solutions. Indeed, some recent works have started to propose solution aiming towards the joint integration of VAEs and contrastive SSL (Gatopoulos & Tomczak, 2020; Zhue et al., 2020; Wu et al., 2023; Nakamura et al., 2023). Therefore, our work generalizes these specific views to a broader class of SSL approaches and likelihood based generative models. In the subsequent sections, we are going to demonstrate an instantiation of this general view based on SSL clustering and energy-based models. Indeed, we are going to provide a unified theory allowing to learn a backbone classifier network to jointly exploit its generative and discriminative properties. 4 Unifying Self-Supervised Clustering and Energy-Based Models Before demonstrating the unification, we provide an alternative lower bound for cluster-based SSL to the one obtained in Lemma 9, enabling to simplify the architectural design of the neural network, while ensuring the avoidance of the main failure modes. Then, we provide our new GEnerative and DIscriminative (GEDI) Lower Bound. 4.1 Lower Bound for Cluster-Based SSL We state the following proposition and leave the proof to Section F in the Supplementary Material. Definition 7. Consider the same conditions in Definition 5, except for choosing identical auxiliary and predictive distributions, namely q(y|x) p(y|x; Θ) Softmax(U T g(x)/τ). Note that the choice of identical auxiliary and predictive distributions is a key difference with respect to existing cluster-based SSL. For instance, Sw AV avoids learning trivial solutions when optimizing the lower bound in Lemma 9 by introducing asymmetric distributions. Specifically, it considers p(y|x; Θ) Softmax(U T g(x)/τ) and q(y|x ) Sinkhorn(Stop Grad(U T g(x )/τ)), where Sinkhorn and Stop Grad are two 3For the sake of notation simplicity, we omit the dependence on θ. Published in Transactions on Machine Learning Research (08/2025) operators performing the Sinkhorn-Knopp algorithm and stopping the gradients, respectively. Therefore, our choice constitutes a simplification of the architectural design, but this relaxation requires the development of a new objective to guaranteeing the avoidance of the main training failure modes. The following proposition provide an alternative lower bound for cluster-based SSL (the proof is given in Appendix J, whereas the analysis about failure modes is deferred to Section 4.3) Proposition 1. Given Definition 7, the expected data log-likelihood for the probabilistic graphical model in Fig. 2(c) can be alternatively lower bounded as follows: Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} Hp(x1:n) i=1 Ep(xi)T (x i|xi) {CE(p(yi|x i; Θ), p(yi|xi; Θ))} | {z } LINV (Θ) j=1 p(yj = yi|xj; Θ) | {z } LP RIOR(Θ) Additionally, the corresponding maximum value for the last two addends in Eq. (3) is given by the following inequality:4 LINV (Θ) + LP RIOR(Θ) Hp(y1:n) (4) The above proposition has interesting implications. First of all, by maximizing the discriminative term LINV (Θ) with respect to Θ, we enforce two properties, namely: (i) label invariance, as we ensure that the predictive distributions of the discriminative model for a sample and its augmented version match each other and (ii) confident predictions, as maximizing the cross-entropy forces also to decrease the entropy of these distributions.5 Secondly, by choosing a uniform prior, viz. p(yi) = Uniform({1, . . . , c}), and by maximizing LP RIOR(Θ) with respect to Θ, we ensure to obtain a balanced cluster assignment. This is also commonly done by approaches based on optimal transport objectives and corresponding surrogates, typically empployed in cluster-based SSL (Cuturi, 2013; Caron et al., 2018; 2020; Amrani et al., 2022). 4.2 GEDI Lower Bound For the GEDI instantiation, we derive the lower bound on the expected log-likelihood by exploiting the bound in Eq. (1) and the one in Proposition 1: Ep(x1:n){log p(x1:n; Θ)} LGEN(Ψ) | {z } Generative term CE(p, pΨ) + LINV (Θ) + LP RIOR(Θ) | {z } Discriminative terms additionally we decompose the discriminative model p(y|x; Θ) to obtain pΨ in a similar manner to what we have already done in Eq. (2), namely: p(y, x; Θ) = e U T :yg(x)/τ p(x; Ψ) = |{z} Ψ = Θ p(x; Θ) = Pc y=1 e U T :yg(x)/τ Γ(Θ) = e E(x;Θ) where E(x; Θ) = log Pc y=1 e U T :yg(x)/τ. We will shortly analyze the properties of the different objective terms in Eq. (5). For the moment, we finalize the GEDI instantiation by devising the corresponding training algorithm. 4Here, we assume that the predictive model p(y|x; Θ) has enough capacity to achieve the optimal solution. 5Indeed, recall that CE(p, q) = Hp + KL(p q). Therefore, maximizing CE(p, q) forces to have both KL(p q) = 0 and Hp = 0. Published in Transactions on Machine Learning Research (08/2025) Algorithm 1: GEDI Training. Input: x1:n, x 1:n, Iters, SGLD and Adam optimizer hyperparameters; Output: Trained model Θ; For iter = 1, . . . , Iters; Generate samples from pΘ using SGLD; Estimate 1Θ = ΘLGEN(Θ) using Eq. (7); Compute 2Θ = ΘLINV (Θ); Compute 3Θ = ΘLP RIOR(Θ) ; Θ P3 i=1 iΘ; Θ Adam maximizing using Θ; Return Θ; Learning a GEDI model. We can train the GEDI model by jointly maximizing the objective in Eq. (5) with respect to the parameters Θ through gradient-based strategies. The overall gradient includes the summation of three terms, viz. ΘCE(p, pΘ), ΘLINV (Θ) and ΘLP RIOR(Θ). While the last two gradient terms can be computed easily by leveraging automatic differentiation, the first one must be computed by exploiting the following identities (obtained by simply substituting Eq. (6) into the definition of cross-entropy and expanding ΘΓ(Θ)): ΘCE(p, pΘ) = y=1 e U T :yg(xi)/τ ) n Θ log Γ(Θ) y=1 e U T :yg(xi)/τ ) y=1 e U T :yg(x)/τ ) Importantly, the first and the second expectations in Eq. (7) are estimated using the training and the generated data, respectively. To generate data from pΘ, we use a sampler based on Stochastic Gradient Langevin Dynamics (SGLD), thus following recent best practices to train energy-based models (Xie et al., 2016; Nijkamp et al., 2019; Du & Mordatch, 2019; Nijkamp et al., 2020). The whole learning procedure is summarized in Algorithm 1. Computational requirements. When comparing our GEDI instantiation with traditional SSL training, more specifically to Sw AV, we observe two main differences in terms of computation. Firstly, our learning algorithm does not require to run the Sinkhorn-Knopp algorithm, thus saving computation. Secondly, our GEDI instantiation requires additional forward and backward passes to draw samples from the energy-based model pΘ. However, the number of additional passes through the discriminative model can be limited by the number of SGLD iterations, necessary to generate data (cf. Experiments). 4.3 Analysis of Loss Landscape and the Triad of Failure Modes Asymmetries have been playing an important role in self-supervised learning in order to avoid trivial solutions/failure modes (Wang et al., 2022). Here, we formalize three main failure modes for cluster-based SSL. Then, we study the GEDI loss landscape and show that these undesired trivial solutions are not admitted by our objective. This result holds without introducing asymmetries in the optimization procedure and/or network architecture. Let s start by defining the most important failure modes, namely: Definition 8 (Failure Mode 1 - Representational Collapse). There exists a constant vector k Rh such that for all x Rd, g(x) = k. Definition 9 (Failure Mode 2 - Cluster Collapse). There exists a cluster j {1, . . . , c} such that for all x Rd, p(y = j|x; Θ) = 1. Definition 10 (Failure Mode 3 - Label Inconsistency). For all possible permutations π : {1, . . . , c} {1, . . . , c}, a dataset D = {(xi, ti, t i)}n i=1, its permuted version Dπ = {(xi, tπ(i), t i)}n i=1 and a loss L(Θ; ), Published in Transactions on Machine Learning Research (08/2025) evaluated at one of the two datasets, we have that L(Θ; D) = L(Θ; Dπ). For GEDI, ti .= U T g(xi) and t i .= U T g(x i). In other words, Definition 8 considers the case where the encoder maps (collapses) every input to the same output. Definition 9 considers the situation where the predictive model assigns all samples to the same cluster with high confidence. And Definition 10 considers the case where a hypothetical adversary swaps the predictions made by the model on different pair of inputs. Ideally, we would like to have an objective that does not admit these failure modes. Now, we state the properties of the loss landscape of GEDI with the following theorem (we leave the proof to Appendix K): Theorem 1. Given definitions 8-10, the following statements tells for a particular loss, which failure modes are admissible solutions: a. LGEN(Θ) admits failure modes 2 and 3. b. LINV (Θ) admits failure modes 1 and 2. c. LP RIOR(Θ) admits failure modes 1 and 3. Proof. (Sketch) The proof proceeds by applying the definitions of the various failure families (Definitions 8 10), evaluating the loss terms in each case, and verifying whether their values correspond to global minima. Intuitively, this is feasible because each loss term defines a notion of distance between the ground truth and model distributions. Importantly, Theorem 1 tells us that LGEN(Θ) can be used to penalize representational collapse, LINV (Θ) can be used to ensure that cluster assignments are consistent with data augmentation, while LP RIOR(Θ) can be used to penalize cluster collapse. Consequently, by maximizing the objective in Eq. (5), we are guaranteed to learn solutions which are non-trivial. A table summarizing all these properties is given below. Table 1: Summary of loss landscape Does penalize ? Repr. collapse Clus. collapse Lab. Inconst. LGEN(Θ) Yes No No LINV (Θ) No No Yes LP RIOR(Θ) No Yes No Eq. (5) Yes Yes Yes 5 Related Work We organize the related work according to different objective categories, namely contrastive, cluster-based and non-contrastive self-supervised approaches. Additionally, we discuss recent theoretical results, augmentation strategies as well as connections to energy-based models. For an exhaustive overview of self-supervised learning, we invite the interested reader to check out two recent surveys (Jing & Tian, 2021; Balestriero et al., 2023). Contrastive objectives and connection to mutual information. Contrastive learning represents an important family of self-supervised learning algorithms, which is rooted in the maximization of mutual information between the data and its latent representation (Linsker, 1988; Becker & Hinton, 1992). Estimating and optimizing mutual information from samples is notoriously difficult (Mc Allester & Stratos, 2020), especially when dealing with high-dimensional data. Most recent approaches focus on devising variational lower bounds on mutual information (Barber & Agakov, 2004). Indeed, several popular objectives, like the mutual information neural estimation (MINE) (Belghazi et al., 2018), deep Info MAX (Hjelm et al., 2018), noise Published in Transactions on Machine Learning Research (08/2025) contrastive estimation (Info NCE) (den Oord et al., 2018; O. Henaff, 2020; Chen et al., 2020; Lee, 2022; Xu et al., 2022) to name a few, all belong to the family of variational lower bounds (Poole et al., 2019). All these estimators have different properties in terms of bias-variance trade-off (Tschannen et al., 2019; Song & Ermon, 2020). Our work model contrastive learning using an equivalent probabilistic graphical model and a corresponding objective function based on the data log-likelihood, thus providing an alternative view to the principle of mutual information maximization. This is similar in spirit to the formulations proposed in the recent works of (Mitrovic et al., 2021; Tomasev et al., 2022; Scherr et al., 2022; Xu et al., 2022). Unlike these works, we are able to extend the log-likelihood perspective to other families of self-supervised approaches and also to highlight and exploit their connections to energy-based models. Cluster-based objectives. There are also recent advances in using clustering techniques in representation learning. For example, Deep Cluster (Caron et al., 2018) uses k-means and the produced cluster assignments as pseudo-labels to learn the representation. The work in (Huang et al., 2022) introduces an additional regularizer for deep clustering, invariant to local perturbations applied to the augmented latent representations. The work in (Asano et al., 2020) shows that the pseudo-label assignment can be seen as an instance of the optimal transport problem. Sw AV (Caron et al., 2020) proposes to use the Sinkhorn-Knopp algorithm to optimize the optimal transport objective (Cuturi, 2013) and to perform a soft cluster assignment. Finally, contrastive clustering (Li et al., 2021) proposes to minimize the optimal transport objective in a contrastive setting, leveraging both positive and negative samples. Our work provides a simple yet concise formulation of cluster-based self-supervised learning based on the principle of likelihood maximization. Additionally, thanks to the connection with energy-based models, we can perform implicit density estimation and leverage the learnt information to improve the clustering performance. Negative-free objectives. It s important to mention that new self-supervised objectives have recently emerged (Zbontar et al., 2021; Grill et al., 2020) as a way to avoid using negative samples, which are typically required by variational bounds on the mutual information, namely BYOL (Grill et al., 2020), Sim Siam (Chen et al., 2020), DINO (Caron et al., 2021), Zero-CL (Zhang et al., 2021), W-MSE (Ermolov et al., 2021), Barlow Twins (Zbontar et al., 2021), VICReg (Bardes et al., 2022a) and its variants (Bardes et al., 2022b), MEC (Liu et al., 2022) and Cor Info Max (Ozsoy et al., 2022). DINO proposes to maximize a cross-entropy objective to match the probabilistic predictions from two augmented versions of the same image. BYOL, Sim Siam, W-MSE consider the cosine similarity between the embeddings obtained from the augmented pair of images. Additionally, W-MSE introduces a hard constraint implemented as a differentiable layer to whiten the embeddings. Similarly, Barlow Twins proposes a soft whitening by minimizing the Frobenius norm between the cross-correlation matrix of the embeddings and the identity matrix. Zero-CL pushes forward the idea of whitening the features by also including an instance decorrelation term. VICReg and its variant extend over Barlow Twins by computing the sample covariance matrix, instead of the correlation one (thus avoiding to use batch normalization), by enforcing an identity covariance and by introducing an additional regularizer term to minimize the mean squared error between the embeddings of the two networks and to promote the invariance of the embeddings. Similarly, the work in (Tomasev et al., 2022) uses an invariance loss function in conjunction to the contrastive Info NCE objective. MEC maximizes the log-determinant of the covariance matrix for the latent representation, thus promoting maximum entropy under the Gaussian assumption. Additionally, Cor Info Max extends MEC by introducing a term that enforces the representation to be invariant under data augmentation. In Section 2.3, we can cast the non-contrastive problem as a minimization of the Kullback Leibler divergence between the latent posterior and a standard normal density prior. In essence, our work allows to compactly represent the family of non-contrastive methods using a likelihood-based objective. Additional objectives. Several works have investigated the relation between different families of selfsupervised approaches leading to hybrid objective functions (Kügelgen et al., 2021; Garrido et al., 2023). In contrast, our work attempts to provide a unified view from a probabilistic perspective and to highlight/exploit its connection to energy-based models. The main idea of generative and discriminative training originally appeared in the context of Bayesian mixture models (Sansone et al., 2016). Specifically, the authors proposed to jointly learn a generative model and cluster data in each class in order to be able to discover subgroups in breast cancer data. Subsequently, the work in (Liu & Abbeel, 2020) pushed the idea forward and apply it to a supervised deep learning Published in Transactions on Machine Learning Research (08/2025) setting. Instead, our work focuses on self-supervised learning and generative models, thus avoiding the need for additional supervision on the categorical variable y. Recently, the work in (Li et al., 2022) uses the maximum coding rate criterion to jointly learn an embedding and cluster it. However, the training proceeds in a multistage fashion. In contrast, our work provides a simple formulation enabling to jointly learn and cluster the embeddings in one shot. Additionally, we can leverage the generative perspective to further boost the self-supervised learning performance, as demonstrated by the generation and out-of-distribution detection experiments. Theory of self-supervised learning. Several works have theoretically analysed self-supervised approaches, both for contrastive (Saunshi et al., 2019; Wang & Isola, 2020; Zimmermann et al., 2021; Tosh et al., 2021; Hao Chen et al., 2021; Saunshi et al., 2022) and non-contrastive methods (Tian et al., 2021; Kang-Jun et al., 2022; Weng et al., 2022; Wen & Li, 2022), to motivate the reasons for their successes, identify the main underlying principles and subsequently provide more principled/simplified solutions. Regarding the former family of approaches, researchers have (i) identified key properties, such as representation alignment (i.e. feature for positive pairs need to be close to each other) and uniformity (to avoid both representational and dimensional collapse) (Wang & Isola, 2020; Assran, Mahmoud and Caron, Mathilde and Misra, Ishan and Bojanowski, Piotr and Bordes, Florian and Vincent, Pascal and Joulin, Armand and Rabbat, Mike and Ballas, Nicolas, 2022; Assran et al., 2023), (ii) formulated and analyzed the problem using data augmentation graphs (Hao Chen et al., 2021) and (iii) examined generalization bounds on the downstream supervised performance (Saunshi et al., 2022; Bao et al., 2022). Regarding the latter family of approaches, the main focus has been devoted to understanding the reasons on why non-contrastive approaches avoid trivial solutions. In this regard, asymmetries, in the form of stop-gradient and diversified predictors, are sufficient to ensure well-behaved training dynamics (Tian et al., 2021; Weng et al., 2022; Wen & Li, 2022). Importantly, the asymmetries are shown to implicitly constrain the optimization during training towards solutions with decorrelated features (Kang-Jun et al., 2022). Recent works have also looked at identifying connections between contrastive and non-contrastive methods (Dubois et al., 2022; Garrido et al., 2023; Balestriero & Le Cun, 2022) to unify the two families. The work in (Dubois et al., 2022) proposes a set of desiderata for representation learning, including large dimensional representations, invariance to data augmentations and the use of at least one linear predictor to ensure good performance on linear probe evaluation tasks. The work in (Garrido et al., 2023) analyzes the relations between contrastive and non-contrastive methods, showing their similarities and differences. Both families of approaches learn in a contrastive manner. However, while contrastive solutions learn by contrasting between samples, non-contrastive ones focus on contrasting between the dimensions of the embeddings. The work in (Balestriero & Le Cun, 2022) studies the minima in terms of representations for the different loss functions proposed in the two families. The authors are able to show that such minima are equivalent to solutions achieved by spectral methods. This provides additional evidence on the similarities between the approaches and the possibility for their integration. Our work provides a unifying view of the different classes, allowing to derive several loss functions in a principled manner using variational inference on the data log-likelihood. Additionally, we provide conditions to learn in a principled manner, thus avoiding trivial solutions and the use of asymmetries. Generative models and self-supervised learning. Recently, works have considered synergies between self-supervised and energy-based models (Le Cun, 2022) for the purposes of out-of-distribution detection (Hendrycks et al., 2019; Winkens et al., 2020; Mohseni et al., 2020). This is a common characteristics of energy-based models and indeed our work highlight the explicit connection with self-supervised learning. To the best of our knowledge, there is only one recent work exploring the integration between self-supervised learning approaches and energy models (Kim & Ye, 2022). The authors propose to use an energy-based model to learn a joint distribution over the two augmented views for the same data. The resulting objective can be decomposed into a conditional distribution term, leading to a contrastive learning criterion, and a marginal distribution term, leading to an energy-based model criterion. Therefore, the work only considers the integration between contrastive methods and energy-based models. In contrast, our work goes a step forward by showing a general methodology to integrate generative and SSL approaches. Moreover, we provide a new lower bound for the class of cluster-based approaches that guarantees the avoidance of important failure modes. Published in Transactions on Machine Learning Research (08/2025) Table 2: Clustering performance in terms of normalized mutual information (NMI) on test set (moons and circles). Higher values indicate better clustering performance. Mean and standard deviations are computed from 5 different runs. Dataset JEM GEDI no prior GEDI no inv GEDI no gen GEDI Moons 0.00 0.00 0.00 0.00 0.11 0.15 0.98 0.00 0.94 0.07 Circles 0.00 0.00 0.00 0.00 0.22 0.13 0.83 0.12 1.00 0.01 There has been also recent interest in combining VAEs with self-supervised learning (Gatopoulos & Tomczak, 2020; Zhue et al., 2020; Wu et al., 2023; Nakamura et al., 2023; Tong et al., 2024). Our formulation is general enough to encompass ELBO-like objectives like the one used in VAEs. Instead of instantiating the generative term by leveraging an energy-based model, one could proceed to lower bound the entropy term following traditional ways to derive an ELBO. We are not interested to pursue this direction, as we aim at devising an objective function which can exploit a simple classifier architecture without the need of additional components, such as a decoder network. 6 Experiments We perform experiments to evaluate the generative/discriminative performance of GEDI and its competitors, namely an energy-based model JEM (Grathwohl et al., 2020), which is trained with persistent contrastive divergence Tieleman (2008) to optimize only the generative term of the GEDI Lower Bound (Eq. (5)), and 2 self-supervised baselines, viz. a negative-free approach based on Barlow Twins (Zbontar et al., 2021) and a cluster-based approach based on Sw AV (Caron et al., 2020), which optimize the objectives in Lemma 6 and Lemma 9, respectively. We also compare with a two-stage generative and discriminative solution originally proposed in (Sansone & Manhaeve, 2022), called GEDI two stage. The whole analysis is divided into four main experimental settings, the first part provides empirical validation and intuition on the results of Proposition 1 and Theorem 1, the second part based on two synthetic datasets, including moons and circles, compares GEDI against the above-mentioned baselines, the third part extends the comparisons to real-world data, including SVHN, CIFAR-10 and CIFAR-100, and finally the last part showcases the utility of the GEDI framework to mitigate the symbol grounding problem arising in neuro-symbolic learning on MNIST data (Harnad, 1990; Sansone & Manhaeve, 2023b; Marconato et al., 2023). We use existing code both as a basis to build our solution and also to run the experiments for the different baselines. In particular, we use the code from (Duvenaud et al., 2021) for training energy-based models and the repository from (da Costa et al., 2022) for all self-supervised approaches. Our code will be publicly released in its entirety upon acceptance. Implementation details as well as additional experiments are reported in the Supplementary Material. 6.1 Part 1: Empirical Validation of the Theory (c) no prior (i) no prior Figure 3: Qualitative visualization of the clustering performance for the different strategies on moons (a-f) and on circles (g-l) datasets. Colors identify different cluster predictions. Only GEDI and GEDI no gen are able to perform well on both datasets. Published in Transactions on Machine Learning Research (08/2025) (a) no prior (d) no prior Figure 4: Visualization of the main training losses. As suggested by Eq. 22 in Proposition 1, we can always debug whether the main failure modes in the triad have occurred by checking the minimum value achieved by the losses and compare it with its corresponding global value. Indeed, the global minimum value of each loss is known, that is ln(c) for LP RIOR and 0 for LINV . In both moons (a-c) and circles (d-f) the values are ln(2) 0.69 and 0, respectively. We consider two well-known synthetic datasets, namely moons and circles. We use a multi-layer perceptron (MLP) with two hidden layers (100 neurons each) for enc and one with a single hidden layer (4 neurons) for proj, we choose h = 2 and T (x |x) = N(0, σ2I) with σ = 0.03 as our data augmentation strategy. We train GEDI for 7k iterations using Adam optimizer with learning rate 1e 3. Similarly we ablate the contribution of the different loss terms by training the GEDI Lower Bound using only LGEN (equivalent to JEM), without LINV (called GEDI no inv) without LP RIOR (called GEDI no prior). Further details about the hyperparameters are available in the Supplementary Material (Section G). We evaluate the clustering performance both quantitatively, by using the Normalized Mutual Information (NMI) score and qualitatively, by visualizing the cluster assignments using different colors. From the results in Table 2 and Figure 3, we can make the following observations: (i) GEDI no prior and JEM are subject to cluster collapse on both datasets. This is expected as failure mode 2 is not penalized during training due to the omission of LP RIOR; (ii) GEDI no inv is subject to the problem of label inconsistency. Indeed, the obtained cluster labels are not informative and consistent with the underlying manifold structure of the data distribution. Again, this confirms the result of Theorem 1, as failure mode 3 could be avoided by the use of LINV ; (iii) GEDI no gen achieves competitive performance to GEDI despite the absence of LGEN. While in theory the objective function for this approach admits representation collapse, as predicted by our Theorem, in practice we never observed such issue. It might be the case that the learning dynamics of gradient-based optimisation are enough to avoid the convergence to this trivial solution. Finally (iv) GEDI is guaranteed to avoid the most important failure modes and therefore solve both tasks. We will see later the benefits of including also the generative term in the optimization. An important consequence of Proposition 1 is that we can use the training losses to debug whether we achieve the global minimum value and therefore avoid the triad of failure modes. Figure 4 showcases this property. 6.2 Part 2: Generative/Discriminative Comparisons on Synthetic Data We consider the same experimental setting used in part 1 and train JEM, Sw AV and GEDI for 7k iterations using Adam optimizer with learning rate 1e 3. Further details about the hyperparameters are available in the Supplementary Material (Section G). We evaluate the clustering performance both qualitatively, by visualizing the cluster assignments using different colors, and quantitatively, by using the Normalized Published in Transactions on Machine Learning Research (08/2025) Table 3: Clustering performance in terms of normalized mutual information (NMI) on test set (moons and circles). Higher values indicate better clustering performance. Mean and standard deviations are computed from 5 different runs. Dataset JEM Barlow Sw AV GEDI no gen GEDI Gain Moons 0.00 0.00 0.22 0.10 0.76 0.36 0.98 0.00 0.94 0.07 +0.22 Circles 0.00 0.00 0.13 0.10 0.00 0.00 0.83 0.12 1.00 0.01 +0.87 Figure 5: Qualitative visualization of the clustering performance for the different strategies on moons (a-e) and on circles (f-j) datasets. Colors identify different cluster predictions. Only GEDI and GEDI no gen are able to perform well on both datasets. Qualitative visualization of the generative performance for the different strategies on moons (k-o) and on circles (p-t) datasets. Colors identify different cluster predictions. All GEDI approaches (except for no gen perform comparably well to the generative solution JEM). Mutual Information (NMI) score. Additionally, we qualitatively demonstrate the generative performance of all approaches. We report all quantitative performance in Table 3. As expected, Barlow fails to solve both tasks due to the different nature to cluster-based approaches. Sw AV correctly solves the task for the moons dataset, while it faces difficulty to solve the one based on circles. This is due to the fact that the optimized bound in Eq. (17) includes several trivial solutions. The introduced asymmetries (such as stop gradient and the clustering layers) are not enough to rule them out. This is confirmed by visually inspecting the latent representation and observing that the encoder collapses to the identity function. Therefore, the projector head can solve the task by simply learning a linear separator. As a consequence, we observe that half of the labels are permuted across the two manifolds (refer to Figure 5). In contrast, GEDI can recover the true clusters in both datasets without any additional asymmetry and have a guarantee to avoid the triad of failure modes. Moreover, GEDI is able to learn a correct density estimator compared to the SSL baselines (cf. Figure 5). This is an important property that can come in handy for out-of-distribution detection, as we will see later in the real-world experiments. Published in Transactions on Machine Learning Research (08/2025) Table 4: Generative and discriminative performance on test set (SVHN, CIFAR-10, CIFAR-100). Normalized mutual information (NMI) and Frechet Inception Distance (FID) are used as evaluation metrics for the discriminative and generative tasks, respectively. Higher values of NMI and lower values of FID indicate better performance. Mean and standard deviations are reported for 5 different initialization seeds. Task Method SVHN CIFAR-10 CIFAR-100 Discriminative JEM 0.00 0.00 0.00 0.00 0.00 0.00 Barlow 0.20 0.02 0.17 0.04 0.61 0.05 Sw AV 0.21 0.01 0.44 0.01 0.51 0.21 (NMI) GEDI no gen 0.27 0.04 0.45 0.00 0.87 0.00 GEDI 0.25 0.04 0.44 0.01 0.87 0.00 JEM 201 36 223 15 271 85 Barlow 334 28 382 22 403 28 Sw AV 480 67 410 31 420 26 (FID) GEDI no gen 488 43 403 9 435 11 GEDI 193 10 214 13 226 9 Table 5: OOD detection in terms of AUROC on test set (CIFAR-10, CIFAR-100). Training is performed on SVHN. Dataset JEM Barlow Sw AV GEDI no gen GEDI CIFAR-10 0.73 0.17 0.26 0.1 0.80 CIFAR-100 0.72 0.24 0.32 0.15 0.80 6.3 Part 3: Real-world Experiments on SVHN, CIFAR-10, CIFAR-100 We consider three well-known computer vision benchmarks, namely SVHN, CIFAR-10 and CIFAR-100. We use a simple 8-layer Resnet network for the backbone encoder for both SVHN and CIFAR-10 (around 1M parameters) and increase the hidden layer size for CIFAR-100 (around 4.1M parameters) as from (Duvenaud et al., 2021). We use a MLP with a single hidden layer for proj (the number of hidden neurons is double the number of inputs), we choose h = 256 for CIFAR-100 and h = 128 for all other cases. Additionally, we use data augmentation strategies commonly used in the SSL literature, including color jitter, and gray scale conversion to name a few. We train JEM, Barlow, Sw AV, GEDI no gen and GEDI using Adam optimizer with learning rate 1e 4 and batch size 64 for 20, 200 and 200 epochs for each respective dataset (SVHN, CIFAR-10 AND CIFAR-100). Further details about the hyperparameters are available in the Supplementary Material (Section I). Similarly to the toy experiments, we evaluate the clustering performance by using the Normalized Mutual Information (NMI) score. Additionally, we evaluate the generative performance qualitatively using (c) GEDI no gen Figure 6: Samples generated by the different discriminative strategies on CIFAR-10 running Stochastic Langevin Dynamics for 500 iterations. Published in Transactions on Machine Learning Research (08/2025) Figure 7: The full model for the neurosymbolic experiment using GEDI. the Frechet Inception Distance (Heusel et al., 2017) as well as the OOD detection capabilities following the methodology in (Grathwohl et al., 2020). From Table 4, we observe that GEDI is able to outperform all other competitors by a large margin, thanks to the properties of both generative and self-supervised models. We observe that the difference gap in clustering performance increases with CIFAR100. This is due to a larger size of the backbone used in the CIFAR100 experiments (cf. the size of the latent representation). In terms of generation performance, GEDI is the only approach that compares favorably with JEM. We provide a qualitative set of samples generated by the different discriminative models in Figure 6. Last but not least, we investigate the OOD detection capabilities of the different methods. Table 11 provides a quantitative summary of the performance for a subset of experiments (the complete set is available in Section J). We observe that GEDI is more robust compared to other discriminative baselines, thanks to its generative nature. Overall, these experiments provide real-world evidence on the benefits of the proposed unification and theoretical results. 6.4 Part 4: Tackling A Non-Trivial Instantiation of the Symbol Grounding Problem For the final task, we consider applying the self-supervised learning approach to a neuro-symbolic (Ne Sy) setting. For this, we borrow an experiment from Deep Prob Log (Manhaeve et al., 2018; 2021). In this task, each example consists of a three MNIST images such that the value of the last one is the sum of the first two, e.g. . This can thus be considered a minimal neuro-symbolic tasks, as it requires a minimal reasoning task (a single addition) on top of the image classification task. This task only contains positive examples. We use the neural network from Manhaeve et al. (2018) as the neural backbone. We optimize LNESY instead of LINV , as the Ne Sy loss also forces a clustering of the digits, but is more informed. To calculate LNESY , we use GEDI to classify the images to produce a probability distribution over the classes 0 to 9, and we use the inference mechanism from Deep Prob Log to calculate the probability that the sum holds as shown in Figure 7. For this setting, this coincides with the Semantic Loss function (Xu et al., 2018). The Ne Sy constraint loss is defined as: LNESY (Θ) = i=1 CE (1, p(c|xi; Θ)) where xi is now an arbitrary number of images in each data point. For this experiment, these are the three MNIST images xi = (x1 i , x2 i , x3 i ). The probability of the constraint p(c|xi; Θ) is defined in terms of a set of rules R and a set of probabilistic facts F = {f1 . . . fn} with probabilities p(fj|xi; Θ). For this setting, there are 30 probabilistic facts, one for each possible classifications of each digit, with their probabilities is thus parameterized by the neural network. The probability of the facts is thus p(yi = 0|x1 i ; Θ), . . . , p(yi = 9|x1 i ; Θ), p(yi = 0|x2 i ; Θ), . . . , p(yi = 9|x3 i ; Θ) Published in Transactions on Machine Learning Research (08/2025) Each subset F F defines a possible world w F = {a | R F |= a}, i.e. the set of all atoms a entailed by the rules and the facts in F. In this setting, this is a total assignment of classes to digits, and the truth value of the constraint. The probability of the constraint is the sum of the probability of each possible world (i.e. classifications of the digits) in which the constraint holds. P(ci|xi; Θ) = X F | c w F P(w F |xi; Θ) (8) Where the probability of the possible world is defined as the product of the probability of each fact that is true in this possible word, and one minus the probability of each fact that does not hold in the possible world. P(w F |xi; Θ) = Y fj F p(fj|xi; Θ) Y fj F\F (1 p(fj|xi; Θ)) (9) It is a specifically interesting use case for representation learning, since when only the constraint probability is optimized, the neural network tends to collapse onto the trivial solution of classifying each digit as a 0, as shown in (Manhaeve et al., 2018; 2021; Sansone & Manhaeve, 2023a;b). This is a logically correct solution, but an undesirable solution. Optimizing the uniformity objective should prevent this collapse. A neural network should be able to correctly classify MNIST digits by using the SSL training objective and the logical constraint. Here, the focus is on the small data regime, and see whether the logical constraint is able to provide additional information. We also look at the performance of Sw AV when combined with the logical constraint. Further details about the hyperparameters and experimental setup are available in Appendix M. We evaluate the model by measuring the accuracy and NMI of the model on the MNIST test dataset for different numbers of training examples. The results are shown in Table 6. Here, # Ex. examples indicates the number of addition examples, which each have 3 MNIST digits. The results show that, without the Ne Sy constraint, the mean accuracy is low for all settings. The NMI is higher, however, and increases as there is more data available. This is expected, since the model is able to learn how to cluster from the data, but unable to learn how to classify. By including the constraint loss, the accuracy improves, as the model now has information on which cluster belongs to which class. Furthermore, it also has a positive effect on the NMI, as we have additional information on the clustering which is used by the model. Combining Sw AV with the logical constraint severely underperforms. It does not sufficiently enforce the uniformity of the predicted labels, which leads to the a local optimum where the predictions are distributed over 2-3 classes. These results show us that the proposed method is beneficial to learn to correctly recognize MNIST images using only a weakly-supervised constraint. Furthermore, we show that the proposed method can leverage the constraint to further improve NMI and classification accuracy. Table 6: The median and standard deviation of the accuracy and NMI of GEDI and Sw AV on the MNIST test set after training on the addition dataset. We also include results for learning without GEDI (only optimizing the constraint), and training GEDI without the Ne Sy constraint. No GEDI Only GEDI Sw AV & Ne Sy GEDI & Ne Sy # Ex. Acc. NMI Acc. NMI Acc. NMI Acc. NMI 100 .10 .00 .00 .00 .08 .01 .22 .03 .14 .00 .19 .00 .42 .02 .35 .01 1000 .10 .00 .00 .00 .09 .03 .29 .06 .10 .00 .36 .00 .96 .01 .91 .01 10000 .10 .00 .00 .00 .11 .05 .44 .10 .10 .00 .40 .00 .97 .00 .93 .00 7 Conclusions and Future Research We have presented a unified perspective on self-supervised clustering and energy-based models. The corresponding GEDI lower bound is guaranteed to prevent three main failure modes: representation collapse, cluster collapse, and label inconsistency of cluster assignments. By shedding new light on the synergies Published in Transactions on Machine Learning Research (08/2025) between self-supervised learning and likelihood-based generative models, we aim to inspire subsequent studies proposing new implementations of our general methodology, such as novel approaches to integrating self-supervised learning and latent variable models. Additionally, we demonstrate that GEDI can better capture the underlying data manifolds and provide more precise predictions compared to existing strategies. We believe that further progress can be achieved by combining self-supervised learning with other areas of mathematics, such as topology and differential geometry, for instance to guarantee a notion of connectedness when dealing with the manifold assumption. Moreover, we show that GEDI can be easily integrated into existing statistical relational reasoning frameworks, opening the door to new neuro-symbolic integrations and enabling the handling of low data regimes, which are currently beyond the reach of existing self-supervised learning solutions. There are several area of improvements and extension that we will target. First and foremost, we plan in the near future to apply the existing solution to larger scale datasets, such as Imagenet. Secondly, the current solution assumes that information about the number of classes and the class prior is known in advance. This might limit the applicability of the proposed solution to real-world cases, such as open-world settings or scenarios with long tailed distributions. It will be certainly interesting to generalize the framework to deal with such new settings. Finally, it will be possible to extend the proposed framework towards more object-centric representation learning, thus going beyond the traditional object classification problem. 8 Author Contributions and Acknowledgements The authors would like to thank Michael Puthawala for initial discussion on GEDI. ES had the idea, developed the theory around the three key challenges, implemented the solution for the toy and real cases and wrote the paper. RM implemented and wrote the experiments for the neuro-symbolic setting and also helped by proof-reading the article. This work received funding from the Horizon Europe research and innovation programme (MSCA-GF grant agreement n 101149800) and also from the Flemish Government under the Onderzoeksprogramma Artificiele Intelligentie (AI) Vlaanderen programme. The computational resources and services used in this work were provided by the computing infrastructure in the Electrical Engineering Department (PSI group) and the Department of Computer Science (DTAI group) at KU Leuven. A. Alemi, B. Poole, I. Fischer, J. Dillon, R. A. Saurous, and K. Murphy. Fixing a Broken ELBO. In ICML, 2018. E. Amrani, L. Karlinsky, and A. Bronstein. Self-Supervised Classification Network. In ECCV, 2022. Y. M. Asano, C. Rupprecht, and A. Vedaldi. Self-Labelling via Simultaneous Clustering and Representation Learning. In ICLR, 2020. Mido Assran, Randall Balestriero, Quentin Duval, Florian Bordes, Ishan Misra, Piotr Bojanowski, Pascal Vincent, Michael Rabbat, and Nicolas Ballas. The Hidden Uniform Cluster Prior in Self-Supervised Learning. In ICLR 2023, 2023. Assran, Mahmoud and Caron, Mathilde and Misra, Ishan and Bojanowski, Piotr and Bordes, Florian and Vincent, Pascal and Joulin, Armand and Rabbat, Mike and Ballas, Nicolas. Masked Siamese Networks for Label-Efficient Learning. In ECCV, 2022. R. Balestriero and Y. Le Cun. Contrastive and Non-Contrastive Self-Supervised Learning Recover Global and Local Spectral Embedding Methods. In Neur IPS, 2022. R. Balestriero, M. Ibrahim, V. Sobal, A. Morcos, S. Shekhar, T. Goldstein, F. Bordes, A. Bardes, G. Mialon, Y. Tian, et al. A Cookbook of Self-Supervised Learning. ar Xiv, 2023. H. Bao, Y. Nagano, and K. Nozawa. On the Surrogate Gap between Contrastive and Supervised Losses. In ICML, 2022. Published in Transactions on Machine Learning Research (08/2025) D. Barber and F. Agakov. The IM Algorithm: A Variational Approach to Information Maximization. In Neur IPS, 2004. A. Bardes, J. Ponce, and Y. Lecun. VICReg: Variance-Invariance-Covariance Regularization for Self Supervised Learning. In ICLR, 2022a. A. Bardes, J. Ponce, and Y. Le Cun. VICREGL: Self-Supervised Learning of Local Visual Features. In Neur IPS, 2022b. L. W. Barsalou. Perceptual symbol systems. Behavioral and Brain Sciences, 1999. S. Becker and G. Hinton. Self-Organizing Neural Network that Discovers Surfaces in Random-Dot Stereograms. Nature, 355(6356):161 163, 1992. M. I. Belghazi, A. Baratin, S. Rajeshwar, S. Ozair, Y. Bengio, A. Courville, and D. Hjelm. Mutual Information Neural Estimation. In ICML, 2018. M. Caron, P. Bojanowski, A. Joulin, and M. Douze. Deep Clustering for Unsupervised Learning of Visual Features. In ECCV, 2018. M. Caron, I. Misra, J. Mairal, P. Goyal, P. Bojanowski, and A. Joulin. Unsupervised Learning of Visual Features by Contrasting Cluster Assignments. In Neur IPS, 2020. M. Caron, H. Touvron, I. Misra, H. Jégou, J. Mairal, P. Bojanowski, and A. Joulin. Emerging Properties in Self-Supervised Vision Transformers. In ICCV, 2021. T. Chen, S. Kornblith, M. Norouzi, and G. Hinton. A Simple Framework for Contrastive Learning of Visual Representations. In ICML, 2020. M. Cuturi. Sinkhorn Distances: Lightspeed Computation of Optimal Transport. In Neur IPS, 2013. V. G. T. da Costa, E. Fini, M. Nabi, N Sebe, and E. Ricci. Solo-learn: A Library of Self-supervised Methods for Visual Representation Learning. JMLR, 23(56):1 6, 2022. A. Van den Oord, Y. Li, and O. Vinyals. Representation Learning with Contrastive Predictive Coding. In ar Xiv, 2018. Y. Du and I. Mordatch. Implicit Generation and Generalization in Energy-Based Models. ar Xiv, 2019. Y. Dubois, T. Hashimoto, S. Ermon, and P. Liang. Improving Self-Supervised Learning by Characterizing Idealized Representations. In Neur IPS, 2022. D. Duvenaud, J. Kelly, K. Swersky, M. Hashemi, M. Norouzi, and W. Grathwohl. No MCMC for Me: Amortized Samplers for Fast and Stable Training of Energy-Based Models. In ICLR, 2021. A. Ermolov, A. Siarohin, E. Sangineto, and N. Sebe. Whitening for Self-Supervised Representation Learning. In ICML, 2021. Q. Garrido, Y. Chen, A. Bardes, L. Najman, and Y. Lecun. On the Duality Between Contrastive and Non-Contrastive Self-Supervised Learning. In ICLR, 2023. I. Gatopoulos and J. M. Tomczak. Self-Supervised Variational Auto-Encoders. In Symposium on Advances in Approximate Bayesian Inference, 2020. W. Grathwohl, K.-C. Wang, J.-H. Jacobsen, D. Duvenaud, M. Norouzi, and K. Swersky. Your Classifier is Secretly an Energy Based Model and You Should Treat It Like One. In ICLR, 2020. J. B. Grill, F. Strub, F. Altché, C. Tallec, P. H. Richemond, E. Buchatskaya, C. Doersch, B. A. Pires, Z. D. Guo, M. G. Azar, et al. Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning. In Neur IPS, 2020. Published in Transactions on Machine Learning Research (08/2025) J. Z. Hao Chen, C. Wei, A. Gaidon, and T. Ma. Provable Guarantees for Self-Supervised Deep Learning with Spectral Contrastive Loss. In Neur IPS, 2021. Stevan Harnad. The Symbol Grounding Problem. Physica D: Nonlinear Phenomena, 1990. D. Hendrycks, M. Mazeika, S. Kadavath, and D. Song. Using Self-Supervised Learning Can Improve Model Robustness and Uncertainty. In Neur IPS, 2019. M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. In Neur IPS, 2017. R. D. Hjelm, A. Fedorov, S. Lavoie-Marchildon, K. Grewal, P. Bachman, A. Trischler, and Y. Bengio. Learning Deep Representations by Mutual Information Estimation and Maximization. In ICLR, 2018. Z. Huang, J. Chen, J. Zhang, and H. Shan. Learning Representation for Clustering Via Prototype Scattering and Positive Sampling. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022. L. Jing and Y. Tian. Self-Supervised Visual Feature Learning with Deep Neural Networks: A Survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021. L. Kang-Jun, S. Masanori, and O. Takayuki. Bridging the Gap from Asymmetry Tricks to Decorrelation Principles in Non-contrastive Self-Supervised Learning. In Neur IPS, 2022. B. Kim and J. C. Ye. Energy-Based Contrastive Learning of Visual Representations. In Neur IPS, 2022. J. Von Kügelgen, Y. Sharma, L. Gresele, W. Brendel, B. Schölkopf, M. Besserve, and F. Locatello. Self Supervised Learning with Data Augmentations Provably Isolates Content from Style. In Neur IPS, 2021. Y. Le Cun. A Path Towards Autonomous Machine Intelligence, 2022. K. Lee. Prototypical Contrastive Predictive Coding. In ICLR, 2022. Y. Li, P. Hu, Z. Liu, D. Peng, J. T. Zhou, and X. Peng. Contrastive Clustering. In AAAI, 2021. Z. Li, Y. Chen, Y. Le Cun, and F. T. Sommer. Neural Manifold Clustering and Embedding. In ar Xiv, 2022. R. Linsker. Self-Organization in a Perceptual Network. Computer, 21(3):105 117, 1988. H. Liu and P. Abbeel. Hybrid Discriminative-Generative Training via Contrastive Learning. In ar Xiv, 2020. X. Liu, Z. Wang, Y. Li, and S. Wang. Self-Supervised Learning via Maximum Entropy Coding. In Neur IPS, 2022. R. Manhaeve, S. Dumancic, A. Kimmig, T. Demeester, and L. De Raedt. Deep Prob Log: Neural Probabilistic Logic Programming. In Neur IPS, 2018. R. Manhaeve, S. Dumancic, A. Kimmig, T. Demeester, and L. De Raedt. Neural probabilistic logic programming in Deep Prob Log. Artifical Intelligence, 298:103504, 2021. E. Marconato, S. Teso, and A. Passerini. Neuro-Symbolic Reasoning Shortcuts: Mitigation Strategies and their Limitations. In ar Xiv, 2023. D. Mc Allester and K. Stratos. Formal Limitations on the Measurement of Mutual Information. In AISTATS, 2020. J. Mitrovic, B. Mc Williams, J. Walker, L. Buesing, and C. Blundell. Representation Learning via Invariant Causal Mechanisms. In ICLR, 2021. S. Mohseni, M. Pitale, J. B. S. Yadawa, and Z. Wang. Self-Supervised Learning for Generalizable Out-of Distribution Detection. In AAAI, 2020. Published in Transactions on Machine Learning Research (08/2025) H. Nakamura, M. Okada, and T. Taniguchi. Representation Uncertainty in Self-Supervised Learning as Variational Inference. In ICCV, 2023. E. Nalisnick, A. Matsukawa, Y. W. Teh, D. Gorur, and B. Lakshminarayanan. Do deep generative models know what they don t know? In ICLR, 2019. E. Nijkamp, M. Hill, S. C. Zhu, and Y. N. Wu. Learning non-convergent non-persistent short-run mcmc toward energy-based model. In Neur IPS, 2019. E. Nijkamp, M. Hill, T. Han, S.-C. Zhu, and Y. N. Wu. On the Anatomy of MCMC-Based Maximum Likelihood Learning of Energy-Based Models. In AAAI, 2020. Olivier O. Henaff. Data-Efficient Image Recognition with Contrastive Predictive Coding. In ICML, 2020. S. Ozsoy, S. Hamdan, S. Arik, D. Yuret, and A. T. Erdogan. Self-Supervised Learning with an Information Maximization Criterion. In Neur IPS, 2022. B. Poole, S. Ozair, A. Van Den Oord, A. Alemi, and G. Tucker. On Variational Bounds of Mutual Information. In ICML, 2019. E. Sansone. The Triad of Failure Modes and a Possible Way Out. In Neur IPS SSL Theory and Practice Workshop, 2023. E. Sansone and R. Manhaeve. GEDI: GEnerative and DIscriminative Training for Self-Supervised Learning. ar Xiv, 2022. E. Sansone and R. Manhaeve. Learning Symbolic Representations Through Joint GEnerative and DIscriminative Training. In ICLR Ne Sy-Ge Ms Workshop, 2023a. E. Sansone and R. Manhaeve. Learning Symbolic Representations Through Joint GEnerative and DIscriminative Training (Extended Abstract). In IJCAI KBCG Workshop, 2023b. E. Sansone, A. Passerini, and F. G.B. De Natale. Classtering: Joint Classification and Clustering with Mixture of Factor Analysers. In ECAI, 2016. N. Saunshi, O. Plevrakis, S. Arora, M. Khodak, and H. Khandeparkar. A Theoretical Analysis of Contrastive Unsupervised Representation Learning. In ICML, 2019. N. Saunshi, J. Ash, S. Goel, D. Misra, C. Zhang, S. Arora, S. Kakade, and A. Krishnamurthy. Understanding Contrastive Learning Requires Incorporating Inductive Biases. In ICML, 2022. F. Scherr, Q. Guo, and T. Moraitis. Self-Supervised Learning Through Efference Copies. In Neur IPS, 2022. J. Song and S. Ermon. Understanding the Limitations of Variational Mutual Information Estimators. ICLR, 2020. Y. Tian, X. Chen, and S. Ganguli. Understanding Self-Supervised Learning Dynamics Without Contrastive Pairs. In ICML, 2021. T. Tieleman. Training Restricted Boltzmann Machines Using Approximations to the Likelihood Gradient. In ICML, 2008. N. Tomasev, I. Bica, B. Mc Williams, L. Buesing, R. Pascanu, C. Blundell, and J. Mitrovic. Pushing the Limits of Self-Supervised Res Nets: Can We Outperform Supervised Learning Without Labels on Image Net? In ICML Pre-Training Workshop, 2022. S. Tong, X. Dai, Y. Chen, M. Li, Z. Li, B. Yi, Y. Le Cun, and Y. Ma. Unsupervised Learning of Structured Representations via Closed-Loop Transcription. In CPAL, 2024. C. Tosh, A. Krishnamurthy, and D. Hsu. Contrastive Estimation Reveals Topic Posterior Information to Linear Models. JMLR, 2021. Published in Transactions on Machine Learning Research (08/2025) M. Tschannen, J. Djolonga, P. K. Rubenstein, S. Gelly, and M. Lucic. On Mutual Information Maximization for Representation Learning. In ICLR, 2019. T. Wang and P. Isola. Understanding Contrastive Representation Learning Through Alignment and Uniformity on the Hypersphere. In ICML, 2020. X. Wang, H. Fan, Y. Tian, D. Kihara, and X. Chen. On the Importance of Asymmetry for Siamese Representation Learning. In CVPR, 2022. Z. Wen and Y. Li. The Mechanism of Prediction Head in Non-contrastive Self-supervised Learning. In Neur IPS, 2022. X. Weng, L. Huang, L. Zhao, R. M. Anwer, S. Khan, and F. S. Khan. An Investigation into Whitening Loss for Self-supervised Learning. In Neur IPS, 2022. J. Winkens, R. Bunel, A. G. Roy, R. Stanforth, V. Natarajan, J. R. Ledsam, P. Mac Williams, P. Kohli, A. Karthikesalingam, S. Kohl, et al. Contrastive Training for Improved Out-of-Distribution Detection. In ar Xiv, 2020. C. Wu, J. Pfrommer, M. Zhou, and J. Beyerer. Generative-Contrastive Learning for Self-Supervised Latent Representations of 3D Shapes from Multi-Modal Euclidean Input. ar Xiv, 2023. J. Xie, Y. Lu, S. C. Zhu, and Y. N. Wu. A Theory of Generative Convnet. In ICML, 2016. H. Xu, H. Xiong, and G. J. Qi. K-Shot Contrastive Learning of Visual Features with Multiple Instance Augmentations. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022. J. Xu, Z. Zhang, T. Friedman, Y. Liang, and G. Broeck. A Semantic Loss Function for Deep Learning with Symbolic Knowledge. In ICML, 2018. J. Zbontar, L. Jing, I. Misra, Y. Le Cun, and S. Deny. Barlow Twins: Self-Supervised Learning via Redundancy Reduction. In ICML, 2021. S. Zhang, F. Zhu, J. Yan, R. Zhao, and X. Yang. Zero-CL: Instance and Feature Decorrelation for Negative Free Symmetric Contrastive Learning. In ICLR, 2021. Y. Zhue, R. M. M. Renqiang, A. Kadav, and H. P. Graf. S3VAE: Self-Supervised Sequential VAE for Representation Disentanglement and Data Generation. In CVPR, 2020. R. S. Zimmermann, Y. Sharma, S. Schneider, M. Bethge, and W. Brendel. Contrastive Learning Inverts the Data Generating Process. In ICML, 2021. B Proof of Lemma 3 Definition 11 (Restated). Define the ground truth joint distribution for the graphical model in Fig. 2(a) as p(x1:n, y1:n) Qn j=1 p(xj)δ(yj j) with yj {1, . . . , n} and δ a delta function. Moreover, define the model distribution as p(x1:n, y1:n; Θ) Qn i=1 p(xj)p(yi|xi; Θ) with p(yi|xi; Θ) esim(g(xi),g(xyi ))/τ Pn k=1 esim(g(xi),g(xk))/τ , sim : Rl Rl R being a similarity function, τ > 0 a temperature parameter and Θ = {θ, {xi}n i } a set of parameters. Lemma 4 (Restated). Given Definition 11, maximizing the Info NCE objective (den Oord et al., 2018) is equivalent to maximize the following log-likelihood objective: Ep(x1:n,y1:n){log p(x1:n, y1:n)} Hp(x1:n) | {z } Neg. entropy term i=1 log esim(g(xi),g(xi))/τ Pn k=1 esim(g(xi),g(xk))/τ | {z } Predictive term LCT (Θ) Published in Transactions on Machine Learning Research (08/2025) Moreover, the maximization of this lower bound is equivalent to solve an instance classification problem. Proof. We first recall the Info NCE objective (cf. Eq. 10 in (Poole et al., 2019)) j=1 p(xj)δ(zi g(xj)) i=1 log ef(xi,zi) 1 n Pn k=1 ef(xi,zk) This objective can be rewritten in the following way: j=1 p(xj)δ(zi g(xj)) i=1 log ef(xi,zi) Pn k=1 ef(xi,zk) j=1 p(xj)δ(zi g(xj)) i=1 log ef(xi,zi) Pn k=1 ef(xi,zk) i=1 log ef(xi,g(xi)) Pn k=1 ef(xi,g(xk)) Next, we highlight an important relation between the ground truth and model log-likelihoods: Ep(x1:n,y1:n){log p(x1:n, y1:n)} = Ep(x1:n,y1:n){log p(x1:n, y1:n; Θ)} + KL{p(x1:n, y1:n) p(x1:n, y1:n; Θ)} Ep(x1:n,y1:n){log p(x1:n, y1:n; Θ)} (13) In other words, maximizing the term Ep(x1:n,y1:n){p(x1:n, y1:n; Θ)} is equivalent to minimize the KL term, thus solving an instance discrimination problem, where the model matches the instance classifier. This proves the last part of the lemma. Now, we are left to show the first part. We can elaborate the above lower bound even further by using the definitions at the basis of the Lemma. Indeed, Ep(x1:n,y1:n){log p(x1:n, y1:n; Θ)} = EQn j=1 p(xj)δ(yj j) i=1 p(xi)p(yi|xi; Θ) j=1 p(xj)δ(yj j){log p(xi)p(yi|xi; Θ)} i=1 Ep(xi)δ(yi i){log p(xi)p(yi|xi; Θ)} i=1 Ep(xi){log p(xi)p(yi = i|xi; Θ)} i=1 Ep(xi){log p(xi)} + i=1 Ep(xi){log p(yi = i|xi; Θ)} = Hp(x1:n) + i=1 Ep(xi){log p(yi = i|xi; Θ)} = Hp(x1:n) + EQn i=1 log p(yi = i|xi; Θ) = Hp(x1:n) + EQn i=1 log esim(g(xi),g(xi))/τ Pn k=1 esim(g(xi),g(xk))/τ Now, by defining f(x, z) sim(x, z)/τ and substituting it into Eq. 14, we can match with Eq. 12, thus concluding the proof. Published in Transactions on Machine Learning Research (08/2025) C Proof of Lemma 6 Definition 12 (Restated). Define the ground truth joint distribution for the graphical model in Fig. 2(b) as p(z, ξ1:n, x1:n, x 1:n) p(z) Qn i=1 p(xi|z)p(ξi)p(x i|xi, ξi) with Gaussian priors p(z) = N(z|0, I), p(ξi) = N(ξi|0, γ 1I) and assume the following conditional independencies p(x i|xi, ξi) = T (x i|xi) and p(xi|z) = p(xi). Moreover, introduce the following auxiliary model distributions, that is q(ξi|xi, x i) N(ξi|enc(xi) enc(x i), I) and q(z|x1:n) N(z|0, Σ), with Σ Pn i=1(g(xi) g)(g(xi) g)T + βI, β > 0 chosen to ensure the positive definiteness of Σ and g = 1/n Pn i=1 g(x)i. Finally, define Θ = {θ} as the set of parameters. Lemma 5 (Restated). Given Definition 12, maximizing the Cor Info Max objective (Ozsoy et al., 2022) is equivalent to maximize the following log-likelihood lower bound: Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} Hp(x1:n) | {z } Neg. entropy term Ep(x1:n){KL(q(z|x1:n) p(z))} | {z } discriminative SSL term LNF (Θ) i=1 Ep(xi)T (x i|xi){KL(q(ξi|xi, x i) p(ξi))} | {z } Continuation of LNF (Θ) where the first Kullback-Leibler term KL(q(z|x1:n) p(z)) KL(N(z|0, Σ) N(z|0, I)), the second term KL(q(ξi|xi, x i) p(ξi)) KL(N(ξi|enc(xi) enc(x i), I) N(ξi|0, γ 1I)) and const being a constant for the optimization over θ. Proof. We first recall the Cor Info Max objective (cf Eq. 6 in (Ozsoy et al., 2022)) LCor Info Max Ep(x1:n){log |Σ|} + γ i=1 Ep(xi)T (x i|xi){ g(xi) g(x i) 2} (15) with γ a positive scalar to weight the two objective terms. Next, we derive the log-likelihood lower bound in the Lemma. Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} = Ep(x1:n,x 1:n) i=1 p(xi|z)p(ξi)p(x i|xi, ξi) = Ep(x1:n,x 1:n) i=1 p(xi)p(ξi)T (x i|xi) = Ep(x1:n,x 1:n) i=1 p(xi)T (x i|xi) + Ep(x1:n) + Ep(x1:n,x 1:n) = Hp(x1:n) + const + Ep(x1:n) + Ep(x1:n,x 1:n) = Hp(x1:n) + const + Ep(x1:n) z q(z|x1:n) p(z) q(z|x1:n) + Ep(x1:n,x 1:n) ξi q(ξi|xi, x i) p(ξi) q(ξi|xi, x i) Hp(x1:n) + const + Ep(x1:n) z q(z|x1:n) log p(z) q(z|x1:n) Published in Transactions on Machine Learning Research (08/2025) i=1 Ep(xi)T (x i|xi) ξi q(ξi|xi, x i) log p(ξi) q(ξi|xi, x i) Hp(x1:n) + const + Ep(x1:n) z q(z|x1:n) log p(z) q(z|x1:n) i=1 Ep(xi)T (x i|xi) ξi q(ξi|xi, x i) log p(ξi) q(ξi|xi, x i) = Hp(x1:n) Ep(x1:n){KL(q(z|x1:n) p(z))} i=1 Ep(xi)T (x i|xi){KL(q(ξi|xi, x i) p(ξi))} + const (16) Now, observe that the first KL term can be rewritten as follows: Ep(x1:n){KL(q(z|x1:n) p(z))} = KL(N(z|0, Σ) N(z|0, I)) 2Ep(x1:n){log |Σ|} + const where the last quality holds whenever g has a batch normalization layer in its output, thus making Tr(Σ) a constant for the optimization. The second KL term can be rewritten as follows: i=1 Ep(xi)T (x i|xi){KL(q(ξi|xi, x i) p(ξi))} = i=1 Ep(xi)T (x i|xi){KL(N(ξi|enc(xi) enc(x i), I) N(ξi|0, γ 1I))} i=1 Ep(xi)T (x i|xi){ enc(xi) enc(x i) 2} + const Substituting both expresssions into Eq. 16 and choosing g enc, we recover the Cor Info Max objective up to a factor 1/2 and an additive constant. D Proof of Lemma 9 Definition 13 (Restated). Define the ground truth joint distribution for the graphical model in Fig. 2(c) as p(x1:n, x 1:n, y1:n) Qn i=1 p(xi)T (x i|xi)p(yi|xi; Θ) with yi {1, . . . , c} being a categorical variable to identify one of c clusters, p(yi|xi; Θ) = e UT :yi G:i/τ P y e UT :y G:i/τ , where U Rh c is the matrix6 of cluster centers, G = [g(x1), . . . , g(xn)] Rh n is a representation matrix and Θ = {θ, U} is the set of parameters. Moreover, introduce the auxiliary clustering distribution q(y1:n|x 1:n) Qn i=1 q(yi|x i). We can state the following Lemma (proof is provided in Appendix D) Lemma 6 (Restated). Given Definition 13, maximizing the Sw AV objective (Caron et al., 2020) is equivalent to maximize the following log-likelihood lower bound: Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} Hp(x1:n) | {z } Neg. entropy term 6We use subscripts to select rows and columns. For instance, U:y identify y th column of matrix U. Published in Transactions on Machine Learning Research (08/2025) i=1 Ep(xi)T (x i|xi) n Eq(yi|x i) log p(yi|xi; Θ) + Hq(yi|x i) o | {z } discriminative SSL term LCB(Θ) Proof. We first recall the Sw AV objective (cf. Eq. 2 and 3 in (Caron et al., 2020)) LSw AV(Θ) = i=1 Ep(xi)T (x i|xi){Eq(yi|x i) log p(yi|xi; Θ)} LSw AV(Q) = Ep(x1:n,x 1:n){Tr(QU T G)} + τEp(x1:n,x 1:n){HQ(y1:n|x 1:n)} where where Q [q(y1|x 1), . . . , q(yn|x n)]T is a prediction matrix of size n c and Tr( ) is the trace operator for any input matrix. Next, we derive the log-likelihood lower bound: Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} = Ep(x1:n,x 1:n) i=1 p(xi)T (x i|xi)p(yi|xi; Θ) = Ep(x1:n,x 1:n) yi p(xi)T (x i|xi)p(yi|xi; Θ) = Hp(x1:n) + const i=1 Ep(xi)T (x i|xi) yi p(yi|xi; Θ) = Hp(x1:n) + const i=1 Ep(xi)T (x i|xi) yi q(yi|x i)p(yi|xi; Θ) Hp(x1:n) + const i=1 Ep(xi)T (x i|xi) yi q(yi|x i) log p(yi|xi; Θ) = Hp(x1:n) + const i=1 Ep(xi)T (x i|xi) n Eq(yi|x i) log p(yi|xi; Θ) + Hq(yi|x i) o We observe that the maximization of this lower bound with respect to Θ is equivalent to maximize LSw AV (Θ). We can also show the equivalence to LSw AV (Q) by expressing the lower bound in vector format, namely: i=1 Ep(xi)T (x i|xi) n Eq(yi|x i) log p(yi|xi; Θ) + Hq(yi|x i) o = i=1 Ep(xi)T (x i|xi) Eq(yi|x i) log e U T :yi G:i/τ P y e U T :y G:i/τ + Hq(yi|x i) i=1 Ep(xi)T (x i|xi) U T :yi G:i + Hq(yi|x i) τ Ep(x1:n,x 1:n){Tr(QU T G)} + Ep(x1:n,x 1:n){HQ(y1:n|x 1:n)} + const = τLSw AV (Q) + const (17) thus concluding the proof. Published in Transactions on Machine Learning Research (08/2025) E Alternative View of Contrastive SSL Let us focus the analysis on a different graphical model from the one in Section 2.1, involving an input vector x and a latent embedding z (cf. Figure 1). Without loss of generality, we can discard the index i and focus on a single observation. We will later extend the analysis to the multi-sample case. Now, consider the conditional distribution of x given z expressed in the form of an energy-based model p(x|z; Θ) = ef(x,z) where f : Ω Sh 1 R is a score function and Γ(z; Θ) = R Ωef(x,z)dx is the normalizing factor.7 Based on this definition, we can obtain the following lower-bound on the data log-likelihood: Ep(x){log p(x)} = KL(p(x) p(x; Θ)) + Ep(x){log p(x; Θ)} Ep(x){log p(x; Θ)} Ep(x)q(z|x){log p(x|z; Θ)} Ep(x){KL(q(z|x) p(z; Θ))} = Ep(x)q(z|x){log p(x|z; Θ)} + Ep(x)q(z|x){log p(z; Θ)} = Ep(x)q(z|x){f(x, z)} + Ep(x)q(z|x) log p(z; Θ) .= ELBOEBM (18) where q(z|x) is an auxiliary density induced by a deterministic encoding function g : Ω Sh 1.8 Eq. (18) provides the basic building block to derive variational bounds on mutual information (Poole et al., 2019) as well as to obtain several popular contrastive SSL objectives. ELBOEBM and variational lower bounds on mutual information. Our analysis is similar to the one proposed in (Alemi et al., 2018), as relating ELBOEBM to the mutual information. However, while the work in (Alemi et al., 2018) shows that the third line in Eq. (18) can be expressed as a combination of an upper and a lower bound on the mutual information and it studies its rate-distortion trade-off, our analysis considers only lower bounds to mutual information and it makes an explicit connection to them. Indeed, several contrastive objectives are based on lower bounds on the mutual information between input and latent vectors (Poole et al., 2019). We can show that ELBOEBM is equivalent to these lower bounds under specific conditions for prior p(z; Θ) and score function f. Specifically, given a uniform prior p(z; Θ) and f(x, z) .= log p(x) + f(x, z) for all admissible pair x, z and for some arbitrary function f, we obtain the following equivalence (see Appendix F for the derivation): ELBOEBM = Ep(x){log p(x)} + Ep(x)q(z|x){ f(x, z)} Ep(x)q(z|x){log Ep(x ){e f(x ,z)}} = H(X) + IUBA(X, Z) (19) where H(X) = Ep(x){log p(x)} and IUBA(X, Z) refers to the popular Unnormalized Barber and Agakov bound on mutual information. Importantly, other well-known bounds can be derived from IUBA(X, Z) (cf. (Poole et al., 2019) and Appendix F for further details). Consequently, maximizing ELBOEBM with respect to the parameters of f (viz. θ) is equivalent to maximize a lower bound on mutual information. ELBOEBM and Info NCE (den Oord et al., 2018). Now, we are ready to show the derivation of the popular Info NCE objective. By specifying a non-parametric prior9 p(z; Θ) = Γ(z;Θ) 1 n Pn k=1 ef(xk,z) Γ(Θ) , where Γ(Θ) = R Γ(z;Θ) 1 n Pn k=1 ef(xk,z) dz, we achieve the following equality (see Appendix G for the derivation): ELBOEBM = EQn j=1 p(xj,zj) i=1 log ef(xi,zi) 1 n Pn k=1 ef(xk,zi) j=1 p(xj){log Γ(Θ)} 7We assume that f is a well-behaved function, such that the integral value Γ(z; Θ) is finite for all z Sh 1. 8q(z|x) = δ(z g(x)). 9The term non-parametric refers to the fact that the samples can be regarded as parameters of the prior and therefore their number increases with the number of samples. Indeed, we have Θ = {θ; {xi}n i=1}. Published in Transactions on Machine Learning Research (08/2025) .= INCE(X, Z) EQn j=1 p(xj){log Γ(Θ)} (20) Notably, ELBOEBM is equivalent to INCE(X, Z) up to the term log Γ(Θ). Maximizing ELBOEBM has the effect to maximize INCE(X, Z) and additionally to minimize Γ(Θ), thus ensuring that the prior p(z; Θ) selfnormalizes. However, in practice, people only maximize the Info NCE objective, disregarding the normalizing term. F Connection to Unnormalized Barber Agakov Bound Firstly, we recall the derivation of the Unnormalized Barber Agakov bound (Poole et al., 2019) for the mutual information IX,Z, adapting it to our notational convention. Secondly, we derive the equivalence relation in Eq. (19). IX,Z = Ep(x,z) = Ep(x)q(z|x) log p(x|z)q(x|z) = Ep(x)q(z|x) +Ep(x)q(z|x) = Eq(z)KL{p(x|z) q(x|z)}+Ep(x)q(z|x) Ep(x)q(z|x) = Ep(x)q(z|x) log p(x)e f(x,z) = Ep(x)q(z|x) log e f(x,z) = Ep(x)q(z|x){ f(x, z)} Ep(x)q(z|x){log Z(z)} = Ep(x)q(z|x){ f(x, z)} Ep(x)q(z|x){log Ep(x ){e f(x ,z)}} .= IUBA(X, Z) where we have introduced both an auxiliary encoder q(z|x) and an auxiliary decoder defined as q(x|z) = p(x)e f(x,z) Z(z) . Now, we can use the assumptions of a uniform prior p(z; Θ) and f(x, z) = log p(x) + f(x, z) to achieve the following inequalities:s ELBOEBM = Ep(x)q(z|x){f(x, z)} + Ep(x)q(z|x) log p(z; Θ) = Ep(x)q(z|x){f(x, z)} Ep(x)q(z|x){log Γ(z; Θ)} = Ep(x)q(z|x){log p(x) + f(x, z)} Ep(x)q(z|x){log Ep(x ){e f(x ,z)} = Ep(x)q(z|x){log p(x)} + Ep(x)q(z|x){ f(x, z)} Ep(x)q(z|x){log Ep(x ){e f(x ,z)} = H(X) + Ep(x)q(z|x){ f(x, z)} Ep(x)q(z|x){log Ep(x ){e f(x ,z)} = H(X) + IUBA(X, Z) F.1 Other Bounds Notably, IUBA(X, Z) cannot be tractably computed due to the computation of Z(z). Therefore, several other bounds have been derived (Poole et al., 2019) to obtain tractable estimators or optimization objectives, namely: Published in Transactions on Machine Learning Research (08/2025) 1. IT UBA(X, Z),10 which can be used for both optimization and estimation of mutual information (obtained using the inequality log s η s log η 1 for all scalar s, η > 0) IUBA(X, Z) = Ep(x,z){ f(x, z)} Ep(x)q(z|x){log Z(z)} Ep(x,z){ f(x, z)} Ep(x)q(z|x) η(z) + log η(z) 1 .= IT UBA(X, Z) 2. INW J(X, Z), which can be used for both optimization and estimation of mutual information (obtained from IT UBA(X, Z) by imposing η(z) = e) IT UBA(X, Z) = Ep(x,z){ f(x, z)} Ep(x)q(z|x) η(z) + log η(z) 1 = Ep(x,z){ f(x, z)} Ep(x)q(z|x) = Ep(x,z){ f(x, z)} 1 e Ep(x)q(z|x){Z(z)} = Ep(x,z){ f(x, z)} Ep(x )p(x)q(z|x){e f(x ,z) 1} .= INW J(X, Z) G Derivation of Info NCE Consider a prior p(z) = Γ(z;Θ) 1 n Pn k=1 ef(xk,z) Γ(Θ) , where Γ(Θ) = R Γ(z;Θ) 1 n Pn k=1 ef(xk,z) dz, ELBOEBM = Ep(x)q(z|x){f(x, z)} Ep(x)q(z|x){log Γ(z; Θ)} + Ep(x)q(z|x){log p(z; Θ)} = Ep(x1)q(z|x1){f(x1, z)} Ep(x1)q(z|x1){log Γ(z; Θ)} + Ep(x1)q(z|x1){log p(z; Θ)} = Ep(x1)p(x2) p(xn)q(z|x1){f(x1, z)} Ep(x1)p(x2) p(xn)q(z|x1){log Γ(z; Θ)} + Ep(x1)p(x2) p(xn)q(z|x1){log p(z; Θ)} = Ep(x1)p(x2) p(xn)q(z|x1){f(x1, z)} Ep(x1)p(x2) p(xn)q(z|x1) k=1 ef(xk,z) Ep(x1)p(x2) p(xn){log Γ(Θ)} = Ep(x1)p(x2) p(xn)q(z|x1){log ef(x1,z)} Ep(x1)p(x2) p(xn)q(z|x1) k=1 ef(xk,z) Ep(x1)p(x2) p(xn){log Γ(Θ)} = Ep(x1,z)p(x2) p(xn) log ef(x1,z) 1 n Pn k=1 ef(xk,z) Ep(x1)p(x2) p(xn){log Γ(Θ)} = Ep(x1,z1)p(x2,z2) p(xn,z K) log ef(x1,z1) 1 n Pn k=1 ef(xk,z1) Ep(x1)p(x2) p(xn){log Γ(Θ)} i=1 Ep(x1,z1)p(x2,z2) p(xn,z K) log ef(xi,zi) 1 n Pn k=1 ef(xk,zi) Ep(x1)p(x2) p(xn){log Γ(Θ)} = Ep(x1,z1)p(x2,z2) p(xn,z K) i=1 log ef(xi,zi) 1 n Pn k=1 ef(xk,zi) Ep(x1)p(x2) p(xn){log Γ(Θ)} .= INCE(X, Z) EQn j=1 p(xj){log Γ(Θ)} 10T in TUBA stands for tractable. Published in Transactions on Machine Learning Research (08/2025) H Derivation of Proto CPC: A Lower Bound of Info NCE Proto CPC (Lee, 2022) can be viewed as a lower bound of the Info NCE objective. To save space, we use notation E to refer to EQn j=1 p(xj,zj). Therefore, we have that i=1 log ef(xi,zi) 1 n Pn k=1 ef(xk,zi) i=1 log e PC c=1 pt c(xi) log ps c(xi) 1 n Pn k=1 e PC c=1 ptc(xk) log psc(xi) c=1 pt c(xi) log e zs c,xi /τs c=1 pt c(xi) log PC c =1 e zs c ,xi /τs 1 n Pn k=1 e PC c=1 ptc(xk) log psc(xi) c=1 pt c(xi) log e zs c,xi /τs c=1 pt c(xi) log PC c =1 e zs c ,xi /τs 1 n Pn k=1 e PC c=1 pt c(xk) log e zs c,xi /τs c=1 pt c(xk) log PC c =1 e zs c ,xi /τs c=1 pt c(xi) log e zs c,xi /τs c =1 e zs c ,xi /τs 1 n Pn k=1 e PC c=1 pt c(xk) log e zs c,xi /τs c =1 e zs c ,xi /τs i=1 log e PC c=1 pt c(xi) log e zs c,xi /τs 1 n Pn k=1 e PC c=1 ptc(xk) log e zsc,xi /τs i=1 log e PC c=1 pt c(xi)zs c,xi/τs 1 n Pn k=1 e PC c=1 ptc(xk)zsc,xi/τs i=1 log e PC c=1 pt c(xi)zs c,xi/τs 1 n Pn k=1 PC c=1 ptc(xk)ezs c,xi/τs i=1 log e PC c=1 pt c(xi)zs c,xi/τs PC c=1 1 n Pn k=1 ptc(xk)ezsc,xi/τs i=1 log e PC c=1 pt c(xi)zs c,xi/τs PC c=1 qtcezsc,xi/τs i=1 log ept(xi)T zs xi/τs PC c=1 qtcezsc,xi/τs .= IP roto CP C Specifically, the second equality comes from the fact that f(x, z) = PC c=1 pt c(x) log ps c(x), where s, t stand for student and teacher networks, respectively, and p c(x) is the c-th entry of the vector obtained by applying a softmax on the embedding of the corresponding network. The inequality in the derivation is obtained by applying Jensen s inequality to the denominator. Finally, qt c = 1 n Pn k=1 pt c(xk) corresponds to the prototype for class c. Published in Transactions on Machine Learning Research (08/2025) I Details about Negative-Free Methods We can specify different definitions for LNF (Θ), namely: 1. Barlow Twins. The approach enforces the cross-correlation matrix to be close to the identity matrix: LNF (Θ) = CCorr(G, G ) Λ I 2 F where F is the Frobenius norm, Λ = Jλ + (1 λ)I, I is the identity matrix, λ is a positive hyperparameter, J is a matrix of ones, CCorr(G, G ) = HT H is the sample cross-correlation matrix, G = [g(x1), . . . , g(x n)]T Rn h and G = [g(x 1), . . . , g(x n)]T Rn h and g(xi) = BN(Proj(enc(xi))) Note that BN is a batch normalization layer, Proj is a projection head usually implemented using a multi-layer perceptron and enc is the encoder. 2. W-MSE. This approach is similar to Barlow Twins. The main difference lie in the fact that sample cross-correlation matrix is enforced to be an identity matrix thanks to a whitening layer: LNF (Θ) = Tr(G G ) where Tr(G G ) computes the trace of the outer product for matrices G = [g(x1), . . . , g(x n)]T Rn h and G = [g(x 1), . . . , g(x n)]T Rn h. g(xi) = L2-Norm(Whitening(Proj(enc(xi)))) L2-Norm is a normalization layer based on L2 norm. 3. VICReg. This approach attempts to simplify the architecture of Barlow Twins by introducing an invariance regularization term in the score function, thus avoiding to use a batch normalization layer: LNF (Θ) = λTr((G G ) (G G )) µ[v(G) + v(G )] ν[w(G) + w(G )] where λ, µ, ν are positive hyperparameters and G = [g(x1), . . . , g(x n)]T Rn h and G = [g(x 1), . . . , g(x n)]T Rn h. The first addend enforces the representation to be invariant to the data augmentation, whereas the other two addends enforce the sample covariance matrix to be diagonal. Indeed, the second addend forces the diagonal elements of the sample covariance matrices to be unitary, namely: j=1 max{0, 1 q V ar{G:j} + ϵ} where ϵ > 0 is used to avoid numerical issues, V ar computes the variance for each column of matrix G. The third addend ensures that the off-diagonal elements of the sample covariance matrix approach to zero: c(G) = Cov(G) (J I) 2 F with Cov(H) = GT G. Therefore, these last two addends have a similar behaviour to the score function of Barlow Twins. Finally, the resulting function is simplified. g(xi) = Proj(enc(xi)) Published in Transactions on Machine Learning Research (08/2025) J Proof of Proposition 1 Proposition 2. (Restated) Given Definition 7, the expected data log-likelihood for the probabilistic graphical model in Fig. 2(c) can be alternatively lower bounded as follows: Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} Hp(x1:n) i=1 Ep(xi)T (x i|xi) {CE(p(yi|x i; Θ), p(yi|xi; Θ))} | {z } LINV (Θ) j=1 p(yj = yi|xj; Θ) | {z } LP RIOR(Θ) +const (21) Additionally, the corresponding maximum value for the last two addends in Eq. (21) is given by the following inequality:11 LINV (Θ) + LP RIOR(Θ) Hp(y1:n) (22) Proof. Let s start to rewrite the log-likelihood term. Ep(x1:n,x 1:n){log p(x1:n, x 1:n)} = Ep(x1:n,x 1:n) i=1 p(xi)T (x i|xi)p(yi|xi; Θ) = Ep(x1:n,x 1:n) yi p(xi)T (x i|xi)p(yi|xi; Θ) = Hp(x1:n) + const i=1 Ep(xi)T (x i|xi) yi p(yi|xi; Θ) (23) Now add the zero quantity log P y1:n p(y1:n) to the previous equation: Eq. (23) = Hp(x1:n) + const + i=1 Ep(xi)T (x i|xi) yi p(yi|xi; Θ) y1:n p(y1:n) (24) We can lower bound the previous equation by exploiting the fact that P z p(z)q(z) for any given auxiliary discrete distribution q, viz.: Eq. (24) Hp(x1:n) + const + i=1 Ep(xi)T (x i|xi) yi q(yi|x i)p(yi|xi, Θ) y1:n p(y1:n)q(y1:n) (25) Now, by applying Jensen s inequality to the last two addends in Eq. (25), by defining q(y1:n) = 1 n Pn j=1 p(yj|xj, Θ) and considering Definition 10, we obtain the following lower bound: Eq. (24) Hp(x1:n) + const + i=1 Ep(xi)T (x i|xi) yi p(yi|x i, Θ) log p(yi|xi, Θ) y1:n p(y1:n) log 1 j=1 p(yj|xj, Θ) (26) 11Here, we assume that the predictive model p(y|x; Θ) has enough capacity to achieve the optimal solution. Published in Transactions on Machine Learning Research (08/2025) Additionally, by substituting p(y1:n) = Qn i=1 p(yi) into the above bound, we achieve the following equality: Eq. (26) = Hp(x1:n) + const + i=1 Ep(xi)T (x i|xi) yi p(yi|x i; Θ) log p(yi|xi; Θ) yi p(yi) log j=1 p(yj = yi|xj; Θ) And by rewriting the last two addends in Eq. (27) using the definition of cross-entropy, we obtain our final result. Now, we can conclude the proof by looking at the maxima for LINV and LP RIOR. Indeed, we observe that both terms compute a negative cross-entropy between two distributions. By leveraging the fact that CE(p, q) = Hp + KL(p q) for arbitrary distributions p, q, we can easily see that the maximum of LINV is attained when the term is 0 (corresponding to minimal entropy and minimal KL), whereas the maximum of LP RIOR is attained when the term is equal to Hp(yi) (corresponding to minimal KL). K Proof of Theorem 1 Proof. The overall strategy to prove the statements relies on the evaluation of the loss terms over the three failure modes and on checking whether these attain their corresponding maxima. Let s start by proving statement a and recalling that LGEN(Θ, D) = CE(p, pΘ). Firstly, we test for failure mode 1 (i.e. representational collapse). We observe that for all x Rd Pc y=1 e U T :yg(k)/τ thus pΘ(x) assigns constant mass everywhere. Clearly, pΘ is different from p. Therefore, CE(p, pΘ) < CE(p, p) and failure mode 1 is not admissible. Secondly, we test for failure mode 2 (i.e. cluster collapse). We can equivalently rewrite the definition of cluster collapse by stating that there exists j {1, . . . , c}, such that for all x Rd and for all y = j, U T :j g(x) U T :yg(x) . Additionally, we observe that Pc y=1 e U T :yg(x)/τ R Pc y=1 e U T :yg(x)/τdx e U T :j g(x)/τ e(U T :yg(x) U T :j g(x))/τ R efj(ξx)/τ e(U T :yg(x) U T :j g(x))/τ = e U T :j g(x)/τ R e U T :j g(x)/τdx (28) where we have used the failure mode condition to obtain the last equality. Now, note that Eq. (28) defines a standard energy-based model. Consequently, given enough capacity for the predictive model, it is trivial to verify that there exists Θ such that the condition about failure mode is met and pΘ is equal to p. Cluster collapse is therefore an admissible solution. Thirdly, we test for the inconsistency of cluster assignments. Indeed, we have that LGEN(Θ, D) = i=1 Ep(xi) {log pΘ(xi)} Published in Transactions on Machine Learning Research (08/2025) Pc y=1 eti(y)/τ R Pc y=1 eti(y)/τdx i=1 Ep(x1:n) Pc y=1 eti(y)/τ R Pc y=1 eti(y)/τdx where ti(y) = U T :yg(xi). Similarly, we have that LGEN(Θ, Dπ) = |{z} from Eq. (29) i=1 Ep(x1:n) Pc y=1 etπ(i)(y)/τ R Pc y=1 etπ(i)(y)/τdx i=1 Ep(xπ(i)) Pc y=1 etπ(i)(y)/τ R Pc y=1 etπ(i)(y)/τdx = |{z} l .= π(i) Pc y=1 etl(y)/τ R Pc y=1 etl(y)/τdx = LGEN(Θ, D) Hence, failure mode 3 is an admissible solution. Let s continue by proving statement b and recalling that LINV (Θ, D) = i=1 Ep(xi)T (x i|xi) {CE(p(yi|x i; Θ), p(yi|xi; Θ))} (30) Firstly, we test for representational collapse. In this case, we have that for all i {1, . . . , n} p(yi|xi; Θ) = p(yi|x i; Θ) = Softmax(U T k/τ) Based on this result, we observe that the cross-entropy terms in Eq. (30) can be made 0 by proper choice of k. Therefore, representational collapse is an admissible solution. Secondly, we test for cluster collapse. Here, it is easy to see that the cross-entropy terms in Eq. (30) are all 0. Therefore, also cluster collapse is admissible. Thirdly, we test for the inconsistency of cluster assignments. On one hand, we have that the cross-entropy terms for LINV (Θ, D)) in Eq. (30) can be rewritten in the following way: CE(p(yi|x i; Θ), p(yi|xi; Θ)) et i(yi)/τ Pc y=1 et i(y)/τ , eti(yi)/τ Pc y=1 eti(y)/τ and the optimal solution is achieved only when t i = ti for all i {1, . . . , n}. On the other hand, the cross-entropy terms for LINV (Θ, Dπ) are given by the following equality: CE(p(yi|x i; Θ), p(yi|xi; Θ)) et i(yi)/τ Pc y=1 et i(y)/τ , etπ(i)(yi)/τ Pc y=1 etπ(i)(y)/τ However, the optimal solution cannot be achieved in general as t i = tπ(i) for some i {1, . . . , n}.12 Therefore, LINV is not permutation invariant to cluster assignments. Let s conclude by proving statement c and recalling that LP RIOR(Θ, D) = l=1 p(yl = yi|xl; Θ) 12Indeed, note that t i = tπ(i) for all i occurs only when we are in one of the first two failure modes. Published in Transactions on Machine Learning Research (08/2025) Firstly, we test for representational collapse. One can easily observe that if enc(x) = k for all x Rd, p(y|x; Θ) becomes uniform, namely p(y|x; Θ) = 1/c for all y {1, . . . , c}. Consequently, 1 n Pn l=1 p(yl = yi|xl; Θ) = 1/c for all i {1, . . . , n}. Now, since p(yi) = 1/c for all i {1, . . . , n}, the cross-entropy terms in Eq. (33) reach their maximum value Hp(yi) for all i {1, . . . , n}. Therefore, representational collapse attains the global maximum of LP RIOR and is an admissible solution. Secondly, we test for cluster collapse. By using the definition of cluster collapse, we observe that l=1 p(yl = yi|xl; Θ) = 0 yi = j 1 yi = j (34) Therefore, the resulting distribution is non-uniform, differently from p(yi). The cross-entropy terms in Eq. (33) are not optimized and cluster collapse is not admissible. Thirdly, we test for the inconsistency of cluster assignments. We observe that l=1 p(yl = yi|xl; Θ) = 1 etl(yi)/τ Pc y=1 etl(y)/τ etπ(l)(yi)/τ Pc y=1 etπ(l)(y)/τ (35) which is permutation invariant to cluster assignments. Consequently, also LP RIOR(Θ, D) = LP RIOR(Θ, Dπ). This concludes the proof. L Hyperparameters for Synthetic Data For the backbone enc, we use a MLP with two hidden layers and 100 neurons per layer, an output layer with 2 neurons and Re LU activation functions. For the projection head proj (f for GEDI and its variants), we use a MLP with one hidden layer and 4 neurons and an output layer with 2 neurons (batch normalization is used in all layers for Barlow and Sw AV as required by their original formulation). All methods use a batch size of 400. Baseline JEM (following the original paper): Number of iterations 20K; learning rate 1e 3; Adam optimizer(β1 = 0.9, β2 = 0.999); SGLD steps 10; buffer size 10000; reinitialization frequency 0.05; SGLD step-size 0.012 2 ; SGLD noise 0.01. For the self-supervised learning methods, please refer to Table 7. We also provide an analysis of sensitivity to hyperparameters for GEDI (Figure 8). Table 7: Hyperparameters used in the synthetic experiments. Methods Barlow Sw AV GEDI no gen GEDI Iters 20k Learning rate 1e 3 Optimizer Adam β1 = 0.9, β2 = 0.999 Data augmentation noise σ 0.03 SGLD steps T - - 1 1 Buffer size |B| - - 10000 10000 Reinitialization frequenc - - 0.05 0.05 SGLD step size - - 0.012 2 SGLD noise - - 0.01 0.01 Weight for LGEN - - 1 1 Weight for LINV - - 50 50 Weight for LP RIOR - - 10 10 M Hyperparameters for SVHN, CIFAR10, CIFAR100 For the backbone enc, we use a Res Net with 8 layers as in (Duvenaud et al., 2021), where its architecture is shown in Table 8. For the projection head proj (f for GEDI and its variants), we use a MLP with one hidden Published in Transactions on Machine Learning Research (08/2025) (b) Circles Figure 8: Sensitivity analysis on the discriminative performance of GEDI fr different loss weights (in the range {0, 10, 20, 30, 40, 50}). Performance are averaged over 5 different random seeds. Yellow means perfect NMI. Table 8: Resnet architecture. Conv2D(A,B,C) applies a 2d convolution to input with B channels and produces an output with C channels using stride (1, 1), padding (1, 1) and kernel size (A, A). Name Layer Res. Layer Conv2D(3,3,F) Avg Pool2D(2) Leaky RELU(0.2) Conv2D(3,F,F) Conv2D(1,3,F) no padding Avg Pool2D(2) Leaky RELU(0.2) Conv2D(3,F,F) Leaky RELU(0.2) Conv2D(3,F,F) Avg Pool2D(2) Leaky RELU(0.2) Conv2D(3,F,F) Leaky RELU(0.2) Conv2D(3,F,F) Leaky RELU(0.2) Conv2D(3,F,F) Leaky RELU(0.2) Conv2D(3,F,F) Avg Pool2D(all) layer and 2 F neurons and an output layer with F neurons (batch normalization is used in all layers for Barlow and Sw AV as required by their original formulation + final L2 normalization). F = 128 for SVHN, CIFAR-10 (1 million parameters) and F = 256 for CIFAR-100 (4.1 million parameters). For JEM, we use the same settings of (Duvenaud et al., 2021). All methods use a batch size of 64. Baseline JEM (following the original paper): number of epochs 20, 200, 200 for SVHN, CIFAR-10, CIFAR-100, respectively; learning rate 1e 4; Adam optimizer; SGLD steps 20; buffer size 10000; reinitialization frequency 0.05 ; SGLD step-size Published in Transactions on Machine Learning Research (08/2025) Table 9: Hyperparameters (in terms of sampling, optimizer, objective and data augmentation) used in all experiments. Class Name param. SVHN CIFAR-10 CIFAR-100 MNIST Addition Data augment. Color jitter prob. 0.1 0.1 0.1 0.1 0.1 Gray scale prob. 0.1 0.1 0.1 0.1 0.1 Random crop Yes Yes Yes Yes Yes Additive Gauss. noise (std) 0.03 0.03 0.03 0.2 0.2 Random horizontal flip No Yes Yes No No SGLD iters 20 20 20 20 20 Buffer size 10k 10k 10k 10k 10k Reinit. frequency 0.05 0.05 0.05 0.05 0.05 SGLD step-size 1 1 1 1 1 SGLD noise 0.01 0.01 0.01 0.01 0.01 Batch size 64 64 64 60 60 Epochs 20 200 200 * * Adam β1 0.9 0.9 0.9 0.9 0.9 Adam β2 0.999 0.999 0.999 0.999 0.999 Learning rate 1e 4 1e 4 1e 4 1e 4 1e 4 L2 regularization 0 0 0 1e 4 1e 4 Weights for losses LGEN 1 1 1 1 1 LINV 50 50 50 50 0 LP RIOR 25 25 50 25 400 LNESY - - - 0 25 *: For the 3 different dataset sizes (100, 1000, 10000), we trained for 500, 100 and 25 epochs respectively. 1; SGLD noise 0.01; data augmentation (Gaussian noise) 0.03. For the self-supervised learning methods, please refer to Table 9. N Additional Experiments on SVHN, CIFAR-10, CIFAR-100 We conduct a linear probe evaluation of the representations learnt by the different models Table 10. These experiments provide insights on the capabilities of learning representations producing linearly separable classes. From Table 10, we observe a large difference in results between Barlow and Sw AV. Our approach provides interpolating results between the two approaches. We also provide additional qualitative analyisis on the generation performance on SVHN and CIFAR-100. Please, refer to Figure 9 and Figure 10. Finally, we evaluate the performance in terms of OOD detection, by following the same methodology used in (Grathwohl et al., 2020). We use the OOD score criterion proposed in (Grathwohl et al., 2020), namely s(x) = log pΨ(x) x 2. From Table 5, we observe that GEDI achieves almost optimal performance. While these results are exciting, it is important to mention that they are not generally valid. Indeed, when training on CIFAR-10 and performing OOD evaluation on the other datasets, we observe that all approaches achieve similar performance both on CIFAR-100 and SVHN, suggesting that all datasets are considered in-distribution, see Table 11. A similar observation is obtained when training on CIFAR-100 and evaluating on CIFAR-10 and SVHN, see Table 12. Importantly, this is a phenomenon which has been only recently observed by the scientific community on generative models. Tackling this problem is currently out of the scope of this work. For further discussion about the issue, we point the reader to the works in (Nalisnick et al., 2019). Published in Transactions on Machine Learning Research (08/2025) (c) GEDI no gen Figure 9: Qualitative visualization of the generative performance for the different discriminative strategies on SVHN. Results are obtained by running Stochastic Langevin Dynamics for 500 iterations. (c) GEDI no gen Figure 10: Qualitative visualization of the generative performance for the different discriminative strategies on CIFAR-100. Results are obtained by running Stochastic Langevin Dynamics for 500 iterations. Table 10: Supervised linear evaluation in terms of accuracy on test set (SVHN, CIFAR-10, CIFAR-100). The linear classifier is trained for 100 epochs using SGD with momentum, learning rate 1e 3 and batch size 100. Mean and standard deviations are computed over results from 5 different initialization seeds. Dataset JEM Barlow Sw AV GEDI no gen GEDI SVHN 0.20 0.00 0.75 0.01 0.44 0.04 0.59 0.02 0.54 0.01 CIFAR-10 0.24 0.00 0.65 0.00 0.50 0.02 0.63 0.01 0.63 0.01 CIFAR-100 0.03 0.00 0.27 0.01 0.14 0.01 0.31 0.01 0.31 0.01 Table 11: OOD detection in terms of AUROC on test set (SVHN, CIFAR-100). Training is performed on CIFAR-10. Dataset JEM Barlow Sw AV GEDI no gen GEDI SVHN 0.44 0.32 0.62 0.11 0.57 CIFAR-100 0.53 0.56 0.51 0.51 0.61 Table 12: OOD detection in terms of AUROC on test set (SVHN, CIFAR-10). Training is performed on CIFAR-100. Dataset JEM Barlow Sw AV GEDI no gen GEDI SVHN 0.44 0.45 0.30 0.55 0.53 CIFAR-10 0.49 0.43 0.47 0.46 0.48 O Details on the MNIST addition experiment. The hyperparameters for the experiment without and with the Ne Sy constraint are as reported in Table 9. The data was generated by uniformly sampling pairs a, b such that 0 a 9, 0 b 9 and 0 Published in Transactions on Machine Learning Research (08/2025) a + b 9. For each triplet (a, b, c), we assigned to a, b, c, random MNIST images with corresponding labels, without replacement. For the experiment with the Ne Sy constraint, we use a slightly different uniformity regularization. We maximize the entropy of the mean output distribution for each batch, cf. Manhaeve et al. (2021). O.1 Replacing LINV with LNESY For these experiments, we replace LINV with LNESY as the Ne Sy loss also forces a clustering of the digits, making LINV redundant. Figure 11 shows that both the Ne Sy Loss and the clustering loss progress similarly. Figure 11: Comparing LNESY with LINV on the Addition experiment, where only the Ne Sy loss is optimized.