# state_representation_learning_using_an_unbalanced_atlas__0444e1e9.pdf Published as a conference paper at ICLR 2024 STATE REPRESENTATION LEARNING USING AN UNBALANCED ATLAS Li Meng University of Oslo Oslo, Norway li.meng@its.uio.no Morten Goodwin University of Agder Kristiansand, Norway morten.goodwin@uia.no Anis Yazidi Oslo Metropolitan University Oslo, Norway anisy@oslomet.no Paal Engelstad University of Oslo Oslo, Norway paal.engelstad@its.uio.no The manifold hypothesis posits that high-dimensional data often lies on a lowerdimensional manifold and that utilizing this manifold as the target space yields more efficient representations. While numerous traditional manifold-based techniques exist for dimensionality reduction, their application in self-supervised learning has witnessed slow progress. The recent MSim CLR method combines manifold encoding with Sim CLR but requires extremely low target encoding dimensions to outperform Sim CLR, limiting its applicability. This paper introduces a novel learning paradigm using an unbalanced atlas (UA), capable of surpassing state-of-the-art self-supervised learning approaches1. We investigated and engineered the Deep Infomax with an unbalanced atlas (DIM-UA) method by adapting the Spatiotemporal Deep Infomax (ST-DIM) framework to align with our proposed UA paradigm. The efficacy of DIM-UA is demonstrated through training and evaluation on the Atari Annotated RAM Interface (Atari ARI) benchmark, a modified version of the Atari 2600 framework that produces annotated image samples for representation learning. The UA paradigm improves existing algorithms significantly as the number of target encoding dimensions grows. For instance, the mean F1 score averaged over categories of DIM-UA is 75% compared to 70% of ST-DIM when using 16384 hidden units. 1 INTRODUCTION Self-supervised learning (SSL) is a field in machine learning (ML) that aims to learn useful feature representations from unlabelled input data. SSL includes mainly contrastive methods (Oord et al., 2018; Chen et al., 2020; He et al., 2020) and generative models (Kingma & Welling, 2013; Gregor et al., 2015; Oh et al., 2015). Generative models rely on using generative decoding and reconstruction loss, whereas typical contrastive methods do not involve a decoder but apply contrastive similarity metrics to hidden embeddings instead (Liu et al., 2021). State representation learning (SRL) (Anand et al., 2019; Jonschkowski & Brock, 2015; Lesort et al., 2018) focuses on learning representations from data typically collected in a reinforcement learning (RL) environment. A collection of images can be sampled through an agent interacting with the environment according to a specified behavior policy. Such images are interesting as study subjects due to their innate temporal/spatial correlations. Moreover, RL can also benefit from self-supervised learning just as computer vision (CV) and natural language processing (NLP) do, and successful pretraining of neural network (NN) models may lead to improvements in downstream RL tasks. A manifold can be learned by finding an atlas that accurately describes the local structure in each chart (Pitelis et al., 2013). In SSL, using an atlas can be viewed as a generalization of both dimen- 1Code is available at https://github.com/mengli11235/DIM-UA. Published as a conference paper at ICLR 2024 sionality reduction and clustering (Korman, 2018; 2021a;b). Namely, it generalizes the case where only one chart exists and where the charts do not overlap in an atlas. In MSim CLR (Korman, 2021b), NNs can encode an atlas of a manifold by having chart embeddings and membership probabilities. Uniform prior Unbalanced atlas Figure 1: The entropy of the output vector recorded epoch-wise when pretrained on the CIFAR10 dataset for a total of 1000 epochs, utilizing 8 charts and a dimensionality of 256. One primary issue of MSim CLR is its reliance on a uniform prior, which allocates inputs into each chart embedding uniformly. We postulate that although this uniform prior may more effectively represent the data distribution when d is exceedingly small, it concurrently introduces higher prediction uncertainty. Simultaneously, it also suffers from a problem akin to that faced by bootstrapped methods in RL. It has been noted that multiple NN heads inside a model, in the absence of additional noise, tend to output similar results after being trained a large number of epochs (Osband & Van Roy, 2015; Osband et al., 2016; Ecoffet et al., 2019; Meng et al., 2022). To rectify the aforementioned problems, this study introduces a novel SSL paradigm that leverages an unbalanced atlas (UA). In this context, UA denotes the absence of a uniform prior distribution, with the membership probability distribution deliberately trained to deviate significantly from uniformity. As illustrated in Fig. 1, it is evident that the entropy of the output vector during pretraining when using UA is markedly lower than that with a uniform prior, which suggests a heightened degree of confidence in its predictions. Our contribution is summarized as follows: (1) We modify the SRL algorithm ST-DIM (Anand et al., 2019) with our UA paradigm and introduce a new algorithm called DIM-UA. This furthers the research into the integration of RL and SSL with a novel manifold-based learning paradigm. DIM-UA achieves the state-of-the-art performance on samples collected from 19 Atari games of the Atari ARI benchmark. (2) We also provide detailed ablations and additional experiments on CIFAR10 to examine different underlying effects of possible design choices. (3) We demonstrate that our UA paradigm is capable of effectively representing a manifold with a large number (e.g., 256) of hidden dimensions, whereas previous research (Korman, 2021a;b) only showed promise with a small number (e.g., 8) of hidden dimensions. The UA paradigm thereby showcases its capability to build larger models, transcending the constraints imposed by model backbones. 2 RELATED WORK Dimensionality reduction with manifolds It is common for nonlinear dimensionality reduction (NLDR) algorithms to approach their goals based on the manifold hypothesis. For example, the manifold structure of an isometric embedding can be discovered by solving for eigenvectors of the matrix of graph distances (Tenenbaum et al., 2000). A sparse matrix can also be used instead with a locally linear embedding (Roweis & Saul, 2000). Correspondence between samples in different data sets can be recovered through the shared representations of the manifold (Ham et al., 2003). Manifold regularization provides an out-of-sample extension compared to graph-based approaches Published as a conference paper at ICLR 2024 (Belkin et al., 2006). Manifold sculpting simulates surface tension progressively in local neighborhoods to discover manifolds (Gashler et al., 2007). Self-supervised learning There are relevant works on generative models, such as variational autoencoders (VAEs) (Kingma & Welling, 2013) and adversarial autoencoders (AAEs) (Makhzani et al., 2015). Meanwhile, contrastive methods have shown promise in the field of SSL. Contrastive Predictive Coding (CPC) learns predictive representations based on the usefulness of the information in predicting future samples (Oord et al., 2018). Sim CLR provides a simple yet effective framework using data augmentations (Chen et al., 2020). Momentum Contrast (Mo Co) utilizes a dynamic dictionary, which can be much larger than the mini-batch size (He et al., 2020). The recent trend within research in contrastive learning has been on removing the need for negative pairs. BYOL utilizes a momentum encoder to prevent the model from collapsing due to a lack of negative pairs (Grill et al., 2020). Sim Siam further shows that a stop-gradient operation alone is sufficient (Chen & He, 2021). Barlow Twins, on the other hand, achieves so by minimizing the redundancy of vector components outputted by two identical networks that take distorted versions of inputs (Zbontar et al., 2021). Self-supervised learning with manifolds Representing non-Euclidean data in NN models is a key topic in geometric deep learning (Bronstein et al., 2017). Learning manifolds using NNs was explored in (Korman, 2018), in which AAEs were used to learn an atlas as latent parameters. Constant-curvature Riemannian manifolds (CCMs) of different curvatures can be learned similarly using AAEs (Grattarola et al., 2019). Mixture models of VAEs can be used to express the charts and their inverses to solve inverse problems (Alberti et al., 2023). A combination of autoencoders and Barlow Twins can capture both the linear and nonlinear solution manifolds (Kadeethum et al., 2022). Uα Z Uβ ψα ψβ ψαβ Figure 2: A manifold Z embedded in a higher dimension. Two domains are denoted by Uα and Uβ in Z. ψα and ψβ are the corresponding charts that map them to a lower dimensional Euclidean space. An atlas is then a collection of these charts that together cover the entire manifold. Our method extends the work from Korman (2021b) and builds a manifold representation using multiple output embeddings and membership probabilities of those embeddings. First, an illustration of a manifold Z is in Fig. 2. Instead of directly learning an encoder of Z, we can learn encoder functions of ψα(Uα) and ψβ(Uβ) together with a score function. For a given input, we can specify which encoder to use according to the output of the score function. We formally model a distribution of input data by a manifold as follows: x is the input from input space X, Z is the latent space, f is an embedding function: X Z. I is the identity mapping, d the number of dimensions for each chart output embedding, n the number of charts, and N denotes {1, 2, ..., n}. ψi: Z Rd is the inverse mapping of a coordinate map: Rd Z, whereas q = (q1, q2, ..., qn) : Z [0, 1]n is the chart membership function. The output of our model is then given by Eq. 1. Published as a conference paper at ICLR 2024 Output(x) = X qi(f(x))I(ψi(f(x))) (1) At inference time, the one-hot encoding of q(x) is used instead (Eq. 2). Output(x) = I(ψi(f(x))), where i = argmaxj qj(f(x)) (2) 3.1 UNBALANCED ATLAS Like other SSL methods with manifolds (Korman, 2021b;a), UA uses a maximal mean discrepancy (MMD) objective (Gretton et al., 2012; Tolstikhin et al., 2017), which is defined by Eq. 3. MMDk(P1, P2) = Z S k(s, )d P1(s) Z S k(s, )d P2(s) Hk (3) Here, k is a reproducing kernel, Hk is the reproducing kernel Hilbert space of real-valued functions mapping S to R, and P1, P2 are distributions on the space. In our paradigm, the input x is designed to be represented in charts with higher membership probabilities. Thus, we take an MMD loss that moves the conditional membership distribution far away from the uniform distribution. We use the kernel k N : N N R, (i, j) δij, and δij = 1 if i = j else 0, and thus have Eq. 4. LN (q) = Ez MMDk N (q(z), UN ) = Ez i=1 (qi(z) 1 Here, UN denotes the uniform distribution on N, z is the embedding of f(x). Unlike MSim CLR, we do not use an MMD objective to make the prior distribution to be uniform, but take another approach to improve the model stability and head diversity when d is not trivial. In Fig. 2, Uα Uβ has transitive maps in their respective chart coordinates with domains restricted to ψα(Uα Uβ) and ψβ(Uα Uβ), which are ψαβ = ψβ ψ 1 α and ψβα = ψα ψ 1 β . What interests us the most is this intersection between Uα and Uβ. More precisely, since the overlapping representations in each head of the model become a dominant negative factor when d grows larger, we aim at modelling a manifold with dilated prediction targets in pretraining to avoid convergent head embeddings and collapsing solutions. We use the average values of chart outputs to model a Minkowski sum (Mamatov & Nuritdinov, 2020; Wang et al., 2020), which serves a key purpose in our paradigm. While convergent head embeddings should be avoided, the learning process should not break the convergence entirely. Proposition 1 implies that the Minkowski sum of the output embeddings contains the Minkowski sum of all mappings of intersections, which means that using dilated prediction targets by taking the Minkowski sum does not omit any mapped intersected embedding. Proposition 1. Let U = {U1, U2, ..., Un} be a collection of open subsets of Z whose union is all of Z, and n T i=1 Ui is not empty. For each i {1, 2, ..., n}, there is a homeomorphism ψi : Ui Vi to an open set Vi Rd. We have the Minkowski sum Vi + Vj = {a + b | a Vi, b Vj}. Then Pn i=1 ψi( n T j=1 Uj) Pn i=1 Vi. Proof. For any vector a Pn i=1 ψi( n T j=1 Uj), there exists ai ψi( n T j=1 Uj) such that a = Pn i=1 ai. Because ψi( n T j=1 Uj) Vi, we also have ai Vi, i {1, 2, ..., n}. Then Pn i=1 ai Pn i=1 Vi, a Pn i=1 Vi and thus Pn i=1 ψi( n T j=1 Uj) Pn i=1 Vi. Published as a conference paper at ICLR 2024 Proposition 2 further states that the average Minkowski sum of the output embeddings together with the Minkowski sum of the average of mappings of intersections can be used instead and still keeps Proposition 1 true, under the assumption that each mapping of the intersection is convex. More generally, Proposition 1 holds true with scalar multiplications when convexity is assumed. However, it should be noted that convexity is not guaranteed here. In Eq. 1, we approach this assumption by using an identity mapping I instead of a linear mapping from Korman (2021b). More about the convexity assumption is addressed in Appendix B. Proposition 2. Let U = {U1, U2, ..., Un} be a collection of open subsets of Z whose union is all of i=1 Ui is not empty. For each i {1, 2, ..., n}, there is a homeomorphism ψi : Ui Vi to an open set Vi Rd. The multiplication of set V and a scalar λ is defined to be λV = {λa | a V }. We take the Minkowski sum. If each ψi( n T j=1 Uj) is convex, then Pn i=1 1 nψi( n T n Pn i=1 Vi. Proof. Follows Proposition 1 and the property of scalar multiplication, 1 n Pn i=1 ψi( n T 1 n Pn i=1 Vi. Since scalar multiplication is preserved for convex sets, we have Pn i=1 1 nψi( n T 1 n Pn i=1 Vi. We experiment with our UA paradigm using the SRL algorithm ST-DIM (Anand et al., 2019), and propose DIM-UA. ST-DIM develops on Deep Info Max (DIM)(Hjelm et al., 2018) that uses info NCE (Oord et al., 2018) as the mutual information estimator between patches. Its objective consists of two components. One is the global-local objective (LGL) and the other one is the local-local objective (LLL), defined by Eq. 5 and Eq. 6 respectively. n=1 log exp(gm,n(xt, xt+1)) P xt Xnext exp(gm,n(xt, xt )) (5) n=1 log exp(hm,n(xt, xt+1)) P xt Xnext exp(hm,n(xt, xt )) (6) Here, xt and xt+1 are temporally adjacent observations, whereas Xnext is the set of next observations and xt is randomly sampled from the minibatch. M and N are the height and width of local feature representations. We denote the encoder as f, the output (global feature) vector of input xt as Output(xt), and the local feature vector of xt at point (m, n) as fm,n(xt). Wg and Wh are linear layers that will be discarded in probing. Then, we have the score function gm,n(xt, xt+1) = Output(xt)T Wgfm,n(xt+1), and hm,n(xt, xt+1) = fm,n(xt)T Whfm,n(xt+1) of ST-DIM. For DIM-UA, we need to redefine the score function of LGL by Eq. 7 because the UA paradigm utilizes dilated prediction targets during pretraining, where ψi(f(xt)) is the output of xt from the i-th head following encoder f for each i in N. gm,n(xt, xt+1) = [ 1 i=i ψi(f(xt))]T Wgfm,n(xt+1) (7) Published as a conference paper at ICLR 2024 According to Eq. 4 , we have the MMD objective LQ defined as Eq. 8, where qi(f(xt)) is the membership probability of the i-th head for each i in N when the input is xt. i=1 ((qi(f(xt)) 1 n)2 + (qi(f(xt+1)) 1 Thereby, the UA objective (Eq. 9) is a sum of above objectives, where τ is a hyper-parameter. LUA = LGL + LLL + τLQ (9) 4 EXPERIMENTAL DETAILS The performance of DIM-UA and other SRL methods is evaluated on 19 games of the Atari ARI benchmark. There are five categories of state variables in Atari ARI (Anand et al., 2019), which are agent localization (Agent Loc.), small object localization (Small Loc.), other localization (Other Loc.), miscellaneous (Misc.), and score/clock/lives/display (Score/.../Display). We follow the customary SSL pipeline and record the probe accuracy and F1 scores on the downstream linear probing tasks. The encoder is first pretrained with SSL, and then is used to predict the ground truth of an image with an additional linear classifier. Notably, the weights of the encoder are trained only during the pretraining and are fixed in the probing tasks. The data for pretraining and probing are collected by an RL agent running a certain number of steps using a random policy since it was found that the samples collected by a random policy could be more favorable than those collected by policy gradient policies for SSL methods (Anand et al., 2019). Previous SSL methods in Anand et al. (2019) have used a single output head with 256 hidden units. One of the major interests in our experiment is to discover the effect of choosing different values for the number of dimensions d and the number of charts n. Therefore, we scale up the number of hidden units, while keeping the model architecture, to observe the performance of using a single output head without UA and of using multiple heads with UA. To make a fair comparison, we compare the performance when the total number of hidden units are equal, i.e., when 1 d for a single output head and n d for multiple output heads are equal in our Atari ARI experiment. In contrast, we also modify Sim CLR using our UA paradigm and follow the customs from MSim CLR to compare the performance of different methods with d being equal (Korman, 2021b) in additional experiments on CIFAR10. The experiments are conducted on a single Nvidia Ge Force RTX 2080 Ti and 8-core CPU, using Py Torch-1.7 (Paszke et al., 2019). An illustration of the model backbone, hyper-parameters, and pseudocode of the algorithm are accompanied in Appendix A. In this section, we show the empirical results of our experiments and compare DIM-UA with other SSL methods to verify the efficacy of our UA paradigm. Meanwhile, we pay special attention to the performance of models when choosing different values for n and d. For a straightforward comparison, we first observe the probe F1 scores together with standard deviations of each game averaged across categories in Table 1. ST-DIM* denotes ST-DIM with one output head of 16384 units, while DIM-UA uses four output heads with 4096 units in each head. We compare them to various methods of using a single output head of 256 hidden units here, which are taken from Anand et al. (2019). Each table entry is an average of 5 independent pretraining/probing runs using images sampled from different seeds. The probe accuracy scores are also included in Appendix A. Using 16384 units ( ST-DIM* ) does not necessarily guarantee better performance than using 256 units (ST-DIM). In 7 out of 19 games, ST-DIM* has lower F1 scores than ST-DIM. In particular, the model collapses due to overfitting when using too many units to represent the global features on Freeway. As a result, ST-DIM* only gains an F1 score of 0.3 with standard deviation 0.355 on Freeway. The mean F1 score of ST-DIM* is only 0.7 compared to 0.72 of ST-DIM. On the other Published as a conference paper at ICLR 2024 Table 1: Probe F1 scores of each game averaged across categories Game VAE CPC ST-DIM ST-DIM* DIM-UA Asteroids 0.36 0.42 0.49 0.48 0.005 0.5 0.007 Bowling 0.50 0.90 0.96 0.96 0.021 0.96 0.018 Boxing 0.20 0.29 0.58 0.61 0.008 0.64 0.007 Breakout 0.57 0.74 0.88 0.88 0.02 0.9 0.016 Demon Attack 0.26 0.57 0.69 0.71 0.01 0.74 0.012 Freeway 0.01 0.47 0.81 0.3 0.355 0.86 0.02 Frostbite 0.51 0.76 0.75 0.73 0.005 0.75 0.004 Hero 0.69 0.90 0.93 0.93 0.008 0.94 0.004 Montezuma Revenge 0.38 0.75 0.78 0.81 0.016 0.84 0.014 Ms Pacman 0.56 0.65 0.72 0.74 0.017 0.76 0.011 Pitfall 0.35 0.46 0.60 0.69 0.031 0.73 0.029 Pong 0.09 0.71 0.81 0.78 0.015 0.85 0.004 Private Eye 0.71 0.81 0.91 0.91 0.009 0.93 0.009 Qbert 0.49 0.65 0.73 0.78 0.026 0.79 0.02 Seaquest 0.56 0.66 0.67 0.68 0.007 0.69 0.007 Space Invaders 0.52 0.54 0.57 0.59 0.007 0.62 0.013 Tennis 0.29 0.60 0.60 0.57 0.018 0.64 0.025 Venture 0.38 0.51 0.58 0.57 0.014 0.58 0.01 Video Pinball 0.45 0.58 0.61 0.6 0.031 0.62 0.023 Mean 0.41 0.63 0.72 0.7 0.033 0.75 0.013 ST-DIM DIM-UA Figure 3: The mean F1 and accuracy scores of 19 games when the total number of hidden units varies. The number of heads for DIM-UA is set to 4 here. hand, DIM-UA achieves higher scores and more stable performance. The F1 scores of DIM-UA are equal or higher than those of both ST-DIM* and ST-DIM in every game. The mean F1 score of DIM-UA is 0.75, the highest among all methods in Table 1. 5.1 ABLATIONS In this subsection, our goal is to observe the behavior of models with different settings while the total number of units (n d) on the horizontal axis of figures varies. Fig. 3 examines the difference between ST-DIM and DIM-UA. Fig. 4 examines the effects of changing the number of heads for DIM-UA. In addition, we compare DIM-UA with two methods designed on ST-DIM in Fig. 5. One method uses the paradigm from MSim CLR, denoted by +MMD . The other is similar to ours, which minimizes the loss of Eq. 9, but without modifying the score function as in Eq. 7 (namely, DIM-UA without using dilated prediction targets), denoted by -UA . The 6 games mentioned here are Asteroids, Breakout, Montezuma Revenge, Private Eye, Seaquest, and Video Pinball. Published as a conference paper at ICLR 2024 2 heads 4 heads 8 heads Figure 4: The mean F1 and accuracy scores of DIM-UA on 6 games when the number of output heads is 2, 4, or 8. -UA DIM-UA +MMD Figure 5: The mean F1 and accuracy scores on 6 games with different adaptations. All methods use 4 output heads. In Fig. 3, ST-DIM performs better than DIM-UA when the number of hidden units is small. Their scores become close to each other when the number of units is around 2048. DIM-UA continues to improve as the total number of units grows, whereas the performance of ST-DIM drops at the same time. It is expected for DIM-UA to have lower F1 and accuracy scores when the encoding dimensions are low since the diversity among output heads demands more epochs of training. However, the efficacy of our UA paradigm is clearly demonstrated, as it allows the model to extend its capability by continuously expanding the encoding dimensions. Since we expect DIM-UA to converge slower because of the diversity among output heads, we expect this to become more obvious as the number of heads increases. We can verify that in Fig. 4, where the model with two output heads has the highest F1 and accuracy scores when the total number of hidden units is below 2048. On the other hand, it obtains the lowest F1 and accuracy score when the total number of units grows to 16384. Meanwhile, the model with eight output heads gives the worst results when the number of units is small but shows no sign of plateau, even with very high encoding dimensions. Increasing n while keeping d the same in our UA paradigm helps with the manifold representation but also lowers the performance if d is not large enough. In Fig. 5, it is not surprising that +MMD obtains the worst results in spite of the number of units, since MSim CLR was only found out to be helpful when the number of units is extremely small (e.g., 2, 4). -UA obtains better results than DIM-UA when the number of units is 512 but gets overrun by DIM-UA when the number becomes even larger. This empirically demonstrates that the dilated prediction targets in our UA paradigm are critical to achieve effective manifold representations. Published as a conference paper at ICLR 2024 Table 2: Linear evaluation accuracy on CIFAR10 Method Head Dimension 256 512 1024 Sim CLR - 0.881 0.002 0.883 0.002 0.881 0.003 MSim CLR 2 0.877 0.002 0.878 0.001 0.866 0.003 MSim CLR 4 0.873 0.001 0.873 0.001 0.861 0.002 MSim CLR 8 0.864 0.001 0.859 0.005 0.857 0.002 Sim CLR-UA 2 0.882 0.001 0.884 0.001 0.885 0.001 Sim CLR-UA 4 0.885 0.001 0.884 <0.001 0.88 0.001 Sim CLR-UA 8 0.882 <0.001 0.886 0.002 0.876 0.005 5.2 ADDITIONAL EXPERIMENTS ON CIFAR10 We modify Sim CLR using the UA paradigm (Sim CLR-UA) and perform additional experiments on CIFAR10, following the parameter settings and evaluation protocol from Korman (2021b); Chen et al. (2020). Sim CLR-UA uses multiple heads with dilated prediction targets instead in pretraining, and adds LQ in Eq. 8 to the contrastive loss of Sim CLR. Here, Res Net50 is used as the backbone, which is significantly larger than the default backbone in ST-DIM. We also slightly modify our model here based on an ablation study on CIFAR10. Please check Appendix B for more details. In Table 2, each entry is an average accuracy score obtained from three independent pretraining and evaluation runs, together with the standard deviation. Sim CLR obtains an accuracy score of 88.3% when it uses 512 hidden units. In the mean time, Sim CLR-UA achieves the highest accuracy score of 88.6% among three methods when it uses eight heads with 512 units in each. The second best score is also achieved by Sim CLR-UA, which is 88.5% when using four heads with 256 units in each head, or using two heads with 1024 units in each. We acknowledge that our improvement over Sim CLR is small under this experimental setup. Nonetheless, there is a significant increase in accuracy when comparing Sim CLR-UA to MSim CLR, especially in the case where the number of heads is larger. For instance, the highest evaluation accuracy score of MSim CLR is 87.8%, 87.3% and 86.4% respectively, when using two, four or eight heads. In contrast, Sim CLR-UA obtains the highest accuracy score when using eight heads. This supports our hypothesis that UA can be a universal paradigm to effectively create manifold representations in SSL. 6 DISCUSSION We have demonstrated that our UA paradigm helps improve the performance of both ST-DIM and Sim CLR when encoding dimensions are high. Furthermore, we argue that training NNs with multiple output heads is inherently slower and more demanding than training with a single output head, which has restrained the study in its domain. It is evident that our paradigm can overcome this headwind by generating effective manifold representations. Moreover, our UA paradigm has gained a significant amount of improvement compared to the most related state-of-the-art manifold representation paradigm MSim CLR in the experiments on Atari ARI and CIFAR10. Notably, the UA paradigm also exhibits the potential of modeling a manifold using further higher dimensions while increasing the number of output heads (Fig. 4). It can be an important contribution because this means the performance of the model scales with the size of output heads. Using 16384 hidden units in total is not very efficient economically when the entire model is small, but the additional overhead introduced by doing so can be relatively insignificant when the model itself is large. In particular, this trade-off may also be worthwhile in challenging downstream tasks where the smallest increase in probe accuracy can make a difference. Our work has illustrated that SSL methods with manifolds have great potential, and more topics can be researched in this area. The relationship between the number of hidden units and the number of output heads in an NN model demands more study (see Appendix B for more discussion on this). The convexity assumption is crucial in representing the manifold. Future research may focus on representing a manifold using an unbalanced atlas more efficiently, e.g., designing new objectives and convexity constraints. Published as a conference paper at ICLR 2024 ACKNOWLEDGMENTS This work was performed on the [ML node] resource, owned by the University of Oslo, and operated by the Department for Research Computing at USIT, the University of Oslo IT-department. Giovanni S Alberti, Johannes Hertrich, Matteo Santacesaria, and Silvia Sciutto. Manifold learning by mixture models of vaes for inverse problems. ar Xiv preprint ar Xiv:2303.15244, 2023. Ankesh Anand, Evan Racah, Sherjil Ozair, Yoshua Bengio, Marc-Alexandre Cˆot e, and R Devon Hjelm. Unsupervised state representation learning in atari. Advances in neural information processing systems, 32, 2019. Mikhail Belkin, Partha Niyogi, and Vikas Sindhwani. Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Journal of machine learning research, 7 (11), 2006. Michael M Bronstein, Joan Bruna, Yann Le Cun, Arthur Szlam, and Pierre Vandergheynst. Geometric deep learning: going beyond euclidean data. IEEE Signal Processing Magazine, 34(4):18 42, 2017. Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International conference on machine learning, pp. 1597 1607. PMLR, 2020. Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 15750 15758, 2021. Adrien Ecoffet, Joost Huizinga, Joel Lehman, Kenneth O Stanley, and Jeff Clune. Go-explore: a new approach for hard-exploration problems. ar Xiv preprint ar Xiv:1901.10995, 2019. Michael Gashler, Dan Ventura, and Tony Martinez. Iterative non-linear dimensionality reduction with manifold sculpting. Advances in neural information processing systems, 20, 2007. Daniele Grattarola, Lorenzo Livi, and Cesare Alippi. Adversarial autoencoders with constantcurvature latent manifolds. Applied Soft Computing, 81:105511, 2019. Karol Gregor, Ivo Danihelka, Alex Graves, Danilo Rezende, and Daan Wierstra. Draw: A recurrent neural network for image generation. In International conference on machine learning, pp. 1462 1471. PMLR, 2015. Arthur Gretton, Karsten M Borgwardt, Malte J Rasch, Bernhard Sch olkopf, and Alexander Smola. A kernel two-sample test. The Journal of Machine Learning Research, 13(1):723 773, 2012. Jean-Bastien Grill, Florian Strub, Florent Altch e, Corentin Tallec, Pierre Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent-a new approach to self-supervised learning. Advances in neural information processing systems, 33:21271 21284, 2020. Ji Hun Ham, Daniel D Lee, and Lawrence K Saul. Learning high dimensional correspondences from low dimensional manifolds. International Conference on Machine Learning Workshop, 2003. Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9729 9738, 2020. R Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Phil Bachman, Adam Trischler, and Yoshua Bengio. Learning deep representations by mutual information estimation and maximization. ar Xiv preprint ar Xiv:1808.06670, 2018. Rico Jonschkowski and Oliver Brock. Learning state representations with robotic priors. Autonomous Robots, 39(3):407 428, 2015. Published as a conference paper at ICLR 2024 Teeratorn Kadeethum, Francesco Ballarin, Daniel O malley, Youngsoo Choi, Nikolaos Bouklas, and Hongkyu Yoon. Reduced order modeling for flow and transport problems with barlow twins self-supervised learning. Scientific Reports, 12(1):20654, 2022. Diederik P Kingma and Max Welling. Auto-encoding variational bayes. ar Xiv preprint ar Xiv:1312.6114, 2013. Eric O Korman. Autoencoding topology. ar Xiv preprint ar Xiv:1803.00156, 2018. Eric O Korman. Atlas based representation and metric learning on manifolds. ar Xiv preprint ar Xiv:2106.07062, 2021a. Eric O Korman. Self-supervised representation learning on manifolds. In ICLR 2021 Workshop on Geometrical and Topological Representation Learning, 2021b. Timoth ee Lesort, Natalia D ıaz-Rodr ıguez, Jean-Franois Goudou, and David Filliat. State representation learning for control: An overview. Neural Networks, 108:379 392, 2018. Xiao Liu, Fanjin Zhang, Zhenyu Hou, Li Mian, Zhaoyu Wang, Jing Zhang, and Jie Tang. Selfsupervised learning: Generative or contrastive. IEEE Transactions on Knowledge and Data Engineering, 35(1):857 876, 2021. Alireza Makhzani, Jonathon Shlens, Navdeep Jaitly, Ian Goodfellow, and Brendan Frey. Adversarial autoencoders. ar Xiv preprint ar Xiv:1511.05644, 2015. Mashrabjon Mamatov and Jalolxon Nuritdinov. Some properties of the sum and geometric differences of minkowski. Journal of Applied Mathematics and Physics, 8(10):2241 2255, 2020. Li Meng, Morten Goodwin, Anis Yazidi, and Paal Engelstad. Improving the diversity of bootstrapped dqn by replacing priors with noise. IEEE Transactions on Games, 2022. Junhyuk Oh, Xiaoxiao Guo, Honglak Lee, Richard L Lewis, and Satinder Singh. Action-conditional video prediction using deep networks in atari games. Advances in neural information processing systems, 28, 2015. Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. ar Xiv preprint ar Xiv:1807.03748, 2018. Ian Osband and Benjamin Van Roy. Bootstrapped thompson sampling and deep exploration. ar Xiv preprint ar Xiv:1507.00300, 2015. Ian Osband, Charles Blundell, Alexander Pritzel, and Benjamin Van Roy. Deep exploration via bootstrapped dqn. Advances in neural information processing systems, 29:4026 4034, 2016. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pp. 8024 8035. Curran Associates, Inc., 2019. Nikolaos Pitelis, Chris Russell, and Lourdes Agapito. Learning a manifold as an atlas. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1642 1649, 2013. Sam T Roweis and Lawrence K Saul. Nonlinear dimensionality reduction by locally linear embedding. science, 290(5500):2323 2326, 2000. Joshua B Tenenbaum, Vin de Silva, and John C Langford. A global geometric framework for nonlinear dimensionality reduction. science, 290(5500):2319 2323, 2000. Ilya Tolstikhin, Olivier Bousquet, Sylvain Gelly, and Bernhard Schoelkopf. Wasserstein autoencoders. ar Xiv preprint ar Xiv:1711.01558, 2017. Published as a conference paper at ICLR 2024 Xiangfeng Wang, Junping Zhang, and Wenxing Zhang. The distance between convex sets with minkowski sum structure: application to collision detection. Computational Optimization and Applications, 77:465 490, 2020. Jure Zbontar, Li Jing, Ishan Misra, Yann Le Cun, and St ephane Deny. Barlow twins: Self-supervised learning via redundancy reduction. In International Conference on Machine Learning, pp. 12310 12320. PMLR, 2021. A DETAILS OF DIM-UA Table 3 provides values of some crucial hyper-parameters experimented on Atari ARI that are kept the same across different methods. In addition, τ in Eq. 9 is set to 0.1 for DIM-UA. Fig. 6 illustrates the standard backbone in the Atari ARI experiment. The output values of the last convolutional layer in the backbone are taken as the feature map to get the local feature vector fm,n(xt) of input xt at location (m, n). For the original ST-DIM, a fully connected layer of 256 units immediately follows the backbone. For DIM-UA, a projection head (ψ) and a membership probability head (q) branch from the backbone. Table 3: The values of hyper-parameters on Atari ARI Hyper-parameter Value Image size 160 210 Minibatch size 64 Learning rate 3e-4 Epochs 100 Pretraining steps 80000 Probe training steps 35000 Probe testing steps 10000 Figure 6: An illustration of the standard backbone used by ST-DIM. Pytorch-style pseudocode of the DIM-UA algorithm is provided in Algorithm 1. On a side note, the output of an unbalanced atlas at inference time relies on a single output head, since Eq. 4 moves the membership probability far away from the uniform distribution. As a result, the rest of the output heads does not play a role at inference time. This is different from MSim CLR, which partitions inputs into each head by simultaneously forcing a uniform prior and low entropy on conditional distributions. The role of those remaining output heads in an unbalanced atlas is comparable to the moving average network of BYOL, which produces prediction targets to help stabilize Published as a conference paper at ICLR 2024 Algorithm 1: Pytorch-style pseudocode for DIM-UA. # f: encoder network # fm: f but only up to the last conv layer # pj: projection head # mp: membership probability head # c1, c2: classifier layers only used in pretraining # # B: batch size # N: number of heads # D: number of hidden units # C: local feature map channels # H, W: local feature map height and width # # mean: mean function along a specified dimension # matmul: matrix multiplication # cross entropy: cross entropy loss # mmd: mmd loss # size: get the size along a specified dimension # range: get the range vector of an integer # t: transpose # for xt, xt+1 in loader: # load B samples of xt, xt+1 # get feature maps ot, yt, yt+1 = mean(pj(f(xt)), 1), fm(xt), fm(xt+1) # B D, B H W C # get the membership probabilities qt, qt+1 = mp(f(xt)), mp(f(xt+1)) # B N # get the feature map size sb, sm, sn = size(yt, 0), size(yt, 1), size(yt, 2) # B, H, W # initialize the loss values lossg = 0, lossl = 0 # mmd loss loss = 0.05 (mmd(qt) + mmd(qt+1)) # spatial-temporal loss for m in range(sm): for n in range(sn): # global-local loss logitsg = matmul(c1(ot), yt+1[:, m, n, :].t()) # B B lossg + = cross entropy(logitsg, range(sb)) # cross entropy loss # local-local loss logitsl = matmul(c2(yt[:, m, n, :]), yt+1[:, m, n, :].t()) # B B lossl + = cross entropy(logitsl, range(sb)) # cross entropy loss lossg / = (sm sn), lossl / = (sm sn) # averaged by map size loss + = lossg + lossl # total loss # optimization step loss.backward() optimizer.step() the bootstrap step. The UA paradigm accomplishes a similar goal by using dilated prediction targets of output heads instead. The probe accuracy scores are shown in Table 4, which are overall similar to the F1 scores in Table 1. The accuracy scores of DIM-UA are equal or higher than those of both ST-DIM* and ST-DIM in every game. The mean accuracy score of DIM-UA is 0.76, the highest among all methods. Published as a conference paper at ICLR 2024 Table 4: Probe accuracy scores of each game averaged across categories Game VAE CPC ST-DIM ST-DIM* DIM-UA Asteroids 0.41 0.48 0.52 0.51 0.005 0.53 0.007 Bowling 0.56 0.90 0.96 0.96 0.021 0.96 0.017 Boxing 0.23 0.32 0.59 0.61 0.008 0.64 0.007 Breakout 0.61 0.75 0.89 0.89 0.019 0.91 0.015 Demon Attack 0.31 0.58 0.70 0.72 0.009 0.74 0.011 Freeway 0.07 0.49 0.82 0.33 0.34 0.86 0.017 Frostbite 0.54 0.76 0.75 0.73 0.004 0.75 0.004 Hero 0.72 0.90 0.93 0.93 0.008 0.94 0.004 Montezuma Revenge 0.41 0.76 0.78 0.81 0.015 0.84 0.014 Ms Pacman 0.60 0.67 0.73 0.75 0.016 0.77 0.01 Pitfall 0.35 0.49 0.61 0.7 0.028 0.74 0.027 Pong 0.19 0.73 0.82 0.79 0.014 0.85 0.004 Private Eye 0.72 0.81 0.91 0.91 0.01 0.93 0.009 Qbert 0.53 0.66 0.74 0.79 0.025 0.8 0.019 Seaquest 0.61 0.69 0.69 0.69 0.006 0.7 0.006 Space Invaders 0.57 0.57 0.59 0.6 0.009 0.63 0.014 Tennis 0.37 0.61 0.61 0.58 0.016 0.65 0.024 Venture 0.43 0.52 0.59 0.58 0.012 0.59 0.009 Video Pinball 0.47 0.59 0.61 0.6 0.03 0.63 0.022 Mean 0.46 0.65 0.73 0.71 0.031 0.76 0.013 B MORE ABOUT CIFAR10 EXPERIMENT The convexity assumption is crucial in order to effectively model a manifold (e.g., scalar multiplication is not preserved when a set is non-convex). However, the universal approximation theorem implies that weights of multiple linear layers can approximate any non-convex functions. Thus, whether multiple linear layers should be used here or not could be an interesting ablation topic. Moreover, clamping can be introduced to define open sets in our method. An ablation study of Sim CLR-UA is performed on CIFAR10, using 4 heads with 512 units in each head. The results are shown in Table 5, where FC1 denotes the linear layer immediately following the Res Net50 backbone and FC2 denotes the projection layers following coordinate mappings. The range of clamping is set to ( 10, 10). As the results suggest, the combination of clamping and FC2 yields the best accuracy and is hence used to obtain results in Table 2. Referring to the performance of Sim CLR-UA in Table 2, the accuracy reaches the highest when the number of heads is eight with 512 units in each head, but when the number of heads is four, the optimal number of units is 256. It also appears that a small number of hidden units can be sufficient. This finding is different from what is observed in the Atari ARI experiment, where using eight heads and 2048 units in each head is not sufficient to guarantee the optimal (Fig. 4). This may be attributed to the image size and the number of ground truth labels in an image, and more challenging tasks may demand better representations. However, we do think there should be a limited number of heads needed, related to the intrinsic dimension (ID) of data. Techniques to find ID can potentially be used to decide the optimal number of heads. Table 5: Ablation study on CIFAR10 FC1 Clamp FC2 Accuracy 0.857 0.876 0.883 0.872 0.875 0.884 0.872 Published as a conference paper at ICLR 2024 After observing the results in Table 2, it appears that the performance of Sim CLR-UA could be further enhanced. Whilst comparing MSim CLR with Sim CLR-UA, we maintained most hyperparameters identical for both. However, Sim CLR-UA may attain higher performance given a different set of hyper-parameters. In Fig. 1, the initial entropy of using UA is notably lower than when using a uniform prior. Ideally, it may be advantageous for the entropy to remain high during the initial stages and to decrease gradually. We suggest that the hyper-parameter τ, which regulates the LQ loss, could be set smaller or set to 0 initially and gradually increased over time. We conduct an additional small-scale experiment to validate this hypothesis. In this context, we use Res Net18 as the backbone instead of Res Net50, and the training duration is set at 100 epochs as opposed to 1000. The model incorporates 8 heads, with each head containing 512 hidden units. If τ is linearly scaled, it would increment linearly, from zero up to its final value over the pretraining epochs. The outcome of this experiment is detailed in Table 6. The table clearly illustrates that implementing linear scaling or utilizing smaller τ values can genuinely enhance the performance of Sim CLR-UA. For τ values of 0.1 and 0.2, adopting a linearscaling scheme is instrumental for optimizing the performance. However, for small τ values of 0.05 and 0.02, such a scheme is not needed. Thus, the performance of Sim CLR-UA, as presented in Table 2, could potentially be boosted further, since it uses a relatively large τ value of 0.1 without any linear scaling. Table 6: Changing τ in Sim CLR-UA τ Linear scaling Accuracy 0.2 0.791 0.2 0.797 0.1 0.791 0.1 0.8 0.05 0.799 0.05 0.796 0.02 0.802 0.02 0.785