# time_to_augment_selfsupervised_visual_representation_learning__7ae22b6a.pdf Published as a conference paper at ICLR 2023 TIME TO AUGMENT SELF-SUPERVISED VISUAL REPRESENTATION LEARNING Arthur Aubret1 Markus R. Ernst2 Céline Teulière1 Jochen Triesch2 1Clermont Auvergne Université, CNRS, Clermont Auvergne INP, Institut Pascal 2Frankfurt Institute for Advanced Studies {arthur.aubret, celine.teuliere}@uca.fr {mernst, triesch}@fias.uni-frankfurt.de Biological vision systems are unparalleled in their ability to learn visual representations without supervision. In machine learning, self-supervised learning (SSL) has led to major advances in forming object representations in an unsupervised fashion. Such systems learn representations invariant to augmentation operations over images, like cropping or flipping. In contrast, biological vision systems exploit the temporal structure of the visual experience during natural interactions with objects. This gives access to augmentations not commonly used in SSL, like watching the same object from multiple viewpoints or against different backgrounds. Here, we systematically investigate and compare the potential benefits of such time-based augmentations during natural interactions for learning object categories. Our results show that incorporating time-based augmentations achieves large performance gains over state-of-the-art image augmentations. Specifically, our analyses reveal that: 1) 3-D object manipulations drastically improve the learning of object categories; 2) viewing objects against changing backgrounds is important for learning to discard background-related information from the latent representation. Overall, we conclude that time-based augmentations during natural interactions with objects can substantially improve self-supervised learning, narrowing the gap between artificial and biological vision systems. 1 INTRODUCTION Learning object representations without supervision is a grand challenge for artificial vision systems. Recent approaches for visual self-supervised learning (SSL) acquire representations invariant to data-augmentations based on simple image manipulations like crop/resize, blur or color distortion (Grill et al., 2020; Chen et al., 2020). The nature of these augmentations determines what information is retained and what information is discarded, and therefore how useful these augmentations are for particular downstream tasks (Jaiswal et al., 2021; Tsai et al., 2020). Biological vision systems, in contrast, appear to exploit the temporal structure of visual input during interactions with objects for unsupervised representation learning. According to the slowness principle (Wiskott and Sejnowski, 2002; Li and Di Carlo, 2010; Wood and Wood, 2018), biological vision systems strive to discard high-frequency variations (e.g. individual pixel intensities) and retain slowly varying information (e.g. object identity) in their representation. Incorporating this idea into SSL approaches has led to recent time-contrastive learning methods that learn to map inputs occurring close in time onto close-by latent representations (Oord et al., 2018; Schneider et al., 2021). How these systems generalize depends on the temporal structure of their visual input. In particular, visual input arising from embodied interactions with objects may lead to quite different generalizations compared to what is possible with simple image manipulations. For instance, human infants learning about objects interact with them in various ways (Smith et al., 2018). First, infants rotate, bring closer/farther objects while playing with them (Byrge et al., 2014). Second, as they gain mobility, they can move in the environment while holding an object, viewing it in different contexts and against different backgrounds. We refer to (simulations of) such interactions as natural interactions. Equal contribution. Published as a conference paper at ICLR 2023 Here, we systematically study the impact of such natural interactions on representations learnt through time-contrastive learning in different settings. We introduce two new simulation environments based on the near-photorealistic simulation platform Three DWorld (TDW) (Gan et al., 2021) and combine them with a recent dataset of thousands of 3D object models (Toys4k) (Stojanov et al., 2021). Then we validate our findings on two video datasets of real human object manipulations, Toy Box (Wang et al., 2018) and CORe50 (Lomonaco and Maltoni, 2017). Our experiments show that adding time-based augmentations to conventional data-augmentations considerably improves category recognition. Furthermore, we show that the benefit of time-based augmentations during natural interactions has two main origins. First, 3-D object rotations boost generalization across object shapes. Second, viewing objects against different backgrounds while moving with them reduces harmful effects of background clutter. We conclude that exploiting natural interactions via time-contrastive learning greatly improves self-supervised visual representation learning. 2 RELATED WORK Data-augmented self-supervised learning. The general idea behind most recent approaches for self-supervised learning is that two semantically close/different inputs should be mapped to close/distant points in the learnt representation space. Applying a certain transformation to an image such as flipping it horizontally generates an image that will be very different at the pixel level, but has a similar semantic meaning. A learning objective for SSL will therefore try to ascertain that the representations of an image and its augmented version are close in latent space, while being far from the representations of other unrelated images. A concrete approach may work as follows: sample an image x, apply transformations to it taken from a predefined set of transformations, resulting in new images x , also called a positive pair. The same procedure is applied to a batch of different images. Embeddings of positive pairs are brought together while keeping the embeddings of the batch overall distant from one another. There are three main categories of approaches for doing so: contrastive learning methods (Chen et al., 2020; He et al., 2020) explicitly push away the embeddings of a batch of inputs from one another; distillation-based methods (Grill et al., 2020; Chen and He, 2021) use an asymmetric embedding architecture, allowing the model to discard the push away part; entropy maximization methods (Bardes et al., 2022; Ermolov and Sebe, 2020) maintain a high entropy in the embedding space. Image manipulations as data-augmentations. Most self-supervised learning approaches have used augmentations based on simple image manipulations to learn representations. Frequently used are color distortion, cropping/resizing a part of an image, the horizontal flipping of an image, gray scaling the image, and blurring the image (Chen et al., 2020). Other augmentations can be categorized in three ways (Shorten and Khoshgoftaar, 2019; Jaiswal et al., 2021): 1) geometric augmentations include image rotations (Chen et al., 2020) or image translations (Shorten and Khoshgoftaar, 2019); 2) Context-based augmentations include jigsaw puzzle augmentations (Noroozi and Favaro, 2016; Misra and Maaten, 2020), pairing images (Inoue, 2018), greyed stochastic/saliency-based occlusion (Fong and Vedaldi, 2019; Zhong et al., 2020), or automatically modifying the background (Ryali et al., 2021); 3) Color-oriented transformations can be the selection of color channels (Tian et al., 2020) or Gaussian noise (Chen et al., 2020). A related line of work also proposes learning how to generate/select data-augmentations (Cubuk et al., 2019; Tian et al., 2020), but since it takes advantage of labels, the approach is no longer self-supervised. Time-based data-augmentations. Several works have proposed using the temporality of interactions to learn visual representations. A recent line of work proposes to learn embeddings of video frames using the temporal contiguity of frames: Knights et al. (2021) propose a learning objective that makes codes of adjacent frames within a video clip similar, however the system still needs to have information about where each video starts and ends. In contrast, our setups expose the system to a continuous stream of visual inputs. Other methods showed the importance of time-based augmentations based on videos for object tracking (Xu and Wang, 2021), category recognition (Gordon et al., 2020; Parthasarathy et al., 2022; Orhan et al., 2020) or adversarial robustness (Kong and Norcia, 2021). Unlike us they do not make an in-depth analysis of the impact of different kinds of natural interactions. Schneider et al. (2021) showed the importance of natural interactions with objects for learning object Published as a conference paper at ICLR 2023 Figure 1: Top view of the VHE used to situate our agent in a house. Blue triangles/orange circles show possible agent/object positions, respectively. In this episode, the agent is located in the office (blue arrow) and interacts with a plush toy. Successive first person views of object interactions are shown in the lower left. representations, but only considered small datasets of few objects, without complex backgrounds and without studying generalization over categories of objects. Wang et al. (2021) have shown that one can replace crop augmentations by saccade-like magnifications of images, establishing a link between conventional image-based and more natural time-based augmentations. Aubret et al. (2022) studied the impact of embodiment constrains on object representations learnt through time-contrastive learning during natural interactions. They highlight that complex backgrounds negatively impact the representations of objects. Here we show that allowing the agent to move with objects so that they are seen against changing backgrounds improves robustness against background complexity. We note that a large body of work also considers time-contrastive learning in the context of reinforcement learning (Oord et al., 2018; Laskin et al., 2020; Stooke et al., 2021; Okada and Taniguchi, 2021) and intrinsic motivation (Guo et al., 2021; Yarats et al., 2021; Li et al., 2021; Aubret et al., 2021). Our goal is to study the potential of time-based augmentations during natural interactions with objects for visual representation learning and to evaluate the utility of the learned representations for object categorization.1 We focus on two kinds of natural interactions: object manipulations (3D translations and rotations) and ego-motion. We introduce two computer-rendered environments, the Virtual Home Environment (VHE) and the 3D Shape Environment (3DShape E) to compare and disentangle the effects of these different types of natural interactions. We also validate our findings using two real-world first-person video datasets. The Toy Box environment (Wang et al., 2018) permits studying the relative benefits of object translations and rotations. The CORe50 environment (Lomonaco and Maltoni, 2017) allows us to explore the effects of smooth visual tracking and to mimic the effects of ego-motion. 3.1 VIRTUAL HOME ENVIRONMENT To simulate a diverse set of natural interactions of an infant in its home, we take inspiration from (Aubret et al., 2022) and place an agent within an environment that resembles a residential house, where it can play with objects and change positions. Figure 1 depicts a bird s eye view showing the floor plan of the aforementioned house. In comparison with Aubret et al. (2022), we integrate a two order of magnitude higher number of fully textured toy objects, as well as novel interactions with these objects. The agent is spawned in a randomly sampled location, staying there for 100-timestep-long 1The source code is available at https://github.com/trieschlab/Time To Augment SSL Published as a conference paper at ICLR 2023 Figure 2: Examples of time-based augmentations while playing in the VHE (left, A-C): A) object rotation; B) ego-motion; C) depth change. CORe50 (right, D-E): D) close-by frames of objects in 2 different recording sessions; E) different exemplars of the object category cellphones shown in different contexts. Toy Box (left, F-H): F) depth change, G) lateral translation, H) object rotation. episodes (or sessions). Its position in the world determines what background will be seen. Each background contains a unique combination of static objects and floor/wall textures. Toys4k dataset. The agent interacts with objects from the Toys4k dataset, which is composed of 4,179 diverse 3-D toy objects distributed in an unbalanced way into 105 categories (Stojanov et al., 2021). It was designed to match the objects typically encountered by infants. We place textured objects in different locations of the full house. We modify/remove 3D models, only keeping quick-to-process and correct-quality objects in our simulations (cf. Appendix C.1 for more details). To study the ability of our agent to generalize over categories, we use two thirds of the objects of each category for training and keep the last third for testing. We construct a test set composed of 5 randomly rotated views of the test objects in different locations of the house. Except for rotation, no additional transformations have been applied to build the test set. We also introduce 3 novel houses to test the robustness of the object-representation with respect to novel out-of-distribution backgrounds (cf. Appendix C.2). Object manipulations. At only a few months old, infants start to hold objects and move and rotate them (Byrge et al., 2014). Similarly, our agent can manipulate the objects in two different ways. First, it can rotate the object around the vertical axis (yaw) by a random angle drawn uniformly from the interval [0, rot] degrees, where the hyperparameter rot sets the maximal speed of object rotations (cf. Figure 2A). Second, the agent can also bring the object closer or move it further away, as shown in Figure 2C; we model this as a change of distance uniformly sampled from [ d, d], where d = 7.5 cm is the maximal distance change. The distance is bounded between 0.65 m and 1.1 m to reflect morphological limitations of the agent s arms. Ego-motion. When infants become mobile, they can play with objects while moving in their environment. To simulate this, our agent can turn its body towards a neighboring location while holding the same object so that it is seen against a different background. As shown in Figure 2B, the new background will usually have similar floors and walls. Sometimes an infant will also engage with a new object. We model both possibilities by deterministically turning the agent every Ns time steps and switching to a new object with probability 1 po = 0.1. A small/medium Ns respectively simulates in an unnatural/natural fashion an agent that always/sometimes moves with its object. When this ego-motion is disabled, the agent simultaneously picks up a new object and turns every Ns = 10 time steps. For some experiments (Table 1 and Figure 5A, B), we also allow the agent to visit another room every Ns steps while continuing to interact with the same object. This allows us to investigate the effect of more drastic background changes, but may be a less valid model of a human infant s interaction with objects. 3.2 3D SHAPE ENVIRONMENT In order to study the influence of natural interactions on shape generalization while minimizing the impact of object colors or background, we introduce a simple 3D Shape Environment (3DShape E). It differs from the VHE in two respects: 1) we import untextured versions of toy objects from the Toys4k dataset, making them appear grey; 2) we place the agent in an empty environment so that objects are seen against a blank background. Examples of inputs are given in Figure 4E and augmentations are the same as in the VHE environment (Figure 2A-C). Published as a conference paper at ICLR 2023 3.3 TOYBOX ENVIRONMENT The Toy Box dataset (Wang et al., 2018) contains first-person videos of an observer manipulating 360 objects (12 categories, 30 objects each). In each clip, a person applies one type of transformation: a translation (x, y or z), rotation (around x, y or z axis), a hodgepodge, nothing specific, or in some cases, nothing if there is no object. We sample two frames per second from each clip (cf. Appendix C.3 for more details). Two successively sampled frames form a positive pair. At the end of a clip, a positive pair is formed from the last frame sampled from the clip and the first image of the randomly sampled next clip. 3.4 CORE50 ENVIRONMENT The CORe50 dataset (Lomonaco and Maltoni, 2017) contains first-person views of an observer manipulating objects. The dataset comprises 167,866 images of 50 different objects belonging to 10 distinct object categories. All objects were filmed in 11 sessions corresponding to different natural environments. The images from all but one object per class are used to form the training and validation sets. Specifically, every 10th image enters the validation set, the others form the training set. The images of the held-out object of each class form the test set. This allows us to test generalization to unknown objects from familiar categories. The videos in this dataset show real object manipulations, but we also use it to mimic the effects of ego-motion through the way we sample images from the different video clips. In general, successively sampled images form positive pairs. We create sequences of images by starting from a particular frame of one of the video clips showing an object manipulation and defining a probability po that the manipulation of the same object continues. Otherwise an interaction with a new object will start. In case the interaction continues, the next frame in our sampled sequence will be randomly chosen as either the next or the previous frame in the video clip. Thus, as long as the object manipulation continues, the sampled sequence of frames corresponds to a random walk through the frames of the video. A high/low po leads to long/short random walks. Specifically, the expected number of successive frames No showing the same object is given by: E [No|po] = 1 + x=0 xpx o(1 po) = 1 + po 1 po . (1) To mimic the effects of ego-motion causing an object to be seen against changing backgrounds, we also define a probability ps that the sequence of views stays within the same session, i.e., the same video clip. Thus, there is a chance of 1 ps that the sequence of views of the same object jumps to a different video clip showing the same object in a different context. In this case, a positive pair is generated from two images of the same object recorded in different recording sessions. This cross session sampling (CSS) is analogous to the role of ego-motion in the VHE. As an alternative to the random walk procedure described above we also consider a variant where the next view of the same object is chosen uniformly at random from the current video clip. This leads to images forming a positive pair, which stem from distant time points in the original video clip. While this is not a realistic model of viewing sequences experienced by an infant, it allows us to assess the benefits of positive pairs containing more distinct views of the same object. This is analogous to greater speed of object rotation in the VHE. For implementation details of our sampling procedure see Appendix B.5. In brief, it is an extension of the procedure used in Schneider et al. (2021) to allow CSS. 3.5 EVALUATION AND TRAINING PROCEDURE FOR ALL ENVIRONMENTS Learning algorithms. We consider three SOTA representatives of existing algorithms for dataaugmented SSL: Sim CLR for contrastive learning (Chen et al., 2020), BYOL (Grill et al., 2020) for distillation-based methods and VICReg (Bardes et al., 2022) for entropy maximization methods. We refer to the versions of these algorithms using time-based augmentations as, e.g., Sim CLR through time (Sim CLR-TT), etc. (Schneider et al., 2021). Published as a conference paper at ICLR 2023 Source images Figure 3: Sequences of input images during object rotations (rot(360)) and ego-motion and their augmented positive pair according to the augmentation method (same column). Baseline refers to conventional augmentations; *-TT refers to time-based augmentations; *-TT+ refers to combined augmentations. Positive pair construction. We consider three ways to augment an image. First, we consider time-based augmentations during natural interactions (Sim CLR-TT, etc.). For this, unless stated otherwise, we set the positive pair of a source image as the next image in the temporal sequence. Second, as a comparison baseline, we augment a sampled image with a widely adopted SOTA set of augmentations (Chen et al., 2020; Shijie et al., 2017), i.e. crop/resize, grayscale, color jittering and horizontal flip with their default parameters, except for the crop/resize augmentation (cf. Appendix D). Third, we propose to combine the two kinds of augmentations (Sim CLR-TT+, etc.). For this we apply the conventional SOTA augmentations to the image that temporally follows the sampled source image. We refer to Figure 3 for examples of the different ways to augment source images. Controlling for the number of augmentations. While we can generate an arbitrary number of positive pairs from one sample image with conventional augmentations, we only have access to one positive time-based pair per sample image for the VHE, 3DShap E, and the Toy Box environment. To control for the impact of the number of positive pairs, we therefore only construct one positive pair per sample for all algorithms and augmentation methods: 1) for the Toy Box environment, we sample positive pairs from a pre-built dataset of the same size as the original one, but composed of conventionally augmented images; 2) for the VHE and 3DShape E the agent collects images, augments them and stores them in a buffer of size 100,000. It simultaneously trains on images randomly sampled from the buffer at every time step. We show in Appendix B.1 that controlling for the number of augmentations does not impact the results of our analysis. Controlling for the number/distribution of source images. We want to make sure that we compare the quality of the augmentations, and not the amount/diversity of data that feeds the SSL algorithms. Thus, for the comparison baseline, we consider the same natural interactions to generate source images, but we do not use temporal information to generate the positive pairs. This ensures that the number/distribution of non-augmented images is exactly the same across methods (source images in Figure 3). We make an exception with depth changes in 3DShape E since we found it to be harmful for the baselines. Evaluation. We compute the representation with a Res Net18 on Toy Box and CORe50 and with a simpler convolutional neural network on 3DShape E and VHE. We evaluate the learnt representation after 480,000 steps for the VHE and 3DShape E, 60 epochs for Toy Box and 100 epochs for CORe50. The quality of the learned representation is assessed by training linear readouts on top of the learned representation in a supervised fashion (Chen et al., 2020). Depending on the question, we train linear readouts to either predict the identity of an object, the category of an object, or the location/session of the object. We average all results over 3 random seeds. Because of the small number of objects per category in the CORe50 dataset, we apply a cross-validation with 5 splits. Hyperparameters are given in Appendix D. Published as a conference paper at ICLR 2023 Method VHE 3DShape E Toy Box CORe50 Sim CLR 0.328 0.002 0.527 0.005 0.313 0.001 0.544 0.074 Sim CLR-TT 0.369 0.005 0.598 0.007 0.189 0.009 0.449 0.152 Sim CLR-TT+ 0.537 0.001 0.621 0.001 0.378 0.005 0.610 0.090 BYOL 0.352 0.017 0.496 0.012 0.354 0.007 0.552 0.063 BYOL-TT 0.369 0.007 0.570 0.001 0.213 0.014 0.448 0.160 BYOL-TT+ 0.516 0.002 0.536 0.001 0.393 0.031 0.614 0.128 VICReg 0.169 0.006 0.512 0.002 0.381 0.019 0.410 0.073 VICReg-TT 0.264 0.012 0.383 0.167 0.158 0.008 0.245 0.084 VICReg-TT+ 0.435 0.013 0.574 0.008 0.418 0.016 0.503 0.057 Table 1: Top 1 accuracy standard deviation (over 3 seeds, CORe50 over 5 training/testing splits) under linear evaluation of previous SOTA (BYOL, Sim CLR, VICReg) versus our time-based augmentations (*-TT) and their combination (*-TT+). In 3DShape E, *-TT+ use crop/resize, color jittering and rotations. In VHE, *-TT+ use rotations, Ns = 1 with room changes, color jittering and gray scaling. In both VHE and 3D Shape Environments, *-TT use rotations, Ns = 1 with room changes (VHE only) and depth changes. In CORe50 we use ps = 0.5, po = 0.9. We refer to Appendix B.2 for the ablation study that motivates these choices. In Toy Box, we always apply all their respective transformations. Unlike *-TT+, we apply all standard augmentations for the baseline, as we found that removing some of them (crop/resize and flip) was harmful. In this section, we first study the benefits of using time-based augmentations during natural interactions with objects for self-supervised visual representation learning. Then, we provide an in-depth analysis of the impact of 3-D object rotations and ego-motion on the learned representations. An ablation study is performed in Appendix B. 4.1 TIME-BASED AUGMENTATIONS DERIVED FROM NATURAL INTERACTIONS BOOST PERFORMANCE OF SSL In Table 1, we compare the quality of the learnt representation when using conventional dataaugmentations (Sim CLR, BYOL, VICReg), time-based augmentations (*-TT), and their combination (*-TT+) in our four test environments. We observe that our combined approach (*-TT+) increases the average test category accuracy with respect the the baselines (Sim CLR, BYOL, VICReg) by a range of [0.037; 0.266] points across all environments and all SSL methods. To formally validate our approach, we ran additional experiments for Sim CLR and Sim CLR-TT+ and applied a t-test to statistically compare their performance, supporting the superiority of Sim CLR-TT+ (p < 0.05 in our four test environments). Time-based augmentations during natural interactions (*-TT) do not perform very well on their own, presumably because they do not create invariance to color/grayscale changes in our test environments (cf. Appendix B for ablation studies of augmentations). In the next sections, we analyse the impact of 3-D rotations and ego-motion on the learnt representations. 4.2 OBJECT MANIPULATIONS INCLUDING 3-D ROTATIONS SUPPORT SHAPE GENERALIZATION To assess the importance of object manipulations to generalize over object shapes, we conducted experiments in 3DShape E. Figure 4A shows that increasing the speed of object rotations systematically increases the category recognition accuracy. Similarly, in the CORe50 Environment (Figure 4D) the uniform sampling of views clearly outperforms the default random walk sampling procedure, in particular for short object manipulations (low po). Interestingly, we observe the inverse effect on individual object recognition accuracy (Figure 4B). We conclude that fast object rotations (in the extreme case constructing positive pairs from randomly rotated views of an object) tend to discard object-specific details and focus the representation on more general shape features, which will be similar for different objects of the same category. Published as a conference paper at ICLR 2023 0.0 0.1 0.2 0.3 Million steps rot(10) rot(30) rot(90) rot(360) 3DShape E category acc. A 0.0 0.1 0.2 0.3 Million steps rot(10) rot(30) rot(90) rot(360) 3DShape E object acc. B Sim CLR TT+ Sim CLR Method Front views Side views 3/4 views Majority 3DShape E category acc., single single C Figure 4: A) 3DShape E test category recognition accuracy for different maximal rotation speeds S, without other augmentations. B) 3DShape E test object recognition accuracy for different maximum rotation speeds, without other augmentations. The shaded area displays the +/- standard deviation of test category accuracy over seeds. C) Test category accuracy with a linear classifier trained on one single orientation of objects, and tested on the same orientation of test objects. We used 5 seeds, all our conventional augmentations but only rotations ([0, 360] degrees) as time-based augmentation. We also apply the majority vote on datasets of six different orientations: 0 , 10 , -45 , 80 , -80 , -90 . D) Category accuracy for the CORe50 Environment when varying the duration of object manipulations (ps = 0.95). The shaded area shows standard error. A-D) We used Sim CLR-TT. E) Example of three views for eight objects from four categories in 3DShape E. In order from left to right for each object, we show orientations of: 0 (side view); -45 (3/4 view); 90 (back view). The red rectangles highlight visually similar views of different objects. Finally, we investigate the relation between rotation invariance and object shape. To do so, we design a new test setup. In the single-single setup, we train the linear classifier on the representations of a single-view dataset that contains only one view from a common orientation for each training object (back, side, . . . ). We assess the classifier on images of the test objects shown in the same orientation. This means that the linear classifier can not learn the invariance to rotation and does not explicitly need it for classification. We consider datasets of orientations of -90 (front view), -45 (3/4 view), 0 (side view) as they contain the extreme cases in terms of the amount of visible surface (Blanz et al., 1999). We hypothesize that including 3-D rotations in time-based augmentations allows the model to encode 3-D shape features in the learned representation, which results in an improved category recognition. To test this hypothesis, we first evaluate category recognition based on several views of an object (thus its whole shape). In Figure 4C, we compute a majority vote among classification results from the different single-single classifiers as follows: if more views (e.g., 90 , 80 , 10 ) are classified as category A than category B (e.g., 0 , -90 ), category A is selected. We observe a large increase of accuracy for Sim CLR. We conclude that encoding the 3-D shape improves category recognition. Importantly, we do not observe such an increase of accuracy for Sim CLR-TT+ with the majority vote. We deduce that Sim CLR-TT+ better infers the 3-D shape of the object from one view, which strongly improves category recognition. 4.3 MOVING WITH OBJECTS MITIGATES THE EFFECT OF BACKGROUND CLUTTER To evaluate the impact of ego-motion while holding objects, we trained our agent in the VHE with different levels of motion. In Figure 5A, we see that increasing the frequency 1/Ns and the range (room) of motion considerably boosts the test category accuracy in backgrounds from training and novel houses. At the same time, we observe a decrease of the accuracy when classifying the background (Figure 5B). We conclude that moving while holding objects improves the representation s invariance with respect to the background. Moreover, we analysed the frequency of session changes compared to the duration of object manipulations as one of the hyperparameters of our approach in the CORe50 Environment, see Figure 5C, Published as a conference paper at ICLR 2023 Test houses Training house N = 10 N = 5 N = 1 N = 1, room change VHE category acc. A Test houses Training house VHE background acc. B Figure 5: A and B. Impact of the frequency (Ns) of body motions and whether they change rooms on the final test category accuracy (A) and the final background accuracy (B). We also used object rotations ([0, 360] degrees), color jittering and gray scaling. Vertical bars indicate one standard deviation. C and D. Mean CORe50 accuracy for category (C) and session (D) classification dependent on the duration of object manipulations (No) and the frequency of session switches (Ns). Sim CLRTT, random walk procedure, mean based on 3 training/testing splits. D. Qualitatively, for category classification we observe low accuracies for high-frequency object switches, low session changes and high accuracies for low-frequency object switches, high session changes. Accuracy seems to more or less saturate at the main diagonal. We infer from this that intermediate values may actually be better than going to the extreme, but it could also be partly because Ns is bounded by No. The classification of sessions consistently shows that session information is being disregarded when the frequency of session changes rises. 5 CONCLUSION We investigated the benefits of time-based data-augmentations during infant-inspired natural object interactions for self-supervised visual representation learning. We studied these interactions through a novel 3-D simulation environment where an agent plays with objects in a house, and we validated our findings on two real-world video datasets of object manipulations. We find that combining timebased augmentations with conventional data-augmentations (*-TT+ algorithms) greatly improves the ability of the learned representations to generalize over object categories. Our analysis shows that 1) 3-D object rotations are crucial to build good representations of shape categories. 2) Ego-motion while holding an object so that it is seen against different backgrounds prevents the background from cluttering the learned representation. Our work raises the question whether time-based augmentations during natural interactions with objects could fully replace conventional ones. Indeed, we showed that 3-D rotations are superior to the flip data-augmentation for shape generalization (Appendix B.3). However, we found that color-based transformations like color jittering and grayscaling remain important. It is an interesting question for future work if time-based augmentations during changes of direct and indirect lighting or shadows cast by other objects could replace augmentations based on simple image color manipulations. Our work has a number of limitations. First, while we have portrayed the time-based augmentations as stemming from object manipulations, we did not render a hand holding the object or similar in the VHE and 3DShape E. Second, the CORe50 and Toy Box environments are only medium-sized. Large scale experiments may give additional insights. Finally, in this work, we only considered naive behavioural strategies for, e.g., turning objects or ego-motion. This contrasts with the way infants learn about their environment (Bambach et al., 2016). For example, they are biased towards creating planar views of objects and often prefer unfamiliar objects (Roder et al., 2000). Thus, we expect that learning to actively select successive views will further unveil the potential of time-based augmentations. ACKNOWLEDGEMENTS This work was sponsored by a public grant overseen by the French National Agency through the IMob S3 Laboratory of Excellence (ANR-10-LABX-0016) and the IDEX-ISITE initiative CAP 20-25 (ANR-16-IDEX-0001). Financial support was also received from Clermont Auvergne Metropole through a French Tech-Clermont Auvergne professorship. We gratefully acknowledge support Published as a conference paper at ICLR 2023 from GENCI IDRIS (Grant 2022-AD011011623R1) for providing computing and data-processing resources needed for this work. Additional support was received by the Deutsche Forschungsgemeinschaft (DFG project Abstract REpresentations in Neural Architectures (ARENA) ), as well as the projects The Adaptive Mind and The Third Wave of Artificial Intelligence funded by the Excellence Program of the Hessian Ministry of Higher Education, Science, Research and Art (HMWK). JT was supported by the Johanna Quandt foundation. Aubret, A., Hassas, S., et al. (2021). Distop: Discovering a topological representation to learn diverse and rewarding skills. ar Xiv preprint ar Xiv:2106.03853. Aubret, A., Teulière, C., and Triesch, J. (2022). Toddler-inspired embodied vision for learning object representations. In IEEE 21st International Conference on Development and Learning (ICDL). Bambach, S., Crandall, D. J., Smith, L. B., and Yu, C. (2016). Active viewing in toddlers facilitates visual object learning: An egocentric vision approach. In Proceedings of the 38th Annual Conference of the Cognitive Science Society, pages 1631 1636. Cognitive Science Society. Bardes, A., Ponce, J., and Lecun, Y. (2022). VICReg: Variance-invariance-covariance regularization for self-supervised learning. In Proceedings of the 10th International Conference on Learning Representations (ICLR). Blanz, V., Tarr, M. J., and Bülthoff, H. H. (1999). What object attributes determine canonical views? Perception, 28(5):575 599. Byrge, L., Sporns, O., and Smith, L. B. (2014). Developmental process emerges from extended brain body behavior networks. Trends in Cognitive Sciences, 18(8):395 403. Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. (2020). A simple framework for contrastive learning of visual representations. In III, H. D. and Singh, A., editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 1597 1607. PMLR. Chen, X. and He, K. (2021). Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 15750 15758. Cubuk, E. D., Zoph, B., Mane, D., Vasudevan, V., and Le, Q. V. (2019). Autoaugment: Learning augmentation strategies from data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Ermolov, A. and Sebe, N. (2020). Latent world models for intrinsically motivated exploration. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 5565 5575. Curran Associates, Inc. Fong, R. and Vedaldi, A. (2019). Occlusions for effective data augmentation in image classification. In 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), pages 4158 4166. Gan, C., Schwartz, J., Alter, S., Mrowca, D., Schrimpf, M., Traer, J., De Freitas, J., Kubilius, J., Bhandwaldar, A., Haber, N., et al. (2021). Threedworld: A platform for interactive multi-modal physical simulation. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 1). Gordon, D., Ehsani, K., Fox, D., and Farhadi, A. (2020). Watching the world go by: Representation learning from unlabeled videos. ar Xiv preprint ar Xiv:2003.07990. Grill, J.-B., Strub, F., Altché, F., Tallec, C., Richemond, P., Buchatskaya, E., Doersch, C., Avila Pires, B., Guo, Z., Gheshlaghi Azar, M., Piot, B., kavukcuoglu, k., Munos, R., and Valko, M. (2020). Bootstrap your own latent - a new approach to self-supervised learning. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 21271 21284. Curran Associates, Inc. Published as a conference paper at ICLR 2023 Guo, Z. D., Azar, M. G., Saade, A., Thakoor, S., Piot, B., Pires, B. A., Valko, M., Mesnard, T., Lattimore, T., and Munos, R. (2021). Geometric entropic exploration. ar Xiv preprint ar Xiv:2101.02055. He, K., Fan, H., Wu, Y., Xie, S., and Girshick, R. (2020). Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Inoue, H. (2018). Data augmentation by pairing samples for images classification. ar Xiv preprint ar Xiv:1801.02929. Jaiswal, A., Babu, A. R., Zadeh, M. Z., Banerjee, D., and Makedon, F. (2021). A survey on contrastive self-supervised learning. Technologies, 9(1). Knights, J., Harwood, B., Ward, D., Vanderkop, A., Mackenzie-Ross, O., and Moghadam, P. (2021). Temporally coherent embeddings for self-supervised video representation learning. In 2020 25th International Conference on Pattern Recognition (ICPR), pages 8914 8921. IEEE. Kong, N. and Norcia, A. (2021). Are models trained on temporally-continuous data streams more adversarially robust? In SVRHM 2021 Workshop @ Neur IPS. Laskin, M., Srinivas, A., and Abbeel, P. (2020). Curl: Contrastive unsupervised representations for reinforcement learning. In International Conference on Machine Learning, pages 5639 5650. PMLR. Li, N. and Di Carlo, J. J. (2010). Unsupervised natural visual experience rapidly reshapes size-invariant object representation in inferior temporal cortex. Neuron, 67(6):1062 1075. Li, S., Zheng, L., Wang, J., and Zhang, C. (2021). Learning subgoal representations with slow dynamics. In International Conference on Learning Representations (ICLR). Lomonaco, V. and Maltoni, D. (2017). Core50: a new dataset and benchmark for continuous object recognition. In Conference on Robot Learning, pages 17 26. PMLR. Loshchilov, I. and Hutter, F. (2018). Decoupled weight decay regularization. In International Conference on Learning Representations (ICLR). Misra, I. and Maaten, L. v. d. (2020). Self-supervised learning of pretext-invariant representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Noroozi, M. and Favaro, P. (2016). Unsupervised learning of visual representations by solving jigsaw puzzles. In European Conference on Computer Vision (ECCV), pages 69 84. Springer International Publishing. Okada, M. and Taniguchi, T. (2021). Dreaming: Model-based reinforcement learning by latent imagination without reconstruction. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 4209 4215. IEEE. Oord, A. v. d., Li, Y., and Vinyals, O. (2018). Representation learning with contrastive predictive coding. ar Xiv preprint ar Xiv:1807.03748. Orhan, E., Gupta, V., and Lake, B. M. (2020). Self-supervised learning through the eyes of a child. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 9960 9971. Curran Associates, Inc. Parthasarathy, N., Eslami, S., Carreira, J., and Hénaff, O. J. (2022). Self-supervised video pretraining yields strong image representations. ar Xiv preprint ar Xiv:2210.06433. Roder, B. J., Bushnell, E. W., and Sasseville, A. M. (2000). Infants preferences for familiarity and novelty during the course of visual processing. Infancy, 1(4):491 507. Published as a conference paper at ICLR 2023 Ryali, C., Schwab, D. J., and Morcos, A. S. (2021). Learning background invariance improves generalization and robustness in self-supervised learning on imagenet and beyond. In Neur IPS 2021 Workshop on Image Net: Past, Present, and Future. Schneider, F., Xu, X., Ernst, M. R., Yu, Z., and Triesch, J. (2021). Contrastive learning through time. In SVRHM 2021 Workshop @ Neur IPS. Shijie, J., Ping, W., Peiyi, J., and Siping, H. (2017). Research on data augmentation for image classification based on convolution neural networks. In 2017 Chinese automation congress (CAC), pages 4165 4170. IEEE. Shorten, C. and Khoshgoftaar, T. M. (2019). A survey on image data augmentation for deep learning. Journal of Big Data, 6(1):60. Smith, L. B., Jayaraman, S., Clerkin, E., and Yu, C. (2018). The developing infant creates a curriculum for statistical learning. Trends in Cognitive Sciences, 22(4):325 336. Stojanov, S., Thai, A., and Rehg, J. M. (2021). Using shape to categorize: Low-shot learning with an explicit shape bias. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 1798 1808. Stooke, A., Lee, K., Abbeel, P., and Laskin, M. (2021). Decoupling representation learning from reinforcement learning. In Meila, M. and Zhang, T., editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 9870 9879. PMLR. Tian, Y., Sun, C., Poole, B., Krishnan, D., Schmid, C., and Isola, P. (2020). What makes for good views for contrastive learning? In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 6827 6839. Curran Associates, Inc. Tsai, Y.-H. H., Wu, Y., Salakhutdinov, R., and Morency, L.-P. (2020). Self-supervised learning from a multi-view perspective. In International Conference on Learning Representations (ICLR). Wang, B., Mayo, D., Deza, A., Barbu, A., and Conwell, C. (2021). On the use of cortical magnification and saccades as biological proxies for data augmentation. In SVRHM 2021 Workshop@ Neur IPS. Wang, X., Ma, T., Ainooson, J., Cha, S., Wang, X., Molla, A., and Kunda, M. (2018). The toybox dataset of egocentric visual object transformations. ar Xiv preprint ar Xiv:1806.06034. Wiskott, L. and Sejnowski, T. J. (2002). Slow feature analysis: Unsupervised learning of invariances. Neural Computation, 14(4):715 770. Wood, J. N. and Wood, S. M. (2018). The development of invariant object recognition requires visual experience with temporally smooth objects. Cognitive Science, 42(4):1391 1406. Xu, J. and Wang, X. (2021). Rethinking self-supervised correspondence learning: A video frame-level similarity perspective. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 10075 10085. Yarats, D., Fergus, R., Lazaric, A., and Pinto, L. (2021). Reinforcement learning with prototypical representations. In Meila, M. and Zhang, T., editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 11920 11931. PMLR. Zhong, Z., Zheng, L., Kang, G., Li, S., and Yang, Y. (2020). Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 13001 13008. Published as a conference paper at ICLR 2023 A SAMPLING VIEWS IN THE CORE50 ENVIRONMENT To simulate No and Ns we do not directly sample from the data set, but rather build a buffer of positive pairs and iterate through it. Building the training buffer is done in cycles, see Figure 6A. In every cycle, each different object of the data set is chosen exactly once the order in which the objects are presented is a new random permutation during each cycle. For each sampled object, the procedure consists of three steps: 1. We randomly sample a view v1. 2. We sample a view from the same object v2 to form the same-object pair (v1, v2) using a) random walk, b) uniform. If we reach Ns, we sample v2 from a different session (same object but with a different background). 3. We repeat 2. to form (v2, v3) and continue until we reached No of object manipulation steps. After that, we sample a view from the next object according to the cycle (which is also a positive pair) and repeat the procedure. Thus, a cycle consists of a total of Nobj No object views. During the manipulation of one object, multiple cross-session transitions can occur, see Figure 6B. Several cycles of Nobj No views are stored in the buffer after another. During training, batches of pairs of subsequent views are sampled uniformly from the buffer (dashed brackets) and filled into the batch. We consider an epoch to be an entire run through the buffer, which is chosen to match the size of the underlying CORe50 dataset. In consequence, every training image on average is presented once each epoch. randomwalk, No = 5, Ns = 2 randomwalk, No = 3 1 2 3 4 5 6 9 8 9 4 3 2 9 Figure 6: Dynamic sampling procedure for the CORe50 dataset. A) Buffer with deterministic No = 3. Digits correspond to frame number of the video. B) Buffer with cross session sampling, colors represent different sessions. B COMPLEMENTARY ANALYSIS B.1 ADDITIONAL CONTROL EXPERIMENTS The number of conventional positive pairs per sample does not matter. To fairly compare the quality of time-based augmentations with conventional ones in Table 1, we controlled for the number of conventional positive pairs in VHE, 3DShape E and Toy Box by fixing 1 positive pair per image. Here, we aim to check whether time-based augmentations are useful even if we increase the number of positive pairs per sample (a new one for each training minibatch). In this case, the number of positive pairs per sample equals the number of epochs. In the Toy Box environment (Figure 7A), we observe increasing the number of conventional positive pairs improves both the baseline (Sim CLR) and the combined augmentations (Sim CLR-TT+) and does not change their relative ordering. In addition, it does not significantly impact the accuracy of the 3DShape E (Figure 7B) and VHE (Figure 7C). Overall, this confirms that time-based augmentations provide complementary and better information about the object shape in comparison to conventional data-augmentations. The linear classifier learns about rotation invariance, but it does not fully explain our results. Here, we want to assess the importance of the role of linear classifiers to learn the invariance over rotations in our results. We focus on the VHE and reuse the single-single setup used in Section 4.2 and compare the results with a new multi-single setup. Through the multi-single setup, we train Published as a conference paper at ICLR 2023 More aug One aug Time Used augmentations Sim CLR TT+ Sim CLR Sim CLR TT Toy Box category acc. A More aug One aug Time Used augmentations 3DShape E category acc. B More aug One aug Time Used augmentations VHE category acc. C Figure 7: Test category accuracy according to whether we use one positive pair per image (One aug) or several positive pairs, one per minibatch (More aug). We compare on A) the Toy Box environment; B) 3DShape E and C) VHE. Sim CLR TT+ Sim CLR Method Front views Side views 3/4 views Majority 3DShape E category acc., multi single A Sim CLR TT+ Sim CLR Method Front views Side views 3/4 views Majority 3DShape E category acc., single single B Figure 8: A) Test category accuracy with a linear classifier trained on the multi-view dataset (several orientations of objects), and tested on test objects with one orientation. B) Replicated from Figure 4C for readability. Test category accuracy with a linear classifier trained on one single orientation of objects, and tested on the same orientation of test objects. We used 5 seeds, all our conventional augmentations but only rotations ([0, 360] degrees) as time-based augmentation. We apply the majority vote on datasets of six different orientations: 0 , 10 , 45 , 80 , -80 , -90 . a linear classifier on representations gained from the usual multi-view dataset, but test it on the single-view datasets. We observe an overall decrease of performance in Sim CLR when testing with the multi-single setup (Figure 8A) in comparison to single-single setup (Figure 8B). In contrast, Sim CLR-TT+ (with rotations only) is more robust to the training set of the linear classifier. It suggests that learning rotation invariance with the linear classifier is mostly harmful for Sim CLR. However, the remaining performance discrepancy in the single-single setup shows that it does not explain the entire performance discrepancy in Table 1. TT+ significantly outperforms its Sim CLR counterpart. As mentioned in the main text we underline the results presented in Section 4.1 by statistically testing our Sim CLR-TT+ method against its Sim CLR counterpart for all of our 4 different datasets/environments. We did so by rerunning the experiment with additional random-seeds and comparing test-set accuracy with a two sample independent t-test. We found significant differences for all of our comparisons p < .05, the details of this analysis can be found in Table 2. Sim CLR-TT+ also improves over Sim CLR in VHE and 3DShape E with a Res Net18. We wanted to verify whether TT+ during object rotations improves over TT in VHE and 3DShape E with a Res Net18 neural network. In Table 3, we repeat the Sim CLR-based experiments described in Section 4 with the following modifications: 1 we use a Res Net18 backbone; 2we set the object manipulation time to 50 time steps; 3we increase the number of positive pairs per sample (like Published as a conference paper at ICLR 2023 Sim CLR Sim CLR-TT+ M SD M SD df t(df) p VHE 0.326 0.004 0.534 0.004 18 112.22 2.2 10 16 3DShape E 0.523 0.004 0.617 0.002 17 55.477 2.2 10 16 Toy Box 0.315 0.004 0.378 0.007 10 18.409 4.8 10 09 CORe50 0.554 0.070 0.629 0.081 18 2.228 0.039 Table 2: Detailed results of the statistical comparison between Sim CLR and Sim CLR-TT+ regarding test-set accuracy. P-values are bold. Environment Sim CLR Sim CLR-TT Sim CLR-TT+ 3DShape E 0.617 0.007 0.598 0.001 0.674 0.001 VHE 0.626 0.002 0.43 0.007 0.634 0.003 Table 3: Results of different augmentations with Sim CLR and a Res Net18 on the VHE and 3DShape E environments. in Appendix B.1); 4we change the maximum rotation speed in the VHE from 360 degrees to 10 degrees, as we found it to work better in this case. We hypothesize that high rotation speeds favor the encoding of color information, since we did not observe this effect in 3DShape E. We observe that Sim CLR-TT+ gives the best performances in both environments. B.2 ABLATION OF DATA-AUGMENTATIONS In this section, we do an ablation experiment of data-augmentations in Toy Box, 3DShape E and VHE since we can choose to remove positive pairs that come from specific natural interactions. We also consider conventional data-augmentations in VHE and 3DShape E since we found the complete set of augmentations to be suboptimal. Shape Environment. In Figure 9A, we test different sets of augmentations in the 3DShape E 3D, progressively and cumulatively combined. Since the combinatorial combination of all augmentations is computationally infeasible, we run a greedy forward algorithm to select the order of augmentations: R RC RCJ RCJD RCJG RCJH RD Used augmentations Average accuracy Test category acc. A 0.0 0.1 0.2 0.3 0.4 0.5 Million steps Average accuracy d = 0.075 d = 0.15 d = 0.3 d = 0.45 no depth Test category acc. B Figure 9: A) Ablation analysis of all data-augmentations in the 3D Shape environment. Horizontal bars indicate the minimum and maximum accuracy over 3 seeds. We note: R=Rotations; D=Depth changes; J=Color Jittering; G=Grayscale; C=Crop and Resize; H=Horizontal flip. B) Analysis of the impact of the maximal change of depth distance d. Published as a conference paper at ICLR 2023 None R RE REJ REJG REJGC REJGD REJGH Used augmentations Average accuracy VHE category acc. A All Ho Ho,R Ho,Tr T Used augmentations Toy Box category acc. B Figure 10: A) Ablation analysis of all data-augmentations in the VHE. Horizontal bars indicate the minimum and maximum accuracy over 3 seeds. We note: R=Rotations; D=Depth changes; J=Color Jittering; G=Grayscale; C=Crop and Resize; H=Horizontal flip; E=Ego-motion with Ns=1 and room changes. B) Ablation study on the Toy Box environment. We apply Sim CLR-TT+ but remove hodgepodge (-Ho), translations (-Tr), rotations (-R) and all time-based augmentations (-T). Experiments use the same training data, and we only change how we sample the positive pair (except for None and Depth experiments). at each step, we keep the augmentation with the largest increase in accuracy and stop when there is no more improvement. We find three main augmentations: rotations, crop and resize and color jittering. Adding other augmentations does not seem to have a significant large effect on the accuracy. To better understand the impact of time-based augmentations during depth changes on the representation, we also combine rotations with depth changes (R+D). Adding depth changes achieve similar performance to the addition of crop and resize. We hypothesize it mimics the resize part of the crop. An additional analysis (Figure 9B) shows that the results are robust to the maximal speed of depth changes. Virtual Home Environment. We test different sets of augmentations in the VHE, following the same procedure as in Appendix B.2. In Figure 10A, we observe that rotations, ego-motion and color-based transformations are crucial for category recognition. Yet, we can not conclude other augmentations are useless, since it may be the redundancy with other augmentations that is harmful, as we found with depth changes in Figure 9A. Toy Box environment. Figure 10B shows the impact of removing time-based augmentations in the Toy Box environment. We observe that removing augmentations based on translations (-Ho,Tr) and rotations (-Ho,R) both hurt the quality of the representation. Removing only some translations and translations (-Ho) do not affect the performance, presumably because other clips (with rotations and translations alone) already provide information about the shape of the object. B.3 ROTATIONS ALLOW FOR BETTER SHAPE GENERALIZATION THAN THE FLIP AUGMENTATION Here, we investigate how object rotations relate to the standard horizontal flip augmentation, since the flip resembles a 180-degrees rotation for symmetric objects and uniform background. In Figure 11A, we combine the flip augmentation and object rotations with other conventional augmentations. We see that when the object rotation speed is large enough (rot(360)), there is no additional benefit in using the horizontal flip augmentation. However, rotations always boost performance compared to the pure flipping augmentation. Thus, we conclude that using 3-D object rotations rather than flipping results in considerably better representation of object shape. B.4 HIGH-FREQUENCY OBJECT SWITCHES HURT CATEGORY RECOGNITION. Here, we study the impact of the object manipulation duration No. In 3DShape E (Figure 11B), we observe a sweet spot at No = 10 when using a simple convolutional neural network (Conv Net). This effect seems to be architecture-dependent since we observe a consistent increase of accuracy as we Published as a conference paper at ICLR 2023 0.0 0.1 0.2 0.3 Million steps rot(360) + flip rot(360) flip rot(90) + flip rot(90) 3DShape E category acc. A 2 5 10 20 50 Manipulation duration N Conv Net Res Net18 3DShape E category acc. B Figure 11: A) 3DShape E test category accuracy according to different combinations of flip and rotation speeds. B) 3DShape E test category accuracy according to different manipulation duration No with a simple Conv Net and a Res Net18. increase No with a Res Net18. Importantly, in both cases, we observe that high-frequency object switches (low No) hurt the downstream object categorization. In CORe50 (Table 4), the combined randomwalk (TT+ rw) method is sensitive to a high frequency of object switches within the same session, which considerably improves validation set category accuracy (A). We assume that this is because the additional contrastive information is considerably higher for the randomwalk method, whereas uniform view sampling already achieves some degree of background variation due to sampling contrasts from one whole session compared to the neighboring frames. For test set accuracies, we observe a similar trend, albeit not as pronounced (B). Note how test-accuracies for the random walk procedure are considerably higher than for the uniform sampling. B.5 INFLUENCE OF THE SESSION SWITCH FREQUENCY IN THE CORE50 ENVIRONMENT Detailed analysis of varying the session frequency in the CORe50 environment. To complete the results from the CORe50 dataset we add overview tables and a more thorough analysis regarding the randomwalk procedure. The mean and standard errors of the analysis corresponding to Figure 4D can be found in Table 4 together with a supervised and Sim CLR baseline trained with the same hyperparameters. The uniform view sampling (uni) performs best on the validation set, whereas the two combined methods (*-TT+) deliver the best generalization capabilities across splits. algorithm 2 5 10 20 50 supervised .530 .080 " " " " Sim CLR .544 .030 " " " " -TT rw .229 .018 .327 .033 .386 .046 .429 .055 .427 .052 -TT uni .426 .055 .435 .060 .440 .056 .450 .059 .440 .056 -TT+ rw .501 .032 .589 .030 .615 .029 .629 .031 .638 .032 -TT+ uni .618 .084 .608 .036 .620 .036 .616 .036 .616 .033 Table 4: Category accuracy and standard error on CORe50, based on Sim CLR, ps = 0.95. Standard error based on five different training/testing splits. Training occurred for 100 epochs, batchsize 512. Best performance is highlighted in bold, cf. Figure 4D. Published as a conference paper at ICLR 2023 Figure 12: Analysis of the Sim CLR-TT+ approach for different values of ps, CORe50 Environment. Error bars indicate standard error based on three different training/testing splits. We also report means and standard errors for category and session accuracy (Table 5) of our randomwalk procedure when evaluated with different values for ps and po. Category accuracy E [Ns] 2 5 10 20 50 2 .993 .001 .994 .001 .994 .001 .997 .000 .998 .000 5 .898 .009 .974 .002 .986 .000 .989 .001 .992 .001 10 .572 .005 .909 .005 .956 .004 .972 .002 .981 .003 20 .439 .004 .629 .008 .814 .010 .908 .009 .939 .004 50 .401 .007 .454 .006 .522 .007 .617 .003 .711 .006 Session accuracy E [Ns] 2 5 10 20 50 2 .572 .010 .535 .011 .540 .015 .507 .008 .518 .010 5 .665 .012 .644 .014 .658 .018 .671 .011 .704 .011 10 .646 .016 .695 .019 .718 .016 .718 .014 .707 .009 20 .670 .026 .673 .003 .716 .021 .740 .013 .735 .013 50 .704 .012 .689 .025 .732 .012 .728 .015 .731 .020 Table 5: Category accuracy and standard error on CORe50, Sim CLR-TT algorithm with randomwalk sampling procedure. Standard error based on three different training/testing splits. Training occurred for 100 epochs, batchsize 512, cf. Figure 5C, D. Influence of session changes on combined augmentations. Since our main analysis with the CORe50 dataset aims to give a solid overview over the presented approaches, we have picked a medium level of session changes as the default (ps = 0.95, E [Ns|ps] = 20). The results show that the combined approaches (*-TT+) outperform the conventional Sim CLR method, but we were also interested in how sensitive the novel methods are to the ps parameter. Results of that analysis are depicted in Figure 12. In line with results from our Virtual Home Environment experiments, we observe that for very low values of Ns the internal representation contains less information about the session for the randomwalk approach (C,D). The uniform method, however, shows robustness to this effect. Published as a conference paper at ICLR 2023 C.1 3D OBJECTS IN THE VIRTUAL HOME ENVIRONMENT We import two versions of objects into the Three DWorld Software (TDW): 3DShape E. This is the grey version of objects. We export the 3D models to .obj format and remove the texture files. We discard objects that took too long to process in TDW. Textured objects. This is the colored version of objects. In order to import them in TDW, we apply a series of operators: 1) we reduce the complexity of some meshes to make the next step computationally feasible and reduce the TDW processing complexity; 2) we manually bake most of the 3D models to obtain a single texture file with the meshes; 3) we import the models into TDW; 4) we resize the models to marginalize the effect of size on category recognition and avoid harmful collisions. We visually check the quality of all models. When we do not manage to obtain good-looking textures for some objects, we remove them from our dataset (< 15% of the dataset). The tree category does not contain enough good-looking objects, so we remove the whole category. C.2 TEST DATASETS OF THE VIRTUAL HOME ENVIRONMENT. Our House test set is composed of 7,740 images, from 1,548 objects distributed into 105 categories. By showcasing the category elephant in Figure 13 we exhibit the diversity of objects within one category. The elephant category includes flying elephants, bipedal elephants, red elephants, differently positioned elephants or geometrically simplified elephants. We also show the sequence of input images in the VHE according to different rotation speeds. In Figure 14, we display the three test houses used in Figure 5B,C. House scenes were taken from pre-provided bundles in TDW and we added the agent and objects in the scene similarly to Figure 1. The underlying Three DWorld (TDW) software used is licenced under BSD 2-Clause. C.3 TOYBOX ENVIRONMENT The Toy Box environment is a collection of video-clips. We sample two frames per second from each clip, resulting in 159,627 images. Converted clips contain 2 to 50 sampled frames, which corresponds to the No parameter in the CORe50 environment. Two thirds of objects are used for the training set and the rest for the test set. D HYPERPARAMETERS For all conducted experiments, we apply a weight decay of 10 6 and update weights with the Adam W optimizer (Loshchilov and Hutter, 2018) and a learning rate of 5 10 4. Virtual Home Environment and 3D Shape Environment. The agent in the Virtual Home Environment perceives the world around it as 128 128 pixel RGB images. Unless stated otherwise, these images are encoded by a succession of convolutional layers with the following [channels, kernel size, stride, padding] structure: [64, 8, 4, 2], [128, 4, 2, 1], [256, 4, 2, 1], [256, 4, 2, 1]. The output is grouped by an average pooling layer and a linear layer ending with 128 units. Each convolution layer is followed by a non-linear Re LU activation function and a dropout layer (p = 0.5) to prevent over-fitting. We do not use projection heads. We consider a temperature hyperparameter of 0.1 for Sim CLR. We use a batch size of 256 and a buffer size of 100,000. Average training time was 72 hours per run for VHE, and 20 hours for the 3DShape E. All experiments ran on GPUs of type NVIDIA V100. We did a hyperparameter search of the best crop minimal size in {0.08, 0.2, 0.5, 0.75} for the Combined and Sim CLR aug experiments. CORe50 Dataset. The additional diversity of the CORe50 training data required a more capable encoder network, which we chose to be a Res Net-18 (He et al., 2016). The encoder transforms the Published as a conference paper at ICLR 2023 Figure 13: All objects of the category elephant used in our Virtual Home Environment. Figure 14: Top view of houses, without ceiling, used only when testing category recognition in unfamiliar houses experiments. A) House 2; B) House 3; C) House 4. Published as a conference paper at ICLR 2023 input images into a 128-dimensional latent representation, which is followed by a 2-layer linear projection head [256, 128]. For the projection head, we use batch normalization and a Re LU activation function in-between the two layers. We chose a batch size of 512 and training was done for 100 epochs on a GPU of type NVIDIA V100 or NVIDIA RTX2070 SUPER. Average training time was 17.5h per run. For BYOL we use τ = 0.996 and for VICReg the default error weighting as described in Bardes et al. (2022). The CORe50 dataset is licensed under CC-BY-4.0. Toy Box environment. We use the same hyperparameters as for the CORe-50 experiments. However, we made an ablation study of the crop minimal size hyperparameter to adapt to the small size and off-center position of the objects. We found a value of 0.5 to be the best in {0.08, 0.2, 0.5, None}.