# unsupervised_object_detection_with_theoretical_guarantees__d9ee436f.pdf Unsupervised Object Detection with Theoretical Guarantees Marian Longa Visual Geometry Group University of Oxford mlonga@robots.ox.ac.uk João F. Henriques Visual Geometry Group University of Oxford joao@robots.ox.ac.uk Unsupervised object detection using deep neural networks is typically a difficult problem with few to no guarantees about the learned representation. In this work we present the first unsupervised object detection method that is theoretically guaranteed to recover the true object positions up to quantifiable small shifts. We develop an unsupervised object detection architecture and prove that the learned variables correspond to the true object positions up to small shifts related to the encoder and decoder receptive field sizes, the object sizes, and the widths of the Gaussians used in the rendering process. We perform detailed analysis of how the error depends on each of these variables and perform synthetic experiments validating our theoretical predictions up to a precision of individual pixels. We also perform experiments on CLEVR-based data and show that, unlike current SOTA object detection methods (SAM, Cut LER), our method s prediction errors always lie within our theoretical bounds. We hope that this work helps open up an avenue of research into object detection methods with theoretical guarantees. 1 Introduction Unsupervised object detection using deep neural networks is a long-standing area of research at the intersection of machine learning and computer vision. Its aim is to learn to detect objects from images without the use of training labels. Learning without supervision has multiple advantages, as obtaining labels for training data is often costly and time consuming, and in some cases may be impractical or unethical. For example, in medical imaging, unsupervised object detection can help save specialists time by automatically flagging suspicious abnormalities [19], and in autonomous driving it may help automatically detect pedestrians on a collision course with the vehicle [3]. It is thus important to understand and develop better unsupervised object detection methods. While successful, current object detection methods are often empirical and possess few to no guarantees about their learned representations. In this work we aim to address this gap by designing the first unsupervised object detection method that we prove is guaranteed to learn the true object positions up to small shifts, and performing a detailed analysis of how the maximum errors of the learned object positions depend on the encoder and decoder receptive field sizes, the object sizes, and the sizes of the Gaussians used for rendering. This is especially important in sensitive domains such as medicine, where incorrectly detecting an object could be costly. Our method guarantees to detect any object that moves in a video or that appears at different locations in images, as long as the objects are distinct and the images are reconstructed correctly. We base our unsupervised object detection method on an autoencoder with a convolutional neural network (CNN) encoder and decoder, and modify it to make it exactly translationally equivariant (sec. 3). This allows us to interpret the latent variables as object positions and lets us train the network without supervision. We then use the equivariance property to formulate and prove a theorem that relates the maximum position error of the learned latent variables to the size of the encoder and 38th Conference on Neural Information Processing Systems (Neur IPS 2024). decoder receptive fields, the size of the objects, and the width of the Gaussian used in the decoder (sec. 4). Next, we derive corollaries describing the exact form of the maximum position error as a function of these four variables. These corollaries can be used as guidelines when designing unsupervised object detection networks, as they describe the guarantees of the learned object positions that can be obtained under different settings. We then perform synthetic and CLEVR-based [12] experiments to validate our theory (sec. 5). Finally, we discuss the implications of our results for designing reliable object detection methods (sec. 6). Concretely, the contributions of this paper are: 1. An unsupervised object detection method that is guaranteed to learn the true object positions up to small shifts. 2. A proof and detailed theoretical analysis of how the maximum position error of the method depends on the encoder and decoder receptive field sizes, object sizes, and widths of the Gaussians used in the rendering process. 3. Synthetic experiments, CLEVR-based experiments, and real video experiments validating our theoretical results up to precisions of individual pixels. 2 Related Work Object Detection. Object detection is an area of research in computer vision and machine learning, dealing with the detection and location of objects in images. Popular supervised approaches to object detection include Segment Anything (SAM) [13], Mask R-CNN [7], U-Net [18] and others [4]. While successful, these methods typically require large amounts of annotated segmentation masks and bounding boxes, which may be costly or impossible to obtain in certain applications. Popular unsupervised and self-supervised object detection methods include Cut LER [23], Slot Attention [17], Mo Net [2] and others [5]. These methods aim to learn object-centric representations for object detection and segmentation without using training labels. Finally, unsupervised object localisation methods such as FOUND [21] and others [22] aim to localise objects in images, typically using vision transformer (Vi T) self-supervised features. Compared to both current supervised and unsupervised object detection and localisation methods, our work is the only one that has provable theoretical guarantees of recovering the true object positions up to small shifts. It also requires no supervision. Identifiability in Representation Learning. Identifiability in representation learning refers to the issue of being able to learn a latent representation that uniquely corresponds to the true underlying latent representation used in the data generation process. Some recent works aim to reduce the space of indeterminacies of the learned representations, and thus achieve identifiability, by incorporating various assumptions into their models. Xi et al. [24] categorise these assumptions for generative models into constraints on the distribution of the latent variables and constraints on the generator function. Some of their categories include non-Gaussianity of the latent distribution [20], dependence on an auxiliary variable [9, 10], use of multiple views [16], use of interventions [1, 15], use of mechanism sparsity [14], and restrictions on the Jacobian of the generator [6]. In contrast, in our work we achieve identifiability by making our network equivariant to translations, imposing an interpretable latent space structure, and requiring the data to obey our theorem s assumptions. In this section we describe the proposed method for unsupervised object detection with guarantees. On a high level, our architecture is based on an autoencoder that is fully equivariant to translations, which we achieve by making the encoder consist of a CNN followed by a soft argmax function to extract object positions, and making the decoder consist of a Gaussian rendering function followed by another CNN to reconstruct an image from the object positions (fig. 1). In the following sections we describe the different parts of the architecture in detail. Autoencoder with CNN Encoder and Decoder. We start with an autoencoder, a standard unsupervised representation learning model, consisting of an encoder network ψ that maps an image x to a low-dimensional latent variable z, followed by a decoder network ϕ that maps this variable back to an image ˆx, with the objective of minimising the difference between x and ˆx. Typically, the encoder and decoder networks are parametrised by multi-layer perceptrons (MLPs) or convolutional neural networks (CNNs) paired with fully-connected (FC) layers, however neither of these parametrisations Soft Argmax 𝑒!, , 𝑒" (𝑥 Render Gaussians 𝑧!,$, 𝑧!,% 𝑧",$, 𝑧",% Input Image Rendered Gaussians Image Latent Variables CNN CNN Reconstruction Loss Positional Encodings Figure 1: Network architecture. Encoder: (1) an image x is passed through a CNN ψ to obtain n embedding maps e1, ..., en, (2) a maximum of each map is found using softargmax to obtain latent variables [z1,x, z1,y, ..., zn,x, zn,y]. Decoder: (1) Gaussians ˆe1, ..., ˆen are rendered at the positions given by the latent variables, (2) the Gaussian maps are concatenated with positional encodings and passed through a CNN ϕ to obtain the predicted image ˆx. Finally, x and ˆx are used to compute reconstruction loss L(ˆx, x). by default can guarantee that the learned latent variables will correspond to the true object positions (because of the universal approximation ability of MLPs and FC layers [8]). To obtain such guarantees, we would thus like to modify the autoencoder to make it exactly translationally equivariant, that is, a shift of an object in the input image x should correspond to a proportional shift of the latent variable z, and a shift of the latent variable z should correspond to a shift in the predicted image ˆx. We start with an autoencoder where the encoder and decoder are both CNNs. CNNs consist of layers computing the convolution between a feature map x and a filter F, defined in one dimension as (x F)[i] = X j x[j]F[j i] (1) Intuitively, this corresponds to sliding the filter F across the feature map x and at each position of the filter i computing the dot product between the feature map x and the filter F. We can prove that convolutional layers are equivariant to translations, since ((τ x) F)[i] = X j x[j t]F[j i] = X j x[j]F[j (i t)] = τ (x F)[i] (2) where τ is the translation operator that translates a feature map by t pixels, and we have used the substitution j j +t at the second equality. Therefore, the encoder and decoder are both equivariant to translations, but this property only holds for translations of feature maps (i.e. spatial tensors). From Encoder Feature Maps to Latent Variables. So far we have only worked with images and feature maps, but the latter do not directly express positions of any detected objects. It would be preferable to convert these feature maps into scalar variables that can be interpreted as object positions that are equivariant to image translations. To do this, we first define a translation τ of a (1D) feature map x and a translation τ of a scalar z as τ(x)[i] = x[i t], τ (z) = z + t (3) where i is the position in the feature map x, τ shifts an image by t pixels, and τ shifts a scalar by t units. To relate translations in feature maps to translations in latent variables, we can use a function that computes a scalar property of a feature map x, such as argmax, defined as argmax(x) = {i : x[j] x[i] j}. Using these definitions we can now prove the equivariance of argmax, i.e. that shifting the feature map x by τ corresponds to shifting the latent variable argmax(x) by τ : argmax(τ x) = {i : τ x[j] τ x[i] j} = {i : x[j t] x[i t] j} = {i + t : x[j] x[i] j} = argmax(x) + t = τ argmax(x) (4) where at the first equality we use the definition of argmax, at the second equality we use the definition of τ (eq. 3, left), at the third equality we use the substitution i i + t, at the fourth equality we use the definition of argmax, and at the last equality we use the definition of τ (eq. 3, right). However, because the argmax operation is not differentiable, for neural network training we approximate it via a differentiable soft argmax function, defined in 2D as softargmax(x) = where σ is the softmax function defined in one dimension as σ(x)[i] = exp(x[i])/ P j exp(x[j]), σ1(x) and σ2(x) is the softmax function evaluated along the first and second dimensions of x, Θ is a temperature hyperparameter, [i, j] is the image index, I is the image width, J is the image height, and the term 1/2 ensures that the densities correspond to pixel centres. As the temperature Θ in eq. 5 approaches zero, softargmax reduces to the classical argmax function. From Latent Variables to Decoder Feature Maps. Similar to mapping from encoder feature maps to latent variables, we would now like to relate shifts in latent variables z to shifts of decoder feature maps x. To do this, we can invert the action of the argmax operation. Because argmax is a many-toone function, finding an exact inverse is not possible, but we can obtain a pseudo-inverse using the Dirac delta function defined as delta(z)[i] = δ(i z). We can show that delta is a pseudo-inverse of argmax because argmax delta z = i : δ(j z) δ(i z) ; j = z. Now, similar to the argmax function, we can prove that the delta function is equivariant to the latent variable shift τ on the input and the feature map shift τ on the output, i.e. delta(τ z)[i] = δ(i τ z) = δ(i z t) = delta(z)[i t] = τ delta(z)[i] (6) where at the first equality we have used the definition of delta, at the second equality we have used the definition of τ (eq. 3, right), at the third equality we have used the definition of delta, and at the last equality we have used the definition of τ (eq. 3, left). Again, because the delta function is not differentiable, we can approximate it using a differentiable render function as render(z)[i] = N(i z, σ2) (7) where N(i z, σ2) is a Gaussian evaluated at i z with variance given by the hyperparameter σ2. As the variance σ2 in eq. 7 approaches zero, the render function reduces to the hard delta function. Additionally, because the decoder is translationally equivariant, we also condition it on positional encodings of the size of the images to provide it with sufficient information to reconstruct different parts of the background, assuming the background is static. Alternatively, if background is varying, the decoder can be conditioned on a randomly-sampled nearby video frame instead, which will provide information about the background but not the objects positions (following Jakab et al. [11]). We also note that since the latent variables z are ordered, this allows the encoder and decoder to learn to associate each variable with the semantics of each object and achieve successful reconstruction. We thus now have all the elements we need to create an equivariant architecture where the encoder and decoder are defined, respectively, by z = softargmax ψ x, ˆxt = ϕ render zt. (8) This is shown in fig. 1. Having designed an exactly translationally equivariant architecture now allows us to obtain theoretical guarantees about the learned latent variables, which we discuss next. 1 4 Theoretical Results In this section we present our main theorem stating the maximum bound on the position errors of the latent variables learned with our method in terms of the encoder and decoder receptive field sizes, the object size, and the Gaussian standard deviation (thm. 4.1). We continue by deriving specialised corollaries relating the maximum position error to the encoder receptive field size (cor. 4.2), decoder receptive field size (cor. 4.3), object size (cor. 4.4), and Gaussian standard deviation (cor. 4.5). Theorem 4.1. Error Bound. Consider a set of images x X with objects of size so, CNN encoder ψ with receptive field size sψ, CNN decoder ϕ with receptive field size sϕ, soft argmax function softargmax, rendering function render with Gaussian standard deviation σG and G N(0, σ2 G), and latent variables z, composed as z = softargmax ψ x and ˆx = ϕ render z (fig. 1). Assuming (1) the objects are reconstructed at the same positions as in the original images, (2) each object appears in at least two different positions in the dataset, and (3) there are no two identical objects in any image, then the learned latent variables z correspond to the true object positions up to object permutations and maximum position errors of 2 + G . (9) 1We note that a similar architecture was proposed by Jakab et al. [11], with empirical success in keypoint detection. However, we derive our architecture by enforcing strict translation equivariance properties, which makes our theoretical results possible. (a) Encoder Error (b) Decoder Error Figure 2: Position errors. (a) Maximum position error due to encoder, given by sψ/2 + so/2 1. The maximum error occurs when the encoder and the object are as far away from each other as possible while still overlapping by one pixel. (b) Maximum position error due to decoder, given by sϕ/2 so/2 + G. The maximum error occurs when some part of the Gaussian at position z + G is within the decoder receptive field (RF) but is as far away from the rendered object as possible. For proof see appendix A. Intuitively, the assumptions ensure that each latent variable corresponds to the position of each object in the image. The error in the learned object positions then arises from both the encoding and decoding process. In the encoding process, the maximum error occurs when the encoder and the object are as far away from each other as possible while still overlapping, i.e. sψ/2 + so/2 1 (fig. 2a). Conversely, in the decoding process, the maximum error occurs when the rendered object and the latent variable are as far away from each other as possible while both still being inside the decoder receptive field, i.e. sϕ so/2 (fig. 2b). Additionally, there is an extra error of G as the latent variable is rendered by a Gaussian and the decoder can capture any part of this Gaussian. Finally, because we assume each object is reconstructed at the same position as in the original image, the errors from the encoder and decoder must cancel each other out. Therefore, the overall maximum position error is given by the lower of the two expressions for the encoder and the decoder, leading to eq. 9. Next, we present corollaries relating this error bound to different factors. Corollary 4.2. Error Bound vs. Encoder RF Size. The maximum position error as a function of the encoder receptive field (RF) size sψ for a given sϕ, so, σG, is 2 1 if 1 sψ sϕ 2so + 2, sϕ 2 + G if sψ > sϕ 2so + 2. For an illustration see fig. 3a. There are two regions of the curve (separated by a dashed line). In the left-most region, sψ < sϕ 2so + 2, the error is dominated by the encoder error, and in the right-most region, sψ sϕ 2so + 2, the error is dominated by the decoder error. Initially, for sψ = 1, the error is given by so/2 1/2, because the 1 1 px encoder can match any pixel that is part of the object and so can be at most half of the object size away from the true object position that is at the centre of the object. As the encoder RF size increases up to sϕ 2so + 2, the position error increases linearly with it as sψ/2 + so/2 1, because now any part of the encoder RF can match any part of the object (fig. 2a). This bound is deterministic due to the deterministic encoding process. At sψ = sϕ 2so + 2 (vertical dashed line in fig. 3a), the maximum errors from encoder and decoder both become equal to sϕ/2 so/2. For sψ > sϕ 2so + 2, the position error is dominated by the error from the decoder which is constant at sϕ/2 so/2 + G with G N(0, σ2 G), and so even though the encoder RF size is increasing, this has no effect as the limiting factor is now the decoder. Due to the Gaussian rendering step in the decoding process, this bound is now probabilistic, and is distributed normally with variance σ2 G.The results of corollary 4.2 can be extended to multiple objects with different sizes (see appendix B, cor. B.1) Corollary 4.3. Error Bound vs. Decoder RF Size. The maximum position error as a function of the decoder receptive field (RF) size sϕ for a given sψ, so, σG, is 2 + G if so sϕ < sψ + 2so 2, sψ 2 1 if sϕ sψ + 2so 2. For an illustration see fig. 3b. Similar to corollary 4.2, there are two regions of the curve, one for sϕ < sψ + 2so 2 (left), where the error is dominated by the decoder error, and another for sϕ sψ + 2so 2 (right), where the error is dominated by the encoder error. Note that this is opposite to cor. 4.2. Initially, for sϕ = so, the decoder receptive field has the same size as the object, (a) Error vs. Encoder RF. (b) Error vs. Decoder RF. (c) Error vs. Object Size. (d) Error vs. Gaussian S.D. Figure 3: Theoretical bounds for the maximum position error as a function of the encoder receptive field size sψ, decoder receptive field size sϕ, object size so, and Gaussian standard deviation σG, as the remaining factors are fixed. Each bound consists of a region due to the encoder error (solid line) and the decoder error (probabilistic bound). Standard deviations are represented by shades of blue. and so to achieve perfect reconstruction it needs to be at the same position as the object, resulting in 0 position error plus any error G caused by the non-zero width of the Gaussian. As the decoder RF size increases up to sψ + 2so 2, the position error increases linearly with it as sϕ/2 so/2 + G, because now the object can be at an increasing number of positions within the decoder and still achieve perfect reconstructions (fig. 2b). At sϕ = sϕ + 2so 2, the maximum errors from encoder and decoder both become equal to sψ/2 + so/2 1. For sϕ > sψ + 2so 2, the position error is dominated by the error from the encoder which is constant at sψ/2 + so/2 1, and so even though the decoder RF size is increasing, this has no effect as the limiting factor is now the encoder. Similar to corollary 4.2, the results of corollary 4.3 can be extended to objects with multiple different sizes (see appendix B, cor. B.2). Corollary 4.4. Error Bound vs. Object Size. The maximum position error as a function of the object size so for a given sψ, sϕ, σG, is 2 1 if 1 so sϕ 2 + G if sϕ 2 + 1 < so sϕ. For an illustration see fig. 3c. Again, there are two regions of the curve, one for so < sϕ/2 sψ/2+1 (left), where the error is dominated by the encoder error, and one for so sϕ/2 sψ/2 + 1 (right), where the error is dominated by the decoder error. Initially, for so = 1, the error is given by sψ/2 1/2, because any pixel of the encoder receptive field can match the 1 1 px object and so the error can be at most half of the encoder receptive field size. As the object size increases up to sϕ/2 sψ/2 + 1, the position error increases linearly with it as sψ/2 + so/2 1, because now any part of the encoder RF can match any part of the object (fig. 2a). At so = sϕ/2 sψ/2 + 1, the maximum errors from encoder and decoder both become equal to sψ/4 + sϕ/4 1/2. For so > sϕ/2 sψ/2 + 1, the position error is dominated by the error from the decoder and decreases linearly as sϕ/2 so/2 + G, because now there is a decreasing number of positions where the object can still fit inside the decoder receptive field (fig. 2b). At so = sϕ, the object reaches the same size as the decoder, and thus the position error decreases to 0 with an additional error due to the width of the Gaussian, G. Interestingly, the triangular shape of the error curve means that small and large objects will both incur small position errors, while medium sized objects will incur higher errors. Corollary 4.5. Error Bound vs. Gaussian Size. The maximum position error as a function of the Gaussian standard deviation σG for a given sψ, sϕ, so, is 2 + G if σG < sψ 2 + so 1, sψ 2 1 if σG sψ For an illustration see fig. 3d. Firstly, there is an overall maximum bound for the position error due to the encoder, given by the constant sψ/2 + so/2 1, which is independent of the Gaussian standard deviation. Then, initially for σG = 0, the rendered Gaussian is effectively a delta function, and so the position error is dominated by the decoder error given by sϕ/2 so/2, which describes the maximum distance between the object and the delta function with both of them fitting inside the decoder receptive field (fig. 3b). As the Gaussian standard deviation increases, the position error increases linearly as sϕ/2 so/2 + G with G N(0, σ2 G). Then, depending on what part of the Gaussian the decoder is convolved with, there are different bounds for the maximum (a) Error vs. Encoder RF. (b) Error vs. Decoder RF. (c) Error vs. Object Size. (d) Error vs. Gaussian S.D. Figure 4: Synthetic experiment results showing position error as a function of the encoder receptive field size sψ, decoder receptive field size sϕ, object size so, and Gaussian standard deviation σG, as the remaining factors are fixed to sψ = 9, sϕ = 25, so = 9, σG = 0.8 (in a,b,c) or to sψ = 9, sϕ = 11, so = 7 (in d). Theoretical bounds are denoted by a blue line (with 4 shaded regions denoting 1 to 4 standard deviations of the probabilistic bound) and experimental results by red dots. position error. If the decoder is convolved with a part of the Gaussian that is within n standard deviations of its centre, the maximum position error increases linearly as sϕ/2 so/2 + nσG up to σG = (sψ/2 sϕ/2 + so 1)/n, after which point the position error becomes dominated by the encoder value of sψ/2 + so/2 1. In fig. 3d, the maximum position error bound when the decoder is convolved with a part of the Gaussian within its first and second standard deviations, is denoted by darker and lighter shades of blue, respectively. 5 Experimental Results In this section we validate our theoretical results on synthetic experiments (sec. 5.1) and CLEVR data (sec. 5.2). Additional real video experiments are in app. E. We first validate corollaries 4.2-4.4 via synthetic experiments in sec. 5.1, demonstrating very high agreement up to sizes of individual pixels. We then apply our method to CLEVR-based [12] data containing multiple objects of different sizes in varying scenes (sec. 5.2) and show that compared to current SOTA object detection methods (SAM [13], Cut LER [23]), only our method predicts positions within theoretical bounds. 5.1 Synthetic Experiments In this section we validate corollaries 4.2-4.5 via synthetic experiments. Our dataset consists of a small white square on a black background. In each experiment we fix all but one of the encoder RF size sψ, decoder RF size sϕ, object size so, and Gaussian standard deviation σG, and vary the remaining variable. We perform each experiment 20 times, corresponding to 20 random initializations of the trained parameters, and record the position error as the difference between the predicted object position z and the ground truth object position z GT . For more details see appendix C.1. Position Error vs. Encoder RF Size. In this experiment we aim to empirically validate corollary 4.2, by measuring the experimental position errors as a function of the encoder receptive field size. We vary the encoder RF sizes sψ {1, 3, . . . , 31} and fix sϕ = 25, so = 9, σG = .8 and record position errors . We visualise the data points (red) and the theoretical bounds (blue) in fig. 4a. We can observe that all the data points lie at or below the theoretical boundary, which validates corollary 4.2. In particular, we observe that the deterministic boundary in the region to the left of the dashed line (corresponding to the encoder bound) is well respected, with some of the trained networks achieving exactly the maximum error predicted by theory. Position Error vs. Decoder RF Size. In this experiment we aim to validate corollary 4.3 by measuring the experimental position errors as a function of the decoder receptive field size. We vary the decoder RF sizes sϕ {1, 3, . . . , 31} and fix sϕ = 9, so = 9, σG = .8 and record position errors . We visualise the results in fig. 4b. The figure shows the theory to be a strong fit to the data, validating corollary 4.3. In particular, we note that the data points fit the Gaussian distribution in the decoder part of the curve (left of the dashed line) and are very close to (1 px below) the deterministic upper bound in the encoder part of the curve (right). (a) Error vs. Encoder RF. (b) Error vs. Decoder RF. (c) Error vs. Object Size. (d) Error vs. Gaussian S.D. Figure 5: CLEVR experiment results showing position error as a function of the encoder receptive field size sψ, decoder receptive field size sϕ, and object size so, and Gaussian standard deviation σG, as the remaining factors are fixed to sψ = 9, sϕ = 25, so [6, 10], σG = 0.8 for (a)-(c) and to sψ = 5, sϕ = 13, so [6, 10] for (d). Theoretical bounds are denoted by blue, experimental results in red, SAM baseline in green, and Cut LER baseline in orange. Position Error vs. Object Size. In this experiment we aim to validate corollary 4.4 by measuring the experimental position errors as a function of the object size. We vary the object sizes so {1, 3, . . . , 25} and fix sψ = 9, sϕ = 25, σG = .8 and record position errors . We visualise the results in fig. 4c. As all the data points lie at or below the theoretical boundary, this validates corollary 4.4. We note that the empirical distribution of errors follows very closely the shape of the theoretical bound, very strictly on the left side of the dashed line (encoder bound) and according to the distribution predicted on the right side (decoder bound). Position Error vs. Gaussian Size. In this experiment we aim to validate corollary 4.5 by measuring the experimental position errors as a function of the Gaussian standard deviation. We vary the Gaussian standard deviations σG {0.1, 0.2, . . . , 2.1, 2.25, 2.5, ..., 5}, fixing sψ = 9, sϕ = 11, so = 7 and record position errors . We visualise the data points (red) and the theoretical bounds (blue) in fig. 4d. As all the data points lie at or below the theoretical boundary, this validates corollary 4.5. In particular, we note that all the data points lie below the encoder bound (solid blue line), and all the data points lie within the bound denoted by four standard deviations away from the Gaussian. This means that in practice, the decoder can be convolved with any part of the Gaussian that lies within 4 standard deviations (corresponding to 3.2 px) from its centre. We also note that as the Gaussian standard deviation increases, the position error increases as expected, denoted by the positive slope of the data points between the third and fourth standard deviations (lightest shade of blue). 5.2 CLEVR Experiments In this section we validate our theory on CLEVR-based [12] image data of 3D scenes. Our dataset consists of 3 spheres of different colours at random positions, with a range of sizes due to perspective distortion. We train and evaluate each experiment similarly to those in sec. 5.1, recording position errors for the learned objects, and compare our results with SAM [13] and Cut LER [23] baselines. We compute the theoretical bounds according to our theory in sec. 4 and app. B, and visualise the results in figs. 5a-5c. For details see app. C.2. For experiments with different shapes see app. D. Once again, the experimental results demonstrate high agreement with our theory, now even for more complex images with multiple objects and a range of object sizes (fig. 5, red, blue). Furthermore, while the SAM and Cut LER baselines generally achieve low position errors, this is not guaranteed, and in some cases their errors are much higher than our bound (fig. 5, green, orange). We report the proportion of position errors from fig. 5c that lie within 2 standard deviations of our theoretical bound in table 6b and fig. 6a, showing that compared to SOTA object detection methods, only for our method are the position errors always guaranteed to be within our theoretical bound. 6 Discussion In light of our theoretical results, in this section we present some conclusions that can be drawn when designing new unsupervised object detection methods: (a) Results visualisation. Errors Within Bound (%) Object Size 1.5 (px) Method All 9 12 15 18 21 24 Ours 100.0 100.0 100.0 100.0 100.0 100.0 100.0 Cut LER 78.4 100.0 100.0 75.0 57.1 83.3 37.5 SAM 43.5 83.3 60.0 54.5 23.1 0.0 16.7 (b) Results table. Figure 6: Proportion of position errors within 2 standard deviations of the theoretical bound (%), reported for different object sizes and methods. Results from table (b) are visualised in plot (a). 1. If the size of the objects that will be detected is known, to minimise the error on the learned object positions, one should aim to design the decoder receptive field size to be as small as possible while still encompassing the object. As the decoder RF grows beyond the object size, the error bound increases linearly with it up to a certain point (fig. 3b). 2. To minimise the error stemming from the encoder for a given object size, the encoder RF size should be kept as small as possible while still detecting the object (the RF size may be smaller than the object size), as again the error bound grows linearly with it up to a certain point (fig. 3a). 3. To minimise the error, the width of the rendering Gaussian should be kept as small as possible while still permitting gradient flow, as increasing it even slightly may result in a dramatic increase to the decoder term of the position error (fig. 3d). This is because, in practice, the decoder is able to detect parts of the Gaussian that are even 4 standard deviations away from its centre (fig. 4d). 4. In the case that one does not know a priori the exact size of the objects to be detected, one can still design a network that minimises the position errors for a given range of sizes. In that case, one should set up the decoder RF size to be as close as possible to the size of the largest object, and keep the encoder RF size as small as possible while still detecting all objects. The position errors for different object sizes will then be distributed according to the curve in fig. 3c, where the smallest and largest objects will achieve lowest errors and medium-size objects will achieve the greatest error, approximately given by a half of the average of the encoder and decoder RF sizes. Finally, we discuss some limitations of our method. Firstly, the method can only detect dynamic objects, for example if they move in a video or if they appear at multiple locations in images. Secondly, in its current form the method learns representations that can not be used for videos with different backgrounds than the one used at training time; however, this can be overcome by conditioning the decoder on an unrelated video frame instead of the positional encodings, as in Jakab et al. [11]. Thirdly, the guarantees of our method are conditional on the images being successfully reconstructed, which depends on the network architecture and optimisation method. 7 Conclusion We have presented the first unsupervised object detection method that is provably guaranteed to recover the true object positions up to small shifts. We proved that the object positions are learned up to a maximum error related to the encoder and decoder receptive field sizes, the object sizes, and bandwidth of the Gaussians used to render the objects. We derived expressions for how the position error depends on each of these factors and performed synthetic experiments that validated our theory up to sizes of individual pixels. We then performed experiments on CLEVR-based data, showing that unlike current SOTA methods, the position errors our method are always guaranteed to be within our theoretical bounds. We hope our work will provide a starting point for more research into object detection methods that possess theoretical guarantees, which are lacking in current practice. Acknowledgements. The authors acknowledge the generous support of the Royal Academy of Engineering (RF\201819\18\163), the Royal Society (RG\R1\241385) and EPSRC (Visual AI, EP/T028572/1). [1] Johann Brehmer, Pim De Haan, Phillip Lippe, and Taco Cohen. Weakly supervised causal representation learning. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho, editors, Advances in Neural Information Processing Systems, 2022. [2] Christopher P. Burgess, Loic Matthey, Nicholas Watters, Rishabh Kabra, Irina Higgins, Matt Botvinick, and Alexander Lerchner. Monet: Unsupervised scene decomposition and representation, 2019. [3] Abdelkader Dairi, Fouzi Harrou, Mohamed Senouci, and Ying Sun. Unsupervised obstacle detection in driving environments using deep-learning-based stereovision. Robotics and Autonomous Systems, 100:287 301, 2018. [4] Eran Goldman, Roei Herzig, Aviv Eisenschtat, Jacob Goldberger, and Tal Hassner. Precise detection in densely packed scenes. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 5227 5236, 2019. [5] Klaus Greff, Raphaël Lopez Kaufman, Rishabh Kabra, Nick Watters, Christopher Burgess, Daniel Zoran, Loic Matthey, Matthew Botvinick, and Alexander Lerchner. Multi-object representation learning with iterative variational inference. In International conference on machine learning, pages 2424 2433. PMLR, 2019. [6] Luigi Gresele, Julius von Kügelgen, Vincent Stimper, Bernhard Schölkopf, and Michel Besserve. Independent mechanism analysis, a new concept? In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 28233 28248. Curran Associates, Inc., 2021. [7] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In Proceedings of the IEEE international conference on computer vision, pages 2961 2969, 2017. [8] Kurt Hornik, Maxwell Stinchcombe, and Halbert White. Multilayer feedforward networks are universal approximators. Neural networks, 2(5):359 366, 1989. [9] Aapo Hyvarinen and Hiroshi Morioka. Unsupervised feature extraction by time-contrastive learning and nonlinear ica. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. [10] Aapo Hyvarinen and Hiroshi Morioka. Nonlinear ICA of Temporally Dependent Stationary Sources. In Aarti Singh and Jerry Zhu, editors, Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, volume 54 of Proceedings of Machine Learning Research, pages 460 469. PMLR, 20 22 Apr 2017. [11] Tomas Jakab, Ankush Gupta, Hakan Bilen, and Andrea Vedaldi. Unsupervised learning of object landmarks through conditional image generation. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. [12] Justin Johnson, Bharath Hariharan, Laurens Van Der Maaten, Li Fei-Fei, C Lawrence Zitnick, and Ross Girshick. Clevr: A diagnostic dataset for compositional language and elementary visual reasoning. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 2901 2910, 2017. [13] Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alexander C Berg, Wan-Yen Lo, et al. Segment anything. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 4015 4026, 2023. [14] Sebastien Lachapelle, Pau Rodriguez, Yash Sharma, Katie E Everett, Rémi LE PRIOL, Alexandre Lacoste, and Simon Lacoste-Julien. Disentanglement via mechanism sparsity regularization: A new principle for nonlinear ICA. In First Conference on Causal Learning and Reasoning, 2022. [15] Phillip Lippe, Sara Magliacane, Sindy Löwe, Yuki M Asano, Taco Cohen, and Stratis Gavves. CITRIS: Causal identifiability from temporal intervened sequences. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato, editors, Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pages 13557 13603. PMLR, 17 23 Jul 2022. [16] Francesco Locatello, Ben Poole, Gunnar Raetsch, Bernhard Schölkopf, Olivier Bachem, and Michael Tschannen. Weakly-supervised disentanglement without compromises. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 6348 6359. PMLR, 13 18 Jul 2020. [17] Francesco Locatello, Dirk Weissenborn, Thomas Unterthiner, Aravindh Mahendran, Georg Heigold, Jakob Uszkoreit, Alexey Dosovitskiy, and Thomas Kipf. Object-centric learning with slot attention. Advances in neural information processing systems, 33:11525 11538, 2020. [18] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In Medical image computing and computer-assisted intervention MICCAI 2015: 18th international conference, Munich, Germany, October 5-9, 2015, proceedings, part III 18, pages 234 241. Springer, 2015. [19] Thomas Schlegl, Philipp Seeböck, Sebastian M Waldstein, Ursula Schmidt-Erfurth, and Georg Langs. Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In International conference on information processing in medical imaging, pages 146 157. Springer, 2017. [20] Shohei Shimizu, Patrik O. Hoyer, Aapo Hyvärinen, and Antti Kerminen. A linear non-gaussian acyclic model for causal discovery. Journal of Machine Learning Research, 7(72):2003 2030, 2006. [21] Oriane Siméoni, Chloé Sekkat, Gilles Puy, Antonín Vobeck y, Éloi Zablocki, and Patrick Pérez. Unsupervised object localization: Observing the background to discover objects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3176 3186, 2023. [22] Oriane Siméoni, Éloi Zablocki, Spyros Gidaris, Gilles Puy, and Patrick Pérez. Unsupervised object localization in the era of self-supervised vits: A survey. International Journal of Computer Vision, pages 1 28, 2024. [23] Xudong Wang, Rohit Girdhar, Stella X Yu, and Ishan Misra. Cut and learn for unsupervised object detection and instance segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 3124 3134, 2023. [24] Quanhan Xi and Benjamin Bloem-Reddy. Indeterminacy in generative models: Characterization and strong identifiability. In Francisco Ruiz, Jennifer Dy, and Jan-Willem van de Meent, editors, Proceedings of The 26th International Conference on Artificial Intelligence and Statistics, volume 206 of Proceedings of Machine Learning Research, pages 6912 6939. PMLR, 25 27 Apr 2023. A Proof of Theorem 4.1 Theorem Maximum Position Error. Consider a set of images x X with objects of size so, CNN encoder ψ with receptive field size sψ, CNN decoder ϕ with receptive field size sϕ, soft argmax function softargmax, rendering function render with Gaussian standard deviation σG and G N(0, σ2 G), and latent variables z, composed as z = softargmax ψ x and ˆx = ϕ render z (fig. 1). Assuming (1) the objects are reconstructed at the same positions as in the original images, (2) each object appears in at least two different positions in the dataset, and (3) there are no two identical objects in any image, then the learned latent variables z correspond to the true object positions up to object permutations and maximum position errors of 2 + G . (10) Proof. By assumption (1), the positions (z1, z2) objects in the original image x have to be the same as the positions (ˆz1, ˆz2) of the objects in the reconstructed image ˆx. In practice, this occurs whenever the reconstruction loss is minimised. By assumption (2) (each object appears at a minimum of 2 different positions), the latent variables used by the decoder have to contain some information about each object, and thus the encoder has to learn to match all the objects. This is because the decoder CNN ϕ takes as its input the rendered Gaussians ˆe = render z concatenated with positional encodings or a randomly sampled nearby frame (fig. 1), and if some object in the dataset only appeared at a single position the model could achieve perfect reconstruction solely by using the positional encodings or the conditioned image without having to use the Gaussian maps ˆe. However, because the dataset contains each object at a minimum of 2 positions, relying purely on positional encodings or on the conditioned image is now not sufficient, as without any information about the object passed to the decoder, it would be impossible for it to predict where to render the object. More formally, because ˆx = ϕ render softargmax ψ x, this means that for objects in x to be reconstructed at the same positions as in ˆx (assumption 1), the encoder ψ needs to match some part of each object in x. Because z = softargmax ψ x is equivariant to translations of x (sec. 3), and because the encoder ψ has to match some part of each object in x (as shown previously), and also each image consists of distinct objects (assumption 3) on a known background, the image x with an object at the position (u1, u2) is encoded by softargmax ψ to the latent variables (z1, z2) = (u1 + ψ1, u2 + ψ2), | ψ1|, | ψ2| sψ where the shifts ψ1 and ψ2 arise because any part of the encoder filter (of size sψ) can match any part of the object (of size so). See fig. 2a for an illustration. Next, because ˆx = ϕ render z is equivariant to translations of z (sec. 3), the latent variables z = (u1 + ψ1, u2 + ψ2) are mapped to a predicted image ˆx = ϕ render z with an object at position (ˆz1, ˆz2) = (u1 + ψ1 + ϕ1, u2 + ψ2 + ϕ2), | ϕ1|, | ϕ2| sϕ where the shifts ϕ1 and ϕ2 arise because any part of the decoder filter (of size sϕ) can match any part of the the rendered Gaussian ˆet = render z, where G N(0, σ2 G). See fig. 2b for illustration. Finally, by assumption (1), the position of each object in the original image x has to be equal to the position of the object in the reconstructed image, i.e. (u1, u2) = (u1 + ψ1 + ϕ1, u2 + ψ2 + ϕ2) (13) This results in the conditions ψ1 + ϕ1 = 0, ψ2 + ϕ2 = 0 (14) and therefore | ψ1| = | ϕ1|, | ψ2| = | ϕ2| (15) In words, the shift in the latent variables acquired from the encoder ψ has to be balanced by an opposite shift of the same magnitude in the decoder ϕ in order to reconstruct the object at the same position. Because the shift due to the encoder is of maximum magnitude of sψ/2 + so/2 1 and the shift due to the decoder has a maximum magnitude of sϕ/2 so/2 + G, this means that the maximum magnitude of the shift of the latent variables has to be the minimum of these two expressions, i.e. the learned latent variables (z1, z2) correspond to the ground truth latent variables (u1, u2) up to (z1, z2) = (u1 + 1, u2 + 2), | 1|, | 2| min sψ 2 + G . (16) Additionally, because the order in which the objects get mapped to each latent variable is arbitrary, there is an additional indeterminacy arising due to variable permutations. B Theoretical Results for Multiple Object Sizes The results of corollary 4.2 can be extended to objects with a range of different sizes. The bound can be obtained by taking the maximum over all the bounds for objects with sizes so [smin o , smax o ], leading to the following corollary. Corollary B.1. Error vs. Encoder RF Size for Multiple Object Sizes. The maximum position error as a function of the encoder receptive field (RF) size sψ for a given sϕ, so [smin o , smax o ], σG, is 2 1 if 1 sψ sϕ 2smax o + 2, sψ 2 + G if sϕ 2smax o + 2 < sψ sϕ 2smin o + 2, sϕ 2 + G if sψ > sϕ 2smin o + 2. For an illustration see fig. 7a. Similar to corollary 4.2, the results of corollary 4.3 can be extended to objects with a range of different sizes by taking the maximum over the bounds for objects with sizes so [smin o , smax o ], leading to the following corollary. Corollary B.2. Error vs. Decoder RF Size for Multiple Object Sizes. The maximum position error as a function of the decoder receptive field (RF) size sϕ for a given sψ, so [smin o , smax o ], σG, is 2 + G if smin o sϕ sψ + 2smin o 2, sψ 2 + G if sψ + 2smin o 2 < sϕ sψ + 2smax o 2 sψ 2 1 if sϕ > sψ + 2smax o 2. For an illustration see fig. 7b. (a) Position error vs. encoder RF size for a range of object sizes. (b) Position error vs. decoder RF size for a range of object sizes. Figure 7: Theoretical bounds for the maximum position error as a function of the encoder receptive field size sψ and the decoder receptive field size sϕ for objects of sizes ranging from smin o to smax o , as the remaining factors are kept constant. Each theoretical bound consists of two regions separated by a dashed line, one where the maximum error is due to the encoder (deterministic, represented by a solid line), and one where the maximum error is due to the decoder (probabilistic, represented by a solid line and shaded regions). Areas within one and two standard deviations of the mean are represented by a darker and lighter shades of blue respectively. C Experiment Training Details C.1 Synthetic Experiments Dataset. In all of the synthetic experiments, we use the following setup. The dataset consists of black images of size simg simg = 80 80 px, each with a white square with dimensions so so px, centered at positions (x, y) where x, y {spad + so/2, spad + so/2 + 1, . . . , simg spad so/2}, where spad max(sψ, sϕ) 1 (to prevent unwanted edge effects). We divide the dataset into 4 quadrants and assign images from 3 quadrants to the training set and the remaining quadrant to the test set. Evaluation. For each experiment, we fix all but one variable from the following set: encoder RF size sψ, decoder RF size sϕ, object size so, and Gaussian standard deviation σG, and vary the remaining variable. For each value of the investigated variable we perform 20 experiments with different random seeds, noting down the learned position error as the absolute difference between the learned position z and the centre of the object z GT (maximum over horizontal and vertical differences, and over the test set). We discard a result if the reconstruction accuracy of the run is below 99.9%, to only consider runs where the object has been detected successfully (this is because the square is on the order of 7 7 px, thus only comprising 0.8% of the image). Architecture. We parametrise both the encoder ψ and the decoder ϕ as 5-layer CNNs with Batch Normalisation and Re LU activations, 32 channels, and filter sizes in {1 1, 3 3, 5 5, 7 7}, such that their receptive field sizes equal sψ and sϕ respectively. We train each network for 500 epochs using the Adam optimiser with learning rate 10 3 and batch size 128. We train each experiment on a single GPU for around 6 hours with <6GB memory on an internal cluster. C.2 CLEVR Experiments Dataset. Our CLEVR experiments use data generated with the CLEVR [12] image generation script. Our training and test sets consist of 150 and 50 images respectively, containing red, green and blue metallic spheres on a random background, at random positions, and with a different range of sizes (fig. 8). For experiments measuring position error as a function of encoder and decoder RF sizes and Gaussian s.d., we use a dataset with object sizes between 6-10 px after perspective distortion (fig. 8a). For the experiment measuring position error as a function of object size, we use 5 datasets with object sizes 9-14 px, 11-17 px, 13-19 px, 15-24 px, 17-27 px after perspective distortion (figs. 8b, 8c). (a) Object sizes 6-10 px. (b) Object sizes 9-14 px. (c) Object sizes 17-27 px. Figure 8: Samples from our CLEVR datasets, with object sizes (a) 6-10 px, (b) 9-14 px, (c) 17-27 px. Evaluation. For each experiment, we fix all but one variable from the following set: encoder RF size sψ, decoder RF size sϕ, object size so, and Gaussian standard deviation σG, and vary the remaining variable. For each value of the investigated variable we perform experiments with different random seeds, noting down the learned position errors as the absolute difference between the learned position z of each object and the centre of the object z GT (maximum over horizontal and vertical differences and over the test set, for each object). We only consider results for objects that have been learned successfully. For experiments measuring position error as a function of encoder and decoder RF sizes and Gaussian s.d., we consider an object to be learned if the position error is less than 35 px, and if only a single variable corresponds to the object, and if the position error is stable over consecutive training iterations. For the experiment measuring position error as a function of object size, we only consider runs where all 3 objects have been learned, i.e. where the reconstruction accuracy is higher than 98.0% for the dataset with object sizes 6-10 px, 98.8% for dataset with sizes 11-17 px, 98.0% for dataset with sizes 13-19 px, 97.5% for dataset with sizes 15-24 px, and 98.0% for dataset with sizes 17-27 px. Architecture. We parametrise both the encoder ψ and the decoder ϕ as 5-layer CNNs with Batch Normalisation and Re LU activations, 32 channels, and filter sizes in {1 1, 3 3, 5 5, 7 7}, such that their receptive field sizes equal sψ and sϕ respectively. We train each network until convergence using the Adam optimiser with learning rate 10 2 and batch size 128 for the experiments measuring position error as a function of encoder and decoder RF size and Gaussian s.d., and with learning rate 10 3 and batch size 150 for the experiment measuring position error as a function of object size. We train each model on a single GPU for less than day with <8GB memory on an internal cluster. Baselines. We also evaluate the results for two State-of-the-Art baselines, SAM [13] and Cut LER [23]. First, for each of our 6 datasets (containing objects of sizes 6-10 px, ..., 17-27 px), we combine its training and test set to create 4 sets of 50 images each. We then apply SAM and Cut LER to all images in each 50-image set, noting down the learned position errors as the absolute difference between the predicted position z of each object and the centre of the object z GT (maximum over horizontal and vertical differences and over the 50-image set, for each object). For SAM, we take the predicted object positions to be the centres of the bounding boxes corresponding to the second, third and fourth predicted masks (first mask corresponding to the background). For Cut LER, we take the predicted object positions to be the centres of the predicted bounding boxes if the method predicts 3 bounding boxes (one for each object), otherwise we discard the prediction. For both methods, we discard any result where position error is greater than 35 px, to be consistent with the evaluation for our method. Finally, this results in 12 position error values (3 objects 4 data splits), for each method and each of our 6 datasets. D CLEVR Experiments with Different Shapes To demonstrate that our method applies to objects of any shape, in this section we include experiments on our CLEVR dataset with three distinct objects a red metallic sphere, a blue rubber cylinder and a green rubber cube. The dataset has objects of size 9-19 px after perspective distortion, and a sample from the dataset is shown in fig. 9a. We perform training and evaluation in the same way as in app. C.2. We plot the position error as a function of the encoder and decoder receptive field sizes in figs. 9b and 9c, respectively. We can observe that all our experimental data (red) lies within the bounds predicted by our theory (blue), successfully validating our theory for objects with different shapes. (a) Dataset Sample. (b) Error vs. Encoder RF. (c) Error vs. Decoder RF. Figure 9: CLEVR experiment results using (a) a dataset with 3 objects of different shapes (a sphere, a cube, and a cylinder), showing position error as a function of (b) the encoder receptive field size sψ and (c) decoder receptive field size sϕ, as the remaining factors are fixed to sψ = 9, sϕ = 25, object size so [9, 19] and Gaussian s.d. σG = 0.8. Theoretical bounds are denoted by blue, experimental results in red, SAM baseline in green, and Cut LER baseline in orange. E Experiments with Real Videos In this section we present experimental results of applying our method to real You Tube videos, including overhead traffic footage and mini pool game footage. E.1 Traffic Data In this experiment we aim to learn the position of a car from an overhead traffic video. The training and test sets for this experiment consist of 25 frames each, from a video of an overhead view of a car moving for a short distance in a single lane (fig. 10a). We train the architecture from fig. 1 on this training set and validate it on the test set. It achieves a mean squared error between the ground truth object positions and the learned object positions of 7.2 10 5 (in units normalised by the image size), demonstrating that the object position has been learned successfully with a very low error. We then modify the learned position variable and decode it to generate videos of the car at novel positions (figs. 10b, 10c). (a) Training data. (b) Generated data (steady speed in a different lane). (c) Generated data (lane change and acceleration). Figure 10: Road traffic video used for training, together with two videos generated after training by modifying and decoding the learned latent variables (video frames are superimposed). The car object is detected successfully and is used to generate realistic videos with objects at unseen positions. E.2 Mini Pool In this experiment we aim to learn the positions of balls from a video of a game of mini pool. The training and test sets for this experiment consist of 15 and 11 frames respectively from a video of a game of mini pool, cropped to a portion where two balls are moving at the same time (fig. 11a). We train the architecture from fig. 1 on this training set and validate it on the test set, where it achieves a mean squared error between the ground truth object positions and the learned object positions of 8.2 10 3 (in normalised units). This demonstrates that the object positions were learned successfully with a very low error. We then modify the learned position variables and decode them to generate videos of the balls at novel positions (figs. 11b, 11c). In practice, it was important to set the encoder and decoder receptive field sizes to be greater than but close to the size of the objects, as for larger RF sizes the position error increased unnecessarily and the images were rendered further away from the position given by the latent variables. Also, for large receptive field sizes, the decoder filter contained too much background which caused low quality reconstructions when rendering the balls near the mini pool table edges. (a) Training data. (b) Generated data (linear motion at unseen positions). (c) Generated data (collision and slowing down). Figure 11: Mini pool video used for training, together with two videos generated after training by modifying and decoding the learned latent variables (video frames are superimposed). Both ball objects are detected successfully and are used to generate realistic videos. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: We clearly state our contributions and scope in the abstract and introduction. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: We discuss the limitations in sec. 6. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: We provide proof assumptions and sketch in sec. 4 and full proof in appendix A. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: We provide training details in Appendix C. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [No] Justification: We will release the code upon publication. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: We provide experimental setting details in Appendix C. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: We provide uncertainty ranges for all Gaussian distributions and error bars for data points where appropriate (figs. 3, 4, 5). Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: We provide information about compute resources in appendix C. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: We preserve anonymity in the paper submission and adhere to other points in the Neur IPS Code of Ethics. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: Our paper focuses on fundamental research and theory and does not present direct malicious/unintended uses etc. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: Our models and data do not pose risks for misuse. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [NA] Justification: Our paper does not use existing assets; we are the authors of all code. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: Our paper does not release new assets. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: Our paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: Our paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.