# unsupervised_discovery_of_objectcentric_neural_fields__d04f4370.pdf Published in Transactions on Machine Learning Research (02/2025) Unsupervised Discovery of Object-Centric Neural Fields Rundong Luo rundongluo@cs.cornell.edu Cornell University Hong-Xing Yu koven@cs.stanford.edu Stanford University Jiajun Wu jiajunwu@cs.stanford.edu Stanford University Reviewed on Open Review: https: // openreview. net/ forum? id= Sc Ev13W2f1 We study inferring 3D object-centric scene representations from a single image. While recent methods have shown potential in unsupervised 3D object discovery, they are limited in generalizing to unseen spatial configurations. This limitation stems from the lack of translation invariance in their 3D object representations. Previous 3D object discovery methods entangle objects intrinsic attributes like shape and appearance with their 3D locations. This entanglement hinders learning generalizable 3D object representations. To tackle this bottleneck, we propose the unsupervised discovery of Object-Centric neural Fields (u OCF), which integrates translation invariance into the object representation. To allow learning object-centric representations from limited real-world images, we further introduce a learning method that transfers object-centric prior knowledge from a synthetic dataset. To evaluate our approach, we collect four new datasets, including two real kitchen environments. Extensive experiments show that our approach significantly improves generalization and sample efficiency, and enables unsupervised 3D object discovery in real scenes. Notably, u OCF demonstrates zero-shot generalization to unseen objects from a single real image. The project page is available at https://red-fairy.github.io/u OCF/. 1 Introduction Creating factorized, object-centric 3D scene representations is a fundamental ability in human vision and a long-standing topic of interest in computer vision and machine learning. Some recent work has explored unsupervised learning of 3D factorized scene representations from images alone (Stelzner et al., 2021; Yu et al., 2022; Smith et al., 2023; Jia et al., 2023). These methods have delivered promising results in 3D object discovery and reconstruction from a simple synthetic image. However, existing methods fail to generalize to unseen spatial configurations and objects. A fundamental bottleneck is that their representations lack the invariance to the 3D positions of the objects. In particular, existing methods represent 3D objects as implicit functions in the viewer s coordinate frame, so that any change related the coordinate frame (e.g., slight changes in an object s location or subtle camera movements) may lead to significant changes in the object representation even if the object remains the same. Therefore, existing methods do not generalize when an object appears at an unseen location during inference. To address this fundamental bottleneck, we propose the unsupervised discovery of Object-Centric neural Fields (u OCF). Unlike existing methods, u OCF explicitly infers an object s 3D location, disentangling it from the object s latent representation. This design builds translation invariance into the object representation, so Equal Contribution. Work done while R. Luo was a visiting student at Stanford University. Published in Transactions on Machine Learning Research (02/2025) Reconstruction (ours) Unseen input image Rearrangement (ours) Reconstruction (u ORF) Insertion (ours) Figure 1: We propose the unsupervised discovery of Object-Centric neural Fields (u OCF), which infers factorized 3D scene representations from an unseen real image, thus enabling scene reconstruction and manipulation from novel views. We compare u OCF with the state-of-the-art method, u ORF (Yu et al., 2022). that the object s latent only represents the intrinsics of the object (e.g., shape and appearance). This design significantly improves generalization. As showcased in Figure 1, u OCF can generalize to unseen real-world scenes. We train u OCF on sparse multi-view images without object annotations. During inference, u OCF takes in a single image and generates a set of object-centric neural radiance fields (Ne RFs) (Mildenhall et al., 2020) and a background Ne RF. Another advantage of our translation-invariant 3D object representation is that it facilitates learning 3D object priors from simple scenes and generalizes to more complex scenes with unseen spatial configurations and objects. This further boosts sample efficiency and thus it is particularly beneficial when we deal with real scenes where training data is often limited. We introduce an object prior learning method to this end. To evaluate our approach, we introduce new challenging datasets for 3D object discovery, including two real kitchen datasets and two synthetic room datasets. The two real datasets feature real-world kitchen backgrounds and objects from multiple categories. The synthetic room datasets feature furniture with diverse, realistic shapes and textures. Across all these datasets, u OCF yields high-fidelity discovery of object-centric neural fields, allowing applications such as unsupervised 3D object segmentation and scene manipulation from a real image. u OCF shows strong generalization to unseen spatial configurations and high sample efficiency, and we showcase that it even allows zero-shot 3D object discovery on a few simple real scenes with unseen objects. In summary, our contributions are threefold: First, we highlight the overlooked role of translation invariance in unsupervised 3D object discovery. We instantiate the idea by proposing the unsupervised discovery of Object-Centric neural Fields (u OCF), which builds translation invariance to the object representation. Second, we introduce a 3D object prior learning method, which leverages u OCF s translation-invariant property to learn category-agnostic object priors from simple scenes and generalize to different object categories and scene layouts. Lastly, we collect four challenging datasets, Room-Texture, Room-Furniture, Kitchen-Matte, and Kitchen Shiny, and show that u OCF significantly outperforms existing methods on these datasets, unlocking zero-shot, single-image object discovery. All code and data will be made public. 2 Related Works Unsupervised 2D object discovery. Before the advent of deep learning, traditional methods for object discovery (often referred to as co-segmentation) primarily focused on locating visually similar objects across Published in Transactions on Machine Learning Research (02/2025) Input image Feature map Input view rotation Query point position Volume Rendering Density-Weighted Comp. Reconstruction (a) Encoder (c) Object Ne RF Decoder Relative PE Cross-Attn. Global residual (b) Latent Inference Module Learnable object queries Inferred object latents Momentum Update Figure 2: With a single forward pass, u OCF processes a single image input to infer a set of object-centric radiance fields along with their 3D locations and background radiance field. u OCF is trained on sparse multi-view images from a collection of scenes and uses a single image as input during inference. a collection of images (Sivic et al., 2005; Russell et al., 2006), where objects are defined as visual words or clusters of patches (Grauman & Darrell, 2006; Joulin et al., 2010). This clustering concept was later incorporated into deep learning techniques for improved grouping results (Li et al., 2019; Vo et al., 2020). The incorporation of deep probabilistic inference propelled the field towards factorized scene representation learning (Eslami et al., 2016). These methods decompose a visual scene into several components, where objects are often modeled as latent codes that can be decoded into image patches (Kosiorek et al., 2018; Crawford & Pineau, 2019; Jiang et al., 2020; Lin et al., 2020), phase values (Löwe et al., 2022; Gopalakrishnan et al., 2024), scene mixtures (Greff et al., 2016; 2017; 2019; Burgess et al., 2019; Engelcke et al., 2019; Locatello et al., 2020; Biza et al., 2023; Didolkar et al., 2023), or layers (Monnier et al., 2021). Notably, Seitzer et al. (2023) leveraged DINO features to extract robust image representations, while Daniel & Tamar (2022; 2023) introduced approaches to disentangle object latents into 2D position and appearance attributes. Despite their efficacy in scene decomposition, they do not model the objects 3D nature. Unsupervised 3D object discovery. To capture the 3D nature of scenes and objects, recent works have explored learning 3D-aware representations from point clouds (Wang et al., 2022), videos (Henderson & Lampert, 2020), and multi-view images of either single scenes (Liang et al., 2022) or large datasets for generalization (Eslami et al., 2018; Chen et al., 2020; Sajjadi et al., 2022). More recent research focuses on inferring object-centric factorized scene representations from single images (Stelzner et al., 2021; Yu et al., 2022; Smith et al., 2023). Among these, Yu et al. (2022) proposed a method for the unsupervised discovery of object radiance fields (u ORF) from single images. Follow-up works (Smith et al., 2023) improved rendering efficiency by replacing Ne RF with light fields (Sitzmann et al., 2021) or enhanced segmentation accuracy through bi-level query optimization (Jia et al., 2023). However, these representations often lack translation invariance, which limits their robustness and generalization capabilities. In this work, we address this limitation by incorporating translation invariance, resulting in significant improvements in both generalization and sample efficiency. For a more comprehensive review of related methods, we refer readers to Villa-Vásquez & Pedersoli (2024). Object-centric 3D reconstruction. Decomposing visual scenes on an object-by-object basis and estimating their semantic/geometric attributes has been explored in several recent works (Wu et al., 2017; Yao et al., 2018; Kundu et al., 2018; Ost et al., 2021). Some approaches, such as Auto RF (Müller et al., 2022), successfully reconstruct specific objects (e.g., cars) from annotated images. Others decompose visual scenes into the background and individual objects represented by neural fields (Yang et al., 2021; Wu et al., 2022). Our work differs because of its emphasis on unsupervised learning. Another line of recent work focuses on lifting 2D segmentation to reconstructed 3D scenes (Fan et al., 2022; Cen et al., 2023a;b). In contrast, our work aims at single-image inference, whereas these studies concentrate on multi-view reconstruction. Generative neural fields. Neural fields have revolutionized 3D scene modeling. Early works have shown promising geometric representations (Sitzmann et al., 2019; Park et al., 2019). The seminal work on neural Published in Transactions on Machine Learning Research (02/2025) radiance fields (Mildenhall et al., 2020) has opened up a burst of research on neural fields. We refer the reader to recent survey papers (Tewari et al., 2020; Xie et al., 2022) for a comprehensive overview. In particular, compositional generative neural fields such as GIRAFFE (Niemeyer & Geiger, 2021) and others (Nguyen Phuoc et al., 2020; Wang et al., 2023b) also allow learning object representations from image collections. Yet, they target unconditional generation and cannot tackle inference. Given a single input image, our goal is to infer object-centric radiance fields (i.e., each discovered object is represented in its local object coordinate rather than the world or the viewer coordinates) and the objects 3D locations. The object-centric design not only boosts generalizability due to representation invariance, but also allows learning object priors from scenes with different spatial layouts and compositional configurations. The following provides an overview of our approach and then introduces the technical details. 3.1 Model Overview As shown in Figure 2, u OCF consists of an encoder, a latent inference module, and a decoder. Encoder. From an input image I, the encoder extracts a feature map f RN C, where N = H W is the spatial size of the feature map and C represents the number of channels. We set it as a frozen DINOv2-Vi T (Oquab et al., 2023) followed by two convolutional layers. Latent inference module. The latent inference module infers the latent representation and position of the objects in the underlying 3D scene from the feature map. We assume that the scene is composed of a background environment and no more than K foreground objects. Therefore, the output includes a background latent zb R1 D and a set of foreground object latent zf = [zf T 1 zf T 2 zf T K ]T RK D with their corresponding positions {pwd i }K i=1 , where pwd i R3 denotes a position in the world coordinate. Note that some object latent may be empty when the scene has < K objects. Decoder. Our decoder employs the conditional Ne RF formulation g(x|z), which takes the 3D location x and the latent z as input and generates the radiance color and density for rendering. We use two MLPs, gb and gf, for the background environment and the foreground objects, respectively. 3.2 Object-Centric 3D Scene Modeling Object-centric latent inference. Our Latent Inference Module (LIM) aims at binding a set of learnable object queries ( qf = [qf T 1 qf T 2 qf T K ]T RK D) to the visual features of each foreground object, and another query to the background features (qb R1 D). The binding is modeled via the cross-attention mechanism with learnable linear functions Kb, Kf, Qb, Qf, Vb, Vf: Ai,j = exp(Mi,j) P k exp(Mi,k), where M = 1 Qb(qb) Kb(f)T Qf(qf) Kf(f)T T RN (K+1). (1) We then calculate the update signals for queries via an attention-weighted mean of the input: ub = (W(:,1))T Vb(f) R1 D; uf = (W(:,2:))T Vf(f) RK D, (2) where Wi,j = Ai,j P l Al,j is the normalized attention map. Queries are then updated by: qb qb + ub, qf qf + uf; qb qb + tb(qb), qf qf + tf(qf), (3) where tb and tf are MLPs. We repeat this procedure for T iterations, followed by concatenating the updated object queries with the corresponding attention-weighted mean of the input feature map f (global residual), finally delivering the background latent zb and foreground latent {zf i}K i=1. Our LIM is related to the Slot Attention (Locatello et al., 2020) while differs in several critical aspects. We discuss their relationship in Appendix C.1. Object location inference. To infer objects position along with their latent representation, we assign a normalized image position pimg i [ 1, 1]2 initialized as zero to each foreground object query, then Published in Transactions on Machine Learning Research (02/2025) Stage 1: Learn 3D object prior from synthetic scenes with simple composition Stage 2: Learn to discover objects from scenes with diverse object category and spatial layout Figure 3: Our object-centric design allows learning 3D object priors that generalize across different scene configurations. We first train our model to learn 3D object priors on simple synthetic scenes (e.g., single synthetic object), and then we leverage the 3D object priors to learn to discover objects in more complex scenes with different object categories and spatial layouts. Note that no object annotation is needed in either stage. iteratively update them by momentum m with the attention-weighted mean over the normalized 2D grid Eabs [ 1, 1]N 2: pimg i (W(:,i+1))T Eabs (1 m) + pimg i m. (4) To incorporate the inferred positions, we adopt the relative positional encoding (Biza et al., 2023) Epos i := concat([Eabs pimg i , pimg i Eabs]) RN 4, i {1, 2, , K}, and Epos 0 := concat([Eabs, Eabs]), where concat is the concatenation along the last dimension. Then, we re-write M in Eq. (1) as: Qb(qb) Kb(f + h1(Epos 0 ))T Qf(qf 1) Kf(f + h1(Epos 1 ))T Qf(qf K) Kf(f + h1(Epos K ))T where h1 : R4 RD is a linear function. Overall, LIM achieves a gradual binding between the queries and the objects in the scene through an iterative update of the queries and their locations. To address potential issues of duplicate object identification, we invalidate one of two similar object queries with high similarity and positional proximity by the start of the last iteration. Finally, a small bias term is added to the position to handle potential occlusion, i.e., pimg i pimg i + tanh(h2((W(:,i+1))T )) α , where scaling hyperparameter α = 0.2 and h2 : RN R2 is a linear function. The 2D positions pimg i are then unprojected into the 3D world coordinate to obtain pwd i . To do this, we extend the rays by depth d si, where d is the depth estimated by a monocular depth estimator (Ranftl et al., 2022) and {si}K i=1 are scaling terms predicted by a linear layer using the camera parameters and object latent as input. Compositional neural rendering. The object positions allow us to put objects in their local coordinates rather than the viewer or world coordinates, thereby obtaining object-centric neural fields. Technically, for each 3D point x in the world coordinate, we transform it to the ith object s local coordinate by xi = R (x pwd i ), where R denotes the input camera rotation matrix. We then retrieve the color and density of x in the foreground radiance fields as (ci, σi) = gf(xi|zf i) and in the background radiance field as (c0, σ0) = gb(x|zb). These values are aggregated into the scene s composite density and color (c, σ) using density-weighted means: i 0 ωiσi, c = X i 0 ωici, where ωi = σi P j 0 σj . (6) Finally, we compute the pixel color by volume rendering. Our pipeline is trivially differentiable, allowing backpropagation through all parameters simultaneously. Discussion on extrinsics disentanglement. An object s canonical orientation is ambiguous without assuming its category (Wang et al., 2019). Thus, we choose not to disentangle objects orientation since we target category-agnostic object discovery. Further, we observe that u OCF has learned meaningful representations that can smoothly interpolate an object s scale and orientation. Please refer to Appendix B for visualization and analysis. Published in Transactions on Machine Learning Research (02/2025) (d) Kitchen-Shiny (a) Room-Texture (c) Kitchen-Matte (b) Room-Furniture Figure 4: Samples from our collected datasets, where Room-Texture and Room-Furniture consist of synthetic images, and Kitchen-Matte and Kitchen-Shiny consist of real photos. 3.3 Object Prior Learning Unsupervised discovery of 3D objects in complex scenes is inherently difficult due to multiple challenging ambiguities. A major ambiguity is what defines an object. While existing methods define objects via visual appearance similarity (Yu et al., 2022) or priors from 2D segments (Chen et al., 2024), they suffer from under-segmentation due to visual cluttering (Yu et al., 2022) or over-segmentation inherited from the 2D supervision (Chen et al., 2024). We explore addressing this challenge by learning 3D object priors from synthetic data. Existing methods have difficulties learning generalizable 3D object priors, as their object representation is sensitive to spatial configurations: a minor shift in camera pose or object location, rather than the object itself, can lead to drastic changes in the object representation. Thus, such learned object priors do not generalize when there are unseen spatial configurations. Our 3D object-centric representation mitigates this issue by translation invariance. In particular, we introduce 3D object prior learning. We show an illustration in Figure 3. The main idea is to pre-train u OCF on synthetic scenes that are constructed with a single object to ease the learning, similar to curriculum learning. After the pre-training stage, we proceed to training u OCF on the more complex scenes that may have different object categories and spatial layouts. Note that either training stage does not require any object annotation. The pre-training synthetic single-object dataset can be easily scaled up. 3.4 Model Training Object-centric sampling. To improve the reconstruction quality, we leverage an object s local coordinates to concentrate the sampled points in proximity to the object. Specifically, we start dropping distant samples from the predicted object positions after a few training epochs when the model has learned to distinguish the foreground objects and predict their positions. This approach enables us to quadruple the number of samples with the same amount of computation, leading to significantly improved robustness and visual quality. In both training stages, we train our model across scenes, each with calibrated sparse multi-view images. For each training step, the model receives an image as input, infers the objects latent representations and positions, renders multiple views from the input and reference poses, and compares them to the ground truth images to calculate the loss. Model supervision consists of the MSE reconstruction loss ℓrecon and the perceptual loss ℓperc (Johnson et al., 2016) between the reconstructed and ground truth images. In addition, we incorporate the depth ranking loss (Wang et al., 2023a) with pre-trained monocular depth estimators and background occlusion regularization (Yang et al., 2023) to minimize common floating artifacts in few-shot Ne RFs. The overall loss function is thus formulated as follows: L = ℓrecon + λpercℓperc + λdepthℓdepth + λoccℓocc. (7) We leave further architectural details and illustrations in Appendix C.1. 4 Experiments We evaluate our method on three tasks: unsupervised object segmentation in 3D, novel view synthesis, and scene manipulation in 3D. Below, we briefly describe the data collection process and experimental configurations, with additional details provided in Appendices C.2 and C.3. Sample code and data are included in the supplementary material, and we plan to release the full code and datasets for public use. Published in Transactions on Machine Learning Research (02/2025) Table 1: Object segmentation and view synthesis on Room-Texture and Room-Furniture. Room-Texture Room-Furniture Method Object segmentation Novel view synthesis Object segmentation Novel view synthesis ARI FG-ARI NV-ARI PSNR SSIM LPIPS ARI FG-ARI NV-ARI PSNR SSIM LPIPS u ORF (Yu et al., 2022) 0.670 0.093 0.578 24.23 0.711 0.254 0.686 0.497 0.556 27.49 0.780 0.258 BO-QSA (Jia et al., 2023) 0.697 0.354 0.604 25.26 0.739 0.215 0.682 0.479 0.579 27.29 0.774 0.261 COLF (Smith et al., 2023) 0.235 0.532 0.011 22.98 0.670 0.504 0.514 0.458 0.439 28.73 0.781 0.386 u OCF (ours) 0.785 0.563 0.704 28.85 0.798 0.136 0.861 0.739 0.808 29.77 0.830 0.127 GT segment. BO-QSA u OCF (ours) Input view Novel view Input view Novel view Figure 5: Scene segmentation qualitative results. Novel view images are for reference only. Data. We collect two synthetic datasets and two real-world datasets to evaluate our method. Examples of these datasets are shown in Figure 4. Room-Texture. Room-Texture features 324 object models from the armchair category of the ABO(Collins et al., 2022) dataset. Each scene includes 2 4 objects arranged on backgrounds randomly selected from a collection of floor textures. The dataset comprises 5,000 scenes for training and 100 for evaluation. Each scene is rendered from four viewpoints centered on the scene. Room-Furniture. In Room-Furniture, objects are selected from 1,425 ABO (Collins et al., 2022) models spanning seven categories, including bed , cabinet , chair , dresser , ottoman , sofa , and plant pot . Other configurations match that of Room-Texture. Kitchen-Matte. This dataset includes scenes featuring single-color matte dinnerware set against two types of backgrounds: a plain tabletop or a complex kitchen environment. The dataset contains 735 scenes for training and 102 for evaluation. Each scene includes 3 4 objects positioned randomly and is captured from three viewpoints (for tabletop scenes) or two viewpoints (for kitchen backdrops). Kitchen-Shiny. This dataset contains scenes with textured, shiny dinnerware. Similar to Kitchen-Matte, the first half of the dataset features a plain tabletop, while the latter half includes a kitchen background. The dataset consists of 324 scenes for training and 56 for evaluation. Published in Transactions on Machine Learning Research (02/2025) Table 2: Novel view synthesis on Kitchen-Shiny and Kitchen-Matte. Method Kitchen-Shiny Kitchen-Matte PSNR SSIM LPIPS PSNR SSIM LPIPS u ORF (Yu et al., 2022) 19.23 0.602 0.336 26.07 0.808 0.092 BO-QSA (Jia et al., 2023) 19.78 0.639 0.318 27.36 0.832 0.067 COLF (Smith et al., 2023) 18.30 0.561 0.397 20.68 0.643 0.236 u OCF (ours) 28.58 0.862 0.049 29.40 0.867 0.043 Table 3: Novel view synthesis on Kitchen-Shiny with a larger number of object queries K. PSNR SSIM LPIPS K = 4 28.58 0.862 0.049 K = 5 28.28 0.846 0.059 K = 6 28.04 0.848 0.058 K = 10 28.20 0.840 0.065 Input view Novel view u ORF GT u OCF (ours) COLF BO-QSA Input view Novel view Figure 6: Novel view synthesis qualitative results on Kitchen-Shiny (top) and Room-Furniture (bottom). Implementation details. To learn object priors, we generate a synthetic dataset of over 8,000 scenes. Each scene contains one object, sampled from a high-quality subset of Objaverse-LVIS (Deitke et al., 2023), placed against a room background. These objects span over 100 categories. The synthetic dataset is easy to generate and scalable, making it ideal for learning object priors for all our experiments. In the second stage, the number of foreground object queries is set to K = 4. We initialize the model with the pre-trained weights from the object prior learning stage and train it on multi-object scenes. Once trained, our model can perform direct inference on images with spatial configurations that differ from those seen during training. Additionally, our model can adapt to unseen environments through efficient test-time optimization (see Sec. 4.2 for details). Baselines. We compare our method against u ORF (Yu et al., 2022), BO-QSA (Jia et al., 2023), and COLF (Smith et al., 2023). To ensure a fair comparison, we increase the latent dimensions and training iterations for all methods. For baseline models, we retain their original implementation without incorporating our proposed object-centric learning stage to maintain consistency. Details on baselines enhanced with our object-centric learning stage can be found in Appendix D. Metrics. We report the PSNR, SSIM, and LPIPS metrics for novel view synthesis. For scene segmentation, we use three variants of the Adjusted Rand Index (ARI): the conventional ARI (calculated on all input image pixels), the Foreground ARI (FG-ARI, calculated on foreground input image pixels), and the Novel View ARI (NV-ARI, calculated on novel view pixels). All scores are computed on images of resolution 128 128. Published in Transactions on Machine Learning Research (02/2025) Table 4: Scene manipulation results on the Room-Texture dataset. Method Object Translation Object Removal PSNR SSIM LPIPS PSNR SSIM LPIPS u ORF (Yu et al., 2022) 23.65 0.654 0.284 23.81 0.664 0.282 BO-QSA (Jia et al., 2023) 25.21 0.700 0.226 24.58 0.698 0.247 u OCF (ours) 27.66 0.774 0.156 28.99 0.802 0.136 Object translation Input image u OCF (ours) u ORF BO-QSA u ORF u OCF (ours) BO-QSA Object removal Figure 7: Qualitative results of single-image 3D scene manipulation on the Kitchen-Shiny dataset. 4.1 Baseline Comparison on Multiple Tasks Unsupervised object segmentation in 3D. We evaluate the object discovery quality by object segmentation in 3D. We render a density map di for each latent i and assign each pixel p a segmentation label sp = arg max K i=0di p in the input view and novel views. We show our results in Table 1 and examples in Figure 5. From Table 1, we see that our u OCF outperforms all existing methods in all metrics. From Figure 5, we observe that no prior method can produce reasonable segmentation results in real-world Kitchen-Shiny scenes. Specifically, u ORF binds all objects to the background, resulting in empty object segmentation; BO-QSA fails to distinguish different object instances; COLF produces meaningless results on novel views. A fundamental issue in these methods is that they lack appropriate object priors to handle the ambiguity in disentangling multiple objects. In contrast, u OCF can discover objects in real-world scenes. Moreover, u OCF can handle scenes where objects occlude each other. We provide more visualization results in Appendix D. Novel view synthesis. We evaluate the scene and object reconstruction quality by novel view synthesis. For each test scene, we use a single image as input and other views as references. We show our results in Table 2 and examples in Figure 6. We also show additional results in Appendix D. Our method significantly surpasses the baselines in all metrics. Importantly, while previous methods often fail to distinguish foreground objects and thus produce blurry reconstruction of objects, our approach consistently produces high-fidelity scene and object reconstruction and novel view synthesis results. Scene manipulation in 3D. We further evaluate object discovery by single-image 3D scene manipulation. Since u OCF explicitly infers 3D locations of discovered objects, it readily supports: 1) object translation by modifying an object s position, and 2) object removal by excluding objects during compositional rendering. For quantitative evaluation, we create a test set by randomly selecting an object in each of the Room Texture scenes, and shift its position (object translation) or remove it (object removal). During inference, we determine the object to manipulate by selecting the object with the highest Io U score with the ground truth mask. As shown in Table 4, u OCF outperforms baselines across all metrics in both object translation and object removal due to its better performance in object discovery. We further show qualitative examples from the Kitchen-Shiny dataset in Figure 7. We observe that u ORF merges all objects into the background, and thus the manipulation results are identical to the original reconstruction; BO-QSA fails to distinguish foreground objects, resulting in blurry manipulation results (we show more visualization in Appendix D). Published in Transactions on Machine Learning Research (02/2025) Input image Reconstruction Objects Novel view 324 training scenes 100 training scenes 50 training scenes 10 training scenes Figure 8: Qualitative results on sample efficiency. With fewer training scenes, our u OCF can still produce reasonable object discovery thanks to the object-centric modeling and learned object priors. u OCF (ours) BO-QSA u ORF GT u OCF (ours) BO-QSA u ORF (a) Room-Texture à HM3D (b) Room-Texture à Phone capture Figure 9: Zero-shot generalization results. We load the model trained on one dataset and test it on an image from another dataset after a fast test-time optimization on the input view only. First/second/third row: scene reconstruction/novel view/objects. In contrast, our u OCF delivers much higher-quality manipulation results. We show additional visualization results in the supplementary video. 4.2 Generalization Analysis In the experiments above, all test scenes have unseen novel spatial configurations, where u OCF shows strong generalization. We further evaluate the sample efficiency on spatial generalization, and we showcase the generalization to unseen objects. Published in Transactions on Machine Learning Research (02/2025) Input view Novel view u OCF (ours) u OCF (variant 1) u OCF (variant 2) u OCF (variant 3) Figure 10: Ablation on translation invariance and object prior learning. Three variants: (1) without both translation invariance and object prior learning, (2) without translation invariance, (3) without object prior learning. Sample efficiency. We train u OCF with a small subset of (e.g., only 10) the training scenes, and test it on the test set. As shown by the qualitative example in Figure 8, even when we only have a few training scenes, u OCF still demonstrates a good generalization ability to discover objects. This is mainly due to the translation invariance and learned object priors, which reduce the dependence on massive training scenes. Generalization to unseen objects. We evaluate the zero-shot generalization ability of u OCF by training it on one dataset and test it on a single image of unseen background and objects. Specifically, we test our model on five real-world examples (one from the HM3D dataset (Ramakrishnan et al., 2021) and four captured with a phone) using a model trained solely on the synthetic Room-Texture dataset. As shown in Figures 9 and , existing methods struggle to adapt to novel objects in unseen settings. In contrast, u OCF demonstrates remarkable generalizability, requiring only a lightweight single-image test-time optimization for 1000 iterations. This process takes approximately 3 minutes, a fraction of the 6 days needed for the full training of the model. These results highlight u OCF s ability to adapt effectively from synthetic training data to real-world scenarios. Details on this experiment can be found in Appendix C.3. 4.3 Ablation Study Key technical contributions. We conduct ablation studies to analyze the impact of our key technical contributions: the translation-invariant design and object prior learning. As shown in Table 5 and Figure 10, incorporating translation invariance dramatically reduces the LPIPS metric from 0.186 to 0.049. Similarly, leveraging object prior learning significantly decreases LPIPS from 0.125 to 0.049. These results demonstrate that both contributions are not only individually impactful but also complementary. Removing either one severely degrades performance, justifying their importance in achieving high-quality results. Other technical improvements. We also evaluate the impact of other technical improvements through ablation studies. As shown in Table 6, the inclusion of DINO Vi T and standard attention enhances overall performance. However, these components contribute relatively modestly; for instance, removing DINO or standard attention slightly increases LPIPS from 0.049 to 0.060 or 0.062, respectively. Excluding depth and occlusion losses also degrades visual quality, leading to a noticeable drop in performance. Additionally, removing the object-centric sampling strategy slightly reduces overall reconstruction quality, further highlighting the significance of these improvements. Different K values. We evaluate the effectiveness of different K values and different scales of data used for 3D object prior learning. We evaluate our method s robustness to different K values, and we show results in Table 3 and Figure 11. From Table 3, we can see that even when we set K = 10 which is much higher than the number of possible maximal objects (i.e., 4), our model is robust and gives comparable results. From Figure 11, we observe that even if there are more object queries than the number of objects in the scene, u OCF learns to generate empty object queries instead of over-segmenting the objects. Published in Transactions on Machine Learning Research (02/2025) Table 5: Ablation studies on our key technical contributions on Kitchen-Shiny. Method PSNR SSIM LPIPS w/o trans. invar. or object prior learning 20.68 0.645 0.303 w/o trans. invar. 23.70 0.724 0.186 w/o object prior learning 26.81 0.806 0.125 u OCF (ours) 28.58 0.862 0.049 Table 6: Ablation study on model architecture and loss function on Kitchen-Shiny. Method PSNR SSIM LPIPS w/o DINO 26.25 0.831 0.060 w/o standard attention 27.82 0.844 0.062 w/o object-centric sampling 27.31 0.852 0.072 w/o ℓdepth and ℓocc 26.79 0.819 0.081 u OCF (ours) 28.58 0.862 0.049 Input image Object 5 Object 6+ Reconstruction Object 3 Object 4 Object 1 Object 2 Figure 11: Qualitative results of u OCF on scenes with larger object queries K. The order of the object reconstructions is rearranged for better visualization. 5 Conclusion We study the importance of translation invariance for unsupervised 3D object discovery, instantiated as our model for the unsupervised discovery of Object-Centric neural Fields (u OCF). Our results show that our translation-invariant design and the 3D object prior learning can substantially improve the spatial generalization and sample efficiency. Our results demonstrate that unsupervised 3D object discovery can be extended to real scenes while obtaining satisfactory performances. Limitations. Although u OCF shows promising unsupervised 3D object discovery results, it is currently limited to simple real scenes such as the kitchen scenes. Extending to more complex real scenes with complex spatial layouts and a large number of objects from different categories is an important future direction. We leave more discussion on technical limitations in Appendix E. Acknowledgments This work is in part supported by NSF RI #2211258 and #2338203, ONR MURI N00014-22-1-2740, and ONR YIP N00014-24-1-2117. Jonathan T Barron, Ben Mildenhall, Matthew Tancik, Peter Hedman, Ricardo Martin-Brualla, and Pratul P Srinivasan. Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. In IEEE/CVF International Conference on Computer Vision (ICCV), 2021. Ondrej Biza, Sjoerd van Steenkiste, Mehdi SM Sajjadi, Gamaleldin F Elsayed, Aravindh Mahendran, and Thomas Kipf. Invariant slot attention: Object discovery with slot-centric reference frames. In International Conference on Machine Learning (ICML), 2023. Christopher P Burgess, Loic Matthey, Nicholas Watters, Rishabh Kabra, Irina Higgins, Matt Botvinick, and Alexander Lerchner. Monet: Unsupervised scene decomposition and representation. ar Xiv:1901.11390, 2019. Jiazhong Cen, Jiemin Fang, Chen Yang, Lingxi Xie, Xiaopeng Zhang, Wei Shen, and Qi Tian. Segment any 3d gaussians. ar Xiv:2312.00860, 2023a. Published in Transactions on Machine Learning Research (02/2025) Jiazhong Cen, Zanwei Zhou, Jiemin Fang, Wei Shen, Lingxi Xie, Xiaopeng Zhang, and Qi Tian. Segment anything in 3d with nerfs. In Advances in Neural Information Processing Systems (Neur IPS), 2023b. Chang Chen, Fei Deng, and Sungjin Ahn. Learning to infer 3d object models from images. ar Xiv:2006.06130, 2020. Honglin Chen, Wanhee Lee, Hong-Xing Yu, Rahul Mysore Venkatesh, Joshua B Tenenbaum, Daniel Bear, Jiajun Wu, and Daniel LK Yamins. Unsupervised 3d scene representation learning via movable object inference. Transactions on Machine Learning Research (TMLR), 2024. Jasmine Collins, Shubham Goel, Kenan Deng, Achleshwar Luthra, Leon Xu, Erhan Gundogdu, Xi Zhang, Tomas F Yago Vicente, Thomas Dideriksen, Himanshu Arora, Matthieu Guillaumin, and Jitendra Malik. Abo: Dataset and benchmarks for real-world 3d object understanding. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Eric Crawford and Joelle Pineau. Spatially invariant unsupervised object detection with convolutional neural networks. In AAAI Conference on Artificial Intelligence (AAAI), 2019. Tal Daniel and Aviv Tamar. Unsupervised image representation learning with deep latent particles. In International Conference on Machine Learning (ICML), 2022. Tal Daniel and Aviv Tamar. Ddlp: Unsupervised object-centric video prediction with deep dynamic latent particles. Transactions on Machine Learning Research (TMLR), 2023. Matt Deitke, Dustin Schwenk, Jordi Salvador, Luca Weihs, Oscar Michel, Eli Vander Bilt, Ludwig Schmidt, Kiana Ehsani, Aniruddha Kembhavi, and Ali Farhadi. Objaverse: A universe of annotated 3d objects. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023. Aniket Didolkar, Anirudh Goyal, and Yoshua Bengio. Cycle consistency driven object discovery. ar Xiv:2306.02204, 2023. Martin Engelcke, Adam R Kosiorek, Oiwi Parker Jones, and Ingmar Posner. Genesis: Generative scene inference and sampling with object-centric latent representations. ar Xiv:1907.13052, 2019. SM Eslami, Nicolas Heess, Theophane Weber, Yuval Tassa, David Szepesvari, Koray Kavukcuoglu, and Geoffrey E Hinton. Attend, infer, repeat: Fast scene understanding with generative models. In Advances in Neural Information Processing Systems (Neur IPS), 2016. SM Ali Eslami, Danilo Jimenez Rezende, Frederic Besse, Fabio Viola, Ari S Morcos, Marta Garnelo, Avraham Ruderman, Andrei A Rusu, Ivo Danihelka, Karol Gregor, et al. Neural scene representation and rendering. Science, 2018. Zhiwen Fan, Peihao Wang, Yifan Jiang, Xinyu Gong, Dejia Xu, and Zhangyang Wang. Nerf-sos: Any-view self-supervised object segmentation on complex scenes. ar Xiv:2209.08776, 2022. Anand Gopalakrishnan, Aleksandar Stanić, Jürgen Schmidhuber, and Michael Curtis Mozer. Recurrent complex-weighted autoencoders for unsupervised object discovery. In Advances in Neural Information Processing Systems (Neur IPS), 2024. Kristen Grauman and Trevor Darrell. Unsupervised learning of categories from sets of partially matching image features. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2006. Klaus Greff, Antti Rasmus, Mathias Berglund, Tele Hotloo Hao, Jürgen Schmidhuber, and Harri Valpola. Tagger: Deep unsupervised perceptual grouping. In Advances in Neural Information Processing Systems (Neur IPS), 2016. Klaus Greff, Sjoerd Van Steenkiste, and Jürgen Schmidhuber. Neural expectation maximization. In Advances in Neural Information Processing Systems (Neur IPS), 2017. Published in Transactions on Machine Learning Research (02/2025) Klaus Greff, Raphaël Lopez Kaufman, Rishabh Kabra, Nick Watters, Christopher Burgess, Daniel Zoran, Loic Matthey, Matthew Botvinick, and Alexander Lerchner. Multi-object representation learning with iterative variational inference. In International Conference on Machine Learning (ICML), 2019. Paul Henderson and Christoph H Lampert. Unsupervised object-centric video generation and decomposition in 3d. In Advances in Neural Information Processing Systems (Neur IPS), 2020. Yicong Hong, Kai Zhang, Jiuxiang Gu, Sai Bi, Yang Zhou, Difan Liu, Feng Liu, Kalyan Sunkavalli, Trung Bui, and Hao Tan. Lrm: Large reconstruction model for single image to 3d. ar Xiv preprint ar Xiv:2311.04400, 2023. Baoxiong Jia, Yu Liu, and Siyuan Huang. Improving object-centric learning with query optimization. In International Conference on Learning Representations (ICLR), 2023. Jindong Jiang, Sepehr Janghorbani, Gerard de Melo, and Sungjin Ahn. Scalor: Generative world models with scalable object representations. In International Conference on Learning Representations (ICLR), 2020. Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and superresolution. In European Conference on Computer Vision (ECCV), 2016. Armand Joulin, Francis Bach, and Jean Ponce. Discriminative clustering for image co-segmentation. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2010. Adam R Kosiorek, Hyunjik Kim, Ingmar Posner, and Yee Whye Teh. Sequential attend, infer, repeat: Generative modelling of moving objects. In Advances in Neural Information Processing Systems (Neur IPS), 2018. Abhijit Kundu, Yin Li, and James M Rehg. 3d-rcnn: Instance-level 3d object reconstruction via render-andcompare. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2018. Bo Li, Zhengxing Sun, Qian Li, Yunjie Wu, and Anqi Hu. Group-wise deep object co-segmentation with coattention recurrent neural network. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Shengnan Liang, Yichen Liu, Shangzhe Wu, Yu-Wing Tai, and Chi-Keung Tang. Onerf: Unsupervised 3d object segmentation from multiple views. ar Xiv:2211.12038, 2022. Zhixuan Lin, Yi-Fu Wu, Skand Vishwanath Peri, Weihao Sun, Gautam Singh, Fei Deng, Jindong Jiang, and Sungjin Ahn. Space: Unsupervised object-oriented scene representation via spatial attention and decomposition. In International Conference on Learning Representations (ICLR), 2020. Francesco Locatello, Dirk Weissenborn, Thomas Unterthiner, Aravindh Mahendran, Georg Heigold, Jakob Uszkoreit, Alexey Dosovitskiy, and Thomas Kipf. Object-centric learning with slot attention. In Advances in Neural Information Processing Systems (Neur IPS), 2020. Sindy Löwe, Phillip Lippe, Maja Rudolph, and Max Welling. Complex-valued autoencoders for object discovery. Transactions on Machine Learning Research (TMLR), 2022. Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In European Conference on Computer Vision (ECCV), 2020. Tom Monnier, Elliot Vincent, Jean Ponce, and Mathieu Aubry. Unsupervised layered image decomposition into object prototypes. In IEEE/CVF International Conference on Computer Vision (ICCV), 2021. Norman Müller, Andrea Simonelli, Lorenzo Porzi, Samuel Rota Bulo, Matthias Nießner, and Peter Kontschieder. Autorf: Learning 3d object radiance fields from single view observations. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Published in Transactions on Machine Learning Research (02/2025) Thu H Nguyen-Phuoc, Christian Richardt, Long Mai, Yongliang Yang, and Niloy Mitra. Blockgan: Learning 3d object-aware scene representations from unlabelled images. In Advances in Neural Information Processing Systems (Neur IPS), 2020. Michael Niemeyer and Andreas Geiger. Giraffe: Representing scenes as compositional generative neural feature fields. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021. Maxime Oquab, Timothée Darcet, Théo Moutakanni, Huy Vo, Marc Szafraniec, Vasil Khalidov, Pierre Fernandez, Daniel Haziza, Francisco Massa, Alaaeldin El-Nouby, et al. Dinov2: Learning robust visual features without supervision. ar Xiv:2304.07193, 2023. Julian Ost, Fahim Mannan, Nils Thuerey, Julian Knodt, and Felix Heide. Neural scene graphs for dynamic scenes. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021. Jeong Joon Park, Peter Florence, Julian Straub, Richard Newcombe, and Steven Lovegrove. Deepsdf: Learning continuous signed distance functions for shape representation. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Santhosh K Ramakrishnan, Aaron Gokaslan, Erik Wijmans, Oleksandr Maksymets, Alex Clegg, John Turner, Eric Undersander, Wojciech Galuba, Andrew Westbury, Angel X Chang, et al. Habitat-matterport 3d dataset (hm3d): 1000 large-scale 3d environments for embodied ai. ar Xiv:2109.08238, 2021. René Ranftl, Katrin Lasinger, David Hafner, Konrad Schindler, and Vladlen Koltun. Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer. Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2022. Bryan C Russell, William T Freeman, Alexei A Efros, Josef Sivic, and Andrew Zisserman. Using multiple segmentations to discover objects and their extent in image collections. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2006. Mehdi SM Sajjadi, Daniel Duckworth, Aravindh Mahendran, Sjoerd van Steenkiste, Filip Pavetic, Mario Lucic, Leonidas J Guibas, Klaus Greff, and Thomas Kipf. Object scene representation transformer. In Advances in Neural Information Processing Systems (Neur IPS), 2022. Maximilian Seitzer, Max Horn, Andrii Zadaianchuk, Dominik Zietlow, Tianjun Xiao, Carl-Johann Simon Gabriel, Tong He, Zheng Zhang, Bernhard Schölkopf, Thomas Brox, et al. Bridging the gap to real-world object-centric learning. In International Conference on Learning Representations (ICLR), 2023. Vincent Sitzmann, Michael Zollhöfer, and Gordon Wetzstein. Scene representation networks: Continuous 3d-structure-aware neural scene representations. In Advances in Neural Information Processing Systems (Neur IPS), 2019. Vincent Sitzmann, Semon Rezchikov, Bill Freeman, Josh Tenenbaum, and Fredo Durand. Light field networks: Neural scene representations with single-evaluation rendering. In Advances in Neural Information Processing Systems (Neur IPS), 2021. Josef Sivic, Bryan C Russell, Alexei A Efros, Andrew Zisserman, and William T Freeman. Discovering objects and their location in images. In IEEE International Conference on Computer Vision (ICCV), 2005. Cameron Smith, Hong-Xing Yu, Sergey Zakharov, Fredo Durand, Joshua B Tenenbaum, Jiajun Wu, and Vincent Sitzmann. Unsupervised discovery and composition of object light fields. Transactions on Machine Learning Research (TMLR), 2023. Karl Stelzner, Kristian Kersting, and Adam R Kosiorek. Decomposing 3d scenes into objects via unsupervised volume segmentation. ar Xiv:2104.01148, 2021. Ayush Tewari, Ohad Fried, Justus Thies, Vincent Sitzmann, Stephen Lombardi, Kalyan Sunkavalli, Ricardo Martin-Brualla, Tomas Simon, Jason Saragih, Matthias Nießner, et al. State of the art on neural rendering. Computer Graphics Forum, 2020. Published in Transactions on Machine Learning Research (02/2025) José-Fabian Villa-Vásquez and Marco Pedersoli. Unsupervised object discovery: A comprehensive survey and unified taxonomy. ar Xiv preprint ar Xiv:2411.00868, 2024. Huy V Vo, Patrick Pérez, and Jean Ponce. Toward unsupervised, multi-object discovery in large-scale image collections. In European Conference on Computer Vision (ECCV), 2020. Guangcong Wang, Zhaoxi Chen, Chen Change Loy, and Ziwei Liu. Sparsenerf: Distilling depth ranking for few-shot novel view synthesis. In IEEE/CVF International Conference on Computer Vision (ICCV), 2023a. He Wang, Srinath Sridhar, Jingwei Huang, Julien Valentin, Shuran Song, and Leonidas J Guibas. Normalized object coordinate space for category-level 6d object pose and size estimation. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Qian Wang, Yiqun Wang, Michael Birsak, and Peter Wonka. Blobgan-3d: A spatially-disentangled 3d-aware generative model for indoor scenes. ar Xiv:2303.14706, 2023b. Tianyu Wang, Miaomiao Liu, and Kee Siong Ng. Spatially invariant unsupervised 3d object-centric learning and scene decomposition. In European Conference on Computer Vision (ECCV), 2022. Jiajun Wu, Joshua B Tenenbaum, and Pushmeet Kohli. Neural scene de-rendering. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2017. Qianyi Wu, Xian Liu, Yuedong Chen, Kejie Li, Chuanxia Zheng, Jianfei Cai, and Jianmin Zheng. Objectcompositional neural implicit surfaces. In European Conference on Computer Vision (ECCV), 2022. Yiheng Xie, Towaki Takikawa, Shunsuke Saito, Or Litany, Shiqin Yan, Numair Khan, Federico Tombari, James Tompkin, Vincent Sitzmann, and Srinath Sridhar. Neural fields in visual computing and beyond. Computer Graphics Forum, 2022. Bangbang Yang, Yinda Zhang, Yinghao Xu, Yijin Li, Han Zhou, Hujun Bao, Guofeng Zhang, and Zhaopeng Cui. Learning object-compositional neural radiance field for editable scene rendering. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021. Jiawei Yang, Marco Pavone, and Yue Wang. Freenerf: Improving few-shot neural rendering with free frequency regularization. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023. Shunyu Yao, Tzu Ming Harry Hsu, Jun-Yan Zhu, Jiajun Wu, Antonio Torralba, William T Freeman, and Joshua B Tenenbaum. 3d-aware scene manipulation via inverse graphics. In Advances in Neural Information Processing Systems (Neur IPS), 2018. Alex Yu, Vickie Ye, Matthew Tancik, and Angjoo Kanazawa. pixelnerf: Neural radiance fields from one or few images. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021. Hong-Xing Yu, Leonidas J Guibas, and Jiajun Wu. Unsupervised discovery of object radiance fields. In International Conference on Learning Representations (ICLR), 2022. Published in Transactions on Machine Learning Research (02/2025) A Appendix Overview This supplementary document is structured as follows: We begin with the proof of concept in Appendix B and provide the implementation details in Appendix C. Then, we discuss the limitations of our approach in Appendix E and present additional qualitative results in Appendix D. Accompanying this document is our project page with an overview video attached in the supplementary file. B Proof of Concept We conduct a toy experiment (Figure 12) to demonstrate that our model has successfully learned object position, rotation, and scale. In this experiment, we begin with two images (input 1 and input 2) of a chair placed at the scene s center, exhibiting different sizes (on the left) or rotation angles (on the right), all captured from the same viewing direction. We extract the object latents from these images, interpolate them, and then send the interpolated latents to the decoder. As shown between the two input images, we observe a smooth transition in object size and rotation, indicating that the latent representation has effectively captured the scale and rotation of objects. In the second row, we placed the chairs in different positions. As shown on the right, we obtained a smooth transition again, proving that our model could disentangle object positions from the latent representation. Input 1 Input 2 Latent interpolation size: big small rotation: 0 𝜋/2 Input 1 Rotation: 𝜋/2 Position (0, 0) Input 2 Rotation: 0 Position (0, 0) Latent linear interpolation & position translation Input 1 Input 2 Latent interpolation Figure 12: Proof of concept. We demonstrate that u OCF has effectively learned objects scale and orientation along with the translation-invariant object representation by interpolating the representation of two identical objects with different orientations and scales to obtain transitional results. C Implementation C.1 Model Architecture Encoder. Our encoder module consists of a frozen DINO encoder and two convolutional layers. We illustrate its architecture in Figure 13(a). Latent inference module. While motivated by the background-aware slot attention module proposed by (Yu et al., 2022), our latent inference module exhibits three key differences: (1) The object queries are initialized with learnable embeddings instead of being sampled from learnable Gaussians, which enhances training stability; (2) We jointly extract object positions and their latent representations and add objectspecific positional encoding to utilize the extracted position information; (3) We remove the Gated Recurrent Unit (GRU) and replace it with the transformer architecture to smooth the gradient flow. C.2 Data Collection This section introduces the details of our datasets. Published in Transactions on Machine Learning Research (02/2025) Resize to 876*876 (Frozen) DINOv2 Vi T (876*876*3à64*64*768) 3*3 conv, stride=1, Re LU (64*64*768à64*64*64) Feature map 𝑓 (a) Encoder architecture 3*3 conv, stride=1 (64*64*768à64*64*64) Dropped samples Remaining samples Object bounding box (b) Object-centric sampling Figure 13: (a) Architecture of our encoder module. (b) Object-centric sampling: We drop the samples distant from the predicted object position for efficient sampling.. Room-Texture. In Room-Texture, objects are chosen from 324 ABO objects (Collins et al., 2022) from the armchair category. The single-object subset contains four scenes for each object instance, resulting in 1296 scenes in total. The multiple-object subset includes 5, 000 scenes for training and 100 for evaluation, with each scene containing 2-4 objects set against a background randomly chosen from a collection of floor textures. Each scene is rendered from 4 directions toward the center. Room-Furniture. In Room-Furniture, objects are chosen from 1, 425 ABO (Collins et al., 2022) object models, spanning across seven categories, including bed , cabinet , chair , dresser , ottoman , sofa , and plant pot . Each scene contains 2-4 objects set against a background randomly chosen from a collection of floor textures. We render 5000 scenes for training and 100 scenes for evaluation. Kitchen-Matte. In Kitchen-Matte, objects are diffuse and have no texture. The dataset comprises 16 objects and 6 tablecloths in total. We captured 3 images for each tabletop scene and 2 for each kitchen scene. This dataset contains 735 scenes for training and 102 for evaluation, each containing 3-4 objects. We calibrate the cameras using the Open CV library. Kitchen-Shiny. In Kitchen-Shiny, objects are specular, and the lighting is more complex. The dataset comprises 12 objects and 6 tablecloths, and the other settings are identical to Kitchen-Matte. This dataset contains 324 scenes for training and 56 for evaluation, each containing 4 objects. C.3 Training Configuration This section discusses the training configuration of u OCF. We employ Mip-Ne RF (Barron et al., 2021) as our Ne RF backbone and estimate the depth maps by Mi Da S (Ranftl et al., 2022). An Adam optimizer with default hyper-parameters and an exponential decay scheduler is used across all experiments. The initial learning rate is 0.0003 for the first stage and 0.00015 for the second stage. Loss weights are set to λperc = 0.006, λdepth = 1.5, and λocc = 0.1. The position update momentum m is set to 0.5, and the latent inference module lasts T = 6 iterations. Most hyperparameters are inherited from Yu et al. (2022), while the loss weights of our proposed losses are chosen to ensure a similar scale between all loss terms. All experiments are run on a single RTX-A6000 GPU. Training lasts approximately 6 days, including 1.5 days for object prior learning and another 4.5 days for training on multi-object scenes. Coarse-To-Fine progressive training. We employ a coarse-to-fine strategy in our second training stage to facilitate training at higher resolutions. Reference images are downsampled to a lower resolution (64 64) Published in Transactions on Machine Learning Research (02/2025) during the coarse training stage and replaced by image patches with the same size as the low-resolution images randomly cropped from the high-resolution (128 128) input images during the fine training stage. Locality constraint and object-centric sampling. We employ the locality constraint (a bounding box for foreground objects in the world coordinate) proposed by (Yu et al., 2022) in both training stages but only adopt it before starting object-centric sampling. The number of samples along each ray before and after starting object-centric sampling is set to 64 and 256, respectively. We provide an illustration of our object-centric sampling strategy in Figure 13(b). Training configuration on Room-Texture. During stage 1, we train the model for 100 epochs directly on images of resolution 128 128. We start with the reconstruction loss only, add the perceptual 10th epoch, and start the object-centric sampling at the 20th epoch. During stage 2, we train the model for 60 epochs on the coarse stage and another 60 on the fine stage. We start with the reconstruction loss only, add the perceptual loss at the 10th epoch, and start the object-centric sampling from the 20th epoch. Training configuration on Kitchen-Matte and Kitchen-Shiny. Both kitchen datasets share the same training configuration with Room-Texture in stage 1. During stage 2, we train the model for 750 epochs, where the fine stage starts at the 250th epoch. We add the perceptual loss at the 50th epoch and start the object-centric sampling from the 150th epoch. Training configuration for zero-shot generalization. For the test-time adaptation, we fine-tune our model on the input view only using our proposed loss function (Eq. 7) for 1000 iterations on resolution 128 128. We use the Adam optimizer and the learning rate is set to 1 10 4. This optimization takes about 3 minutes on a single A6000 gpu. D Additional Experiments Additional real-world dataset. We introduce an additional real-world dataset, named Planters, which features tabletop scenes containing four plant pots or vases arranged on tablecloths. The dataset includes 745 scenes for training and 140 scenes for evaluation, with each scene captured from three different camera poses. As shown in the quantitative results in Table 7 and the qualitative results in Figure 17, our method significantly outperforms existing approaches. It achieves superior scene reconstruction and novel view synthesis, delivering results with noticeably higher visual quality. Method PSNR SSIM LPIPS u ORF 24.49 0.748 0.163 u ORF-BO-QSA 28.09 0.847 0.108 COLF 19.22 0.588 0.464 u OCF (ours) 29.00 0.864 0.062 Table 7: Quantitative results on the Planters dataset. Baseline performance with object-centric learning. To ensure fair comparisons, we conducted additional experiments where object prior learning was incorporated into the baseline methods. The results in Table 8 demonstrate that even with the incorporation of object prior learning, u OCF significantly outperforms existing methods due to its translation-invariant object representation, which enhances generalization and data efficiency. Additional zero-shot evaluation. We evaluate our method on two additional phone-captured scenes to further demonstrate the generalizability of u OCF. To ensure fair comparisons, we incorporate object prior learning into the baseline methods. However, as shown in Figure 15, the baseline methods still struggle generalizing to unseen environments. In contrast, our method produces accurate object segmentation and novel view synthesis results, further validating its effectiveness. Visualization on discovered objects. We visualize the discovered objects in Figure 16. Notably, u ORF (Yu et al., 2022) puts all objects within the background, whereas BO-QSA (Jia et al., 2023) binds the same object to all queries, resulting in identical foreground reconstruction. In contrast, u OCF accurately differentiates between the foreground objects and the background. Published in Transactions on Machine Learning Research (02/2025) Method LPIPS SSIM PSNR u ORF 0.336 0.602 19.23 u ORF + object prior learning 0.193 0.714 22.78 BO-QSA 0.318 0.639 19.78 BO-QSA + object prior learning 0.129 0.766 24.00 COLF 0.397 0.561 18.30 COLF + object prior learning 0.290 0.709 21.66 u OCF (ours) 0.049 0.862 28.58 Table 8: Baseline performance with object-centric learning. Our method maintains its superiority even when baselines methods employ our proposed object-prior learning approach due to its translation-invariance representation. Visualization on object segmentation in 3D. We show scene segmentation results on the kitchen datasets in Figure 18. Unlike compared methods that yield cluttered results, u OCF consistently yields high-fidelity segmentation results. Additional novel view synthesis results. We show more qualitative results for novel view synthesis in Figures 18, 19, and 20. Our method produces much better results than compared methods regarding visual quality. E Limitations Analysis Limitation on reconstruction quality. Scene-level generalizable Ne RFs (Yu et al., 2021; Sajjadi et al., 2022; Yu et al., 2022) commonly face challenges in accurately reconstructing detailed object textures. Our approach also has difficulty capturing extremely high-frequency details. As shown in Figure 14(a), our method fails to replicate the mug s detailed texture. Future research may benefit from stronger object priors learned from larger-scale datasets, such as Large Reconstruction Models (Hong et al., 2023). Failure in position prediction. Our two-stage training pipeline, despite its robustness in many situations, is not immune to errors, particularly in object position prediction. Due to the occlusion between objects, using the attention-weighted mean for determining object positions can sometimes lead to inaccuracies. Although a bias term can rectify this in most instances (Figure 6), discrepancies persist under a few conditions, as depicted in Figure 14(b). Training instability. Like other slot-based object discovery methods, our approach also faces challenges with training instability. For example, the model occasionally collapses within the first few training epochs, even when using identical hyperparameters. To address this, we perform multiple trials for each experiment, terminating early if signs of collapse appear during the initial training stages. We observed that approximately half of the experiments fail at this stage. However, once the model progresses beyond this critical phase, the final results remain consistent across different trials. For baseline methods that struggle on our proposed complex datasets, we conduct at least five trials to ensure that their failure stems from limitations in their design rather than issues with random initialization. Published in Transactions on Machine Learning Research (02/2025) (a) Failure on reconstruct texture (b) Error in position prediction GT Reconstruction GT Reconstruction Figure 14: Failure case visualizations. Our method may fail to reconstruct intricate object texture or predict biased object position. u OCF (ours) u ORF BO-QSA u OCF (ours) u ORF BO-QSA Input image Input image Reconstruction Segmentation Object 1 Object 2 Object 3 Object 4 Novel view Figure 15: Qualitative zero-shot generalization results. Published in Transactions on Machine Learning Research (02/2025) Input image Input image Background Object 1 u OCF (ours) Object 3 Object 2 Object 4 u ORF BO-QSA u OCF (ours) Figure 16: Visualization on discovered objects on Kitchen-Shiny. Published in Transactions on Machine Learning Research (02/2025) u OCF (ours) Input view Novel view Input view Novel view Input view Novel view Input view Novel view Figure 17: Qualitative comparison results on the Planters dataset. Published in Transactions on Machine Learning Research (02/2025) GT u ORF COLF BO-QSA u OCF (ours) Input view Novel view Input view Novel view Input view Novel view Input view Novel view Figure 18: Additional segmentation and view synthesis results on the Room-Texture dataset. Published in Transactions on Machine Learning Research (02/2025) GT u ORF COLF BO-QSA u OCF (ours) Input view Novel view Input view Novel view Input view Novel view Input view Novel view Figure 19: Additional view synthesis results on the Kitchen-Matte dataset. Published in Transactions on Machine Learning Research (02/2025) Input view Novel view u OCF (ours) Input view Novel view Input view Novel view Input view Novel view Figure 20: Additional view synthesis results on the Kitchen-Shiny dataset.