# gsn_generalisable_segmentation_in_neural_radiance_field__7dab5d8b.pdf GSN: Generalisable Segmentation in Neural Radiance Field Vinayak Gupta 1, Rahul Goel2, Sirikonda Dhawal2, P. J. Narayanan2 1Indian Institute of Technology, Madras 2International Institute of Information Technology, Hyderabad vinayakguptapokal@gmail.com, {rahul.goel,dhawal.sirikonda}@research.iiit.ac.in, pjn@iiit.ac.in Traditional Radiance Field (RF) representations capture details of a specific scene and must be trained afresh on each scene. Semantic feature fields have been added to RFs to facilitate several segmentation tasks. Generalised RF representations learn the principles of view interpolation. A generalised RF can render new views of an unknown and untrained scene, given a few views. We present a way to distil feature fields into the generalised GNT representation. Our GSN representation generates new views of unseen scenes on the fly along with consistent, per-pixel semantic features. This enables multi-view segmentation of arbitrary new scenes. We show different semantic features being distilled into generalised RFs. Our multi-view segmentation results are on par with methods that use traditional RFs. GSN closes the gap between standard and generalisable RF methods significantly. Project Page: https://vinayak-vg.github.io/GSN/ Introduction Capturing, digitising, and authoring detailed structures of real-world scene settings is tedious. The digitised real-world scenes open many applications in graphics and augmented reality. Efforts to capture real-world scenes included special hardware setup (Narayanan, Rander, and Kanade 1998), structure from motion (Sf M) from unstructured collections (Snavely, Seitz, and Szeliski 2006; Agarwal et al. 2009), and recently to Neural Radiance Fields (Ne RF) (Mildenhall et al. 2020). Ne RF generates photo-realistic novel views using a neural representation learned from a set of discrete images of a scene. Follow-up efforts enhanced performance, reduced memory, etc. (Kobayashi, Matsumoto, and Sitzmann 2022; Huang et al. 2022; M uller et al. 2022; Chen et al. 2022a). Grid-based representations led to faster learning (Fridovich Keil and Yu et al. 2022; Sun, Sun, and Chen 2022; Chen et al. 2022a). These use scene-specific representations that do not generalise to unseen scenes. Pixel Ne RF (Yu et al. 2021), IBRNet (Wang et al. 2021b), MVSNe RF (Chen et al. 2021) and GNT (Varma et al. 2023) generalise RFs to unknown scenes by formulating novel Copyright 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Work done during an internship at IIIT Hyderabad. view synthesis as multi-view interpolation. Pixel Ne RF introduced a scene-class prior and used a CNN to render novel views. IBRNet used an MLP and a ray transformer to estimate radiance and volume density, collecting information on the fly from multiple source views. MVSNe RF builds on the work of MVSNet (Yao et al. 2018) to create a neural encoding volume based on the nearest source views. GNT improves the IBRNet formulation to leverage epipolar constraints using attention-based view transformers and ray transformers. Radiance Fields are trained using a few dozen views of a scene and memorise a specific scene. New views are generated from it using a volumetric rendering procedure. Generalised RF methods, on the other hand, learn the principles of view generation using epipolar constraints and interpolation of proximate views. They are trained using views from multiple scenes. Novel views of an unknown scene for any camera pose are generated by directly interpolating a few views. The distinction between learning a scene and learning a view-based rendering algorithm is significant. Users are often interested in segmenting parts of the captured 3D scene and manipulating them, besides new view generation. Initial segmentation efforts like N3F (Tschernezki et al. 2022) and DFF (Kobayashi, Matsumoto, and Sitzmann 2022) distilled a semantic field alongside a radiance field. More recent ISRF (Goel et al. 2023) leveraged proximity in 3D space and the distilled semantic space for interactive segmentation to extract good multi-view segmentation masks. Can we combine semantics into the generalised Radiance Fields to facilitate semantic rendering along with view generation of unseen scenes? In this paper, we extend the GNT formulation to include different semantic features efficiently. This makes the generation of consistent semantic features possible, along with generating new views of an unknown scene on the fly. Our generalisable representation can take advantage of semantic features like DINO (Caron et al. 2021), CLIP (Radford et al. 2021), and SAM (Kirillov et al. 2023), to facilitate more consistent segmentation of the scene in new views. The key points about our method are : Integrate semantic features into a generalised radiance field representation deeply, facilitating the rendering of per-pixel semantic features for each novel view. This is done without retraining or fine-tuning to new scenes. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Facilitate object segmentation, part segmentation, and other operations simultaneously for new views, exploiting the per-pixel semantic features. Provide segmentation quality similar to the best scenespecific segmentation methods for radiance fields. Allow for distilling multiple semantic fields in the generalised setting. This results in finer and cleaner semantic features that perform better than the original. Related Work Since Ne RF (Mildenhall et al. 2020) has come out, much research has been surrounding it. For a good overview of the study around radiance fields, we recommend the readers to refer to these excellent surveys (Tewari et al. 2022; Xie et al. 2022). Neural Radiance Fields Given a point (x, y, z) in 3D and a viewing direction specified using polar angles (θ, ϕ), a radiance field F maps these parameters to an RGB color(radiance) value: F(x, y, z, θ, ϕ) : R3 S2 R3. With assistance from a density field, these values can be accumulated along a ray shot through a pixel using the volumetric rendering equation Eq. (1). ˆC(r) = PK i=i Tiαici where αi = 1 e σiδi and Ti = Qi 1 j=1(1 αj). (1) In Eq. (1), for a sampled point i along a ray, δi is the inter-sample distance, Ti is the accumulated transmittance, and ci is the directional colour for the point. For more details about this formulation, please refer to Ne RF (Mildenhall et al. 2020). Ne RF models the density field and the radiance field using an MLP. Since the advent of Ne RFs, there have been various works to improve Ne RFs to handle reflections better (Verbin et al. 2022), to handle sparse input views (Niemeyer et al. 2022; Xu et al. 2022), improve quality in the case of unconstrained images (Martin-Brualla et al. 2021; Chen et al. 2022b), to model large scale scenes (Tancik et al. 2022; Turki, Ramanan, and Satyanarayanan 2022), in controllability for editing(Wang et al. 2022; Yuan et al. 2022) and to deal with dynamic scenes (Park et al. 2021; Fridovich-Keil et al. 2023). Several efforts have attempted to improve efficiency and rendering speed of radiance fields using dense grids (Sun, Sun, and Chen 2022), hash grids (M uller et al. 2022), decomposed grids (Chen et al. 2022a), and gaussians (Kerbl et al. 2023). However, all these methods perform learning/optimisation that is scene-specific and cannot generalise to new scenes. Generalised Neural Radiance Fields A fundamental shortcoming of most radiance field methods is that they need to be trained separately for each scene, i.e., they are scene-specific. Multiple attempts have been made to generalize them to wild scenes without training. MVSNe RF (Chen et al. 2021), a deep neural network to reconstruct radiance fields by using constraints from multi-view stereo. Their approach generalises across scenes by constructing a neural encoding volume using a cost volume built on the three nearest source views and uses an MLP to regress the corresponding density and colour per point on the ray. IBRNet (Wang et al. 2021b) uses an MLP and a ray transformer that estimates the volumetric density and radiance for a given 3D point from a view direction. Instead of learning the scene structure, the neural networks learn how the source views of the scene can be interpolated to produce a new view. This allows them to generalise to arbitrary scenes. GNT (Varma et al. 2023) improves IBRNet (Wang et al. 2021b) using a series of view transformers and ray transformers. They completely substitute the volumetric rendering equation with a ray transformer and show that it acts as a better feature aggregator and produces superior results for the task of novel view synthesis. Semantics in Radiance Fields Multi-view segmentation is a basic problem in computer vision. It could be a requirement for several downstream tasks (Wang et al. 2021a). Neural radiance fields are ideal for this task since they provide good-quality 3D reconstruction. Since we can render images from new views, it also enables segmentation from unseen views. Multiple efforts have successfully incorporated semantics into radiance fields. N3F (Tschernezki et al. 2022) showed incorporation of DINO (Caron et al. 2021) features into Ne RFs, which can be used to perform segmentation. Concurrently, DFF (Kobayashi, Matsumoto, and Sitzmann 2022) distilled DINO features as well as LSeg (Li et al. 2022) features into a radiance field and portrayed similar results. LERF (Kerr et al. 2023) brings CLIP features into RFs to enable language queries similar to DFF. NVOS (Ren et al. 2022) uses graph-cut to segment 3D neural volumetric representations. ISRF (Goel et al. 2023) also proposes distilling DINO features into a voxelised radiance field. They show state-of-the-art segmentation by leveraging the spatial growth of a 3D mask. However, these efforts suffer from the same issue: their representation and semantic distillation are scene-specific. The model needs to be retrained and re-distilled for a each new scene. Contrary to previous distillation works, we propose a method to distil features in a generalisable radiance field. This allows us to generate semantic feature images alongside RGB images for a new scene from any viewpoint. These semantic features can be used for several downstream tasks. In particular, we demonstrate their capability on a segmentation task. We also show that the features predicted by our generalised architecture are finer and cleaner than the original features of the images due to the flow of information from other nearby views. We begin by describing the Generalized Ne RF Transformer (GNT) (Varma et al. 2023). We modify the GNT architecture to aid our semantic feature distillation. We first describe The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Figure 1: Overview: Stage I: 1) We aggregate the features from the source views in View Transformer constrained by the epipolar geometry 2) The point aggregated features are passed on to the ray transformer along with input positions to aggregate the information along the ray. 3) The ray aggregated features and input view direction are passed onto an MLP and pooled to obtain pixel-wise colour. Stage II: 4) The view-independent features from the ray transformer are passed on to the stage-II block and aggregated by the view transformer and the ray transformer using the source view features extracted from the image using a pre-trained model like DINO. 5) The features out of the ray transformer are concatenated with input positions and pooled to predict pixel-wise features of the corresponding target-view pixel. our two-stage training-distillation procedure. Finally, we describe how we perform multi-view segmentation using the distilled features. Generalised Ne RF Transformer GNT (Varma et al. 2023) propose to solve the problem of generalization of radiance fields using the following two modules: 1. A view transformer aggregates the image features from nearby source views along epipolar lines to predict pointwise features. The role of this module is to interpolate features from nearby views. 2. A ray transformer accumulates these point-wise features, from the previous module, along a ray. This module acts as a substitute for the volumetric rendering equation. These two modules are stacked one after another sequentially for b = 4 blocks. The output of the last ray transformer is passed through an MLP (after average pooling), which directly predicts the pixel s colour through which the ray was shot. The following equations summarise the View Transformer and Ray-Transformer formulation of GNT: F(x, θ) = V (F1(Π1(x), θ), , FN(ΠN(x), θ)) R(o, d, θ) = R(F(o + t1d, θ), , F(o + t Md, θ)) C(r) = MLP Mean R(o, d, θ) Here, V ( ) is the view transformer encoder, Πi(x) projects position x R3 onto the ith image plane by applying extrinsic matrix, and Fi(z, θ) Rd computes the feature vector at position z R2 and viewing direction θ via bi-linear interpolation on the feature grids, r = o + td with t1, , t M being uniformly sampled between near and far plane, R( ) is a standard transformer encoder. F and R are the corresponding outputs of the view transformer and ray transformer, respectively, and C( ) is the predicted pixel colour. Modifications to GNT As explained in the previous section, the Ray-Transformer takes position and view direction as an input parameter. The repeated stacking of the View-Transformer and Ray Transformer means that the information of view direction is being passed very early to the architecture. However, as shown by ISRF (Goel et al. 2023), semantic features are view-agnostic. Looking at an object from a different view direction should not change the semantic features of the object. We remove the view direction as input to the Ray-Transformer to facilitate this. The view-direction input is re-introduced later in the RGB-prediction branch as shown in Fig. 1. Shifting this to a later stage degrades the RGB rendering quality, but very imperceptibly. On the positive side, we obtain a base architecture that generates view-independent feature images, significantly improving segmentation quality. Results related to this decision are shown in the ablation study section. Feature Distillation As shown by Fig. 1, we have a two-stage training mechanism: 1. Stage I: In the first stage, the GNT model with our proposed modification is trained to generate novel views on multiple scenes simultaneously using the RGB Loss LRGB = CGT CP red 2. 2. Stage II: In the second stage, we first obtain pixel-wise features for all the images across all the scenes. We branch off from our direction-independent base model and add another set of view and ray transformers that The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) CHESS TABLE COLOR FOUNTAIN STOVE SHOE RACK Figure 2: Comparison: Row 1 shows the reference scenes. Row 2 shows the segmentation results of N3F/DFF (Tschernezki et al. 2022; Kobayashi, Matsumoto, and Sitzmann 2022) with the corresponding patch query. Row 3 shows segmentation results of ISRF (Goel et al. 2023) with strokes. Row 4 shows the segmentation results of our GSN method. It is to be noted that the previous methods rely on scene-specific training to enable segmentation. For more details (highlighted boxes), please refer to the Results section in the manuscript. predict the features. The output features from this student feature head f P red are compared against the teacher feature images f GT to perform distillation: LF eat. = f GT f P red 2. Segmentation Similar to N3F, DFF and ISRF, we mainly use DINO (Caron et al. 2021) features to perform segmentation. As shown in Fig. 4, other feature sets may be used. Given a set of input user strokes, our task is to segment the marked object from all the different views. The preliminary methods like N3F (Tschernezki et al. 2022) propose to use average feature matching to perform segmentation in 3D during the rendering procedure. ISRF (Goel et al. 2023) significantly improves this by first clustering the features marked by the user s stroke and then doing 2D-3D NNFM (Nearest Neighbour Feature Matching). We follow a similar strategy as ISRF but in a multi-view setting. The semantic features underlying the user s input stroke are collected and are clustered into k clusters using K-Means clustering. Then, all the pixels in the feature image are classified using the k clusters as to whether they belong to the marked object. This is done using NNFM with all the k clusters. The threshold for the feature distance is identical for all the views and is chosen separately for each scene. Implementation Details We use the code provided by (Varma et al. 2023) and implement our methods on top of it. We use 2 RTX 3090 GPUs for distributed training and conducting our experiments. For extraction of features, we run PCA on every image on the scene to reduce the feature dimension to 64 and normalise these features. For Stage I training, we use four blocks of viewray transformers and one block of view-ray transformer in The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Original GNT GSN (Ours) DINO Features Clustered Features Figure 3: Row 1 shows DINO (Caron et al. 2021) features computed in various settings on HORNS from LLFF (Mildenhall et al. 2019). For visual simplification, a 3-dimensional PCA has been done on the features. Col. 1 and Col. 2 show DINO features computed on the original image and on GNT s (Varma et al. 2023) output respectively. Col. 3 shows features predicted by our GSN method. The boxes highlight clear feature differences. We demonstrate better feature quality by doing K-Means Clustering on the feature images as shown in Row 2. Our method gives clear, noise-free clusters as shown by the boxes. Scene Metric N3F ISRF Ours Mean Io U 0.344 0.912 0.828 Accuracy 0.820 0.990 0.981 m AP 0.334 0.916 0.833 Colorfountain Mean Io U 0.871 0.927 0.927 Accuracy 0.979 0.989 0.989 m AP 0.871 0.927 0.93 Mean Io U 0.416 0.827 0.839 Accuracy 0.954 0.992 0.993 m AP 0.387 0.824 0.835 Mean Io U 0.589 0.861 0.911 Accuracy 0.913 0.980 0.987 m AP 0.582 0.869 0.912 Table 1: This table shows the segmentation metrics against previous works of N3F (Tschernezki et al. 2022), DFF (Kobayashi, Matsumoto, and Sitzmann 2022), and ISRF (Goel et al. 2023). We calculate the mean Io U, accuracy and mean average precision for four scenes from the dataset provided by LLFF (Mildenhall et al. 2019). Our method performs better than the preliminary works of N3F and DFF while on par with ISRF. The ground truth segmentation masks were hand-annotated for comparison. Stage II training. We used 512 rays, 192 points and 200,000 iterations for Stage-I and trained it for two days. Stage II is trained only for 5,000 iterations with 512 rays and 192 points and trained only for 4 hours. We use a learning rate of 1e-3 for the Stage II training and decrease the learning rate of Stage-I by a factor of 10 during the Stage II training. The weight factor of the RGB loss LRGB is set to 0.1, and feature loss LF eat is set to 1 during Stage II training. We select 10 source views for every target view, and the training is done at an image resolution of 756x1008. It takes around 60 seconds to render an image. The dataset provided by IBRNet (Wang et al. 2021b) and some scenes from the LLFF Dataset (Mildenhall et al. 2019) are used for training our generalised model. For testing, we use the LLFF (Mildenhall et al. 2019) dataset, which was unseen during the training phase. These images are passed through DINO Vi T-b8 to obtain feature images. The number of clusters for segmentation is set to 11, although a slight variation does not make a difference. This section shows qualitative and quantitative results against previous works that perform multi-view segmentation using radiance fields. We demonstrate that our work improves the feature prediction while at the same time generalising by interpolating the features across multiple views. Our method can generalise to any set of features, and we show this by incorporating DINO (Caron et al. 2021), CLIP (Radford et al. 2021), SAM (Kirillov et al. 2023) and DINOv2 (Oquab et al. 2023) features. Note: Every result (for our method) shown in this paper/- supplementary is a scene not present in the training set, i.e. a previously unseen scene. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Figure 4: Other Semantic Fields: Col. 1 shows the FLOWER scene with DINOv2 (Oquab et al. 2023) features distilled into it. We show part segmentation of the flowers, i.e., the parts of each flower are coloured the same, depicting the distillation of appropriate features. Col. 2 shows the result on the FORTRESS scene when SAM (Kirillov et al. 2023) features are distilled into our GSN model, and the SAM decoder is used to segment the image. Col. 3 shows the distillation of CLIP (Radford et al. 2021) features into the T-REX scene. We use the text-prompt a fossil of dinosaur to localise the object in the rendered image. The heat map shows how well the pixel corresponds to the text prompt. This figure depicts that our generalised GSN model can incorporate various semantic features. We compare our results against the previous works of N3F (Tschernezki et al. 2022), DFF (Kobayashi, Matsumoto, and Sitzmann 2022), and ISRF (Goel et al. 2023). N3F and DFF use similar methods for their segmentation, so we treat them as one method as done by previous works (Goel et al. 2023). These works utilise feature distillation from 2D feature images into a 3D radiance field and use these distilled features to perform semantic segmentation. All these works require scene-specific training before segmentation can be achieved, while we propose a generalised framework. Despite this disadvantage, we perform on par with ISRF and better than N3F/DFF. Fig. 2 shows our segmentation results against N3F/DFF and ISRF. Due to an average feature matching method, the results of N3F/DFF are filled with noise alongside occasional bleeding-in of unmarked objects (CHESS TABLE ). ISRF uses K-Means Clustering followed by NNFM to improve the results significantly. K-Means Clustering helps in reducing the noise, and NNFM helps in the selection of the correct object. As explained before, we follow a similar methodology in 2D to achieve segmentation. Tab. 1 shows quantitative results on four scenes taken from the LLFF (Mildenhall et al. 2019) dataset. We calculate mean Io U, accuracy and mean average precision score on four scenes from the LLFF (Mildenhall et al. 2019) dataset. These metrics show that our method is better than N3F/DFF and comes close to the state-of-the-art method ISRF and, in some cases, even performs better than them. Please refer to the highlighted boxes in Fig. 2 for the following discussion. In CHESS TABLE scene, ISRF performs better than us. The area below the table is not segmented well because of fewer views of that area. Our method heavily relies on appropriate epipolar geometry prediction, and fewer views lead to bad epipolar geometry. Similarly, in COLOR FOUNTAIN scene, our segmentation suffers under the basin due to fewer views. ISRF cleverly counters this using their 3D region growing. In the STOVE scene, we are able to recover more content than ISRF. In the SHOE RACK scene, the white sole is recovered more in our case than compared to ISRF. An important observation needs to be noted. It can be seen that the N3F/DFF and ISRF show the part of the shoe that is occluded by the shoerack because they are doing 3D segmentation. Since we are doing multiview segmentation instead of 3D segmentation, the occluder is negated in our result. Student Surpasses Teacher There are three methods to obtain a good semantic feature representation for a given set of multi-view images. The first method is to directly pass the image through a feature extractor and get the features. If one wants features from an unseen The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Figure 5: Left and right images show segmentation results with original GNT and our GSN architectures respectively. We input the same single stroke and threshold for segmentation. In our case, the features are more coherent, leading to more accurate segmentation. view, one can use novel view synthesis methods on a scene (Varma et al. 2023; Mildenhall et al. 2020) and then pass the novel view through a feature extractor. We show results on these methods in Fig. 3 and compare them against our approach, which also produces features on novel views and that too in a generalised fashion. The figure shows the finegrained features that we produce. Essentially, our method can interpolate the coarse features from different views and combine them with the learnt geometry to produce finegrained features. Our student GSN method can surpass the teacher feature extractor methods to produce better features in a multi-view setting. The feature images shown in the figure are obtained by doing a PCA of the features to 3 dimensions which correspond to RGB channels of the image. For results on more scenes, please refer to the supplementary document. Other Semantic Fields The generalised feature distillation we propose can be applied to any set of features. We demonstrate this by distilling three different features using our method. We have already shown results on DINO(Caron et al. 2021) in Fig. 2. In Fig. 4, we show the distillation of DINOv2 (Oquab et al. 2023) features, CLIP (Radford et al. 2021) features, and SAM (Kirillov et al. 2023) features. We show part-segmentation using DINOv2, segmentation using CLIP space using a text prompt and segmentation using SAM features space by passing them through the SAM decoder. These results indicate that our generalisation approach can be adapted to any feature space by fine-tuning Stage II. For more details on the exact method, please refer to supplementary. Ablation Study Architecture Modification An essential modification to the original GNT architecture is displacing the view directions as input. We train the original GNT architecture and attempt to do segmentation with it using one user stroke. Fig. 5 shows the segmentation results obtained on the original GNT model and with our modification when the same user stroke is provided. In our case, the segmented object covers a more significant part of the Figure 6: The left and right images show the segmentation result when we using volumetric rendering and ray transformer (our method) respectively. In both cases, we tweak the threshold of segmentation to the point such that the chair (left of table) barely starts to bleed in. In case of ray transformer, results are improved indicating that it is a better feature aggregator. underlying object. In the case of the original GNT architecture, the semantic features are conditioned upon the view directions, which is inherently wrong since semantic features are view-agnostic. In common words, they are diffuse and not specular. This leads to incorrect feature distillation and poorer segmentation results in GNT architecture. Volumteric Rendering One may suspect that the volumetric-rendering variant mentioned in GNT (Varma et al. 2023) should give better results since it performs segmentation in 3D. We test this with our GSN method and observe that the results are subpar to our method, as shown in Fig. 6. This indicates that the raytransformer blocks are better aggregators for image-based rendering methods than volumetric rendering equation for density, colour and feature values. Conclusion We present a novel method for multi-view segmentation. The essential advantage of our approach is its generalisability, i.e. it can perform segmentation on arbitrarily new scenes without any training. This differentiates it from the previous methods. We compare our results against the earlier methods and show that we perform at par with them while being generalisable to unseen scenes. This is a big step in bringing the applications of generalisable neural radiance fields closer to scene-specific radiance fields. The features predicted by our method can be used for several downstream tasks. Limitations & Future Work. Since we rely on a transformer-based architecture, the rendering process is inherently slow compared to the several scene-specific radiance field methods. Improving the rendering speed can significantly improve the human interaction experience required for our stroke-based segmentation method. We leave the rendering speed improvement of generalisable radiance fields as future work. Currently, our method performs multiview segmentation since it uses image-based rendering. Some applications require 3D segmentation instead of multiview segmentation. Thus, a generalisable 3D segmentation framework is promising for future work. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Acknowledgements We thank Prof Kaushik Mitra of IIT Madras, for advice and computational support. We also thank Mukund Varma of UCSD for discussions and insights on GNT. References Agarwal, S.; Snavely, N.; Simon, I.; Seitz, S. M.; and Szeliski, R. 2009. Building Rome in a day. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Caron, M.; Touvron, H.; Misra, I.; J egou, H.; Mairal, J.; Bojanowski, P.; and Joulin, A. 2021. Emerging Properties in Self-Supervised Vision Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). Chen, A.; Xu, Z.; Geiger, A.; Yu, J.; and Su, H. 2022a. Tenso RF: Tensorial Radiance Fields. In Proceedings of the European Conference on Computer Vision (ECCV). Chen, A.; Xu, Z.; Zhao, F.; Zhang, X.; Xiang, F.; Yu, J.; and Su, H. 2021. Mvsnerf: Fast generalizable radiance field reconstruction from multi-view stereo. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). Chen, X.; Zhang, Q.; Li, X.; Chen, Y.; Feng, Y.; Wang, X.; and Wang, J. 2022b. Hallucinated neural radiance fields in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Fridovich-Keil, S.; Meanti, G.; Warburg, F. R.; Recht, B.; and Kanazawa, A. 2023. K-Planes: Explicit Radiance Fields in Space, Time, and Appearance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Fridovich-Keil and Yu; Tancik, M.; Chen, Q.; Recht, B.; and Kanazawa, A. 2022. Plenoxels: Radiance Fields without Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Goel, R.; Sirikonda, D.; Saini, S.; and Narayanan, P. 2023. Interactive Segmentation of Radiance Fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Huang, Y.-H.; He, Y.; Yuan, Y.-J.; Lai, Y.-K.; and Gao, L. 2022. Stylized Ne RF: Consistent 3D Scene Stylization as Stylized Ne RF via 2D-3D Mutual Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Kerbl, B.; Kopanas, G.; Leimk uhler, T.; and Drettakis, G. 2023. 3D Gaussian Splatting for Real-Time Radiance Field Rendering. ACM Transactions on Graphics. Kerr, J.; Kim, C. M.; Goldberg, K.; Kanazawa, A.; and Tancik, M. 2023. LERF: Language Embedded Radiance Fields. In International Conference on Computer Vision (ICCV). Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A. C.; Lo, W.- Y.; Doll ar, P.; and Girshick, R. 2023. Segment Anything. ar Xiv:2304.02643. Kobayashi, S.; Matsumoto, E.; and Sitzmann, V. 2022. Decomposing Ne RF for Editing via Feature Field Distillation. In Adv. Neural Inform. Process. Syst. Li, B.; Weinberger, K. Q.; Belongie, S.; Koltun, V.; and Ranftl, R. 2022. Language-driven Semantic Segmentation. In Int. Conf. Learn. Represent. Martin-Brualla, R.; Radwan, N.; Sajjadi, M. S. M.; Barron, J. T.; Dosovitskiy, A.; and Duckworth, D. 2021. Ne RF in the Wild: Neural Radiance Fields for Unconstrained Photo Collections. In CVPR. Mildenhall, B.; Srinivasan, P. P.; Ortiz-Cayon, R.; Kalantari, N. K.; Ramamoorthi, R.; Ng, R.; and Kar, A. 2019. Local Light Field Fusion: Practical View Synthesis with Prescriptive Sampling Guidelines. ACM Trans. Graph. Mildenhall, B.; Srinivasan, P. P.; Tancik, M.; Barron, J. T.; Ramamoorthi, R.; and Ng, R. 2020. Ne RF: Representing Scenes as Neural Radiance Fields for View Synthesis. In Proceedings of the European Conference on Computer Vision (ECCV). M uller, T.; Evans, A.; Schied, C.; and Keller, A. 2022. Instant Neural Graphics Primitives with a Multiresolution Hash Encoding. ACM Trans. Graph. Narayanan, P. J.; Rander, P. W.; and Kanade, T. 1998. Constructing Virtual Worlds Using Dense Stereo. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV). Niemeyer, M.; Barron, J. T.; Mildenhall, B.; Sajjadi, M. S. M.; Geiger, A.; and Radwan, N. 2022. Reg Ne RF: Regularizing Neural Radiance Fields for View Synthesis from Sparse Inputs. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR). Oquab, M.; Darcet, T.; Moutakanni, T.; Vo, H. V.; Szafraniec, M.; Khalidov, V.; Fernandez, P.; Haziza, D.; Massa, F.; El-Nouby, A.; Howes, R.; Huang, P.-Y.; Xu, H.; Sharma, V.; Li, S.-W.; Galuba, W.; Rabbat, M.; Assran, M.; Ballas, N.; Synnaeve, G.; Misra, I.; Jegou, H.; Mairal, J.; Labatut, P.; Joulin, A.; and Bojanowski, P. 2023. DINOv2: Learning Robust Visual Features without Supervision. ar Xiv:2304.07193. Park, K.; Sinha, U.; Hedman, P.; Barron, J. T.; Bouaziz, S.; Goldman, D. B.; Martin-Brualla, R.; and Seitz, S. M. 2021. Hyper Ne RF: A Higher-Dimensional Representation for Topologically Varying Neural Radiance Fields. ACM Trans. Graph. Radford, A.; Kim, J. W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; Krueger, G.; and Sutskever, I. 2021. Learning Transferable Visual Models From Natural Language Supervision. In International Conference on Machine Learning. Ren, Z.; Agarwala, A.; Russell, B.; Schwing, A. G.; and Wang, O. 2022. Neural Volumetric Object Selection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Snavely, N.; Seitz, S. M.; and Szeliski, R. 2006. Photo Tourism: Exploring Photo Collections in 3D. ACM Trans. Graph. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Sun, C.; Sun, M.; and Chen, H. 2022. Direct Voxel Grid Optimization: Super-fast Convergence for Radiance Fields Reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Tancik, M.; Casser, V.; Yan, X.; Pradhan, S.; Mildenhall, B.; Srinivasan, P. P.; Barron, J. T.; and Kretzschmar, H. 2022. Block-nerf: Scalable large scene neural view synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Tewari, A.; Thies, J.; Mildenhall, B.; Srinivasan, P.; Tretschk, E.; Wang, Y.; Lassner, C.; Sitzmann, V.; Martin Brualla, R.; Lombardi, S.; Simon, T.; Theobalt, C.; Nießner, M.; Barron, J. T.; Wetzstein, G.; Zollh ofer, M.; and Golyanik, V. 2022. Advances in Neural Rendering. Comput. Graph. Forum. Tschernezki, V.; Laina, I.; Larlus, D.; and Vedaldi, A. 2022. Neural Feature Fusion Fields: 3D Distillation of Self-Supervised 2D Image Representations. In International Conference on 3D Vision (3DV). Turki, H.; Ramanan, D.; and Satyanarayanan, M. 2022. Mega-NERF: Scalable Construction of Large-Scale Ne RFs for Virtual Fly-Throughs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Varma, M.; Wang, P.; Chen, X.; Chen, T.; Venugopalan, S.; and Wang, Z. 2023. Is Attention All That Ne RF Needs? In The Eleventh International Conference on Learning Representations. Verbin, D.; Hedman, P.; Mildenhall, B.; Zickler, T.; Barron, J. T.; and Srinivasan, P. P. 2022. Ref-Ne RF: Structured View-Dependent Appearance for Neural Radiance Fields. CVPR. Wang, C.; Chai, M.; He, M.; Chen, D.; and Liao, J. 2022. Clip-nerf: Text-and-image driven manipulation of neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Wang, P.; Liu, L.; Liu, Y.; Theobalt, C.; Komura, T.; and Wang, W. 2021a. Neu S: Learning Neural Implicit Surfaces by Volume Rendering for Multi-view Reconstruction. Neur IPS. Wang, Q.; Wang, Z.; Genova, K.; Srinivasan, P. P.; Zhou, H.; Barron, J. T.; Martin-Brualla, R.; Snavely, N.; and Funkhouser, T. 2021b. IBRNet: Learning Multi-View Image-Based Rendering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Xie, Y.; Takikawa, T.; Saito, S.; Litany, O.; Yan, S.; Khan, N.; Tombari, F.; Tompkin, J.; Sitzmann, V.; and Sridhar, S. 2022. Neural Fields in Visual Computing and Beyond. Comput. Graph. Forum. Xu, D.; Jiang, Y.; Wang, P.; Fan, Z.; Shi, H.; and Wang, Z. 2022. Sinnerf: Training neural radiance fields on complex scenes from a single image. In European Conference on Computer Vision. Yao, Y.; Luo, Z.; Li, S.; Fang, T.; and Quan, L. 2018. Mvsnet: Depth inference for unstructured multi-view stereo. In Proceedings of the European conference on computer vision (ECCV). Yu, A.; Ye, V.; Tancik, M.; and Kanazawa, A. 2021. pixel Ne RF: Neural Radiance Fields from One or Few Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Yuan, Y.-J.; Sun, Y.-T.; Lai, Y.-K.; Ma, Y.; Jia, R.; and Gao, L. 2022. Nerf-editing: geometry editing of neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)