# pynerf_pyramidal_neural_radiance_fields__43047005.pdf Py Ne RF: Pyramidal Neural Radiance Fields Haithem Turki Carnegie Mellon University hturki@cs.cmu.edu Michael Zollhöfer Meta Reality Labs Research zollhoefer@meta.com Christian Richardt Meta Reality Labs Research crichardt@meta.com Deva Ramanan Carnegie Mellon University deva@cs.cmu.edu Neural Radiance Fields (Ne RFs) can be dramatically accelerated by spatial grid representations [6, 9, 20, 25]. However, they do not explicitly reason about scale and so introduce aliasing artifacts when reconstructing scenes captured at different camera distances. Mip-Ne RF and its extensions propose scale-aware renderers that project volumetric frustums rather than point samples but such approaches rely on positional encodings that are not readily compatible with grid methods. We propose a simple modification to grid-based models by training model heads at different spatial grid resolutions. At render time, we simply use coarser grids to render samples that cover larger volumes. Our method can be easily applied to existing accelerated Ne RF methods and significantly improves rendering quality (reducing error rates by 20 90% across synthetic and unbounded real-world scenes) while incurring minimal performance overhead (as each model head is quick to evaluate). Compared to Mip-Ne RF, we reduce error rates by 20% while training over 60 faster. 1 Introduction Recent advances in neural volumetric rendering techniques, most notably around Neural Radiance Fields [19] (Ne RFs), have lead to significant progress towards photo-realistic novel view synthesis. However, although Ne RF provides state-of-the-art rendering quality, it is notoriously slow to train and render due in part to its internal MLP representation. It further assumes that scene content is equidistant from the camera and rendering quality degrades due to aliasing and excessive blurring when that assumption is violated. Recent methods [6, 9, 20, 25] accelerate Ne RF training and rendering significantly through the use of grid representations. Others, such as Mip-Ne RF [2], address aliasing by projecting camera frustum volumes instead of point-sampling rays. However, these anti-aliasing methods rely on the base Ne RF MLP representation (and are thus slow) and are incompatible with grid representations due to their reliance on non-grid-based inputs. Inspired by divide-and-conquer Ne RF extensions [22, 23, 27, 30] and classical approaches such as Gaussian pyramids [1] and mipmaps [34], we propose a simple approach that can easily be applied to any existing accelerated Ne RF implementation. We train a pyramid of models at different scales, sample along camera rays (as in the original Ne RF), and simply query coarser levels of the pyramid for samples that cover larger volumes (similar to voxel cone tracing [8]). Our method is simple to implement and significantly improves the rendering quality of fast rendering approaches with minimal performance overhead. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). Figure 1: Comparison of methods. (a) Ne RF traces a ray from the camera s center of projection through each pixel and samples points x along each ray. Sample locations are then encoded with a positional encoding to produce a feature γ(x) that is fed into an MLP. (b) Mip-Ne RF instead reasons about volumes by defining a 3D conical frustum per camera pixel. It splits the frustum into sampled volumes, approximates them as multivariate Gaussians, and computes the integral of the positional encodings of the coordinates contained within the Gaussians. Similar to Ne RF, these features are then fed into an MLP. (c) Accelerated grid methods, such as i NGP, sample points as in Ne RF, but do not use positional encoding and instead featurize each point by interpolating between vertices in a feature grid. These features are then passed into a much smaller MLP, which greatly accelerates training and rendering. (d) Py Ne RF also uses feature grids, but reasons about volumes by training separate models at different scales (similar to a mipmap). It calculates the area covered by each sample in world coordinates, queries the models at the closest corresponding resolutions, and interpolates their outputs. Contribution: Our primary contribution is a partitioning method that can be easily adapted to any existing grid-rendering approach. We present state-of-the-art reconstruction results against a wide range of datasets, including on novel scenes we designed that explicitly target common aliasing patterns. We evaluate different posssible architectures and demonstrate that our design choices provide a high level of visual fidelity while maintaining the rendering speed of fast Ne RF approaches. 2 Related Work The now-seminal Neural Radiance Fields (Ne RF) paper [19] inspired a vast corpus of follow-up work. We discuss a non-exhaustive list of such approaches along axes relevant to our work. Grid-based methods. The original Ne RF took 1 2 days to train, with extensions for unbounded scenes [3, 40] taking longer. Once trained, rendering takes seconds per frame and is far below interactive thresholds. NSVF [17] combines Ne RF s implicit representation with a voxel octree that allows for empty-space skipping and improves inference speeds by 10 . Follow-up works [10, 11, 39] further improve rendering to interactive speeds by storing precomputed model outputs into auxiliary grid structures that bypass the need to query the original model altogether at render time. Plenoxels [25] and DVGO [26] accelerate both training and rendering by directly optimizing a voxel grid instead of an MLP to train in minutes or even seconds. Tenso RF [6] and K-Planes [9] instead use the product of low-rank tensors to approximate the voxel grid and reduce memory usage, while Instant-NGP [20] (i NGP) encodes features into a multi-resolution hash table. The main goal of our work is to combine the speed benefits of grid-based methods with an approach that maintains quality across different rendering scales. Figure 2: We visualize renderings from a pyramid of spatial grid-based Ne RFs trained for different voxel resolutions. Models at finer pyramid levels tend to capture finer content. Divide-and-conquer. Several works note the diminishing returns in using large networks to represent scene content, and instead render the area of interest with multiple smaller models. De RF [22] and Kilo Ne RF [23] focus on inference speed while Mega-Ne RF [30], Block-Ne RF [27], and SUDS [31] use scene decomposition to efficiently train city-scale neural representations. Our method is similar in philosophy, although we partition across different resolutions instead of geographical area. Aliasing. The original Ne RF assumes that scene content is captured at roughly equidistant camera distances and emits blurry renderings when the assumption is violated. Mip-Ne RF [2] reasons about the volume covered by each camera ray and proposes an integrated positional encoding that alleviates aliasing. Mip-Ne RF 360 [3] extends the base method to unbounded scenes. Exact-Ne RF [14] derives a more precise integration formula that better reconstructs far-away scene content. Bungee-Ne RF [36] leverages Mip-Ne RF and further adopts a coarse-to-fine training approach with residual blocks to train on large-scale scenes with viewpoint variation. LIRF [37] proposes a multiscale image-based representation that can generalize across scenes. The methods all build upon the original Ne RF MLP model and do not readily translate to accelerated grid-based methods. Concurrent work. Several contemporary efforts explore the intersection of anti-aliasing and fast rendering. Zip-Ne RF [4] combines a hash table representation with a multi-sampling method that approximates the true integral of features contained within each camera ray s view frustum. Although it trains faster than Mip-Ne RF, it is explicitly not designed for fast rendering as the multi-sampling adds significant overhead. Mip-Vo G [12] downsamples and blurs a voxel grid according to the volume of each sample in world coordinates. We compare their reported numbers to ours in Section 4.2. Tri-Mip RF [13] uses a similar prefiltering approach, but with triplanes instead of a 3D voxel grid. Classical methods. Similar to Py Ne RF, classic image processing methods, such as Gaussian [1] and Laplacian [5] hierarchy, maintain a coarse-to-fine pyramid of different images at different resolutions. Compared to Mip-Ne RF, which attempts to learn a single MLP model across all scales, one could argue that our work demonstrates that the classic pyramid approach can be efficiently adapted to neural volumetric models. In addition, our ray sampling method is similar to Crassin et al. s approach [8], which approximates cone tracing by sampling along camera rays and querying different mipmap levels according the spatial footprint of each sample (stored as a voxel octree in their approach and as a Ne RF model in ours). (a) Point Sampling (c8, σ8) = f8(x, d) (c9, σ9) = f9(x, d) (b) Model Evaluation c = 0.4c8 + 0.6c9 σ = 0.4σ8 + 0.6σ9 (c) Interpolation Figure 3: Overview. (a) We sample frustums along the camera ray corresponding to each pixel and derive the scale of each sample according to its width in world coordinates. (b) We query the model heads closest to the scale of each sample. (c) We derive a single color and density value for each sample by interpolating between model outputs according to scale. 3.1 Preliminaries Ne RF. Ne RF [19] represents a scene within a continuous volumetric radiance field that captures geometry and view-dependent appearance. It encodes the scene within the weights of a multilayer perceptron (MLP). At render time, Ne RF casts a camera ray r for each image pixel. Ne RF samples multiple positions xi along each ray and queries the MLP at each position (along with the ray viewing direction d) to obtain density and color values σi and ci. To better capture high-frequency details, Ne RF maps xi and d through an L-dimensional positional encoding (PE) γ(x) = [sin(20 x), cos(20 x), . . . , sin(2L x), cos(2L x)] instead of directly using them as MLP inputs. It then composites a single color prediction ˆC(r) per ray using numerical quadrature PN 1 i=0 Ti(1 exp( σiδi)) ci, where Ti = exp( Pi 1 j=0 σjδj) and δi is the distance between samples. The training process optimizes the model by sampling batches R of image pixels and minimizing the loss P """C(r) ˆC(r) . We refer the reader to Mildenhall et al. [19] for details. Anti-aliasing. The original Ne RF suffers from aliasing artifacts when reconstructing scene content observed at different distances or resolutions due to its reliance on point-sampled features. As these features ignore the volume viewed by each ray, different cameras viewing the same position from different distances may produce the same ambiguous feature. Mip-Ne RF [2] and variants instead reason about volumes by defining a 3D conical frustum per camera pixel. It featurizes intervals within the frustum with a integrated positional encoding (IPE) that approximates each frustum as a multivariate Gaussian to estimate the integral E[γ(x)] over the PEs of the coordinates within it. Grid-based acceleration. Various methods [6, 9, 20, 25, 26] eschew Ne RF s positional encoding and instead store learned features into a grid-based structure, e.g. implemented as an explicit voxel grid, hash table, or a collection of low-rank tensors. The features are interpolated based on the position of each sample and then passed into a hard-coded function or much smaller MLP to produce density and color, thereby accelerating training and rendering by orders of magnitude. However, these approaches all use the same volume-insensitive point sampling of the original Ne RF and do not have a straightforward analogy to Mip-Ne RF s IPE as they no longer use positional encoding. 3.2 Multiscale sampling Assume that each sample x (where we drop the i index to reduce notational clutter) is associated with an integration volume. Intuitively, samples close to a camera correspond to small volumes, while samples far away from a camera correspond to large volumes (Figure 3). Our crucial insight for enabling multiscale sampling with grid-based approaches is remarkably simple: we train separate Ne RFs at different voxel resolutions and simply use coarser Ne RFs for samples covering larger volumes. Specifically, we define a hierarchy of L resolutions that divide the world into voxels of length 1/N0, ..., 1/NL 1, where Nl+1 = s Nl and s is a constant scaling factor. We also define a function fl(x, d) at each level that maps from sample location x and viewing direction d to color c and density σ. fl can be implemented by any grid-based Ne RF; in our experiments, we use a Algorithm 1 Py Ne RF rendering function Input: m rays r, L pyramid levels, hierarchy mapping function M, base resolution N0, scaling factor s Output: m estimated colors c x, d, P(x) sample(r) . Sample points x along each ray with direction d and area P(x) M(P(x)) logs(P(x)/N0) . Equation 1 l min(L 1, max(0, d M(P(x))e)) . Equation 2 w l M(P(x)) . Equation 5 model_out zeros(len(x)) . Zero-initialize model outputs for each sample x for i in unique(l) do . Iterate over sample levels model_out[l = i] += w[l = i]fi(x[l = i], d[l = i]) model_out[l = i] += (1 w)[l = i]fi 1(x[l = i], d[l = i]) end for c composite(model_out) . Composite model outputs into per-ray color c return c hash table followed by small density and color MLPs, similar to i NGP. We further define a mapping function M that assigns the integration volume of sample x to the hierarchy level l. We explore different alternatives, but find that selecting the level whose voxels project to the 2D pixel area P(x) used to define the integration volume works well: M(P(x)) = logs(P(x)/N0) (1) l = min(L 1, max(0, d M(P(x))e)) (2) σ, c = fl(x, d), [Gauss Py Ne RF] (3) where d e is the ceiling function. Such a model can be seen as a (Gaussian) pyramid of spatial grid-based Ne RFs (Fig. 2). If the final density and color were obtained by summing across different pyramid levels, the resulting levels would learn to specialize to residual or band-pass frequencies (as in a 3D Laplacian pyramid [5]): fi(x, d). [Laplacian Py Ne RF] (4) Our experiments show that such a representation is performant, but expensive since it requires l model evaluations per sample. Instead, we find a good tradeoff is to linearly interpolate between two model evaluations at the levels just larger than and smaller than the target integration volume: σ, c = wfl(x, d) + (1 w)fl 1(x, d), where w = l M(P(x)). (Default) [Py Ne RF] (5) This adds the cost of only a single additional evaluation (increasing the overall rendering time from 0.0045 to 0.005 ms per pixel) while maintaining rendering quality (see Section 4.6). Our algorithm is summarized in Algorithm 1. Matching areas vs volumes. One might suspect it may be better to select the voxel level l whose volume best matches the sample s 3D integration volume. We experimented with this, but found it more effective to match the projected 2D pixel area rather than volumes. Note that both approaches would produce identical results if the 3D volume was always a cube, but volumes may be elongated along the ray depending on the sampling pattern. Matching areas is preferable because most visible 3D scenes consist of empty space and surfaces, implying that when computing the composite color for a ray r, most of the contribution will come from a few samples x lying near the surface of intersection. When considering the target 3D integration volume associated with x, most of the contribution to the final composite color will come from integrating along the 2D surface (since the rest of the 3D volume is either empty or hidden). This loosely suggests we should select levels of the voxel hierarchy based on (projected) area rather than volume. Hierarchical grid structures. Our method can be applied to any accelerated grid method irrespective of the underyling storage. However, a drawback of this approach is an increased on-disk serialization footprint due to training a hierarchy of spatial grid Ne RFs. A possible solution is to exploit hierarchical grid structures that already exist within the base Ne RF. Note that multi-resolution grids such as those Table 1: Synthetic results. Py Ne RF outperforms all baselines and trains over 60 faster than Mip Ne RF. Both Py Ne RF and Mip-Ne RF properly reconstruct the brick wall in the Blender-A dataset, but Mip-Ne RF fails to accurately reconstruct checkerboard patterns. Multiscale Blender [2] Blender-A "PSNR "SSIM #LPIPS #Avg Error #Train Time (h) "PSNR "SSIM #LPIPS #Avg Error # Train Time (h) Plenoxels [25] 24.98 0.843 0.161 0.080 0:28 18.13 0.511 0.523 0.190 0:40 K-Planes [9] 29.88 0.946 0.058 0.022 0:32 21.17 0.593 0.641 0.405 1:22 Tenso RF [6] 30.04 0.948 0.056 0.021 0:27 27.01 0.785 0.197 0.054 1:20 i NGP [20] 30.21 0.958 0.040 0.022 0:20 20.85 0.767 0.244 0.089 0:56 Nerfacto [28] 29.56 0.947 0.051 0.022 0:25 27.46 0.796 0.195 0.053 1:07 Mip-Vo G [12] 30.42 0.954 0.053 Mip-Ne RF [2] 34.50 0.974 0.017 0.009 29:49 31.33 0.894 0.098 0.063 30:12 Py Ne RF 34.78 0.976 0.015 0.008 0:25 41.99 0.986 0.007 0.004 1:10 used by i NGP [20] or K-Planes [9] already define a scale hierarchy that is a natural fit for Py Ne RF. Rather than learning a separate feature grid for each model in our pyramid, we can reuse the same multi-resolution features across levels (while still training different MLP heads). Multi-resolution pixel input. One added benefit of the above is that one can train with multiscale training data, which is particularly helpful for learning large, city-scale Ne RFs [27, 30, 31, 36, 38]. For such scenarios, even storing high-resolution pixel imagery may be cumbersome. In our formulation, one can store low-resolution images and quickly train a coarse scene representation. The benefits are multiple. Firstly, divide-and-conquer approaches such as Mega-Ne RF [31] partition large scenes into smaller cells and train using different training pixel/ray subsets for each (to avoid training on irrelevant data). However, in the absence of depth sensors or a priori 3D scene knowledge, Mega-Ne RF is limited in its ability to prune irrelevant pixels/rays (due to intervening occluders) which empirically bloat the size of each training partition by 2 [30]. With our approach, we can learn a coarse 3D knowledge of the scene on downsampled images and then filter higher-resolution data partitions more efficiently. Once trained, lower-resolution levels can also serve as an efficient initialization for finer layers. In addition, many contemporary Ne RF methods use occupancy grids [20] or proposal networks [3] to generate refined samples near surfaces. We can quickly train these along with our initial low-resolution model and then use them to train higher-resolution levels in a sample-efficient manner. We show in our experiments that such course-to-fine multiscale training can speed up convergence (Section 4.5). Unsupervised levels. A naive implementation of our method will degrade when zooming in and out of areas that have not been seen at training time. Our implementation mitigates this by maintaining an auxiliary data structure (similar to an occupancy grid [20]) that tracks the coarsest and finest levels queried in each region during training. We then use the structure at inference time to only query levels that were supervised during training. 4 Experiments We first evaluate Py Ne RF s performance by measuring its reconstruction quality on bounded synthetic (Section 4.2) and unbounded real-world (Section 4.3) scenes. We demonstrate Py Ne RF s generalizability by evaluating it on additional Ne RF backbones (Section 4.4) and then explore the convergence benefits of using multiscale training data in city-scale reconstruction scenarios (Section 4.5). We ablate our design decisions in Section 4.6. 4.1 Experimental Setup Training. We implement Py Ne RF on top of the Nerfstudio library [28] and train on each scene with 8,192 rays per batch by default for 20,000 iterations on the Multiscale Blender and Mip-Ne RF 360 datasets, and 50,000 iterations on the Boat dataset and Blender-A. We train a hierarchy of 8 Py Ne RF levels backed by a single multi-resolution hash table similar to that used by i NGP [20] in Section 4.2 and Section 4.3 before evaluating additional backbones in Section 4.4. We use 4 features per level with a hash table size of 220 by default, which we found to give the best quality-performance trade-off on the A100 GPUs we use in our experiments. Each Py Ne RF uses a 64-channel density MLP with one hidden layer followed by a 128-channel color MLP with two hidden layers. We use similar model capacities in our baselines for fairness. We sample rays using an occupancy Figure 4: Synthetic results. Py Ne RF and Mip-Ne RF provide comparable results on the first three scenes that are crisper than those of the other fast renderers. Mip-Ne RF does not accurately render the tiles in the last row while Py Ne RF recreates them near-perfectly. grid [20] on the Multiscale Blender dataset, and with a proposal network [3] on all others. We use gradient scaling [21] to improve training stability in scenes with that capture content at close distance (Blender-A and Boat). We parameterize unbounded scenes with Mip-Ne RF 360 s contraction method. Metrics. We report quantitative results based on PSNR, SSIM [33], and the Alex Net implementation of LPIPS [41], along with the training time in hours as measured on a single A100 GPU. For ease of comparison, we also report the average error metric proposed by Mip-Ne RF [2] composed of the geometric mean of MSE = 10 PSNR/10, 1 SSIM, and LPIPS. 4.2 Synthetic Reconstruction Datasets. We evaluate Py Ne RF on the Multiscale Blender dataset proposed by Mip-Ne RF along with our own Blender scenes (which we name Blender-A ) intended to further probe the anti-aliasing ability of our approach (by reconstructing a slanted checkerboard and zooming into a brick wall). Baselines. We compare Py Ne RF to several fast-rendering approaches, namely Instant-NGP [20] and Nerfacto [28], which store features within a multi-resolution hash table, Plenoxels [25] which optimizes an explicit voxel grid, and Tenso RF [6] and K-Planes [9], which rely on low-rank tensor decomposition. We also compare our Multiscale Blender results to those reported by Mip-Vo G [12], a contemporary fast anti-aliasing approach, and to Mip-Ne RF [2] on both datasets. Results. We summarize our results in Table 1 and show qualitative examples in Figure 4. Py Ne RF outperforms all fast rendering approaches as well as Mip-Vo G by a wide margin and is slightly better than Mip-Ne RF on Multiscale Blender while training over 60 faster. Both Py Ne RF and Mip-Ne RF properly reconstruct the brick wall in the Blender-A dataset, but Mip-Ne RF fails to accurately reconstruct checkerboard patterns. 4.3 Real-World Reconstruction Datasets. We evaluate Py Ne RF on the Boat scene of the ADOP [24] dataset, which to our knowledge is one of the only publicly available unbounded real-world captures that captures its primary object of Figure 5: Real-world results. Py Ne RF reconstructs higher-fidelity details (such as the spokes on the bicycle and the lettering within the boat) than other methods. Table 2: Real-world results. Py Ne RF outperforms all baselines in PSNR and average error, and trains 40 faster than Mip-Ne RF 360 and 100 faster than Exact-Ne RF (the next best methods). Boat [24] Mip-Ne RF 360 [3] "PSNR "SSIM #LPIPS #Avg Error #Train Time (h) "PSNR "SSIM #LPIPS #Avg Error # Train Time (h) Plenoxels [25] 17.05 0.505 0.617 0.185 2:14 21.88 0.606 0.524 0.117 1:00 K-Planes [9] 18.00 0.501 0.590 0.168 2:41 21.53 0.577 0.500 0.120 1:08 Tenso RF [6] 14.75 0.398 0.630 0.234 2:30 18.07 0.439 0.677 0.181 1:07 i NGP [20] 15.34 0.433 0.646 0.222 1:42 21.14 0.568 0.521 0.126 0:40 Nerfacto [28] 19.27 0.570 0.425 0.135 2:12 22.47 0.616 0.431 0.105 1:02 Mip-Ne RF 360 w/ GLO [3] 20.03 0.595 0.416 0.124 37:28 22.76 0.664 0.342 0.095 37:35 Mip-Ne RF 360 w/o GLO [3] 15.92 0.480 0.501 0.194 37:10 22.70 0.664 0.342 0.095 37:22 Exact-Ne RF w/ GLO [14] 20.21 0.601 0.425 0.123 109:11 21.40 0.619 0.416 0.121 110:06 Exact-Ne RF w/o GLO [14] 16.33 0.489 0.510 0.187 107:52 22.56 0.619 0.410 0.121 108:11 Py Ne RF 20.43 0.601 0.422 0.121 2:12 23.09 0.654 0.358 0.094 1:00 interest from different camera distances. For further comparison, we construct a multiscale version of the outdoor scenes in the Mip-Ne RF 360 [3] dataset using the same protocol as Multiscale Blender [2]. Baselines. We compare Py Ne RF to the same fast-rendering approaches as in Section 4.2, along with two unbounded Mip-Ne RF variants: Mip-Ne RF 360 [3] and Exact-Ne RF [14]. We report numbers on each variant with and without generative latent optimization [18] to account for lighting changes. Results. We summarize our results in Table 2 along with qualitative results in Figure 5. Once again, Py Ne RF outperforms all baselines, trains 40 faster than Mip-Ne RF 360, and 100 faster than Exact-Ne RF (the next best alternatives). 4.4 Additional Backbones Methods. We demonstrate how Py Ne RF can be applied to any grid-based Ne RF method by evaluating it with K-Planes [9] and Tenso RF [6] in addition to our default i NGP-based implementatino. We take advantage of the inherent multi-resolution structure of i NGP and K-Planes by reusing the same feature grid across Py Ne RF levels and train a separate feature grid per level in our Tenso RF variant. Results. We train the Py Ne RF variants along with their backbones across the datasets described in Section 4.2 and Section 4.3, and summarize the results in Table 3. All Py Ne RF variants show clear improvements over their base methods. 4.5 City-Scale Convergence Dataset. We evaluate Py Ne RF s convergence properties on the the Argoverse 2 [35] Sensor dataset (to our knowledge, the largest city-scale dataset publicly available). We select the largest overlapping subset of logs and filter out moving objects through a pretrained segmentation model [7]. The resulting training set contains 400 billion rays across 150K video frames. Methods. We use SUDS [31] as the backbone model in our experiments. We begin training our method on 8 downsampled images (containing 64 fewer rays) for 5,000 iterations and then on Table 3: Additional backbones. We train the Py Ne RF variants along with their backbones across the datasets described in Section 4.2 and Section 4.3 All Py Ne RF variants outperform their baselines by a wide margin. Synthetic Real-World "PSNR "SSIM #LPIPS #Avg Error "PSNR "SSIM #LPIPS #Avg Error i NGP [20] 28.86 0.916 0.087 0.032 19.94 0.541 0.537 0.146 K-Planes [9] 27.90 0.865 0.131 0.047 20.54 0.553 0.520 0.136 Tenso RF [6] 29.12 0.902 0.100 0.042 17.21 0.421 0.696 0.200 Py Ne RF 36.22 0.979 0.013 0.004 22.65 0.645 0.369 0.098 Py Ne RF-K-Planes 35.42 0.975 0.014 0.005 22.00 0.622 0.405 0.108 Py Ne RF-Tenso RF 35.67 0.976 0.015 0.005 21.35 0.568 0.482 0.122 Table 4: City-scale convergence. We track rendering quality over the first four hours of training. Py Ne RF achieves the same rendering quality as SUDS 2 faster. Time (h) 1:00 2:00 3:00 4:00 SUDS [31] 16.01 17.41 18.08 18.53 Py Ne RF 17.17 18.44 18.59 18.73 Time (h) 1:00 2:00 3:00 4:00 SUDS [31] 0.570 0.600 0.602 0.606 Py Ne RF 0.614 0.618 0.619 0.621 Time (h) 1:00 2:00 3:00 4:00 SUDS [31] 0.531 0.496 0.470 0.466 Py Ne RF 0.521 0.485 0.469 0.465 # Avg Error Time (h) 1:00 2:00 3:00 4:00 SUDS [31] 0.182 0.160 0.150 0.145 Py Ne RF 0.165 0.146 0.144 0.142 progressively higher resolutions (downsampled to 4 , 2 , and 1 ) every 5,000 iterations hereafter. We compare to the original SUDS method as a baseline. Metrics. We report the evolution of the quality metrics used in Section 4.2 and Section 4.3 over the first four hours of the training process. Results. We summarize our results in Table 4. Py Ne RF converges more rapidly than the SUDS baseline, achieving the same rendering quality at 2 hours as SUDS after 4. 4.6 Diagnostics Methods. We validate our design decisions by testing several variants. We ablate our MLP-level interpolation described in Equation 5 and compare it to the Gausss Py Ne RF and Laplacian Py Ne RF variants described in Section 3.2 along with another that instead interpolates the learned grid feature vectors (which avoids the need for an additional MLP evaluation per sample). As increased storage footprint is a potential drawback method, we compare our default strategy of sharing the same multi-resolution feature grid across Py Ne RF levels to the naive implementation that trains a separate grid per level. We also explore using 3D sample volumes instead of projected 2D pixel areas to determine voxel levels l. Results. We train our method and variants as described in Section 4.2 and Section 4.3, and summarize the results (averaged across datasets) in Table 5. Our proposed interpolation method strikes a good balance its performance is near-identical to the full Laplacian Py Ne RF approach while training 3 faster (and is significantly better than the other interpolation methods). Our strategy of reusing the same feature grid across levels performs comparably to the naive implementation while training faster due to fewer feature grid lookups. Using 2D pixel areas instead of 3D volumes to determine voxel level l provides an improvement. Table 5: Diagnostics. The rendering quality of our interpolation method is near-identical to the full residual approach while training 3 faster, and is significantly better than other alternatives. Reusing the same feature grid across levels performs comparably to storing separate hash tables per level while training faster. Method Our Interp. Shared Features 2D Area "PSNR "SSIM #LPIPS # Avg # Train Time (h) Gauss Py Ne RF (Eq. 3) 7 X X 28.72 0.803 0.201 0.056 0:43 Laplacian Py Ne RF (Eq. 4) 7 X X 29.48 0.813 0.190 0.052 2:44 Feature grid interpolation 7 7 X 28.45 0.767 0.244 0.070 0:46 Separate hash tables X 7 X 29.41 0.813 0.196 0.054 0:52 Levels w/ 3D Volumes X X 7 29.19 0.811 0.184 0.054 0:48 Py Ne RF X X X 29.44 0.812 0.191 0.053 0:48 5 Limitations Although our method generalizes to any grid-based method (Section 4.4), it requires a larger on-disk serialization footprint due to training a hierarchy of spatial grid Ne RFs. This can be mitigated by reusing the same feature grid when the underlying backbone uses a multi-resolution feature grid [9, 20], but this is not true of all methods [6, 25]. 6 Societal Impact Our method facilitates the rapid construction of high-quality neural representations in a resource efficient manner. As such, the risks inherent to our work is similar to those of other neural rendering papers, namely privacy and security concerns related to the intentional or inadvertent capture or privacy-sensitive information such as human faces and vehicle license plate numbers. While we did not apply our approach to data with privacy or security concerns, there is a risk, similar to other neural rendering approaches, that such data could end up in the trained model if the employed datasets are not properly filtered before use. Many recent approaches [15, 16, 29, 31, 42] distill semantics into Ne RF s representation, which may be used to filter out sensitive information at render time. However this information would still reside in the model itself. This could in turn be mitigated by preprocessing the input data used to train the model [32]. 7 Conclusion We propose a method that significantly improves the anti-aliasing properties of fast volumetric renderers. Our approach can be easily applied to any existing grid-based Ne RF, and although simple, provides state-of-the-art reconstruction results against a wide variety of datasets (while training 60 100 faster than existing anti-aliasing methods). We propose several synthetic scenes that model common aliasing patterns as few existing Ne RF datasets cover these scenarios in practice. Creating and sharing additional real-world captures would likely facilitate further research. Acknowledgements. HT and DR were supported in part by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior/ Interior Business Center (DOI/IBC) contract number 140D0423C0074. The U.S. Government is authorized to reproduce and distribute reprints for Governmental purposes notwithstanding any copyright annotation thereon. Disclaimer: The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of IARPA, DOI/IBC, or the U.S. Government. [1] E. Adelson, C. Anderson, J. Bergen, P. Burt, and J. Ogden. Pyramid methods in image processing. RCA Eng., 29, 11 1983. [2] J. T. Barron, B. Mildenhall, M. Tancik, P. Hedman, R. Martin-Brualla, and P. P. Srinivasan. Mip-Ne RF: A multiscale representation for anti-aliasing neural radiance fields. In ICCV, 2021. [3] J. T. Barron, B. Mildenhall, D. Verbin, P. P. Srinivasan, and P. Hedman. Mip-Ne RF 360: Unbounded anti-aliased neural radiance fields. In CVPR, 2022. [4] J. T. Barron, B. Mildenhall, D. Verbin, P. P. Srinivasan, and P. Hedman. Zip-Ne RF: Anti-aliased grid-based neural radiance fields. In ICCV, 2023. [5] P. Burt and E. Adelson. The laplacian pyramid as a compact image code. IEEE Transactions on Communications, 31(4):532 540, 1983. doi: 10.1109/TCOM.1983.1095851. [6] A. Chen, Z. Xu, A. Geiger, J. Yu, and H. Su. Tenso RF: Tensorial radiance fields. In ECCV, 2022. [7] B. Cheng, M. D. Collins, Y. Zhu, T. Liu, T. S. Huang, H. Adam, and L.-C. Chen. Panoptic-deeplab: A simple, strong, and fast baseline for bottom-up panoptic segmentation. In CVPR, 2020. [8] C. Crassin, F. Neyret, M. Sainz, S. Green, and E. Eisemann. Interactive indirect illumination using voxel cone tracing: A preview. In Symposium on Interactive 3D Graphics and Games, 2011. [9] S. Fridovich-Keil, G. Meanti, F. R. Warburg, B. Recht, and A. Kanazawa. K-planes: Explicit radiance fields in space, time, and appearance. In CVPR, 2023. [10] S. J. Garbin, M. Kowalski, M. Johnson, J. Shotton, and J. Valentin. Fast Ne RF: High-fidelity neural rendering at 200fps. In ICCV, 2021. [11] P. Hedman, P. P. Srinivasan, B. Mildenhall, J. T. Barron, and P. Debevec. Baking neural radiance fields for real-time view synthesis. In ICCV, 2021. [12] D. Hu, Z. Zhang, T. Hou, T. Liu, H. Fu, and M. Gong. Multiscale representation for real-time anti-aliasing neural rendering. arxiv:2304.10075, 2023. [13] W. Hu, Y. Wang, L. Ma, B. Yang, L. Gao, X. Liu, and Y. Ma. Tri-miprf: Tri-mip representation for efficient anti-aliasing neural radiance fields. In ICCV, 2023. [14] B. K. S. Isaac-Medina, C. G. Willcocks, and T. P. Breckon. Exact-Ne RF: An exploration of a precise volumetric parameterization for neural radiance fields. CVPR, 2023. [15] J. Kerr, C. M. Kim, K. Goldberg, A. Kanazawa, and M. Tancik. LERF: Language embedded radiance fields. In ICCV, 2023. [16] S. Kobayashi, E. Matsumoto, and V. Sitzmann. Decomposing nerf for editing via feature field distillation. In Advances in Neural Information Processing Systems, volume 35, 2022. [17] L. Liu, J. Gu, K. Z. Lin, T.-S. Chua, and C. Theobalt. Neural sparse voxel fields. NIPS, 2020. [18] R. Martin-Brualla, N. Radwan, M. S. M. Sajjadi, J. T. Barron, A. Dosovitskiy, and D. Duckworth. Ne RF in the wild: Neural radiance fields for unconstrained photo collections. In CVPR, 2021. [19] B. Mildenhall, P. P. Srinivasan, M. Tancik, J. T. Barron, R. Ramamoorthi, and R. Ng. Ne RF: Representing scenes as neural radiance fields for view synthesis. In ECCV, 2020. [20] T. Müller, A. Evans, C. Schied, and A. Keller. Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph., 41(4):102:1 15, July 2022. doi: 10.1145/3528223.3530127. [21] J. Philip and V. Deschaintre. Floaters No More: Radiance Field Gradient Scaling for Improved Near- Camera Training. In T. Ritschel and A. Weidlich, editors, Eurographics Symposium on Rendering. The Eurographics Association, 2023. ISBN 978-3-03868-229-5. doi: 10.2312/sr.20231122. [22] D. Rebain, W. Jiang, S. Yazdani, K. Li, K. Yi, and A. Tagliasacchi. De RF: Decomposed radiance fields. In CVPR, 2021. [23] C. Reiser, S. Peng, Y. Liao, and A. Geiger. Kilo Ne RF: Speeding up neural radiance fields with thousands of tiny MLPs. In ICCV, 2021. [24] D. Rückert, L. Franke, and M. Stamminger. Adop: Approximate differentiable one-pixel point rendering. ACM Trans. Graph., 41(4):99:1 14, July 2022. ISSN 0730-0301. [25] Sara Fridovich-Keil and Alex Yu, M. Tancik, Q. Chen, B. Recht, and A. Kanazawa. Plenoxels: Radiance fields without neural networks. In CVPR, 2022. [26] C. Sun, M. Sun, and H. Chen. Direct voxel grid optimization: Super-fast convergence for radiance fields reconstruction. In CVPR, 2022. [27] M. Tancik, V. Casser, X. Yan, S. Pradhan, B. Mildenhall, P. P. Srinivasan, J. T. Barron, and H. Kretzschmar. Block-Ne RF: Scalable large scene neural view synthesis. In CVPR, pages 8248 8258, June 2022. [28] M. Tancik, E. Weber, E. Ng, R. Li, B. Yi, J. Kerr, T. Wang, A. Kristoffersen, J. Austin, K. Salahi, A. Ahuja, D. Mc Allister, and A. Kanazawa. Nerfstudio: A modular framework for neural radiance field development. In ACM SIGGRAPH 2023 Conference Proceedings, SIGGRAPH 23, 2023. [29] V. Tschernezki, I. Laina, D. Larlus, and A. Vedaldi. Neural Feature Fusion Fields: 3D distillation of self-supervised 2D image representation. In International Conference on 3D Vision (3DV), 2022. [30] H. Turki, D. Ramanan, and M. Satyanarayanan. Mega-NERF: Scalable construction of large-scale Ne RFs for virtual fly-throughs. In CVPR, pages 12922 12931, June 2022. [31] H. Turki, J. Y. Zhang, F. Ferroni, and D. Ramanan. SUDS: Scalable urban dynamic scenes. In CVPR, [32] J. Wang, B. Amos, A. Das, P. Pillai, N. Sadeh, and M. Satyanarayanan. A scalable and privacy-aware Io T service for live video analytics. In Multimedia Systems Conference, pages 38 49, 2017. [33] Z. Wang, A. Bovik, H. Sheikh, and E. Simoncelli. Image quality assessment: from error visibility to structural similarity. IEEE Transactions on Image Processing, 13(4):600 612, 2004. [34] L. Williams. Pyramidal parametrics. Computer Graphics, 17:1 11, July 1983. [35] B. Wilson, W. Qi, T. Agarwal, J. Lambert, J. Singh, S. Khandelwal, B. Pan, R. Kumar, A. Hartnett, J. K. Pontes, D. Ramanan, P. Carr, and J. Hays. Argoverse 2: Next generation datasets for self-driving perception and forecasting. In Neur IPS Datasets and Benchmarks, 2021. [36] Y. Xiangli, L. Xu, X. Pan, N. Zhao, A. Rao, C. Theobalt, B. Dai, and D. Lin. Bungee Ne RF: Progressive neural radiance field for extreme multi-scale scene rendering. In ECCV, 2022. [37] H. Xin, Z. Qi, F. Ying, L. Xiaoyu, W. Xuan, and W. Qing. Local implicit ray function for generalizable radiance field representation. In CVPR, 2023. [38] L. Xu, Y. Xiangli, S. Peng, X. Pan, N. Zhao, C. Theobalt, B. Dai, and D. Lin. Grid-guided neural radiance fields for large urban scenes. In CVPR, 2023. [39] A. Yu, R. Li, M. Tancik, H. Li, R. Ng, and A. Kanazawa. Plen Octrees for real-time rendering of neural radiance fields. In ICCV, 2021. [40] K. Zhang, G. Riegler, N. Snavely, and V. Koltun. Ne RF++: Analyzing and improving neural radiance fields. ar Xiv:2010.07492, 2020. [41] R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang. The unreasonable effectiveness of deep features as a perceptual metric. In CVPR, 2018. [42] S. Zhi, T. Laidlow, S. Leutenegger, and A. Davison. In-place scene labelling and understanding with implicit scene representation. In ICCV, 2021.