# hfneus_improved_surface_reconstruction_using_highfrequency_details__bfd60c26.pdf HF-Neu S: Improved Surface Reconstruction Using High-Frequency Details Yiqun Wang KAUST Ivan Skorokhodov KAUST Peter Wonka KAUST Neural rendering can be used to reconstruct implicit representations of shapes without 3D supervision. However, current neural surface reconstruction methods have difficulty learning high-frequency geometry details, so the reconstructed shapes are often over-smoothed. We develop HF-Neu S, a novel method to improve the quality of surface reconstruction in neural rendering. We follow recent work to model surfaces as signed distance functions (SDFs). First, we offer a derivation to analyze the relationship between the SDF, the volume density, the transparency function, and the weighting function used in the volume rendering equation and propose to model transparency as a transformed SDF. Second, we observe that attempting to jointly encode high-frequency and low-frequency components in a single SDF leads to unstable optimization. We propose to decompose the SDF into base and displacement functions with a coarse-to-fine strategy to increase the high-frequency details gradually. Finally, we design an adaptive optimization strategy that makes the training process focus on improving those regions near the surface where the SDFs have artifacts. Our qualitative and quantitative results show that our method can reconstruct fine-grained surface details and obtain better surface reconstruction quality than the current state of the art. Code available at https://github.com/yiqun-wang/HFS. 1 Introduction 3D reconstruction from a set of images is a fundamental challenge in computer vision [9]. In the recent past, the seminal framework Ne RF [19] inspired a lot of follow up work by modeling 3D objects as a density function σ(x) and view-dependent color c(x, v) for each point x R3 in the volume. The density function and view-dependent color function are implicit functions modeled by a neural network. The results of this approach are very strong and therefore Ne RF inspired a large amount of follup up work, e.g. [18, 17, 24, 35, 20, 2]. In particular, one direction of work tries to constrain the density field to make it more consistent with a density field stemming from a surface. In the original formulation, almost arbitrary densities can be modeled by the neural network and there is no guarantee that a meaningful surface can be extracted from the density. Two noteworthy recent approaches, Neus [30] and Vol SDF [32], proposed to embed a signed distance field in the volume rendering equation. Therefore, instead of modeling the density σ with a neural network, these approaches model a signed distance function f with a neural network. This leads to greatly improved surface reconstruction. We build on this exciting recent work and seek further improvement in the quality of surfaces that are being reconstructed. To this end, we propose our method HF-Neu S consisting of three new building blocks. First, we analyze the relationship between the signed distance function on the one hand and the volume density, the transparency, and the weighting function on the other hand. We conclude from our derivation that it would be best to model a function that maps signed distances to the transparency and propose a class of functions that fulfill the theoretical requirements. Second, we observe that 36th Conference on Neural Information Processing Systems (Neur IPS 2022). Figure 1: Qualitative evaluation on the Lego, Robot, and Ficus models. First column: reference images. Second to the fifth column: Ne RF, Vol SDF, Neu S, and OURS. Figure 2: The challenge of using high-frequencies directly in the Neu S framework. First column: reference image. Second to the fourth column: Neu S, Neu S with high-frequency details, and OURS. it is challenging to learn high-frequency details directly with a single signed distance function as shown in Fig. 2. We therefore propose to decompose the signed distance function into a base function and a displacement function following related work. We adapt this idea to the differentiable Ne RF rendering framework and the Ne RF training scheme. Third, the functions that translate distance to transparency can be chosen to have a parameter, which we call scale s. It controls the slope of the function (or the deviation of the derivative), which further controls the localization precision of the surface and how much out-of-surface colors influence the result. In previous work, this parameter s is set globally but is trainable, so it can change from iteration to iteration. We propose a novel spatially adaptive weighting scheme to influence this parameter, so that the optimization focuses more on problematic regions in the distance field. The three building blocks are the three main contributions of the paper. In the results, we can see that HF-Neu S has a clear improvement in surface reconstruction. On the 15 scene DTU benchmark we can improve from the current best values of 0.87 (Neu S) and 0.86 (Vol SDF) to 0.77 the Chamfer distance (See Figs. 1 and 4 for a visual comparison). The benchmark as well as the metric were proposed by previous work. 2 Related Work Multi-view 3D reconstruction. 3D reconstruction based on multiple views is a fundamental challenge in the field of 3D vision. Classical 3D reconstruction algorithms usually reconstruct discrete 3D representations. The methods can be roughly categorized into voxel-based methods and point-based methods. Voxel-based methods [6, 27, 14, 4, 11, 22] first discretize the threedimensional space uniformly into voxels, and then decide whether the surface occupies a particular voxel. Point-based methods [1, 7, 26, 25, 8] usually use the correlation between multiple views to reconstruct depth maps and fuse multiple depth maps into a point cloud. The point cloud needs to be subsequently reconstructed into a mesh model using explicit algorithms like ball-pivoting [3] and Delaunay trianglulation [15] or implicit algorithms like Poisson surface reconstruction [13]. Neural implicit surfaces. Recently, neural implicit representations have received a lot of attention. The corresponding methods aim to reconstruct continuous implicit function representations of shapes directly from 2D images. A required building block is differentiable rendering, which maps the 3D scene representation to a 2D image for a given camera pose. DVR [21] utilizes surface rendering to model the occupancy function of a 3D shape, which uses a root search approach to obtain the location of the surface and predicts a 2D image. IDR [33] models the signed distance function of the shape and uses a sphere tracking algorithm to render 2D images. A significant milestone in 3D reconstruction was the development of Ne RF [19]. It uses volume rendering to map a 3D density field and a 3D directional color field to a 2D image. The proposed representation is flexible enough so that realistic images can be synthesized. To model more complex scenes, Ne RF++ [35] proposes to model the background scene with an additional neural radiance field, which handles the foreground and background separately, and achieves better results for large scenes. However, the density function is not as easy to control as the occupancy function or the signed distance function, and it is difficult to guarantee the smoothness of the generated 3D shape. Subsequently, UNISURF [23] embeds the occupancy function into the volume rendering equation of Ne RF. They use a decay strategy to control which region to sample around the surface during training without explicitly modeling volumetric density. Using signed distance functions, Vol SDF [32] embeds a signed distance function into the density formulation and proposes a sampling strategy that satisfies a derived error bound on the transparency function. Neu S [30] derive an unbiased density function equation using logistic sigmoid functions and introduce a learnable parameter to control the function s slope during rendering and sampling. Concurrent to our work, Neural Patch [5] uses the homography matrix to warp the source patches adjacent to the reference image to constrain colors in the volume to come from closeby patches. However, the calculation of patch warping relies on the accurate surface normal, so it cannot be trained from scratch. Therefore, it is only used as a fine-tuning or post-processing method for other algorithms to optimize the surface. We consider Vol SDF and Neu S as the current state of the art and we will compare to these two methods. High-frequency detail reconstruction. It is generally difficult for neural networks to learn highfrequency information from raw signals. Inspired by the field of natural language processing, positional encoding [19, 29] is used to guide the network to reconstruct high-frequency details. Positional encoding spreads the original signal into different frequency bands using sine and cosine functions of different frequency. Subsequently, SIREN [28] proposes to use the sin function as activation function in the network. Mip Ne RF [2] presents an integrated positional encoding to control frequency in different scales. Park et al. [24] proposed to use a coarse-to-fine learning strategy to gradually increase high-frequency information, which was subsequently used for pose estimation [17]. Hertz et al. [10] further propose a spatially adaptive progressive coding strategy. For surface reconstruction, implicit displacement fields were proposed for single-view 3D reconstruction [16]. Based on the supervision of ground truth SDF values of sampled points, the method utilizes separate networks to model the base SDF and implicit displacement fields. Subsequently, Wang et al. [34] utilize the SIREN network to learn the base implicit function and implicit displacement function, respectively, for point cloud reconstruction tasks. In contrast to our proposed algorithm, these methods require 3D supervision. Further, they do not involve the Ne RF formulation or volume rendering. In our work, we build on these ideas to develop a new state-of-the-art algorithm for multi-view reconstruction. As input we consider a set of N images I = {I1, I2...IN}, and their corresponding intrinsic and extrinsic camera parameters Π = {π1, π2...πN}. HF-Neu S aims to reconstruct the representation of 3D surface S as implicit functions. Specifically, we encode surfaces as signed distance fields. We will explain our method in three parts: 1) First, we show how to embed the signed distance function into the formulation of volume rendering and discuss how to model the relationship between distance and transparency. 2) Then, we propose to utilize an additional displacement signed distance function to add high-frequency details to the base signed distance function. 3) Finally, we observe that the function that maps signed distances to transparency is controlled by a parameter s that determines the slope of the function. We propose a scheme to set this parameter s in a spatially varying manner depending on the gradient norm of the distance field, rather than keeping it constant for the complete volume within a single training iteration. 3.1 Modeling transparency as transformed SDF We first review the integral formula for volume rendering and derive a relationship between transparency and the weighting function (the product between density and transparency). Based on this analysis, we discuss the criteria for functions that are suitable to map signed distances to transparency and propose a class of functions that fulfill the theoretical requirements. Given a ray r(t) = o + td, the volume rendering equation is used to calculate the radiance C of the pixel corresponding to the ray r. The volume rendering equation is an integral along the ray and involves the following quantities defined for each point in the volume: the volume density σ and the (directional) color c. In addition, the volume has compact support and the boundaries of the volume are encoded by tn and tf. C(r) = Z tf tn T(t)σ(r(t))c(r(t), d)dt (1) The transparency T(t) is derived from the volume density as explained below. The function T(t) denotes the accumulated transmittance along the ray from tn to t T(t) = exp Z t tn σ(r(s))ds , (2) and T(t) is a monotonic decreasing function with a starting value of T(tn) = 1. The product T(t)σ(r(t)) can be regarded as a weighting function w (t) in the volume rendering equation as in Eq. (1). In order to involve a signed distance function f, we have to define a function Ψ to transform a signed distance function so that it can be used to compute the density related terms in the rendering equation. One way is to directly model a density function σ(r(t)) = Ψ (f (r(t))) as proposed by VOLSDF [32]. Taking this approach, a sampling method is required to satisfy an error bound of the sampling to make it less than an error threshold by gradually reducing the scale parameter. Another way is to model the weighting function w((t)) = Ψ (f (r(t))) as proposed by Neu S. The Neu S paper showcases a complex derivation to get the expression for the density function σ. We rethink this problem to obtain a simplified derivation by focusing on transparency instead of the weighting function and also a better understanding of the problem, as follows: dt = T(t)σ(r(t)) (3) An interesting observation is that the derivative of the transparency function T (t) is the negative weighting function. The weighting function has the property of having a maximum on the surface. We take the derivative of the weighting function and set it to 0 to find the extrema (maxima), as follows. d (T(t)σ(r(t))) dt = d2 (T(t)) dt2 = d (T (t)) Assuming a planar surface and a single ray-plane intersection, we can see that the extremum point, denoted as ts, of the weighting function w(t) will also be the extremum point of the derivative of the transparency function T (t). The point ts is expected to be the intersection of the ray and the surface. Therefore, we consider defining the transparency function directly as T(t) = Ψ (f (r(t))). If the transparency function is designed in such a way that its derivative T (t) reaches a minimum on the surface, it follows that the weighting function has a maximum on the surface. Therefore, one can directly model a transparency function under the condition that its derivative has a minimum on the surface. This is conceptually simpler than modeling the weighting function w(t) as proposed by Neu S. We compute the derivative of Ψ (f (r(t))) as follows. d (Ψ (f (r(t)))) dt = Ψ (f (r(t))) df dt = Ψ (f (r(t))) f (r(t)) d (5) where f (r(t)) d is the product of the surface normal and the ray direction, which is a constant in case of a planar surface and a single ray-plane intersection. The signed distance function is zero Figure 3: Comparing Neu S and Vol SDF with our transparency model. Ground truth is on the top left. For each method, the left shows the reconstructed image and the right the reconstructed surface. on the surface. Hence Ψ has an extremum at f = 0. This also means Ψ has the steepest slope at the surface of the shape. On the other hand, the signed distance function is positive outside of the object, and negative when entering the interior of the object. We generally assume that t = tn is outside so that the signed distance starts positive and decays to a negative value along a ray, which is a monotonic decreasing function. According to the characteristics of transparency T(t) = Ψ (f (r(t))), the transparency starts at 1 at t = tn and is a monotonic decreasing function to 0 inside the object. This inverse property results in the Ψ function being a monotonic increasing function from 0 to 1. Therefore, we have our design criteria for Ψ: Ψ should be a monotonic increasing function from 0 to 1, with the steepest slope at 0. A very intuitive idea to satisfy this criteria is to use a sigmoid function and normalize the function to have an output in the interval [0, 1]. We simply use the logistic sigmoid function proposed by Neu S [30] for a fair comparison. However, our idea is more general and other sigmoid functions could be used. Our designed transparency function is as follows, T(t) = Ψs (f (r(t))) = 1 1 + e sf(r(t)) , (6) where Ψs is the logistic sigmoid function with parameter s controlling the slope of the function. Note that the parameter s is also the standard deviation of the function Ψ s. We will use this fact later when discussing the adaptive version of the framework. Given the differentiable transparency function T(t), the volume density σ can be easily calculated following Eq. 3. σ(r(t)) = T (t) For discretization, we bring Eq. 5 and Eq. 6 into Eq.7, and take advantage of the properties of the derivative of the logistic sigmoid function Ψ s = sΨs(1 Ψs). We can get the σ formula for the discretization computation: σ(r(ti)) = s (Ψs (f (r(ti))) 1) f (r(ti)) d (8) Then the volume rendering integral can be approximated using α-composition, where αi = 1 exp ( σi (ti+1 ti)). For multiple surface intersections, we follow the same strategy as Neu S [30], where αi = clamp (αi, 0, 1). Compared with Neu S, we obtain a simpler formula for the density σ for the discretization computation, reducing the numerical problems caused by division in Neu S. Furthermore, our approach does not need to involve two different sampling points, namely section points and mid-points, which makes it easier to satisfy the unbiased weighting function. Since there is no need to calculate the SDF and the color separately for the two different point sets, the color and the geometry are more consistent compared to Neu S. Compared to Vol SDF [32], since the transparency function is explicit, our method can use an inverse distribution sampling computed with the inverse CDF to satisfy the approximation quality. Thus no complex sampling scheme as in Vol SDF is required. A visual comparison is shown in Fig. 3. 3.2 Implicit displacement field without 3D supervision In order to enable a multi-scale fitting framework, we propose to model the signed distance function as a combination of a base distance function and a displacement function [34, 16] along the normal of the base distance function. The implicit displacement function is an additional implicit function. The reason for this design is that it is difficult for a single implicit function to learn low-frequency and high-frequency information at the same time. The implicit displacement function can complement the base implicit function, so that it is easier to learn high-frequency information. Compared with the task of learning implicit functions from point clouds, reconstructing 3D shapes from multiple images makes it more difficult to learn high-frequency content. We propose to use neural networks to learn frequencies at multiple scales, and to gradually increase the frequency content in a coarse-to-fine manner. Suppose f is the combined implicit function that represents the surface we want to obtain. The function fb is the base implicit function that represents the base surface. Following [34], the displacement implicit function fd is used to map the point xb on the base surface to the surface point x along the normal nb and vice versa fd is used to map the point x on the base surface to the surface point xb along the normal nb, thus fd (xb) = fd(x). Because of the nature of implicit functions, the relationship between the two functions can be expressed as follows, fb(xb) = f(xb + fd (xb) nb) = 0 (9) where xb = fb(xb) fb(xb) , is the normal of xb on the base surface. To compute the expression for the implicit function f, we bring the formula xb = x fd (xb) nb into the Eq. (9) and obtain the expression for the combined implicit function: f(x) = fb(x fd (x) nb) (10) Therefore, we can use the base implicit function and the displacement implicit function to represent the combined implicit function. However, two challenges arise. First, the Eq. 10 is only satisfied if the point x is on the surface. Second, the normal at the point xb is difficult to estimate when only knowing the position x. We rely on two assumptions to solve the problem. One assumption is that this deformation can be applied to all iso-surfaces, i.e. fb(xb) = f(xb + fd (xb) nb) = c. In this way the equation is assumed to be valid for all points in the volume and not only on the surface. Another assumption is that xb and x are not too far away, thus nb can be replaced with normal n on the point x in the Eq. (10). We control the magnitude of the implicit displacement function using a displacement constraint 4Ψ s(fb). To precisely control the frequency, we use positional encoding to encode the base implicit function and the displacement implicit function separately. We would like to note some differences to [34]. We use positional encoding instead of Siren [28], so that the frequency can be explicitly controlled by a coarse-to-fine strategy rather than simply using two Siren networks with two different frequency levels. This is useful when 3D supervision is not given. More details are shown in the supplementary. Positional encoding decomposes the input position x into multiple selected frequency bands. γ(x) = [γ0(x), γ1(x), ..., γL 1(x)] (11) where each component consists of a sin and a cos function with different frequency. γj(x) = sin 2jπx , cos 2jπx Directly learning high-frequency positional encoding makes the network susceptible to noise, because wrongly learned high-frequencies hinder the learning of low frequencies. This problem is less pronounced if 3D supervision is available, however high-frequency information of images is easily introduced into the surface generation as noise. We use the coarse-to-fine strategy proposed by Park et al. [24] to gradually increase the frequency of the positional encoding. γj(x, α) = ωj (α) γj(x) = (1 cos (clamp (αL j, 0, 1) π)) 2 γj(x) (13) where α [0, 1] is the parameter to control the frequency information involved. In each iteration, α is increased by 1/nmax until it touches 1, where nmax is the maximum number of iterations. We utilize two kinds of positional encoding γ(x, αb), γ(x, αd) with different parameter αb and αd. We set αb = 0.5αd and only control αd for simplicity. We also use two MLP functions MLPb, MLPd for fitting the base and displacement functions. f(x) = MLPb(γ(x, αb) 4Ψ s(fb)MLPd (γ(x, αd)) n), (14) where n = fb(x) fb(x) that can be computed by the gradient of MLPb and Ψ s(fb) = Ψ s(MLPb(γ(x, αb))). The s of the displacement constraint should be clamped during training. We show how to control the adaptive s in the supplemental materials. We bring this implicit function into Eq. (6) for calculating the transparency so that the radiance (color) ˆCs of images can be computed by the volume rendering equation. To train the network, we employ the loss function L = Lrad + Lreg, which includes the radiance loss and the Eikonal regularization loss of the signed distance functions. For the regularization loss, we constrain both the base implicit function and the detailed implicit function. ˆCs Cs 1 + 1 h ( fb(xk) 2 1)2 + ( f(xk) 2 1)2i 3.3 Modeling an adaptivate transparency function In previous subsections, the transparency function is parametrized as a sigmoid function controlled by the scale s. This parameter controls the slope of the sigmoid function and it is also the standard deviation of the derivative. We can also say that it controls the smoothness of the function. When s is large, the value of the sigmoid function drops sharply as the position moves away from the surface. On the contrary, the value decreases smoothly when s is small. However, choosing a single parameter s per iteration gives the same behavior at all spatial locations in the volume. Since two signed distance functions need to be reconstructed, especially after the high frequency is superimposed, it is easy to break the Eikonal constraint, i.e., make the SDF s gradient norm deviate from 1 in some positions. Even with the regularization loss, it is impossible to avoid this problem. We propose to use the gradient norm of the signed distance field to weight the parameter s in a spatially varying manner, increasing s when the gradient norm along the ray direction is larger than 1. The intuition is that the implicit function with the larger gradient norm undergoes more abrupt changes, which indicates a region that should be improved. Making s larger in such regions makes the distance function more precise by magnifying its errors, especially near the surface. In order to adaptively modify the scale s, we propose the following equation: 1 + e s exp K P i=1 ωi fi 1 f(r(t)) where f is the gradient of the signed distance function, and K is the number of sampling points, ωi is the normalized Ψ s(fi) as the weight and K P i=1 ωi = 1. While this method can be used to control the transparency function, it can also be used for the hierarchical sampling stage proposed by standard Ne RF [19]. By locally increasing s, more samples will be generated near the surface where the signed distance values change more rapidly. This mechanism also helps to optimization to focus on these regions in the volume. 4 Experiments Baselines. We compare HF-Neu S to the following three state-of-the-art baselines: (1)Neu S [30] is the most relevant baseline for our work. We consider it to be the best published method. (2)Vol SDF [32] is concurrent work to Neu S. We consider it to be the second best published method. Overall it also performs very well. (3)Ne RF focuses on image synthesis and is included for completeness. Ne RF Table 1: Quantitative results on the DTU dataset. Metric Method 24 37 40 55 63 65 69 83 97 105 106 110 114 118 122 Mean Ne RF 1.90 1.60 1.85 0.58 2.28 1.27 1.47 1.67 2.05 1.07 0.88 2.53 1.06 1.15 0.96 1.49 VOLSDF 1.14 1.26 0.81 0.49 1.25 0.70 0.72 1.29 1.18 0.70 0.66 1.08 0.42 0.61 0.55 0.86 Neu S 1.37 1.21 0.73 0.40 1.20 0.70 0.72 1.01 1.16 0.82 0.66 1.69 0.39 0.49 0.51 0.87 OURS 0.76 1.32 0.70 0.39 1.06 0.63 0.63 1.15 1.12 0.80 0.52 1.22 0.33 0.49 0.50 0.77 Ne RF 26.24 25.74 26.79 27.57 31.96 31.50 29.58 32.78 28.35 32.08 33.49 31.54 31.0 35.59 35.51 30.65 VOLSDF 26.28 25.61 26.55 26.76 31.57 31.50 29.38 33.23 28.03 32.13 33.16 31.49 30.33 34.90 34.75 30.38 Neu S 28.20 27.10 28.13 28.80 32.05 33.75 30.96 34.47 29.57 32.98 35.07 32.74 31.69 36.97 37.07 31.97 OURS 29.15 27.33 28.37 28.88 32.89 33.84 31.17 34.83 30.06 33.37 35.44 33.09 32.12 37.13 37.32 32.33 Figure 4: Qualitative evaluation on DTU (first and third rows) and Blended MVS (second row). is not really a surface reconstruction method and does not reconstruct high-quality surfaces, but it is very good in image-based metrics. We use a threshold of 25 (as proposed by Neu S [30]) to extract surfaces for the comparisons. For all three methods, we use the default parameters and the number of iterations recommended in their respective papers. We do not include older methods in the comparison, such as UNISURF [23] or IDR [33], because Neu S and Vol SDF have better results. Datasets. We conduct experiments on the DTU dataset [12]. We follow previous work and choose the same 15 models for comparison. DTU is a multi-view stereo dataset. Each scene consists of 49 or 64 views with 1600 1200 resolution. We further choose 9 challenging scenes from other datasets: 6 scenes from the Ne RF-synthetic dataset [19] and 3 scenes from Blended MVS [31](CC-4 License). The image resolution of Ne RF synthetic dataset [19] is 800 800 and 100 views are provided for each scene. The dataset contains objects with very obvious detailed and sharp features, such as the Lego and Microphone scenes. We chose this dataset for the analysis of reconstructions of high-frequency details. The Blended MVS dataset is similar to the DTU dataset, but with a richer background. This dataset provides image resolution of 768 576. We also select models with high-frequency details or sharp features which are difficult to reconstruct. In all three datasets, ground truth surfaces and camera poses are provided. Evaluation metrics. To evaluate the quality of the reconstruction, we follow previous work and used Chamfer distance (lower values are better) and PSNR (higher values are better). For the DTU dataset, we use the official evaluation protocol, which means computing the mean of accuracy (distance from the reconstructed surface to the ground truth surface) and completeness (distance from the ground truth surface to the reconstructed surface). For DTU and Blended MVS, the background is not part of the ground truth surface. Therefore, we remove the background for computing the Chamfer distance, following previous work. The Ne RF-synthetic dataset [19] has no background, so we only remove disconnected parts for all competing methods. Table 2: Quantitative results on Ne RF-synthetic and Blended MVS datasets. Metric(10 2) Method Chair Ficus Lego Materials Mic Ship Mean Bread Dog Robot Mean Fidelity Ne RF 2.12 5.17 3.05 1.51 4.77 3.54 3.36 0.102 0.693 2.325 1.07 VOLSDF 1.26 1.54 2.83 1.35 3.62 2.92 2.37 0.074 0.354 1.453 0.63 Neu S 0.74 1.21 2.35 1.30 3.89 2.33 1.97 0.068 0.173 1.036 0.43 OURS 0.69 1.12 0.94 1.08 0.72 2.18 1.12 0.065 0.155 0.922 0.38 Ne RF 33.00 30.15 32.54 29.62 32.91 28.34 31.09 31.27 27.46 25.33 28.02 VOLSDF 25.91 24.41 26.99 28.83 29.46 25.65 26.86 31.05 28.24 25.46 28.25 Neu S 27.95 25.79 29.85 29.36 29.89 25.46 28.05 31.32 28.71 25.87 28.63 OURS 28.69 26.46 30.72 29.87 30.35 25.87 28.66 31.89 29.42 26.15 29.15 Table 3: Ablation study results. Chamfer Distance PSNR Datasets Base Base+H Base+C2F IDF+H IDF+C2F FULL Base Base+H Base+C2F IDF+H IDF+C2F FULL DTU 1.08 1.20 1.07 1.25 0.89 0.78 31.77 32.73 32.56 32.69 32.13 32.49 Ne RF-Synthetic 2.51 3.61 2.95 2.83 1.35 0.91 28.39 30.52 30.63 30.12 29.88 30.31 Blended MVS 0.43 fail 0.63 0.47 0.41 0.38 28.63 fail 27.35 28.20 28.95 29.15 Implementation details. We use MLPs to model two signed distance functions fb and fd. Each MLP consists of 8 layers. Related work like Neu S [30] and IDR [33] also use MLPs with 8 layers. We use Adam with learning rate 5e 4 for the network training using NVIDIA TITAN A100 40GB graphics cards. For adaptive sampling, we first uniformly sample 64 points on the ray, then calculate the SDF and its gradient at these points. We utilize the Eq. 16 to calculate the gain of the s parameter, and then adaptively update the weight according to the gain and sample an additional 64 points. For the coarse-to-fine strategy, we observe that using α0 d = 0 at the beginning for surface reconstruction produces smoothed results. We utilize α0 d = 0.5 and α0 b = 0.5α0 d = 0.25 for both signed distance functions. We set L = 16 for the parameter of the frequency band of positional encoding. For other parameter settings, please see the supplemental materials. Comparison. In table 1, we show quantitative results with other competitors on 15 scenes of the DTU dataset [12]. The values shown in the upper part of the table measure the fidelity of the surface reconstruction, the Chamfer distance. The numbers indicate that HF-Neu S significantly outperforms Ne RF. In most scenes, HF-Neu S is better than Vol SDF and Neu S so that the overall average distance is also improved. In the lower part of the table we show the PSNR values. It can be seen that our PSNR surpasses all other methods. We further compare the visual quality achieved by different methods. As shown in Fig. 4, HF-Neu S can reconstruct high-frequency details. For example, the windows have better geometric details, and the feathers of the bird are more distinct. Most of the scenes in the DTU dataset have smooth surfaces, and high-frequency details are not obvious. We selected 9 challenging models from the Ne RF-synthetic dataset [19] and Blended MVS dataset [31], which have more high-frequency details. For example, the Lego model has uneven repeating bumps, and the power cord of the Mic model has a very thin structure (Fig. 1). The robot model has richer edge and corner features (Fig. 4 second row). As shown in Table 2, the gap between our surface reconstruction quality and that of all other methods widens. This shows that HF-Neu S is especially advantageous for surface reconstructions with high-frequency information. We can also observe that Ne RF is very good in the image-based metric (PSNR) while performing poorly in the surface reconstruction metric (Chamfer distance). This observation is consistent with previous work. Compared with the Ne RF-Synthetic dataset, the Blended BMS dataset has a more complex background, this also restricts the performance of Ne RF to a certain extent. Besides outperforming other baselines in terms of quantitative error, we also achieve better results in terms of qualitative visual effects. As shown in Fig. 1, HF-Neu S can more accurately reconstruct the details of each Lego block and even some of the tiny holes that are not reconstructed by any other method. For the Robot scene, HF-Neu S can reconstruct more accurate facial contours and sharper horns. Finally, for the Mic model, HF-Neu S can clearly reconstruct the power cord, while other methods will mess up this structure. Ablation study. We verify the influence of different modules on the reconstruction results, including the coarse-to-fine module, the implicit displacement function module, and the position-adaptive s control module. In Table 3, Base refers to the baseline method, which is Neu S. "H" means we use high-frequency positional encoding. Here we set L=16 to represent high frequencies. "C2F" refers to the coarse-to-fine optimization strategy with high-frequency positional encoding. We set the initial α to 0.5. "IDF" represents using the implicit displacement function in reconstruction. For each dataset, we chose the mean of the three scenes as the quantitative metric. From the results of the Blended MVS dataset, we can observe that the divergence of network training can be prevented based on the coarse-to-fine strategy. From the DTU and Ne RF-synthetic datasets, introducing high-frequency directly can easily lead to overfitting on these datasets. This means that an increase in PSNR cannot guarantee the improvement of the fidelity of surface reconstruction. Although the coarse-to-fine module can alleviate this mismatch to some degree, it is difficult to further improve the performance. However, adding the implicit displacement function component improves the fidelity of the surface reconstruction and PSNR at the same time. During reconstruction, the network with adaptive s can help to improve the reconstruction quality upon more complex scenes. Limitation. As shown in Fig. 5, our method still has challenges. We show a reference ground truth image, our corresponding reconstructed image, and our reconstructed surface. For the grid of ropes of the ship, some overfitting to ground-truth radiance is still observed. Specifically, the grid of ropes is visible in the image, but the surface is not reconstructed accurately. Another limitation is that the individual thin ropes are missing. We also visualize a bad case of Table 1 where the error is larger than that of the other methods as shown in Fig. 14 DTU Bunny in the supplementary material. In this case, the lighting of this model varies and the texture is not as pronounced, thus it is difficult to reconstruct the details of the belly. Further, integrating our proposed IDF increases computation time. Figure 5: Limitation. First column: the reference ground truth images. Second column: our synthetic images. Last column: our reconstructed surface. 5 Conclusion We introduce HF-Neu S, a new method for multi-view surface reconstruction with high-frequency details. We propose a new derivation to explain the relationship between signed distance and transparency and propose a class of functions that can be used. By decomposing the signed distance field into a combination of two independent implicit functions, and using adaptive scale constraints to focus on optimizing the regions where the implicit function distribution is not ideal, a more refined surface can be reconstructed compared to previous work. The experimental results show that the method outperforms the current state of the art in terms of quantitative metrics and visual inspection. An interesting direction for future work is to explore the reconstruction of scenes under different lighting modalities. Finally, we do not expect negative social impacts that will be directly linked to our research. Negative social impacts of surface reconstruction in general are possible though. Acknowledgements We would like to acknowledge support from the SDAIA-KAUST Center of Excellence in Data Science and Artificial Intelligence and the NSFC No.62202076. [1] C. Barnes, E. Shechtman, A. Finkelstein, and D. B. Goldman. Patchmatch: A randomized correspondence algorithm for structural image editing. ACM Trans. Graph., 28(3):24, 2009. [2] J. T. Barron, B. Mildenhall, M. Tancik, P. Hedman, R. Martin-Brualla, and P. P. Srinivasan. Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5855 5864, 2021. [3] F. Bernardini, J. Mittleman, H. Rushmeier, C. Silva, and G. Taubin. The ball-pivoting algorithm for surface reconstruction. IEEE transactions on visualization and computer graphics, 5(4):349 359, 1999. [4] A. Broadhurst, T. W. Drummond, and R. Cipolla. A probabilistic framework for space carving. In Proceedings eighth IEEE international conference on computer vision. ICCV 2001, volume 1, pages 388 393. IEEE, 2001. [5] F. Darmon, B. Bascle, J.-C. Devaux, P. Monasse, and M. Aubry. Improving neural implicit surfaces geometry with patch warping. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6260 6269, 2022. [6] J. S. De Bonet and P. Viola. Poxels: Probabilistic voxelized volume reconstruction. In Proceedings of International Conference on Computer Vision (ICCV), pages 418 425, 1999. [7] Y. Furukawa and J. Ponce. Accurate, dense, and robust multiview stereopsis. IEEE transactions on pattern analysis and machine intelligence, 32(8):1362 1376, 2009. [8] S. Galliani, K. Lasinger, and K. Schindler. Gipuma: Massively parallel multi-view stereo reconstruction. Publikationen der Deutschen Gesellschaft für Photogrammetrie, Fernerkundung und Geoinformation e. V, 25(361-369):2, 2016. [9] R. Hartley and A. Zisserman. Multiple view geometry in computer vision. Cambridge university press, 2003. [10] A. Hertz, O. Perel, R. Giryes, O. Sorkine-Hornung, and D. Cohen-Or. Sape: Spatially-adaptive progressive encoding for neural optimization. In Advances in Neural Information Processing Systems, 2021. [11] S. Izadi, D. Kim, O. Hilliges, D. Molyneaux, R. Newcombe, P. Kohli, J. Shotton, S. Hodges, D. Freeman, A. Davison, et al. Kinectfusion: real-time 3d reconstruction and interaction using a moving depth camera. In Proceedings of the 24th annual ACM symposium on User interface software and technology, pages 559 568, 2011. [12] R. Jensen, A. Dahl, G. Vogiatzis, E. Tola, and H. Aanæs. Large scale multi-view stereopsis evaluation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 406 413, 2014. [13] M. Kazhdan, M. Bolitho, and H. Hoppe. Poisson surface reconstruction. In Proceedings of the fourth Eurographics symposium on Geometry processing, volume 7, 2006. [14] K. N. Kutulakos and S. M. Seitz. A theory of shape by space carving. International journal of computer vision, 38(3):199 218, 2000. [15] P. Labatut, J.-P. Pons, and R. Keriven. Efficient multi-view reconstruction of large-scale scenes using interest points, delaunay triangulation and graph cuts. In 2007 IEEE 11th international conference on computer vision, pages 1 8. IEEE, 2007. [16] M. Li and H. Zhang. D2im-net: Learning detail disentangled implicit fields from single images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10246 10255, 2021. [17] C.-H. Lin, W.-C. Ma, A. Torralba, and S. Lucey. Barf: Bundle-adjusting neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5741 5751, 2021. [18] L. Liu, J. Gu, K. Zaw Lin, T.-S. Chua, and C. Theobalt. Neural sparse voxel fields. Advances in Neural Information Processing Systems, 33:15651 15663, 2020. [19] B. Mildenhall, P. P. Srinivasan, M. Tancik, J. T. Barron, R. Ramamoorthi, and R. Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In European conference on computer vision, pages 405 421. Springer, 2020. [20] M. Niemeyer and A. Geiger. Giraffe: Representing scenes as compositional generative neural feature fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11453 11464, 2021. [21] M. Niemeyer, L. Mescheder, M. Oechsle, and A. Geiger. Differentiable volumetric rendering: Learning implicit 3d representations without 3d supervision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3504 3515, 2020. [22] M. Nießner, M. Zollhöfer, S. Izadi, and M. Stamminger. Real-time 3d reconstruction at scale using voxel hashing. ACM Transactions on Graphics (To G), 32(6):1 11, 2013. [23] M. Oechsle, S. Peng, and A. Geiger. Unisurf: Unifying neural implicit surfaces and radiance fields for multi-view reconstruction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5589 5599, 2021. [24] K. Park, U. Sinha, J. T. Barron, S. Bouaziz, D. B. Goldman, S. M. Seitz, and R. Martin-Brualla. Nerfies: Deformable neural radiance fields. International Conference on Vomputer Vision, 2021. [25] J. L. Schonberger and J.-M. Frahm. Structure-from-motion revisited. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4104 4113, 2016. [26] J. L. Schönberger, E. Zheng, J.-M. Frahm, and M. Pollefeys. Pixelwise view selection for unstructured multi-view stereo. In European Conference on Computer Vision, pages 501 518. Springer, 2016. [27] S. M. Seitz and C. R. Dyer. Photorealistic scene reconstruction by voxel coloring. International Journal of Computer Vision, 35(2):151 173, 1999. [28] V. Sitzmann, J. Martel, A. Bergman, D. Lindell, and G. Wetzstein. Implicit neural representations with periodic activation functions. Advances in Neural Information Processing Systems, 33:7462 7473, 2020. [29] M. Tancik, P. Srinivasan, B. Mildenhall, S. Fridovich-Keil, N. Raghavan, U. Singhal, R. Ramamoorthi, J. Barron, and R. Ng. Fourier features let networks learn high frequency functions in low dimensional domains. Advances in Neural Information Processing Systems, 33:7537 7547, 2020. [30] P. Wang, L. Liu, Y. Liu, C. Theobalt, T. Komura, and W. Wang. Neus: Learning neural implicit surfaces by volume rendering for multi-view reconstruction. Advances in Neural Information Processing Systems, 2021. [31] Y. Yao, Z. Luo, S. Li, J. Zhang, Y. Ren, L. Zhou, T. Fang, and L. Quan. Blendedmvs: A large-scale dataset for generalized multi-view stereo networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1790 1799, 2020. [32] L. Yariv, J. Gu, Y. Kasten, and Y. Lipman. Volume rendering of neural implicit surfaces. Advances in Neural Information Processing Systems, 34, 2021. [33] L. Yariv, Y. Kasten, D. Moran, M. Galun, M. Atzmon, B. Ronen, and Y. Lipman. Multiview neural surface reconstruction by disentangling geometry and appearance. Advances in Neural Information Processing Systems, 33:2492 2502, 2020. [34] W. Yifan, L. Rahmann, and O. Sorkine-hornung. Geometry-consistent neural shape representation with implicit displacement fields. In International Conference on Learning Representations, 2022. [35] K. Zhang, G. Riegler, N. Snavely, and V. Koltun. Nerf++: Analyzing and improving neural radiance fields. ar Xiv preprint ar Xiv:2010.07492, 2020. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] (c) Did you discuss any potential negative societal impacts of your work? [Yes] (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] see Section 4 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [Yes] See Section4 (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [Yes] All datasets are public. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [Yes] See Section 5 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]