# geometry_processing_with_neural_fields__a770da97.pdf Geometry Processing with Neural Fields Guandao Yang Cornell University Serge Belongie University of Copenhagen Bharath Hariharan Cornell University Vladlen Koltun Intel Labs Most existing geometry processing algorithms use meshes as the default shape representation. Manipulating meshes, however, requires one to maintain high quality in the surface discretization. For example, changing the topology of a mesh usually requires additional procedures such as remeshing. This paper instead proposes the use of neural fields for geometry processing. Neural fields can compactly store complicated shapes without spatial discretization. Moreover, neural fields are infinitely differentiable, which allows them to be optimized for objectives that involve higher-order derivatives. This raises the question: can geometry processing be done entirely using neural fields? We introduce loss functions and architectures to show that some of the most challenging geometry processing tasks, such as deformation and filtering, can be done with neural fields. Experimental results show that our methods are on par with the well-established mesh-based methods without committing to a particular surface discretization. Code is available at https://github.com/stevenygd/NFGP. 1 Introduction In many graphics applications, users may want to edit digital shapes using just a few clicks, such as making a character bow by dragging the head downwards. Such manipulation from sparse input requires geometry processing algorithms. Most of these algorithms use polygonal meshes to represent shapes [11]. Polygonal meshes were created initially for researchers as a representation of real-world shapes that they can both interpret and manipulate [14, 52, 78]. However, developing algorithms that automatically manipulate meshes is often difficult since they involve discretizing the surface. For example, changing the topology of a mesh, such as turning a sphere into a torus, will break such discretization and require additional repair procedures such as remeshing [2, 3]. Given that shape editing is increasingly performed by algorithms, it seems worthwhile to search for a shape representation that is more amenable to automatic geometry processing. An alternative to the polygonal mesh is an implicit representation, in which the surface is represented by a level set of a field: {x|f(x) = c} [22, 39, 63]. Since it is easy to change shape topology using implicit representations, people have applied them for geometry processing tasks such as shape merging [8, 51, 54]. These works store implicit fields using voxels or octrees, which introduce memory-intensive spatial discretization. Recent research addresses this by using continuous neural networks to represent implicit fields [19, 45, 46, 56]. These neural fields possess several advantages in addition to the merits inherited from implicit representation: they are compact to store [23, 45] and can produce high-quality continuous surfaces at arbitrary resolutions [45, 56]. The community has achieved compelling results using neural fields in a variety of applications [47, 64, 73, 84]. Notwithstanding these results, it is still unclear whether shape editing tasks challenging for implicit fields (e.g., deformation and filtering) can be performed with neural fields. In this paper, we ask: can geometry processing be done entirely using neural fields? Email: gy46@cornell.edu. This work is done while Guandao was interning at Intel Labs. 35th Conference on Neural Information Processing Systems (Neur IPS 2021). Geometry processing tasks are challenging because they are typically under-constrained: user input is very sparse. Therefore, we need to leverage priors that characterize how a natural surface behaves. Such surface priors are usually instantiated by minimizing energy functionals that describe the physical properties of surfaces using operators from differential geometry. For example, one can encourage surfaces to be smooth by minimizing the curvature, which can be measured by the Laplace Beltrami operator. Such operators, however, require a parameterization of the surface, which is not available for level sets of a neural field. Mesh algorithms usually approximate these geometric operators using the geodesic neighbors of the surface point. This allows the algorithms to focus on intrinsic geometric properties that are independent of shape parameterization. However, it is not easy to obtain geodesic neighbors in a neural field since the surface of interest is encoded implicitly by a set of points that evaluate the field function to the same value. We posit that computing these geometric operators in neural fields requires a fundamentally different approach. The idea is to approximate the local surface of the level set using the derivatives of the underlying field. We can evaluate intrinsic geometry properties of the level set, such as curvature, using only the field derivatives [39, 54]. This allows us to develop loss functions that describe surface priors such as elasticity or rigidity. To achieve this, we use the fact that neural fields are designed to be infinitely differentiable [66, 74]. The infinite differentiability of neural fields makes it possible to optimize loss functions that involve higher-order derivatives using gradient descent methods. Thus, unlike mesh-based geometry processing algorithms that approximate these objectives using surface discretizations, we can directly optimize in terms of the derivatives of the field. Our formulation provides a proof of concept that geometry processing can be done entirely using neural fields. In particular, we focus on two tasks: shape filtering (e.g., smoothing and sharpening) and topology-preserving deformations. These two tasks are not only essential for downstream applications but also bring out the known challenges associated with implicit representations. We first tackle shape sharpening and smoothing and show how these can be done by optimizing a loss function based on geometric properties computed via the neural field s derivatives. To achieve shape deformation, we propose to warp the neural field using a deformation vector field modeled by an invertible neural network [7]. With this invertible deformation field, we derive a training objective that models the implicitly represented surfaces as elastic shells. We compare our method with well-established mesh-based baselines [70, 76]. Our method can match the quality of the mesh-based counterparts without discretizing the surface. We hope that our work can inspire future generations of geometry processing algorithms using neural fields. 2 Related Work This paper builds on two bodies of work: polygon mesh geometry processing and neural field representations. We will focus on prior work on two geometry processing tasks: shape filtering and topology-preserving deformation. Please refer to Botsch et al. [11] for further reading. Mesh smoothing (and sharpening). The goal of mesh smoothing is to remove high-frequency noise and produce a smooth surface. Noise removal can be achieved by Fourier analysis [76, 79] or by modeling surface motion with a diffusion equation [26]. Smoothness can be promoted by minimizing energy functionals inspired by differential operators [50, 72, 82]. Sharpening is the inverse of smoothing. In this paper, we adopt the goal of smoothing from prior work [20, 76] and design a corresponding differentiable objective for neural fields. Implicit field smoothing. Implicit representations can be smoothed by evolving the level set according to its curvature normal [49, 54]. Many network architectures, such as Re LU MLP, bias toward smoothness when modeling an implicit field [61, 66, 74]. Regularization objectives and initialization schemes have been proposed to obtain smooth level sets from scanned surface points [5, 31]. These works mostly use smoothness as a prior when training neural fields. Our work formulates an objective that allows smoothing or sharpening a neural field as desired. Mesh deformation. Mesh deformation is done by either deforming the shape (i.e. vertices) [9, 69] or the space [6, 36, 41, 48, 65]. These algorithms usually draw inspiration from physics and differential geometry to develop energy functions that encourage natural deformation [10, 15, 24, 70, 71, 77]. They also linearize such objectives so that they can be solved efficiently using linear solvers [9]. Recent research also applies deep learning to optimize for non-linear deformation losses [32, 34, 38, 80, 81, 83]. Deforming a mesh, however, requires estimating the deformation objectives with spatial discretization and maintaining vertex connectivity in a way that preserves topology [42]. We deform neural fields with a continuous invertible field to circumvent these issues. Implicit field deformation. Prior works have studied the deformation of implicit fields for tasks like physics simulation [39, 51, 54, 55]. These works use voxels or octrees to represent implicit fields, while our paper advocates for using neural fields for deformation. There is some past work on deforming neural fields to match a target shape [12, 18, 37, 53, 58] or image [43, 57, 60] without any correspondences between the two. In contrast, we perform such deformation with very few localized user-provided correspondences, but without any other information about the target shape. Our setting is very convenient for artists, but it requires the algorithm to provide a strong surface prior due to the lack of a densely specified target state. Other prior works on editing neural fields are trained on a dataset of shapes from a particular object category [25, 29, 30, 33, 85]. Since surface priors are specified implicitly by the dataset, it is unclear how to apply these methods to out-of-distribution shapes. In contrast, our paper requires no training dataset and instead enforces priors inspired by more general physical properties such as smoothness and elasticity. Remelli et al. [62] developed a way to differentiate iso-surface extraction, which allows deforming neural fields using objectives carried over from mesh-based algorithms. Our method avoids iso-surface extraction entirely and formulates differentiable deformation objectives directly on the implicit field. 3 Computing Surface Properties of Neural Fields In this section, we will discuss how to compute surface properties using neural fields and establish notation for the following sections. Let Ωbe a surface enclosing the region Ω. A signed distance field (SDF) for surface Ωis defined as d(x) = s(x) miny Ω x y . The sign function s(x) evaluates to 1 if x Ωand 1 otherwise. We will assume that a neural field approximate an SDF whose zero-isosurface represents the surface of interest. For a neural field f, we denote its zero-isosurface using Mf. This section summarizes how to compute surface normals and curvatures from neural fields. This is well-understood and we summarize this here for convenience [39, 54, 59]. Surface normal. The surface normal is the vector perpendicular to the tangent plane and describes the local orientation of the surface. For an SDF, the outward oriented surface normal of its level set is the field gradient: nd(x) = xd(x). It can be shown that the SDF s gradient norm is always one: xd(x) = 1. This property can be used to ensure that the field remains a valid SDF throughout any manipulation [4, 31]. The closest point in Ωfrom x can be found by x d(x)nd(x). This property can be used to sample points from the isosurface without creating a mesh [5, 13]. Finally, since the tangent plane is perpendicular to the surface normal, we can project vectors onto the tangent plane by subtracting their projection onto the normal using the following projection matrix x: Pd(x) = I nd(x)nd(x)T . This projection matrix allows us to characterize the tangent plane. This matrix is important for measuring how much a given deformation stretches the tangent direction [35]. Curvature. Intuitively, curvature describes how much a surface deviates from a plane. This can be captured by the total derivative of the surface normal, which is also known as the shape operator: SD(x) = Dnd(x). For an SDF, the normal itself is given by the derivative of the field, thus the shape operator for an SDF is the Hessian of the field function: Hd(x) = D2d(x). The shape operator can be used to capture many different notions of curvature, including mean curvature (half the trace of SD(x), denoted by κ), Gaussian curvature (the determinant of SD(x)), and principal curvatures (the eigenvalues of SD(x)). We will use these to define objectives that smooth or sharpen the surface. 4 Shape Smoothing and Sharpening In this section, we will show that shape smoothing and sharpening can be done directly on neural fields without producing meshes. In our setup, the input shape is represented as the zero-isosurface MF of the neural field F. The algorithm needs to output neural field Gθ whose zero-isosurface MGθ satisfies two goals. First, MGθ should preserve the global structure of MF . Second, the surface MGθ should make desired changes to match the curvature of shape MF . Figure 2: Smoothing and sharpening results. (A) zoomed-in input; (B) our method; (C) baseline without remeshing; (D) baseline with remeshing. We can see that the baseline without remeshing finds it difficult to smooth or sharpen the mesh extracted from neural fields. Our method is able to achieve good results without discretizing the surface. To achieve these goals, we first instantiate Gθ using the same architecture and parameter values as F. Then we optimize Gθ using the following objective: |Gθ(x) F(x)|2 + λg ( x Gθ(x) 1)2 dx + Z λk (κGθ(x) βκF (x))2 dx. Figure 1: L: SIREN learned to fit the SDF of a square; R: Zooming in on the surface. The first integral encourages the network Gθ to preserve the original shape. The second term regularizes Gθ to remain a valid SDF by enforcing that the norm of the gradient is 1 (i.e., the Eikonal constraint [4, 31]). The integration is over the region U where the original neural field F is supervised over. The final term aims to smooth or sharpen the surface by increasing or decreasing the curvature κ. Setting β < 1 will decrease the curvature of the output surface κGθ, resulting in a smoother shape. β > 1 will lead to surfaces with higher curvature details and will thus sharpen shapes. In this paper, we will use mean curvature since it s easy to compute with neural fields: κf(x) = tr D(nf(x)). One challenge is that the computation of curvature can be very noisy for neural fields using periodic activations. Figure 4 shows how an ostensibly smooth isosurface learned through SIREN [66] is actually quite rough when zoomed in. The curvature evaluated on such a rough surface can be too noisy to be used for training. To alleviate this issue, we only compute the curvature regularization in areas where the curvature of the level set is less than a certain threshold τ. Formally, we define this area as Vτ = {x U| max(|κGθ|, |κF |) < τ}. We use rejection sampling to sample points from Vτ when computing the loss during training. Results. We follow prior works [20, 76] to use Armadillo [40] and a sphere with one half of it corrupted by Gaussian noise. We compare our algorithm to two baselines. The first baseline directly applies the smoothing algorithm developed by Taubin [76] to the mesh extracted from the neural field using Marching cube [44]. The second baseline applies the filtering algorithms on meshes simplified by quadratic decimation [27]. The input neural fields are created following the procedure of Park et al. [56]. The results are shown in Figure 2. The first baseline fails to smooth or sharpen the surface appropriately. It only modifies the surface with high-frequency noise. While the second baseline can filter the surface correctly, it introduces discretization artifacts due to the decimation process. This suggests it is challenging to filter the neural field surface with mesh processing algorithms since these algorithms are sensitive to the quality of surface discretization. On the other hand, our algorithm can produce good filtering results without the need to maintain a good discretization of the surface. 5 Deformation To deform a shape, the user will choose a set of deformation handles {hi}n i=1. For each handles hi, the user will specify a target location ti describing where the handle will be dragged to. Users can set ti = hi to enforce a part of the surface to be unchanged. Given this input, there are two goals for the shape deformation algorithm: 1) satisfying user inputs and 2) ensuring that the deformation resembles the natural behavior of real objects. The input shape will be represented by the zero-isosurface of a neural field F in our setting. The algorithm must therefore output a neural field Gθ whose zero-isosurface represents the deformed shape that satisfies the aforementioned two goals. In this setting, deformation can be formulated as solving a constrained optimization problem: min θ Ldfm(Gθ, F) s.t. Lconst(Gθ, ti, hi) = 0, 1 i n, (2) where the objective Ldfm(Gθ, F) measures how natural the deformation is. The constraints Lconst(Gθ, ti, hi) = 0 ensure that the result deformation matches user input. One way to ensure a natural-looking deformation is to assume that the surface behaves like a thin elastic shell, which is resistant to stretching and bending [1, 70]. Traditional mesh-based algorithms model elasticity by minimizing the thin shell energy that penalizes stretching or bending happened to any local patch. Naively adapting such training objectives from mesh-based algorithms to neural fields is challenging because these algorithms assume dense correspondences between two surfaces and efficient access to geodesic neighbors. In this section, we will address these challenges by following strategy. To measure how much bending or stretching happened to the whole surface, we will first divide the surface into infinitesimal patches and then sum the amount of bending or stretching that happened to each patch. To achieve that, we first develop a sampling schema to sample these infinitesimal patches uniformly from the input shapes (Sec 5.1). Then we warp the input shape with an invertible neural network to achieve correspondences between the input and output surfaces (Sec 5.2). Finally, we derive ways that measure the amount of bending and stretching between two infinitesimal surface patches (Sec 5.3). 5.1 Sampling Figure 3: Sampling from dino. L: no rejection; R: ours (with rejection). Our first step is to sample infinitesimal surface patches uniformly to the surface area from the zero-isosurface of the neural field. Each infinitesimal surface patch can be represented by a surface point and its local derivatives. With this said, the problem amounts to sampling points from the zero-isosurface. Formally, we are given a neural field F that approximates the signed distance field of the input shape. The goal is to find a way to sample an arbitrary number of points on its zero-isosurface MF . One way to sample points from MF is to run Langevin dynamics using the gradient of the field F [5, 13]: xt+1 = xt F( xt)n F ( xt), xt N(xt, σI). x0 are initialized by sampling uniformly within [ 1, 1]3. Here we also leverage an SDF property: the closest surface point to x can be computed by x n(x)F(x) (Sec 3). While this process produces points guaranteed to be on the zero-isosurface, the mixing time of the procedure is very long without appropriate coarse-to-fine annealing [68]. Running this process for limited iterations results in point samples that are concentrating near the high curvature area, as shown on the left side of Figure 5.1. To alleviate this issue, we instantiate the sampling procedure with x0 sampled from a rough shape. This can be done by first sampling uniformly from the bounded space and rejecting points that are too far away from the isosurface: x0 {x U( 1, 1)|F(x) < τs}. We found that τs can be set to a relatively large number (e.g., about 0.1). This prevents high rejection rates without sacrificing sampling quality. Figure 5.1 shows that our method creates uniform samples. We quantitatively evaluate the effectiveness of the sampling scheme. Specifically, we compare our sampling scheme with directly applying Langevin dynamics [13] (LD in Table 1). To do that, we compute the Chamfer Distance (CD) and Earth Mover Distance (EMD) between points sampled by the valuated algorithms and points sampled uniformly from the extracted mesh. For both metrics, lower values indicate better performance Since larger surfaces can lead to larger CD and EMD values, we normalize the metrics by the values obtained by comparing two sets of uniform samples from the mesh (i.e., CDr and EMDr) The results show that initializing x0 by rejection sampling significantly improves both CD and EMD compared to naively applying Langevin dynamics, at a minor cost in sampling time. 5.2 Invertible Deformation Field Table 1: Our sampling method achieves better results than the baseline. Dino Armadillo Metrics LD Ours LD Ours CDr ( ) 1.54 1.04 1.36 1.02 EMDr ( ) 3.38 1.15 3.30 1.08 Time 0.15 0.21 0.12 0.18 After we obtained surface points samples from zero isosurfaces of the input neural fields, the next step is to deform these surface points to the output shapes. One way to deform neural fields is to warp the coordinate space of the input network using a deformation field predicted by a neural network Dθ [25, 43, 57]. Under this framework, the output network can be defined as Gθ(x) = F(Dθ(x)). Then the constraint Lconst(Gθ, t, h) can be defined as Lconst(Gθ, t, h) = Dθ(t) h . Ideally, we would like the deformation Dθ to be continuous and invertible. This will allow us to create well-behaved one-to-one correspondences between points on the deform shape (i.e., x) and points on the input shape (i.e., Dθ(x)). To achieve this, we make Dθ to be an invertible network composed of a sequence of invertible residual blocks [7]. Invertible residual block. Prior works [7, 17] have shown that a sufficient condition for the residual block f(x) = x + g(x) to be invertible is that the Lipschitz constant of function g is less than 1. The architecture of the network g(x) is usually composed of spectral normalized linear layers and Lipschitz continuous nonlinearities such as ELU [21]. The inverse of such residual block can be computed by finding the fixed point of function y 7 y g(x) [7]. Intuitively, deforming using one invertible residual block amounts to moving point x with deformation vector g(x). Lipschitz continuous positional encoding. Applying such architecture directly without modification fails to produce deformations with many different local rotations. Recent research suggests that periodic functions are essential for coordinate MLPs to predict complex signals [47, 66, 74]. Periodic functions are usually used as positional encoding or activations of the form sin(ax + b), where a controls the frequency of the activation. To use such periodic function as part of the invertible residual block without restricting the frequency, we normalize the output of the periodic function sin(ax + b)|a| 1, bringing it is Lipschitz constant below 1. Formally, the positional encoding with normalized Lipschitz constant applied to one dimension is xi, cos(20πxi) 20π , sin(20πxi) 20π , . . . , cos(2Lπxi) 2Lπ , sin(2Lπxi) We will apply such encoding to each of the dimensions of the input coordinate for the body of the invertible residual block: Rθ(x) = x + gθ 1 d[γ1 (x) , . . . , γd (x)] . No inverse No PE Ours Figure 4: Architecture ablation on 2D SDF deformation. Blue points are handles; Red points are targets. Ablation. We conduct ablation studies on two architecture choices: invertibility and positional encoding. In this experiment, we deform a neural field representing the SDF of a 2D rectangle. We optimize each ablation case to satisfies the user specified constraints with following loss: 1 n Pn i=1 Dθ(hi) ti 2. The results are shown in Figure 4. If we replace the invertible architecture with SIREN as done in Deng et al. [25], the deformation field will tend to break topology (No inverse). Removing the positional encoding will fail to create a complex field (No PE). Our architecture can create a reasonable guess (No loss). This shows that our network architecture prioritizes natural deformation that preserves topology. 5.3 Implicit Thin Shell Loss At this point, we have obtained the correspondences y = Dθ(x), with x in the output (or deformed) space and y in the input space. Our next step is to design Ldfm that compares these corresponding patches to ensure natural deformation. Recall that one way to ensure a natural-looking deformation is to minimize the amount the resistance to bending or stretching that happened during the deformation [1, 70]. We will develop novel loss functions for measuring stretching and bending. Please refer to the supplement for the connection between our loss and the thin shell energy. 5.3.1 Stretching Loss Stretching can be captured by the change of dot product in tangent space. Intuitively, a local surface patch is stretched when the lengths of some tangent vectors change. The change of tangent vectors norm can be measured by the tangent dot-product. With this said, the amount of stretch can be measured by the change of tangent dot product. Note that we do not have access to analytical surface parameterization, so we need to express the tangent dot product with neural fields and its derivative. We can describe the tangent dot product to the spatial dot product using the projection matrix. The tangent vector of x can be parameterized by projecting a vector v to the tangent space. To achieve that, we need to multiply v with the projection matrix PGθ(x) = I n Gθ(x)n Gθ(x)T , where n Gθ(x) is the surface normal of point x. PGθ(x)v is a tangent vector in the tangent plane of x. Now we are ready to compute the change of tangent dot-product. Let ti and tj be two arbitrary tangent vectors near point x at the deformed shape. Further assume that these vectors can be parameterized as ti = PGθ(x)v1 and tj = PGθ(x)v2. These tangent vector will be transformed by Dθ into t i = JDθ(x)ti and t j = JDθ(x)tj. These are tangent vectors at point y at the input shape. The change of tangent dot-product with respect to these two vectors can be computed as follows: |t T 1 t2 t T 1 t 2| = v T 1 PGθ(x)T I JDθ(x)T JDθ(x) PGθ(x)v2 . (4) To minimize the stretch, we need to enforce that dot-product stays the same for all tangent vectors. This can be done by minimize the matrix norm of PT Gθ I JT DθJDθ PGθ. Here, we drop the function arguments x for the matrix for notation clarity. With these, we define the the stretch loss as: PT Gθ I JT DθJDθ PGθ 2 F dx. (5) 5.3.2 Bending Loss Bending can be characterized by the change of surface curvature (e.g., making the surface more or less curved). Intuitively, curvature can be described as the change of tangent dot product along the surface normal direction [16, 59]. Let t1 = PGθ(x)v1 and t2 = PGθ(x)v2 be two tangent vectors at x. Consider the tangent dot product of the surface family x + tn(x). Note that this set of surface corresponding to the level sets {p|Gθ(p) = t} if Gθ is approximating an SDF [54]. Then the change of dot product along the surface normal direction can be given by Hessian and directional derivative: d dtt T 1 t2|t=0 = t T 1 HGθ(x)t2. Intuitively, the larger this value is, the faster the surface changes when going along the surface normal direction, which means the surface has larger curvature. Measuring bending amounts to measure the change of tangent dot-product derivative along the surface normal direction. Similar to the previous section, assume we have tangent vectors ti = PGθ(x)vi and tj = PGθ(x)vj. These tangent vectors are are transformed by the Jacobian of the deformation field JDθ(x) to t i = JDθ(x)ti and t j = JDθ(x)tj. Since t i and t j are tangent vectors at point y on the surface MF , the derivative of tangent dot-product is given by the hessian matrix of field F: d dtt T 1 t 2|t=0 = t T 1 HF (Dθ(x))t 2 = t T 1 JT Dθ(x)HF (Dθ(x))JDθ(x)t2. We will drop the function argument to JDθ, HGθ, and PGθ for notation clarity. The change of tangent dot-product derivative is: d dt t T 1 t2 t T 1 t 2 = v T 1 PT Gθ HGθ JT DθHF (Dθ(x))JDθ PGθv2 . (6) If there is almost no bending happened between the infinitesimal patches around x and y, then the quantity d dt t T 1 t2 t T 1 t 2 t=0 should stay close to 0 for all pairs of tangent vectors. We will quantify this by the matrix norm of PT Gθ HGθ JT DθHF (Dθ(x))JDθ PGθ. Finally, we arrive at our bending loss, which minimizes the norm of the matrix that modulate the change of tangent dot-product derivatives moving along the surface normal direction: PT Gθ HGθ JT DθHF (Dθ(x))JDθ PGθ s F dx. (7) 5.3.3 Computing Ls and Lb Computing Ls and Lb during training requires approximating a surface integral of the form R x MGθ L(x)dx using Monte Carlo integration, which requires an efficient way to sample points uniformly from surface MGθ. While it is feasible to sample uniformly from an SDF, this is challenging to do with Gθ since there is no guarantee that Gθ remains a valid SDF during the course of training. Thanks to the invertibility of Dθ, we can apply change of variable x = D 1 θ (y) to the integration, where y are points on the input surface MF : Z x MGθ L(x)dx = Z y MF L(x) det JDθ(x)PGθ(x) + n F (y)n Gθ(x)T 2 dy. (8) JDθPGθ R3 3 maps tangent vector from MGθ to MF . We apply the extension trick from Iglesias et al. [35] to create a matrix whose determinant equals the change of surface area. This is achieved by adding the surface normal component n F n T G and keeping it unchanged. With this change of variable, we now can compute the losses with point samples from the input field F. Since the input field F approximates an SDF, we can use the method introduced in Section 5.1 to sample uniform points. 5.4 Optimization Putting the losses together, we deform a neural field by solving the constrained optimization problem arg min θ λs Ls(Gθ) + λb Lb(Gθ), s.t. 1 i n, Dθ(ti) hi = 0 (9) where λs, λb are hyperparameters that determine the material properties. One way to solve this constrained optimization problem for a neural network is to make the constraints a soft loss adding to the objective function: Lconst(Dθ) = 1 n Pn 1=1 Dθ(ti) hi . Our final objective is: L(θ) = Ls(Gθ) + λb Lb(Gθ) + λc Lc(Dθ). (10) We set λc to be a high value to enforce that the model satisfies user-specified input as much as possible. The user can tune λb and λs depending on the application scenario. We optimize L(θ) with Adam optimizer to obtain the output field Gθ. Hyperparameters are provided in the supplement. 6 Deformation Results In this section, we will demonstrate the results of our methods to shape deformation. The shapes for deformation are taken from Sorkine-Hornung and Alexa [70]. To create neural fields from these meshes, we follow the procedure of Park et al. [56] to compute ground-truth SDF for locations sampled within [ 1, 1]3. We then fit a SIREN [66] to the ground-truth SDF to generate our initial neural fields. Our main baseline is ARAP [70, 86]. The simplest way to use ARAP to edit neural fields is applying it on a mesh extracted from the input neural fields using marching cubes [44, 45]. We first present results comparing with ARAP applied to the extracted mesh. Then we will conduct an analysis to show our losses encourage the right behaviors. Comparing to ARAP baseline. We follow ARAP [70] to create a set of basic shape deformation operations to evaluate our algorithm. In this setting, the user will first specify a set of handles used for manipulation (shown in Figure 5 in blue). In general, the user can apply three basic types of operation on a handle: 1) make it static (i.e., no deformation), 2) translate it, or 3) rotate it around a center. Usually, the first operation is used in combination with the latter two to produce useful deformation. We first show how our algorithm deforms simple objects (e.g., a cylinder and a rectangle bar) when the user rotates or translates one end while fixing the other. We then extend this set of operations to shapes with more detail (e.g., Cactus, Armadillo, and Dino). Finally, we test our algorithms when all three operations are specified together on a single shape. The results are shown in Figure 5. We can see that applying ARAP directly on the extracted mesh creates undesirable volume distortion. Our algorithm is able to produce deformation results that are natural while satisfying the user s intention. Figure 5: Deformation results. (A) Input shape. (B) Baseline. (C) Ours. Red points are user specified handles hi. Blue points denote user specified target points ti. First row: single rotation or translation on simple shapes. Second and third rows: single rotation or translation on more complex objects. Fourth row: multiple edits on complex objects. Comparing to different ARAP variants. This artifact is partially due to the surface discretization made by the marching cubed algorithm doesn t agree with the assumptions made by ARAP [41]. To verify this hypothesis, we also include three additional baselines in Figure 6. First, we show the results of applying ARAP on the low-poly mesh used to create the input neural fields (Original). Second, we first apply Garland and Heckbert [28] to simplify the surface of the extracted mesh before applying ARAP to it (Remeshed). Finally, we also compare to the results of SR-ARAP [41], which adds smoothness regularization to ARAP to remove some unnatural distortion (SR-ARAP). The figure shows that ARAP works much better when the surface is discretized appropriately. Discretizing the surface, however, usually requires case-by-case manual adjustment to perform well. Our algorithm does not suffer from such issues as we optimize the thin shell objective directly without committing to a particular discretization of the surface. Loss analysis. Here we provide an experiment to analyze the behavior of stretching loss and bending loss. Specifically, we compare models optimized for only the stretching loss Ls, only the bending loss Lb, and both losses together. Similar to Section 5.2, we deform a neural field that approximates the SDF of a 2D rectangle. The results are shown in Figure 7. Optimizing only for Ls produces a shape that best preserves surface area (or the perimeter in the 2D case), which explains the wiggling. Optimizing only for Lb produces a shape that tries to prevent wiggling of the surface, which changes the curvature of the surface. The user can choose appropriate λs and λb to produce deformation with resistance to both bending and stretching. User Input MC Original Remeshed SR-ARAP [41] Ours Figure 6: Additional ARAP baselines. Directly applying ARAP to extracted mesh produces poor results (i.e., MC). While this can be alleviated by remshing (i.e. Original or Remeshed) or regularization (i.e. SR-ARAP), this shows mesh deformation algorithms can be sensitive to discretization quality. Our method do not discretize the surface, thus circumvent such issue. Red points are handles; Blue points are targets. 7 Discussion, Limitations, and Future work Ls only Ls + Lb Lb only Figure 7: Analysis of stretching and bending losses. Blue points are handles; Red points are targets. Advantages of using neural fields for geometry processing. The main strength of using neural fields for geometry processing is that one can disentangle the processing algorithm from the surface discretization. Users do not need to worry about discretization while manipulating a shape represented in neural fields. This might open the door to more automatic geometry processing pipelines. Additionally, neural fields are modular and can be combined with other neural networks. Neural fields can be easily incorporated into current deep learning pipelines. This makes them well-suited for data-driven applications. Limitations and future directions. The main limitation of our method is the slow optimization speed that hinders interactive editing. Typically, ARAP can solve each of such deformation problems within half an hour for large meshes, and it can achieve interactive speed for meshes with a small number of triangles. The Taubin smoothing can finish in a couple of seconds. Our deformation method right now requires a Titan X GPU with 12GB memory to train for 10 hours. Our smoothing and sharpening method takes about 10 minutes on the same GPU. The current algorithm does not ensure that the output field remains an SDF, which prevents performing multiple editing operations. We believe many of these limitations can be addressed in future work. One can potentially use techniques like progressive training or meta-learning [67, 75] to improve the training speed. To allow consecutive editing, we can design regularization loss that enforces the output fields to approximate an SDF [4] or design sampling methods robust with noisy SDF. Other interesting directions include providing deeper theoretical analysis and applying neural fields in other geometry processing tasks. Societal impact. Our work can lead to more efficient geometry processing algorithms. Such algorithms can empower artists to generate creative content. Potential negative impact includes misuse of geometry processing algorithms to create offensive content. 8 Conclusion Our work provides a proof of concept that geometry processing can be done entirely with neural fields without discretizing the surface. We develop network architectures and training objectives for filtering and deforming shapes represented by neural fields and demonstrate the advantages of using neural fields for geometry processing. We hope that our work can inspire a new generation of geometry processing algorithms using neural fields. Acknowledgement. Guandao s Ph D was supported in part by a research gift from Magic Leap and a donation from NVIDIA. We want to thank Wenqi Xian, Professor Steve Marschner, and members of Intel Labs for providing insightful feedback for this project. [1] M. Alexa, D. Cohen-Or, and D. Levin. As-rigid-as-possible shape interpolation. Proceedings of the 27th annual conference on Computer graphics and interactive techniques, 2000. 5, 7 [2] Pierre Alliez, Mark Meyer, and Mathieu Desbrun. Interactive geometry remeshing. ACM Transactions on Graphics (TOG), 21(3):347 354, 2002. 1 [3] Pierre Alliez, David Cohen-Steiner, Olivier Devillers, Bruno Lévy, and Mathieu Desbrun. Anisotropic polygonal remeshing. ACM SIGGRAPH 2003 Papers, 2003. 1 [4] Matan Atzmon and Y. Lipman. Sal: Sign agnostic learning of shapes from raw data. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 2562 2571, 2020. 3, 4, 10 [5] Matan Atzmon, Niv Haim, Lior Yariv, Ofer Israelov, Haggai Maron, and Y. Lipman. Controlling neural level sets. In Neur IPS, 2019. 2, 3, 5 [6] D. Bechmann. Space deformation models survey. Comput. Graph., 18:571 586, 1994. 2 [7] Jens Behrmann, D. Duvenaud, and J. Jacobsen. Invertible residual networks. In ICML, 2019. 2, [8] J. Bloomenthal and B. Wyvill. Introduction to implicit surfaces. 1997. 1 [9] Mario Botsch and O. Sorkine-Hornung. On linear variational surface deformation methods. IEEE Transactions on Visualization and Computer Graphics, 14:213 230, 2008. 2, 3 [10] Mario Botsch, Mark Pauly, Markus H. Gross, and Leif P. Kobbelt. Primo: coupled prisms for intuitive surface modeling. In SGP 06, 2006. 2 [11] Mario Botsch, Leif Kobbelt, Mark Pauly, Pierre Alliez, and Bruno Lévy. Polygon mesh processing. CRC press, 2010. 1, 2 [12] Aljavz Bovzivc, Pablo Rodríguez Palafox, Michael Zollhofer, Justus Thies, Angela Dai, and Matthias Nießner. Neural deformation graphs for globally-consistent non-rigid reconstruction. Ar Xiv, abs/2012.01451, 2020. 3 [13] Ruojin Cai, Guandao Yang, Hadar Averbuch-Elor, Zekun Hao, Serge Belongie, Noah Snavely, and Bharath Hariharan. Learning gradient fields for shape generation. In Proceedings of the European Conference on Computer Vision (ECCV), 2020. 3, 5 [14] Edwin Catmull. A system for computer generated movies. In Proceedings of the ACM annual conference-Volume 1, pages 422 431, 1972. 1 [15] Isaac Chao, Ulrich Pinkall, Patrick Sanan, and Peter Schröder. A simple geometric model for elastic deformations. In SIGGRAPH 2010, 2010. 2 [16] Hunter Chase. Fundamental forms of surfaces and the gauss-bonnet theorem. 2012. 7 [17] Ricky T. Q. Chen, Jens Behrmann, D. Duvenaud, and J. Jacobsen. Residual flows for invertible generative modeling. Ar Xiv, abs/1906.02735, 2019. 6 [18] Xu Chen, Yufeng Zheng, M. Black, Otmar Hilliges, and Andreas Geiger. Snarf: Differentiable forward skinning for animating non-rigid neural implicit shapes. Ar Xiv, abs/2104.03953, 2021. 3 [19] Zhiqin Chen and Hao Zhang. Learning implicit fields for generative shape modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5939 5948, 2019. 1 [20] Ming Chuang and M. Kazhdan. Interactive and anisotropic geometry processing using the screened poisson equation. ACM SIGGRAPH 2011 papers, 2011. 2, 4 [21] Djork-Arné Clevert, Thomas Unterthiner, and S. Hochreiter. Fast and accurate deep network learning by exponential linear units (elus). ar Xiv: Learning, 2016. 6 [22] B. Curless and M. Levoy. A volumetric method for building complex models from range images. Proceedings of the 23rd annual conference on Computer graphics and interactive techniques, 1996. 1 [23] T. Davies, Derek Nowrouzezahrai, and A. Jacobson. On the effectiveness of weight-encoded neural implicit 3d shapes. 2020. 1 [24] A. de Boer, Martijn van der Schoot, and Hester Bijl. Mesh deformation based on radial basis function interpolation. Computers & Structures, 85:784 795, 2007. 2 [25] Yu Deng, J. Yang, and Xin Tong. Deformed implicit field: Modeling 3d shapes with learned dense correspondence. 2021. 3, 6 [26] Mathieu Desbrun, M. Meyer, P. Schröder, and A. Barr. Implicit fairing of irregular meshes using diffusion and curvature flow. In SIGGRAPH 99, 1999. 2 [27] Michael Garland and Paul S. Heckbert. Surface simplification using quadric error metrics. Proceedings of the 24th annual conference on Computer graphics and interactive techniques, 1997. 4 [28] Michael Garland and Paul S Heckbert. Surface simplification using quadric error metrics. In Proceedings of the 24th annual conference on Computer graphics and interactive techniques, pages 209 216, 1997. 9 [29] Kyle Genova, Forrester Cole, Daniel Vlasic, Aaron Sarna, William T Freeman, and Thomas Funkhouser. Learning shape templates with structured implicit functions. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7154 7164, 2019. 3 [30] Kyle Genova, Forrester Cole, Avneesh Sud, Aaron Sarna, and Thomas Funkhouser. Local deep implicit functions for 3d shape. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4857 4866, 2020. 3 [31] Amos Gropp, Lior Yariv, Niv Haim, Matan Atzmon, and Y. Lipman. Implicit geometric regularization for learning shapes. In ICML, 2020. 2, 3, 4 [32] Kunal Gupta and Manmohan Chandraker. Neural mesh flow: 3d manifold mesh generationvia diffeomorphic flows. Ar Xiv, abs/2007.10973, 2020. 3 [33] Zekun Hao, Hadar Averbuch-Elor, Noah Snavely, and Serge J. Belongie. Dualsdf: Semantic shape manipulation using a two-level representation. pages 7628 7638, 2020. 3 [34] Jingwei Huang, C. Jiang, Baiqiang Leng, Bin Wang, and L. Guibas. Meshode: A robust and scalable framework for mesh deformation. Ar Xiv, abs/2005.11617, 2020. 3 [35] José A. Iglesias, M. Rumpf, and O. Scherzer. Shape-aware matching of implicit surfaces based on thin shell energies. Foundations of Computational Mathematics (New York, N.y.), 18:891 927, 2018. 3, 8 [36] A. Jacobson, I. Baran, Jovan Popovic, and O. Sorkine-Hornung. Bounded biharmonic weights for real-time deformation. Communications of the ACM, 57:99 106, 2014. 2 [37] Timothy Jeruzalski, Boyang Deng, Mohammad Norouzi, J. P. Lewis, G. Hinton, and Andrea Tagliasacchi. Nasa: Neural articulated shape approximation. In ECCV, 2020. 3 [38] Chiyu Max Jiang, Jingwei Huang, Andrea Tagliasacchi, and Leonidas J. Guibas. Shapeflow: Learnable deformation flows among 3d shapes. In Neur IPS, 2020. 3 [39] M. W. Jones, J. A. Bærentzen, and M. Srámek. 3d distance fields: a survey of techniques and applications. IEEE Transactions on Visualization and Computer Graphics, 12:581 599, 2006. 1, 2, 3 [40] Venkat Krishnamurthy and M. Levoy. Fitting smooth surfaces to dense polygon meshes. Proceedings of the 23rd annual conference on Computer graphics and interactive techniques, 1996. 4 [41] Zohar Levi and C. Gotsman. Smooth rotation enhanced as-rigid-as-possible mesh animation. IEEE Transactions on Visualization and Computer Graphics, 21:264 277, 2015. 2, 9, 10 [42] Y. Lipman, O. Sorkine-Hornung, D. Cohen-Or, D. Levin, Christian Rössl, and H. Seidel. Differential coordinates for interactive mesh editing. Proceedings Shape Modeling Applications, 2004., pages 181 190, 2004. 3 [43] Steven Liu, Xiuming Zhang, Zhoutong Zhang, Richard Zhang, Jun-Yan Zhu, and Bryan C. Russell. Editing conditional radiance fields. Ar Xiv, abs/2105.06466, 2021. 3, 6 [44] W. Lorensen and H. Cline. Marching cubes: A high resolution 3d surface construction algorithm. In SIGGRAPH 87, 1987. 4, 8 [45] Lars M. Mescheder, Michael Oechsle, M. Niemeyer, S. Nowozin, and Andreas Geiger. Occupancy networks: Learning 3d reconstruction in function space. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4455 4465, 2019. 1, 8 [46] Mateusz Michalkiewicz, Jhony K Pontes, Dominic Jack, Mahsa Baktashmotlagh, and Anders Eriksson. Deep level sets: Implicit surface representations for 3d shape inference. ar Xiv preprint ar Xiv:1901.06802, 2019. 1 [47] Ben Mildenhall, Pratul P. Srinivasan, Matthew Tancik, J. T. Barron, R. Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In ECCV, 2020. 1, 6 [48] Tim Milliron, R. Jensen, Ronen Barzel, and A. Finkelstein. A framework for geometric warps and deformations. ACM Trans. Graph., 21:20 51, 2002. 2 [49] I. Mitchell. The flexible, extensible and efficient toolbox of level set methods. Journal of Scientific Computing, 35:300 329, 2008. 2 [50] Henry P. Moreton and C. Séquin. Functional optimization for fair surface design. Proceedings of the 19th annual conference on Computer graphics and interactive techniques, 1992. 2 [51] K. Museth, D. Breen, R. Whitaker, and A. Barr. Level set surface editing operators. Proceedings of the 29th annual conference on Computer graphics and interactive techniques, 2002. 1, 3 [52] Martin E Newell. The utilization of procedure models in digital image synthesis. Technical report, UTAH UNIV SALT LAKE CITY SCHOOL OF COMPUTING, 1975. 1 [53] M. Niemeyer, Lars M. Mescheder, Michael Oechsle, and Andreas Geiger. Occupancy flow: 4d reconstruction by learning particle dynamics. 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pages 5378 5388, 2019. 3 [54] S. Osher and Ronald Fedkiw. Level set methods and dynamic implicit surfaces. In Applied mathematical sciences, 2003. 1, 2, 3, 7 [55] S. Osher and N. Paragios. Geometric level set methods in imaging, vision, and graphics. In Springer New York, 2003. 3 [56] Jeong Joon Park, Peter R. Florence, J. Straub, Richard A. Newcombe, and S. Lovegrove. Deepsdf: Learning continuous signed distance functions for shape representation. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 165 174, 2019. 1, 4, 8 [57] Keunhong Park, Utkarsh Sinha, Jonathan T. Barron, Sofien Bouaziz, Dan B Goldman, Steven M. Seitz, and Ricardo Martin-Brualla. Deformable neural radiance fields. ar Xiv preprint ar Xiv:2011.12948, 2020. 3, 6 [58] Sida Peng, Junting Dong, Qianqian Wang, Shangzhan Zhang, Qing Shuai, Hujun Bao, and Xiaowei Zhou. Animatable neural radiance fields for human body modeling. Ar Xiv, abs/2105.02872, 2021. 3 [59] H. Piaggio. Differential geometry of curves and surfaces. Nature, 169:560 560, 1952. 3, 7 [60] Albert Pumarola, Enric Corona, Gerard Pons-Moll, and F. Moreno-Noguer. D-nerf: Neural radiance fields for dynamic scenes. Ar Xiv, abs/2011.13961, 2020. 3 [61] Nasim Rahaman, Aristide Baratin, Devansh Arpit, Felix Draxler, Min Lin, Fred Hamprecht, Yoshua Bengio, and Aaron Courville. On the spectral bias of neural networks. In International Conference on Machine Learning, pages 5301 5310. PMLR, 2019. 2 [62] Edoardo Remelli, Artem Lukoianov, Stephan R. Richter, Benoit Guillard, Timur M. Bagautdinov, Pierre Baqué, and P. Fua. Meshsdf: Differentiable iso-surface extraction. Neur IPS, 2020. 3 [63] A. Rosenfeld and J. Pfaltz. Sequential operations in digital picture processing. J. ACM, 13: 471 494, 1966. 1 [64] S. Saito, Zeng Huang, R. Natsume, S. Morishima, Angjoo Kanazawa, and Hao Li. Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pages 2304 2314, 2019. 1 [65] T. Sederberg and S. Parry. Free-form deformation of solid geometric models. Proceedings of the 13th annual conference on Computer graphics and interactive techniques, 1986. 2 [66] V. Sitzmann, Julien N. P. Martel, Alexander W. Bergman, David B. Lindell, and G. Wetzstein. Implicit neural representations with periodic activation functions. Ar Xiv, abs/2006.09661, 2020. 2, 4, 6, 8 [67] Vincent Sitzmann, Eric Chan, Richard Tucker, Noah Snavely, and Gordon Wetzstein. Metasdf: Meta-learning signed distance functions. Ar Xiv, abs/2006.09662, 2020. 10 [68] Yang Song and S. Ermon. Generative modeling by estimating gradients of the data distribution. Ar Xiv, abs/1907.05600, 2019. 5 [69] Olga Sorkine. Differential representations for mesh processing. In Computer Graphics Forum, volume 25, pages 789 807. Wiley Online Library, 2006. 2 [70] O. Sorkine-Hornung and M. Alexa. As-rigid-as-possible surface modeling. In Symposium on Geometry Processing, 2007. 2, 5, 7, 8 [71] O. Sorkine-Hornung, D. Cohen-Or, Y. Lipman, M. Alexa, Christian Rössl, and H. Seidel. Laplacian surface editing. In SGP 04, 2004. 2 [72] O. Stein, E. Grinspun, M. Wardetzky, and A. Jacobson. Natural boundary conditions for smoothing in geometry processing. ACM Transactions on Graphics (TOG), 37:1 13, 2018. 2 [73] Towaki Takikawa, Joey Litalien, K. Yin, K. Kreis, Charles T. Loop, Derek Nowrouzezahrai, A. Jacobson, M. Mc Guire, and S. Fidler. Neural geometric level of detail: Real-time rendering with implicit 3d shapes. Ar Xiv, abs/2101.10994, 2021. 1 [74] Matthew Tancik, Pratul P. Srinivasan, Ben Mildenhall, Sara Fridovich-Keil, Nithin Raghavan, Utkarsh Singhal, R. Ramamoorthi, J. T. Barron, and Ren Ng. Fourier features let networks learn high frequency functions in low dimensional domains. Ar Xiv, abs/2006.10739, 2020. 2, 6 [75] Matthew Tancik, Ben Mildenhall, Terrance Wang, Divi Schmidt, Pratul P. Srinivasan, Jonathan T. Barron, and Ren Ng. Learned initializations for optimizing coordinate-based neural representations. In CVPR, 2021. 10 [76] G. Taubin. A signal processing approach to fair surface design. Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, 1995. 2, 4 [77] Demetri Terzopoulos, John C. Platt, A. Barr, and K. Fleischer. Elastically deformable models. In SIGGRAPH 87, 1987. 2 [78] Jason Torchinsky. The first real object ever 3D scanned and rendered was a VW beetle. Jalopnik, 2013. https://jalopnik.com/ the-first-real-object-ever-3d-scanned-and-rendered-was-494241353. 1 [79] B. Vallet and B. Lévy. Spectral geometry processing with manifold harmonics. Computer Graphics Forum, 27, 2008. 2 [80] Weiyue Wang, D. Ceylan, R. Mech, and U. Neumann. 3dn: 3d deformation network. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 1038 1046, 2019. 3 [81] Yifan Wang, Noam Aigerman, Vladimir G. Kim, S. Chaudhuri, and O. Sorkine-Hornung. Neural cages for detail-preserving 3d deformations. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 72 80, 2020. 3 [82] William Welch and A. Witkin. Variational surface modeling. Proceedings of the 19th annual conference on Computer graphics and interactive techniques, 1992. 2 [83] Zhan Xu, Yang Zhou, E. Kalogerakis, Chris Landreth, and Karan Singh. Rignet: Neural rigging for articulated characters. Ar Xiv, abs/2005.00559, 2020. 3 [84] Lior Yariv, Yoni Kasten, Dror Moran, Meirav Galun, Matan Atzmon, Basri Ronen, and Yaron Lipman. Multiview neural surface reconstruction by disentangling geometry and appearance. Advances in Neural Information Processing Systems, 33, 2020. 1 [85] Zerong Zheng, Tao Yu, Qionghai Dai, and Y. Liu. Deep implicit templates for 3d shape representation. Ar Xiv, abs/2011.14565, 2020. 3 [86] Qian-Yi Zhou, Jaesik Park, and Vladlen Koltun. Open3D: A modern library for 3D data processing. ar Xiv:1801.09847, 2018. 8