# metasdf_metalearning_signed_distance_functions__8c07170c.pdf Meta SDF: Meta-learning Signed Distance Functions Vincent Sitzmann Stanford University sitzmann@cs.stanford.edu Eric R. Chan Stanford University erchan@stanford.edu Richard Tucker Google Research richardt@google.com Noah Snavely Google Research snavely@google.com Gordon Wetzstein Stanford University gordon.wetzstein@stanford.edu vsitzmann.github.io/metasdf/ Neural implicit shape representations are an emerging paradigm that offers many potential benefits over conventional discrete representations, including memory efficiency at a high spatial resolution. Generalizing across shapes with such neural implicit representations amounts to learning priors over the respective function space and enables geometry reconstruction from partial or noisy observations. Existing generalization methods rely on conditioning a neural network on a lowdimensional latent code that is either regressed by an encoder or jointly optimized in the auto-decoder framework. Here, we formalize learning of a shape space as a metalearning problem and leverage gradient-based meta-learning algorithms to solve this task. We demonstrate that this approach performs on par with auto-decoder based approaches while being an order of magnitude faster at test-time inference. We further demonstrate that the proposed gradient-based method outperforms encoder-decoder based methods that leverage pooling-based set encoders. 1 Introduction Humans possess an impressive intuition for 3D shapes; given partial observations of an object we can easily imagine the shape of the complete object. Computer vision and machine learning researchers have long sought to reproduce this ability with algorithms. An emerging class of neural implicit shape representations, for example using signed distance functions parameterized by neural networks, promises to achieve these abilities by learning priors over neural implicit shape spaces [1, 2]. In such methods, each shape is represented by a function Φ, e.g., a signed distance function. Generalizing across a set of shapes thus amounts to learning a prior over the space of these functions Φ. Two questions arise: (1) How do we parameterize functions Φ, and (2) how do we infer the parameters of such a Φ given a set of (partial) observations? Existing methods assume that the space of functions Φ is low-dimensional and represent each shape as a latent code, which parameterizes the full function Φ via concatenation-based conditioning or hypernetworks. These latent codes are either directly inferred by a 2D or 3D convolutional encoder, taking either a single image or a classic volumetric representation as input, or via the auto-decoder framework, where separate latent codes per training sample are treated as free variables at training time. Convolutional encoders, while fast, require observations on a regular grid, and are not equivariant to 3D transformations [3]. They further do not offer a straightforward way to accumulate information from a variable number of observations, usually resorting to permutation-invariant pooling of perobservation latent codes [4]. Recently proposed pooling-based set encoders may encode sets of These authors contributed equally to this work. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. Specialized SDFs Meta-SDF Few GD steps Few GD steps Few GD steps Few GD steps Figure 1: We propose to leverage gradient-based meta-learning to learn the initialization of an implicit neural representation such that it can be adapted to any specific instance in few gradient descent steps. variable cardinality [5, 6], but have been found to underfit the context. This is corroborated by theoretical evidence, showing that to guarantee universality of the embedded function, the embedding requires a dimensionality of at least the number of context points [7]. In practice, most approaches thus leverage the auto-decoder framework for generalization, which is agnostic to the number of observations and does not require observations on a regular grid. This has yielded impressive results on few-shot reconstruction of geometry, appearance and semantic properties [8, 9]. To infer the parameters of a single Φ from a set of observations at test time, encoder-based methods only require a forward pass. In contrast, the auto-decoder framework does not learn to infer an embedding from observations, and instead requires solving an optimization problem to find a lowdimensional latent embedding at test time, which may take several seconds even for simple scenes, such as single 3D objects from the Shape Net dataset. In this work, we identify a key connection between learning of neural implicit function spaces and meta-learning. We then propose to leverage recently proposed gradient-based meta-learning algorithms for the learning of shape spaces, with benefits over both encoder and auto-decoder based approaches. Specifically, the proposed approach performs on par with auto-decoder based models while being an order of magnitude faster at inference time, does not require observations on a regular grid, naturally interfaces with a variable number of observations, outperforms pooling-based setencoder approaches, and does not require the assumption of a low-dimensional latent space. We note that we do not propose any improvements to gradient-based meta-learning algorithms in general. 2 Related Work Neural Implicit Scene Representations. Implicit parameterizations of scenes are an emerging topic of interest in the machine learning community. Parameterized as multilayer perceptrons, these continuous representations have applications in modeling shape parts [10, 11], objects [1, 12 14], or scenes [8, 15, 16]. These and related neural implicits are typically trained from 3D data [1, 2, 12 19], but more recent work has also shown how 2D image data can be directly used to supervise the training procedure [8, 20 22], leveraging differentiable neural rendering [23]. Earlier work on compositional pattern producing networks explored similar strategies to parameterize 2D images [24, 25]. Learning shape spaces. A large body of prior work has explored learning priors over the parameters of classic geometry representations, such as meshes, voxel grids, or point clouds [5, 26 28]. Learning a prior over neural implicit representations of geometry, however, requires learning a prior over a space of functions. To this end, existing work assumes a low-dimensional latent shape space, and leverages auto-decoders [1, 8] or convolutional encoders [2, 19, 21] to regress a latent embedding. While generally, a single latent code represents the full object, making them a function of spatial coordinates enables capturing of local features [18, 19]. This embedding is then decoded into a function either via concatenation-based conditioning [1, 2, 19] or via hypernetworks [8, 29]. In this work, we instead propose a meta-learning-based approach to learning a shape of signed distance functions. Meta-Learning. Meta-learning is usually formalized in the context of few-shot learning. Here, the goal is to train a learner that can quickly adapt to new, unseen tasks given only few training examples, often referred to as context observations. One class of meta-learners proposes to learn an optimizer or update rule [30 32]. Conditional and attentive neural processes [4, 6, 33] instead encode context observations into a low-dimensional embedding via a permutation-invariant set encoder, and a decoder network is conditioned on the resulting latent embedding. A recent class of algorithms proposes to learn the initialization of a neural network, which is then specialized to a new task via few steps of gradient descent [34, 35]. This class of algorithms has recently seen considerable interest, resulting in a wide variety of extensions and improvements. Rusu et al. [36] blend feed-forward and gradient-descent based specialization via optimization in a latent space. Rajeswaran et al. [37] propose to obtain gradients via an implicit method instead of backpropagating through unrolled iterations of the inner loop, leading to significant memory savings. In this work, we leverage this class of gradient-descent based meta learners to tackle the learning of a shape space. We refer the reader to a recent survey paper for an exhaustive overview [38]. 3 Meta-learning Signed Distance Functions We are interested in learning a prior over implicit shape representations. As in [1], we consider a dataset D of N shapes. Each shape is itself represented by a set of points Xi consisting of K point samples from its ground-truth signed distance function (SDF) SDFi: D = {Xi}N i=1, Xi = {(xj, sj) : sj = SDFi(xj)}K j=1 (1) Where xj are spatial coordinates, and sj are the signed distances at these spatial coordinates. We aim to represent a shape by directly approximating its SDF with a neural network Φi [1], which represents the surface of the shape implicitly as its zero-level set L0(Φi): L0(Φi) = {x R3|Φi(x) = 0}, Φi : R3 7 R (2) We may alternatively choose to approximate the binary occupancy of points rather than their signed distances; the surface would then be represented by the binary classifier s decision boundary [2]. Generalizing across a set of shapes thus amounts to learning a prior over the space of functions X where Φi X. Two questions arise: (1) How do we parameterize a single element Φi of the set X, and (2) how do we infer the parameters of such a Φi given a set of (partial) observations Xi? Parameterizing Φi: Conditioning via concatenation and hypernetworks. Existing approaches to generalizing over shape spaces rely on decoding latent embeddings of shapes. In conditioning via concatenation, the target coordinates x are concatenated with the latent shape embedding zi and are fed to a feedforward neural network whose parameters are shared across all shape instances. The latent code conditions the output of the shared network and allows a single shared network to represent multiple shape instances [1, 2]. In an alternative formulation, we can instead use a hypernetwork, which takes as input the latent code zi and generates the parameters of a shape-representing MLP, Φi. Φi can then be sampled at coordinates x to produce output signed distances. As we demonstrate in the supplement, conditioning via concatenation is a special case of a hypernetwork [29], where the hypernetwork is parameterized as a single linear layer that predicts only the biases of the network Φi. Indeed, recent related work has demonstrated that hypernetworks mapping a latent code zi to all parameters of the MLP Φi can similarly be used to learn a space of shapes [8]. Inferring parameters of Φi from observations: Encoders and Auto-decoders. Existing approaches to inferring latent shape embeddings zi rely on encoders or auto-decoders. The former rely on an encoder, such as a convolutional or set encoder, to generate a latent embedding that is then decoded into an implicit function using conditioning via concatenation or hypernetworks. In auto-decoder (i.e., decoder-only) architectures, the latent embeddings are instead treated as learned parameters rather than inferred from observations at training time. At training time, the auto-decoder framework enables straightforward generalization across a set of shapes. At test time, we freeze the weights of the model and perform a search to find a latent code ˆz that is compliant with a set of context observations ˆX. A major weakness of the auto-decoder framework is that solving this optimization problem is slow, taking several seconds per object. 3.1 Shape generalization as meta-learning Meta-learning aims to learn a model that can be quickly adapted to new tasks, possibly with limited training data. When adapting to a new task, the model is given context observations, with which Algorithm 1 Meta SDF: Gradient-based meta-learning of shape spaces Precondition: Distribution D over SDF samples, outer learning rate β, number of inner-loop steps k 1: Initialize inner, per-parameter learning rates α, meta-parameters θ 2: while not done do 3: Sample batch of shape datasets Xi D 4: for all Xi do 5: Split Xi into Xtrain i , Xtest i 6: Initialize φ0 i = θ, Ltrain = 0 7: for j = 0 to k do 8: Ltrain 1 |Xtrain i | P (x,s) Xtrain i ℓ1 Φ(x; φj i), s 9: φj+1 i φj i α φj i Ltrain 10: Ltest Ltest + 1 |Xtest i | P (x,s) Xtest i ℓ1 Φ(x; φk i ), s 11: θ, α (θ, α) β (θ,α)Ltest return θ, α it can modify itself, e.g. through gradient descent. The adapted model can then be used to make predictions on unseen target observations. Formally, supervised meta-learning assumes a distribution over tasks T p(T ), where each task is of the form T = (L, {(xl, yl)}l C, {(xj, yj)}j T), with loss function L, model inputs x, model outputs y, the set of all indices belonging to a context set C, and the set of all indices belonging to a target set T. At training time, a batch of tasks Ti is drawn, with each task split into context and target observations. The model adapts itself using the context observations and makes predictions on the target observations. The model parameters are optimized to minimize the losses L on all target observations in the training set, over all tasks. Our key idea is to view the learning of a shape space as a meta-learning problem. In this framework, a task Ti represents the problem of finding the signed distance function Φi of a shape. Simply put, by giving a model a limited number of context observations, each of which consists of a world coordinate location x and a ground truth signed-distance s, we aim to quickly specialize it to approximating the underlying SDFi. Context and target observations are samples from an objectspecific dataset Xi: Ti = ({(xl, sl)}l C, {(xj, sj)}j T). Because each task consists of fitting a signed distance function, we can choose a global loss, such as ℓ1. Note that the auto-decoder framework could be viewed as a meta-learning algorithm as well, though it is not generally discussed as such. In this view, the auto-decoder framework is an outlier, as it does not perform model specialization in the forward pass. Instead, specialization requires stochastic gradient descent until convergence, both at training and at test time. Learning a shape space with gradient-based meta-learning. We propose to leverage the family of MAML-like algorithms [34]. We consequently view an instance-specific SDFi, approximated by Φi, as the specialization of an underlying meta-network with parameters θ. In the forward pass, we sample a batch of shape datasets Xi, and split each dataset into a training set Xtrain i and test set Xtest i . We then perform k gradient descent steps according to the following update rule: φj+1 i = φj i λ X (x,s) Xtrain i L(Φ(x; φj i), s), φ0 i = θ (3) where φj i are the parameters for shape i at inner-loop step j, and Φ( ; φj i) indicates that the network Φ is evaluated with these parameters. The final specialized parameters φk i are now used to make predictions on the test set, Xtest i , and to compute an outer-loop loss Ltest. Finally, we backpropagate the loss Ltest through the inner-loop update steps to the parameters θ. We further found that the proposed method benefited from the added flexibility of per-parameter learning rates, as proposed by Li et al. [39]. The full algorithm is formalized in Algorithm 1. This formulation has several advantages over the auto-decoder framework. First, as we demonstrate in Section 4, inference at test time is an order of magnitude faster, while performing on par or slightly better both qualitatively and quantitatively. Further, as MAML optimizes the inference algorithm as well, it can be trained to infer the specialized network from different kinds of context observations. We will demonstrate that this enables, for instance, reconstruction of an SDF given only points on the zero-level set, while the auto-decoder framework requires heuristics in the form of surrogate GT Hyper Net Concat. Cond. NP Meta SDF Figure 3: Reconstruction of SDFs of two test digits from classes unseen at training time. GT Hyper Net Concat. Cond. NP Meta SDF Figure 4: Reconstruction of SDFs of two rotated test digits. GT Hyper Net Concat. Cond. NP Meta SDF Figure 5: Reconstruction of SDFs of two examples with three digits. Unseen Digit Rotation Composit. Hyper Net 0.144 0.189 0.284 Concat 0.062 0.151 0.137 Cond. NP 0.253 0.316 0.504 Meta SDF 0.023 0.032 0.059 Table 1: ℓ1-error for reconstruction of out-ofdistribution samples. All results are 10 3. GT Hyper Net Concat Cond. NP Meta SDF Zero-lvl set Dense Zero-level set Hyper Net 0.044 1.562 Concat 0.050 1.813 Cond. NP 0.108 0.156 Meta SDF 0.015 0.057 Figure 6: Qualitative and quantitative comparison of performance of different generalization strategies in inferring the signed distance function of an MNIST digit either from dense observations (top row), or only points on the zero-level set (bottom row). All quantitative results are ℓ1-error 10 3. losses to achieve this goal. We empirically show that we outperform pooling-based set-encoder based methods, consistent with recent work that found these encoders to underfit the context [33]. Finally, both auto-decoder and auto-encoder-based methods assume a low-dimensional latent space, while MAML naturally optimizes in the high-dimensional space of all parameters θ of the meta-network. In this section, we first apply the proposed Meta SDF approach to the learning of 2D SDFs extracted from MNIST digits, and subsequently, on 3D shapes from the Shape Net dataset. All code and datasets will be made publicly available. 4.1 Meta-learning 2D Signed Distance Functions Figure 2: MNIST digit with extracted SDF, bold zero-level set. We study properties of different generalization methods on 2D signed distance functions (SDFs) extracted from the MNIST dataset. From every MNIST digit, we extract a 2D SDF via a distance transform, such that the contour of the digit is the zero-level set of the corresponding SDF, see Fig. 2. Following [1], we directly fit the SDF of the MNIST digit via a fully connected neural network. We benchmark three alternative generalization approaches. First, two auto-decoder based approaches, where the latent code is decoded into a function Φ either using conditioning via concatenation as in [1] or via a fully connected hypernetwork as in [8]. Second, we compare to a conditional neural process (CNP) [6], representative of permutation-invariant set-encoders. Implementation. All models are implemented as fully connected Re LU-MLPs with 256 hidden units and no normalization layers. Φ is implemented with four layers. The set encoder of CNPs similarly uses four layers. Hypernetworks are implemented with three layers as in [8]. The proposed approach performs 5 inner-loop update steps, where we initialize α as 1 10 1. All models are optimized using the ADAM optimizer [40] with a learning rate of 1 10 4. Inference with partial observations. We demonstrate that we may train the Meta SDF to infer continuous SDFs either from dense ground-truth samples, or from samples from only the zero-level set. This is noteworthy, as inferring an SDF from only the zero-level set points amounts to solving a boundary value problem defined by a particular Eikonal equation, see [14]. In Deep SDF [1], the authors also demonstrate test-time reconstruction from zero-level set points, but require augmenting the loss with heuristics to ensure the problem is well posed. Gropp et al. [14] instead propose to explicitly account for the Eikonal constraint in the loss. We train all models on SDFs of the full MNIST training set, providing supervision via a regular grid of 64 64 ground-truth SDF samples. For CNPs and the proposed approach, we train two models each, conditioned on either (1) the same 64 64 ground-truth SDF samples or (2) a set of 512 points sampled from the zero-level set. We then test all models to reconstruct SDFs from the unseen MNIST test set from these two different kinds of context. Fig. 6 shows a qualitative and quantitative comparison of results. All approaches qualitatively succeed in reconstructing SDFs with a dense context set. Quantitatively, Meta SDFs perform on par with auto-decoder based methods for conditioning on dense SDF values. Both methods outperform CNPs by an order of magnitude. This is consistent with prior work showing that pooling-based set encoders tend to underfit the context set. When conditioned on the zero-level set only, auto-decoder based methods fail to reconstruct test SDFs. In contrast, both CNPs and Meta SDFs succeed in reconstructing the SDF, with the proposed approach again significantly outperforming CNPs. Inference speed. We compare the time required to fit a single unseen 2D MNIST SDF for both hypernetworks and the concatenation-based approach. Even for this simple example, both methods require on the order of 4 seconds to converge on the latent code of an unseen SDF at test time. In contrast, the proposed meta-learning based approach infers functional representations in 5 gradient descent steps, or in about 50 ms. Figure 7: T-SNE embedding of parameters of specialized networks. Out-of-distribution generalization. We investigate the capability of different optimization methods to reconstruct out-of-distribution test samples. We consider three modes of out-of-distribution generalization: First, generalization to MNIST digits unobserved at training time. Here, we train on digits 0 5, holding out 6 9. Second, generalization to randomly rotated MNIST SDFs, where we train on the full MNIST training set. And last, generalization to signed distance functions extracted from the triple-MNIST dataset [41], a dataset of compositions of three scaled-down digits into a single image. In this last experiment, we train on a dataset comprised of both the full MNIST training set, as well as the double MNIST dataset, testing on the triple-MNIST dataset. The double-MNIST dataset contains digits of the same size as the target triple-MNIST dataset. Figures 3, 4, and 5 show qualitative results, while Table 1 reports quantitative performance. The proposed Meta SDF approach far outperforms all alternative approaches in this task. Qualitatively, it succeeds in generating both unseen digits and rotated digits, although no method succeeds to accurately reconstruct the zero-level set in this compositionality experiment. This suggests that the proposed approach is more flexible to out-of-distribution samples. Interpretation as representation learning. It has previously been observed that learning implicit representations can be seen as a form of representation learning [9]. Here, we corroborate this observation, and demonstrate that weights of specialized SDF networks found with the proposed metalearning approach encode information about the digit class, performing unsupervised classification. Fig. 7 shows T-SNE [42] embeddings of the parameters of 10, 000 test-set MNIST SDFs. While classes are not perfectly linearly separable, it is apparent that the parameters of the SDF-encoding neural network Φ carry information about the class of the MNIST digit. 4.2 Meta-Learning a 3D Shape Space We now demonstrate that the proposed approach scales to three dimensions, where results on 3D shape representation are consistent with the 2D MNIST results above. GT Deep SDF Meta SDF GT Meta SDF Figure 8: Top row: comparison of reconstructed shapes of Deep SDF and Meta SDF on the Shape Net planes class. Both models receive dense signed distance values at training and test time. The two methods perform approximately on par, with the proposed method performing slightly better on out-of-distribution samples while speeding up inference by one order of magnitude. Bottom row: Additional reconstructions of Meta SDF from the Shape Net tables class. Meta SDF ONet Input Figure 9: Comparison of reconstructed shapes of proposed method and a Point Net-encoder based method as proposed in [2]. Both models are conditioned only on zero-level set points at training and test time. The proposed method generally outperforms the Point Net-based approach, with strongest gains in uncommon airplanes. We first benchmark the proposed approach in the case of a densely sampled SDF. To this end, we follow the experimental setup of Park et al. [1], and benchmark with their Deep SDF model. Deep SDF combines the auto-decoder framework with conditioning via concatenation to learn a prior over 3D shapes in the Shape Net [43] dataset. The dataset contains ground-truth samples from the analytically computed SDFs of the respective mesh. The application of the proposed framework to this dataset follows the 2D examples in Sec. 4.1, with two minor differences. Park et al. [1] propose to clamp the ℓ1 loss to increase accuracy of the SDF close to the surface. We empirically found that clamping in the inner loop of a gradient-based meta-learning algorithm leads to unstable training. Instead, we found that a multitask loss function that tracks both SDF values as well as occupancy, i.e. the sign of the signed distance function, provides stable training, increases the level of detail of the reconstructed shapes for all models, and provides faster convergence. Instead of predicting a single signed distance output, we predict two values: the distance to the surface and the sign of the distance function (inside or outside the object). The ℓ1 SDF loss is then combined with a binary cross-entropy sign loss with the uncertainty loss weighting scheme described by Kendall et al. [44]. This strategy can be interpreted as combining the loss terms of the concurrently proposed Deep SDF and Occupancy Networks [1, 45]. Tables 2 and 3 report mean and median Chamfer distance as well as MIOU of reconstructed meshes Table 2: Mean / median / std Chamfer Distance (CD) for Shape Net V2. All results are 10 3. Full context Levelset only Deep SDF Meta SDF Point Net Meta SDF Lvl. Planes 0.07 / 0.03 / 0.21 0.05 / 0.02 / 0.13 0.55 / 0.22 / 1.07 0.13 / 0.07 / 0.19 Tables 0.31 / 0.04 / 1.48 0.13 / 0.06 / 0.24 0.69 / 0.21 / 1.21 0.32 / 0.09 / 0.95 Benches 0.31 / 0.07 / 1.70 0.30 / 0.09 / 1.38 0.74 / 0.25 / 1.45 0.31 / 0.11 / 0.92 Table 3: Mean / median / std MIOU evaluation for Shape Net V2. Full context Levelset only Deep SDF Meta SDF Point Net Meta SDF Lvl. Planes 0.85 / 0.87 / 0.08 0.86 / 0.89 / 0.09 0.67 / 0.68 / 0.20 0.76 / 0.80 / 0.13 Tables 0.84 / 0.87 / 0.13 0.85 / 0.88 / 0.14 0.66 / 0.69 / 0.22 0.75 / 0.78 / 0.18 Benches 0.76 / 0.78 / 0.15 0.74 / 0.79 / 0.18 0.56 / 0.60 / 0.22 0.68 / 0.72 / 0.17 on the test set, while Fig. 8 displays corresponding reconstructed shapes. As in the 2D experiments, the proposed approach generally performs on par with the baseline auto-decoder approach. However, reconstruction of a 3D shape at test time is significantly faster, improving by more than one order of magnitude from 8 seconds to 0.4 seconds. To evaluate the impact of our proposed composite loss, we compare Deep SDF trained using our composite loss function against the same architecture trained using the original ℓ1 loss with clamping. As opposed to the proposed approach, where the composite loss is critical to stable training, Deep SDF profits little, with mean and median Chamfer Distances on Shape Net planes at 6.7 10 5 and 3.3 10 5, respectively. Next, as in our 2D experiments, we demonstrate that the same Meta SDF model with no changes to architecture or loss function can learn to accurately reconstruct 3D geometry conditioned only on zero-level set points, i.e., a point cloud. We compare to a baseline with a Point Net [5] encoder and conditioning via concatenation, similar to the model proposed in Occupancy Networks [45]. We match the parameter counts of both models. Tables 2 and 3 report the Chamfer distance and MIOU of test-set reconstructions for this experiment, while Fig. 9 shows qualitative results. Consistent with our results on 2D SDFs, the proposed Meta SDF approach outperforms the set-encoder based approach. Again, performance differences are most significant with out-of-distribution shapes. We note that auto-decoder-based models cannot perform zero-level set reconstructions without the aid of additional heuristics. Meta SDF Point Net Planes 0.63 / 0.28 / 0.95 0.76 / 0.28 / 1.59 Benches 1.68 / 0.72 / 4.69 1.74 / 0.63 / 3.91 Figure 10: Qualitative and quantitative results of SDF reconstruction from depth maps. All quantitative results are Chamfer Distance 10 3. Depth map completion. We demonstrate that the prior learned by Meta SDF enables reconstruction from partial observations in the form of a depth map. We render out depth maps of Shape Net models and transform them into world coordinates using ground-truth camera parameters. We then reconstruct the full SDF using the proposed Meta SDF model trained on level-set points, without any further finetuning. Figure 10 shows quantitative and qualitative results on objects from the Shapenet benches and airplane classes, compared with the Point Net-encoder based approach. We note that autodecoder based approaches such as Deep SDF require additional heuristics to solve this problem, but may produce compelling results when these heuristics are designed appropriately. Visualizing inner loop steps. To gain additional insight into the process of the inner-loop optimization, in Fig. 11, we visualize the evolution of the signed distance function over inner loop iterations, beginning with the unspecialized meta-network. The underlying model was trained for level-set Initialization Step 1 Step 2 Step 3 Step 4 Step 5 Figure 11: Evaluation of the SDF level sets throughout the inner-loop iterations. Red contours are level sets in increments of 0.05. inference on the Shape Net planes class. Intriguingly, even though this model is class-specific, the initialization does not resemble an airplane, and only encodes valid signed distances starting from the second iteration. At initialization and after the first update step, function values increase significantly faster than those of the ground-truth signed distance function, visualized by the collapse of the level sets onto the surface. In contrast, the mean zero-vector of the auto-decoder based Deep SDF model encodes an object that resembles an average plane (see supplemental material). Future investigation of this difference may reveal additional insight into how gradient-based meta learning algorithms succeed in conditioning the optimization problem for fast convergence. Reconstruction of objects from different class. We evaluate the performance of models trained on the Shape Net tables class in reconstructing objects from the Shape Net benches class. Please see the supplemental material for quantitative and qualitative results. In contrast to our 2D results, Deep SDF outperforms Meta SDF in the case of dense conditioning, while Meta SDF continues to outperform the Point Net-encoder based approach in the case of level set conditioning. We hypothesize that in the case of dense conditioning in 3D, the unconstrained number of optimization steps of the auto-decoder based approach offers an advantage over the constant number of optimization steps of Meta SDF. This could potentially be addressed by increasing the number of inner-loop optimization steps of Meta SDF at test time. 5 Discussion In summary, Meta SDF is a meta-learning approach to learning priors over the space of SDFs represented by fully connected neural networks. This approach performs on par with auto-decoder based approaches while being an order of magnitude faster at inference time, outperforms poolingbased set-encoder methods, and makes weaker assumptions about the dimensionality of the latent space. However, several limitations remain. The current implementation requires backpropagation through the unrolled inner-loop gradient descent steps, and thus the computation of second-order gradients, which is highly memory-intensive. This may be addressed by recently proposed implicit gradient methods that offer memory complexity independent of the number of inner-loop optimization steps [37]. More generally, significant progress has recently been made in gradient-based metalearning algorithms. We leverage Meta-SGD [39] as the inner-loop optimizer in this paper, but recent work has realized significant performance gains in few-shot learning via optimization in latent spaces [36] or by learning iterative updates in an infinite-dimensional function space [46], which could further improve Meta SDFs. Another promising line of future investigation are recently proposed attention-based set encoders [33], which have been shown to perform better than set encoders that aggregate latent codes via pooling, though the memory complexity of the attention mechanism makes conditioning on large (5k points) context sets as in our 3D experiments computationally costly. Gradient-based meta-learning may also be a promising approach to learn algorithms that infer latent variables given an image in generative adversarial networks. Finally, future work may apply this approach to representations of scenes, i.e., both 3D geometry and appearance, via neural rendering [8, 22]. Our approach advances the understanding of generalization strategies for emerging neural implicit shape representations by drawing the connection to the meta-learning community. We hope that this approach inspires follow-on work on learning more powerful priors of implicit neural shape representations. Broader Impact Emerging neural implicit representations are a powerful tool for representing signals, such as 3D shape and appearance. Generalizing across these neural implicit representations requires efficient approaches to learning distributions over functions. We have shown that gradient-based meta-learning approaches are one promising avenue to tackling this problem. As a result, the proposed approach may be part of the backbone of this emerging neural signal representation strategy. As a powerful representation of natural signals, such neural implicits may in the future be used for the generation and manipulation of signals, which may pose challenges similar to those posed by generative adversarial models today. Acknowledgments and Disclosure of Funding We would like to offer special thanks to Julien Martel, Matthew Chan, and Trevor Chan for fruitful discussions and assistance in completing this work. Vincent Sitzmann was supported by a Stanford Graduate Fellowship. Gordon Wetzstein was supported by an NSF CAREER Award (IIS 1553333), a Sloan Fellowship, and a PECASE from the ARO. [1] Jeong Joon Park, Peter Florence, Julian Straub, Richard Newcombe, and Steven Lovegrove. Deepsdf: Learning continuous signed distance functions for shape representation. Proc. CVPR, 2019. [2] Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. Occupancy networks: Learning 3d reconstruction in function space. In Proc. CVPR, 2019. [3] Marc Finzi, Samuel Stanton, Pavel Izmailov, and Andrew Gordon Wilson. Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data. ar Xiv preprint ar Xiv:2002.12880, 2020. [4] SM Ali Eslami, Danilo Jimenez Rezende, Frederic Besse, Fabio Viola, Ari S Morcos, Marta Garnelo, Avraham Ruderman, Andrei A Rusu, Ivo Danihelka, Karol Gregor, et al. Neural scene representation and rendering. Science, 360(6394):1204 1210, 2018. [5] Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. Proc. CVPR, 2017. [6] Marta Garnelo, Dan Rosenbaum, Chris J Maddison, Tiago Ramalho, David Saxton, Murray Shanahan, Yee Whye Teh, Danilo J Rezende, and SM Eslami. Conditional neural processes. Proc. ICML, 2018. [7] Edward Wagstaff, Fabian B Fuchs, Martin Engelcke, Ingmar Posner, and Michael Osborne. On the limitations of representing functions on sets. Proc. ICML, 2019. [8] Vincent Sitzmann, Michael Zollhöfer, and Gordon Wetzstein. Scene representation networks: Continuous 3d-structure-aware neural scene representations. In Proc. Neur IPS, 2019. [9] Amit Kohli, Vincent Sitzmann, and Gordon Wetzstein. Semantic implicit neural scene representations with semi-supervised training. Proc. 3DV, 2020. [10] Kyle Genova, Forrester Cole, Daniel Vlasic, Aaron Sarna, William T Freeman, and Thomas Funkhouser. Learning shape templates with structured implicit functions. In Proc. ICCV, pages 7154 7164, 2019. [11] Kyle Genova, Forrester Cole, Avneesh Sud, Aaron Sarna, and Thomas Funkhouser. Deep structured implicit functions. ar Xiv preprint ar Xiv:1912.06126, 2019. [12] Mateusz Michalkiewicz, Jhony K Pontes, Dominic Jack, Mahsa Baktashmotlagh, and Anders Eriksson. Implicit surface representations as layers in neural networks. In Proc. ICCV, pages 4743 4752, 2019. [13] Matan Atzmon and Yaron Lipman. Sal: Sign agnostic learning of shapes from raw data. ar Xiv preprint ar Xiv:1911.10414, 2019. [14] Amos Gropp, Lior Yariv, Niv Haim, Matan Atzmon, and Yaron Lipman. Implicit geometric regularization for learning shapes. ar Xiv preprint ar Xiv:2002.10099, 2020. [15] Chiyu Max Jiang, Avneesh Sud, Ameesh Makadia, Jingwei Huang, Matthias Nießner, and Thomas Funkhouser. Local implicit grid representations for 3d scenes. ar Xiv preprint ar Xiv:2003.08981, 2020. [16] Songyou Peng, Michael Niemeyer, Lars Mescheder, Marc Pollefeys, and Andreas Geiger. Convolutional occupancy networks. ar Xiv preprint ar Xiv:2003.04618, 2020. [17] Zhiqin Chen and Hao Zhang. Learning implicit fields for generative shape modeling. In Proc. CVPR, pages 5939 5948, 2019. [18] Shunsuke Saito, Zeng Huang, Ryota Natsume, Shigeo Morishima, Angjoo Kanazawa, and Hao Li. Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. In Proc. ICCV, pages 2304 2314, 2019. [19] Qiangeng Xu, Weiyue Wang, Duygu Ceylan, Radomir Mech, and Ulrich Neumann. Disn: Deep implicit surface network for high-quality single-view 3d reconstruction. In Proc. Neur IPS, pages 492 502, 2019. [20] Michael Oechsle, Lars Mescheder, Michael Niemeyer, Thilo Strauss, and Andreas Geiger. Texture fields: Learning texture representations in function space. In Proc. ICCV, 2019. [21] Michael Niemeyer, Lars Mescheder, Michael Oechsle, and Andreas Geiger. Differentiable volumetric rendering: Learning implicit 3d representations without 3d supervision. In Proc. CVPR, 2020. [22] Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. Proc. ECCV, 2020. [23] Ayush Tewari, Ohad Fried, Justus Thies, Vincent Sitzmann, Stephen Lombardi, Kalyan Sunkavalli, Ricardo Martin-Brualla, Tomas Simon, Jason Saragih, Matthias Nießner, et al. State of the art on neural rendering. Eurographics, 2020. [24] Kenneth O Stanley. Compositional pattern producing networks: A novel abstraction of development. Genetic programming and evolvable machines, 8(2):131 162, 2007. [25] Alexander Mordvintsev, Nicola Pezzotti, Ludwig Schubert, and Chris Olah. Differentiable image parameterizations. Distill, 3(7):e12, 2018. [26] Daniel Maturana and Sebastian Scherer. Voxnet: A 3d convolutional neural network for real-time object recognition. In Proc. IROS, page 922 928, September 2015. [27] Gernot Riegler, Ali Osman Ulusoy, and Andreas Geiger. Octnet: Learning deep 3d representations at high resolutions. In Proc. CVPR, 2017. [28] Dominic Jack, Jhony K. Pontes, Sridha Sridharan, Clinton Fookes, Sareh Shirazi, Frédéric Maire, and Anders Eriksson. Learning free-form deformations for 3d object reconstruction. Co RR, abs/1803.10932, 2018. [29] David Ha, Andrew Dai, and Quoc V Le. Hypernetworks. Proc. ICLR, 2017. [30] Jürgen Schmidhuber. Evolutionary principles in self-referential learning, or on learning how to learn: the meta-meta-... hook. Ph D thesis, Technische Universität München, 1987. [31] Marcin Andrychowicz, Misha Denil, Sergio Gomez, Matthew W Hoffman, David Pfau, Tom Schaul, Brendan Shillingford, and Nando De Freitas. Learning to learn by gradient descent by gradient descent. In Proc. NIPS, pages 3981 3989, 2016. [32] Sachin Ravi and Hugo Larochelle. Optimization as a model for few-shot learning. In Proc. ICLR, 2016. [33] Hyunjik Kim, Andriy Mnih, Jonathan Schwarz, Marta Garnelo, Ali Eslami, Dan Rosenbaum, Oriol Vinyals, and Yee Whye Teh. Attentive neural processes. Proc. ICLR, 2019. [34] Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. In Proc. ICML, pages 1126 1135. JMLR. org, 2017. [35] Alex Nichol and John Schulman. Reptile: a scalable metalearning algorithm. ar Xiv preprint ar Xiv:1803.02999, 2, 2018. [36] Andrei A Rusu, Dushyant Rao, Jakub Sygnowski, Oriol Vinyals, Razvan Pascanu, Simon Osindero, and Raia Hadsell. Meta-learning with latent embedding optimization. Proc. ICLR, 2019. [37] Aravind Rajeswaran, Chelsea Finn, Sham M Kakade, and Sergey Levine. Meta-learning with implicit gradients. In Proc. NIPS, pages 113 124, 2019. [38] Timothy Hospedales, Antreas Antoniou, Paul Micaelli, and Amos Storkey. Meta-learning in neural networks: A survey. ar Xiv preprint ar Xiv:2004.05439, 2020. [39] Zhenguo Li, Fengwei Zhou, Fei Chen, and Hang Li. Meta-sgd: Learning to learn quickly for few-shot learning. ar Xiv preprint ar Xiv:1707.09835, 2017. [40] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. Proc. ICLR, 2014. [41] Shao-Hua Sun. Multi-digit mnist for few-shot learning, 2019. URL https://github.com/ shaohua0116/Multi Digit MNIST. [42] Laurens van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(Nov):2579 2605, 2008. [43] Angel X Chang, Thomas Funkhouser, Leonidas Guibas, Pat Hanrahan, Qixing Huang, Zimo Li, Silvio Savarese, Manolis Savva, Shuran Song, Hao Su, et al. Shapenet: An information-rich 3d model repository. ar Xiv preprint ar Xiv:1512.03012, 2015. [44] Alex Kendall, Yarin Gal, and Roberto Cipolla. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proc. CVPR, pages 7482 7491, 2018. [45] Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. Occupancy networks: Learning 3d reconstruction in function space. In Proc. CVPR, 2019. [46] Jin Xu, Jean-Francois Ton, Hyunjik Kim, Adam R Kosiorek, and Yee Whye Teh. Metafun: Meta-learning with iterative functional updates. ar Xiv preprint ar Xiv:1912.02738, 2019.