# diffusing_differentiable_representations__c47877a8.pdf Diffusing Differentiable Representations Yash Savani Carnegie Mellon University ysavani@andrew.cmu.edu Marc Finzi Carnegie Mellon University mfinzi@andrew.cmu.edu J. Zico Kolter Carnegie Mellon University zkolter@andrew.cmu.edu We introduce a novel, training-free method for sampling differentiable representations (diffreps) using pretrained diffusion models. Rather than merely modeseeking, our method achieves sampling by pulling back the dynamics of the reverse-time process from the image space to the diffrep parameter space and updating the parameters according to this pulled-back process. We identify an implicit constraint on the samples induced by the diffrep and demonstrate that addressing this constraint significantly improves the consistency and detail of the generated objects. Our method yields diffreps with substantially improved quality and diversity for images, panoramas, and 3D Ne RFs compared to existing techniques. Our approach is a general-purpose method for sampling diffreps, expanding the scope of problems that diffusion models can tackle. 1 Introduction Diffusion models have emerged as a powerful tool for generative modeling [Ho et al., 2020, Song et al., 2022, 2021, Karras et al., 2022, Rombach et al., 2022], and subsequent work has extended these models to generate complex objects such as 3D assets, tiled images, and more. Although such approaches typically require training or at least fine-tuning of the diffusion models for these new modalities [Wang et al., 2023, Luo et al., 2023], two notable exceptions, Wang et al. [2022] and Poole et al. [2022], have developed methods for training-free production of 3D objects by directly using an image-based diffusion model. Both methods work by optimizing a differentiable representation (diffrep) in this case, a Neural Radiance Field (Ne RF) [Mildenhall et al., 2020] to produce rendered views consistent with the output of the image-based diffusion model. Unfortunately, the nature of both these methods is that they optimize the diffrep to produce the most likely representation consistent with the images; that is, they perform mode-seeking rather than actually sampling from the diffusion model. This results in overly smoothed outputs that lack detail and do not reflect the underlying distribution of the diffusion model. In this paper, we present a novel method for sampling directly in the diffrep space using pretrained diffusion models. The method is training-free, can handle arbitrary diffreps, and performs true sampling according to the underlying diffusion model rather than merely mode-seeking. The key idea of our approach is to rewrite the reverse diffusion process itself in the diffrep parameter space. This is achieved by pulling back the dynamics of the reverse time process from the image space to the parameter space and solving a (small) optimization problem, implied by the pulled-back dynamics, over the parameters for each diffusion step. To further encourage solver convergence, we identify constraints that the diffrep induces on the samples and use these constraints to guide the reverse process to generate samples from the high-density regions of the diffusion model while satisfying the constraints along the sampling trajectory. Our experiments use a pretrained image-based diffusion model (Stable Diffusion 1.5 [Rombach et al., 2022]) to generate samples from various classes of diffreps. For example, by sampling SIREN representations [Vincent, 2011] with wrap-around boundary constraints, we can sample total 360degree panorama scenes, and by sampling Ne RF representations [Mildenhall et al., 2020], we can 38th Conference on Neural Information Processing Systems (Neur IPS 2024). generate 3D models of many objects. In both settings as well as in baseline comparisons on simple SIREN-based (non-panorama) image generation our approach substantially outperforms previous methods such as Score Jacobian Chaining (SJC) [Wang et al., 2022]. Though the problem setting is considerably more difficult without fine-tuning or retraining, sampling diffreps using pretrained diffusion models with our method substantially improves and extends the state-of-the-art in training-free generation methods. 2 Related work and preliminaries 2.1 Differentiable representations (diffreps) Diffreps are a powerful tool for representing complex scenes or manifolds using parameterized differentiable functions to map coordinates of the scene into a feature space that encodes the properties of the scene at those coordinates. Popular instantiations of diffreps include: SIRENs [Sitzmann et al., 2020], which implicitly model an image using an MLP with sinusoidal activations to map 2D (x, y) pixel coordinates into RGB colors, and Ne RFs [Mildenhall et al., 2020], which implicitly model a 3D scene using an MLP that transforms 3D (x, y, z) voxel coordinates with view directions (ϕ, ψ) into RGBσ values. We can render images from different views of the scene by numerically integrating the Ne RF outputs along the unprojected rays from the camera. Faster or alternative diffreps also exist for 3D scenes such as the Instant NGP [Müller et al., 2022] model or Gaussian splats [Kerbl et al., 2023] although in this work we focus on the basic Ne RF architecture. Many other kinds of visual assets can also be formulated as diffreps. For example, we can use diffreps to model panoramas, spherical images, texture maps for 3D meshes, compositions of multiple images, scenes from a movie, and even the output of kinematic and fluid simulations. Many interesting diffreps can be used to render not just one image from the scene but multiple coupled images or even a distribution over images. Given a diffrep, parameterized by θ Θ, for a scene such as a SIREN panorama or a Ne RF we can render an image of the scene from a view π Π using a differentiable render function f(θ, π) = image X. To accommodate the multiview setting in our discussion, we consider the curried Haskell form of the render function f : Θ (Π X), where f(θ) : Π X is a map itself from a view π Π to f(θ)(π) = f(θ, π) = image X. In the case of Ne RFs or SIREN panoramas, the view π is a continuous random variable drawn from a distribution that, with abuse of notation, we will also call Π. To formalize this, let H X Π be a vector space of functions from Π to X. With this definition, we can write the signature of f as f : Θ H. Although H is usually larger than X (and can be even infinite in some cases), we can simplify our notation by equipping H with an inner product to make it a Hilbert space. This allows us to use familiar matrix notation with H and hide the view dependence of f(θ) in H. To complete this formalization, we lift the inner product on X to define an inner product on H: h, g H : h, g := Eπ Π[ h(π), g(π) ] = Eπ Π[h(π) g(π)]. This formulation allows us to handle the multiview nature of Ne RFs and SIREN panoramas in a mathematically rigorous way while preserving the convenient notation. The partial application of f returns an entire view-dependent function f(θ) H. Given a set of images (xi)i [N] of the scene from different views (πi)i [N], the standard diffrep task is to find a θ Θ that minimizes L(θ) = P i [N] f(θ)(πi) xi . Because the diffreps and f are both differentiable, this can be accomplished using first-order solvers, such as gradient descent or Adam. For pedagogical convenience, we identify H with X in the following sections. This identification simplifies the presentation, provides a more interpretable and intuitive perspective on our method, and is precise when Π is a singleton. Because H is a Hilbert space, the main points of our arguments still hold when we consider a larger Π. We describe the specific changes needed to adapt our method to the general multiview setting when H = X, in subsection 3.2. 2.2 Diffusion models Diffusion models implicitly define a probability distribution via reversing a forward noising process. While many different presentations of image diffusion models exist in the literature (e.g., DDPM [Ho et al., 2020], DDIM [Song et al., 2022], Score-Based Generative Modeling through SDEs [Song et al., 2021], EDM [Karras et al., 2022]), they are all equivalent for our intended purposes. Given a noise schedule σ(t) for t [0, T] and a score function log pt(x(t)) for the distribution pt over noisy images x(t) at time t, we can sample from p0 by first initializing x(T) N(0, σ2(T)I) and then following the reverse time probability flow ODE (PF-ODE) given in Karras et al. [2022] to transform the easy-to-sample x(T) into a sample from p0: dt = σ(t)σ(t) log pt(x(t)). (1) The PF-ODE is constructed so that the perturbation kernel (conditional distribution) is given by p0t(x(t)|x(0)) = N(x(t); x(0), σ2(t)I). Using the reparameterization trick, we can write this as x(t) = x(0) + σ(t)ϵ, where ϵ N(0, I). We approximate the score function with a (learned) noise predictor ˆϵ(x(t), t) x(t) E[x(0)|x(t)] the Tweedie formula [Robbins, 1956, Efron, 2011]: log pt(x(t)) = E[x(0)|x(t)] x(t) σ2(t) ˆϵ(x(t),t) 2.3 Training differentiable representations using diffusion priors 2.3.1 Training-free methods Poole et al. [2022] laid the foundation, demonstrating that generating 3D assets from purely 2D image-trained diffusion models was even possible. In Dream Fusion (SDS), they perform gradient ascent using Et,ϵ[w(t)J (ˆϵ(f(θ) + σ(t)ϵ, t) ϵ)], derived from the denoising objective, where J is the Jacobian of the differentiable render function f and ˆϵ is the learned noise predictor. Independently, Wang et al. [2022] introduced Score Jacobian Chaining (SJC), which performs gradient ascent using log p(θ) := Et,ϵ[J log pt(f(θ)+σ(t)ϵ)] with a custom sampling schedule for the ts. The custom schedule can be interpreted as gradient annealing to align the implicit σ(t) of the diffrep with the t used to evaluate the score function. Comparison of methods Using the Tweedie formula, we can rewrite the SDS objective as Et,ϵ h w(t) σ(t) J log pt(f(θ) + σ(t)ϵ) i 1. This expression is identical to the log p(θ) term from SJC if we let w(t) = σ(t) for all t. Both methods follow this gradient to convergence, which leads to a critical point a local maximum or mode in the log p(θ) landscape. To approximate the objective, both approaches use Monte Carlo sampling. Both approaches optimize the objective using gradient ascent (GA). While GA on the transformed score resembles solving the PF-ODE particularly in their discretized forms the two procedures serve fundamentally different purposes. GA cares only about finding a local maximum in the log p(θ) landscape, regardless of the trajectory taken to get there. In contrast, the PF-ODE follows a specific path to ensure that generated samples are typical of the distribution. For a discussion of the limitations of GA and why it does not produce representative samples from the distribution, see Appendix A. Recent developments Several subsequent works have built upon the SDS and SJC methods by incorporating additional inputs, fine-tuning, and regularization to enhance the quality of the generated 3D assets. Zero-1-to-3 [Liu et al.] expands on SJC by fine-tuning the score function to leverage a single real input image and additional view information. Magic 123 [Qian et al.] further builds on this by incorporating additional priors derived from 3D data. Fantasia3D [Chen et al.] separates geometry modeling and appearance into distinct components, using SDS to update both. Hi FA [Zhu et al.] introduces a modified schedule and applies regularization to improve the SDS process. Finally, Latent Ne RF [Metzer et al.] utilizes SDS but parameterizes the object in the latent space of a stable diffusion autoencoder, rendering into latent dimensions rather than RGB values. While these methods rely on the SDS/SJC framework for mode-seeking, our work takes a wholly different approach, focusing on developing a more faithful sampling procedure to replace SDS/SJC for both 3D generation and broader differentiable function sampling. 1Note that E[ϵ] = 0, so the additional ϵ term in the SDS objective behaves as an unbiased variance reduction term for the score estimate since Cov(ˆϵ, ϵ) > 0. 2.3.2 Pretrained and fine-tuning methods Unlike SDS and SJC which are entirely zero-shot and can be performed with a frozen diffusion model several other methods have been developed to achieve improved quality and diversity of the generated diffreps, albeit at the cost of additional fine-tuning. While some of these methods require additional data [Zhang et al.], Prolific Dreamer (VSD) [Wang et al., 2023] and Diff Instruct [Luo et al., 2023] are two examples in which diffusion models can be fine-tuned or distilled using only synthetically generated data. VSD specifically addresses the problem of generating 3D assets from a 2D diffusion model using particle-based variational inference to follow the Wasserstein gradient flow, solving the KL objective from SDS in the W2(Θ) parameter distribution space. This approach produces high-quality, diverse samples even at lower guidance levels. In our work, we restrict ourselves to the training-free setting. This choice more readily enables application to new modalities and is independent of the score function estimator s architecture. 2.4 Constrained sampling methods Our method requires generating multiple consistent images of a scene from different views. For instance, we generate various images of a 3D scene from different camera locations and orientations to determine Ne RF parameters. These images must be consistent to ensure coherent updates to the diffrep parameters. While the diffrep inherently enforces this consistency, we can improve our method s convergence rate by encouraging the reverse process to satisfy consistency conditions using constrained sampling methods. Several approaches enable constrained sampling from an unconditionally trained diffusion model. The naïve approach of projecting onto constraints [Song et al., 2021] leads to samples lacking global coherence. In contrast, Lugmayr et al. [2022] propose Re Paint a simple method for inpainting that intermingles forward steps with reverse steps to harmonize generated noisy samples with constraint information. For references to more general methods used to constrain the distribution by conditioning the score function, see Appendix B. 3 Pulling back diffusion models to sample differentiable representations The score model associated with the noise predictor implicitly defines a distribution on the data space X, and the reverse time PF-ODE from Eq. 1 provides a means to sample from that distribution. In this section, we will show how to use the score model and the PF-ODE to sample the parameters of a differentiable representation (diffrep) so that the rendered views resemble samples from the implicit distribution. SDS and SJC derive their parameter updates using the pullback of the sample score function through the render map f. This pullback is obtained using the chain rule: d(log p) dθ = d(log p) dx dx dθ . The pullback score is then used as is for gradient ascent in the parameter space. However, since p is a distribution not a simple scalar field the change of variables formula requires an additional log det J volume adjustment term for the correct pulled-back score function in the parameter space, where J is the Jacobian of f. A careful examination of this approach through the lens of differential geometry reveals even deeper issues. In differential geometry, it is a vector field (a section of the tangent bundle TX) not a differential 1-form (a section of the cotangent bundle T X) that defines the integral curves (flows) on a manifold. To derive the probability flow ODE in the parameter space, we must pull back the sample vector field dx dt TX to the parameter vector field dθ dt TΘ using f. The pullback of the vector field through f is given by (f dx dt )|θ = (J J) 1J dx dt |f(θ), Thus, as we derive in Appendix C, the pullback of the probability flow ODE is: dt = σ(t)σ(t)(J J) 1J log pt(f(θ)). (2) We note that this is in contrast to the score chaining from Poole et al. [2022], Wang et al. [2022]. The confusion stems from comparing the different types of pulled-back elements. While dx dt in Eq. 1 2 (J J) 1 3 dθ dt = (J J) 1J dx log p(θ) = J log p(x) log p(x) (a) (J J) 1J dx (b) J log p(x) (c) λJ log p(x) Figure 1: (Left) Commutative diagram showing how the PF-ODE vector field gets pulled back through f, respecting the differential geometry. The process involves: 1 converting dx dt to the cotangent vector field log p(x) (up to scaling terms) with the Euclidean metric I, 2 pulling back log p(x) via the chain rule using the Jacobian J, and then 3 transforming the pulled back differential form score function into the corresponding vector field using the inverse of pulled back metric (J J) 1. When used in a PF-ODE, SJC and SDS take the bottom path with the chain rule, however they do not complete the path by neglecting the T Θ TΘ transformation. (Right) SIREN image renders generated using the PF-ODE schedule with the prompt An astronaut riding a horse using the: (a) complete pulled-back dx dt vector field, (b) pulled-back covector field from SJC (omitting step 3 ) J log p(x), (c) Scaled pulled-back covector field from SJC λ = 0.0001. is a vector field, the score function log pt(x(t)) is a covector field a differential form. Pulling back the score function as a differential form correctly yields J log pt(f(θ)) the term used in SJC. The issue lies with the hidden (inverse) Euclidean metric within Eq. 1, which converts the differential-form score function into the corresponding vector field. In canonical coordinates, the Euclidean metric is the identity. Consequently, the components of the score function remain unchanged when transformed into a vector field. Therefore we can safely ignore the metric term in the PF-ODE formulation of Eq. 1 for the sample space X. However, this is not true for the diffrep parameter space Θ. To convert the pulled-back score function into the corresponding pulled-back vector field, we must use the pulled-back inverse Euclidean metric given by (J J) 1. This yields the pulled-back form of the PF-ODE in Eq. 22. Fig. 1 illustrates this procedure via a commutative diagram and provides an example of what goes wrong if you use the incorrect pulled-back term, as suggested by SDS and SJC. In the SDS and SJC approaches, the (J J) 1 term behaves like a PSD preconditioner for gradient ascent. It may accelerate the convergence rate to the mode but does not fundamentally change the solution. However, this term is critical for the pulled-back PF-ODE since it impacts the entire trajectory and the ultimate sample (see Appendix A for more discussion). For an explanation of why the log det J term is absent in the mathematically correct update, see Appendix C. A more intuitive way to understand why the correct update is given by Eq. 2 rather than SDS and SJC s version is to consider the case where the input and output dimensions are equal. From the chain rule, we have dθ dt . Using the inverse function theorem, we compute dθ dθ 1 = J 1, where J is f s Jacobian. This leads to dθ dt = J 1 dx dt . For invertible f, this equation and Eq. 2 are equivalent. For non-invertible f, we can interpret the pullback as the solution to the least-squares minimization problem arg min dθ 3.1 Efficient implementation Separated noise Following the pulled-back PF-ODE in Eq. 2, we can find θt such that f(θt) represents a sample x(t) p0t(x(t)|x(0)) = N(x(t); x(0), σ2(t)I). However, this approach has a limitation: x(t) is noisy, and most diffrep architectures are optimized for typical, noise-free images. Consequently, J is likely ill-conditioned, leading to longer convergence times. 2One may be tempted to impose that the metric is Euclidean both in the X space and the Θ space, but these two choices are incompatible because the render transformation maps between the two spaces. Figure 2: The parameters of the diffrep θt Θ (torus) are used to render the noiseless signal f(θt) = bx0(t), which are then combined with the noise σ(t)ϵ to generate the noisy sample x(t). We can pull back each step of the reverse diffusion process to update the parameters θt + θt. To address this issue, we factor x(t) into a noiseless signal and noise using the reparameterization of the perturbation kernel: x(t) = bx0(t) + σ(t)ϵ(t). By letting ϵ(t) = ϵ remain constant throughout the sampling trajectory and starting with bx0(T) = 0, we can update bx0 with dbx0 dt σ(t)ϵ. This decomposition allows f(θt) to represent the noiseless bx0(t) instead of x(t), substantially improving the conditioning of J. Fig. 2 illustrates how this separation works in practice. Efficient optimization The Jacobian J in Eq. 2 represents the derivative of the image with respect to the diffrep parameters. For all but the smallest examples, explicitly forming J J is computationally intractable. While iterative linear solvers together with modern automatic differentiation frameworks (e.g., [Potapczynski et al., 2024]) could solve Eq. 2, we found it faster to compute the parameter update as the solution to a non-linear optimization problem. This approach avoids the costly JVP (Jacobian vector product) required to multiply with J J. For each step with size t, we solve: θ = arg min δθ f(θ) f(θ δθ) σ(t) t(ˆϵ(f(θ) + σ(t)ϵ, t) ϵ) 2 . (3) As t 0, using the linearization f(θ θ) = f(θ) J θ + O( θ 2), this optimization objective approaches the linear least-squares solution. The optimal θ t converges to the exact dθ dt from Eq. 2. In practice, we employ the Adam optimizer to solve Eq. 3 for whatever discrete t is used in the diffusion sampling process. This procedure avoids the explicit formation of the linear least-squares solution. 3.2 Coupled images, stochastic functions, and 3D rendering Thus far, we have only considered the case where the output of the render function is a single image x X. However, our interest lies in render functions f that map parameters θ Θ to an entire view-dependent image space H X Π. Using the inner product defined in subsection 2.1, we derive the view-dependent pullback PF-ODE: dθ dt = f dbx0 dt σ(t)ϵ = σ(t)σ(t)E J π Jπ 1E h J π log pt(x(π), π) ϵ(π) where the expectations are taken over the views π Π. Here, log pt(x(π), π) represents the view-specific score function. We use the original score function log pt(x(π)), with additional view information in the prompt. ϵ(π) denotes the separately managed noise rendered from view π. Examining the optimization form of this equation provides further insight. We can interpret it as minimizing the norm in the function space 2 H = Eπ Π 2 : This is directly analogous to Eq. 3, but with a different inner product. Explicitly written: θ = arg min θ Eπ Π f(θ, π) f(θ θ, π) σ t(ˆϵt(π) ϵ(π)) 2 , (5) which we can empirically minimize using view samples π Π. 3.3 Consistency and implicit constraints Our method successfully pulls back the PF-ODE dx dt using the pull-back of the score function log pt(x(t)). However, our true goal is to pull back log pt(x(t)|bx0(s) range(f)) for all s t. This ensures that f can render the noiseless components bx0 throughout the remaining reverse process. For example, when pulling back the PF-ODE for different views of a 3D scene, we want to ensure consistency across all the views. When f is sufficiently expressive and invertible, log pt(x(t)|bx0(s) range(f)) log pt(x(t)). This would allow us first to sample x(0) using the PF-ODE from Eq. 1 and then invert f to find θ(0). However, for most significant applications of our method, f is not invertible. In noninvertible cases, consider the sampling trajectory x(t) in X when bx0(t) / range f, particularly when no nearby sample has high probability in range f. Each step of Eq. 2 follows the direction in Θ that best approximates the direction of the score model in a least-squares sense. When f is not invertible, J is not full rank, and the update to θ will not precisely match the trajectory in x. Consequently, f(θt θ) x(t t) will be significant. The score model, unaware of this, will continue to point towards high-probability regions outside the range of f. For example, consider an f that only allows low-frequency Fourier modes in generated samples. The unconstrained score model might favor images with high-frequency content (e.g., hair, explosions, detailed textures), resulting in blurry, unresolved images. However, suppose the score model was aware of this constraint against high-frequency details. In that case, it might guide samples toward more suitable content, such as landscapes, slow waves, or impressionist art dominated by expressible low-frequency components. This issue becomes more pronounced when sampling multiple coupled images through f, such as with panoramas and Ne RFs, where various views must be consistent. To address this challenge, we adapt the Re Paint method [Lugmayr et al., 2022], initially designed for inpainting, to guide our PF-ODE towards more renderable samples. Re Paint utilizes the complete Langevin SDE diffusion process, interspersing forward and reverse steps in the sampling schedule. This approach allows the process to correct for inconsistencies during sampling. The forward steps harmonize samples at the current time step by carrying information from earlier steps that implicitly encode the constraint. Re Paint requires a stochastic process for conditioning, so we employ the DDIM [Song et al., 2022] sampling procedure with controllable noise at each step. We derive the forward updates of DDIM in Appendix D. 3.4 Summary Our method introduces significant advancements for sampling diffreps using pretrained diffusion models by performing actual sampling instead of merely mode-seeking, thus capturing the full diversity of high-dimensional distributions. We derive the correct pullback of the PF-ODE, incorporating the essential (J J) 1 term to ensure unbiased sampling. Efficiency is enhanced through separated noise handling and a faster suboptimization process, allowing practical application for complex representations. Additionally, our approach extends to coupled images and stochastic functions with implicit constraints, making it suitable for tasks like panorama generation and 3D rendering. A comprehensive summary of the contributions in this section and the pseudocode for the DDRep algorithm using DDIM sampling with Re Paint is presented in Appendix E. 0 3 10 30 100 CFG Scale Ours SD ref SJC Figure 3: The left figure contains sample renders using the prompt A woman is standing at a crosswalk at a traffic intersection. from the reference SD (top), our method (middle), and SJC (bottom) over the CFG scales [0,3,10,30,100] from left to right. The right plot is the KID metric (closer to 0 is better) measured on the SIRENs sampled from our method, the SD reference samples, and the SIRENs sampled using SJC. Figure 4: Samples generated using SD ref (top), Our method (middle), and SJC (bottom) using the same prompt An office cubicle with four different types of computers with eight different seeds. 4 Experiments All our experiments used the Hugging Face implementation of Stable Diffusion 1.5 (SDv1.5) from Rombach et al. [2022] with the default hyperparameters as the noise predictor model. We conducted our experiments on a single NVIDIA A6000 GPU. We used the complete Langevin SDE given by the DDIM [Song et al., 2022] procedure for all our experiments, with η = 0.75 as the stochastic interpolation hyperparameter. We interspersed forward and reverse steps to harmonize the diffrep constraints in the sampling procedure. For the suboptimization problem described in Eq. 3, we used the Adam optimizer for 200 steps. 4.1 Image SIRENs We compare the results of our method with those of SJC [Wang et al., 2022] when used to produce SIRENs [Vincent, 2011] for images. While we could directly fit the SIREN to the final image x(0) obtained by first solving the PF-ODE in the image space we use this scenario to compare different training-free generation methods quantitatively. The SIREN we used was a 3-layer MLP mapping the 2D (x, y) [0, 1]2 pixel coordinates to the R4 latent space of SDv1.5. The MLP had a 128-dimensional sinusoidal embedding layer, a depth of 3, a width of 256, and used sin( ) as the activation function. To render an image, we used a 64 64 grid of equally spaced pixel coordinates from [0, 0] to [1, 1] with a shape of (64, 64, 2). The MLP CFG Scale 10.0 No Re Paint Figure 5: Comparison of landscape panoramas sampled using our method. The top panorama is sampled using the Re Paint method, while the bottom is sampled without Re Paint. Both approaches use 460 function evaluations (NFEs) to ensure fairness and a CFG scale of 10.0. The prompt for these panoramas was Landscape picture of a mountain range in the background with an empty plain in the foreground 50mm f/1.8 . output was a latent image with shape (64, 64, 4), which we decoded using the SDv1.5 VQ-VAE into an RGB image with shape (512, 512, 3). We calculated the final metrics on these rendered images. We used the first 100 captions (ordered by id) from the MS-COCO 2017 validation dataset [Lin et al.] as prompts. We generated three sets of images: (1) reference images using SDv1.5, (2) images rendered from SIRENs sampled using our method, and (3) images rendered from SIRENs generated using SJC [Wang et al., 2022]. We compared these sets against 100 baseline images generated using SDv1.5 with the same prompts but a different seed. We repeated this experiment at five CFG scales [Ho and Salimans] (0, 3, 10, 30, 100). We used the KID metric from Bi nkowski et al. to compare the generated images, as it converges within 100 image samples while maintaining good comparison ability. For runtime metrics, see Appendix F. The example images in Fig. 3 (left) demonstrate that the samples generated by our method are almost indistinguishable from the reference. The results in Fig. 3 (right) clearly show that our method produces images on par with SDv1.5 reference samples. To further support this observation, we measured the PSNR, SSIM, and LPIPS scores of our method s samples against the SDv1.5 reference images (Table 1). The results show that our method produces SIREN images nearly identical to the reference set. To demonstrate that our method achieves sampling instead of mode-seeking, we plot eight samples using the same prompt for all models in Fig. 4 with different seeds at CFG scale 10. The samples generated by SJC all look very similar, while those generated by our method and SD show significantly more diversity. 4.2 SIREN Panoramas In this section, we demonstrate how our method can generate SIREN panoramas using the same architecture as the SIREN from subsection 4.1 but with a modified render function. We begin by sampling a view grid of pixel locations with the shape (8, 64, 64, 2) where the batch size is 8, the height and width are 64, and the last 2 dimensions are the (x, y) coordinates. The y values are equally spaced in the range [0,1], and x values are equally spaced in the range [r, r +1/ar] mod 1.0 for r U(0, 1). The mod 1.0 constraint ensures horizontal wrapping, making the panorama continuous over the entire 360 azimuth. We render the view by sampling from the SIREN at the discrete pixel locations given by the view grid, then decode the latent output using the SDv1.5 VQ-VAE. For our experiments, we assumed that the images were taken by a camera lens with a 45 field of view (equivalent to a 50mm focal length), corresponding to an aspect ratio of ar = 8. The SIREN panorama imposes implicit constraints to ensure consistency across overlapping views. Fig. 5 and Fig. 7 illustrate the impact of the consistency conditioning described in subsection 3.3. Figure 6: Ne RFs generated using our method with the prompts (top) a photo of a delicious hamburger, centered, isolated, 4K. , (middle) a DSLR photo of a rose , and (bottom) a DSLR photo of a yellow duck". Both Re Paint and non-Re Paint methods use the same number of function evaluations (460 NFEs) to ensure fair comparisons. Using Re Paint substantially improves the quality and consistency of the generated panorama. This improvement is even more pronounced at low CFG scales, where the diffusion model is encouraged to be more creative and is typically less likely to produce consistent images, as seen in Fig. 7. For runtime metrics and additional panoramas generated using our approach, see Appendix F. 4.3 3D Ne RFs For our Ne RF experiments, we employed the vox Ne RF architecture used by SJC [Wang et al., 2022], chosen for its speed and efficiency. While we did not utilize any additional loss terms from their method, we found that incorporating the object-centric scene initialization from Wang et al. [2023] significantly improved our results. This initialization bootstrapped the diffusion process and led to faster convergence. Specifically, we used the initialization function σinit(x) = 10 1 x 0.5 , as detailed in Wang et al. [2023]. For more experimental details, see Appendix F. The Ne RFs generated using our method, as illustrated in Fig. 6, demonstrate our approach s capability to produce detailed, high-quality 3D representations from text prompts. These results showcase the effectiveness of our method in generating complex 3D structures while maintaining consistency across multiple views. 5 Conclusion We have presented a comprehensive, training-free approach for pulling back the sampling process of a diffusion model through a generic differentiable function. By formalizing the problem, we have addressed two critical issues with prior approaches: enabling true sampling rather than just mode seeking (crucial at low guidance levels), and addressing a latent consistency constraint using Re Paint to improve generation quality. For future work and limitations, see Appendix G. Our method opens up new possibilities for generating complex and constrained outputs using pretrained diffusion models. We hope that future research will build on these insights to further expand the capabilities and applications of diffusion models in areas such as 3D content generation, game design, and beyond. Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising Diffusion Probabilistic Models, December 2020. URL http://arxiv.org/abs/2006.11239. ar Xiv:2006.11239 [cs, stat]. Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising Diffusion Implicit Models, October 2022. URL http://arxiv.org/abs/2010.02502. ar Xiv:2010.02502 [cs]. Yang Song, Jascha Sohl-Dickstein, Diederik P. Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-Based Generative Modeling through Stochastic Differential Equations, February 2021. URL http://arxiv.org/abs/2011.13456. ar Xiv:2011.13456 [cs, stat]. Tero Karras, Miika Aittala, Timo Aila, and Samuli Laine. Elucidating the Design Space of Diffusion-Based Generative Models, October 2022. URL http://arxiv.org/abs/2206. 00364. ar Xiv:2206.00364 [cs, stat]. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High Resolution Image Synthesis with Latent Diffusion Models, April 2022. URL http://arxiv. org/abs/2112.10752. ar Xiv:2112.10752 [cs]. Zhengyi Wang, Cheng Lu, Yikai Wang, Fan Bao, Chongxuan Li, Hang Su, and Jun Zhu. Prolific Dreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation, May 2023. URL http://arxiv.org/abs/2305.16213. ar Xiv:2305.16213 [cs]. Weijian Luo, Tianyang Hu, Shifeng Zhang, Jiacheng Sun, Zhenguo Li, and Zhihua Zhang. Diff Instruct: A Universal Approach for Transferring Knowledge From Pre-trained Diffusion Models, May 2023. URL http://arxiv.org/abs/2305.18455. ar Xiv:2305.18455 [cs]. Haochen Wang, Xiaodan Du, Jiahao Li, Raymond A. Yeh, and Greg Shakhnarovich. Score Jacobian Chaining: Lifting Pretrained 2D Diffusion Models for 3D Generation, December 2022. URL http://arxiv.org/abs/2212.00774. ar Xiv:2212.00774 [cs]. Ben Poole, Ajay Jain, Jonathan T. Barron, and Ben Mildenhall. Dream Fusion: Text-to-3D using 2D Diffusion, September 2022. URL http://arxiv.org/abs/2209.14988. ar Xiv:2209.14988 [cs, stat]. Ben Mildenhall, Pratul P. Srinivasan, Matthew Tancik, Jonathan T. Barron, Ravi Ramamoorthi, and Ren Ng. Ne RF: Representing Scenes as Neural Radiance Fields for View Synthesis, August 2020. URL http://arxiv.org/abs/2003.08934. ar Xiv:2003.08934 [cs]. Pascal Vincent. A connection between score matching and denoising autoencoders. Neural Computation, 23(7):1661 1674, July 2011. ISSN 0899-7667. doi: 10.1162/NECO_a_00142. URL https://doi.org/10.1162/NECO_a_00142. Vincent Sitzmann, Julien N. P. Martel, Alexander W. Bergman, David B. Lindell, and Gordon Wetzstein. Implicit Neural Representations with Periodic Activation Functions, June 2020. URL http://arxiv.org/abs/2006.09661. ar Xiv:2006.09661 [cs, eess]. Thomas Müller, Alex Evans, Christoph Schied, and Alexander Keller. Instant Neural Graphics Primitives with a Multiresolution Hash Encoding. ACM Transactions on Graphics, 41(4):1 15, July 2022. ISSN 0730-0301, 1557-7368. doi: 10.1145/3528223.3530127. URL http: //arxiv.org/abs/2201.05989. ar Xiv:2201.05989 [cs]. Bernhard Kerbl, Georgios Kopanas, Thomas Leimkühler, and George Drettakis. 3D Gaussian Splatting for Real-Time Radiance Field Rendering, August 2023. URL https://arxiv.org/ abs/2308.04079v1. Herbert Robbins. An Empirical Bayes Approach to Statistics. In Proceedings of the Third Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics, volume 3.1, pages 157 164. University of California Press, January 1956. URL https://projecteuclid.org/ ebooks/berkeley-symposium-on-mathematical-statistics-and-probability/ Proceedings-of-the-Third-Berkeley-Symposium-on-Mathematical-Statistics-and/ chapter/An-Empirical-Bayes-Approach-to-Statistics/bsmsp/1200501653. Bradley Efron. Tweedie s Formula and Selection Bias. Journal of the American Statistical Association, 106(496):1602 1614, 2011. ISSN 0162-1459. URL https://www.ncbi.nlm.nih.gov/pmc/ articles/PMC3325056/. Ruoshi Liu, Rundi Wu, Basile Van Hoorick, Pavel Tokmakov, Sergey Zakharov, and Carl Vondrick. Zero-1-to-3: Zero-shot One Image to 3D Object. URL http://arxiv.org/abs/2303.11328. Guocheng Qian, Jinjie Mai, Abdullah Hamdi, Jian Ren, Aliaksandr Siarohin, Bing Li, Hsin-Ying Lee, Ivan Skorokhodov, Peter Wonka, Sergey Tulyakov, and Bernard Ghanem. Magic123: One Image to High-Quality 3D Object Generation Using Both 2D and 3D Diffusion Priors. URL https://arxiv.org/abs/2306.17843v2. Rui Chen, Yongwei Chen, Ningxin Jiao, and Kui Jia. Fantasia3D: Disentangling Geometry and Appearance for High-quality Text-to-3D Content Creation. URL http://arxiv.org/abs/2303. 13873. Junzhe Zhu, Peiye Zhuang, and Sanmi Koyejo. Hi FA: High-fidelity Text-to-3D Generation with Advanced Diffusion Guidance. URL http://arxiv.org/abs/2305.18766. Gal Metzer, Elad Richardson, Or Patashnik, Raja Giryes, and Daniel Cohen-Or. Latent-Ne RF for Shape-Guided Generation of 3D Shapes and Textures. URL https://arxiv.org/abs/2211. 07600v1. Cheng Zhang, Qianyi Wu, Camilo Cruz Gambardella, Xiaoshui Huang, Dinh Phung, Wanli Ouyang, and Jianfei Cai. Taming Stable Diffusion for Text to 360{\deg} Panorama Image Generation. URL http://arxiv.org/abs/2404.07949. Andreas Lugmayr, Martin Danelljan, Andres Romero, Fisher Yu, Radu Timofte, and Luc Van Gool. Re Paint: Inpainting using Denoising Diffusion Probabilistic Models, August 2022. URL http: //arxiv.org/abs/2201.09865. ar Xiv:2201.09865 [cs]. Andres Potapczynski, Marc Finzi, Geoff Pleiss, and Andrew G Wilson. Cola: Exploiting compositional structure for automatic and efficient numerical linear algebra. Advances in Neural Information Processing Systems, 36, 2024. Tsung-Yi Lin, Michael Maire, Serge Belongie, Lubomir Bourdev, Ross Girshick, James Hays, Pietro Perona, Deva Ramanan, C. Lawrence Zitnick, and Piotr Dollár. Microsoft COCO: Common Objects in Context. URL http://arxiv.org/abs/1405.0312. Jonathan Ho and Tim Salimans. Classifier-Free Diffusion Guidance. URL http://arxiv.org/ abs/2207.12598. Mikołaj Bi nkowski, Danica J. Sutherland, Michael Arbel, and Arthur Gretton. Demystifying MMD GANs. URL http://arxiv.org/abs/1801.01401. Hyungjin Chung, Byeongsu Sim, Dohoon Ryu, and Jong Chul Ye. Improving Diffusion Models for Inverse Problems using Manifold Constraints, October 2022. URL http://arxiv.org/abs/ 2206.00941. ar Xiv:2206.00941 [cs, stat]. Hyungjin Chung, Jeongsol Kim, Michael T. Mccann, Marc L. Klasky, and Jong Chul Ye. Diffusion Posterior Sampling for General Noisy Inverse Problems, February 2023. URL http://arxiv. org/abs/2209.14687. ar Xiv:2209.14687 [cs, stat]. Arpit Bansal, Hong-Min Chu, Avi Schwarzschild, Soumyadip Sengupta, Micah Goldblum, Jonas Geiping, and Tom Goldstein. Universal guidance for diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 843 852, 2023. Marc Finzi, Anudhyan Boral, Andrew Gordon Wilson, Fei Sha, and Leonardo Zepeda-Núñez. User-defined Event Sampling and Uncertainty Quantification in Diffusion Models for Physical Dynamical Systems, June 2023. URL http://arxiv.org/abs/2306.07526. ar Xiv:2306.07526 [cs]. Litu Rout, Yujia Chen, Abhishek Kumar, Constantine Caramanis, Sanjay Shakkottai, and Wen-Sheng Chu. Beyond First-Order Tweedie: Solving Inverse Problems using Latent Diffusion, December 2023. URL http://arxiv.org/abs/2312.00852. ar Xiv:2312.00852 [cs, stat]. A Gradient Ascent (GA) does not sample It is possible to reparameterize the Gradient Ascent (GA) procedure from Poole et al. [2022] and Wang et al. [2022] in a way that resembles solving the PF-ODE by carefully selecting optimizer hyperparameters and stopping times. However, this reparameterization is highly specific. The crucial difference lies in the interpretation of the trajectory: in GA, the path taken is typically irrelevant as long as it converges to the critical point, allowing us the flexibility to multiply the gradient by any positive semidefinite (PSD) matrix without affecting the solution. For the PF-ODE, however, the trajectory is everything. The stopping time, weighting terms, and score evaluations must be precisely aligned to ensure that the solution looks like a typical sample. Even small deviations in the path, like evolving the PF-ODE for too long or too short, can lead to either oversmoothed samples (with too much entropy removed) or under-resolved, noisy samples. Thus, while GA might superficially look like a discretized version of the PF-ODE, the underlying processes are inherently distinct: one finds the mode, and the other describes the full trajectory required to generate samples from the distribution. In the case of SDS and SJC, they both use Adam as the optimizer, ignoring any of the nuanced considerations needed to transform the PF-ODE into the corresponding optimization procedure. In practice, this distinction means that while both SDS and SJC may yield plausible diffreps at high classifier-free guidance levels, they tend to falter when producing coherent samples under lower CFG weights particularly for complex, multimodal distributions with many degrees of freedom. High CFG levels may mask these limitations by strongly biasing the output towards the mode, which aligns more closely with the GA approach, but at the cost of reduced sample diversity and potentially missed details captured by the full distribution. The problem with mode-seeking To appreciate the limitations of using the mode as a proxy for sampling, we consider the multi-modal and the high-dimensional settings. In the multimodal setting, the sample distribution may contain several distinct high-density regions. However, mode-seeking algorithms focus on only one of these regions, sacrificing sample diversity. In low-dimensional spaces, the mode often resembles a typical sample. However, this intuition breaks down in high-dimensional spaces (d 1), such as the space of images. This counterintuitive behavior can be explained using the thin-shell phenomenon. For a high-dimensional standard Gaussian distribution, samples are not concentrated near the mode (the origin), as one might expect. Instead, they predominantly reside in an exponentially thin shell at the boundary of a d-radius ball centered on the mode. This phenomenon explains why the mode can be an anomalous and atypical point in high-dimensional distributions. As an illustrative example, consider sampling a normalized pure-noise image. Despite having a mode of 0, we would almost never expect to sample a uniformly gray image. This provides some insight as to why the mode of a distribution with several degrees of freedom will lack the quality and details present only in samples from the thin shell. B General conditioning While Re Paint can be used to condition the reverse trajectory, it tends to be less generalizable for several tasks. For more general conditioning including inverse problems MCG [Chung et al., 2022] and its extension DPS [Chung et al., 2023] have gained popularity. These methods add a corrective term to the score, encouraging reverse steps to minimize constraint violations for the predicted final sample ˆx0 according to the Tweedie formula. MCG and DPS have seen considerable application and extension to broader settings [Bansal et al., 2023]. Finzi et al. [2023] showed this correction to be an asymptotically exact approximation of log pt(constraint|x), while Rout et al. [2023] derive the relevant term from Bayes rule and higher-order corrections. C PF-ODE Pullback Derivation and Geometric Details This section provides the details for the derivation of Eq. 2 using differential geometry. First, we provide a full expansion of Eq. 1. For any point in our sample space x X, we can define a Riemannian metric Mx that describes the local geometry of the space around x. Given tangent vectors y, z at point x, we define the inner product using this metric y, z Mx = y T Mxz where Mx is represented using a positive-definite matrix. We implicitly use the Euclidean metric for the image space X, which corresponds to the identity matrix Mx = Id, x X. However, in curved spaces such as the parameter space Θ, the corresponding Riemannian metric is not necessarily the identity and can vary based on where it is evaluated. Here, we use matrix notation rather than Einstein summation or coordinate-free notation for familiarity within the machine learning context. However, we note that matrix notation can obscure some important aspects of differential geometry. For example, while the components of the Euclidean metric in the Euclidean coordinate basis form the identity matrix, the metric is not the identity function, as it transforms between two distinct vector spaces. As such, when using the matrix notation, it becomes very important to keep track of the types of objects rather than just their values. Because we use the Euclidean metric for the image space, its role in Eq. 1 is hidden. Explicitly incorporating the metric into the process, we get the following expanded metric PF-ODE dt = σ(t)σ(t)M 1 x log d Pt/dλx, (6) where Pt is the probability measure at time t and λx is the Lebesgue measure on X, and d Pt/dλx is the Radon-Nikodym derivative. Notice that when we use Mx = Id, x X, this equation is equivalent to Eq. 1. If we pull back the dynamics of this process through the render function f : X Θ to the parameter space Θ, we must pull back both the score function and the metric Mx to get the pulled back the vector field f dx dt . The different object types transform as follows with the pullback f : Scalar Functions s C(X) : (f s)(θ) = s(f(θ)) (7) Cotangent Vectors ω T X : (f ω)(θ) = J ω(f(θ)) (8) Metric Tensor M T X T X : (f M)(θ) = J Mf(θ)J (9) Tangent vectors v TX : (f v)(θ) = (J Mf(θ)J) 1J Mf(θ)v(f(θ)), (10) where J is the Jacobian of f evaluated at θ. For a scalar function s(x), the gradient s(x) is co-vector field (also known as a differential 1-form) and can be converted into a vector field using the inverse metric: M 1 x s(x). We can see that Eq. 10 can be expressed concisely in terms of the pullback metric f v = (f M) 1f (Mv), which geometrically corresponds to converting the vector field to a covector field with M, pulling back the covector field (using the chain rule), and then converting back to a vector field with the inverse of the pulled back metric. This sequence is illustrated in Fig. 1. From Eq. 9, the pullback of the Euclidean metric is J IJ = J J. From these expressions, we can see that the form of f dx dt should be: dt = σ(t)σ(t)(J Mf(θ)J) 1J Mf(θ)M 1 f(θ) log d Pt/dλx dt = σ(t)σ(t)(J Mf(θ)J) 1 log d Pt/dλx dt = σ(t)σ(t)(J J) 1 log pt (on Euclidean X where Mx = I). Thus the pullback of the reverse time PF-ODE is dt = (J J) 1J dx dt = σ(t)σ(t)(J J) 1J log pt(f(θ)). (11) Change of variables contribution. Some readers may be surprised to see that the Jacobian log determinant does not show up in this transformation. Though somewhat technical, this can be seen by unpacking the Radon-Nikodym derivative d Pt/dλx. In general, on a manifold with metric M, the Lebesgue measure is given by dλ = det Mxdx. Therefore, when written in terms of the density d Pt/dx = pt(x), one has d Pt/dλx = pt(x)/ det Mx. When evaluated only in the Euclidean space where the coordinates are chosen such that Mx = I, this detail can be safely ignored. However, when pulling back to the parameter space, these hidden terms matter. The Radon-Nikodym derivative d Pt/dλx is a scalar field and is not transformed under the pullback, however pt is a scalar density and satisfies (f pt)(θ) = pt(f(θ)) q the familiar change of variables formula. When the input and output dimensions are the same pΘ(θ) = p X (f(θ)) det J. However, det Mx is also a scalar density and similarly transforms: det f Mx = p det J Mx J = Assembling these two components together, we see that the determinant contribution from the change of variables formula cancels with the contribution from the change of the Lebesgue measure in the sampling ODE. f log(d Pt/dλx) = f log pt/ p f log(d Pt/dλx) = J log(f pt) log f p f log(d Pt/dλx) = J log pt + 1 2 log det J J 1 2 log det J Mx J f log(d Pt/dλx) = J log pt. This is why the change of variables term 1 2 log det J J term does not appear in the pullback of the PF-ODE. D Details on constrained sampling For our experiments, we used the stochastic non-Markovian reverse process from DDIM Song et al. [2022]. For the Re Paint steps, we also need to take forward steps to encourage the samples that are renderable . In this section, we prove the correct form for the forward process looks like x(t) = s(t)ex0(t dt) + σ(t) σ(t dt) σ2(t dt) τ 2(t)eϵ0(t dt) + τ(t)ϵf(t) . (12) Where σ(t) is the standard deviation and can be calculated using σ(t) = 1 1/α2(t)+1 and s(t) is the scale of signal. It can be calculated using s(t) = 1 α2(t)+1 where α(t) is the noise-to-signal ratio. In the main body of the text, we let σ(t) notate the noise-to-signal ratio, and we assume s(t) = 1 but that is not necessarily true in the general case. Theorem D.1. Eq. 12 is the correct forward update for the non-Markovian process. Proof. We will use the notation x(t|0) = x(t) = x(t)|x(0), x(t|t dt, 0) = x(t)|x(t dt), x(0), and x(t dt|t, 0) = x(t dt)|x(t)x(0). The same holds if we replace x with ϵ0. Using the reparametrization trick we can write x(t) = x(t|0) = s(t)x(0) + σ(t)ϵ0(t|0), so ln p(x(t|0)) = 1 2ϵ2 0(t|0) + c. Using DDIM and the reparameterization trick, we can also write ln p(x(t dt|t, 0)) = 1 1 τ(t)x(t dt) s(t dt) σ2(t dt) τ 2(t) τ(t) ϵ0(t dt) σ2(t dt) τ 2(t) Now we can use Bayes theorem p(x(t|t dt, 0)) = p(x(t dt|t, 0))p(x(t|0)) p(x(t dt|0)) ln p(x(t|t dt, 0)) = 1 τ(t) ϵ0(t dt) σ2(t dt) τ 2(t) + ϵ2 0(t) ϵ2 0(t dt) σ2(t dt) τ 2(t) τ(t) ϵ0(t dt) σ2(t dt) τ 2(t) σ(t dt) ϵ0(t dt) Using reparameterization, we can say that (conditioned on x(0)), x(t|t dt, 0) = s(t)x(0) + σ(t)ϵ0(t|t dt, 0), where ϵ0(t|t dt, 0) N σ2(t dt) τ 2(t) σ(t dt) ϵ0(t dt), τ 2(t) σ2(t dt) E Summary of key ideas and algorithm Our method introduces several crucial innovations for sampling differentiable representations using pretrained diffusion models: 1. True sampling vs. mode-seeking: Unlike SJC and SDS, which primarily find modes of the distribution, our method aims to generate true samples. This distinction is critical for capturing the full diversity and characteristics of the underlying distribution, especially in high-dimensional spaces. 2. Correct pullback of the PF-ODE: We derive the mathematically correct form of the pulled-back Probability Flow ODE, which includes a crucial (J J) 1 term. While this term acts as a preconditioner in optimization-based approaches (affecting only convergence rates), it is essential for unbiased sampling in our PF-ODE framework. Omitting this term leads to incorrect results. 3. Efficient implementation: We introduce techniques for efficient implementation, including separated noise handling and a faster suboptimization procedure. These innovations allow for the practical application of our method to sample complex differentiable representations. 4. Generalization to coupled images and stochastic functions: Our method extends naturally to scenarios involving coupled images and stochastic functions, making it applicable to a wide range of problems, including 3D rendering and panorama generation. 5. Handling multimodality and constraints: Our sampling approach naturally handles multimodal distributions, unlike mode-seeking methods. However, this introduces challenges when the samples are incompatible with the constraints imposed by the differentiable representation. We address this by conditioning the sampling process with the consistency constraint using an adapted Re Paint method. These key ideas collectively enable our method to generate high-quality, diverse samples of differentiable representations by encouraging the diffusion model to maintain consistency with the constraints imposed by the representation. The pseudocode for the DDRep algorithm using DDIM sampling with Re Paint is presented in Alg. 1. F Additional experimental details This section presents additional details and runtime metrics for our algorithm with the experimental settings given in section 4. Algorithm 1 DDRep with Re Paint 1: Input: cfg, reverse_steps, jump_interval, jump_len, jump_repeat, η 2: Init: θ = arg minθ Eπ Π f(θ, π) , ϵ(π) N(0, I), π Π 3: schedule = Re Paint Schedule(reverse_steps, jump_interval, jump_len, jump_repeat) 4: for (reverse, t) in schedule do 5: σlangevin = η p σ 2(t) + 1 q 1 σ2(t t)+1 σ2(t)+1 6: ϵlangevin N(0, σ2 langevin) 7: Sample Views: π Π 8: if reverse then 9: x(t) = f(θ, π) + σ(t)ϵ(π) 10: ˆϵt(π) = ˆϵ(x(t), t, view_prompt(π), cfg) 11: bxnext = x(t) σ(t)ˆϵt(π) + q 1 σ2 langevinσ(t t)ˆϵt(π) + σ(t t)ϵlangevin 12: else 13: bxnext = f(θ, π) + σ(t) q 1 σ2 langevinϵ(π) + σ(t)ϵlangevin 14: end if 15: Update Noise Parameters: ϵ(π) = q 1 σ2 langevinϵ(π) + ϵlangevin 16: Update Diffrep Parameters: θ = arg minθ f(θ, π) bxnext + σnextϵ(π) 17: end for cfg scale 0 3 10 30 100 PSNR 29.712 4.231 29.931 5.767 27.593 6.574 23.453 6.397 13.586 4.024 SSIM 0.899 0.086 0.896 0.107 0.888 0.115 0.826 0.145 0.523 0.183 LPIPS 0.044 0.050 0.053 0.074 0.061 0.074 0.098 0.098 0.336 0.138 Table 1: Image similarity scores (mean std) for the images produced using our method and those produced using the SD reference. Image SIRENs Sampling a batch of eight reference images using SDv1.5 takes 39 seconds. Generating SIRENs using SJC depends on the number of iterations used; running SJC for 3000 iterations to generate eight SIRENs takes 694 seconds, corresponding to a per-iteration time of 2.31 seconds. In comparison, sampling a batch of eight SIRENs using our method takes 82 seconds. These timings demonstrate that our method offers a significant speed advantage over SJC while producing comparable results to the reference SDv1.5 samples. The PSNR, SSIM, and LPIPS scores of our method against the SDv1.5 reference images are presented in Table 1. SIREN Panoramas Generating a SIREN panorama with our method takes 218 seconds. Fig. 7 shows how using Re Paint to condition the sampling procedure can generate more coherent panoramas for low CFG scales. Additional examples of SIREN panoramas generated using our method are provided in Fig. 8. 3D Ne RFs Our algorithm processed a batch of eight views per iteration, which was the maximum capacity for the A6000 VRAM. The generation of each Ne RF required approximately 7.2 seconds per NFE (Number of Function Evaluations). We implemented 100 inference steps using the DDIM procedure, with one Re Paint forward step for each reverse step. This configuration resulted in 199 NFEs per Ne RF, translating to about 24 minutes of sampling time for each Ne RF. It is worth noting that run times can fluctuate based on the specific sampling procedure and GPU architecture used. For comparison, generating a comparable Ne RF using SJC takes approximately 34 minutes for 10,000 steps, based on the code provided by the authors. CFG Scale 3.0 No Re Paint Figure 7: Comparison of landscape panoramas sampled using our method. The top panorama is sampled using the Re Paint method, while the bottom is sampled without Re Paint. Both approaches use 460 function evaluations (NFEs) to ensure fairness and a CFG scale of 3.0. The prompt for these panoramas was Landscape picture of a mountain range in the background with an empty plain in the foreground 50mm f/1.8 . Figure 8: Additional SIREN Panoramas with prompts (top) Urban skyline at twilight, city lights twinkling in the distance." (middle) A futuristic kitchen" (bottom) Desert oasis, palm trees surrounding a crystal-clear pool." G Future work and limitations Our work has demonstrated the zero-shot generation of implicit images, panoramas, and Ne RFs. The versatility of our approach opens up a wide range of potential applications, including vector graphics generation, embedding diverse content in scenes through geometric transformations, differentiable heightmaps, and applications using differentiable physics renderers. While we focused on Ne RFs in this paper, more advanced differentiable representations like Gaussian splats and full-scene representations have emerged. Our method should be applicable to these cases, but further work is needed to adapt our approach to these newer representations. Looking ahead, one could envision generating entire maps or game environments using diffusion models. However, this would likely require innovations in amortization and strategies for splitting and combining subproblems. In these cases, the function f may be even more restrictive than for SIRENs and Ne RFs, underscoring the importance of conditionally sampling the implicit constraint. Limitations Our method faces two significant limitations. First, the additional steps required for Re Paint introduce computational overhead. A priori, we cannot determine how many steps are needed to harmonize the constraint into the diffrep. Transitioning to a conditional sampling method like MCG could potentially allow us to skip these extra forward steps and directly integrate the constraint into the solver. However, this requires further investigation to ensure compatibility with our method. Second, the substantial stochasticity in the Monte Carlo estimate of the pullback score over several views, particularly when the Jacobian of the render function is ill-conditioned, poses a challenge. While increasing the number of sample views could reduce variance in our estimate, it would also slow down our algorithm and consume more memory. Further research into view sampling techniques, such as importance sampling, could help decrease variance without compromising computational efficiency. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: We justify all the claims we make in the abstract and introduction throughout the rest of the paper. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: We have a Limitations and conclusion section. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: All theoretical results explicitly state the full set of assumptions and have a complete and correct proof. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: All the details needed to reproduce the experiments are provided in the paper. We will also release the code to reproduce the experiments. The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [No] Justification: The code is not currently in a state ready for distribution. We will release the code after we have some time to clean it up. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: All training and test details are provided. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: We report the mean and standard deviation for all our metrics and show samples using multiple seeds. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: We report that all experiments were conducted using a single A6000 GPU and we provide the wall clock time needed to run each experiment. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: The research in the paper conforms, in every respect, wit the Neur IPS Code of Ethics. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [No] Justification: Our contribution is methodological and we don t know how it may be used. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: Wo do not release any data or models in this paper. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: All assets used in the paper are properly credited. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: The paper does not release new assets. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.