# folddiff_folding_in_point_cloud_diffusion__59a47503.pdf Published in Transactions on Machine Learning Research (08/2025) Fold Diff: Folding in Point Cloud Diffusion Yuzhou Zhao yz4450@princeton.edu Department of Electrical and Computer Engineering Princeton University Duke University J. Matías Di Martino matias.di.martino@duke.edu Department of Computer Science Universidad Católica del Uruguay Duke University Amirhossein Farzam a.farzam@duke.edu Department of Electrical and Computer Engineering Duke University Guillermo Sapiro guillermos@princeton.edu Department of Electrical and Computer Engineering Princeton University Apple Reviewed on Open Review: https: // openreview. net/ forum? id= pm Rab MH1JW Diffusion denoising has emerged as a powerful approach for modeling data distributions, treating data as particles with their position and velocity modeled by a stochastic diffusion process. While this framework assumes data resides in a fixed vector spaces (e.g., images as pixel-ordered vectors), point clouds present unique challenges due to their unordered representation. Existing point cloud diffusion methods often rely on voxelization to address this issue, but this approach is computationally expensive, with cubically scaling complexity. In this work, we investigate the misalignment between point cloud irregularity and diffusion models, analyzing it through the lens of denoising implicit priors. First, we demonstrate how the unknown permutations inherent in point cloud structures disrupt denoising implicit priors. To address this, we then propose a novel folding-based approach that reorders point clouds into a permutation-invariant grid, enabling diffusion to be performed directly on the structured representation. This construction is exploited both globally and locally. Globally, folded objects can represent point cloud objects in a fixed vector space (like images), therefore it enables us to extend the work of denoising as implicit priors to point clouds. Locally, the folded tokens are efficient and novel token representations that can improve existing transformer-based point cloud diffusion models. Our experiments show that the proposed folding operation integrates effectively with both denoising implicit priors as well as advanced diffusion architectures, such as UNet and Diffusion Transformers (Di Ts). Notably, Di T with locally folded tokens achieves competitive generative performance compared to state-of-the-art models while significantly reducing training and inference costs relative to voxelization-based methods. 1 Introduction Modern representations of images and surfaces are often high-dimensional, encompassing thousands of pixels or 3D points. These representations typically reside on low-dimensional manifolds, either explicitly defined Published in Transactions on Machine Learning Research (08/2025) or inferred. Recent advancements in diffusion models (Ho et al., 2020; Sohl-Dickstein et al., 2015; Song & Ermon, 2019; Song et al., 2021; Peebles & Xie, 2023) have demonstrated their efficacy in capturing implicit data distributions, and are conceptually linked to thermal systems. The forward diffusion process simulates thermal agitation, gradually transforming data into a standard Gaussian; while the reverse process mimics cooling, reconstructing data via Langevin-style probabilistic gradient ascent. However, this thermaldynamics-inspired framework treats data as particles and assumes that the data reside in a structured vector space (e.g., images as pixel-ordered vectors). The irregularity and lack of inherent order in point clouds pose significant challenges to the direct application of diffusion or Langevin dynamics. The problems are illustrated in Fig. 1, where a diffusion process or Langevin dynamics describes the motion of a particle in an arbitrary space. When we consider RGB images of height H and width W as particles, their positions are described by vectors in RH W 3 and updated by diffusion denoising or Langevin dynamics. Each entry in such vectors follows the same permutation across objects, and such a correspondence is preserved during denoising. This correspondence is missing in unordered point clouds. Thus, unlike image particles, the motions of particles that represent the point cloud objects cannot be properly described due to the absence of this structured space. Permutation invariant choices for denoising loss further impose irretrievable permutation matrices on the noisy point clouds. This observation aligns with the findings reported in Zhou et al. (2021) and Mo et al. (2023), where similar challenges were encountered in designing diffusion models for unstructured point clouds. In our work, we build upon these observational results and present a formulated theoretical explanation of their failures based on denoising implicit priors. (a) Structured space with consistent axis. (b) Unstructured space with unknown permutation. Figure 1: (a) Particle motions can be properly described in a shared and structured space (e.g., RGB images with the same height and width). (b) Particle motions cannot be properly described in unstructured spaces (e.g., unknown permutation between unordered point cloud objects) or during unstructured motion (e.g., unknown permutation imposed by permutation invariant denoising). Axes correspondence are noted by colors. The theoretical framework of denoising implicit priors investigates diffusion from its fundamental building blocks: Denoising Autoencoders (DAEs). As first shown in Vincent (2011), a DAE that minimizes the l2 norm between a recovered signal and a clean one is equivalent to an energy model that tries to match its score to that of a non-parametric Parzen density estimator of the data. The theoretical connection between DAEs and score-matching (Vincent, 2011) guarantees that each denoising step moves the data towards higher probability regions. Thus, authentic samples can be drawn from the implicit priors that DAEs learned from l2 noise regression. In this work, we employ the theoretical framework of denoising implicit priors to, for the first time, offer a comprehensive theoretical analysis of diffusion denoising in unstructured spaces. In Sec. 3, we theoretically prove that such unknown permutations breaks the proportional relationship between an l2 denoising residual and the probability score, which are also empirically demonstrated in our experiments (Sec. 5.3). Recent diffusion models for point clouds have opted for voxel grids (Zhou et al., 2021; Zeng et al., 2022; Mo et al., 2023) or triplanes (Shue et al., 2023) as structured representations for the diffusion processes. However, these methods suffers from cubically or quadratically scaling computational complexities as well as quantization errors. In this work, we proposed a compact alternative based on the folding operation (Yang et al., 2018), which reconstructs a permutation-invariant, grid-like representation of the input point cloud. The folded reconstruction can be viewed as a fixed permutation that approximates the original geometry, providing a structured representation for point cloud diffusion. Although folding is frequently employed to reconstruct local point patches in self-supervised learning models (Yu et al., 2022; Chen et al., 2023), prior research has neither examined nor leveraged the permutation-invariant properties of its reconstructions in depth. In our work, we provide a thorough discussion of these properties and demonstrate how they effectively resolve the permutation challenge in point cloud diffusion (see Sec. 4.2). Published in Transactions on Machine Learning Research (08/2025) Our novel framework, Fold Diff, is thus different from previous methods that performs diffusion on a voxelized space that incur cubic-scale computational costs and introduce quantization errors. It directly addresses the permutation challenge within the point clouds with a permutation-equivariant reconstruction, thereby being both minimalist and efficient, while fully inspired by our theoretical contribution and exploiting the folding operation (Yang et al., 2018). Fold Diff is a two-stages framework, with a folding-based stage for learning structured permutations, and a denoising-based or DDPM-based stage for learning generative priors. It enjoys a linearly scaling computational complexity with the number of points in each object during diffusion denoising. The resulting grid-like representation, akin to a Geometry Image (GI) (Gu et al., 2002), pairs effectively with image denoisers to learn implicit priors of the original manifold of 3D objects. By combining image denoisers trained on folded objects with the Langevin-style sampling algorithm from Kadkhodaie & Simoncelli (2021), 3D objects can be sampled from the denoising implicit priors. Additionally, we test our framework on the Denoising Diffusion Probabilistic Models (DDPMs) (Ho et al., 2020) with both UNet (Ronneberger et al., 2014) and Di T (Peebles & Xie, 2023) backbones, and observed qualitatively and quantitatively competitive generative performances at significantly reduced cost. Notably, the use of folded tokens significantly reduces the token count in Di Ts relative to its voxelization-based variant, leading to substantial improvements in training and sampling efficiency for point cloud diffusion. Our main contributions can be summarized as follows: As our core theoretical contribution, we present a thorough analysis of the intractability of applying diffusion or Langevin dynamics to unordered data structures, grounded in the theory of denoising implicit priors. As our core technical contribution, we propose a novel folding-based point cloud diffusion framework, dubbed Fold Diff 1, offering greater efficiency than popular voxelization-based methods while enabling a seamless unification of 2D and 3D diffusion methods. We empirically validate our novel framework on denoising implicit priors, UNet-based DDPMs and Di T-based DDPMs, demonstrating competitive generative performance with a lower training and inferencing cost. 2 Related work Diffusion models and denoising implicit priors. Diffusion models (Ho et al., 2020; Sohl-Dickstein et al., 2015; Song & Ermon, 2019; Song et al., 2021; Peebles & Xie, 2023) have recently emerged as powerful generative models. Standard Denoising Diffusion Probabilistic Models (DDPMs) (Ho et al., 2020) utilize a fixed Markovian forward process that incrementally corrupts the data with Gaussian noise, paired with a learned noise-conditioned denoiser that constructs the reverse Markovian process. DDPMs are theoretically equivalent to stacks of Denoising Autoencoders (DAEs) (Kadkhodaie et al., 2024), and are deeply connected to score matching and probability gradient ascent (Vincent et al., 2008). DAEs estimate the data distribution s score at varying noise levels, enabling denoising implicit priors on the data manifold. In particular, Kadkhodaie & Simoncelli (2021) proposed a stochastic gradient ascent procedure to sample from image DAEs. Our work builds upon this idea, but addresses the unique challenges in extending it for the first time to the point cloud domain. We propose a novel folding-based solution to address point cloud irregularity (unordered data), which is compatible with any variants of diffusion-based modeling (e.g., standard DDPM (Ho et al., 2020), score-based diffusion models (Song & Ermon, 2019; Song et al., 2021; 2023), or flow-based diffusion models (Lipman et al., 2022; Liu et al., 2022)); and also with multiple architectures (e.g., UNet (Ronneberger et al., 2014) or transformers (Peebles & Xie, 2023; Ma et al., 2024)). A parallel line of research investigates relation-conditioned diffusion process (Gao et al., 2023; Xu et al., 2024; Yang et al., 2024), which conceptually aligns with our Di T-fold that models a combinatorial distribution of folded patches. Formally, a relation-conditioned diffusion aims to learn a conditional score function y log qt(yt|R), where y = (x1, x2, ...) is a combinatorial structure (e.g., an image or a point cloud y as an assemble of local patches xi), and R is the correlation matrix for all xi, xj (Yang et al., 2024). To encourage 1Code is available at https://github.com/yzdn13l/Fold Diff. Published in Transactions on Machine Learning Research (08/2025) capturing patch-wise correlations, Gao et al. (2023) integrates latent masking within Di T, while Xu et al. (2024) introduces asynchronous time scheduling for local patches. Notably, Xu et al. (2024) explored the generation of combinatorial 3D semantic parts with the more advanced flow-based diffusion (Ma et al., 2024). While these modifications may provide potential benefits for our folding-based Di T, we maintain an unmodified training scheme to ensure a fair comparison in our experiments. Diffusion models for point clouds. Generative models for point clouds aim to capture distribution (objects in RN 3) of distributions (points in R3). Diffusion process was first adopted in Luo & Hu (2021b) to model the distribution of points in R3 conditioned on a Point Net (Qi et al., 2017a) encoded shape latent. Our work instead examines the diffusion models at the object level, where the particles are point cloud objects in RN 3. Due to the irregularity of point clouds, prior works perform diffusion either on a voxelized space following Point Voxel Diffusion (Zhou et al., 2021), or on a encoded triplane representation following Triplane diffusion (Shue et al., 2023). These two strctured representations are easy to optimize and intuitively analogous to 2D pixels, but they both suffer from quantization errors and higher computational complexity. Comparing to generative models modeled on raw point clouds in RN 3, voxelization-based methods (Zhou et al., 2021; Zeng et al., 2022; Mo et al., 2023) perform diffusion in the voxel space in RC V V V . Here, C denotes the number of feature channels per voxel, and V represents the resolution of the voxel grid along each spatial dimension. Similarly, triplane-based methods (Shue et al., 2023) perform diffusion in the space of encoded triplane latents in R3C H W , where H and W specify the spatial resolution of each triplane feature map of dimension C. Our proposed folded structure fundamentally differs from previous structured representations, offering computational complexity at the order of the original point cloud (RN 3) while being fully compatible with established 2D diffusion architectures. A closely related area of point cloud diffusion is 3D molecule generation, which also faces the challenge of permutation equivariance. In this study, we address only permutation equivariance, deferring SE(3) equivariance better managed by SE(3) diffusion models (Peng et al., 2024; Yim et al., 2023) to future works. Early studies primarily rely on Graph Neural Networks (Liu et al., 2018; Shi et al., 2020; Vignac et al., 2023) to ensure permutation equivariance in molecule graph generation. However, graph-based methods scale poorly for point cloud generation because of the lack of intrinsic connectivity in point clouds, which rules out the application of spectral-based graph diffusion methods (Elhag et al., 2024); and point clouds, often consist of thousands of continuously valued points in R3, exhibiting far greater representational complexity than 3D molecules, which typically involve only tens to hundreds of atoms with limited atom and bond types. Notably, some graph-free techniques generate molecules in a permutation-invariant manner by inferring atomic densities within voxel grids (Masuda et al., 2020), which is less efficient comparing to our approach. Recent works also explore the capability of diffusion models as powerful pretrained models for supervised learning tasks (Xiang et al., 2023; Zheng et al., 2023). In this work, we do not investigate this application in point cloud, but instead focus on the detrimental effects of an unstructured space on the diffusion process. Point cloud denoisers. The unordered nature of point clouds poses a key challenge in applying imagebased deep learning methods to 3D data. Point Net (Qi et al., 2017a) pioneered using permutation invariant operations, e.g., pointwise convolutions followed by channel-wise pooling, to address this. Subsequent works improved the expressiveness (Qi et al., 2017b; Wang et al., 2018) or efficiency (Liu et al., 2019) of this approach. State-of-the-art point cloud denoisers (Rakotosaona et al., 2020; Luo & Hu, 2020; 2021a; de Silva Edirimuni et al., 2023) adopt a similar permutation invariant encoder-decoder architecture, but focus on recovering local surface details rather than global shape of an object. We do not consider these local denoisers since we are interested in modeling the distribution of point cloud objects globally. As described in Sec. 5.3, we performed experiments on object-level denoisers paired with the Chamfer Distance loss. We provide theoretical and empirical analysis on why this object-level denoiser cannot be directly used to sample shapes using denoiser as prior techniques (Sec. 3.1). We solve this problem with the approach here introduced (Sec. 4.1), and demonstrate the empirical applications of these ideas (Sec. 4.3). Published in Transactions on Machine Learning Research (08/2025) Folding-based autoencoders and geometry images. Folding-based decoders (Yang et al., 2018; Groueix et al., 2018; Pang et al., 2020) became a popular design choice due to their expressiveness power. In general, the decoding process simulates a deformation process from a genus-0 primitive surface (i.e., a 2D grid lattice) to a shape that reconstructs the inputting point cloud. This 2D-to-3D deformation is deeply connected with Geometry Images (GIs) (Gu et al., 2002), which aims to simplify shape analysis with imageprocessing tools including CNNs (Zhang et al., 2022; Sinha et al., 2016; 2017; Maron et al., 2017). In our work, folding is adopted to obtain a permutational-invariant order of the inputting point cloud, creating a structured vector space for diffusion systems. Recent works have also explored Geometry Images for 3D shape diffusion in the UV domain (Yan et al., 2024; Elizarov et al., 2025), where the surface of a 3D mesh is segmented into disjoint local patches and arranged into a geometry image in the UV domain. The segmented geometry images can be paired with texture maps and text prompts, enabling textured text-to-3D generation with well-established text-to-image techniques. Though similar in concept, these methods mainly focus on the generation of UV images processed from 3D meshes, while our method focuses on the generation of local or global geometry images learned from 3D point clouds. 3 Denoising implicit priors on point clouds 3.1 Denoising on images and point clouds. We represent the original input data (image or point cloud) as a vector x RN 3. The N dimensions, can be associated to an RGB image of N pixels or a point cloud of N (3D) points. The goal is to efficiently represent the prior probability P(x) (i.e., the manifold on the original space where data lies). In both image and point cloud denoising, it is common to assume that the noisy observation is a distribution conditioned on the clean observation, expressively y = x + z, where y RN 3 is the noisy observation, x RN 3 is the clean observation, and z N(0, σ2IN) is the Gaussian noise modeled as both additive and Gaussian (Kadkhodaie & Simoncelli, 2021). Thus, a denoiser is optimized to approximate the residual z to recover the clean image x given y. If x is drawn from a prior distribution P(x), its created noisy version can be considered a Gaussian convolved distribution P(y) that can be rewritten via marginalization, P(y) = Z P(y|x)P(x)dx = Z g(y x)P(x)dx, (1) where g(z) is the multivariate Gaussian PDF with variance σ2 that models the noise residual z. Even though the additive noise is modeled by a Gaussian distribution in both scenarios, point cloud denoising differs from image denoising in its optimization objectives. A key observation is that point clouds are unstructured data while images are structured. In image denoising, each noisy pixel yi has a one-to-one correspondence with its ground truth clean pixel xi. In this case, optimizing a denoiser fθ(y) = y x is equivalent to minimizing the objective argmin θ Ey P (y) [L( x, x)] , (2) where L is commonly the l2 norm, and x = y fθ(y) is the recovered signal. Unlike images, a clean point cloud is a discrete set of samples from a surface S embedded in 3D space. Hence, each noisy point yi can have multiple origins from the clean surface S, which breaks the point-wise correspondence between noisy and clean point clouds. If we naively treat every point cloud with its given permutation as a vector and supervise denoising with the l2 norm, the point-wise correspondence is still unknown during test time. As reported in Rakotosaona et al. (2020), a denoiser optimized to minimize the objective given in Equation 2 is not recovering the original clean point but an average of all possible candidates that the noisy point originates from. This leads to worse denoising performance as it does not guarantee the average to be a point on the surface. Another option is to minimize the distance from denoised points to the underlying surface through argmin θ Ey P (y) [l( x, S)] . (3) Published in Transactions on Machine Learning Research (08/2025) In practice, we only have access to clean points x as discretized samples from the underlying surface. Proximity to the surface is thus the Euclidean distance between a noisy point yi and its nearest neighbor (NN) in x, |yi NN(yi, x)| . (4) A common loss function l( x, S) (Rakotosaona et al., 2020; Roveri et al., 2018) is then x NN(y, x) 2 2 = fθ(y) (y NN(y, x)) 2 2. (5) This strategy presents two main challenges: (1) nearest neighbors need to be computed online, which is computationally very demanding; and (2) the process of denoising introduces an unknown permutation in the points representations. Next, we discuss the challenges of this permutation matrix. 3.2 Denoising implicit priors in structured and unstructured spaces. Denoising CNNs can be viewed as least squares estimators that recover the true signal by computing the conditional mean of the posterior, ˆx(y) = Z x P(x|y)dx = Z x P(y|x)P(x) P(y) dx, (6) where ˆx(y) is the best (in the l2 sense) approximation of the recovered signal. This solution can be expressed as (Miyasawa, 1961) ˆx(y) = y + σ2 y log P(y). (7) The predicted residual ˆx(y) y is then proportional to the score function y log P(y). This implies that the training objective of a CNN denoiser is equivalent to score matching. The noise residual provides a direction to move up a probability gradient toward a clean image density. Thus, as proposed in Kadkhodaie & Simoncelli (2021), we can gradually converge to the manifold P(x) using Langevin style gradient ascent. Unfortunately, Equation 7 no longer holds when the data is unstructured, and therefore the whole idea of denoiser as prior collapses. To formalize this problem, let s consider a noisy observed point set Y = {xi + zi : i = 1, . . . , N} , where we assume zi N(0, σ2) for all i, and x = (x1, . . . , x N) is the correct vectorization of clean points. Any vectorization of Y can be rewritten as y = Π (x + z) , (8) where Π is a permutation matrix. Since zi N(0, σ2), we have Πz N(0, σ2Π), i.e., z and Πz have the same distribution, and since z is not directly observed, without loss of generality, we can write y = x + z, (9) where x = Πx is a permutation of the original vectorization and z N(0, σ2I). Note that Π = IN N if and only if yi = xi + zi for all i. Suppose we naively vectorize Y to obtain y by a random permutation. Given a permutation Π, we can rewrite the marginal distribution of the observation P(y|Π) and the least squares estimate ˆx(y) as P(y|Π) = Z P(y|x, Π)P(x|Π)dx = Z g(y x)P(x|Π)dx, (10) ˆx(y) = Z x P(x|y, Π)dx = Z x P(y|x, Π)P(x|Π) P(y|Π) dx. (11) The following proposition then follows. Proposition 1. The denoiser residual f(y) = ˆx y is proportional to y log P(y) if and only if Π = IN N. The proof is presented in the Appendix A.1. This proposition implies that Equation 7 can only be applied to a structured data format that always follows its original permutation, which is not generally the case in point cloud denoising, validating the empirical observations in Sec. 5.3. Published in Transactions on Machine Learning Research (08/2025) 4 Fold Diff for 3D shape generation 4.1 Folding Net as reordering module Spatial coordinates Max pooling Latent vector Spatial coordinates Chamfer Distance Spatial coordinates Max pooling Latent vector Sinkhorn Distance Cross attention Spatial coordinates (a) Original Folding Net. (b) Modified Folding Net. Figure 2: Overviews of the original Folding Net and our proposed modifications. To solve the challenge of permutation and irregularity in point cloud denoising, we need a canonical permutation Π for all unordered point clouds x. Folding Net (Yang et al., 2018) aligns with our goal. The original Folding Net architecture is illustrated in Fig. 2(a), which encompasses a permutation-invariant encoder (Qi et al., 2017a) E(x) that outputs a 1D latent vector z, and a folding-based decoder D(z, g) that simulates the z-conditioned deforming forces acting on the predefined 2D grid g and outputs a reconstruction ˆx. Since both z and g are permutation-invariant to the original point cloud, ˆx also inherits this property. Consequently, ˆx has a structured ordering defined by g, enabling us to treat the folded reconstruction as particles suitable for diffusion systems. 4.2 Modified Folding Net In our framework, we introduce two modifications to the original Folding Net. Fig. 2 shows an overview of the original Folding Net and our proposed alternative. First, while Folding Net was originally optimized using Chamfer Distance (CD), we replace it with the Sinkhorn Distance (Cuturi, 2013; Feydy et al., 2019). Since our goal is to use the Folding Net to find a fixed permutation that reconstructs the inputting point clouds, it can be formulated as an optimal transport problem that aims to transport the source unordered points to the target reordered reconstructions, and the Sinkhorn Distance is a differentiable solution. In Appendix A.2, we provide a thorough theoretical analysis on the optimal transport interpretation of Folding Net. Second, we incorporate an attention block following Wen et al. (2020) to improve the reconstruction quality without breaking the permutation-invariant property. As illustrated in Fig. 2(b), the queries are generated by a multilayer perceptron (MLP) that processes the 2D coordinates concatenated with the 1D max-pooled feature. Subsequently, the attention module automatically selects the most informative encoded point-wise features corresponding to these queries, thereby establishing a mapping a permutation from an input point to a specific position on the 2D grid. This process yields integrated features that guide the decoder in producing more precise reconstructions, while the decoder in the original design (Fig. 2(a)) relies solely Published in Transactions on Machine Learning Research (08/2025) Implicit Prior Diffusion Sampling Figure 3: An overview of the proposed Fold Diff framework. Folding operation provides a reordered reconstruction that is permutation invariant to the input. The resulting structured global or local representation can be paired with various 2D diffusion techniques and enables efficient point cloud diffusion sampling without the need for voxelization. on the global feature. Given that the attention mechanism is permutation-equivariant with respect to the queries, and considering that our queries are permutation-invariant features, the permutation-invariance of the reordered reconstruction is preserved. In Appendix A.2, we provide the training details of our local Folding Net, and empirically justify the above modifications with experiments. While the modified Folding Net demonstrated in Fig. 2 is sufficient for locally folded tokens, we use a hierarchical Folding Net for globally folded objects. The architecture is inspired by Point Net++ (Qi et al., 2017b) and the Laplacian Pyramid, and the details are covered in Appendix A.3. The resulting reconstruction is then reshaped into a H W geometry image (GI), where each pixel encodes the (x, y, z) coordinates of the reconstructed geometry. The GI is then naturally compatible with 2D UNet denoiser and UNet-DDPMs trained with l2 noise residuals. 4.3 The Fold Diff framework To address the irregularity in point clouds for diffusion models without incurring the high computational complexities and quantization errors of voxelization-based approaches, we propose Fold Diff (demonstrated in Fig. 3), which is a two-stages pipeline that first trains a folding-based autoencoder and then performs denoising diffusion on folded point clouds. We explore three variants of our framework: (1) 2D UNet denoising implicit priors paired with globally folded objects; (2) 2D UNet-based DDPMs paired with globally folded objects; and (3) Di T-based DDPMs paired with locally folded tokens. Here we define our naming schemes for better clarity and consistency: Globally folded objects/point clouds refers to the structured global H W point cloud reconstructed by a hierarchical Folding Net with the architecture discussed in Appendix A.3. These global representations are primarily paired with our UNet-based methods for 3D point cloud generation. Locally folded tokens/patches refers to the structured local k k point cloud reconstructed by a shared lightweight Folding Net with the architecture discussed in Sec. 4.2. These token representations are primarily paired with our Di T-based methods for 3D point cloud generation. Sampling folded objects from UNet implicit priors. Modeling the manifold of reordered reconstructions instead of unordered inputs circumvents the challenges posed by permutation matrices as shown in Proposition 1. The globally folded reconstructions can be trivially reshaped into an image-like representation (i.e., Geometry Image), making them compatible with 2D CNN-based denoisers. As demonstrated in Equation 7, the predicted noise residual gives us a direction (i.e., score estimation) toward the clean manifold, thereby providing access to the density p(x). A general denoiser can be considered a score approximator Published in Transactions on Machine Learning Research (08/2025) adaptive to different noise levels. It s connection to denoising diffusion provides us a proof-of-concept algorithm to illustrate the necessity of structured permutation. Following Kadkhodaie & Simoncelli (2021), we trained a 2D CNN UNet denoiser (Ronneberger et al., 2014) as our implicit prior. Architectural details of the UNet are covered in Appendix A.3. The training objective is the l2-norm between the denoised residuals and the ground truth additive Gaussian noise on each reordered point. To sample high-probability objects from a denoiser, we follow the Langevin-style stochastic gradient ascent algorithm in Kadkhodaie & Simoncelli (2021). A detailed explanation of the sampling algorithm is included in Appendix A.4. Sampling folded objects from Unet-based DDPMs. Extending this approach, we perform Denoising Diffusion Probabilistic Models (DDPMs) sampling using noise-conditioned UNet denoisers trained on globally folded point clouds. Unlike general denoisers, noise-conditioned UNets are specifically trained to predict the score at fixed noise levels rather than adapting to a wide range of noise levels. This specialization allows for more accurate score estimation, enhancing the sampling quality and diversity. In DDPMs, the forward diffusion process corrupts data x0 by adding Gaussian noise over T timesteps, forming latent variables x1, x2, . . . , x T . The joint distribution of the forward process is defined as q(x1:T |x0) = QT t=1 q(xt|xt 1), where q(xt|xt 1) = N(xt; αtxt 1, (1 αt)I). With the reparametrization trick, samples xt q(xt|xt 1) can be rewritten as αtx0 + 1 αtϵ0, where ϵ0 N(0, I). The reverse process learns a score estimation network pθ(xt 1|xt) such that pθ(x0:T ) = p(x T ) QT t=1 pθ(xt 1|xt), where p(x T ) N(0, I). The parameters θ are optimized to minimize the evidence lower bound (ELBO) on a negative log-likelihood of x0 under pθ(x0:T ). The ELBO then reduces to minimizing the l2-norm between the ground truth noise and the predicted noise up to a irrelevant constant multiplier, which coincides with the denoising score matching results as introduced in Sec 3. Sampling from Di T-based DDPMs with folded tokens. If UNet based diffusion predicts the dynamics of a particle (i.e., an image or a point cloud), Diffusion Transformer (Di T) (Peebles & Xie, 2023) predicts the dynamics of a particle group (i.e., a group of tokens in an image or a point cloud). Here, an indestructible particle is no longer the entire object, but the tokens. In this case, the movement direction (i.e., probability gradient) of each particle (i.e., each token) is dependent not only on noise levels, but also on the interactions within a particle group (i.e., group of tokens), and such interactions are captured by the transformer (Vaswani et al., 2017). The original Di T for images consists of two main components. The first is a pretrained tokenizer, which compresses the high-resolution image x into a lower-resolution latent image z. This latent image is then divided into patch embeddings in the raster order. The second component is the latent diffusion transformer, which replaces the UNet in DDPMs. In the latent diffusion transformer, the l2 noise on each token is predicted conditioned on token interactions captured by self-attention. The original Di T added positional embeddings to the image patch embeddings to retain positional information. We dub our Di T as Di T-fold, where the tokenizer is a lightweight Folding Net as described in Sec. 4.2, and the folded reconstructions serve as our tokens. We applied sinusoidal positional embeddings on the barycenter of each folded token, allowing the transformer to capture the spatial relationship between tokens. Additionally, without the constraints of voxel grids, our approach allows for any integer number of token length L, with each token a k k local geometry image. We considered choices such that Lk2 2N for objects of N points. This design offers greater flexibility in sequence length compared to Di T architectures based on voxelization such as Mo et al. (2023), which is restricted to voxel dimensions V {16, 32, 64}, and patch widths p {2, 4, 8}, resulting in token lengths L = (V/p)3 {23, 43, 83, 163, 323}. In their largest configuration (V = 64, p = 2), the token count reaches 32,768 significantly exceeding the number of points (N = 2048) in the experiments leading to highly inefficient training and inference. Our proposed Fold Diff provides a more compact and efficient alternative to point-voxel representations for point cloud diffusion, despite some computational overhead for training a lightweight folding tokenizer. Such overhead can be minimized by training a universal local patch tokenizer applicable across all object categories. A welloptimized, lightweight tokenizer can then be directly paired with Di Ts of any size to model diverse target distributions efficiently. Aside from the above modifications, our implementations align with the original Di T (Peebles & Xie, 2023). Published in Transactions on Machine Learning Research (08/2025) LION Di T-3D Figure 4: Single-class point cloud generation for airplanes, cars, and chairs. The proposed Di T-fold avoids the quantization error introduced from voxelization, thus generates smoother shapes. For fair comparisons, we employ the standard DDPM for both voxel-based and our folding-based methodologies. However, our proposed Fold Diff framework is also compatible with flow-based diffusion training and sampling techniques, as evidenced by our ablation studies. 5 Experiments 5.1 Experimental setup Datasets. We compare the performance of different methods on single-category 3D shape generation using Shape Net (Chang et al., 2015) chairs, cars, and airplanes as primary datasets. We use the same dataset splits of previous works (Luo & Hu, 2021b; Yang et al., 2019; Zhou et al., 2021; Zeng et al., 2022; Mo et al., 2023). Throughout our experiments, each object contains 2048 uniformly sampled points. During evaluation, both the generated shapes and the reference shapes are inversely transformed with the global mean and variance of the training set. Implementation details. All models are trained with Py Torch. UNet implicit priors and DDPMs are trained on globally folded objects. Di T-based DDPMs use the locally folded tokens with a shared lightweight Folding Net (see Sec. 4.2). Model details are covered in Appendix A.3. Evaluation metrics. Following previous works (Luo & Hu, 2021b; Yang et al., 2019; Zhou et al., 2021; Zeng et al., 2022; Mo et al., 2023), we adopt coverage score (COV) and 1-Nearest-Neighbor classifier Accuracy (1-NNA) to quantitatively evaluate the sampling quality and diversity of different methods. Specifically, 1-NNA evaluates whether the generated distribution is identical to the reference distribution (e.g. test-set distribution). Closer to 50% implies better similarity between distributions, thus capturing both the sampling quality and diversity. COV calculates the proportion of shapes from the reference dataset to be matched with at least one generated shape. Higher COV implies better diversity. 1-NNAs and COVs are computed from the full pair-wise distance matrix of elements from both test set and the sampled set. Chamfer Distance (CD) and Earth Mover s Distance (EMD) were used as our distance metrics to compute 1-NNA and COV. Yang et al. (2019) provided a more detailed discussion over different evaluation metrics. 5.2 3D shape generation We evaluate the generative performance and efficiency of Di T-DDPM-based Fold Diff against previous point cloud generative frameworks, as demonstrated in Table 1. We compared previous works with similar Gflops (Yang et al., 2019; Kim et al., 2020; 2021; Klokov et al., 2020; Luo & Hu, 2021b; Zhou et al., 2021; Mo et al., 2023). Among them, point-voxel architectures (Zhou et al., 2021; Zeng et al., 2022; Mo et al., 2023) demonstrate the most competitive generative performance, with Di T-3D-XL (Mo et al., 2023) recently achieving state-of-the-art results. Since LION (Zeng et al., 2022) has Gflops of 247.38 and Di T-3D-XL(64/2) in Mo et al. (2023) has Gflops of 1.12 104, they are omitted in our experiments for fair comparison. The point-voxel representation provides a structured space for point cloud diffusion, yet suffers from a cubically scaling complexity and quantization errors. In contrast, our folded representation avoids the above restrictions. As illustrated in Fig. 4, our framework reduces quantization noise, producing smoother point Published in Transactions on Machine Learning Research (08/2025) Table 1: Comparison results (%) on shape metrics of our Di T-fold and baseline models. (*) denotes reimplemented performances on 2048 uniformly sampled points. Here we considered the methods with Gflops from 0 to 100. We report the average performance of 3 generative runs. Method Gflops Chair Airplane Car 1-NNA ( ) COV ( ) 1-NNA ( ) COV ( ) 1-NNA ( ) COV ( ) CD EMD CD EMD CD EMD CD EMD CD EMD CD EMD r-GAN Achlioptas et al. (2017) 83.69 99.70 24.27 15.13 98.40 96.79 30.12 14.32 94.46 99.01 19.03 6.539 l-GAN (CD) Achlioptas et al. (2017) 68.58 83.84 41.99 29.31 87.30 93.95 38.52 21.23 66.49 88.78 38.92 23.58 l-GAN (EMD) Achlioptas et al. (2017) 71.90 64.65 38.07 44.86 89.49 76.91 38.27 38.52 71.16 66.19 37.78 45.17 Point Flow Yang et al. (2019) 2.10 62.84 60.57 42.90 50.00 75.68 70.74 47.90 46.41 58.10 56.25 46.88 50.00 Soft Flow Kim et al. (2020) 11.29 59.21 60.05 41.39 47.43 76.05 65.80 46.91 47.90 64.77 60.09 42.90 44.60 Set VAE Kim et al. (2021) 1.61 58.84 60.57 46.83 44.26 76.54 67.65 43.70 48.40 59.94 59.94 49.15 46.59 DPF-Net Klokov et al. (2020) 3.02 62.00 58.53 44.71 48.79 75.18 65.55 46.17 48.89 62.35 54.48 45.74 49.43 DPM Luo & Hu (2021b) 2.09 60.05 74.77 44.86 35.50 76.42 86.91 48.64 33.83 68.89 79.97 44.03 34.94 PVD Zhou et al. (2021) 81.12 57.09 60.87 36.68 49.24 73.82 64.81 48.88 52.09 54.55 53.83 41.19 50.56 PVD* Zhou et al. (2021) 81.12 56.85 55.91 46.68 51.41 73.57 67.97 47.22 48.20 57.79 53.07 47.51 53.30 Di T-3D-S(32/4) Mo et al. (2023) 21.95 56.31 55.82 47.21 50.75 Di T-3D-B(32/4) Mo et al. (2023) 87.42 55.59 54.91 50.09 52.80 Di T-3D-S(32/4)* Mo et al. (2023) 21.95 59.19 55.82 44.96 51.16 78.57 67.86 45.57 47.78 63.99 61.65 40.06 50.28 Di T-fold-S(256/16) (Ours) 12.80 55.48 53.56 48.03 51.69 72.19 66.06 47.28 51.11 58.74 52.26 48.53 53.79 Di T-fold-S(512/9) (Ours) 23.92 55.06 53.25 46.83 53.78 67.24 64.78 50.99 51.97 56.63 52.85 47.73 52.40 clouds than PVD (Zhou et al., 2021), LION (Zeng et al., 2022), and Di T-3D (Mo et al., 2023). Moreover, we observed a generalized generative behavior as covered in Appendix A.5 by comparing a generated shape to its nearest neighbors in the training set. Due to constraints on training budget, we compare the point-voxel framework with our Fold Diff framework under the same model size: Di T-S. We denote different configurations of Di T-fold as Di T-fold(L/k2), where L stands for token length, and each token is a k k local geometry image. We denote different configurations of our baseline, Di T-3D (Mo et al., 2023), as Di T-3D(V /p), where V is the voxel dimension, and p is the patch width. We re-implement point-voxel diffusion methods with similar Gflops including PVD and Di T3D-S for a more comprehensive comparison in Shape Net (Chang et al., 2015) chairs, cars, and airplanes, while the results for other methods are reported as in their original paper. For fairness, we report both the available published performance metrics and re-implemented results (denoted with *). In our default configuration, we use Di T-S with 256 tokens, each containing 16 folded points, which corresponds to 4 4 local geometry images. During training, each token is captured from 16 nearest neighbors of 256 farthest sampled centers. During sampling, tokens are directly inferred and contains 3D position of points. We train the Di T-S with folded tokens for 10,000 epochs with Adam W optimizers using a learning rate of 2 10 4. Following Di T (Peebles & Xie, 2023), we maintain an exponential moving average (EMA) of model weights over training with a decay of 0.9999 and the EMA weights were used during sampling for evaluation. This technique was also used in re-implemented Di T-3D. As shown in Table 1, with similar or lower computational costs, our default configuration Di T-fold-S(256/16) demonstrates superior performance over traditional methods (Yang et al., 2019; Kim et al., 2020; 2021; Klokov et al., 2020; Luo & Hu, 2021b). More importantly, Di T-fold-S(256/16) surpasses the performances of pointvoxel methods, PVD (Zhou et al., 2021) and Di T-3D-S (Mo et al., 2023), with equivalent or less Gflops. In particular, Di T-fold-S(256/16) outperforms re-implemented Di T-3D-S(32/4) with only 1/2 of the Gflops. It also achieves on-par performance on Shape Net chairs with the reported Di T-3D-B(32/4) performance with only 1/8 of the Gflops and smaller model. We further boost the performance of Di T-fold by scaling to 512 tokens. Due to restrictions on our training budget, we didn t scale our models to Di T-B, Di T-L, and Di T-XL. Nevertheless, Fold Diff demonstrates its superiority over the point-voxel diffusion framework. Please refer to Appendix A.6.1 for more visualizations. 5.3 Implicit priors and DDPMs Here we demonstrate the necessity of a structured space in both denoising implicit priors and DDPMs, also showcasing the flexibility of Fold Diff as a general framework. Published in Transactions on Machine Learning Research (08/2025) Table 2: Generative performances of three Fold Diff variants using denoising implicit priors (DIPs) and denoising diffusion probabilistic models (DDPMs) with structured Shape Net objects. UNet-based methods use globally folded reconstructions for training and sampling, while Di T-based methods use locally folded reconstructions. Architecture Method Chair Airplane Car 1-NNA ( ) COV ( ) 1-NNA ( ) COV ( ) 1-NNA ( ) COV ( ) CD EMD CD EMD CD EMD CD EMD CD EMD CD EMD 2D UNet DIP 85.73 87.24 28.85 28.40 96.13 96.09 20.82 22.63 90.62 92.14 34.09 36.93 2D UNet DDPM 59.29 62.54 42.13 45.92 84.44 85.31 39.51 38.52 67.33 70.03 42.05 40.91 Di T-fold DDPM 55.48 53.56 48.03 51.69 72.19 66.06 47.28 51.11 58.74 52.26 48.53 53.79 Table 3: Ablation studies. MMD-CD is multiplied with 1 10 3, MMD-EMD is multiplied with 1 10 2. Folding Configs 1-NNA (%) COV (%) MMD token source attention loss CD EMD CD EMD CD EMD Chair CD 56.38 55.83 47.83 53.45 2.55 1.45 Chair CD 56.35 54.39 47.71 51.90 2.59 1.47 Chair,Airplane,Car CD 55.93 54.15 47.53 52.52 2.51 1.43 Chair,Airplane,Car EMD 56.22 54.76 48.19 53.47 2.55 1.45 Chair,Airplane,Car Sinkhorn 55.48 53.56 48.03 51.69 2.51 1.44 (a) Ablating Folding Nets. Configs 1-NNA (%) COV (%) MMD Method EMA CD EMD CD EMD CD EMD Di T-3D-S(32/4) 68.35 69.18 39.73 40.33 2.90 1.63 Di T-3D-S(32/4) 59.19 55.82 44.96 51.16 2.54 1.49 Di T-fold-S(256/16) 56.85 55.20 45.78 50.60 2.62 1.48 Di T-fold-S(256/16) 55.48 53.56 48.03 51.69 2.51 1.44 (b) Ablating EMA. Reproduced failures in unstructured spaces. In Sec. 3, we demonstrated that the theoretical foundation of denoising implicit priors fails with unknown permutations. It supports the failures reported by Zhou et al. (2021) and Mo et al. (2023) when they perform diffusion on unstructured point clouds. Here we reproduce these empirical results using object-level point cloud denoisers or diffusion models. In total, we experiment on four different options in the unstructured space: (1) Point Net denoising implicit prior, (2) Point Net++ denoising implicit prior, (3) Point Net++ DDPM as reported by Zhou et al. (2021), and (4) Di T-raw DDPM as reported by Mo et al. (2023). Implementation details are covered in Appendix A.3. For the denoising implicit prior methods, the training loss is the surface proximity (Equation 5). For the DDPM methods, the training loss is the mean squared error (MSE) between the predicted clean point clouds and the ground truth clean point clouds. Langevin stochastic gradient ascent fails to generate any visible shape from the denoising prior trained on such unstructured space of point clouds. Similarly, DDPM methods trained on unstructured point clouds also fail to generate any visible shape, agreeing with the empirical findings in Zhou et al. (2021) and Mo et al. (2023). Implicit priors and DDPMs with Fold Diff. The Fold Diff framework offers a structured space for denoising implicit priors. We prepared a dataset of folded objects or folded token groups from Shape Net Chang et al. (2015) chairs, airplanes, and cars. In total, we experiment on three different variants of Fold Diff (as described in Sec. 4.3) in the structured domain: (1) UNet denoising implicit prior with globally folded objects; (2) UNet DDPM with globally folded objects; and (3) Di T DDPM with locally folded tokens. As reported in Table 2, all three variants of Fold Diff are capable of generating authentic shapes, and DDPMs provides better sampling quality and diversity than denoising implicit priors. Among the three variants, Di Ts with locally folded tokens produce the most authentic and diverse shapes comparing to UNets with globally folded shapes. The better performance of the local folding paradigm can be explained by it avoiding rediscovering a globally autoencoded manifold learned by the hierarchical Folding Net. This leaves more freedom for a combinatorial diffusion generation, as Di Ts predict the denoising direction for particle groups instead of fixed particles of the globally folded structure. 5.4 Ablation studies Local Folding Net. Here we ablate our Di T-fold with different local Folding Nets discussed in Sec. 4.2. We evaluate the default Di T-fold-S(256/16) on the chair subset with different Folding Net configurations. Minimum Matching Distance (MMD) were included as additional metrics to evaluate the quality of generated Published in Transactions on Machine Learning Research (08/2025) Configs Sampling 1-NNA (%) COV (%) MMD Backbone Model Speed (s/object) CD EMD CD EMD CD EMD Di T-fold-S DDPM 0.31 55.48 53.56 48.03 51.69 2.51 1.44 Si T-fold-S PF-ODE 0.07 57.10 54.10 47.53 53.47 2.50 1.44 Table 4: The compatibility of Fold Diff with flow-based diffusion. shapes for experiments. Lower MMD implies better fidelity. The experimental results, summarized in Table 3a, reveal several key insights. As observed, introducing attention mechanisms into the local folding tokenizers is necessary for accurately reconstructing local patches, leading to improved overall performance. Changing the optimization goal from the Chamfer Distance (CD) to the Sinkhorn Distance further boosts the performance, as the Sinkhorn Distance is better at continuously approximating a permutation from reconstructions. Moreover, training the local folding tokenizer on a diverse set of object categories (e.g., chairs, airplanes, and cars) enhances both the accuracy and generalization capability of the learned tokens. This diversity proves to be a critical factor in achieving better performance in Di T-fold. Exponential Moving Average. Exponential Moving Average (EMA) is a widely adopted weight averaging technique in deep learning, where an EMA of the raw weights is maintained during training and used for evaluation or inference. This approach typically leads to improved generalization compared to using the raw, last-step weights. On Shape Net chairs, we examined the impact of this technique. As observed in Table 3b, EMA is crucial for enhancing generative quality and diversity for both the baseline (Di T-3D-S) and our model (Di T-fold-S). Notably, without using EMA weights, our model exhibits more robust performance compared to the point-voxel-based baseline. Flow-based Diffusion. Fold Diff is compatible with various diffusion model variants. In this study, we demonstrate its compatibility with Scalable Interpolant Transformers (Si T)(Ma et al., 2024), a recently proposed flow-based diffusion transformer. Flow-matching is an emerging variant of generative diffusion models that facilitates faster sampling via probability flow ordinary differential equation (PF-ODE). As presented in Table 4, we compare the generative performance of both Di T-fold-S and Si T-fold-S models on Shape Net chairs. The results indicate that while both models achieve comparable generative quality, the Si T model significantly outperforms in sampling speed. This finding underscores the efficacy of our proposed method when applied to advanced diffusion models such as Si T. 6 Conclusion In this work, we provide a theoretical explanation to the impact of irregularity in point cloud diffusion. To solve this problem, we propose Fold Diff, a novel framework for 3D point cloud diffusion with linear complexity. By transforming point cloud diffusion from an unstructured space to a reordered structured space, Fold Diff eliminates the need for voxelization and integrates seamlessly with 2D denoising priors and DDPM-based methods. The framework leverages the folding operation to output a reordered reconstruction of the inputting geometry, enabling direct integration with off-the-shelf 2D denoising implicit priors and DDPM-based methods. This facilitates efficient modeling of the distribution of clean 3D objects and highprobability sampling from the learned distribution. Leveraging folded tokens with Diffusion Transformers, Fold Diff outperforms voxel-based approaches in both performance and efficiency, offering a new direction for tokenizers and transformer architectures in point cloud modeling. Future work includes exploring similar frameworks for meshes, as well as integrating texture information into the generative process. Additionally, our evaluations were conducted under a relatively small model setting due to computational constraints (i.e., Di T-S), and while sufficient to show the power of the proposed framework, it mainly lays the groundwork for future research to scale up our results or design optimized Di T variants tailored to the folding operation, potentially unifying approaches for 3D generative modeling with 2D methods. Published in Transactions on Machine Learning Research (08/2025) Acknowledgments Work partially supported by ONR, NSF, Simons Foundation, and gifts/awards from Google, Amazon, and Apple. Panos Achlioptas, Olga Diamanti, Ioannis Mitliagkas, and Leonidas J. Guibas. Learning Representations and Generative Models for 3D Point Clouds. In International Conference on Machine Learning, 2017. Angel X. Chang, Thomas A. Funkhouser, Leonidas J. Guibas, Pat Hanrahan, Qi-Xing Huang, Zimo Li, Silvio Savarese, Manolis Savva, Shuran Song, Hao Su, Jianxiong Xiao, Li Yi, and Fisher Yu. Shape Net: An Information-Rich 3D Model Repository. Co RR, abs/1512.03012, 2015. Guangyan Chen, Meiling Wang, Yi Yang, Kai Yu, Li Yuan, and Yufeng Yue. Point GPT: Auto-regressively Generative Pre-training from Point Clouds. Advances in Neural Information Processing Systems, 36, 2023. Marco Cuturi. Sinkhorn Distances: Lightspeed Computation of Optimal Transport. In C.J. Burges, L. Bottou, M. Welling, Z. Ghahramani, and K.Q. Weinberger (eds.), Advances in Neural Information Processing Systems, volume 26. Curran Associates, Inc., 2013. Dasith de Silva Edirimuni, Xuequan Lu, Zhiwen Shao, Gang Li, Antonio Robles-Kelly, and Ying He. Iterative PFN: True Iterative Point Cloud Filtering. 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13530 13539, 2023. Ahmed Elhag, Yuyang Wang, Josh Susskind, and Miguel Angel Bautista Martin. Manifold diffusion fields. In ICLR, 2024. Slava Elizarov, Ciara Rowles, and Simon Donné. Geometry Image Diffusion: Fast and Data-Efficient Textto-3D with Image-Based Surface Representation. In International Conference on Learning Representations (ICLR), 2025. URL https://openreview.net/pdf?id=Glm7Kj47n N. Haoqiang Fan, Hao Su, and Leonidas J. Guibas. A point set generation network for 3d object reconstruction from a single image. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2463 2471, 2016. Jean Feydy, Thibault Séjourné, François-Xavier Vialard, Shun-ichi Amari, Alain Trouve, and Gabriel Peyré. Interpolating between Optimal Transport and MMD using Sinkhorn Divergences. In The 22nd International Conference on Artificial Intelligence and Statistics, pp. 2681 2690, 2019. Shanghua Gao, Pan Zhou, Mingg-Ming Cheng, and Shuicheng Yan. Masked Diffusion Transformer is a Strong Image Synthesizer. 2023 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 23107 23116, 2023. Thibault Groueix, Matthew Fisher, Vladimir G. Kim, Bryan C. Russell, and Mathieu Aubry. A Papier Mache Approach to Learning 3D Surface Generation. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 216 224, 2018. Xianfeng Gu, Steven J. Gortler, and Hugues Hoppe. Geometry Images. In Proceedings of the 29th annual conference on Computer graphics and interactive techniques, 2002. Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising Diffusion Probabilistic Models. In Advances in Neural Information Processing Systems, volume 33, pp. 6840 6851, 2020. Zahra Kadkhodaie and Eero Simoncelli. Stochastic Solutions for Linear Inverse Problems using the Prior Implicit in a Denoiser. In Advances in Neural Information Processing Systems, volume 34, pp. 13242 13254, 2021. Published in Transactions on Machine Learning Research (08/2025) Zahra Kadkhodaie, Florentin Guth, Eero P Simoncelli, and Stéphane Mallat. Generalization in Diffusion Models Arises from Geometry-adaptive Harmonic Representations. In The Twelfth International Conference on Learning Representations, 2024. Hyeongju Kim, Hyeonseung Lee, Woo Hyun Kang, Joun Yeop Lee, and Nam Soo Kim. Soft Flow: Probabilistic Framework for Normalizing Flow on Manifolds. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 16388 16397. Curran Associates, Inc., 2020. Jinwoo Kim, Jae Hyeon Yoo, Juho Lee, and Seunghoon Hong. Set VAE: Learning Hierarchical Composition for Generative Modeling of Set-Structured Data. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 15054 15063, 2021. Roman Klokov, Edmond Boyer, and Jakob J. Verbeek. Discrete Point Flow Networks for Efficient Point Cloud Generation. In European Conference on Computer Vision, 2020. Yaron Lipman, Ricky T. Q. Chen, Heli Ben-Hamu, Maximilian Nickel, and Matt Le. Flow Matching for Generative Modeling. Ar Xiv, abs/2210.02747, 2022. Qi Liu, Miltiadis Allamanis, Marc Brockschmidt, and Alexander L. Gaunt. Constrained Graph Variational Autoencoders for Molecule Design. In Neural Information Processing Systems, 2018. Xingchao Liu, Chengyue Gong, and Qiang Liu. Flow Straight and Fast: Learning to Generate and Transfer Data with Rectified Flow. Ar Xiv, abs/2209.03003, 2022. Zhijian Liu, Haotian Tang, Yujun Lin, and Song Han. Point-Voxel CNN for Efficient 3D Deep Learning. In Advances in Neural Information Processing Systems, volume 32, 2019. Shitong Luo and Wei Hu. Differentiable Manifold Reconstruction for Point Cloud Denoising. Proceedings of the 28th ACM International Conference on Multimedia, 2020. Shitong Luo and Wei Hu. Score-Based Point Cloud Denoising. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 4563 4572, 2021a. Shitong Luo and Wei Hu. Diffusion Probabilistic Models for 3D Point Cloud Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2021b. Nanye Ma, Mark Goldstein, Michael S Albergo, Nicholas M. Boffi, Eric Vanden-Eijnden, and Saining Xie. Si T: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers. In European Conference on Computer Vision, 2024. Haggai Maron, Meirav Galun, Noam Aigerman, M. Trope, Nadav Dym, Ersin Yumer, Vladimir G. Kim, and Yaron Lipman. Convolutional Neural Networks on Surfaces via Seamless Toric Covers. ACM Transactions on Graphics (TOG), 36:1 10, 2017. Tomohide Masuda, Matthew Ragoza, and David Ryan Koes. Generating 3D Molecular Structures Conditional on a Receptor Binding Site with Deep Generative Models. Ar Xiv, abs/2010.14442, 2020. Koichi Miyasawa. An Empirical Bayes Estimator of the Mean of a Normal Population. Bull. Inst. Internat. Statist., 38:181 188, 1961. Shentong Mo, Enze Xie, Ruihang Chu, Lanqing Hong, Matthias Niessner, and Zhenguo Li. Di T-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation. In A. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, and S. Levine (eds.), Advances in Neural Information Processing Systems, volume 36, pp. 67960 67971. Curran Associates, Inc., 2023. Jiahao Pang, Duanshun Li, and Dong Tian. Tearing Net: Point Cloud Autoencoder to Learn Topology Friendly Representations. In 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7449 7458, 2020. Published in Transactions on Machine Learning Research (08/2025) William Peebles and Saining Xie. Scalable Diffusion Models with Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 4195 4205, October 2023. Yong Peng, Linlin Tang, Qing Liao, Liu Yang, Shuhan Qi, and Jiajia Zhang. SE(3)-Diffusion: An Equivariant Diffusion Model for 3D Point Cloud Generation. International Conference on Genetic and Evolutionary Computing., Jan 2024. doi: 10.1007/978-981-97-0068-4_48. Charles R. Qi, Hao Su, Mo Kaichun, and Leonidas J. Guibas. Point Net: Deep Learning on Point Sets for 3D Classification and Segmentation. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 77 85, 2017a. doi: 10.1109/CVPR.2017.16. Charles R. Qi, Li Yi, Hao Su, and Leonidas J. Guibas. Point Net++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 5105 5114, 2017b. Marie-Julie Rakotosaona, Vittorio La Barbera, Paul Guerrero, Niloy J Mitra, and Maks Ovsjanikov. POINTCLEANNET: Learning to Denoise and Remove Outliers from Dense Point Clouds. In Computer Graphics Forum, volume 39, pp. 185 203, 2020. Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention MICCAI, pp. 234 241, 2014. doi: 10.1007/978-3-319-24574-4_28. Riccardo Roveri, A. Cengiz Öztireli, Ioana Pandele, and Markus H. Gross. Point Pro Nets: Consolidation of Point Clouds with Convolutional Neural Networks. Computer Graphics Forum, 37, 2018. Yossi Rubner, Carlo Tomasi, and Leonidas J. Guibas. The earth mover s distance as a metric for image retrieval. International Journal of Computer Vision, 40:99 121, 2000. Chence Shi, Minkai Xu, Zhaocheng Zhu, Weinan Zhang, Ming Zhang, and Jian Tang. Graph AF: a Flowbased Autoregressive Model for Molecular Graph Generation. Ar Xiv, abs/2001.09382, 2020. J. Ryan Shue, Eric Ryan Chan, Ryan Po, Zachary Ankner, Jiajun Wu, and Gordon Wetzstein. 3D Neural Field Generation Using Triplane Diffusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 20875 20886, June 2023. Ayan Sinha, Jing Bai, and Karthik Ramani. Deep Learning 3D Shape Surfaces Using Geometry Images. In European Conference on Computer Vision, 2016. Ayan Sinha, Asim Unmesh, Qi-Xing Huang, and Karthik Ramani. Surf Net: Generating 3D Shape Surfaces Using Deep Residual Networks. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 791 800, 2017. Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep Unsupervised Learning using Nonequilibrium Thermodynamics. In Proceedings of the 32nd International Conference on Machine Learning, volume 37, pp. 2256 2265, 2015. Yang Song and Stefano Ermon. Generative Modeling by Estimating Gradients of the Data Distribution. In Advances in Neural Information Processing Systems, 2019. Yang Song, Jascha Sohl-Dickstein, Diederik P. Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-Based Generative Modeling through Stochastic Differential Equations. In 9th International Conference on Learning Representations, ICLR 2021, 2021. Yang Song, Prafulla Dhariwal, Mark Chen, and Ilya Sutskever. Consistency Models. In International Conference on Machine Learning, 2023. Ashish Vaswani, Noam M. Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is All you Need. In Advances in Neural Information Processing Systems, 2017. Published in Transactions on Machine Learning Research (08/2025) Clement Vignac, Igor Krawczuk, Antoine Siraudin, Bohan Wang, Volkan Cevher, and Pascal Frossard. Digress: Discrete denoising diffusion for graph generation. In The Eleventh International Conference on Learning Representations, 2023. Pascal Vincent. A Connection Between Score Matching and Denoising Autoencoders. Neural Computation, 23:1661 1674, 2011. Pascal Vincent, H. Larochelle, Yoshua Bengio, and Pierre-Antoine Manzagol. Extracting and Composing Robust Features with Denoising Autoencoders. In International Conference on Machine Learning, 2008. Yue Wang, Yongbin Sun, Ziwei Liu, Sanjay E. Sarma, Michael M. Bronstein, and Justin M. Solomon. Dynamic Graph CNN for Learning on Point Clouds. ACM Transactions on Graphics (TOG), 38:1 12, 2018. Xin Wen, Tianyang Li, Zhizhong Han, and Yu-Shen Liu. Point Cloud Completion by Skip-Attention Network With Hierarchical Folding. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1936 1945, 2020. Weilai Xiang, Hongyu Yang, Di Huang, and Yunhong Wang. Denoising diffusion autoencoders are unified self-supervised learners. 2023 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 15756 15766, 2023. Rui Xu, Jiepeng Wang, Hao Pan, Yang Liu, Xin Tong, Shiqing Xin, Changhe Tu, Taku Komura, and Wenping Wang. Combostoc: Combinatorial stochasticity for diffusion generative models, 2024. URL https://arxiv.org/abs/2405.13729. Xingguang Yan, Han-Hung Lee, Ziyu Wan, and Angel X. Chang. An Object is Worth 64x64 Pixels: Generating 3D Object via Image Diffusion, 2024. URL https://arxiv.org/abs/2408.03178. Guandao Yang, Xun Huang, Zekun Hao, Ming-Yu Liu, Serge J. Belongie, and Bharath Hariharan. Point Flow: 3D Point Cloud Generation With Continuous Normalizing Flows. 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 4540 4549, 2019. Ling Yang, Haotian Qian, Zhilong Zhang, Jingwei Liu, and Bin Cui. Structure-Guided Adversarial Training of Diffusion Models. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7256 7266, 2024. Yaoqing Yang, Chen Feng, Yiru Shen, and Dong Tian. Foldingnet: Point Cloud Auto-encoder via Deep Grid Deformation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 206 215, 2018. Jason Yim, Brian L. Trippe, Valentin De Bortoli, Emile Mathieu, Arnaud Doucet, Regina Barzilay, and Tommi Jaakkola. SE(3) Diffusion Model with Application to Protein Backbone Generation. In Proceedings of the 40th International Conference on Machine Learning, ICML 23. JMLR.org, 2023. Xumin Yu, Lulu Tang, Yongming Rao, Tiejun Huang, Jie Zhou, and Jiwen Lu. Point-BERT: Pre-Training 3D Point Cloud Transformers with Masked Point Modeling. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Xiaohui Zeng, Arash Vahdat, Francis Williams, Zan Gojcic, Or Litany, Sanja Fidler, and Karsten Kreis. LION: Latent Point Diffusion Models for 3D Shape Generation. Ar Xiv, abs/2210.06978, 2022. Qijian Zhang, Junhui Hou, Yu Qian, Yiming Zeng, Juyong Zhang, and Ying He. Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud Analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45:9726 9742, 2022. Xiao Zheng, Xiaoshui Huang, Guofeng Mei, Yuenan Hou, Zhaoyang Lyu, Bo Dai, Wanli Ouyang, and Yongshun Gong. Point Cloud Pre-Training with Diffusion Models. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 22935 22945, 2023. Linqi Zhou, Yilun Du, and Jiajun Wu. 3D Shape Generation and Completion through Point-Voxel Diffusion. 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 5806 5815, 2021. Published in Transactions on Machine Learning Research (08/2025) A.1 Proof of Proposition 1 In Sec. 3.2 we explained why gradient denoising in Equation 7 cannot be naively extended to unstructured data. This is formalized through our Proposition 1 below, which implies that Equation 7 holds only if we pick a permutation of the point set that is exactly equal to a true underlying structure. For unstructured data, such permutation matrix is non-retrievable, thus Equation 7 no longer holds. In this appendix we include the proof of this proposition. Proposition 1. The denoiser residual f(y) = ˆx y is proportional to y log P(y) if and only if Π = IN N. Proof. Starting from Equation 10, we can write y P(y|Π) = σ 2 Z ( x y) g(y x)P(x)dx, = σ 2 Z (Πx y) P(y, x|Π)dx, where we use x = Πx and replace g(y Πx) with P(y|x, Π) to obtain the second row. Dividing by σ 2P(y|Π) gives σ2 y log P(y|Π) = Z Πx P(x|y, Π)dx Z y P(x|y, Π)dx, where in the first row we used y log P(y|Π) = y P (y|Π) P (y|Π) , and in the second row we used Equation 11. This gives us the estimator ˆx(y) as ˆx(y) = ΠT y + σ2 y log P(y|Π) . (12) Therefore, we can evaluate how much f(y) = ˆx y differs from being proportional to y log P(y|Π) as f(y) σ2 y log P(y|Π) = (I Π) ˆx(y), which is 0 if and only if Π = I. A.2 Local Folding Net An optimal transport interpretation of folding as reordering. For better understandings of our proposed folding as reordering, we interpret it as an optimal transport problem that aims to transport the source unordered points to the target reordered reconstructions. Formally, the optimization goal can be written as min Π:x ˆx(θ) i,j xi Πi,jˆxj(θ) 2, (13) where Π is the global matching from the source input x to the target reconstruction ˆx, and θ are the Folding Net parameters. One may notice that the objective is exactly an Earth Moving Distance (EMD) (Rubner et al., 2000), which provides an explicit transport plan from the source to the target. However, EMD is not everywhere differentiable. When multiple permutations of the reconstruction ˆx(θ) give the same optimal transport cost, the gradient over θ is not unique and thus not well-defined. Even though EMD may offer the best reconstruction results, the permutation implicitly learned with θ may not be an ideal structured representation for diffusion (see Table 5). One differentiable approximation to EMD is the Chamfer Distance (CD), which is the original loss function for Folding Net (Yang et al., 2018). CD is still favored due to its better efficiency, as it does not explicitly compute an optimal transport plan, but approximates one via nearest neighbor matching. However, as a pseudo-distance (Fan et al., 2016), CD produces less faithful point cloud reconstructions than EMD (Achlioptas et al., 2017). Published in Transactions on Machine Learning Research (08/2025) Table 5: Folding Net reconstruction performance on small local patches of 16 points with different modifications and the generative performance (1-NNA) on Shape Net chairs. Attention Gflops loss CD(10 3) EMD(10 3) Laplacian(10 1) 1-NNA-CD 1-NNA-EMD 1.6799 CD 3.0586 8.8583 4.0846 57.36 55.83 1.8716 CD 1.8379 5.7414 4.1840 55.93 54.15 1.8716 EMD 1.0152 1.0904 6.0602 56.22 54.76 1.8716 Sinkhorn 1.4429 3.3540 5.0698 55.48 53.56 Due to the above limitations of CD and EMD, we pivoted to the Sinkhorn distance (Cuturi, 2013; Feydy et al., 2019), which is an entropy smoothed differentiable approximation of EMD, min Π:x ˆx(θ) i,j xi Πi,jˆxj(θ) 2 1 λH(Π), (14) where H(Π) = P i,j Πi,j log Πi,j is the entropy of the transport plan Π and λ 0 is the regularization strength. With the entropy regularization introduced, Equation 14 is strictly convex and has a unique minimizer. In practice, we empirically observed that the Sinkhorn distance performs the best in balancing local patch reconstructions and the final generative performance (see Table 5). Training details and experiments. To validate our introduced modifications to the local Folding Net, we report the reconstruction errors in Chamfer Distances (CD) and Earth Moving Distances (EMD) on small local patches. We also compute the Laplacian of local folded patches to measure the smoothness of the local geometry image (GI) representation. Training local patches are extracted from chairs, cars, and airplanes in the Shape Net dataset, where each object contains 2048 uniformly sampled points via farthest point sampling. We also report the generative performance, on Shape Net chairs, of different modifications in 1-Nearest Neighbor Accuracy (1-NNA) computed with CD and EMD. In our default setting, each local patch resembles a 4 4 geometry image that aggregates 16 nearest neighbors of a randomly chosen center point. The models were trained for 1,000 epochs with a batch size of 65,536 and a learning rate annealing from 2 10 4 to 2 10 6 with a cosine schedule. The experimental results in Table 5 demonstrates that our modifications balance the reconstruction quality of folded local patches and boosts the final generative performances with only a 0.2 Gflops overhead. A.3 Implementation details Here we cover the implementation details of each model we used for experiments. We will discuss important components for each model. For more details, please refer to our public repo. Architectural details of unstructured Point Net denoiser. We use a Point Net-based denoiser for denoising implicit priors experiments in unstructured spaces as discussed in Sec. 5.3. The network design follows the segmentation-variant of Point Net (Qi et al., 2017a). The encoder is composed of MLP layers with an input dimension of 3 (spatial coordinates) and output dimensions of [64, 64, 64, 128, 512]. The N 512-dimensional point-wise features then maxpooled over the N-points to achieve a 512-dimensional 1D vector. The 1D vector is then concatenated with the local feature at dimension 128, creating a global-local mixture of point-wise latent code in N (512 + 128). The point-wise latent code is then fed into a decoder of output dimensions [256, 128, 64, 3] to predict the point-wise denoising direction. The loss is computed with the mean-squared-error between the predicted denoising direction and the displacement from the noisy point to the nearest clean point. Architectural details of hierarchical Folding Net. This section continues Sec. 4.2 to provide more architectural and training details of the proposed hierarchical folding autoencoder. We first adopt a stack of Published in Transactions on Machine Learning Research (08/2025) Set Abstraction modules as proposed in Point Net++ (Qi et al., 2017b) to encode information in different levels of resolution. After three Set Abstraction operations, we acquire 3 sets of points in decreasing resolutions N1,N2,N3 and their aggregated features N1 C1,N2 C2,N3 C3. During decoding, the first folding block will reconstruct N3 points as our base geometry image (GI) I0. The last-layer features before reconstruction are fed forward. Then we expand base GI P0 to the same resolution as N2, and use the second folding block that outputs a difference image to add on Up(I0) to achieve a higher resolution GI I1. We iterate this process in our decoder until we get a GI with the same resolution as our inputting point cloud. In our implementation, N1 = 2048, N2 = 512, and N3 = 256. The corresponding resolution of GIs are I0 = (16, 16), I1 = (32, 16), and I2 = (64, 32). Set Abstraction blocks and Folding blocks follow the same design in SA-Net Wen et al. (2020). Note that in our architecture, we follow SA-Net to apply cross-attention layers Vaswani et al. (2017) analogously as skip-connections Ronneberger et al. (2014). This design preserves permutation invariance of resulting GIs, as the queries of the cross-attention layers are permutation-invariant features extracted by the encoder. Note that we train seperate hierarchical foldingnet for each category. This is less efficient than the Di T case where a lightweight folding tokenizer can be trained on multi-class objects as a generalized tokenizer and be used for different Di Ts. Architectural details of 2D UNet. This section continues Sec. 4 to provide more architectural and training details of the UNet (Ronneberger et al., 2014) model we applied to model denoiser implicit priors. The UNet contains 4 pairs of downsampling-upampling modules with feature dimensions [32, 64, 128, 256], each pair connected by skip connections. Each module contains 2 convolutional layers with residual connections. The same architecture is used in all experiments for fair comparison. We first construct a GI dataset converted from point clouds that contains 2048 points farthest sampled from Shape Net (Chang et al., 2015) single-class objects, which further normalized into a bounding-box of [ 1, 1]3 across the dataset. In Kadkhodaie & Simoncelli (2021), the UNet is trained with Gaussian noise with standard deviations drawn from σ U[0, 0.4] (relative to image intensity range [0, 1]). Thus, given our GIs have an intensity range [ 1, 1], we train the UNet with absolute noise intensity σ U[0, 0.8]. The UNet used in the diffusion model have similar feature dimensions and architecture, except each layer is conditioned on diffusion time step t sigmoid(U[0, 1]). Architectural details of Di T-fold and the folding tokenizer. This section continues Sec. 5 and covers implementation details of Di T with our Fold Diff framework. The lightweight folding tokenizer follows the modified architecture as shown in Sec. 4.2. We train a generalized tokenizer across the chair, airplane, car subsets of Shape Net (Chang et al., 2015), and use this tokenizer for each of the following experiments related to Di T. The tokenizer contains a shallow Point Net-based (Qi et al., 2017a) encoder of output dimensions [64,128,128]. The max-pooled 128-dimensional 1D vector is then concatenated with the coordinates of a k k 2D grid, and fed into a MLP to output a point-wise latent code of dimension 128 that mixes the grid information with the patch-wise information. Then we compute the attention scores between the point-wise latent code and the 128-dimensional point-wise feature before max-pooling, further merging point-wise and patch-wise information. Two folding layers then takes the merged point-wise latent as input and output a reordered reconstruction of the inputting local patch. A.4 Langevin gradient ascent details This section continues Sec. 4 to provide a more detailed explanation for the Langevine dynamics sampling algorithm in Kadkhodaie & Simoncelli (2021) which is detailed in Algorithm 1. In the hyperparameters, σ0 is the initial noise level, σL is the convergence threshold, h0 controls the step size of each denoising correction, and β controls the proportion of injected noise in each iteration. As shown in the algorithm, we update noisy image yt after each denoising step with yt 1 + htdt + γtzt. Published in Transactions on Machine Learning Research (08/2025) Here ht controls the step size of the denoising correction dt, while γt controls the magnitude of noise injection after each denoising step. Thus the effective noise variance σ2 t of yt after each iteration is σ2 t = (1 ht)2σ2 t 1 + γ2 t , = (1 βht)2σ2 t 1. (15) In the first expression, the first term is the remaining noise variance after the denoiser correction, and the second term is the additional variance from the injected noise. To ensure convergence, σ2 t can be rewritten as the second expression in Equation 15 by enforcing βht 1, with ht = h0t 1+h0(t 1) given an initial parameter h0 [0, 1], and β [0, 1] that controls the proportion of injected noise. When β = 1, we have γ2 t = 0, which indicates no noise injection; when β = 0, a noise with a variance equivalent to the removed noise is injected. The noise injection amplitude γt is relevant to both ht and β with the expression: γ2 t = (1 βht)2 (1 ht)2 σ2 t 1, = (1 βht)2 (1 ht)2 f(yt 1) 2 /N, (16) where f(yt 1) 2 /N = σ2 t 1 denotes the magnitude of predicted noise variance, and N is the number of pixels in the image. To sample high-probability objects from a denoiser, we first sample from random Gaussian noise y0 N(0, σ2 0I). Then the residual of our 2D UNet denoiser is applied as the score y log p(y). In each iteration, we take a small step toward the suggested direction, thus moving closer to the manifold of clean folded reconstructions, then inject an additional Gaussian perturbation to avoid getting stuck in local maxima (Kadkhodaie & Simoncelli, 2021). Algorithm 1: Coarse-to-fine stochastic ascent method for sampling from the implicit prior of a denoiser, using denoiser residual f(y) = ˆx(y) y. (Kadkhodaie & Simoncelli, 2021) Parameters: σ0, σL, h0, β Initialization: t = 1, draw y0 N(0, σ2 0I) while σt 1 σL do ht = h0t 1+h0(t 1) ; // step size for denoising step dt = f(yt 1) ; // denoising direction σ2 t = ||dt||2 N ; // effective noise variance γ2 t = (1 βht)2 (1 ht)2 σ2 t ; // shrinking noise variance Draw zt N(0, I) ; // sampling Gaussian noise yt yt 1 + htdt + γtzt ; // update data with Langevin dynamics t t + 1 end A.5 Nearest neighbors of generated shapes Fig. 5 demonstrates the top 5 nearest neighbors of a sampled chair using both Chamfer Distance (CD) and Earth Moving Distance (EMD). As one can notice, the sampled chair is novel comparing to its nearest neighbors, showcasing a generalized generative behavior of our Di T-fold. A.6 Qualitative visualizations A.6.1 More visualizations of synthesized shapes Here we show more abundant synthesized shapes from Di T-fold trained on Shape Net Chang et al. (2015) Airplane (Fig. 6), Car (Fig. 7), and Chair (Fig. 8). Published in Transactions on Machine Learning Research (08/2025) Chair Top 5 nearest shapes Figure 5: Top 5 nearest neighbors of the sampled chair from the training set. Top: nearest neighbors of Chamfer Distance (CD). Bottom: nearest neighbors of Earth Moving Distance (EMD). Figure 6: Synthesized airplanes using Di T based on the proposed Fold Diff framework. Published in Transactions on Machine Learning Research (08/2025) Figure 7: Synthesized cars using Di T based on the proposed Fold Diff framework. Published in Transactions on Machine Learning Research (08/2025) Figure 8: Synthesized chairs using Di T based on the proposed Fold Diff framework.