# gaussian_mixture_solvers_for_diffusion_models__e6073dfb.pdf Gaussian Mixture Solvers for Diffusion Models Hanzhong Guo 1,3, Cheng Lu4, Fan Bao4, Tianyu Pang2, Shuicheng Yan2, Chao Du 2, Chongxuan Li 1,3 1Gaoling School of Artificial Intelligence, Renmin University of China 2Sea AI Lab, Singapore 3Beijing Key Laboratory of Big Data Management and Analysis Methods, Beijing, China 4Tsinghua University {allanguo, tianyupang, yansc, duchao}@sea.com; lucheng.lc15@gmail.com; bf19@mails.tsinghua.edu.cn; chongxuanli@ruc.edu.cn Recently, diffusion models have achieved great success in generative tasks. Sampling from diffusion models is equivalent to solving the reverse diffusion stochastic differential equations (SDEs) or the corresponding probability flow ordinary differential equations (ODEs). In comparison, SDE-based solvers can generate samples of higher quality and are suited for image translation tasks like stroke-based synthesis. During inference, however, existing SDE-based solvers are severely constrained by the efficiency-effectiveness dilemma. Our investigation suggests that this is because the Gaussian assumption in the reverse transition kernel is frequently violated (even in the case of simple mixture data) given a limited number of discretization steps. To overcome this limitation, we introduce a novel class of SDE-based solvers called Gaussian Mixture Solvers (GMS) for diffusion models. Our solver estimates the first three-order moments and optimizes the parameters of a Gaussian mixture transition kernel using generalized methods of moments in each step during sampling. Empirically, our solver outperforms numerous SDE-based solvers in terms of sample quality in image generation and stroke-based synthesis in various diffusion models, which validates the motivation and effectiveness of GMS. Our code is available at https://github.com/Guohanzhong/GMS. 1 Introduction In recent years, deep generative models and especially (score-based) diffusion models [35, 14, 37, 18] have made remarkable progress in various domains, including image generation [15, 7], audio generation [22, 32], video generation [16], 3D object generation [31], multi-modal generation [42, 5, 12, 43], and several downstream tasks such as image translation [28, 45] and image restoration [19]. Sampling from diffusion models can be interpreted as solving the reverse-time diffusion stochastic differential equations (SDEs) or their corresponding probability flow ordinary differential equations (ODEs) [37]. SDE-based and ODE-based solvers of diffusion models have very different properties and application scenarios In particular, SDE-based solvers usually perform better when given a sufficient number of discretization steps [18]. Indeed, a recent empirical study [24] suggests that SDE-based solvers can potentially generate high-fidelity samples with realistic details and intricate semantic coherence from pre-trained large-scale text-to-image diffusion models [34]. Besides, SDEbased solvers are preferable in many downstream tasks such as stroke-based synthesis [27], image translation [45], and image manipulation [20]. Work done during an internship at Sea AI Lab. Correspondence to Chao Du and Chongxuan Li. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). (a) #Step=5 (b) #Step=10 (c) #Step=20 Figure 1: Sampling on a mixture of Gaussian. GMS (ours) and SN-DDPM excel in fitting the true distribution when the transition kernel is Gaussian, with a sufficient number of sampling steps (c). However, GMS outperforms SN-DDPM [2] when sampling steps are limited and the reverse transition kernel deviates from Gaussian (a-b). Despite their wide applications, SDE-based solvers face a significant trade-off between efficiency and effectiveness during sampling, since an insufficient number of steps lead to larger discretization errors. To this end, Bao et al. [3] estimate the optimal variance of the Gaussian transition kernel in the reverse process instead of using a handcraft variance to reduce the discretization errors. Additionally, Bao et al. [2] explore the optimal diagonal variance when dealing with an imperfect noise network and demonstrate state-of-the-art (SOTA) performance with a few steps among other SDE-based solvers. Notably, these solvers all assume that the transition kernel in the reverse process is Gaussian. In this paper, we systematically examine the assumption of the Gaussian transition kernel and reveal that it can be easily violated under a limited number of discretization steps even in the case of simple mixture data. To this end, we propose a new type of SDE-based solver called Gaussian Mixture Solver (GMS), which employs a more expressive Gaussian mixture transition kernel in the reverse process for better approximation given a limited number of steps (see visualization results on toy data in Fig. 1). In particular, we first learn a noise prediction network with multiple heads that estimate the high-order moments of the true reverse transition kernel respectively. For sampling, we fit a Gaussian mixture transition kernel in each step via the generalized methods of the moments [11] using the predicted high-order moments of the true reverse process. To evaluate GMS, we compare it against a variety of baselines [14, 2, 17, 36] in terms of several widely used metrics (particularly sample quality measured by FID) in the tasks of generation and stroke-based synthesis on multiple datasets. Our results show that GMS outperforms state-of-the-art SDE-based solvers [14, 2, 17] in terms of sample quality with the limited number of discretization steps (e.g., < 100). For instance, GMS improves the FID by 4.44 over the SOTA SDE-based solver [2] given 10 steps on CIFAR10. Furthermore, We evaluate GMS on a stroke-based synthesis task. The findings consistently reveal that GMS achieves higher levels of realism than all aforementioned SDE-based solvers as well as the widely adopted ODE-based solver DDIM [36] while maintaining comparable computation budgets and faithfulness scores (measured by L2 distance). Such empirical findings validate the motivation and effectiveness of GMS. 2 Background In this section, we provide a brief overview of the (score-based) diffusion models, representative SDE-based solvers for diffusion models, and applications of such solvers. 2.1 Diffusion models Diffusion models gradually perturb data with a forward diffusion process and then learn to reverse such process to recover the data distribution. Formally, let x0 Rn be a random variable with unknown data distribution q(x0). Diffusion models define the forward process {xt}t [0,1] indexed by time t, which perturbs the data by adding Gaussian noise to x0 with q(xt|x0) = N(xt|a(t)x0, σ2(t)I). (1) In general, the function a(t) and σ(t) are selected so that the logarithmic signal-to-noise ratio log a2(t) σ2(t) decreases monotonically with time t, causing the data to diffuse towards random Gaussian noise [21]. Furthermore, it has been demonstrated by Kingma et al. [21] that the following SDE shares an identical transition distribution qt|0(xt|x0) with Eq. (1): dxt = f(t)xtdt + g(t)dω, x0 q(x0), (2) where ω Rn is a standard Wiener process and f(t) = d log a(t) dt , g2(t) = dσ2(t) dt 2σ2(t)d log a(t) Let q(xt) be the marginal distribution of the above SDE at time t. Its reversal process can be described by a corresponding continuous SDE which recovers the data distribution [37]: dx = f(t)xt g2(t) xt log q(xt) dt + g(t)d ω, x1 q(x1), (4) where ω Rn is a reverse-time standard Wiener process. The only unknown term in Eq. (4) is the score function xt log q(xt). To estimate it, existing works [14, 37, 18] train a noise network ϵθ(xt, t) to obtain a scaled score function σ(t) xt log q(xt) using denoising score matching [38]: L(θ) = R 1 0 w(t)Eq(x0)Eq(ϵ)[ ϵθ(xt, t) ϵ 2 2]dt, (5) where w(t) is a weighting function, q(ϵ) is standard Gaussian distribution and xt q(xt|x0) follows Eq. (1). The optimal solution of the optimization objective Eq. (5) is ϵθ(xt, t) = σ(t) xt log q(xt). Hence, samples can be obtained by initiating the process with a standard Gaussian variable x1, then substituting xt log q(xt) with ϵθ(xt,t) σ(t) and discretizing reverse SDE Eq. (4) from t = 1 to t = 0 to generate x0. 2.2 SDE-based solvers for diffusion models The primary objective of SDE-based solvers lies in decreasing discretization error and therefore minimizing function evaluations required for convergence during the process of discretizing Eq. (4). Discretizing the reverse SDE in Eq. (4) is equivalent to sample from a Markov chain p(x0:1) = p(x1) Q ti 1,ti St p(xti 1|xti) with its trajectory St = [0, t1, t2, ..., ti, .., 1]. Song et al. [37] proves that the conventional ancestral sampling technique used in the DPMs [14] that models p(xti 1|xti) as a Gaussian distribution, can be perceived as a first-order solver for the reverse SDE in Eq. (4). Bao et al. [3] finds that the optimal variance of p(xti 1|xti) N(xti 1|µti 1|ti, Σti 1|ti(xti)) is Σ ti 1|ti(xti) = λ2 ti + γ2 ti σ2(ti) Eq(x0|xti)[ϵθ(xti, ti)] 2 where, γti = p σ2(ti 1) λ2 ti q α(ti) σ2(ti), λ2 ti is the variance of q(xti 1|xti, x0). Analytic DPM [3] offers a significant reduction in discretization error during sampling and achieves faster convergence with fewer steps. Moreover, SN-DDPM [2] employs a Gaussian transition kernel with an optimal diagonal covariance instead of an isotropic covariance. This approach yields improved sample quality and likelihood compared to other SDE-based solvers within a limited number of steps. 2.3 Applications of SDE-based solvers for stroke-based synthesis The stroke-based image synthesis is a representative downstream task suitable for SDE-based solvers. It involves the user providing a full-resolution image x(g) through the manipulation of RGB pixels, referred to as the guided image. The guided image x(g) possibly contains three levels of guidance: a high-level guide consisting of coarse colored strokes, a mid-level guide comprising colored strokes on a real image, and a low-level guide containing image patches from a target image. SDEdit [27] solves the task by first starting from the guided image x(g) and adding Gaussian noise to disturb the guided images to xt and q(x(g)(t0)|x(g)) N(xt0|a(t0)x(g), σ2(t)I) same with Eq. (1). Subsequently, it solves the corresponding reverse stochastic differential equation (SDE) up to t = 0 to generate the synthesized image x(0) discretizing Eq. (4). Apart from the discretization steps taken by the SDE-based solver, the key hyper-parameter for SDEdit is t0, the time step from which we begin the image synthesis procedure in the reverse SDE. (a) logarithm mean of L2-norm (b) logarithm median of L2-norm Figure 2: Empirical evidence of suboptimality of Gaussian kernel on CIFAR10. (a) and (b) plot the logarithm of the image-wise mean and median of L2-norm between Gaussian-assumed third-order moments and estimated third-order moments. Clearly, as the number of sampling steps decreases, the disparity between the following two third-order moments increases, denoting that the true transition kernel diverges further from the Gaussian distribution. See Appendix D.2 for more details. 3 Gaussian mixture solvers for diffusion models In this section, we first show through both theoretical and empirical evidence, that the true reverse transition kernel can significantly diverge from Gaussian distributions as assumed in previous SOTA SDE-based solvers [3, 2], indicating that the reverse transition can be improved further by employing more flexible distributions (see Sec. 3.1). This motivates us to propose a novel class of SDE-based solvers, dubbed Gaussian Mixture Solvers, which determines a Gaussian mixture distribution via the generalized methods of the moments [11] using higher-order moments information for the true reverse transition (see Sec. 3.2). The higher-order moments are estimated by a noise prediction network with multiple heads on training data, as detailed in Sec. 3.3. 3.1 Suboptimality of Gaussian distributions for reverse transition kernels As described in Sec. 2, existing state-of-the-art SDE-based solvers for DPMs [3, 2] approximate the reverse transition q(xti 1|xti) using Gaussian distributions. Such approximations work well when the number of discretization steps in these solvers is large (e.g., 1000 steps). However, for smaller discretization steps (such as when faster sampling is required), the validity of this Gaussian assumption will be largely broken. We demonstrate this theoretically and empirically below. First, observe that q(xs|xt)1 can be expressed as q(xs|xt) = R q(xs|xt, x0)q(x0|xt)dx0, which is non-Gaussian for a general data distribution q(x0). For instance, for a mixture of Gaussian or a mixture of Dirac q(x0), we can prove that the conditional distributions q(xs|xt) in the reverse process are non-Gaussian, as characterized by Proposition 3.1, proven in Appendix A.1. Proposition 3.1 (Mixture Data Have non-Gaussian Reverse Kernel). Assume q(x0) is a mixture of Dirac or a mixture of Gaussian distribution and the forward process is defined in Eq. (1). The reverse transition kernel q(xs|xt), s < t is a Gaussian mixture instead of a Gaussian. Empirically, we do not know the distributions of the real data (e.g., high-dimensional images) and cannot obtain an explicit form of q(xs|xt). However, even in such cases, it is easy to validate that q(xs|xt) are non-Gaussian. In particular, note that the third-order moment of one Gaussian distribution (M (G) 3 ) can be represented by its first-order moment (M1) and its second-order moment (M2) 2, which motivates us to check whether the first three order moments of q(xs|xt) satisfy the relationship induced by the Gaussian assumption. We perform an experiment on CIFAR-10 and estimate the first three orders of moments ˆ M1, ˆ M2, ˆ M3 for q(xs|xt) from data by high-order noise networks (see details in Sec. D.2). As shown in Fig. 2, we plot the mean and median of l2-norm 1To enhance the clarity of exposition, we introduce the notation s .= ti 1 and t .= ti to denote two adjacent time steps in the trajectory. Consequently, we refer to q(xs|xt) as the reverse transition kernel in this context. 2The scalar case is M (G) 3 = M 3 1 + 3M1(M2 M 2 1 ). See Appendix D.2 for the vector case. between the estimated third-order moment ˆ M3 and third-order moment M (G) 3 calculated under the Gaussian assumption given the different number of steps at different time steps t. In particular, when the number of steps is large (i.e., #Step=1000), the difference between the two calculation methods is small. As the number of steps decreases, the l2-norm increases, indicating that the true reverse distribution q(xs|xt) is non-Gaussian. With the time step closer to t = 0, the l2-norm increases too. Both the theoretical and empirical results motivate us to weaken the Gaussian assumption in SDEbased solvers for better performance, especially when the step size is large. 3.2 Sampling with Gaussian mixture transition kernel There are extensive choices of p(xs|xt) such that it is more powerful and potentially fits q(xs|xt) better than a Gaussian. In this paper, we choose a simple mixture of the Gaussian model as follows: i=1 wi N(xs|µi(xt), Σi(xt)), i=1 wi = 1, (7) where wi is a scalar and µi(xt) and Σi(xt) are vectors. The reasons for choosing a Gaussian mixture model are three-fold. First, a Gaussian mixture model is multi-modal (e.g., see Proposition 3.1), potentially leading to a better performance than Gaussian with few steps. Second, when the number of steps is large and the Gaussian is nearly optimal [35, 3], a designed Gaussian mixture model such as our proposed kernel in Eq. (9) can degenerate to a Gaussian when the mean of two components are same, making the performance unchanged. Third, a Gaussian mixture model is relatively easy to sample. For the sake of completeness, we also discuss other distribution families in Appendix C. Traditionally, the EM algorithm is employed for estimating the Gaussian mixture [26]. However, it is nontrivial to apply EM here because we need to learn the reverse transition kernel in Eq. (7) for all time step pairs (s, t) by individual EM processes where we need to sample multiple xs3 given a xt to estimate the parameters in the Gaussian mixture indexed by (s, t). This is time-consuming, especially in a high-dimensional space (e.g., natural images) and we present an efficient approach. For improved flexibility in sampling and training, diffusion models introduce the parameterization of the noise network ϵ(xt, t) [14] or the data prediction network x0(xt, t) [18]. With such a network, the moments under the q(xs|xt) measure can be decomposed into moments under the q(x0|xt) measure so that sampling any xt to xs requires only a network whose inputs are xt and t, such as the decomposition shown in Eq. (10). In previous studies, Gaussian transition kernel was utilized, allowing for the distribution to be directly determined after obtaining the estimated first-order moment and handcrafted second-order moment [14] or estimated first-order and second-order moment [3]. In contrast, such a feature is not available for the Gaussian mixtures transition kernel in our paper. Here we present the method to determine the Gaussian mixture given a set of moments and we will discuss how to obtain the moments by the parameterization of the noise network in Sec. 3.3. Assume the length of the estimated moments set is N and the number of parameters in the Gaussian mixture is d. We adopt a popular and theoretically sound method called the generalized method of moments (GMM) [11] to learn the parameters by: min θ Q(θ, M1, ..., MN) = min θ [ 1 i=1 g(xi, θ)]T W[ 1 i=1 g(xi, θ)], (8) where θ Rd 1 includes all parameters (e.g., the mean of each component) in the Gaussian mixture defined in Eq. (7). For instance, d = 2M Ddata+M 1 in Eq. (7), where Ddata represents the number of dimensions of the data and θ contains M mean vectors with Ddata dimensions, M variance vectors of with Ddata dimensions (considering only the diagonal elements), and M 1 weight coefficients which are scalar. The component of g(xi, θ) RN 1 is defined by gn(xi, θ) = Mn(xi) M (GM) n (θ), where Mn(xi) is the n-th order empirical moments stated in Sec. 3.3 and M (GM) n (θ) is the n-th order moments of Gaussian mixture under θ, and W is a weighted matrix, Nc is the number of samples. Theoretically, the parameter ˆθGMM obtained by GMM in Eq. (8) consistently converges to the potential optimal parameters θ for Gaussian mixture models given the moments condition because Nc(ˆθGMM θ ) d N(0, V(θGMM)), as stated in Theorem 3.1 in Hansen [11]. 3Note that xs serves as the training sample in the corresponding EM process. Hence, we can employ GMM to determine a Gaussian mixture transition kernel after estimating the moments of the transition kernel. To strike a balance between computational tractability and expressive power, we specifically focus on the first three-order moments in this work and define a Gaussian mixture transition kernel with two components shown in Eq. (9) whose vectors µ(1) t , µ(2) t , and σ2 t are parameters to be optimized. This degree of simplification is acceptable in terms of its impact. Intuitively, such a selection has the potential to encompass exponential modes throughout the entire trajectory. Empirically, we consistently observe that utilizing a bimodal Gaussian mixture yields favorable outcomes across all experimental configurations. p(xs|xt) = 1 3N(µ(1) t , σ2 t ) + 2 3N(µ(2) t , σ2 t ), (9) meanwhile, under the simplification in our paper, the number of parameters d = 3 Ddata in our Gaussian transition kernel is equal to the number of moments condition N = 3 Ddata. According to proposition 3.2, under the selection of arbitrary weighted weights, the asymptotic mean (asymptotically consistent) and asymptotic variance of the estimator remain consistent, proof in Appendix A.2. Hence, any choice of weighted weights is optimal. To further streamline optimization, we set W = I. Proposition 3.2 (Any weighted matrix is optimal for d = N). Assume the number of parameters d equals the number of moments condition N, and ϵn θ (xt, t) p Eq(x0|xt)[diag(ϵ n 1 ϵ)] (where n denotes n-fold outers product) which denotes n-th order noise network converging in probability. The asymptotic variance V(θGMM) and the convergence speed of GMM remain the same no matter which weighted matrix is adopted in Eq. (8). Namely, any weighted matrix is optimal. What s more, we provide a detailed discussion on the selection of parameters such as the choice of different parameters to optimize and the different choices of weight wi in the Gaussian mixture in Appendix E.1. Combining with Sec. 4.1, our empirical findings illustrate the efficacy of the GMS across various benchmarks via using the Gaussian transition kernel in Eq. (9) fitted by the objective function shown in Eq. (31) in Appendix B.1 via the ADAN [41] as optimization method. Details regarding the parameters for this optimizer are provided in Appendix E.5. 3.3 Estimating high-order moments for non-Gaussian reverse process In Sec. 3.2, we have explored the methodology for determining a Gaussian mixture transition kernel given a set of moments M1, ..., Mn. In the subsequent section, we will present an approach for estimating these moments utilizing noise networks and elucidate the process of network learning. Given the forward process described by Eq. (1), it can be inferred that both q(xt|xs) and q(xs|xt, x0) follow Gaussian distributions. Specifically, q(xt|xs) N(xt|at|sxs, σ2 t|s I) [21], where at|s = α(t) α(s) and σ2 t|s = σ2(t) a2 t|sσ2(s). Consequently, we can deduce that Eq(xs|xt)[xs n xs] = Eq(x0|xt)q(xs|xt,x0)[xs n xs], where n denotes n-fold outer product. Thus, we can first utilize the Gaussian property of q(xs|xt, x0) and employ an explicit formula to calculate the moments under the measure of q(xs|xt, x0). What s more, we only consider the diagonal elements of higher-order moments for computational efficiency, similar to Bao et al. [2] since estimating full higher-order moments results in escalated output dimensions (e.g., quadratic growth for covariance and cubic for the third-order moments) and thus requires substantial computational demands. The expression of diagonal elements of the third-order moment of the reverse transition kernel can be derived as: ˆ M3 = Eq(xs|xt)[diag(xs xs xs)] = Eq(x0|xt)Eq(xs|xt,x0)[diag(xs xs xs)] = (10) [(at|sσ2 s σ2 t )3diag(xt xt xt) + 3λ2 t at|sσ2 s σ2 t xt] | {z } Constant term + [ 3a2 t|sσ4 sa2 s|0β2 t|s σ8 t (diag(xt xt)) + as|0βt|s σ2 t I] Eq(x0|xt)[x0] | {z } Linear term in x0 + 3at|sσ2 s σ2 t (as|0βt|s σ2 t )2xt Eq(x0|xt)[diag(x0 x0)] | {z } Quadratic term in x0 + (as|0βt|s σ2 t )3Eq(x0|xt)[diag(x0 x0 x0)] | {z } Cubic term in x0 Algorithm 1 Learning of the high order noise network 1: Input: The n-th order noise network ϵn θ,ϕn(xt, t) and its tunable parameter ϕn 2: repeat 3: x0 q(x0) 4: t Uniform([1, ..., T]) 5: ϵt N(0, I) 6: xt = α(t)x0 + σ(t)ϵt 7: Take a gradient step on ϕ(t) n ϕn ϵn t ϵn θ,ϕn(xt, t) 2 2 8: until converged Algorithm 2 Sampling via GMS with the first three order moments 1: Input: The assembled noise network f 3(xt, t) in Eq. (13) 2: x T N(0, I) 3: for t = T to 1 do 4: zt N(0, I) 5: ˆ M1, ˆ M2, ˆ M3 = h(f 3 [1](xt, t), f 3 [2](xt, t), f 3 [3](xt, t)) in Appendix B 6: π k, µ k, σ k = minπk,µk,σk Q(πk, µk, σk, ˆ M1, ˆ M2, ˆ M3) in Sec. 3.2 7: i π k, xt 1 = µ i + σ i zt 8: end for 9: return x0 where is the outer product is the Hadamard product. Additionally, λ2 t corresponds to the variance of q(xs|xt, x0), and further elaboration on this matter can be found in Appendix B. In order to compute the n-th order moment ˆ Mn, as exemplified by the third-order moment in Eq. (10), it is necessary to evaluate the expectations Eq(x0|xt)[x0], . . . , Eq(x0|xt)[diag(x0 n 1 x0)]. Furthermore, by expressing x0 as x0 = 1 α(t)(xt σ(t)ϵ), we can decompose ˆ Mn into a combination of terms Eq(x0|xt)[ϵ], . . . , Eq(x0|xt)[(ϵ n 1 ϵ)], which are learned by neural networks. The decomposition of third-order moments ˆ M3, as outlined in Eq. (29), is provided to illustrate this concept. Therefore in training, we learn several neural networks {ϵn θ }N n=1 by training on the following objective functions: min {ϵn θ } N n=1 Et Eq(x0)q(ϵ) ϵn ϵn θ (xt, t) 2 2 , (11) where ϵ N(0, I), xt = α(t)x0+σ(t)ϵ, and ϵn denotes ϵ n 1ϵ. After training, we can use {ϵn θ }N n=1 to replace {Eq(x0|xt)[ϵ n 1 ϵ]}N n=1 and estimate the moments { ˆ Mn}N n=1 of reverse transition kernel. However, in the present scenario, it is necessary to infer the network a minimum of n times in order to make a single step of GMS. To mitigate the high cost of the aforementioned overhead in sampling, we adopt the two-stage learning approach proposed by Bao et al. [2]. Specifically, in the first stage, we optimize the noise network ϵθ(xt, t) by minimizing the expression min Et Eq(x0)q(ϵ) ϵ ϵθ(xt, t) 2 2 or by utilizing a pre-trained noise network as proposed by [14, 37]. In the second stage, we utilize the optimized network as the backbone and keep its parameters θ fixed, while adding additional heads to generate the n-th order noise network ϵn θ,ϕn (xt, t). ϵn θ,ϕn (xt, t) = NN (ϵθ (xt, t) , ϕn) , (12) where NN is the extra head, which is a small network such as convolution layers or small attention block, parameterized by ϕn, details in Appendix E.3. We present the second stage learning procedure in Algorithm 1. Upon training all the heads, we can readily concatenate the outputs of different heads, as the backbone of the higher-order noise network is shared. By doing so, we obtain the assembled noise network f N (xt, t). When estimating the k-th order moments, it suffices to extract only the first k components of the assembled noise network. f N (xt, t) = concat([ϵθ(xt, t), ϵ2 θ,ϕ2 (xt, t) , .., ϵk θ,ϕk (xt, t) | {z } Required for estimating ˆ Mk , ..., ϵN θ,ϕN (xt, t)]), (13) Table 1: Comparison with competitive SDE-based solvers w.r.t. FID score on CIFAR10 and Image Net 64 64. our GMS outperforms existing SDE-based solvers in most cases. SN-DDPM denotes Extended Analytic DPM from Bao et al. [2]. CIFAR10 (LS) # TIMESTEPS K 10 20 25 40 50 100 200 1000 DDPM, βt 43.14 25.28 21.63 15.24 15.21 10.94 8.23 5.11 DDPM, βt 233.41 168.22 125.05 82.31 66.28 31.36 12.96 3.04 SN-DDPM 21.87 8.32 6.91 4.99 4.58 3.74 3.34 3.71 GMS (OURS) 17.43 7.18 5.96 4.52 4.16 3.26 3.01 2.76 CIFAR10 (CS) IMAGENET 64 64 # TIMESTEPS K 10 25 50 100 200 1000 25 50 100 200 400 4000 DDPM, βt 34.76 16.18 11.11 8.38 6.66 4.92 29.21 21.71 19.12 17.81 17.48 16.55 DDPM, βt 205.31 84.71 37.35 14.81 5.74 3.34 170.28 83.86 45.04 28.39 21.38 16.38 SN-DDPM 16.33 6.05 4.19 3.83 3.72 4.08 27.58 20.74 18.04 16.72 16.37 16.22 GMS (OURS) 13.80 5.48 4.00 3.46 3.34 4.23 26.50 20.13 17.29 16.60 15.98 15.79 To this end, we outline the GMS sampling process in Algorithm 2, where f 3 [2](xt, t) represents concat([ϵθ(xt, t), ϵ2 θ,ϕ2 (xt, t)]). In comparison to existing methods with the same network structure, we report the additional memory cost of the assembled noise network in Appendix E.6. 4 Experiment In this section, we first illustrate that GMS exhibits superior sample quality compared to existing SDE-based solvers when using both linear and cosine noise schedules [14, 30]. Additionally, we evaluate various solvers in stroke-based image synthesis (i.e., SDEdit) and demonstrate that GMS surpasses other SDE-based solvers, as well as the widely used ODE-based solver DDIM [36]. 4.1 Sample quality on image data In this section, we conduct a quantitative comparison of sample quality using the widely adopted FID score [13]. Specifically, we evaluate multiple SDE-based solvers, including a comparison with DDPM [14] and Extended Analytic DPM [2] (referred to as SN-DDPM) using the even trajectory. As shown in Tab. 1, GMS demonstrates superior performance compared to DDPM and SN-DDPM under the same number of steps in CIFAR10 and Image Net 64 64. Specifically, GMS achieves a remarkable 4.44 improvement in FID given 10 steps on CIFAR10. Appendix E.7 illustrates in more detail the improvement of GMS when the number of sampling steps is limited. Meanwhile, we conduct GMS in Image Net 256 256 via adopting the U-Vi T-Huge [4] backbone as the noise network in Appendix E.8. Furthermore, taking into account the additional time required by GMS, our method still exhibits improved performance, as detailed in Appendix E.9. For integrity, we provide a comparison with other SDE-based solvers based on continuous time diffusion such as Gotta Go Fast [17], EDM [18] and SEED [10] in Appendix E.10 and shows that GMS largely outperforms other SDE-based solvers when the number of steps is less than 100. In Appendix G, we provide generated samples from GMS. 4.2 Stroke-based image synthesis based on SDEdit [27] Evaluation metrics. We evaluate the editing results based on realism and faithfulness similar with Meng et al. [27]. To quantify the realism of sample images, we use FID between the generated images and the target realistic image dataset. To quantify faithfulness, we report the L2 distance summed over all pixels between the stroke images and the edited output images. SDEdit [27] applies noise to the stroke image xg at time step t0 using N(xg t0|α(t)xg, σ2(t)I) and discretize the reverse SDE in Eq. (4) for sampling. Fig. 9 demonstrates the significant impact of t0 on the realism of sampled images. As t0 increases, the similarity to real images decreases. we choose (a) Trade-off between realism and faithfulness (b) t0 = 0.3T (c) t0 = 0.4T (d) t0 = 0.5T Figure 3: Result among different solvers in SDEdit. t0 denotes the time step of the start of reverse. (a): The points on each line represent the same t0 and the number of sampling steps. We select t0 = [300, 400, 500], number of steps = [50, 100]. (b): When t0 = 300, the effect of DDIM diminishes more prominently with the increase in the number of steps. the range from t0 = 0.3T to t0 = 0.5T (T = 1000 in our experiments) for our further experiments since sampled images closely resemble the real images in this range. Fig. 3(a) illustrates that when using the same t0 and the same number of steps, edited output images from GMS have lower faithfulness but with higher realism. This phenomenon is likely attributed to the Gaussian noise introduced by the SDE-based solver during each sampling step. This noise causes the sampling to deviate from the original image (resulting in low faithfulness) but enables the solver to transition from the stroke domain to the real image domain. Fig. 3(b) to Fig. 3(d) further demonstrates this phenomenon to a certain extent. The realism of the sampled images generated by the SDE-based solver escalates with an increase in the number of sampling steps. Conversely, the realism of the sampled images produced by the ODE-based solver diminishes due to the absence of noise, which prevents the ODE-based solver from transitioning from the stroke domain to the real image domain. Additionally, in the SDEdit task, GMS exhibits superior performance compared to SN-DDPM [2] in terms of sample computation cost. Fig. 4 shows the samples using DDIM and GMS when t0 = 400 and the number of steps is 40. 5 Related work Faster solvers. In addition to SDE-based solvers, there are works dedicated to improving the efficiency of ODE-based solvers [25, 23, 8]. Some approaches use explicit reverse transition kernels, such as those based on generative adversarial networks proposed by Xiao et al. [40] and Wang et al. [39]. Gao et al. [9] employ an energy function to model the reverse transition kernel. Zhang and Chen [44] use a flow model for the transition kernel. Non-Gaussian diffusion. Apart from diffusion, some literature suggests using non-Gaussian forward processes, which consequently involve non-Gaussian reverse processes. Bansal et al. [1] introduce a generalized noise operator that incorporates noise. Nachmani et al. [29] incorporate Gaussian mixture or Gamma noise into the forward process. While these works replace both the forward and reverse processes with non-Gaussian distributions, our approach aims to identify a suitable combination of non-Gaussian distributions to model the reverse process. 6 Conclusion This paper presents a novel Gaussian mixture solver (GMS) for diffusion models. GMS relaxes the Gaussian reverse kernel assumption to reduce discretization errors and improves the sample quality under the same sampling steps. Experimental results show that GMS outperforms existing SDE-based solvers, achieving a remarkable 4.44 improvement in FID compared to the state-of-the-art SDE-based solver proposed by Bao et al. [2] given 10 steps. Furthermore, due to the presence of noise, SDE-based solvers prove more suitable for stroke-based synthesis tasks and GMS still outperforms state-of-the-art SDE-based solvers. (a) Real Images (d) DDIM Figure 4: SDEdit samples of GMS and DDIM. Due to the presence of noise, SDE-based solvers, such as GMS OURS, generate images with more details. Limitations and broader impacts. While GMS enhances sample quality and potentially accelerates inference speed compared to existing SDE-based solvers, employing GMS still fall short of real-time applicability. Like other generative models, diffusion models can generate problematic fake content, and the use of GMS may amplify these undesirable effects. Acknowledgement This work was supported by NSF of China (Nos. 62076145); Beijing Outstanding Young Scientist Program (No. BJJWZYJH012019100020098); Major Innovation & Planning Interdisciplinary Platform for the Double-First Class Initiative, Renmin University of China; the Fundamental Research Funds for the Central Universities, and the Research Funds of Renmin University of China (No. 22XNKJ13). C. Li was also sponsored by Beijing Nova Program (No. 20220484044). [1] Arpit Bansal, Eitan Borgnia, Hong-Min Chu, Jie S Li, Hamid Kazemi, Furong Huang, Micah Goldblum, Jonas Geiping, and Tom Goldstein. Cold diffusion: Inverting arbitrary image transforms without noise. ar Xiv preprint ar Xiv:2208.09392, 2022. [2] Fan Bao, Chongxuan Li, Jiacheng Sun, Jun Zhu, and Bo Zhang. Estimating the optimal covariance with imperfect mean in diffusion probabilistic models. ar Xiv preprint ar Xiv:2206.07309, 2022. [3] Fan Bao, Chongxuan Li, Jun Zhu, and Bo Zhang. Analytic-dpm: an analytic estimate of the optimal reverse variance in diffusion probabilistic models. ar Xiv preprint ar Xiv:2201.06503, 2022. [4] Fan Bao, Shen Nie, Kaiwen Xue, Yue Cao, Chongxuan Li, Hang Su, and Jun Zhu. All are worth words: A vit backbone for diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 22669 22679, 2023. [5] Fan Bao, Shen Nie, Kaiwen Xue, Chongxuan Li, Shi Pu, Yaole Wang, Gang Yue, Yue Cao, Hang Su, and Jun Zhu. One transformer fits all distributions in multi-modal diffusion at scale. ar Xiv preprint ar Xiv:2303.06555, 2023. [6] Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved baselines with momentum contrastive learning. ar Xiv preprint ar Xiv:2003.04297, 2020. [7] Prafulla Dhariwal and Alexander Nichol. Diffusion models beat gans on image synthesis. Advances in Neural Information Processing Systems, 34:8780 8794, 2021. [8] Tim Dockhorn, Arash Vahdat, and Karsten Kreis. Genie: Higher-order denoising diffusion solvers. ar Xiv preprint ar Xiv:2210.05475, 2022. [9] Ruiqi Gao, Yang Song, Ben Poole, Ying Nian Wu, and Diederik P Kingma. Learning energybased models by diffusion recovery likelihood. ar Xiv preprint ar Xiv:2012.08125, 2020. [10] Martin Gonzalez, Nelson Fernandez, Thuy Tran, Elies Gherbi, Hatem Hajri, and Nader Masmoudi. Seeds: Exponential sde solvers for fast high-quality sampling from diffusion models. ar Xiv preprint ar Xiv:2305.14267, 2023. [11] Lars Peter Hansen. Large sample properties of generalized method of moments estimators. Econometrica: Journal of the econometric society, pages 1029 1054, 1982. [12] Amir Hertz, Ron Mokady, Jay Tenenbaum, Kfir Aberman, Yael Pritch, and Daniel Cohen-Or. Prompt-to-prompt image editing with cross attention control. ar Xiv preprint ar Xiv:2208.01626, 2022. [13] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017. [14] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840 6851, 2020. [15] Jonathan Ho, Chitwan Saharia, William Chan, David J Fleet, Mohammad Norouzi, and Tim Salimans. Cascaded diffusion models for high fidelity image generation. J. Mach. Learn. Res., 23(47):1 33, 2022. [16] Jonathan Ho, Tim Salimans, Alexey Gritsenko, William Chan, Mohammad Norouzi, and David J Fleet. Video diffusion models. ar Xiv preprint ar Xiv:2204.03458, 2022. [17] Alexia Jolicoeur-Martineau, Ke Li, Rémi Piché-Taillefer, Tal Kachman, and Ioannis Mitliagkas. Gotta go fast when generating data with score-based models. ar Xiv preprint ar Xiv:2105.14080, 2021. [18] Tero Karras, Miika Aittala, Timo Aila, and Samuli Laine. Elucidating the design space of diffusion-based generative models. ar Xiv preprint ar Xiv:2206.00364, 2022. [19] Bahjat Kawar, Michael Elad, Stefano Ermon, and Jiaming Song. Denoising diffusion restoration models. ar Xiv preprint ar Xiv:2201.11793, 2022. [20] Bahjat Kawar, Shiran Zada, Oran Lang, Omer Tov, Huiwen Chang, Tali Dekel, Inbar Mosseri, and Michal Irani. Imagic: Text-based real image editing with diffusion models. ar Xiv preprint ar Xiv:2210.09276, 2022. [21] Diederik Kingma, Tim Salimans, Ben Poole, and Jonathan Ho. Variational diffusion models. Advances in neural information processing systems, 34:21696 21707, 2021. [22] Zhifeng Kong, Wei Ping, Jiaji Huang, Kexin Zhao, and Bryan Catanzaro. Diffwave: A versatile diffusion model for audio synthesis. ar Xiv preprint ar Xiv:2009.09761, 2020. [23] Luping Liu, Yi Ren, Zhijie Lin, and Zhou Zhao. Pseudo numerical methods for diffusion models on manifolds. ar Xiv preprint ar Xiv:2202.09778, 2022. [24] Cheng Lu, 2023. https://github.com/huggingface/diffusers/pull/3344. [25] Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. Dpm-solver: A fast ode solver for diffusion probabilistic model sampling in around 10 steps. ar Xiv preprint ar Xiv:2206.00927, 2022. [26] Geoffrey J Mc Lachlan and Thriyambakam Krishnan. The EM algorithm and extensions. John Wiley & Sons, 2007. [27] Chenlin Meng, Yang Song, Jiaming Song, Jiajun Wu, Jun-Yan Zhu, and Stefano Ermon. Sdedit: Image synthesis and editing with stochastic differential equations. ar Xiv preprint ar Xiv:2108.01073, 2021. [28] Eliya Nachmani and Shaked Dovrat. Zero-shot translation using diffusion models. ar Xiv preprint ar Xiv:2111.01471, 2021. [29] Eliya Nachmani, Robin San Roman, and Lior Wolf. Non gaussian denoising diffusion models. ar Xiv preprint ar Xiv:2106.07582, 2021. [30] Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. In International Conference on Machine Learning, pages 8162 8171. PMLR, 2021. [31] Ben Poole, Ajay Jain, Jonathan T Barron, and Ben Mildenhall. Dreamfusion: Text-to-3d using 2d diffusion. ar Xiv preprint ar Xiv:2209.14988, 2022. [32] Vadim Popov, Ivan Vovk, Vladimir Gogoryan, Tasnima Sadekova, and Mikhail Kudinov. Gradtts: A diffusion probabilistic model for text-to-speech. In International Conference on Machine Learning, pages 8599 8608. PMLR, 2021. [33] Prajit Ramachandran, Barret Zoph, and Quoc V Le. Searching for activation functions. ar Xiv preprint ar Xiv:1710.05941, 2017. [34] Alex Shonenkov, Misha Konstantinov, Daria Bakshandaeva, Christoph Schuhmann, Ksenia Ivanova, and Nadiia Klokova, 2023. https://github.com/deep-floyd/IF. [35] Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning, pages 2256 2265. PMLR, 2015. [36] Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. ar Xiv preprint ar Xiv:2010.02502, 2020. [37] Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. ar Xiv preprint ar Xiv:2011.13456, 2020. [38] Pascal Vincent. A connection between score matching and denoising autoencoders. Neural computation, 23(7):1661 1674, 2011. [39] Zhendong Wang, Huangjie Zheng, Pengcheng He, Weizhu Chen, and Mingyuan Zhou. Diffusion-gan: Training gans with diffusion. ar Xiv preprint ar Xiv:2206.02262, 2022. [40] Zhisheng Xiao, Karsten Kreis, and Arash Vahdat. Tackling the generative learning trilemma with denoising diffusion gans. ar Xiv preprint ar Xiv:2112.07804, 2021. [41] Xingyu Xie, Pan Zhou, Huan Li, Zhouchen Lin, and Shuicheng Yan. Adan: Adaptive nesterov momentum algorithm for faster optimizing deep models. ar Xiv preprint ar Xiv:2208.06677, 2022. [42] Xingqian Xu, Zhangyang Wang, Eric Zhang, Kai Wang, and Humphrey Shi. Versatile diffusion: Text, images and variations all in one diffusion model. ar Xiv preprint ar Xiv:2211.08332, 2022. [43] Lvmin Zhang and Maneesh Agrawala. Adding conditional control to text-to-image diffusion models. ar Xiv preprint ar Xiv:2302.05543, 2023. [44] Qinsheng Zhang and Yongxin Chen. Diffusion normalizing flow. Advances in Neural Information Processing Systems, 34:16280 16291, 2021. [45] Min Zhao, Fan Bao, Chongxuan Li, and Jun Zhu. Egsde: Unpaired image-to-image translation via energy-guided stochastic differential equations. ar Xiv preprint ar Xiv:2207.06635, 2022. A.1 Proof of Proposition 3.1 When q(x0) is a mixture of Dirac distribution which means that q(x0) = PM i=1 wiδ(x xi), PM i=1 wi = 1, which has total M components, and when the forward process q(xt|x0) is a Gaussian distribution as Eq. (1), the reverse process q(xs|xt) would be: q(xs|xt) = Z q(xs|xt, x0)q(x0|xt)dx0 = Z q(xs|xt, x0)q(x0)q(xt|x0)/q(xt)dx0 = 1/q(xt) Z q(xs|xt, x0)q(x0)q(xt|x0)dx0 = 1/q(xt) i=1 wiq(xs|xt, xi 0)q(xt|xi 0) According to the definition of forward process, the distribution q(xt|xi 0) = N(xt| atxi 0, (1 at)I). Due to the Markov property of forward process, when t > s, we have q(xs, xt|x0)q(xs|x0)q(xt|xs). The term q(xs|xt, x0) would be viewed as a Bayesian posterior resulting from a prior q(xs|x0), updated with a likelihood term q(xt|xs). And therefore q(xs|xt, xi 0) = N(xs|µq(xt, x0), Σq(xt, x0)I) = N(xs|at|sσ2 s σ2 t xt + asσ2 t|s σ2 t x0, σ2 sσ2 t|s σ2 t I) It is easy to prove that, the distribution q(xs|xt) is a mixture of Gaussian distribution: i=1 wiq(xs|xt, xi 0)q(xt|xi 0) i=1 wi N(xt| atxi 0, (1 at)I) N(xs|µq(xt, x0), Σq(xt, x0)) When t is large, s is small, σ2 t|s would be large, meaning that the influence of xi 0 would be large. Secondly, when q(x0) is a mixture of Gaussian distribution which means that q(x0) = PM i=1 wi N(xi 0|µi, Σi), PM i=1 wi = 1. For simplicity of analysis, we assume that this distribution is a one-dimensional distribution or that the covariance matrix is a high-dimensional Gaussian distribution with a diagonal matrix Σi = diagi(σ2). Similar to the situation above, for each dimension in the reverse process: q(xs|xt) = 1/q(xt) Z q(xs|xt, x0)q(x0)q(xt|x0)dx0 i=1 wi/q(xt) Z q(xs|xt, x0)N(x0|µi, Σi)q(xt|x0)dx0 i=1 wi/q(xt) Z 1 2πσq e (xi 0 µq)2 2πσi e (xi 0 µi)2 2π 1 at e (xt atx0)2 i=1 wi/q(xt) Z Zi 1 2πσx e (xi 0 µx)2 σ2x e (xs µ(xt))2 And q(xs|xt) could be a Gaussian mixture which has M component. A.2 Proof of Proposition 3.2 Recall that our objective function is to find optimal parameters: ˆθ = argmin θ [QNc(θ) | {z } 1 1 ] = argmin θ [g Nc(θ)T | {z } 1 N WNc |{z} N N g Nc(θ) | {z } 1 N where g M(θ) is the moment conditions talked about in Sec. 3.2, W is the weighted Matrix, Nc is the sample size. When solving such problems using optimizers, which are equivalent to the one we used when selecting θGMM such that QT (θGMM) θ = 0, and its first derivative: θ | {z } d 1 WNc |{z} N N i=1 g(xi, θ)] its second derivative (the Hessian matrix): θ2 | {z } d d θ1θ1 2QNc(θ) θ1θ2 ... 2QNc(θ) θ1θd 2QNc(θ) θ2θ1 ... ... ... ... ... ... ... ... ... ... 2QNc(θ) θiθj = 2[ 1 θi ]T WNc[ 1 θiθj ]WNc[ 1 i=1 g(xi, θ)]. By Taylor s expansion of the gradient around optimal parameters θ0, we have: θ θT (θGMM θ) (17) 7 (θGMM θ) ( 2QNc(θ0) θ θT ) 1 QNc(θ0) Consider one element of the gradient vector QNc(θ0) θm | {z } p E( g(xi,θ0) ]T WNc |{z} p W i=1 g(xi, θ0) | {z } p E[g(xi,θ0)]=0 Consider one element of the Hessian matrix 2QNc(θ0) θi θT j = 2[ 1 θi ]T WNc[ 1 θiθj ]T WNc[ 1 i=1 g(xi, θ0)] p 2ΓT 0,i WΓ0,j. Therefore, it is easy to prove that θGMM θ0 p 0, and uses law of large numbers we could obtain, θm | {z } p E( g(X,θ0) ]T WM |{z} p W i=1 g(xi, θ0) d N(0,E(g(X,θ0)g(X,θ0)T )) ] d 2ΓT 0,m WN(0, Φ0), therefore, we have T(θGMM θ0) ( 2QT (θ0) d N(0, (ΓT 0 WΓ0) 1ΓT 0 WΦ0WΓ0(ΓT 0 WΓ0) 1). When the number of parameters d equals the number of moment conditions N, Γ0 becomes a (nonsingular) square matrix, and therefore, T(θGMM θ0) d N(0, (ΓT 0 WΓ0) 1ΓT 0 WΦ0WΓ0(ΓT 0 WΓ0) 1) (22) = N(0, Γ 1 0 W 1(ΓT 0 ) 1ΓT 0 WΦ0WΓ0Γ 1 0 W 1(ΓT 0 ) 1) = N(0, Γ 1 0 Φ0(ΓT 0 ) 1), which means that the foregoing observation suggests that the selection of WNc has no bearing on the asymptotic variance of the GMM estimator. Consequently, it implies that regardless of the specific method employed to determine WNc, provided the moment estimates are asymptotically consistent, WNc serves as the optimal weight matrix, even when dealing with small samples. The moments utilized for fitting via the generalized method of moments in each step are computed based on the noise network s first n-th order. To ensure adherence to the aforementioned proposition, it is necessary to assume that the n-th order of the noise network converges with probability to Eq(x0|xt)[diag(ϵ n 1 ϵ)], details in Appendix B. Consequently, the n-th order moments derived from the noise networks converge with probability to the true moments. Therefore, any choice of weight matrix is optimal. A.3 Non-Gaussian distribution of transition kernel within large discretization steps we can apply Bayes rule to the posterior distribution q(xt|xt+ t) as follows: q(xt|xt+ t) = q(xt+ t|xt)q(xt) q(xt+ t) = q(xt+ t|xt) exp(log(q(xt)) log(q(xt+ t))) (23) xt+ t xt ft(xt) t 2 2g2 t t + log p(xt) log(xt+ t) where t is the step size, q(xt) is the marginal distribution of xt. When xt+ t and xt are close enough, using Taylor expansion for log p(xt+ t), we could obtain: log p(xt+ t) log p(xt) + (xt+ t xt) xt log p(xt) + t t log p(xt), (24) q(xt|xt+ t) exp xt+ t xt [ft(xt) g2 t xt log p(xt)] t 2 2g2 t t + O( t) By ignoring the higher order terms, the reverse transition kernel will be Gaussian distribution. However, as t increases, the higher-order terms in the Taylor expansion cannot be disregarded, which causes the reverse transition kernel to deviate from a Gaussian distribution. Empirically, from Fig. 2 in our paper, we observe that as T decreases, the reverse transition kernel increasingly deviates from a Gaussian distribution. For more details, please refer to Appendix D.2. B Calculation of the first order moment and higher order moments Suppose the forward process is a Gaussian distribution same with the Eq. (1) as q(xt|xs) = N(xt|a(t)xt 1, σ(t)I). And let 1 t > s 0 always satisfy, q(xt|xs) = N(xt|at|sxs, βt|s I), where at|s = at/as and βt|s = σ2 t a2 t|sσ2 s, σs = p 1 a(s),σt = p 1 a(t). It s easy to prove that the distribution q(xt|x0), q(xs|xt, x0) are also a Gaussian distribution [21]. Therefore, the mean of xs under the measure q(xs|xt) would be Eq(xs|xt)[xs] = Eq(x0|xt)Eq(xs|xt,x0)[xs] (26) = Eq(x0|xt)[ 1 at|s (xt βt|s = 1 at|s (xt βt|s σt Eq(x0|xt)[ϵt]). And for the second order central moment Covq(xs|xt)[xs], we use the total variance theorem, refer to [2], and similar with [2], we only consider the diagonal covariance. Covq(xs|xt)[xs] = Eq(x0|xt)Covq(xs|xt,x0)[xs] + Covq(x0|xt)Eq(xs|xt,x0)[xs] (27) = λ2 t I + Covq(x0|xt) µ(xn, Eq(x0|xn)[x0]) = λ2 t I + asβ2 t|s σ4 t Covq(x0|xt)[x0] = λ2 t I + as|0β2 t|s σ4 t σ2 t at|0 Covq(x0|xt)[ϵt] = λ2 t I + β2 t|s σ2 t at|s (Eq(x0|xt)[ϵt ϵt] Eq(x0|xt)[ϵt] Eq(x0|xt)[ϵt]), since the higher-order moments are diagonal matrix, we use diag(M) to represent the diagonal elements that have the same dimensions as the first-order moments, such as diag(xs xs) = Covq(xs|xt)[xs] and diag(xs xs xs) = ˆ M3 have the same dimensions as xs Moreover, for the diagonal elements of the third-order moments, we have Eq(xs|xt)[diag(xs xs xs)] = Eq(x0|xt)Eq(xs|xt,x0)[xs xs xs], we could use the fact that Eq(xs|xt,x0)[(xs µ(xt, x0)) (xs µ(xt, x0)) (xs µ(xt, x0))] = 0, therefore, ˆ M3 = Eq(xs|xt)[diag(xs xs xs)] = Eq(x0|xt)Eq(xs|xt,x0)[diag(xs xs xs)] (28) = Eq(x0|xt)Eq(xs|xt,x0)[3diag(xt xt µ(xt, x0)) 3diag(xt µ(xt, x0) µ(xt, x0)) + diag(µ(xt, x0) 2 µ(xt, x0))] = [(at|sσ2 s σ2 t )3diag(xt xt xt) + 3λ2 t at|sσ2 s σ2 t xt] | {z } Constant term + [ 3a2 t|sσ4 sa2 s|0β2 t|s σ8 t (diag(xt xt)) + as|0βt|s σ2 t I] Eq(x0|xt)[x0] | {z } Linear term in x0 + 3at|sσ2 s σ2 t (as|0βt|s σ2 t )2xt Eq(x0|xt)[diag(x0 x0)] | {z } Quadratic term in x0 + (as|0βt|s σ2 t )3Eq(x0|xt)[diag(x0 x0 x0)] | {z } Cubic term in x0 where the third equation is derived from the decomposition of higher-order moments of Gaussian distribution, the fourth equation is obtained by splitting It is noted to highlight that in our study, we only consider the diagonal higher-order moments in our method for computational efficiency since estimating full higher-order moments results in escalated output dimensions (e.g., quadratic growth for covariance and cubic for the third-order moments) and thus requires substantial computational demand and therefore all outer products can be transformed into corresponding element multiplications and we have: Eq(x0|xt)[diag(x0 2 x0)] = 1 α 3 2 (t) Eq(x0|xt)[diag((xt σ(t)ϵ) 2 (xt σ(t)ϵ))] (29) α 3 2 (t) Eq(x0|xt)[diag(xt 2 xt 3σ(t)(xt xt) ϵ + 3σ2(t)xt (ϵ ϵ) σ3(t)(ϵ 2 ϵ))] α 3 2 (t) [Eq(x0|xt)[xt 2 xt] 3σ(t)(xt xt) Eq(x0|xt)[ϵ]+ + 3σ2(t)xt Eq(x0|xt)[ϵ ϵ] σ3(t)[ϵ 2 ϵ]], Therefore, when we need to calculate the third-order moment, we only need to obtain Eq(x0|xt)[ϵt], Eq(x0|xt)[ϵt ϵt] and Eq(x0|xt)[ϵt ϵt ϵt]. Similarly, when we need to calculate the n-order moment, we will use Eq(x0|xt)[ϵt], ..., Eq(x0|xt)[ϵt n 1 ϵt]. Bao et al. [2] put forward using a sharing network and using the MSE loss to estimate the network to obtain the above information about different orders of noise. The function h(f 3 [1], f 3 [2], f 3 [3]) in Algo. 2 is defined as h(f 3 [1](xt, t), f 3 [2](xt, t), f 3 [3](xt, t)) = M1(f 3 [1](xt, t)), M2(f 3 [2](xt, t)), M3(f 3 [3](xt, t)), where M1(f 3 [1](xt, t)) = Eq(xs|xt)[xs] in Eq. (26), M2(f 3 [2](xt, t)) = Covq(xs|xt)[xs] in Eq. (27) and M3(f 3 [3](xt, t)) = ˆ M3 = Eq(xs|xt)[diag(xs xs xs)] in Eq. (28) B.1 Objective function for Gaussian mixture transition kernel with two components Recall that the general objective function to fit a Gaussian mixture transition kernel in each sampling step via GMM is shown in Eq. (30). min θ Q(θ, M1, ..., MN) = min θ [ 1 i=1 g(xi, θ)]T W[ 1 i=1 g(xi, θ)], (30) where AT denotes the transpose matrix of matrix A. In our paper, we propose to use the first three moments to fit a Gaussian mixture transition kernel p(xs|xt) = 1 3N(µ(1) t , σ2 t ) + 2 3N(µ(2) t , σ2 t ) in each sampling step from xt to xs. Therefore, the final objective function as follow: min µ(1) t ,µ(2) t ,σ2 t [ 3µ(1) t + 2 3µ(2) t ) M2 ( 1 3[(µ(1) t )2 + σ2 t ] + 2 3[(µ(2) t )2 + σ2 t )]) M3 ( 1 3[(µ(1) t )3 + 3µ(1) t σ2 t ] + 2 3[(µ(2) t )3 + 3µ(2) t σ2 t ]) 3µ(1) t + 2 3µ(2) t ) M2 ( 1 3[(µ(1) t )2 + σ2 t ] + 2 3[(µ(2) t )2 + σ2 t )]) M3 ( 1 3[(µ(1) t )3 + 3µ(1) t σ2 t ] + 2 3[(µ(2) t )3 + 3µ(2) t σ2 t ]), where to simplify the analysis, we use the scalar form of parameters µ(1) t , µ(2) t , σ2 t as representation. C Modeling reverse transition kernel via exponential family Analysis in Sec. 3.1 figures out that modeling reverse transition kernel via Gaussian distribution is no longer sufficient in fast sampling scenarios. In addition to directly proposing the use of a Gaussian Mixture for modeling, we also analyze in principle whether there are potentially more suitable distributions i.e., the feasibility of using them for modeling. We would turn back to analyzing the original objective function of DPMs to find a suitable distribution. The forward process q(xt|xs) = N(xt|at|sxs, βt|s I), consistent with the definition in Appendix B. DPMs goal is to optimize the modeled reverse process parameters to maximize the variational bound L in Ho et al. [14]. And the ELBO in Ho et al. [14] can be re-written to the following formula: L = DKL(q(x T )||p(x T )) + Eq[ X t 1 DKL(q(xs|xt)||p(xs|xt))] + H(x0), (32) where qt .= q(xt) is the true distribution and pt .= p(xt) is the modeled distribution, and the minimum problem could be transformed into a sub-problem, proved in Bao et al. [3]: min {θ} L min {θs|t} T t=1 DKL(q(xs|xt)||pθs|t(xs|xt)). (33) We have no additional information besides when the reverse transition kernel is not Gaussian. But Lemma. C.3 proves that when the reverse transition kernel pθs|t(xs|xt) is exponential family pθt(xs|xt) = p(xt, θs|t) = h(xt) exp θT s|tt(xt) α(θs|t) , solving the sub-problem Eq. (33) equals to solve the following equations, which is to match moments between the modeled distribution and true distribution: Eq(xs|xt)[t(xs)] = Ep(xt,θs|t)[t(xs)]. (34) When t(x) = (x, .., xn)T , solving Eq.(34) equals to match the moments of true distribution and modeled distribution. Meanwhile, Gaussian distribution belongs to the exponential family with t(x) = (x, x2)T and θt = ( µt 2σ2 t )T , details in Lemma. C.2. Therefore, when modeling the reverse transition kernel as Gaussian distribution, the optimal parameters are that make its first two moments equal to the true first two moments of the real reverse transition kernel q(xs|xt), which is consistent with the results in Bao et al. [3] and Bao et al. [2]. The aforementioned discussion serves as a motivation to acquire higher-order moments and identify a corresponding exponential family, which surpasses the Gaussian distribution in terms of complexity. However, proposition C.1 shows that finding such exponential family distribution with higher-order moments is impossible. Proposition C.1 (Infeasibility of exponential family with higher-order moments.). Given the first n-th order moments. It s non-trivial to find an exponential family distribution for min DKL(q||p) when n is odd. And it s hard to solve min DKL(q||p) when n is even. C.1 Proof of Proposition C.1 Lemma C.2. (Gaussian Distribution belongs to Exponential Family). Gaussian distribution p(x) = 2πσ exp (x µ)2 2σ2 is exponential family with t(x) = (x, x2)T and θ = ( µ Proof. For simplicity, we only prove one-dimensional Gaussian distribution. We could obtain: 2πσ exp (x µ)2 2σ2 (x2 2µx + µ2) = exp log 2πσ2 1/2 exp 1 2σ2 (x2 2µx) µ2 = exp log 2πσ2 1/2 exp 1 2σ2 ( 2µ 1)(x x2)T µ2 σ2 1 2σ2 )(x x2)T ( µ2 2 log 2πσ2 ) , where θ = ( µ 2σ2 )T and t(x) = (x, x2)T Lemma C.3. (The Solution for Exponential Family in Minimizing the KL Divergence). Suppose that p(x) belongs to exponential family p(x, θ) = h(x) exp θT t(x) α(θ) , and the solution for minimizing the Eq[log p] is Eq[t(x)] = Ep(x,θ)[t(x)]. Proof. An exponential family p(x, η) = h(x) exp ηT t(x) α(η) f(x, η) = h(x) exp ηT t(x) with log-partition α(η). And we could obtain its first order condition on Eq[log p] as: η log f(x, η) = η(log h(x) + ηT t(x)) = t(x) (36) ηα(η) = η log Z f(x, η)dx = R ηf(x, η)dx R f(x, η)dx (37) = e α(η) Z t(x)f(x, η)dx = Z t(x)p(x, η)dx = Ep(x,η)[t(x)] In order to minimize the DKL(q||p) = R q log(q/p) = Eq[log p], we have: Eq[log p] = Z dq log(h(x)) + Z dq(ηT t(x) α(η) η Eq[log p] = Z dq[ η (ηT x α(η))] = 0 = Z dq(x Ep(x,η)[t(x)]) = Eq[t(x)] Ep(x,η)[t(x)] = 0 = Eq[t(x)] = Ep(x,η)[t(x)] For the second-order condition, we have the following: Z dp(x, η)t(x) (38) = Z η h(x) exp ηT t(x) α(η) t(x)dx = Z h(x)t(x) η exp ηT t(x) α(η) dx = Z h(x)t(x) exp ηT t(x) α(η) dx(t(x) Ep[t(x)]) = Z p(x, η)dx(t2(x) t(x)Ep[t(x)]) = Ep(x,η)[t2(x)] Ep(x,η)[t(x)]2 = Covp(x,η)[t(x)] 0 Therefore, the second-order condition for the cross entropy would be: η2 Eq[log p] = η (Eq[t(x)] Ep(x,η)[t(x)]) (39) = Z η p(x, η)t(x)dx η2 α(η) = Covp(x,η)[t(x)] 0 When we assume that the reverse process is Gaussian, the solution to Eq. (33) equals to match the moment of true distribution and modeled distribution µ = Eq[x], Σ = Covq[x]. Lemma C.4. (Infeasibility of the exponential family with higher-order moments). Suppose given the first N-th order moments Mi, i = 1, .., N and modeled p as an exponential family. It is nontrivial to solve the minimum problem Eq[log p] when N is odd and it s difficult to solve when N is even. Proof. While given the mean, covariance, and skewness of the data distribution, assume that we could find an exponential family that minimizes the KL divergence, so that the distribution would satisfy the following form: L(p, ˆλ) = DKL(q||p) ˆλT ( Z pt m) p L(p, ˆλ) = log p(x) h(x) + 1 ˆλT t = 0 (40) p(x) = h(x) exp ˆλT t 1 where, t(x) = (x, x2, x3), p = h(x) exp λ0 + λ1x + λ2x2 + λ3x3 and R dpx3 = M3. However, when λ3 is not zero, R p = and density can t be normalized. The situation would be the same given an odd-order moment. Similarly, given a more fourth-order moment, we could derive that λ3 = 0 above, and we should solve an equation R dpx4 = M4 and p = h(x) exp λ0 + λ1x + λ2x2 + λ4x4 . Consider such function: dx exp x2 λx4 , λ > 0 (41) When λ 0, we could obtain limλ 0 Z(λ) = π For other cases, the lambda can be expanded and then integrated term by term, which gives Z(λ) P n=0 ( λ)n n! Γ(2n + 1/2), but this function However, the radius of convergence of this level is 0, so when the λ takes other values, we need to propose a reasonable expression for the expansion after the analytic extension. Therefore, for solving the equation R dpx4 = M4, there is no analytical solution first, and the numerical solution also brings a large computational effort. D More information about Fig. 1 and Fig. 2 D.1 Experiment in Toy-data To illustrate the effectiveness of our method, we first compare the results of different solvers on one-dimensional data. The distribution of our toy-data is q(x0) = 0.4N( 0.4, 0.122) + 0.6N(0.3, 0.052) and we define our solvers in each step as p(xs|xt) = 1 3N(µ(1) t , σ2 t ) + 2 3N(µ(1) t , σ2 t ) with vectors µ(1) t , µ(2) t and σ2 t , which can not overfit the ground truth. We then train second and third-order noise networks on the one-dimensional Gaussian mixture whose density is multi-modal. We use a simple MLP neural network with Swish activation [33]. Moreover, we experiment with our solvers in 8-Gaussian. The result is shown in Tab. 2. GMS outperforms Extended Analytic DPM (SN-DDPM) [2] as presented in Tab. 2, with a bandwidth of 1.05σL 0.25, where σ is the standard deviation of data and L is the number of samples. Table 2: Comparison with SN-DDPM w.r.t. Likelihood Eq[log pθ(x)] on 8-Gaussian. GMS outperforms SN-DDPM. # K 5 10 20 40 SN-DDPM -0.7885 0.0661 0.0258 0.1083 GMS -0.6304 0.0035 0.0624 0.1127 D.2 Experiment in Fig. 2 In this section, we will provide a comprehensive explanation of the procedures involved in computing the discrepancy between two third-order moment calculation methods, as depicted in Fig. 2. The essence of the calculation lies in the assumption that the reverse transition kernel follows a Gaussian distribution. By employing the following equations (considering only the diagonal elements of higher-order moments), we can compute the third-order moment using the first two-order moments: Eq(xti 1|xti)[xti 1 xti 1 xti 1]G .= MG = µ µ µ + 3µ Σ, (42) where µ is the first-order moment and Σ is the diagonal elements of second order moment, which can be calculated by the Eq. (26) and Eq. (27). Meanwhile, we can calculate the estimated third-order moment ˆ M3 by Eq. (28). We use the pre-trained noise network from Ho et al. [14] and the second-order noise network form Bao et al. [2] and train the third-order noise network in CIFAR10 with the linear noise schedule. Given that all higher-order moments possess the same dimension as the first-order moment µ, we can directly compare the disparity between different third-order moment calculation methods using the Mean Squared Error (MSE). Thus, to quantify the divergence between the reverse transition kernel q(xs|xt) and the Gaussian distribution, we can utilize the following equation: Ds|t = log Eq(xs|xt)[xs xs xs]G ˆ M3 2 , (43) where ˆ M3 is obtained via Eq. (28), and we can start at different time step t and choose a corresponding s to calculate the Ds|t and draw different time step and step size t s and we can derive Fig. 2. E Experimental details E.1 More discussion on weight of Gaussian mixture From Proposition 3.2, we know that when the number of parameters in the Gaussian mixture equals the number of moment conditions, any choice of weight matrix is optimal. Therefore, we will discuss the choice of parameters to optimize in this section. As we have opted for a Gaussian mixture with two components q(xs|xt) = ω1N(µ(1) s|t , Σ(1) s|t ) + ω2N(µ(2) s|t , Σ(2) s|t ) as our foundational solvers, there exist five parameters (considered scalar, with the vector cases being analogous) available for optimization. Our primary focus is on optimizing the mean and variance of the two components, as optimizing the weight term would require solving the equation multiple times. Additionally, we have a specific requirement that our Gaussian mixture can converge to a Gaussian distribution at the conclusion of optimization, particularly when the ground truth corresponds to a Gaussian distribution. In Tab. 3, we show the result of different choices of parameters in the Gaussian mixture. Table 3: Results among different parameters in CIFAR10 (LS), the number of steps is 50. The weight of Gaussian mixture is ω1 = 1 3 and ω2 = 2 µ(1) s|t ,µ(2) s|t ,Σs|t µ(1) s|t ,Σ(1) s|t ,Σ(2) s|t µs|t,Σ(1) s|t ,Σ(2) s|t CIFAR10 (LS) 4.17 10.12 4.22 When a parameter is not accompanied by a superscript, it implies that both components share the same value for that parameter. On the other hand, if a parameter is associated with a superscript, and only one moment contains that superscript, it signifies that the other moment directly adopts the true value for that parameter. It is evident that the optimization of the mean value holds greater significance. Therefore, our subsequent choices for optimization are primarily based on the first set of parameters µ(1) s|t ,µ(2) s|t ,Σs|t. Another crucial parameter to consider is the selection of weights ωi. In Tab. 4, we show the result while changing the weight of the Gaussian mixture and the set of weight ω1 = 1 2 performs best among different weight. Table 4: Results among different weight choices in CIFAR10 (LS), the number of steps is 50. ω1 = 1 100 , ω2 = 99 100 ω1 = 1 CIFAR10 (LS) 4.63 4.20 4.17 4.26 What s more, we observe that such choice of weights ( 1 2) consistently yielded superior performance among different datasets. Identifying an optimal weight value remains a promising direction for further enhancing the GMS. E.2 Details of pre-trained noise networks In Tab. 5, we list details of pre-trained noise prediction networks used in our experiments. Table 5: Details of noise prediction networks used in our experiments. LS means the linear schedule of σ(t) [14] in the forward process of discrete time step (see Eq. (1)). CS means the cosine schedule of σ(t) [30] in the forward process of discrete timesteps (see Eq. (1)). # TIMESTEPS N NOISE SCHEDULE OPTIMIZER FOR GMM CIFAR10 (LS) 1000 LS ADAN CIFAR10 (CS) 1000 CS ADAN IMAGENET 64X64 4000 CS ADAN E.3 Details of the structure of the extra head In Tab. 6, we list structure details of NN1, NN2 and NN3 of prediction networks used in our experiments. Table 6: NN1 represents noise prediction networks and NN2, NN3 represent networks for estimating the secondand the third-order of noise, which used in our experiments. Conv denotes the convolution layer. Res denotes the residual block. None denotes using the original network without additional parameters NN1 NN2 (NOISE) NN3 (NOISE) CIFAR10 (LS) NONE CONV RES+CONV CIFAR10 (CS) NONE CONV RES+CONV IMAGENET 64X64 NONE RES+CONV RES+CONV E.4 Training Details We use a similar training setting to the noise prediction network in [30] and [2]. On all datasets, we use the ADAN optimizer [41] with a learning rate of 10 4; we train 2M iterations in total for a higher order of noise network; we use an exponential moving average (EMA) with a rate of 0.9999. We use a batch size of 64 on Image Net 64X64 and 128 on CIFAR10. We save a checkpoint every 50K iterations and select the models with the best FID on 50k generated samples. Training one noise network on CIFAR10 takes about 100 hours on one A100. Training on Image Net 64x64 takes about 150 hours on one A100. E.5 Details of Parameters of Optimizer in Sampling In Tab. 7, we list details of the learning rate, learning rate schedule, and warm-up steps for different experiments. Table 7: Details of Parameters of Optimizer used in our experiments. lr Schedule means the learning rate schedule. min lr means the minimum learning rate while using the learning rate schedule, ιt is a function with the second order growth function of sampling steps t. LEARNING RATE LR SCHEDULE MIN LR WARM-UP STEPS CIFAR10 MAX(0.16-ιt*0.16,0.12) COS 0.1 18 IMAGENET 64 64 MAX(0.1-ιt*0.1,0.06) COS 0.04 18 where COS represents the cosine learning rate schedule [6]. We find that the cosine learning rate schedule works best. The cos learning rate could be formulated as follows: i Iw αi if i Iw max((0.5 cos i Iw I Iw π + 1)αt, αmin) else (44) where, at is the learning rate after t steps, Iw is the warm-up steps, αmin is the minimum learning rate, I is the total steps. E.6 Details of memory and time cost In Tab. 8, we list the memory of models (with the corresponding methods) used in our experiments. The extra memory cost higher-order noise prediction network is negligible. Table 8: Model size (MB) for different models. The model of SNDDPM denotes the model that would predict noise and the square of noise; The model of GMDDPM denotes the model that would predict noise, the square of noise, and the third power of noise. NOISE PREDICTION NETWORK (ALL BASELINES) NOISE & SN PREDICTION NETWORKS SNDDPM NOISE & SN PREDICTION NETWORKS (GMDDPM) CIFAR10 (LS) 50.11 MB 50.11 MB 50.52 MB (+0.8%) CIFAR10 (CS) 50.11 MB 50.11 MB 50.52 MB (+0.8%) IMAGENET 64 64 115.46 115.87 MB 116.28 (+0.7%) In Fig. 5, we present a comprehensive breakdown of the time of optimization process within GMS at each sampling step. Subfigure (a) in Fig. 5 illustrates that when the batch size is set to 100, the time of optimizing approximately 320 steps to fit a Gaussian mixture transition kernel at each step is equivalent to the time needed for one network inference, which means that the additional time of GMS for each sampling steps is about 10% when the number of optimizing steps is 30. Meanwhile, Subfigure (a) in Fig. 5 elucidates the relation between sampling time and the number of sampling steps for both GMS and SN-DDPM. It is noteworthy that the optimization steps employed in GMS remain fixed at 25 per sampling step, consistent with our setting in experiments. Evidently, as the number of sampling steps escalates, GMS demonstrates a proportional increase in computational overhead, consistently maintaining this overhead within a 10% margin of the original computational cost. 0 200 400 600 800 1000 Optimizer Steps Time (seconds) Optimizer Steps vs Time Batch=100 Batch=500 Batch=1000 Nosie network inference time (Batch=100) (a) Optimizing time and steps 101 102 103 Sampling Steps Time (in seconds) Sampling Steps vs Time SN-DDPM GMS (b) Sampling time and steps Figure 5: Sampling/optimizing steps and time (in seconds). (a). The correlation between the number of optimizing steps and the corresponding optimizing time for GMS. Notably, the optimizing. (b). The correlation between the number of sampling steps and the corresponding sampling time for a single batch is observed. Notably, the sampling time demonstrates nearly linear growth with the increase in the number of sampling steps for both solvers. Since many parts in the GMS introduce additional computational effort, Fig. 6 provides detailed information on the additional computational time of the GMS and SN-DDPM relative to the DDPM, assuming that the inference time of the noise network is unit one. Emphasizing that the extra time is primarily for reference, most pixels can work well with a Gaussian distribution. By applying a threshold and optimizing only when the disparity between the reverse transition kernel and Gaussian exceeds it, we can conserve about 4% of computational resources without compromising result quality. Figure 6: Time cost distribution for GMS. E.7 The reduction in the FID on CIFAR10 for GMS compared to SN-DDPM To provide a more intuitive representation of the improvement of GMS compared to SN-DDPM at different sampling steps, we have constructed Fig. 7 below. As observed from Fig. 7, GMS exhibits a more pronounced improvement when the number of sampling steps is limited. However, as the number of sampling steps increases, the improvement of GMS diminishes, aligning with the hypothesis of our non-Gaussian reverse transition kernel as stated in the main text. However, we respectively clarify that due to the nonlinear nature of the FID metric, the relative/absolute FID improvements are not directly comparable across different numbers of steps E.8 Additional results with latent diffusion we conduct an experiment on Image Net 256 256 with the backbone noise network as U-Vi THuge [4]. We train extra the second-order and the third-order noise prediction heads with two transformer blocks with the frozen backbone. The training iterations for each head is 150K and other training parameters are the same as the training of the backbone of U-Vi T-Huge (details in [4]). We will evaluate the sampling efficiency of GMS under conditional sampling (classifier-free guidance scale is 0.4) and unconditional sampling in Tab. 9 25 50 75 100 125 150 175 200 Sampling Steps Relative FID Reduction GMS vs. SN-DDPM FID Reduction Relative FID Reduction Figure 7: The reduction in the FID on CIFAR10 for GMS compared to SN-DDPM. With a restricted number of sampling steps, the improvement exhibited by GMS surpasses that achieved with an ample number of sampling steps. Table 9: Comparison with DDPM and SN-DDPM w.r.t. FID score on Image Net 256 256 by using latent diffusion. Uncond denotes the unconditional sampling, and Cond denotes sampling combining 90% conditional sampling with classifier-free guidance equals 0.4 and 10% unconditional sampling, consistent with Bao et al. [4]. COND, CFG=0.4 UNCOND # TIMESTEPS K 15 20 25 40 50 100 200 25 40 50 100 DDPM, βt 6.48 5.30 4.86 4.42 4.27 3.93 3.32 8.62 6.47 5.97 5.04 SN-DDPM 4.40 3.36 3.10 2.99 2.97 2.93 2.82 8.19 5.73 5.32 4.60 GMS (OURS) 4.01 3.07 2.89 2.88 2.85 2.81 2.74 7.78 5.42 5.03 4.45 E.9 Additional results with same calculation cost Since GMS will cost more computation in the process of fitting the Gaussian mixture, we use the maximum amount of computation required (i.e., an additional 10% of computation is needed) for comparison, and for a fair comparison, we let the other solvers take 10% more sampling steps. Our GMS still outperforms existing SDE-based solvers with the same (maximum) computation cost in Tab. 10. Table 10: Fair comparison with competitive SDE-based solvers w.r.t. FID score on CIFAR10. SN-DDPM denotes Extended Analytic DPM from [2]. The number of sampling steps for the GMS is indicated within parentheses, while for other solvers, it is represented outside of parentheses. CIFAR10 (LS) # TIMESTEPS K 11(10) 22(20) 28(25) 44(40) 55(50) 110(100) 220(200) 1000(1000) SN-DDPM 17.56 7.74 6.76 4.81 4.23 3.60 3.20 3.65 GMS (OURS) 17.43 7.18 5.96 4.52 4.16 3.26 3.01 2.76 In order to provide a more intuitive representation for comparing different methods under the same sampling time, we have generated Fig. 8. Subfig. (a) of Fig. 8 illustrates that at each step, there is an additional computational time cost incurred for GMS to fit the Gaussian mixture transition kernel. This computational overhead exhibits a nearly linear growth pattern with the increase in the number of sampling steps. Subfig. (b) of Fig. 8 offers a valuable perspective on the connection between approximate sampling time and sampling quality for two GMS and SN-DDPM. It becomes apparent that GMS consistently exhibits superior performance when compared to SN-DDPM with identical computational resources. Furthermore, the data reveals that the magnitude of improvement introduced by GMS is more substantial when the number of sampling steps is limited, as opposed to scenarios with a higher number of sampling steps. 101 102 103 Sampling Steps Time (in seconds) Sampling Steps vs Time SN-DDPM GMS (a) Sampling steps and sampling time (b) sampling time and quality Figure 8: (a). The correlation between the number of sampling steps and the corresponding sampling time for a single batch is observed. Notably, the sampling time demonstrates nearly linear growth with the increase in the number of sampling steps for both solvers.(b).The relation between the sample quality (in FID) and the sampling time (in seconds) of GMS and SN-DDPM on CIFAR10.) E.10 Compare with continuous time SDE-based solvers For completeness, we compare the sampling speed of GMS and non-improved reverse transition kernel in Tab. 11, and it can be seen that within 100 steps, our method outperforms other SDE-based solvers. It is worth noting that the EDM-SDE [18] is based on the x0 prediction network, while SEED [10] and GMS are based on Ho et al. [14] s pre-trained model which is a discrete-time noise network. Table 11: Comparison with SDE-based solvers [18, 10] w.r.t. FID score on CIFAR10. # TIMESTEPS K 10 20 40 50 100 SEED-2 [10] 481.09 305.88 51.41 11.10 3.19 SEED-3 [10] 483.04 462.61 247.44 62.62 3.53 EDM-SDE [18] 35.07 14.04 9.56 5.12 2.99 GMS (OURS) 17.43 7.18 4.52 4.16 3.26 E.11 Codes and License In Tab.12, we list the code we used and the license. Table 12: codes and license. URL CITATION LICENSE HTTPS://GITHUB.COM/W86763777/PYTORCH-DDPM HO ET AL. [14] WTFPL HTTPS://GITHUB.COM/OPENAI/IMPROVED-DIFFUSION NICHOL AND DHARIWAL [30] MIT Fig. 9 illustrates one of the comprehensive procedures of SDEdit. Given a guided image, SDEdit initially introduces noise at the level of σt0 to the original image x0, where σ denotes the noise schedules of forward process in diffusion models. Subsequently, using this noisy image xt0 and then discretizes it via the reverse SDE to generate the final image. Fig. 9 shows that the choice of t0 will can greatly will greatly affect the realism of sample images. With the increase of t0, the similarity between sample images and the real image is decreasing. Hence, apart from conducting quantitative evaluations to assess the fidelity of the generated images, it is also crucial to undertake qualitative evaluations to examine the outcomes associated with different levels of fidelity. Taking all factors into comprehensive consideration, we have selected the range of t0 from 0.3T to 0.5T in our experiments. Figure 9: t0 denotes the timestep to noise the stroke Besides experiments on LSUN 256 256, we also carried out an experiment of SDEdit on Image Net 64 64. In Tab. 13, we show the FID score for different methods in different t0 and different sample steps. Our method outperforms other SDE-based solvers as well. Table 13: Comparison with competitive methods in SDEdit w.r.t. FID score on Image Net 64 64. ODE-based solver is worse than all SDE-based solvers. With nearly the same computation cost, our GMS outperforms existing methods in most cases. IMAGENET 64X64, t0 = 1200 # K 26(28) 51(55) 101(111) 201(221) DDPM, βn 21.37 19.15 18.85 18.15 DDIM 21.87 21.81 21.95 21.90 SN-DDPM 20.76 18.67 17.50 16.88 GMS 20.50 18.37 17.18 16.83 From Fig. 10 to Fig. 12, we show generated samples of GMS under a different number of steps in CIFAR10 and Image Net 64 64. Here we use K to denote the number of steps for sampling. (a) GMS (K = 10) (b) GMS (K = 20) (c) GMS (K = 25) (d) GMS (K = 40) (e) GMS (K = 50) (f) GMS (K = 100) (g) GMS (K = 200) (h) GMS (K = 1000) Figure 10: Generated samples on CIFAR10 (LS) (a) GMS (K = 25) (b) GMS (K = 50) (c) GMS (K = 100) Figure 11: Generated samples on Image Net 64 64. (a) GMS (K = 200) (b) GMS (K = 400) (c) GMS (K = 4000) Figure 12: Generated samples on Image Net 64 64.