# efficient_diffusion_models_a_survey__c3728e54.pdf Published in Transactions on Machine Learning Research (05/2025) Efficient Diffusion Models: A Survey Hui Shen1, shen.1780@osu.edu Jingxuan Zhang2, jz97@iu.edu Boning Xiong3, bnxiong24@m.fudan.edu.cn Rui Hu4, marshallrui@gmail.com Shoufa Chen5 shoufachen66@gmail.com Zhongwei Wan1 wan.512@osu.edu Xin Wang1 wang.15980@osu.edu Yu Zhang6 zhangyu.ansel@gmail.com Zixuan Gong6 gongzx@tongji.edu.cn Guangyin Bao6 baogy@tongji.edu.cn Chaofan Tao5 tcftrees@gmail.com Yongfeng Huang7 1155187959@link.cuhk.edu.hk Ye Yuan8 yuanye_pku@pku.edu.cn Mi Zhang1 mizhang.1@osu.edu 1The Ohio State University 2Indiana University 3Fudan University 4Hangzhou City University 5The University of Hong Kong 6Tongji University 7The Chinese University of Hong Kong 8Peking University Reviewed on Open Review: ht tp s: // op en re vi ew .n et /f or um ?i d= w H EC k B Ow yt Diffusion models have emerged as powerful generative models capable of producing highquality contents such as images, videos, audio, and text, demonstrating their potential to revolutionize digital content generation. However, these capabilities come at the cost of their significant resource demands and lengthy generation time, underscoring the need to develop efficient techniques for practical deployment. In this survey, we provide a systematic and comprehensive review of research on efficient diffusion models. We organize the literature in a taxonomy consisting of three main categories, covering distinct yet interconnected efficient diffusion model topics from algorithm-level, system-level, and framework perspective, respectively. We have also created a Git Hub repository where we organize the papers featured in this survey at https://github.com/AIo T-MLSys-Lab/Efficient-Diffusion Model-Survey. We hope our survey can serve as a valuable resource to help researchers and practitioners gain a systematic understanding of efficient diffusion model research and inspire them to contribute to this important and exciting field. The marker denotes Co-first authors. Published in Transactions on Machine Learning Research (05/2025) 1 Introduction Diffusion models kickstart a new era in the field of artificial intelligence generative content (AIGC), garnering unprecedented attention (Yang et al., 2023b; Croitoru et al., 2023). Especially in the context of image synthesis tasks, diffusion models have demonstrated impressive and diverse generative capabilities. The powerful cross-modal capabilities of diffusion models have also further fueled the vigorous development of downstream tasks (Chen et al., 2023b). Despite the increasing maturity of diffusion model variants after numerous iterations (Zhang et al., 2023d; Xu et al., 2023a), generating high-resolution complex natural scenes remains both time-consuming and computationally intensive, whether the initial pixel-level approach (Ho et al., 2020) or the latent space variant (Rombach et al., 2022). Therefore, in order to optimize user-level deployment of diffusion models, researchers have never ceased their pursuit of efficient diffusion models. Despite the growing popularity of diffusion models in recent years, one of the significant issues with diffusion model is that its multi-step denoising procedure requires numerous timesteps to reconstruct a high-quality sample from noise. This multi-step process is not only time-consuming but also computationally intensive, resulting in a heavy workload. Therefore, improving the efficiency of diffusion models is crucial. In recent years, various studies have been presented to address this problem, such as controlling the noise added during training (Hang & Gu, 2024; Chen et al., 2023a) and selecting appropriate sampling timesteps (Watson et al., 2021; Sabour et al., 2024), among others. While there are numerous comprehensive surveys on diffusion models (Yang et al., 2023b; Chen et al., 2024; Croitoru et al., 2023; Cao et al., 2024) and those focused on specific fields and tasks (Ulhaq et al., 2022; Lin et al., 2024c; Kazerouni et al., 2023; Lin et al., 2024b; Peng et al., 2024b; Daras et al., 2024), discussions on the efficiency of diffusion models are notably scarce. The only existing survey addressing efficiency (Ma et al., 2024c) serves as an initial exploration in this area. In our work, we provide a more comprehensive and detailed taxonomy of efficient techniques, covering a broader and more recent collection of research papers. The overarching goal of this survey is to provide a holistic view of the technological advances in efficient diffusion models from algorithm-level, system-level, and framework perspectives, as illustrated in Figure 1. These four categories cover distinct yet interconnected research topics, collectively providing a systematic and comprehensive review of efficient diffusion models research. Specifically, Algorithm-Level Methods: Algorithm-level methods are critical for improving the computational efficiency and scalability of diffusion models, as their training and inference processes are often resource-intensive. In 3, we survey efficient techniques that cover research directions related to efficient training, efficient fine-tuning, efficient sampling, and model compression. System-Level Methods: System-level methods aim to optimize the infrastructure and computational resources required for training and deploying diffusion models. In 4, we survey efficient techniques that cover research directions related to optimized hardware-software co-design, parallel computing, and caching techniques. Frameworks: The advent of diffusion models necessitates the development of specialized frameworks to efficiently handle their training, fine-tuning, inference, and serving. While mainstream AI frameworks such as Tensor Flow and Py Torch provide the foundations, they lack built-in support for specific optimizations and features crucial for diffusion models. In 5, we survey existing frameworks specifically designed for efficient diffusion models, covering their unique features, underlying libraries, and specializations. In addition to the survey, we have established a Git Hub repository where we compile the papers featured in this survey at https://github.com/AIo T-MLSys-Lab/Efficient-Diffusion-Model-Survey. We will actively maintain it and incorporate new research as it emerges. Published in Transactions on Machine Learning Research (05/2025) Efficient Diffusion Algorithm ( 3) Efficient Training ( 3.1) Latent Diffusion Loss Formulation Score Matching Rectified Flow Training Tricks Data-Dependent Adaptive Priors Extra Loss Supervision Noise Schedule Efficient Fine-tuning ( 3.2) Control Net Efficient Sampling ( 3.3) Efficient Solver Sampling Scheduling Parallel Sampling Timestep Schedule Truncated Sampling Retrieval-Based Diffusion Knowledge Distillation Vector Field Distillation Generator Distillation Compression ( 3.4) Quantization Post-Training Quantization Quantization-Aware Training Pruning System ( 4) Hardware-Software Co-Design ( 4.1) Parallel Computing ( 4.2) Caching Technique ( 4.3) Frameworks( 5) Flash Attention, x Formers, Deep Speed, One Flow, Stable-Fast, Onediff, Deep Cache, TGATE, x Di T Figure 1: Taxonomy of efficient diffusion model literature. 2 Background and Applications 2.1 Basic formulas for Diffusion models Diffusion models have emerged as powerful generative models capable of producing high-quality samples across various domains. This section explores the theoretical foundations and recent advancements in the diffusion model framework. We begin by examining the fundamental formulation of Denoising Diffusion Probabilistic Models (DDPMs) (Ho et al., 2020), which leverage forward and backward stochastic processes to gradually transform data into noise and then reverse this process for generation. We then explore Score Matching (Hyvärinen & Dayan, 2005) as an alternative formulation that directly optimizes gradient fields of probability densities. Continuing our discussion, we analyze how Stochastic Differential Equations (SDEs) and Ordinary Differential Equations (ODEs) (Song et al., 2020c) provide continuous perspectives on diffusion. Finally, we introduce Flow Matching (Lipman et al., 2022) as a recent paradigm that offers improved efficiency by directly learning vector fields that transform distributions, utilizing ODEs to provide a deterministic process. To better understand the directions for improving efficient diffusion models, it is essential first to comprehend the fundamental framework of diffusion models. Denoising Diffusion Probabilistic Models (DDPMs) (Ho et al., 2020) generate data through a process analogous to thermodynamic diffusion, consisting of two key components: a forward process and a reverse process. These processes work in concert to enable high-quality generative modeling. The forward process in DDPM is a fixed Markov chain involving gradually adding Gaussian noise to the data until it becomes pure noise. q(x0) is denoted as the true data distribution, and assuming that x0 q(x0) represents sampled data from this distribution. The forward process noted as q(x1:T |x0), adds Gaussian Published in Transactions on Machine Learning Research (05/2025) noise step by step, transforming the data from x0 to x T : q(x1:T |x0) := t=1 q(xt|xt 1), q(xt|xt 1) := N(xt; αtxt 1, βt I) (1) βt is defined as the variance of the noise added at each timestep. We then convert this to αt = 1 βt. Additionally, αt = Qt s=1 αs is defined as the cumulative product of αt, following the formulation by Sohl Dickstein et al. (Sohl-Dickstein et al., 2015). This cumulative product allows for modeling the transition from the original data x0 to xt as a Gaussian distribution: q(xt|x0) = N(xt; αtx0, (1 αt)I) (2) This expression describes the distribution of xt given the initial data x0. It indicates that xt can be expressed as a linear combination of x0 and noise, where ϵ N(0, I) represents standard Gaussian noise: xt = αtx0 + 1 αtϵ (3) The reverse process, in contrast, aims to gradually denoise and reconstruct the original data by reversing the noise addition performed in the forward process. This reverse process is modeled as a Markov chain where each step transitions from xt to xt 1 using a learned conditional probability distribution pθ(xt 1|xt). The overall process is expressed as: pθ(x0:T ) := p(x T ) t=1 pθ(xt 1|xt), pθ(xt 1|xt) := N(xt 1; µθ(xt, t), Σθ(xt, t)) (4) where p(x T ) is the initial Gaussian distribution at the final time step T, and pθ(xt 1|xt) represents the conditional probability distribution learned by the model to transition between states. The mean µθ(xt, t) and covariance Σθ(xt, t) are parameterized functions of the state xt, the time step t, and the model parameters θ. In the training process, the optimization objective is to minimize the negative log-likelihood using the variational bound to approximate the true data distribution: E[ log pθ(x0)] Eq log pθ(x0:T ) q(x1:T |x0) log p(x T ) X t 1 log pθ(xt 1|xt) This objective function decomposes the optimization problem into KL divergences for each timestep, progressively optimizing the reverse process. Expanding the KL terms and using the conditional Gaussian form evaluates the difference between the forward and reverse processes, ultimately simplifying the process into a mean squared error form: Lsimple(θ) := Et,x0,ϵ ϵ ϵθ( αtx0 + 1 αtϵ, t) 2 (6) 2.1.2 Score Matching Score matching, introduced by Hyvärinen & Dayan (2005), serves as an effective approach for estimating unnormalized models by minimizing the Fisher divergence between the score function of data distribution sd(x) = x log pd(x) and the score function of model distribution sm(x; θ) = x log pm(x; θ). This approach avoids the need to compute the intractable partition function Zθ, a common problem in Maximum Likelihood Estimation (MLE). While DDPM directly optimizes the noise prediction in Eq.(6), score matching objectives can directly be estimated on a dataset and optimized with stochastic gradient descent. The loss function for score matching takes a different approach, formulated as follows: 2Epd(x) sm(x; θ) sd(x) 2 (7) Published in Transactions on Machine Learning Research (05/2025) Since it typically does not have access to the true score function of the data sd(x), Hyvärinen & Dayan (2005) introduced integration by parts as L(θ) = J(θ) + C to derive an alternative expression that does not require direct access to xd(x): J(θ) = Epd(x) tr( xsm(x; θ)) + 1 2 sm(x; θ) 2 (8) In this expression, tr( ) denotes the trace of the Hessian matrix of sm(x; θ). The constant C is independent of θ and can be ignored for optimization purposes. The final form of the unbiased estimator used for training is: ˆJ(θ; x N 1 ) = 1 tr( xsm(xi; θ)) + 1 2 sm(xi; θ) 2 (9) 2.1.3 Solvers Given that the cost of sampling escalates proportionally with the number of discretized time steps, many researchers have concentrated on devising discretization schemes that reduce the number of time steps. A key insight emerges from reexamining the discrete forward process in the original DDPM formulation Eq.(1), as we reduce the step size between consecutive steps, the process naturally approaches a continuous transformation. Consequently, adopting learning-free methods using SDE or ODE solvers (Song et al., 2020c) has been proposed. SDE Solver. Song et al. (2020c) firstly presents a stochastic differential equation (SDE) that smoothly transforms a complex data distribution to a known prior distribution by slowly injecting noise and a corresponding reverse-time SDE that transforms the prior distribution back into the data distribution by slowly removing the noise.The discrete noise addition steps in Eq.(1) are reformulated into a continuous process: SDE accomplishes the transformation from data to noise in the diffusion training process through the following equation: dx = f(x, t) dt + g(t) d w (10) where w denotes the standard Wiener process, also known as Brownian motion. f(x, t) is a vector-valued function called the drift coefficient of x(t), and g(t) is a scalar function. Similarly, the reverse process Eq.(4) can be generalized to a continuous-time formulation: dx = f(x, t) g(t)2 x log qt(x) dt + g(t)d w (11) w is a standard Wiener process when time flows backward from T to 0, dt is an infinitesimal negative timestep and x log qt(x) represent the score function that we mentioned in Eq.(10). In the diffusion process, reversetime SDE converts noise into data gradually. The complete SDE process is shown in Figure 2. ODE Solver. Unlike SDE solvers, the trajectories generated by ordinary differential equation (ODE) solvers are deterministic (Song et al., 2020c), remaining unaffected by stochastic variations. Consequently, these Forward SDE (data noise) Reverse SDE (noise data) score function Figure 2: Overview of forward SDE process and reverse SDE process (Song et al., 2020c). Published in Transactions on Machine Learning Research (05/2025) deterministic ODE solvers tend to achieve convergence more rapidly compared to their stochastic counterparts, although this often comes at the expense of a marginal reduction in sample quality. The corresponding deterministic process Eq.(12) can be derived from the reverse-time SDE Eq.(11) by removing the stochastic term g(t)d w, resulting in a deterministic process that shares the same marginal probability densities as the reverse-SDE: dx = f(x, t) 1 2g(t)2 x log qt(x) dt (12) The forward process also exhibits a similar distinction between SDE and ODE approaches, yielding a deterministic process that preserves the same marginal distributions: dx = f(x, t)dt (13) 2.1.4 Flow Matching Flow Matching (FM) (Lipman et al., 2022) is a new paradigm for generative modeling based on Continuous Normalizing Flows (CNFs). This approach allows us to train CNFs at an unprecedented scale and offers greater efficiency compared to traditional diffusion models. To better understand the process of flow matching, we need to dive into the preliminaries. First of all, we make Rd denote the data space with data points x = (x1, . . . , xd) Rd. Then, we need to understand several key concepts: Probability density path pt : [0, 1] Rd R>0, which is a time-dependent probability density function, and Time-dependent vector field vt : [0, 1] Rd Rd, which describes how data points change over time. A vector field vt can be used to construct a time-dependent diffeomorphic map, defined by an ordinary differential equation (ODE): d dtϕt(x) = vt(ϕt(x)) (14) ϕ0(x) = x (15) A Continuous Normalizing Flow (CNF) is a generative model that parametrizes the time-dependent vector field vt(x; θ) using a neural network (Chen et al., 2018), where θ represents the learnable parameters. This vector field defines a flow ϕt that transforms the probability distribution from a simple prior density p0 to a more complex target density p1 through the push-forward equation. The core of CNF is using this flow to reshape probability distributions: pt = [ϕt] p0 (16) The push-forward operator is defined as: [ϕt] p0(x) = p0(ϕ 1 t (x)) det ϕ 1 t x (x) (17) The FM loss is formally defined as: LFM(θ) = Et U[0,1], x pt(x) vt(x) ut(x) 2 , (18) where: ut(x) generates a predefined probability density path pt(x), p0(x) = N(x|0, I) is a simple prior distribution (e.g., Gaussian noise), and p1(x) q(x) approximates the data distribution. While FM is conceptually straightforward, directly optimizing this objective is intractable due to the lack of closed-form expressions for pt and ut. To address this, FM constructs conditional probability paths pt(x|x1) and vector fields ut(x|x1) per data sample x1 q(x1), where p0(x|x1) = p(x) at time t = 0, and p1(x|x1) = N(x|x1, σ2I) at t = 1 (a normal distribution with x1 mean and a sufficiently small standard deviation σ > 0). These conditional paths are then aggregated into global counterparts through marginalization: pt(x) = Z pt(x|x1)q(x1)dx1 (19) Published in Transactions on Machine Learning Research (05/2025) This approach avoids explicit modeling of the intractable marginal distributions and enables scalable training, while ensuring that at t = 1, the marginal distribution p1(x) approximates the data distribution q(x). The marginal vector field can then be defined as: ut(x) = Z ut(x|x1)pt(x|x1)q(x1) pt(x) dx1 (20) A key theorem in the paper shows that this constructed marginal vector field ut in Eq.(20) indeed generates the marginal probability path pt in Eq.(19). However, computing the FM objective is still infeasible because it involves complex integrals. To address this issue, the paper proposes the Conditional Flow Matching (CFM) objective: LCFM(θ) = Et,q(x1),pt(x|x1) vt(x) ut(x|x1) 2, (21) The second key theorem in the paper proves that the FM and CFM objectives have identical gradients with respect to θ, so optimizing CFM is equivalent to optimizing FM, but computationally simpler. Compared to score matching, flow matching is faster because score matching typically requires computing Hessian matrices in Eq(9), while CFM cleverly avoids such complex calculations by only computing the L2 distance between vector fields. In conclusion, it s worth noting that Flow Matching can be theoretically connected with other diffusion model formulations. As summarized by Gao et al. (2024), flow matching, score-based models, and traditional diffusion models can be understood within a unified mathematical framework. These approaches essentially represent different parameterizations of the same continuous-time process. While traditional diffusion models like DDPM (Ho et al., 2020) learn to estimate noise or score functions, and score-based models directly optimize gradient fields (Hyvärinen & Dayan, 2005), flow matching learns the vector field itself, directly capturing the transformation map between distributions. 2.2 Applications Building on the foundational principles of diffusion models outlined in 2.1, this section surveys their practical deployment across diverse generative tasks, with a specific emphasis on efficiency-driven innovations. As computational demands and real-world applicability become increasingly critical, researchers have adapted diffusion models to optimize resource usage while preserving or enhancing output quality. Here, we explore key application domains image, video, text, audio, and 3D generation highlighting techniques that reduce inference time, memory footprint, and training complexity. These advancements underscore the adaptability of diffusion models and their growing impact in addressing the challenges of scalable, high-fidelity content generation. 2.2.1 Image Generation Image generation is the primary application domain for efficient diffusion models. Researchers have been developing various approaches to optimize both computational resources and generation quality. The efficiency improvements in this field are well exemplified by several influential works. For example, Stable Diffusion (Rombach et al., 2022) pioneered the concept of efficient image generation by operating in a compressed latent space rather than pixel space, significantly reducing memory and computational requirements while maintaining high-quality outputs. Latent Consistency Models (LCM) (Luo et al., 2023a) further pushed the boundaries by enabling high-quality image generation in just 4 steps through careful design of the consistency loss and distillation process. Progressive distillation (Salimans & Ho, 2022) demonstrated that through a student-teacher framework, diffusion models could achieve comparable quality to 50-step sampling using only 2-8 inference steps. Control Net (Zhang et al., 2023d) introduced an efficient architecture for adding spatial conditioning controls to pretrained diffusion models through zero-initialized convolutions, enabling diverse control capabilities without compromising model efficiency. More recently, Efficient Diffusion (EDM) (Karras et al., 2022) presented a comprehensive framework for training and sampling diffusion models more efficiently, introducing improvements in both training stability and inference speed while maintaining state-of-the-art generation quality. Published in Transactions on Machine Learning Research (05/2025) 2.2.2 Video Generation Following the rapid escalation in image generation, video generation similarly garnered widespread attention (Melnik et al., 2024; Ho et al., 2022b; Xing et al., 2023). The heavy model size and the substantial computational costs have further intensified the focus on developing more efficient methods for video generation (Zhang et al., 2023b; Liu et al., 2023b; Xing et al., 2024; Wang et al., 2023; Lee et al., 2024b). For example, Zhang et al. (2023b) introduced Ada Diff, a lightweight framework designed to optimize a specialized policy gradient method tailored to individual text prompts. This approach facilitates the design of reward functions and enables an effective trade-off between inference time and generation quality. Specifically to the training process, Liu et al. (2023b) proposed an efficient training framework ED-T2V to freeze the pretraining model (Rombach et al., 2022) training additional temporal modules. Similarly, Xing et al. (2024) suggested using spatial and temporal adapters. In their approach, the original T2I model remains frozen during training, and only the newly added adapter modules are updated. Unlike the works above, Wang et al. (2023) presented Video LCM, incorporating consistency distillation in the latent space. Video LCM efficiently distills knowledge from a pretraining model, maintaining fidelity and temporal coherence while improving inference speed. Lee et al. (2024b) introduces a grid diffusion model by representing a video as a grid of images. It employs key grid image generation and autoregressive grid interpolation to maintain temporal consistency. Moreover, Ceylan et al. (2023) leverages self-attention feature injection and guided latent updates, efficiently repurposing image models for video editing, enabling high-quality, consistent edits at minimal computational overhead. Yin et al. (2023) proposes NUMA-XL, using a hierarchical coarse-tofine approach to enable parallelizable, exponential scaling of video length. It achieves a great reduction in inference time and is trained directly on long sequences, ensuring long-term consistency. 2.2.3 Text Generation Efficient diffusion models offer a fresh perspective in text generation through their stochastic and iterative processes. However, they encounter several challenges when applied to discrete data types such as text. For instance, the common use of Gaussian noise is less effective for discrete corruption, and the objectives designed for continuous spaces become unstable in the text diffusion process, particularly at higher dimensions. With these challenges, Chen et al. (2023a) proposed a diffusion model called Masked-Diffuse LM. In the diffusion process, a cross-entropy loss function at each diffusion step is utilized to efficiently bridge the gap between the continuous representations in the model and the discrete textual outputs. Seq Diffu Seq (Yuan et al., 2024) incorporates an encoder-decoder Transformer architecture, achieving efficient text generation through adaptive noise schedule and self-conditioning (Chen et al., 2022a) techniques. Using the same encoderdecoder architecture, Lovelace et al. (2024) presents a methodology where text is encoded into a continuous latent space. Subsequently, continuous diffusion models are employed for sampling within this space. 2.2.4 Audio Generation In the field of audio generation, the application of diffusion models presents several unique challenges. First, audio data exhibits strong temporal continuity, especially in high-resolution audio generation tasks, where the model must accurately reconstruct both time-domain and frequency-domain information. Compared to images or text, even subtle distortions or noise in audio are easily perceptible by humans, directly affecting the listening experience, particularly in speech and music generation tasks. Ensuring high fidelity and maintaining the consistency of details in the generated audio is therefore crucial. Moreover, many audio generation tasks require low-latency feedback, especially in applications like speech synthesis and real-time dialogue, which makes acceleration of diffusion models essential. The multi-dimensional nature of audio data, such as time-domain, frequency-domain, stereo, and spatial audio, further complicates the generation process, requiring the model to handle these dimensions while maintaining consistency during the accelerated generation. To address these challenges, researchers have proposed various methods to accelerate diffusion models in audio generation. Some works focus on reducing the number of diffusion steps to speed up the generation process, such as Chen et al. (2020) in Wave Grad and Kong et al. (2020) in Diff Wave, which optimize the network structure to reduce generation time while maintaining high audio quality. Further optimization comes from the Fast DPM framework (Kong & Ping, 2021), which generalizes discrete diffusion steps to continuous ones, using a bijective mapping between noise levels to accelerate sampling without Published in Transactions on Machine Learning Research (05/2025) compromising quality. Fast DPM s flexibility allows it to adapt to different domains, and in the case of audio synthesis, where stochasticity plays a crucial role, it demonstrates superior performance in high-stochasticity tasks like speech generation. Through these approaches, diffusion models not only accelerate the generation process but also reduce computational costs while ensuring that audio quality remains high, meeting the demands of real-time audio generation applications. 2.2.5 3D Generation As a technique closely aligned with real-world representation, 3D generation holds substantial promise across various sectors, including medical imaging, motion capture, asset production, and scene reconstruction, etc. However, when compared to 2D image generation, distinctive high-resolution elements such as volumetric data or point clouds present unique challenges, significantly escalating computational demands. Several efficient methodologies (Bieder et al., 2023; Zhou et al.; Tang et al., 2023b; Park et al., 2023; Du et al., 2024; Wu et al., 2024) have been proposed, particularly concentrating on enhancing the sampling process and optimizing the architectural framework, which further handles the computational complexity inherent. One of the most prevalent approaches involves designing more efficient sampling schedules (Bieder et al., 2023; Li et al., 2024c; Yu et al., 2024b; Zhou et al.). By utilizing larger sampling step sizes, modifying the sampling strategy between 2D and 3D, or incorporating multi-view parallelism, these techniques address the key bottlenecks in the sampling process, thereby improving sampling efficiency. Moreover, the incorporation of novel architectures, such as state-space models and lightweight feature extractors (Mo, 2024; Tang et al., 2023b), alleviates the computational burden of processing 3D data, significantly enhancing model efficiency. Table 1: Representative applications of diffusion models. Task Datasets Metrics Articles Image Generation Image Net, CIFAR, Met Face, Celeb A HQ, MS COCO, UCI, FFHQ, Diffusion DB, AFHQ, LSUN, SYSTEM-X, LAION FID, s FID, IS, NLL, MSE, CLIP Score, PSNR, LPIPS, MACs, CS, Pick Score, SA, Score Matching Loss Liu et al. (2022b), Liu et al. (2023c), Yan et al. (2024), Lee et al. (2024a), Zhu et al. (2024), etc. Video Generation MSR-VTT, Intern Vid, Web Vid-10M, LAION, UCF-101, CGCaption, DAVIS, Flintstones HD FID, IS, FVD, IQS, NIQE, CLIPSIM, B-FVD-16, Pixel-MSE Zhang et al. (2023b), Liu et al. (2023b), Xing et al. (2024), Wang et al. (2023), Lee et al. (2024b), Ceylan et al. (2023), Yin et al. (2023), etc. Audio Generation SC09, LJSpeech, Speech Commands MOS, FID, IS, m IS, AM Score Chen et al. (2020), Kong et al. (2020), Kong & Ping (2021), etc. Text Generation XSUM, Semantic Content, CCD, IWSLT14, WMT14, ROCStories, E2E, QQP, Wiki-Auto, Quasar-T, AG News Topic Rouge, Semantic Acc, Mem, BLEU, Div, BERTScore, Sacre BLEU, MAUVE Score, Content Fluency, POS Chen et al. (2023a), Yuan et al. (2024), Chen et al. (2022a), Lovelace et al. (2024), etc. 3D Generation Bra TS2020, Shape Net, Objaverse, Cap3D, LLFF, Human ML3D, AMASS, KIT, Human Act12, IBRNet, Instruction-Ne RF2Ne RF Dice, HD95, CD, EMD, 1-NNA, COV, CLIP, Aesthetic, Similarity, R-Precision, FID, DIV, MM-Dist, ACC, Diversity, MModality Bieder et al. (2023), Mo (2024), Li et al. (2024c), Park et al. (2023), Yu et al. (2024b), etc. Published in Transactions on Machine Learning Research (05/2025) Efficient Training Latent Diffusion Video LDM (Blattmann et al., 2023), Magic Video (Zhou et al., 2022), PVDM (Yu et al., 2023), LVDM (He et al., 2022), Latent 3D Graph Diffusion (You et al., 2024), Motion Latent-based Diffusion (Chen et al., 2023c), Audio LDM (Liu et al., 2023a), L3DG (Roessle et al., 2024), Tab Syn (Zhang et al., 2023a), Policy Diffusion (Hegde et al., 2023), RNAdiffusion (Huang et al., 2024) Loss Formulation Score Matching Score Matching (Hyvärinen & Dayan, 2005), Sliced Score Matching (Song et al., 2020b), NCSN (Song & Ermon, 2019), Likelihood Weighting (Song et al., 2021), CLD (Dockhorn et al., 2021) Rectified Flow Rectified Flow (Liu et al., 2022b; Liu, 2022), Insta Flow (Liu et al., 2023c), Pe RFlow (Yan et al., 2024), 2-Rectified Flows++ (Lee et al., 2024a), Slim Flow (Zhu et al., 2024) Training Tricks Data-Dependent Adaptive Priors Prior Grad (Lee et al., 2021), Digress (Vignac et al., 2022), Decomp Diff (Guan et al., 2023), Leapfrog DM (Mao et al., 2023) Extra Loss Supervision REPA (Yu et al., 2024a), VA-VAE (Yao et al., 2025) Noise Schedule DDPM (Ho et al., 2020), IDDPM (Nichol & Dhariwal, 2021), Laplace (Hang & Gu, 2024), Masked-Diffusion LM (Chen et al., 2023a), DDIM (Song et al., 2020a), Res Shift (Yue et al., 2024), Immiscible Diffusion (Li et al., 2024d), Diffu Seq (Gong et al., 2022), Seq Diffu Seq (Yuan et al., 2024) Figure 3: Summary of efficient training techniques for diffusion models. 3 Algorithm-Level Efficiency Optimization 3.1 Efficient Training Efficient training aims to optimize the training process of DMs, reducing computational costs while accelerating convergence. As summarized in Figure 3, enhancing the efficiency of pre-training can be achieved through different and complementary techniques, including latent diffusion, loss formulation, and specialized training tricks. Latent diffusion models compress the optimization process by operating in lower-dimensional latent spaces, though they occasionally struggle with fine-grained detail reconstruction. Loss formulation methods enhance gradient estimation and stability, though sometimes facing numerical challenges. Meanwhile, various training tricks, including data-dependent adaptive priors, extra loss supervision, and optimized noise schedules, further enhance efficiency by leveraging problem-specific knowledge, though each introduces additional hyperparameters requiring careful tuning. 3.1.1 Latent Space Early diffusion models, such as DDPM (Ho et al., 2020), operate directly in pixel space, where images are generated through iterative noise addition and denoising. While these models achieve high-quality results, their reliance on pixel space introduces significant inefficiencies. Each diffusion step involves operations in a high-dimensional space, leading to substantial computational and memory overhead, especially for highresolution images (e.g., 512 512). Additionally, the requirement for hundreds to thousands of iterative steps hinders real-time applications. Although methods like DDIM (Song et al., 2020a) have been proposed to speed up sampling, the pixel-level processing remains a fundamental bottleneck. In contrast, leveraging latent space significantly enhances training efficiency by operating in a lower-dimensional, compact representation, reducing both computational complexity and memory consumption while maintaining high-quality generation. As research progresses, researchers have begun exploring compromise approaches to circumvent the highdimensional pixel space. Consequently, various methods for introducing the latent space have increasingly gained attention from the academic community. To reduce computational complexity, researchers have introduced autoencoders (AEs) and their variants to compress images into a lower-dimensional latent space. Among them, Variational Autoencoders (VAEs ( Kingma et al. (2013))) map images to a Gaussian distribution in the latent space via an encoder, while a decoder reconstructs images from latent variables. Compared to traditional AEs, VAEs learn the data distribution rather than specific samples, mitigating overfitting. However, the generated images often appear blurry due to insufficient latent space constraints, as the KL regularization weight is relatively low. Vector Quantized Variational Autoencoders (VQ-VAEs ( Van Den Oord et al. (2017))) enhance compression efficiency by introducing a discrete latent space through vector quantization. The integration of compression techniques with diffusion models has given rise to various latent space-based diffusion models, with Latent Diffusion Models (LDMs) being the most prominent. Published in Transactions on Machine Learning Research (05/2025) Benefiting from the efficiency advantages of the latent space and the cross-modal cross-attention mechanism of LDMs, a diverse range of image generation(eg. Stable Diffusion( Rombach et al. (2022))) and image editing models(eg. Diff Edit( Couairon et al. (2022))) has emerged. Beyond traditional image synthesis and editing, various other diffusion model-based tasks achieve computational efficiency through the introduction of latent spaces, too. For videos with high-dimensionality, complex temporal dynamics and large spatial variations, learning video distributions within a low-dimensional latent space has proven to be an effective method for efficiently generating high-quality videos. For instance, Video LDM (Blattmann et al., 2023), Magic Video (Zhou et al., 2022), PVDM (Yu et al., 2023) and LVDM (He et al., 2022) apply the latent diffusion model paradigm to video generation. Video LDM focuses on highresolution video generation. Initially, the LDM is pretrained exclusively on images; subsequently, turn the image generator into a video generator by introducing a temporal dimension to the latent space diffusion model and fine-tuning on encoded image sequences, i.e., videos. LVDM proposes hierarchical diffusion in the latent space such that longer videos with more than one thousand frames can be produced. To further overcome the performance degradation issue for long video generation, LVDM introduces conditional latent perturbation and unconditional guidance. You et al. (You et al., 2024) perform 3D graph diffusion in a low-dimensional latent space, which is learned through cascaded 2D 3D graph autoencoders for low-error reconstruction and symmetry-group invariance, resulting in training that is an order of magnitude faster. Motion Latent-based Diffusion (Chen et al., 2023c) is able to produce vivid motion sequences conforming to the given conditional inputs and substantially reduce the computational overhead in both the training and inference stages by performing diffusion process on the motion latent space. Audio LDM (Liu et al., 2023a), a text-to-audio system built on a latent space that learns continuous audio representations from contrastive language-audio pretraining latents, offers advantages in both generation quality and computational efficiency. L3DG (Roessle et al., 2024) utilizes a compressed latent space learned by a vector-quantized variational autoencoder, coupled with a sparse convolutional architecture to efficiently operate on room-scale scenes. As a result, the complexity of the costly 3D Gaussians generation process through diffusion is significantly reduced. Tab Syn (Zhang et al., 2023a) attempt to utilize a diffusion model with a carefully crafted latent space in tabular data synthesis tasks, discovering that generation quality significantly improves while synthesis speed also increases. In reinforcement learning, condensing the archive into a single model while retaining the performance and coverage of the original collection of policies has proved challenging. Hegde et al. (Hegde et al., 2023) propose using latent diffusion models to distill the archive into a single generative model over policy parameters, achieving a compression ratio of 13x. RNAdiffusion (Huang et al., 2024) compresses token-level, biologically meaningful representations of RNA sequences into a set of fixed-length latent vectors and reconstructs RNA sequences from these latent variables. It utilizes a latent diffusion model to achieve controllable and efficiently translated RNA sequence generation. 3.1.2 Loss Formulation In this section, we examine methods that enhance the efficiency of different loss formulations in diffusion models. For score matching (Hyvärinen & Dayan, 2005), we present approaches to reduce its computational costs. For rectified flow (Liu et al., 2022b), we explore how its carefully designed formulation enables straightline sampling trajectories and subsequent improvements that further increase efficiency. We also include Flow Matching s Optimal Transport approach (Lipman et al., 2022) in the rectified flow discussion, as it similarly achieves direct trajectory learning through linear parameterization of probability paths. Score Matching. Compared to DDPM s straightforward optimization in Eq.(6), although score matching Eq.(9) avoids the computation of the partition function Zθ, it still faces computational challenges, particularly in high-dimensional data. The computation of the trace of the Hessian matrix substantially increases the complexity as the dimensionality grows. Specifically, computing the trace requires many more backward passes than the gradient, making score matching computationally expensive for high-dimensional data. Therefore, to address the computational inefficient issue of training process, Song et al. (2020b) observed that one-dimensional problems are typically much easier to solve than high-dimensional ones. Inspired by the idea of the Sliced Wasserstein Distance (Rabin et al., 2012), they proposed Sliced Score Matching. The core idea of sliced score matching is to project both the score function of the model sm(x; θ) and the data sd(x) onto a random direction v, and compare the differences along that direction. The sliced score matching Published in Transactions on Machine Learning Research (05/2025) objective is defined as: L(θ; pv) = 1 2Epv Epd(x) h v sm(x; θ) v sd(x) 2i (22) To eliminate the dependence on sd(x), integration is applied by parts, similar to traditional score matching, resulting in the following form: J(θ; pv) = Epv Epd(x) v xsm(x; θ)v + 1 2(v sm(x; θ))2 (23) which achieves scalability by reducing the complexity of the problem by projecting high-dimensional score functions onto low-dimensional random directions, thereby avoiding the full Hessian computation. While effective for dimensionality reduction, score estimation still faces challenges in low data density regions where data samples are sparse. Roundabout ODE Path Straight ODE Path Rectified Flow Figure 4: Illustration of the rectified flow. Building upon sliced score matching, to address the issue of inaccurate score estimation in low data density regions, Song & Ermon (2019) introduces a novel generative framework that employs Langevin dynamics to produce samples based on estimated gradients of the data distribution pdata(x). They proposed Noise Conditional Score Networks (NCSN) sθ(x, σ), which jointly estimate scores across multiple noise-perturbed data distributions. By conditioning on a geometric sequence of noise levels σ3 > σ2 > σ1, a single network learns to estimate scores for distributions ranging from highly smoothed pσ3(x) that fill low-density regions to concentrated pσ1(x) that preserve the structure of the original data manifold. This unified training approach enables robust score estimation across the entire data space. Following a similar derivation, as Song et al. (2021), Dockhorn et al. (2021) introduces Coupled Langevin Dynamics (CLD), redefining the score matching objective within the CLD framework. Unlike traditional score matching methods that inject noise directly into the data space, CLD simplifies the task by only requiring the model to learn the score of the conditional distribution pt(vt | xt), where noise is injected into an auxiliary variable vt coupled with the data. Figure 5: Illustration of datadependent adaptive priors for diffusion processes across different modalities, demonstrating how tailored priors improve generation on quality. Rectified Flow. As illustrated in Figure 4, Rectified Flow, proposed by (Liu et al., 2022b; Liu, 2022), introduces a method for training ordinary differential equation (ODE) models by learning straight transport paths between two distributions, π0 and π1. The key idea is to minimize the transport cost by ensuring that the learned trajectory between these two distributions follows the most direct route, which can be computationally efficient to simulate. Unlike traditional diffusion models, which may follow roundabout paths, Rectified Flow leverages a simpler optimization problem to create a straight flow, thereby improving both training efficiency and the quality of the generated samples. Flow Matching (Lipman et al., 2022), as discussed in 2.1.4, introduces the Optimal Transport (OT) approach as a significant advancement in generative modeling. The key innovation of the OT method lies in how it parameterizes conditional probability paths. It represents conditional probability paths in Eq(19) using Gaussian distributions with means µt(x) = tx1 and standard deviations σt(x) = 1 (1 σmin)t that change linearly with time. This linear parameterization produces a vector field with constant direction, enabling transitions from noise to data along straight-line trajectories as well. Building upon the foundation of Rectified Flow, Insta Flow (Liu et al., 2023c) applies the Rectified Flow concept to text-to-image generation, achieving a significant breakthrough. Insta Flow represents a major advancement in efficient diffusion models, which are capable of high-quality image generation in just one step. Published in Transactions on Machine Learning Research (05/2025) It applied Rectified Flow to large-scale datasets and complex models like Stable Diffusion, introduced a novel text-conditioned pipeline for one-step image generation, and achieved an FID score of 23.3 on MS COCO 2017-5k. Insta Flow s success highlights the potential of Rectified Flow in dramatically reducing the computational cost of diffusion models while maintaining high output quality. Following Insta Flow, Yan et al. (2024) proposed Pe RFlow, further extending the Rectified Flow concept to create a more flexible and universally applicable acceleration method. Pe RFlow divides the sampling process into multiple time windows, applying the reflow operation to each interval, creating piecewise linear flows that allow for more nuanced trajectory optimization. Through carefully designed parameterizations, Pe RFlow models can inherit knowledge from pretrained diffusion models, achieving fast convergence and superior transfer ability. This approach positions Pe RFlow as a universal plug-and-play accelerator compatible with various workflows based on pretrained diffusion models. While Rectified Flow showed great promise, there was still room for improvement, especially in low Number of Function Evaluations (NFE) settings. Addressing this, Lee et al. (2024a) focused on enhancing the training process of Rectified Flows. They discovered that a single iteration of the Reflow algorithm is often sufficient to learn nearly straight trajectories and introduced a U-shaped timestep distribution and LPIPS-Huber premetric to improve oneround training. These improvements led to significant enhancements in FID scores, particularly in low NFE settings, outperforming state-of-the-art distillation methods on various datasets. Most recently, Zhu et al. (2024) proposed Slim Flow, a method designed to address the joint compression of inference steps and model size within the Rectified Flow framework, introducing Annealing Reflow to address initialization mismatches between large teacher models and small student models, and developing Flow-Guided Distillation to improve performance on smaller student models. 3.1.3 Training Tricks Training tricks encompass a range of practical strategies aimed at improving the efficiency, convergence, and sample quality of diffusion models by optimizing various aspects of the learning process. In this section, we explore three key approaches: data-dependent adaptive priors, which tailor initial distributions to specific data characteristics for faster convergence; extra loss supervision, which introduces additional loss terms to better align the model with data distributions and enhance generation quality; and noise schedule design, which governs the addition and removal of noise to streamline the diffusion process and accelerate sampling. Data-Dependent Adaptive Priors. To enhance the training efficiency of diffusion models and improve the quality of generated samples, data-dependent adaptive priors can be utilized to tailor the prior distribution to specific tasks and datasets. This approach leverages priors that better align with the data distribution, thereby accelerating the training process and ensuring that generated samples more closely match the true data distribution. Recent studies have explored how data-dependent adaptive priors can improve the training of diffusion models. As a method under efficient training (see Figure 5), data-dependent adaptive priors can be applied across various modalities, such as speech, graphs, and trajectories. By aligning the prior with the data distribution specific to each modality, the model can achieve faster convergence during training while producing outputs that better reflect the underlying data structure. In traditional diffusion models, the prior is typically assumed to be a standard Gaussian distribution p(z) = N(0, I). However, this assumption may not align well with the actual data distribution, potentially leading to inefficiencies in training. By constructing datadependent adaptive priors based on the data X, the model can achieve better initialization during training, accelerating convergence without relying solely on the standard Gaussian assumption. Lee et al. (2021) introduced Prior Grad, which enhances the training of diffusion models for speech synthesis by using an adaptive prior derived from conditional data statistics. This method significantly improves the efficiency of the denoising process during training, leading to faster convergence while enhancing the perceptual quality and robustness of generated samples, even with smaller network capacities. Vignac et al. (2022) proposed Di Gress, a discrete denoising diffusion model for graph generation. By leveraging data-dependent priors, this model better captures the discrete nature of graph data, improving training Published in Transactions on Machine Learning Research (05/2025) Proposed leapfrog initializer Diffusion step Denoising step Sample prediction Variance estimation Mean estimation Predicted trajectories Ground-truth future trajectory Random initialization Traditional denoising steps Skipped denoising step Figure 6: The Leapfrog diffusion model (Mao et al., 2023) accelerates inference by using a leapfrog initializer to approximate the denoised distribution, replacing extended denoising sequences while preserving representation capacity. efficiency and the quality of generated graphs, particularly for applications like chemical molecular structures and social networks. In drug design, Guan et al. (2023) introduced Decomp Diff, which improves the training of diffusion models by using decomposed priors to separately model different structural components of drug molecules. This approach enhances the model s ability to capture molecular structure information during training, leading to the generation of more accurate drug candidates. As shown in Figure 6, Mao et al. (2023) proposed the Leapfrog Diffusion Model for stochastic trajectory prediction, introducing a leapfrog initializer based on adaptive priors to skip multiple denoising steps, accelerating training while maintaining accuracy for real-time applications like autonomous driving. Building on this, Fu et al. (2025) developed Mo Flow, a one-step flow matching model with IMLE-based distillation for human trajectory forecasting. Mo Flow employs a novel flow matching loss to ensure accuracy and diversity in predicted trajectories, using data-dependent adaptive priors based on past trajectories and interactions to enhance alignment with the data distribution. Its IMLE distillation achieves a 100x faster one-step student model with comparable performance. Similarly, Jiang et al. (2024) introduced Scene Diffuser, a scene-level diffusion model using amortized diffusion to optimize efficiency for driving simulation, supporting both initialization and rollout. These advances highlight the role of adaptive priors in boosting training efficiency for real-time multimodal applications. Extra Loss Supervision. To further enhance the training efficiency and sample quality of diffusion models, incorporating additional loss supervision has proven to be an effective strategy. Beyond standard denoising objectives, extra loss terms can guide the model toward better alignment with data distributions, accelerate convergence, and improve the robustness of generated outputs. Recent advancements have demonstrated the power of tailoring loss functions to leverage external knowledge or address specific optimization challenges. One notable approach introduces REPresentation Alignment (REPA) (Yu et al., 2024a), a regularization technique that aligns the hidden states of the denoising network with representations from a pre-trained visual encoder, such as DINOv2. By adding a loss term that enforces this alignment, REPA ensures that the diffusion Transformer (e.g., Di T or Si T) leverages high-quality external visual priors rather than learning representations from scratch. This supervision accelerates training achieving over 17.5 speedup in some cases and boosts generation quality, reaching a state-of-the-art FID of 1.42 on Image Net 256x256 with classifier-free guidance. The additional loss acts as a bridge between the model s internal learning and established visual knowledge, making training both faster and more effective. Similarly, another method tackles the optimization dilemma in latent diffusion models (LDMs) by introducing extra loss supervision through the VA-VAE framework (Yao et al., 2025). This approach aligns the latent space of a variational autoencoder with a pre-trained vision foundation model, using a loss term to ensure the latent representations capture the data distribution more effectively. This supervision mitigates the trade-off between reconstruction fidelity and generation quality, a common challenge in LDMs where increasing latent dimensionality often hampers generative performance. Combined with an optimized Lightning Di T architecture, this method achieves an FID of 1.35 on Image Net 256x256, with training convergence accelerated by Published in Transactions on Machine Learning Research (05/2025) over 21 compared to baseline Di T models. The extra loss supervision in VA-VAE enhances the model s ability to balance reconstruction and generation tasks, leading to both efficiency gains and superior sample quality. Noise Schedule. Noise schedule is a crucial component of diffusion models, governing how noise is added during the forward process and removed during the reverse process. Optimizing the noise schedule can significantly enhance the training efficiency of diffusion models by enabling faster convergence and more effective learning of data distributions. Denoising Diffusion Probabilistic Models (DDPM) (Ho et al., 2020) introduced a linear noise schedule that gradually decreases the variance of the noise added in the forward process, as defined in Eq.(1). However, the linear schedule requires calculating complex noise terms across numerous timesteps, which can slow down the training process and highlight the need for more efficient noise schedule designs. As shown in Figure 7, efficient noise schedules can be classified into two main categories: systematic noise addition and dynamic noise adjustment. Figure 7: Illustration of two categories of noise schedules. One approach involves systematically adding noise during the training process at predefined intervals or according to specific levels. DDPM (Ho et al., 2020) employs a linear noise schedule, where the noise variance changes deterministically over time, serving as a foundational example. Building on this, the Improved Denoising Diffusion Probabilistic Model (IDDPM) (Nichol & Dhariwal, 2021) introduces a cosine noise schedule, defined as βt = 1 cos t/T +s cos s 1+s π where t is the current timestep, T is the total number of timesteps, and s is a small positive constant for smoothing initial noise addition. The cosine noise schedule optimizes the noise distribution during training, enabling the model to learn data structures more effectively and achieve faster convergence. However, it allocates computational resources evenly across all noise intensities, which may not prioritize the most critical regions for training. To address this inefficiency, Hang & Gu (2024) proposed the Laplace noise schedule, which enhances training efficiency by increasing the sampling frequency around critical noise regions. This approach ensures that the model focuses computational resources on medium noise intensities, which are more effective for learning data structures and removing noise, leading to faster convergence and improved sample quality during training. The Laplace schedule balances noise addition across timesteps, resulting in a more robust training process. For text generation, traditional diffusion models often add uniform Gaussian noise to each word, which fails to leverage linguistic features and increases computational burden during training. To address this, Chen et al. (2023a) introduced a soft-masking noise strategy that gradually adds noise based on word importance, measured by term frequency and information entropy. Using a square-root noise schedule (Li et al., 2022), this method incrementally increases noise levels, stabilizing the training process and enabling the model to learn linguistic structures more effectively from the initial latent variable X0 to noisy variables X1:T . Published in Transactions on Machine Learning Research (05/2025) In contrast, another set of methods dynamically adjusts the noise schedule based on the model s state or data during training. Song et al. (2020a) developed Denoising Diffusion Implicit Models (DDIM), which improve the noise schedule in DDPM by introducing a non-Markovian forward process, defined as xt 1 = αt 1 xt 1 αtϵθ(xt) + q 1 αt 1 σ2 t ϵθ(xt) + σtϵt, (25) where αt controls the noise addition over time, and σt is dynamically adjusted based on the current state xt and initial state x0. This dynamic adjustment leverages the entire trajectory, enabling more precise noise control during training, which enhances the model s ability to learn data distributions efficiently and improves training convergence. Inspired by DDIM, Yue et al. (2024) proposed Res Shift, which introduces a noise schedule that constructs a shorter Markov chain by shifting residuals between high-resolution (HR) and low-resolution (LR) images instead of adding Gaussian noise. The noise schedule is defined as ηt = η1 bβt 0 , t = 2, . . . , T 1, (26) p (T 1), b0 = exp 1 2(T 1) log ηT with T as the total timesteps, t as the current timestep, p controlling the growth rate of ηt, and η1 and ηT as the initial and final noise levels. This non-uniform noise progression allows the model to focus training on key regions, improving convergence and learning efficiency for super-resolution tasks. To further optimize training, Li et al. (2024d) proposed Immiscible Diffusion, inspired by the physical phenomenon of immiscibility. Unlike traditional methods that diffuse each image across the entire noise space, this approach reassigns noise to images within a mini-batch to minimize the distance between imagenoise pairs. By matching each image with nearby noise, Immiscible Diffusion reduces the complexity of denoising during training, enabling the model to learn data structures more efficiently. Figure 8: Illustration of Hu et al. (2021) s reparameterization approach, where only parameters A and B are trained. For text generation, traditional methods (Gong et al., 2022) often use fixed noise schedules, requiring recalculation of the input sequence at each timestep, which leads to inefficient training. In contrast, Yuan et al. (2024) introduced a dynamic noise adjustment technique that records loss values at each timestep during training and uses linear interpolation to map these losses to noise schedule parameters. This adaptive approach optimizes noise levels at each timestep and token position, improving training efficiency and enabling the model to better capture linguistic features. 3.2 Efficient Fine-Tuning Fine-tuning pre-trained diffusion models demands efficient resource use, leading to distinct strategies: Lo RA, Adapters, and Control Net. Lo RA stands out for its low-rank parameter updates, slashing memory needs by up to 90% and enabling rapid inference, though it struggles with nuanced spatial control. Adapters, by contrast, inject lightweight modules for task-specific tweaks, excelling in conditional guidance (e.g., sketches via T2I-Adapter) but relying on input quality, which can limit robustness. Control Net, however, leverages additional network branches for precise spatial conditioning, offering unmatched control over structure and style, yet at the cost of higher computational load mitigated in variants like Control Net-XS. These trade-offs are explored in detail in the subsections. 3.2.1 Lo RA Low-Rank Adaptation (Lo RA) (Hu et al., 2021) is a model adaptation method that maintains frozen pretrained model weights while enabling efficient task adaptation through the injection of low-rank decomposition matrices into each Transformer layer. The core mathematical foundation of this approach lies in Published in Transactions on Machine Learning Research (05/2025) Efficient Fine-tuning Lo RA (Hu et al., 2021), Concept Sliders (Gandikota et al., 2023), LCM-Lo RA (Luo et al., 2023b), Lo RA-Composer Yang et al. (2024), Lo RA conditioning (Choi et al., 2024b) T2I-Adapter (Mou et al., 2024), IP-Adapter (Ye et al., 2023), Para TAA (Tang et al., 2024), CTRL-Adapter (Lin et al., 2024a), Sim DA (Xing et al., 2024) Control Net Control Net (Zhang et al., 2023d), Control Net++ (Li et al., 2025), Control Net-XS (Zavadski et al., 2023), Control Ne Xt (Peng et al., 2024a), Uni-Control Net (Zhao et al., 2024), Uni Control (Qin et al., 2023) Figure 9: Summary of efficient fine-tuning techniques for diffusion models. its representation of the weight update mechanism: for a pre-trained weight matrix W0 Rd k, Lo RA represents the weight update as: W = W0 + W, where W = BA (28) where B Rd r and A Rr k are trainable low-rank matrices, and the rank r min(d, k). During forward propagation, for an input x Rk, the model computes the hidden representation h Rd as: h = W0x + Wx = W0x + BAx (29) The complete process is illustrated in Figure 8. A key advantage of this design lies in its deployment efficiency, where the explicit computation and storage of W = W0 + BA enables standard inference procedures without introducing additional latency. Originally proposed for fine-tuning Large Language Models (LLMs), Lo RA has demonstrated remarkable parameter efficiency and memory reduction in model adaptation. While predominantly utilized in LLM fine-tuning, recent research has extended its application to diffusion models, indicating its potential as a versatile adaptation technique across different deep learning architectures. LCM-Lo RA (Luo et al., 2023b) proposes a universal acceleration approach for diffusion models. As shown in Figure 10, this method achieves fast sampling by adding an Acceleration vector τLCM to the Base LDM Rombach et al. (2022). This module adopts Lo RA (Hu et al., 2021) to attach low-rank matrices to the original model without architectural modifications. For customized diffusion models that are fine-tuned for specific text-to-image generation tasks, the task-specific Lo RA (τ ) and acceleration Lo RA (τLCM) can be linearly combined through Eq.(30) to achieve fast inference while maintaining generation quality. More importantly, it provides a plug-and-play solution that reduces sampling steps from dozens to around 4, while maintaining compatibility with any pre-trained text-to-image diffusion model. τ LCM = λ1τ + λ2τLCM (30) Beyond the acceleration achieved by LCM-Lo RA, Concept Sliders (Gandikota et al., 2023) extends Lo RA s application to precise control over image generation attributes. This method identifies low-rank directions in the diffusion parameter space corresponding to specific concepts through Lo RA adaptation. The method freezes the original model parameters and trains a Lo RA adapter to learn concept editing directions. Given an input (xt, ct, t), where xt is the noisy image at timestep t. For a target concept ct, the model is guided through a score function to enhance certain attributes c+ while suppressing others c . This training objective can be formulated as: ϵθ (x, ct, t) ϵθ(x, ct, t) + η(ϵθ(x, c+, t) ϵθ(x, c , t)) (31) Here, ϵθ represents the denoising model s prediction, and η is a guidance scale. With this formulation, the method enables smooth control over concept strength through the guidance scale η while maintaining concept independence in the learned directions. By leveraging Lo RA s parameter-efficient nature, it achieves precise attribute manipulation with minimal computational overhead. Besides, Lo RA-Composer Yang et al. (2024) advances Lo RA s application in diffusion models toward seamless multi-concept integration. While previous works focus on acceleration or single-concept control, this Published in Transactions on Machine Learning Research (05/2025) Customized LDM 𝜽 Style vector 𝝉$ = 𝜽$ 𝜽𝐩𝐫𝐞 Accelerating vector 𝝉𝐋𝐂𝐌= 𝜽𝐋𝐂𝐌 𝜽𝐩𝐫𝐞 Combination 𝝉𝐋𝐂𝐌 $ = 𝝀𝟏𝝉$ + 𝝀𝟐𝝉𝐋𝐂𝐌 Customized LCM 𝜽-./ (LCM-Lo RA) (Style-Lo RA) Style: Fast: Figure 10: Illustration of LCM-Lo RA (Luo et al., 2023b). approach tackles the challenging task of combining multiple Lo RA-customized concepts within a single image generation process. It combines multiple Lo RAs in diffusion models by modifying the U-Net s attention blocks. Specifically, it enhances both cross-attention and self-attention layers within U-Net to enable direct fusion of multiple Lo RAs. Compared to traditional methods like Mix-of-Show Gu et al. (2024) that require training a fusion matrix to merge multiple Lo RAs, which increases computational overhead and may degrade generation quality. It directly combines multiple lightweight Lo RAs through modified attention blocks, avoiding the overhead of retraining models. While Lo RA-Composer focuses on fusing multiple Lo RAs for multi-concept control, Choi et al. (2024b) explores the fundamental application of Lo RA in attention layers. Both these works enhance diffusion models by modifying the attention mechanism in U-Net. The latter proposes a structured conditioning approach in U-Net blocks with three key components: (1) conventional convolutional blocks using scale-and-shift conditioning for feature normalization adjustment, (2) attention blocks enhanced by Lo RA adapters that condition both QKV computation and projection layers through learnable low-rank matrices, and (3) two Lo RA conditioning implementations - Time Lo RA/Class Lo RA for discrete-time settings and UC-Lo RA for continuous SNR settings, which utilize MLP-generated weights to combine multiple Lo RA bases. Them method achieves improved performance over traditional conditioning while only increasing the parameter count by approximately 10% through efficient low-rank adaptations in the attention layers. 3.2.2 Adapter Adapters are lightweight modules designed to enable efficient task adaptation by introducing small network layers into pre-trained models, allowing task-specific feature learning while keeping the original weights frozen. As illustrated in Figure 11, adapter layers are placed within the transformer block, positioned between normalization and feed-forward layers. Each adapter module consists of a down-projection, nonlinearity, and up-projection, which generates task-specific transformations without altering the core model s structure. This design significantly reduces memory and computational requirements, making adapters well-suited for tasks requiring lightweight parameter updates, such as text-to-image generation (T2I) and simulated domain adaptation (Sim DA). T2I-Adapter (Mou et al., 2024) is an adapter designed to enhance control in text-to-image generation models by introducing conditional features such as sketches, depth maps, and semantic segmentation maps, allowing for structural guidance in generated images. Unlike approaches that require modifying the model s core architecture, T2I-Adapter uses lightweight modules to incorporate external condition information into the generation process without altering the pre-trained model itself. This method improves the accuracy and consistency of generated images without increasing computational costs. In implementation, T2I-Adapter employs convolutional and residual blocks to align conditional inputs with the spatial dimensions of intermediate features in the UNet model, thus capturing structural information at multiple scales. This integration allows T2I-Adapter to flexibly incorporate conditional features, such as sketches and depth maps, providing Published in Transactions on Machine Learning Research (05/2025) enhanced control over text-to-image generation. Such multi-adapter strategies are particularly suitable for tasks requiring high customization in image generation, enabling simultaneous input of various structural features to refine the output. Figure 11: Architecture of the Adapter module: demonstrating the integration of adapter layers within a transformer block to achieve efficient task adaptation by adding lightweight transformations, while keeping the core model weights frozen. IP-Adapter (Ye et al., 2023) enhances the consistency and visual quality of text-to-image generation by incorporating image prompts. Unlike T2IAdapter (Mou et al., 2024), which provides structural guidance through sketches or depth maps, IPAdapter focuses on capturing visual details from an input image, making it ideal for tasks requiring high visual consistency with a reference image. This adapter processes the input image prompt into latent features, allowing the generation model to capture visual information from the target image and maintain detail alignment throughout the generation process. In its workflow, the image prompt is first mapped into the latent space and then processed through convolution and normalization modules within the adapter, enabling the model to utilize these features during inference. This setup enables the generation model to draw rich visual information from the image prompt, making IP-Adapter particularly suitable for tasks requiring high detail consistency, such as generating images with a style similar to the input image. CTRL-Adapter (Lin et al., 2024a) is designed to enhance attribute control during generation by guiding specific attributes such as emotion or object type, enabling precise customization in generated results. Unlike T2I-Adapter (Mou et al., 2024) and IP-Adapter (Ye et al., 2023), which focus on structural and detail consistency respectively, CTRL-Adapter is tailored to provide diversity control for the generation model. For example, as illustrated in Figure 12, the IP-Adapter architecture employs a decoupled A girl with sunglasses Text Encoder Image Encoder Image Features Text Features Decoupled Cross-Attention Cross Attention Cross Attention Denoising U-Net Frozen modules Figure 12: Architecture of IP-Adapter (Ye et al., 2023) using a decoupled cross-attention strategy, where only newly added modules are trained, and the pre-trained text-to-image model remains frozen. cross-attention strategy, where only newly added modules are trained while the pre-trained text-to-image model remains frozen. In contrast, CTRL-Adapter can adjust the style of generated images based on specified emotions or object types, achieving controllable content generation without altering the core architecture of the model. This makes CTRL-Adapter particularly suitable for tasks requiring high customization in generation, such as emotion-driven text generation or stylized image synthesis. Published in Transactions on Machine Learning Research (05/2025) Sim DA (Xing et al., 2024) is an adapter suited for cross-domain generation tasks, achieving domain adaptation by utilizing simulated data within the adapter to enhance the model s performance on previously unseen data distributions. Unlike CTRL-Adapter (Lin et al., 2024a), which primarily focuses on attribute control, Sim DA is designed to improve the model s generalization ability, allowing it to generate high-quality content even in unfamiliar data environments. Sim DA is particularly useful in generation tasks that require domain transfer, such as adapting a model trained on one image dataset to perform well on another dataset. This enables the model to align with new data characteristics without compromising generation quality. Figure 13: Illustration of Control Net. 3.2.3 Control Net Control Net (Zhang et al., 2023d) and its derivatives represent a significant advancement in adding spatial conditioning controls to pre-trained text-to-image diffusion models. The original Control Net architecture (Zhang et al., 2023d), as illustrated in Figure 13, presents a novel approach to integrating various spatial conditions such as scribbles, edge maps, open-pose skeletons, or depth maps into the generative process while preserving the robust features of pre-trained diffusion models. The architecture employs zero convolution layers that gradually develop parameters without disrupting the pre-trained model s stability. This design enables versatile conditioning, allowing the model to effectively leverage different types of spatial information. Through these conditioning methods, Control Net demonstrates a remarkable ability to guide generation with fine-grained control over structure, style, and composition. Building upon this foundation, several works have proposed improvements and alternatives. Control Net++ (Li et al., 2025) addresses the challenge of alignment between generated images and conditional controls by introducing pixel-level cycle consistency optimization. Through a pre-trained discriminative reward model and an efficient reward strategy involving single-step denoised images, it achieves significant improvements in control accuracy, with notable gains in metrics such as m Io U (11.1%), SSIM (13.4%), and RMSE (7.6%) across various conditioning types. Control Net-XS (Zavadski et al., 2023) reimagines the control system by enhancing the communication bandwidth between the controlling network and the generation process. This redesign not only improves image quality and control fidelity but also significantly reduces the parameter count, resulting in approximately twice the speed during both inference and training while maintaining competitive performance in pixel-level guidance tasks. The field has also seen efforts to unify multiple control capabilities. Uni Control (Qin et al., 2023) introduces a task-aware Hyper Net approach that enables a single model to handle diverse visual conditions simultaneously. Similarly, Uni-Control Net (Zhao et al., 2024) proposes a unified framework supporting both local controls and global controls through just two additional adapters, significantly reducing training costs and model size while maintaining high performance. Most recently, Control Ne Xt (Peng et al., 2024a) has pushed the boundaries of efficiency even further by introducing a streamlined architecture that minimizes computational overhead. It replaces the traditional heavy additional branches with a more concise structure and introduces Cross Normalization (CN) as an alternative to zero convolutions. This approach achieves fast and stable training convergence while reducing learnable parameters by up to 90% compared to previous methods. Published in Transactions on Machine Learning Research (05/2025) Efficient Sampling Efficient Solver Diff Flow (Zhang & Chen, 2021), Di No F (Zand et al., 2023), Jolicoeur-Martineau et al. (2021), GMS (Guo et al., 2024), BFM-Solver2 (Xue et al., 2024a), SA-Solver (Xue et al., 2024b), Restart (Xu et al., 2023b) ODE Solver DDIM (Song et al., 2020a), DPM-solver (Lu et al., 2022), DEIS (Zhang & Chen, 2022), DMCMC (Kim & Ye, 2022), i-DODE (Zheng et al., 2023) Sampling Scheduling Parallel Sampling DEQ (Pokle et al., 2022), Para Di GMS (Shih et al., 2024), Para TAA (Tang et al., 2024), TAA (Walker & Ni, 2011), Stream Diffusion (Kodaira et al., 2023) Timestep Scheduling Fast DPM (Kong & Ping, 2021), Watson et al. (2021), AYS (Sabour et al., 2024) Truncated Sampling Early Exit CATs (Schuster et al., 2021), ASE (Moon et al., 2024b) Retrieval-Guided Initialization RDM (Blattmann et al., 2022), k NN-Diffusion (Sheynin et al., 2022), Re-Imagen (Chen et al., 2022b), Re Di (Zhang et al., 2023c) Knowledge Distillation Vector Field Distillation Denoising Student (Luhman & Luhman, 2021), Progressive Distillation (Salimans & Ho, 2022), Meng et al. (2023), SANA-Sprint (Chen et al., 2025), DMD Yin et al. (2024) Generator Distillation SDS (Poole et al., 2022), VSD (Wang et al., 2024b), Diff-Instruct (Luo et al., 2024), CSD (Decatur et al., 2024) Figure 14: Summary of efficient sampling techniques for diffusion models. 3.3 Efficient Sampling The standard diffusion sampling process is computationally intensive because it requires sequentially executing a large number of denoising steps, with each step dependent on the output of the previous one as formulated in Eq(4). This inherent sequential dependency makes the sampling procedure time-consuming, resulting in significantly slower generation. To address these computational challenges, researchers have developed efficient sampling methods through four principal approaches. As illustrated in Figure 14, these encompass efficient SDE and ODE solvers that reduce the required number of function evaluations, advanced sampling scheduling strategies including parallel sampling techniques and timestep optimization methods, knowledge distillation techniques that transfer diffusion model capabilities to more efficient representations, and truncated sampling approaches that leverage early exit mechanisms and retrieval-based techniques to further accelerate the generation process while preserving output quality. 3.3.1 Efficient Solver As we mentioned in Section. 2.1.3, although diffusion solvers avoid the need to compute the intractable partition function, they still require numerous function evaluations with fixed step sizes for SDEs in Eq(11) or generic numerical integration methods for ODEs in Eq.(12), resulting in slow sampling procedures and substantial computational costs. These limitations create substantial computational costs and lengthy generation times. In the following sections, we present advances in both SDE and ODE methods that address these efficiency constraints. SDE Solver. Nowadays, there are many ways to efficiently implement SDE-based solvers. (Zhang & Chen, 2021)introduces a novel generative modeling and density estimation algorithm called Diffusion Normalizing Flow (Diff Flow). Similar to the SDE of diffusion models Eq.(10), the Diff Flow model also has a forward process: dx = f(x, t, θ)dt + g(t)d w (32) and a backward process: dx = f(x, t, θ) g2(t)s(x, t, θ) dt + g(t)d w (33) As a result of the learnable parameter θ, the drift term f is also learnable in Diff Flow, compared to the fixed liner function as in most diffusion models. Besides, these SDEs are jointly trained by minimizing the KL divergence. This allows the model to better adapt to changes in the data distribution, thus speeding up the convergence of the backward diffusion process. Similar to Diff Flow, Zand et al. (2023) proposes a method called Diffusion with Normalizing Flow priors that also combines diffusion models with normalizing flows. The method first uses a linear SDE in the forward process to convert the data distribution into a noise distribution gradually. In the reverse process, a normalizing flow network is introduced to map the standard Gaussian distribution to latent variables close to the data distribution through a series of reversible Published in Transactions on Machine Learning Research (05/2025) transformations, which allows the samples to return to the data distribution more quickly, rather than relying on a large number of small incremental adjustments. However, the fixed step sizes in existing SDE solvers Eq.(10), which usually require tremendous iterative steps, significantly affect generation efficiency. To address this, Jolicoeur-Martineau et al. (2021) proposes a novel adaptive step-size SDE solver that dynamically adjusts the step size based on error tolerance, thereby reducing the number of evaluations. Specifically, the proposed method dynamically adjusts the step size by estimating the error between first-order and second-order approximations, leveraging a tolerance mechanism that incorporates both absolute and relative error thresholds. Furthermore, the use of extrapolation enhances precision without incurring additional computational overhead. This approach obviates the need for manual step-size tuning and is applicable across a range of diffusion processes, including Variance Exploding and Variance Preserving models. As a result of Gaussian assumption for reverse transition kernels becomes invalid when using limited sampling steps. The Gaussian Mixture Solver (GMS) (Guo et al., 2024) optimized SDE solver by using Gaussian mixture distribution. It addresses the limitations of the traditional process of SDE solvers in Eq.(11), which assume a Gaussian distribution for the reverse transition kernel. Specifically, GMS replaces the Gaussian assumption with a more flexible Gaussian mixture mode and utilizes a noise prediction network with multiple heads to estimate the higher-order moments of the reverse transition kernel. At each sampling step, it employs the Generalized Method of Moments to optimize the parameters of the Gaussian mixture transition kernel, allowing for a more accurate approximation of the true reverse process, even with a limited number of discretization steps. Instead, Xue et al. (2024a) unifies Bayesian Flow Networks (BFNs) with Diffusion Models (DMs) by introducing time-dependent SDEs into the BFN framework. BFNs work by iteratively refining the parameters of distributions at different noise levels through Bayesian inference, rather than directly refining the samples as in traditional diffusion models. To achieve theoretical unification between BFNs and DMs, the authors introduce a time-dependent linear SDE that governs the noise addition process in BFNs. This forward process includes two time-dependent functions: one controlling the drift of parameters and another controlling their diffusion. Based on this forward equation, they derive a corresponding reverse-time SDE for generating data from noise. This reverse process combines the drift term with a score-based correction term. This reverse-time SDE directly aligns with the denoising process in diffusion models, enabling the BFN sampling process to effectively replicate the behavior of diffusion models. By optimizing the solving process of SDE in Eq.(10), Stochastic Adams Solver (SA-Solver) (Xue et al., 2024b) was presented. It is an innovative method designed to efficiently sample from Diffusion SDEs in Diffusion Probabilistic Models (DPMs) (Ho et al., 2020). By addressing the significant computational burden of traditional samplers, SA-Solver achieves this through a clever combination of variance-controlled diffusion SDEs and a stochastic Adams method (Buckwar & Winkler, 2006), which is a multi-step numerical technique that leverages prior evaluations to enhance efficiency. The method introduces a noise control function τ(t), enabling dynamic adjustment of the noise injected during sampling, which in turn strikes an optimal balance between sampling speed and the quality of the generated data. Operating within a predictor-corrector framework, SA-Solver first makes an initial estimate through the predictor step and then refines this estimate using the corrector step, ensuring greater accuracy in the final output. This strategic integration significantly reduces the number of function evaluations required. Xu et al. (2023b) introduces Restart, a novel sampling algorithm for diffusion models that strategically combines the strengths of SDE and ODE solvers. By theoretically analyzing Wasserstein bounds, the authors demonstrate that SDEs contract accumulated errors via stochasticity, while ODEs excel in low-NFE regimes due to smaller discretization errors. Restart bridges these regimes by alternating between noise injection and deterministic ODE backward steps, decoupling stochasticity from drift updates to amplify error contraction while retaining ODE efficiency. ODE Solver. For efficiently implement the ODE-based solvers, Denoising Diffusion Implicit Models (DDIM) (Song et al., 2020a) builds upon the framework of Denoising Diffusion Probabilistic Models (DDPM) (Ho et al., 2020), offering significant enhancements in sampling efficiency, which is one of the first models to leverage ODEs explicitly for the accelerating sampling process. qσ(x1:T |x0) = qσ(x T |x0) t=2 qσ(xt 1|xt, x0) (34) Published in Transactions on Machine Learning Research (05/2025) Unlike DDPM s Markovian forward process Eq.(1) where each state only depends on its immediate predecessor, DDIM utilizes the Non-Markovian Forward Process Eq.(34). These formulas allow each state not only to depend on its immediate predecessor but also on the initial state or a series of previous states. Specifically, Eq.(35) outlines how DDIM generates xt 1 from xt by predicting the denoised observation, which essentially approximates reversing the diffusion process: xt 1 = αt 1 xt 1 αtϵ(t) θ (xt) αt 1 αt 1 σ2 t ϵ(t) θ (xt) + σtϵt (35) During the process, DDIM employs an ODE solver to manage the continuous transformation across the latent space: dx(t) = ϵ(t) θ Eq.(36) is key to the efficient management of the generation process, allowing for fewer steps in the generative sequence by smoothly interpolating between states using an ODE solver, thus significantly reducing the time complexity compared to traditional methods. While DDIM s ODE formulation Eq.(12) and its implementation through Eq.(36) provide a foundation for deterministic sampling, Liu et al. (2022a) identifies two critical issues in the ODE formulation of DDIM: first, the neural network θ and ODE are only well-defined within a narrow data manifold, while numerical methods generate samples outside this region. second, the ODE becomes unbounded as t 0 for linear schedules. Therefore PNDM is proposed to decompose the numerical solver into gradient and transfer components. It achieves second-order convergence, enabling 20x speedup while maintaining quality and reducing FID by 0.4 points at the same step count across different datasets and variance schedules. The DPM-solver (Lu et al., 2022) and Diffusion Exponential Integrator Sampler (DEIS) (Zhang & Chen, 2022) innovate by leveraging the semi-linear structure of the probability flow ODE Eq.(12) to design custom ODE solvers that outperform traditional Runge-Kutta (Hochbruck & Ostermann, 2010) methods in terms of efficiency. Specifically, DPM-solver solves the linear part of the equation and uses neural networks to approximate the nonlinear component. Compared to PNDM, DPM-solver maintains lower FID scores at the same NFE. Further, DEIS employs an Exponential Integrator (Hochbruck & Ostermann, 2010) to discretize ODEs. This method simplifies the probability flow ODE by transforming the probability ODE into a simple non-stiff ODE. Both of the innovations reduce the number of iterations needed producing high-quality samples within just 10 to 20 iterations. To reduce the computational overhead, Zheng et al. (2023) presents an improved technique for maximum likelihood estimation of ODEs. Instead of directly working with the drift and score terms in Eq.(12), it introduces velocity parameterization to predict and optimize velocity changes dxt during the diffusion process directly. The method improves upon previous ODE-based approaches by incorporating secondorder flow matching for more precise trajectory estimation. Additionally, it introduces a negative logsignal-to-noise-ratio (log-SNR) for timing control of the diffusion process, alongside normalized velocity and importance sampling to reduce variance and optimize training. These enhancements significantly improve the model s likelihood estimation performance on image datasets without variational dequantization or data augmentation. While previous methods focus on improving reverse ODE integrators based on Eq.(12), Denoising MCMC (DMCMC) (Kim & Ye, 2022) takes a different approach by integrating Markov Chain Monte Carlo (MCMC) with ODE integrators to optimize the data sampling process. In DMCMC, MCMC first generates initialization points in the product space of data and diffusion time, which are closer to a noise-free state, significantly reducing the noise levels that need to be processed by the ODE integrators. This hybrid approach complements rather than improves the ODE integrators directly, enhancing overall sampling efficiency. Besides, Lu & Song (2024) focuses on improving continuous-time consistency models(CMs) (Song et al., 2023; Song & Dhariwal, 2023) for efficient diffusion sampling by modifying the ODE parameterization and training objectives of continuous-time CMs. The core contribution is Trig Flow, a unified framework that Published in Transactions on Machine Learning Research (05/2025) bridges EDM (Karras et al., 2022) and Flow Matching (Peluchetti, 2023; Lipman et al., 2022; Liu et al., 2022b; Albergo et al., 2023; Heitz et al., 2023). While the traditional probability flow framework is governed by Eq.(12), they propose a simplified parameterization. To model these dynamics, they introduce a neural network Fθ with parameters θ that takes normalized samples and time encodings as input. The time variable t is transformed by cnoise(t) to better condition the network. This results in a concise probability flow ODE: dt = σd Fθ( xt σd , cnoise(t)) (37) By introducing this simplified ODE parameterization, Trig Flow enables training large-scale CMs (up to 1.5B parameters) that achieve state-of-the-art performance with just two sampling steps, significantly reducing computational costs compared to DPM-solver (Lu et al., 2022) and other traditional diffusion models. Moreover, (Du et al., 2024) replaces SDE with ODE s deterministic trajectory, combined with fixed noise and Consistency Distillation Sampling loss, which minimizes stochastic fluctuations and computational redundancy. These designs enable faster convergence while maintaining high fidelity. In conclusion, recent research has produced numerous works on faster diffusion samplers based on solving the ODE Eq.(12). Research shows that ODE samplers are highly effective when only a limited number of NFEs is available, while SDE samplers offer better resilience to prior mismatches (Nie et al., 2023) and exhibit superior performance with a greater Number of Function Evaluations (NFEs) (Lu et al., 2022). 3.3.2 Sampling Scheduling In diffusion models, a sampling schedule outlines a structured approach for timing and managing the sampling steps to improve both the efficiency and quality of the model s output. It focuses on optimizing the sequence and timing of these steps, utilizing advanced techniques to process multiple steps simultaneously or in an improved sequential order. Specifically, this scheduling primarily targets the optimization of the reverse process in DDPM, as described in Eq.(4), where each step requires model prediction to gradually denoise from pure noise to the target sample. This scheduling is crucial for reducing computational demands and enhancing the model s performance in generating high-quality samples. Parallel Sampling. Parallel sampling is a process that schedules sampling tasks in parallel. Traditional diffusion models require a extensive series of sequential denoising steps to generate a single sample, which can be quite slow. For instance, Denoising Diffusion Probabilistic Models (DDPMs) (Ho et al., 2020) might need thousands of these steps to produce one sample. However, parallel sampling leverages the power of a multicore GPU to compute multiple sampling steps. This approach optimizes the use of computational resources and reduces the time needed for model generation. Currently, there is significant work on autoregressive models that employ parallelization to speed up the sampling process. Figure 15: Computation graph of Picard iterations, which introduces skip dependencies (Shih et al., 2024). However, these techniques cannot be directly applied to diffusion models. This is because the computational frameworks and inference efficiency in autoregressive models differ from those in diffusion models. Therefore, designing algorithms tailored to parallelize the sampling process of diffusion models is crucial. An innovative extension of the Denoising Diffusion Implicit Model (DDIM) (Song et al., 2020a) using Deep Equilibrium (DEQ) models is presented (Pokle et al., 2022), where the sampling sequence is conceptualized as a multivariate fixed-point system. This approach focuses on finding the system s fixed point during the forward pass and utilizes implicit differentiation during the backward pass to enhance computational efficiency. By treating the sampling steps as an equilibrium system and solving for their fixed points simultaneously, parallel processing on multiple GPUs is achieved by batching the workload. Notably, it improves efficiency by updating each state xt based on predictions from the noise prediction network Published in Transactions on Machine Learning Research (05/2025) ϵθ, which takes into account all subsequent states xt+1:T , unlike traditional diffusion processes that update states sequentially based only on the immediate next state xt+1. Para Di GMS (Shih et al., 2024) employs Picard iterations to parallelize the sampling process in diffusion models. This method models the denoising process using ordinary differential equations (ODEs) (Song et al., 2020c), where Picard iterations approximate the solution to these ODEs concurrently for multiple state updates. Para Di GMS operates within a sliding window framework, enabling the simultaneous update of multiple state transitions. Each state is iteratively connected to different generations, allowing for information integration from several previous iterations Figure 15. Figure 16: Illustration of timestep schedule optimization process. Building upon these parallel processing concepts, Para TAA (Tang et al., 2024) also adopts an iterative approach, primarily applied in practical deployments for image generation tasks such as text-to-image transformations using Stable Diffusion. Specifically, Para TAA enhances parallel sampling by solving triangular nonlinear equations through fixed-point iteration. Furthermore, the study introduces a novel variant of the Anderson Acceleration (Walker & Ni, 2011) technique, named Triangular Anderson Acceleration, designed to accelerate computation speed and improve the stability of iterative processes. Kodaira et al. (2023) introduces Stream Diffusion for real-time interactive diffusion by combining batched denoising to exploit GPU parallelism and Residual Classifier-Free Guidance(RCFG) to minimize redundant guidance computations, and input-output queues for asynchronous processing. It further optimizes energy efficiency via stochastic similarity filtering, which dynamically skips processing for near-identical frames using cosine similarity thresholds. Timestep Schedule. In the sampling process of diffusion models, the entire process is discrete, and the model progressively restores data from noise through a series of discrete timesteps. Each timestep represents a small denoising step that moves the model from its current state closer to the real data. The timestep schedule refers to the strategy for selecting and arranging these timesteps. It may involve distributing them evenly or performing denser sampling during key stages to ensure the efficiency of the sampling process and the quality of the generated results. Selecting an appropriate method to choose a series of timesteps can enable the sampling process to converge quickly, which the process is shown in Figure 16. Fast DPM (Kong & Ping, 2021) is a unified framework for fast sampling in diffusion models that innovatively generalizes discrete diffusion steps to continuous ones and designs a bijective mapping between these continuous diffusion steps and noise levels. By utilizing this mapping, Fast DPM constructs an approximate diffusion and reverse process, significantly reducing the number of steps required (S T). It allows for the flexible determination of sampling points by selecting specific steps or variances from the original diffusion process, thereby enhancing efficiency. Watson et al. (2021) proposes a dynamic programming algorithm to optimize timestep scheduling in Denoising Diffusion Probabilistic Models (DDPMs). The algorithm efficiently determines the optimal timestep schedule from thousands of possible steps by leveraging the decomposable Published in Transactions on Machine Learning Research (05/2025) property of Evidence Lower Bound (ELBO) across consecutive timesteps and treating timestep selection as an optimization problem. Experiments show that the optimized schedule requires only 32 timesteps to achieve comparable performance to the original model with thousands of steps, effectively balancing efficiency and quality. However, optimizing an exact Evidence Lower Bound (ELBO) is typically not conducive Figure 17: Minimizing an upper bound on the Kullback-Leibler divergence (KLUB) between the true and linearized generative SDEs to find optimal DM sampling schedules (Sabour et al., 2024). to enhancing image quality. To address this, Align Your Steps (AYS) (Sabour et al., 2024) proposes a compute-efficient paradigm for diffusion models by co-optimizing sampling schedules and solvers. Through stochastic calculus-guided Kullback-Leibler Divergence Upper Bound (KLUB) minimization, as shown in 17, AYS derives adaptive schedules that reduce sampling steps by 40% while maintaining quality achieving equivalent FID scores with 20 steps versus 30 steps in baseline methods. The optimized schedules are versatile across modalities and solver types, enabling plug-and-play acceleration without model retraining. 3.3.3 Truncated Sampling Truncated sampling enhances the efficiency of sample generation in diffusion models by strategically reducing redundant computations, thereby lowering computational costs. This optimization category encompasses several approaches, with Early Exit and Retrieval-Guided Initialization representing two primary strategies that target different phases of the diffusion process to improve computational efficiency. Specifically, early Exit focuses on terminating unnecessary computations in later stages of the diffusion process when predictions are confident. Conversely, Retrieval-Guided Initialization improves efficiency in the early stages by leveraging retrieved examples to provide a better initialization, effectively bypassing parts of the iterative refinement process. These approaches allocate computation more effectively by focusing resources on the most critical steps of the sampling process. Early Exit. Recent papers that focus on early exit mechanisms have gained significant attention in the Large Language Models (LLMs) domain (Schuster et al., 2021). By implementing early exit strategies, these methods bypass unnecessary computations in deeper layers when outputs can be generated at earlier stages, thereby substantially reducing inference time and computational resources. Similarly, for computation-intensive diffusion models that typically require numerous denoising steps, early exit techniques present a promising approach to accelerate the generation process dramatically. Moon et al. (2024a) proposes a simple yet effective early exiting framework called Adaptive Score Estimation (ASE) specifically designed to accelerate the sampling process of diffusion models. The method is based on the key observation that score estimation difficulty varies across different timesteps in the diffusion process, with timesteps closer to the noise distribution requiring fewer computational resources than those closer to the data distribution. Published in Transactions on Machine Learning Research (05/2025) The schematic in Figure 18 illustrates the time-dependent exit schedule used in ASE. As shown in the figure, the model progressively skips more building blocks as the diffusion process moves closer to the noise distribution, where the score estimation becomes easier. For example, in the Di T model (Peebles & Xie, 2023), the blocks are dropped progressively as the timestep approaches 1, closer to the noise regime. Conversely, when the timestep is closer to the data regime, more blocks are retained to ensure accurate score estimation. Moreover, the U-Vi T model (Bao et al., 2023) follows a similar dropping schedule, but due to the long skip connections between the encoder and decoder, only decoder blocks are skipped. This preserves critical information from the encoder while still speeding up the process. score function Block 2 Decoder Decoder Block 3 Decoder Block 4 Figure 18: Illustration of the time-dependent exit schedule, where more blocks are skipped as the sampling process moves toward the noise distribution, optimizing computational efficiency Moon et al. (2024a). Through carefully designed time-varying exit schedules, the method significantly accelerates sampling speed while preserving generation quality. It can also be seamlessly integrated with various diffusion model solvers to enhance overall efficiency further. In contrast, Tang et al. (2023a) introduces Dee Diff, which approaches early exiting through a dynamic uncertaintyaware framework. While ASE relies on static, predefined exit schedules that are fixed during inference, Dee Diff incorporates a timestep-aware uncertainty estimation module (UEM) that adaptively assesses the prediction uncertainty of each intermediate layer at every sampling step. The UEM directly estimates uncertainty values from the features at each layer using lightweight, fully-connected networks, which are trained to indicate how well the current layer s output matches the final layer s prediction. During inference, when a layer s uncertainty falls below a predefined threshold, the model exits early at that layer, bypassing deeper layers for that particular timestep. Despite their different implementation strategies, both methods demonstrate that early exiting frameworks can effectively balance computational efficiency and generation quality, offering practical solutions for deploying large-scale diffusion models. Retrieval-Guided Denosing. Retrieval-Guided Initialization combines the efficiency of retrieval mechanisms with the generative power of diffusion models, and can be applied across various generative tasks. This method guides the generation process by retrieving samples or data relevant to the input. For example, as illustrated in Figure 19, when the input is a text prompt, the retriever selects relevant images from a database, which then serve as contextual guidance for the diffusion model to generate a coherent output. By providing a more informed state for the diffusion model, this approach helps the model converge more quickly and generate higher-quality outputs. Retrieval-Guided Denoising is widely used in text-to-image diffusion tasks, offering an efficient way to generate images that align with textual descriptions. The k NN-Diffusion (Sheynin et al., 2022) method integrates large-scale retrieval techniques with the generative power of diffusion models, offering an efficient approach for image generation. The key idea is to use the CLIP (Radford et al., 2021) image encoder to map images into a shared embedding space and employ the k-Nearest-Neighbors (k NN) retrieval to identify the k most similar image embeddings. Notably, this method does not require paired text-image datasets, which is a significant advantage for domains where such datasets are scarce. During training, the model is conditioned on image embeddings, and k NN retrieval is used to extend the distribution of conditioning embeddings, which helps the model generalize better and bridge the gap between text and image distributions. In the sampling phase, the model receives text input, which is converted to text embedding using CLIP. The k NN retrieval mechanism is then applied to find the k most similar image embeddings based on the text embedding. These retrieved image embeddings serve as additional conditional information, guiding the generation process and ensuring that the generated image aligns with the textual description. Consequently, k NN-Diffusion im- Published in Transactions on Machine Learning Research (05/2025) proves both the efficiency of image generation and the computational resource consumption, making it a highly efficient solution for text-to-image generation tasks. Unlike KNN-Diffusion, which primarily relies on image retrieval to guide the sampling process, Re-Imagen (Chen et al., 2022b) enhances the generation process by incorporating multimodal retrieval to balance computational cost and output quality, using both image and text pairs. Re-Imagen s retrieval method is based on an external multimodal knowledge base. During sampling, given an input text, the model queries an external database to retrieve the most relevant image-text pairs. These retrieved pairs are then used as additional conditional inputs for the generation process. To identify the most relevant pairs, the model employs either BM25 (Robertson et al., 2009) or CLIP (Radford et al., 2021) similarity measures, which evaluate the alignment between the input text and the image-text pairs in the database. Therefore, this approach provides more diverse visual information, especially for rare or unseen entities. Additionally, Re-Imagen employs a cascaded diffusion architecture (Ho et al., 2022a), which allows it to generate high-quality images more efficiently compared to k NN-Diffusion, as it reduces the computational cost by progressively refining images at different resolutions. Generator (Diffusion Model) River Database Query Labrador sitting on bench near river. Figure 19: Illustration of the retrieval-based diffusion model. The retriever selects relevant images from a database based on input text. These retrieved images provide contextual guidance for the generator (diffusion model) to produce a new, coherent output image. Besides approaches that rely on similarity measures like CLIP embeddings to retrieve images or texts, Re Di (Retrieval-based Diffusion) (Zhang et al., 2023c) offers a fundamentally different approach to accelerating diffusion model inference. Re Di directly retrieves precomputed trajectories from the diffusion process itself. The method constructs a knowledge base of sample trajectories where each entry contains a key-value pair. For instance, in the forward diffusion process (noise-adding process), an early state sample xk serves as the key, while a later state sample xv serves as the value. During inference, Re Di first generates the initial steps of a trajectory up to x k, uses this as a query to find a similar state xk in the knowledge base, and then skips intermediate computational steps by jumping to the retrieved xv before continuing the generation process. By skipping a large portion of intermediate steps, Re Di significantly reduces the number of function estimations (NFEs) required during sampling. Experiments with Stable Diffusion demonstrate that Re Di achieves a two times speedup with comparable quality and enables effective zero-shot domain adaptation for tasks like image stylization without requiring domain-specific knowledge bases. 3.3.4 Knowledge Distillation Knowledge distillation (Hinton et al., 2015) is a technique that compresses complex models into smaller, efficient versions with minimal performance loss. The process of knowledge distillation can be captured by minimizing the following loss function: LKD = αLCE(y, σ(Ts(x))) + βLMSE(Tt(x), Ts(x)), (38) Published in Transactions on Machine Learning Research (05/2025) Figure 20: Illustrations of the knowledge distillation. where Tt and Ts are the teacher and student models, respectively, σ is the softmax function, LCE is the cross-entropy loss, and LMSE is the mean squared error loss, with α and β as balancing hyperparameters. In DMs, known for generating high-quality data, this approach is increasingly applied to improve efficiency by addressing slow sampling speeds caused by the numerous neural function evaluations in the diffusion process. By distilling the knowledge from DMs into more efficient forms, researchers aim to accelerate sampling while preserving the generative performance of the original models. Follow Luo (2023), knowledge distillation for DMs can be categorized into vector field distillation and generator distillation. Vector Field Distillation. Vector field distillation improves the efficiency of deterministic sampling in diffusion models by transforming the generative ODE into a new generative vector field. This approach reduces the number of NFEs needed to produce samples of similar quality. Luhman & Luhman (2021) first proposes a strategy to distill a DDIM sampler into a Gaussian model that needs only one NFE for sampling. In this approach, a conditional Gaussian model serves as the student model, and the training process involves minimizing the conditional KL divergence between this student model and the DDIM sampler. While this method advances the application of knowledge distillation to diffusion models, it still has computational inefficiencies, as it necessitates generating the final outputs of DDIM or other ODE samplers, which entails hundreds of NFEs for each training batch. Salimans & Ho (2022) proposes a progressive distillation strategy to train a student model to use half the NFEs of the teacher model by learning its two-step prediction strategy, as illustrated in Figure 21. Once the student model accurately predicts the teacher s two-step sampling strategy, it replaces the teacher model, and a new student model is trained to further reduce the sampling steps. This method reduces the NFEs significantly, achieving 250 times greater efficiency with only a 5% drop in performance. Distillation Distillation Distillation Figure 21: The progressive distillation, where the original sampler derived from integrating a learned diffusion model s probability flow ODE, is efficiently condensed into a new sampler that achieves the same task in fewer steps. (Salimans & Ho, 2022). A two-stage distillation strategy is proposed by Meng et al. (2023) to address the challenge of transferring knowledge from classifier-free guided conditional diffusion models like DALL E-2 (Ramesh et al., 2022) and Stable Diffusion (Rombach et al., 2022). In the first stage, a student model is trained with classifier-free guidance to learn from the teacher diffusion model. The second stage employs the progressive diffusion strategy to further reduce the number of diffusion steps for the student model. This two-stage approach is applied to both pixel-space and latent-space models for various tasks, including text-guided generation and image inpainting. Published in Transactions on Machine Learning Research (05/2025) Song et al. (2023) firtstly introduces the Consistency Model (CM), which leverages the self-consistency property of generative ODEs in diffusion models. Instead of directly mimicking the output of the generative ODE, their method focuses on minimizing the difference in the self-consistency function. By randomly diffusing a real data sample and simulating a few steps of the generative ODE to generate another noisy sample on the same ODE path, the model inputs these two noisy samples into a student model. Consequently, Chen et al. (2025) accelerates consistency models via hybrid distillation, enabling 1-4 step generation. It transforms pre-trained flow models into Trig Flow (Lu & Song, 2024) without retraining, preserving trajectory alignment while boosting fidelity via adversarial training LADD. Wu et al. (2024) proposes Multi-scale Latent Updates parameters with SGD: Updates sample in pixel space: Score Distillation Sampling Ancestral Sampling Figure 22: Illustration as (Poole et al., 2022), it utilizes score distillation sampling. Point Consistency Model (MLPCM), which implements one-step generation through consistency distillation, combines multi-scale latent space and 3D attention mechanisms to reduce computational complexity. Moreover, Yin et al. (2024) proposes Distribution Matching Distillation (DMD), which distills multi-step diffusion models into a single-step generator by introducing a distribution-level matching objective that minimizes the KL divergence between real and synthetic data distributions, alongside a regression loss to align large-scale structural features. Generator Distillation. Unlike vector field distillation, which primarily focuses on distilling knowledge into student models with identical input and output dimensions, generator distillation aims to transfer the complex distributional knowledge embedded in a diffusion model into a more efficient generator. The Neural Radiance Field (Ne RF) (Mildenhall et al., 2021) is a powerful technique for reconstructing 3D scenes from 2D images by learning a continuous volumetric scene representation. Ne RFs generate photorealistic views of scenes from novel angles, making them valuable for applications in computer vision and graphics. However, the limited availability of data for constructing Ne RFs is an issue. Therefore, exploring distillation methods to obtain Ne RFs with contents related to given text prompts is a promising way. (Poole et al., 2022) first proposed Score Distillation Sampling (SDS) to distill a 2D text-to-image diffusion model into 3D Ne RFs, as illustrated in Figure 22. Unlike traditional Ne RF construction that requires images from multiple views of the target 3D objects, text-driven construction of Ne RF lacks both the 3D object and the multiple views. The SDS method optimizes the Ne RF by minimizing the diffusion model s loss function using Ne RF-rendered images from a fixed view. Wang et al. (2024b) introduce Variational Score Distillation (VSD), which extends SDS by treating the 3D scene corresponding to a textual prompt as a distribution rather than a single point. Compared to SDS, which generates a single 3D scene and often suffers from limited diversity and fidelity, VSD is capable of generating more varied and realistic 3D scenes, even with a single particle. Luo et al. (2024) propose Diff-Instruct, which can transfer knowledge from pre-trained diffusion models to a wide range of generative models, all without requiring additional data. The key innovation in Diff-Instruct is the introduction of Integral Kullback-Leibler divergence, which is specifically designed to handle the diffusion process and offers a more robust way to compare distributions. Decatur et al. (2024) present Cascaded Score Distillation (CSD), an advancement by addressing a key limitation of standard SDS. Specifically, while traditional SDS only leverages the initial low-resolution stage of a cascaded model, CSD distills scores across multiple resolutions in a cascaded manner, allowing for nuanced control over both fine details and the global structure of the supervision. By formulating a distillation loss that integrates all cascaded stages, which are trained independently, CSD enhances the overall capability of generating high-quality 3D representations. Published in Transactions on Machine Learning Research (05/2025) Compression Quantization Post-Training Quantization PTQ4DM (Shang et al., 2023), Q-Diffsuion (Li et al., 2023), BRECQ (Li et al., 2021), RAQ (Kim et al., 2024b), PTQD (He et al., 2024) Quantization-Aware Training TDQ (So et al., 2024), QALo RA (He et al., 2023) Pruning Diff-Pruning (Fang et al., 2023), LD-Pruner (Castells et al., 2024), Layer Merge (Kim et al., 2024a), LAPTOPDiff (Zhang et al., 2024a) Figure 23: Summary of compression techniques for DMs. 3.4 Compression Model compression enhances efficiency by reducing the sizes and the amount of arithmetic operations of DM. As summarized in Figure 23, model compression techniques for DMs can be grouped into quantization and pruning. These two categories are orthogonal to each other, and compress DMs from different perspectives. Figure 24: Illustrations of the quantization. 3.4.1 Quantization Quantization compresses neural networks by converting model weights and/or activations of high-precision data types XH such as 32-bit floating point into low-precision data types XL such as 8-bit integer (Dettmers et al., 2024). Quantization techniques can be classified into post-training quantization (PTQ) and quantization-aware training (QAT). Post-Training Quantization. PTQ involves selecting operations for quantization, collecting calibration samples, and determining quantization parameters for weights and activations. While collecting calibration samples is straightforward for CNNs and Vi Ts using real training data, it poses a challenge for Diffusion Models (DMs). In DMs, the inputs are generated samples xt at various time steps (t = 0, 1, ..., T), where T is large to ensure convergence to an isotropic Normal distribution. To address this issue, Shang et al. (2023) proposes PTQ4DM, the first DM-specific calibration set collection method, generating calibration data across all time steps with a specific distribution. However, their explorations remain confined to lower resolutions and 8-bit precision. Q-Diffsuion (Li et al., 2023) propose a time step-aware calibration data sampling to improve calibration quality and apply BRECQ (Li et al., 2021), which is a commonly utilized PTQ framework, to improve performance. Furthermore, compared to conventional PTQ calibration methods, they identify the accumulation of quantization error across time steps as another challenge in quantizing DMs Figure 25 Published in Transactions on Machine Learning Research (05/2025) Calibration Training Dataset Random Sampling Synthesized by NN Sample from Training Data (a) Calibration dataset creation Calibration pθ(xt 1|xt) x T xt xt 1 Time Step-aware Uniform Sampling Traditional PTQ Q-Diffusion (Ours) (b) Model Inference pθq(x T 1|x T) pθq(x T 2|x T 1) Quantization errors accumulate at each time step Naive Linear Quantization incurs severe quality loss Figure 25: Traditional PTQ scenarios and Q-Diffusion differ in (a) the creation of calibration datasets and (b) the workflow for model inference (Li et al., 2023). (a). Therefore, they also propose a specialized quantizer for the noise estimation network shown in Figure 25 (b). Based on Q-Diffusion, Kim et al. (2024b) find that inaccurate computation during the early stage of the reverse diffusion process has minimal impact on the quality of generated images. Therefore, they introduce a method that focuses on further reducing the number of activation bits for the early reverse diffusion process while maintaining high-bit activations for the later stages. Lastly, He et al. (2024) presents PTQD, a unified formulation for quantization noise and diffusion perturbed noise. Additionally, they introduce a step-aware mixed precision scheme, which dynamically selects the appropriate bitwidths for synonymous steps. Quantization-Aware Training. Different from PTQ, QAT quantizes diffusion models during the training process, allowing models to learn quantization-friendly representations. Since QAT requires additional training after introducing quantization operators, it is much more expensive and time-consuming than PTQ. So et al. (2024) proposes a novel quantization method that enhances Encoder Latentoriginal Condition embedding Random noise (Latent vector ) t0 Latent vector ti+1 Latent vector tn For conditional tasks (example: Text-to-Image) generations Ngen : time step (over time steps) : with operator modified (e.g. conv) : latent vector generated by : score computed between and ti ith n Unetmodx Unetoriginal xth Latentx Unetx Scorex Latentoriginal Latentx Unetoriginal Figure 26: Pruning evaluates changes in the central tendency and variability to determine the significance of each operator. (Castells et al., 2024). output quality by dynamically adjusting the quantization interval based on time step information. Figure 27: Illustrations of the pruning. The proposed approach integrates with the Learned Step Size Quantization (Esser et al., 2019) framework, replacing the static quantization interval with a dynamically generated output from the Time-Dynamic Quantization module. This dynamic adjustment leads to significant improvements in the quality of the quantized outputs. He et al. (2023) introduces a quantization-aware low-rank adapter that integrates with model weights and is jointly quantized to a low bit-width. This Published in Transactions on Machine Learning Research (05/2025) Optimized Hardware-Software Co-design Chen et al. (2023d), SDA (Yang et al., 2023a), Choi et al. (2024a) Parallel Computing Distri Fusion (Li et al., 2024a), Pipe Fusion (Wang et al., 2024a), Swift Diffusion (Li et al., 2024b),Diffusion Pipe (Tian et al., 2024) Caching Technique NIRVANA (Agarwal et al., 2024), Deep Cache (Ma et al., 2024b), Block Caching (Wimbauer et al., 2024), L2C (Ma et al., 2024a), FORA (Selvaraju et al., 2024), MD-Di T (Shen et al.), -Di T (Chen et al.) Figure 28: Summary of system-level efficiency optimization techniques for diffusion models. approach distills the denoising capabilities of full-precision models into their quantized versions, utilizing only a few trainable quantization scales per layer and eliminating the need for training data. 3.4.2 Pruning Pruning compresses DMs by removing redundant or less important model weights. Currently, most pruning methods for DMs focus on pruning structured patterns such as groups of consecutive parameters or hierarchical structures. For instance, Diff-Pruning (Fang et al., 2023) introduces the first dedicated method designed for pruning diffusion models. Diff-Pruning leverages Taylor expansion over pruned timesteps to estimate the importance of weights. By filtering out non-contributory diffusion steps and aggregating informative gradients, Diff-Pruning enhances model efficiency while preserving essential features. LD-Pruner (Castells et al., 2024), as illustrated in Figure 26, on the other hand, proposes a pruning method specifically designed for Latent Diffusion Models (LDMs) The key innovation of LD-Pruner lies in its utilization of the latent space to guide the pruning process. The method enables a precise assessment of pruning impacts by generating multiple sets of latent vectors one set for the original Unet and additional sets for each modified Unet where a single operator is altered. The importance of each operator is then quantified using a specialized formula that considers shifts in both the central tendency and variability of the latent vectors. This approach ensures that the pruning process preserves model performance while adapting to the specific characteristics of LDMs. Kim et al. (2024a) introduces a technique known as Layer Merge, designed to jointly prune convolution layers and activation functions to achieve a desired inference speedup while minimizing performance degradation. Layer Merge addresses the challenge of selecting which layers to remove by formulating a new surrogate optimization problem. Given the exponential nature of the selection space, the authors propose an efficient solution using dynamic programming. Their approach involves constructing dynamic programming (DP) lookup tables that exploit the problem s inherent structure, thereby allowing for an exact and efficient solution to the pruning problem. Lastly, LAPTOPDiff (Zhang et al., 2024a) introduces a layer-pruning technique aimed at automatically compressing the U-Net architecture of diffusion models. The core of this approach is an effective one-shot pruning criterion, distinguished by its favorable additivity property. This property ensures that the one-shot performance of the pruning is superior to other traditional layer pruning methods and manual layer removal techniques. By framing the pruning problem within the context of combinatorial optimization, LAPTOPDiff simplifies the pruning process while achieving significant performance gains. The proposed method stands out for its ability to provide a robust one-shot pruning solution, offering a clear advantage in compressing diffusion models efficiently. 4 System-Level Efficiency Optimization 4.1 Hardware-Software Co-Design The co-design of hardware and software is pivotal for achieving efficient deployment of diffusion models in real-time and resource-constrained environments. Following algorithm-level optimizations, system-level techniques focus on integrating hardware-specific features, distributed computation, and caching mechanisms. These strategies aim to address the computational complexity and memory demands of large-scale diffusion Published in Transactions on Machine Learning Research (05/2025) models, enabling more practical applications across various platforms like GPUs, FPGAs, and mobile devices. One significant contribution is the work by Chen et al. (2023d), which explores GPU-aware optimizations for accelerating diffusion models directly on mobile devices. Implementing specialized kernels and optimized softmax operations reduces inference latency, achieving near real-time performance on mobile GPUs. Conv Weight Figure 29: Illustration of the Hybrid SA architecture from Yang et al. (2023a). In a related effort, Yang et al. (2023a) propose SDA, a low-bit stable diffusion accelerator designed specifically for edge FPGAs. Utilizing quantization-aware training and a hybrid systolic array architecture as illustrated in Figure 29, SDA effectively balances computational efficiency with flexibility, handling both convolutional and attention operations efficiently. Through a two-level pipelining structure, the nonlinear operators are efficiently integrated with the hybrid SA, enabling coordinated operation that enhances processing speed while reducing resource usage. Finally, SDA achieves a speedup of 97.3x when compared to ARM Cortex-A53 CPU. Furthermore, Choi et al. (2024a) introduces a stable diffusion processor optimized for mobile platforms through patch similarity-based sparsity, mixed-precision strategies and and a Dual-mode Bit-Slice Core (DBSC) architecture that supports mixed-precision computation, which particularly targeting resource-constrained devices such as mobile platforms. Together, these optimizations significantly improve throughput and energy efficiency, making Stable Diffusion more viable for energy-sensitive applications. For GPU-accelerated mobile applications requiring real-time interactivity, Chen et al. (2023d) provides kernel-level optimizations and mobile-specific operator tuning with strongest latency. When targeting embedded edge devices with strict power budgets and static workloads, Yang et al. (2023a) achieves superior performance via hardware-algorithm synergy. Meanwhile, Choi et al. (2024a) firstly approach dynamic energy constraints in always-on scenarios through runtime sparsity adaptation. 4.2 Parallel Computing Figure 30: Illustrations of the parallel computing for diffusion models. Parallel computing Figure 30 plays a critical role in the efficient execution of diffusion models, especially given the computation-intensive nature of these algorithms. Recent advances in parallel computing strategies have enabled significant improvements in inference speed and scalability, often without compromising the quality of the generated output (Li et al., 2024a; Wang et al., 2024a; Li et al., 2024b; Tian et al., 2024). This section highlights several notable contributions that tackle the challenge of parallelizing diffusion models across multiple GPUs and other distributed architectures. Li et al. (2024a) introduced Distri Fusion, a framework designed for distributed parallel inference tailored to high-resolution diffusion models such as SDXL. Their approach involves partitioning the model inputs into distinct patches, which are then processed independently across multiple GPUs. This method leverages Published in Transactions on Machine Learning Research (05/2025) the available hardware resources more effectively, achieving a 6.1x speedup on 8x A100 GPUs compared to single-card operation, all while maintaining output quality. To address potential issues arising from the loss of inter-patch interaction, which could compromise global consistency, Distri Fusion employs dynamic synchronization of activation displacements, striking a balance between preserving coherence and minimizing communication overhead. Computation Asynchronous Communication Figure 31: Illustrations of the diffusion architecture from (Li et al., 2024a). Building on the insights gained from Distri Fusion, Wang et al. (2024a) further refined the distributed inference paradigm with Pipe Fusion. This system not only splits images into patches but also distributes the network layers across different devices, thereby reducing the associated communication costs and enabling the use of PCIe-linked GPUs instead of NVLink-connected ones. Pipe Fusion integrates sequence parallelism, tensor parallelism, displaced patch parallelism, and displaced patch pipeline parallelism, optimizing workflow for a wider range of hardware configurations. For applications involving add-on modules such as Control Net and Lo RA, Li et al. (2024b) developed Swift Diffusion, as illustrated in Figure 31. This framework optimizes the serving workflow of these modules, allowing them to run in parallel on multiple GPUs. As a result, Swift Diffusion delivers a 5x reduction in inference latency and a 2x improvement in throughput, ensuring that enhanced speed does not come at the expense of output quality. Lastly, Tian et al. (2024) focused on the training phase with Diffusion Pipe, demonstrating that pipeline parallelism can produce a 1.41x training speedup, while data parallelism contributes an additional 1.28x acceleration. Although the optimization methods for Diffusion Pipe were not detailed in the notes, the combination of these parallelization strategies offers a promising direction to improve the efficiency of both the training and inference pipelines for diffusion models. These methodologies encompass a hierarchical optimization, progressing from coarse-grained spatial partitioning to fine-grained module-specific adaptations, thereby empowering practitioners to strategically align parallelism configurations with underlying hardware constraints. 4.3 Caching Technique In diffusion models, the computational hotspot often centers around discrete time-step diffusion, which is characterized by strong temporal locality. Consequently, building an efficient caching system for diffusion models is nonnegligible to enhance its performance. Indeed, extensive research has been conducted on optimizing caching systems in Figure 32, resulting in significant advancements in this field. Agarwal et al. (2024) proposed NIRVANA, a novel system designed to enhance the efficiency of text-to-image generation using diffusion models. Specifically, the key innovation lies in its approximate caching technique, which reduces computational costs and latency by reusing intermediate noise states from previous image generation processes. Instead of starting from scratch with every new text prompt, NIRVANA retrieves and reconditions these cached states, allowing it to skip several initial denoising steps. Additionally, the system uses a custom cache management policy called Least Computationally Beneficial and Frequently Used (LCBFU), which optimizes the storage and reuse of cached states to maximize computational efficiency. This Published in Transactions on Machine Learning Research (05/2025) Figure 32: Illustrations of the caching system for diffusion models focus on the U-Net block and the Transformer layer, critical components for effectively implementing caching techniques. makes NIRVANA particularly suited for large-scale, production-level deployments of text-to-image diffusion models. From another perspective, Ma et al. (2024b) introduces an innovative approach called Deep Cache, designed to accelerate the image generation process by leveraging the temporal redundancy in the denoising steps of diffusion models, without the need for additional model training, as illustrated in Figure 33. The key insight is the observation that high-level features, such as the main structure and shape of an image, exhibit minimal changes between adjacent denoising steps. These features can be cached and reused in subsequent steps, thereby avoiding redundant computations. This method takes advantage of the U-Net architecture by combining these cached high-level features with low-level features, updating only the low-level features to reduce computational load, leading to a significant acceleration in the overall process. Wimbauer et al. (2024) proposed Block Caching, a technique that identifies and caches redundant computations within the model s layers during the denoising process. By reusing these cached outputs in subsequent timesteps, the method significantly speeds up inference while maintaining image quality. To optimize this caching process, they introduce an Automatic Cache Scheduling mechanism, which dynamically determines when and where to cache based on the relative changes in layer outputs over time. Additionally, the paper addresses potential misalignment issues from aggressive caching by implementing a Scale-Shift Adjustment mechanism, which fine-tunes cached outputs to align with the model s expectations, thereby preventing visual artifacts. Skip Branch 1 Chosen Path Skip Branch 2 Skip Branch 3 Excluded Path Figure 33: Illustration of the caching system from (Ma et al., 2024b). Recently, the application of diffusion with transformer models has yielded considerable success. Ma et al. (2024a) is concerned with the introduction of a layer caching mechanissm, designated Learning-to-Cache (L2C), to accelerate diffusion transformer models. L2C exploits the redundancy between layers within the transformer architecture, dynamically caching computations from certain layers to reduce redundant calcu- Published in Transactions on Machine Learning Research (05/2025) lations and lower inference costs. The implementation entails transforming the layer selection problem into a differentiable optimization problem, using interpolation to determine whether to perform a full computation or utilize cached results at different timesteps during inference. In contrast to the emphasis on layer caching, Selvaraju et al. (2024) proposed Fast-Forward Caching (FORA), a technique designed to accelerate Diffusion Transformers (Di T) by reducing redundant computations during the inference phase. The key insight behind FORA is the observation that the outputs from the self-attention and MLP layers in a Transformer exhibit high similarity across consecutive time steps in the diffusion process. To leverage this, FORA implements a static caching mechanism where these layer outputs are cached at regular intervals, which are determined by N, and reused for a set number of subsequent steps, thereby avoiding recomputing similar outputs. Shen et al. introduces MD-Di T, a unified framework for efficient diffusion transformers by integrating block skipping and caching strategies, enabling dynamic depth adjustment across timesteps without additional training. It optimizes computation by caching incremental changes from previous timesteps and selectively skipping non-critical blocks. Chen et al. proposes -Di T, a groundbreaking framework for optimizing diffusion transformers through two synergistic innovations: step-sensitive block caching and intelligently allocated computational resources. By exploiting the temporal consistency inherent in diffusion trajectories, the system strategically reuses intermediate residual features across sequential steps. Notably, it harmonizes both computation-skipping and feature-caching mechanisms within a cohesive optimization architecture. Generally, these methods demonstrate an trajectory from coarse temporal reuse to fine-grained layer adaptation, guiding practitioners to employ noise-state caching for prompt-variant scenarios, feature-level caching for structural consistency, and hybrid static-dynamic strategies for transformer-based models. 5 Frameworks Table 2: Comparison of Optimization Support in Selected Diffusion Model Frameworks and Techniques. In this table, Training indicates whether the framework accelerates the training process of diffusion models, while Inference indicates whether it accelerates the generation process. Framework Training Inference Key Features Flash Attention High-efficiency attention computation for Diffusion Transformers (Di T) x Formers Memory-efficient attention and modular ops tailored for diffusion Transformer speedups Deep Speed Scalable distributed training and inference optimizations for large diffusion models One Flow Compiler-optimized pipeline for faster diffusion model training and sampling Stable-Fast q Fast inference optimization for Diffusers with CUDNN fusion Onediff q Diffusion-specific acceleration with Deep Cache and quantization Deep Cache q Reuses cached diffusion features to speed up inference iterations TGATE q Temporal gating to streamline crossattention in diffusion inference x Di T q Parallel inference engine for Diffusion Transformers Frameworks in the efficient diffusion landscape refer to high-performance tools and libraries designed to optimize training and/or inference. DM frameworks can be in general grouped based on whether they Published in Transactions on Machine Learning Research (05/2025) support the tasks of training and inference. Specifically, frameworks that support training aim to provide scalable, efficient, and flexible infrastructure that improves computation efficiency, reduces memory footprint, optimizes communication efficiency, and ensures reliability of the training process. Frameworks that support inference focus on optimizing inference throughput and reducing memory footprint and latency. A subset of frameworks, including Flash Attention (Dao et al., 2022), x Formers (Lefaudeux et al., 2022), Deep Speed (Rasley et al., 2020), and One Flow (Yuan et al., 2021), distinguishes itself by offering comprehensive support for both training and inference phases of diffusion models. These tools, rooted in broader deep learning optimization efforts, cater to the full lifecycle of model development. Flash Attention accelerates both training and inference by optimizing attention computation, reducing memory usage and latency through techniques like tiling and recomputation, making it particularly effective for Diffusion Transformers (Di T). x Formers provides memory-efficient Transformer optimizations with a modular design, enabling flexible and efficient computation across a range of resource-intensive tasks. Deep Speed, originally designed for large-scale model training, extends its distributed capabilities to inference, enabling scalable deployment across multiple GPUs. Similarly, One Flow leverages its compiler-driven architecture to streamline both training and inference workflows, appealing to researchers and practitioners seeking end-to-end optimization. Together, these frameworks provide robust foundations for advancing diffusion model research and deployment, balancing flexibility with high performance. In contrast, a growing array of tools Stable-Fast (chengzeyi, 2024), Onediff (Contributors, 2022), Deep Cache (Ma et al., 2024b), TGATE (Zhang et al., 2024b; Liu et al.), and x Di T (Fang et al., 2024a;b; Fang & Zhao, 2024; Sun et al., 2024) focuses exclusively on accelerating the inference stage of diffusion models, addressing the demand for rapid and resource-efficient generation. Stable-Fast optimizes the Hugging Face Diffusers ecosystem, prioritizing low-latency inference for real-time applications. Onediff builds on this trend by integrating cutting-edge techniques like caching and quantization, tailoring its acceleration to diffusion-specific workloads. Deep Cache and TGATE introduce innovative caching strategies, exploiting temporal redundancies to reduce computational overhead in inference, particularly for U-Net-based models. Meanwhile, x Di T targets the emerging Diffusion Transformer (Di T) architecture, employing parallelization to enhance inference scalability. These inference-centric tools reflect a shift toward specialized optimizations, catering to the practical needs of deployment in constrained environments or novel model paradigms. 6 Future Work Despite the significant progress made in efficient diffusion models, several promising research directions remain open. In particular, we identify the following key areas for future work: Hybridizing Diffusion and Autoregressive Models. One promissing future direction is to explore hybridizing autoregressive and diffusion models to combine the strengths of both paradigms. This integration enables the use of key-value (KV) caching, a technique from autoregressive transformers, to accelerate diffusion and support streamable generation. A representative work is Block Diffusion(Arriola et al., 2025), which segments generation into blocks and applies autoregressive-style caching across denoising steps. This approach opens up promising opportunities to unify architectural benefits, allowing diffusion models to inherit the efficiency and online capabilities of autoregressive methods. Without Classifier-Free Guidance (CFG). Although classifier-free guidance has been widely adopted to enhance generation quality, it introduces substantial computational overhead. Tang et al. (Tang et al., 2025) propose Model-guidance (MG), a novel training objective that eliminates the need for CFG by directly incorporating the posterior probability of conditions rather than solely modeling data distribution. This approach not only doubles inference speed by avoiding the second network forward pass required by CFG but also significantly accelerates model training with 6.5 faster convergence and approximately 60% performance improvement. Notably, when compared to concurrent methods, MG achieves state-of-the-art performance on Image Net 256 with an FID of 1.34 while requiring only about 12% of the computational resources of comparable approaches. The effectiveness of MG suggests a promising direction for future research to explore alternative guidance mechanisms or fundamentally rethink the training paradigm to better balance computational efficiency and generation quality. Published in Transactions on Machine Learning Research (05/2025) Efficient Attention Mechanisms for Video Diffusion. Bidirectional attention has become a dominant component in diffusion architectures, but it introduces substantial computational overhead that scales quadratically with sequence length. This is particularly problematic for video diffusion models where the sequence length grows linearly with the number of frames. Recent works have begun to explore efficient attention mechanisms specifically designed for diffusion transformers (Di Ts). Xia et al. (Xia et al., 2025) propose Ada Spa, which leverages the hierarchical sparsity inherent in Di Ts through a blockified pattern approach and adaptive search methods. Their work demonstrates that sparse characteristics of Di Ts exhibit hierarchical structures between different modalities and remain invariant across denoising steps, enabling significant computational savings. Similarly, Ding et al. (Ding et al., 2025) identify tile-style repetitive patterns in 3D attention maps for video data, introducing sparse 3D attention with linear complexity relative to frame count. Their approach combines efficient attention with consistency distillation techniques to enable up to 7.8 faster generation for high-resolution videos. These developments suggest promising directions for mitigating the computational burden of attention in diffusion models without sacrificing generation quality. 7 Conclusion In this survey, we provide a systematic review of efficient diffusion models, an important area of research aimed at democratizing diffusion models. We start with motivating the necessity for efficient diffusion models. Guided by a taxonomy, we review efficient techniques for diffusion models from algorithm-level and system-level perspectives respectively. Furthermore, we review diffusion models frameworks with specific optimizations and features crucial for efficient diffusion models. We believe that efficiency will play an increasingly important role in diffusion models and diffusion models-oriented systems. We hope this survey could enable researchers and practitioners to quickly get started in this field and act as a catalyst to inspire new research on efficient diffusion models. Shubham Agarwal, Subrata Mitra, Sarthak Chakraborty, Srikrishna Karanam, Koyel Mukherjee, and Shiv Kumar Saini. Approximate caching for efficiently serving {Text-to-Image} diffusion models. In 21st USENIX Symposium on Networked Systems Design and Implementation (NSDI 24), pp. 1173 1189, 2024. Michael S Albergo, Nicholas M Boffi, and Eric Vanden-Eijnden. Stochastic interpolants: A unifying framework for flows and diffusions. ar Xiv preprint ar Xiv:2303.08797, 2023. Marianne Arriola, Aaron Gokaslan, Justin T Chiu, Zhihan Yang, Zhixuan Qi, Jiaqi Han, Subham Sekhar Sahoo, and Volodymyr Kuleshov. Block diffusion: Interpolating between autoregressive and diffusion language models. In The Thirteenth International Conference on Learning Representations, 2025. URL https://arxiv.org/abs/2503.09573. Fan Bao, Shen Nie, Kaiwen Xue, Yue Cao, Chongxuan Li, Hang Su, and Jun Zhu. All are worth words: A vit backbone for diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 22669 22679, 2023. Florentin Bieder, Julia Wolleb, Alicia Durrer, Robin Sandkuehler, and Philippe C Cattin. Memory-efficient 3d denoising diffusion models for medical image processing. In Medical Imaging with Deep Learning, 2023. Andreas Blattmann, Robin Rombach, Kaan Oktay, Jonas Müller, and Björn Ommer. Semi-parametric neural image synthesis. ar Xiv preprint ar Xiv:2204.11824, 2022. Andreas Blattmann, Robin Rombach, Huan Ling, Tim Dockhorn, Seung Wook Kim, Sanja Fidler, and Karsten Kreis. Align your latents: High-resolution video synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 22563 22575, 2023. Evelyn Buckwar and Renate Winkler. Multistep methods for sdes and their application to problems with small noise. SIAM journal on numerical analysis, 44(2):779 803, 2006. Published in Transactions on Machine Learning Research (05/2025) Hanqun Cao, Cheng Tan, Zhangyang Gao, Yilun Xu, Guangyong Chen, Pheng-Ann Heng, and Stan Z. Li. A survey on generative diffusion models. IEEE Transactions on Knowledge and Data Engineering, 36(7): 2814 2830, 2024. doi: 10.1109/TKDE.2024.3361474. Thibault Castells, Hyoung-Kyu Song, Bo-Kyeong Kim, and Shinkook Choi. Ld-pruner: Efficient pruning of latent diffusion models using task-agnostic insights. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 821 830, 2024. Duygu Ceylan, Chun-Hao P Huang, and Niloy J Mitra. Pix2video: Video editing using image diffusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 23206 23217, 2023. Jiaao Chen, Aston Zhang, Mu Li, Alex Smola, and Diyi Yang. A cheaper and better diffusion language model with soft-masked noise. ar Xiv preprint ar Xiv:2304.04746, 2023a. Junsong Chen, Shuchen Xue, Yuyang Zhao, Jincheng Yu, Sayak Paul, Junyu Chen, Han Cai, Enze Xie, and Song Han. Sana-sprint: One-step diffusion with continuous-time consistency distillation. ar Xiv preprint ar Xiv:2503.09641, 2025. Minshuo Chen, Song Mei, Jianqing Fan, and Mengdi Wang. An overview of diffusion models: Applications, guided generation, statistical rates and optimization. ar Xiv preprint ar Xiv:2404.07771, 2024. Nanxin Chen, Yu Zhang, Heiga Zen, Ron J Weiss, Mohammad Norouzi, and William Chan. Wavegrad: Estimating gradients for waveform generation. ar Xiv preprint ar Xiv:2009.00713, 2020. Pengtao Chen, Mingzhu Shen, Peng Ye, Jianjian Cao, Chongjun Tu, Christos-Savvas Bouganis, Yiren Zhao, and Tao Chen. δ-dit: Accelerating diffusion transformers without training via denoising property alignment. Ricky TQ Chen, Yulia Rubanova, Jesse Bettencourt, and David K Duvenaud. Neural ordinary differential equations. Advances in neural information processing systems, 31, 2018. Shoufa Chen, Peize Sun, Yibing Song, and Ping Luo. Diffusiondet: Diffusion model for object detection. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 19830 19843, 2023b. Ting Chen, Ruixiang Zhang, and Geoffrey Hinton. Analog bits: Generating discrete data using diffusion models with self-conditioning. ar Xiv preprint ar Xiv:2208.04202, 2022a. Wenhu Chen, Hexiang Hu, Chitwan Saharia, and William W Cohen. Re-imagen: Retrieval-augmented text-to-image generator. In The Eleventh International Conference on Learning Representations, 2022b. Xin Chen, Biao Jiang, Wen Liu, Zilong Huang, Bin Fu, Tao Chen, and Gang Yu. Executing your commands via motion diffusion in latent space. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 18000 18010, 2023c. Yu-Hui Chen, Raman Sarokin, Juhyun Lee, Jiuqiang Tang, Chuo-Ling Chang, Andrei Kulik, and Matthias Grundmann. Speed is all you need: On-device acceleration of large diffusion models via gpu-aware optimizations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4651 4655, 2023d. chengzeyi. Stable fast. https://github.com/chengzeyi/stable-fast, 2024. Jiwon Choi, Wooyoung Jo, Seongyon Hong, Beomseok Kwon, Wonhoon Park, and Hoi-Jun Yoo. A 28.6 mj/iter stable diffusion processor for text-to-image generation with patch similarity-based sparsity augmentation and text-based mixed-precision. ar Xiv preprint ar Xiv:2403.04982, 2024a. Joo Young Choi, Jaesung R Park, Inkyu Park, Jaewoong Cho, Albert No, and Ernest K Ryu. Simple drop-in lora conditioning on attention layers will improve your diffusion model. ar Xiv preprint ar Xiv:2405.03958, 2024b. Published in Transactions on Machine Learning Research (05/2025) One Diff Contributors. Onediff: An out-of-the-box acceleration library for diffusion models. https://gith ub.com/siliconflow/onediff, 2022. Guillaume Couairon, Jakob Verbeek, Holger Schwenk, and Matthieu Cord. Diffedit: Diffusion-based semantic image editing with mask guidance. ar Xiv preprint ar Xiv:2210.11427, 2022. Florinel-Alin Croitoru, Vlad Hondru, Radu Tudor Ionescu, and Mubarak Shah. Diffusion models in vision: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(9):10850 10869, 2023. Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher Ré. Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in neural information processing systems, 35:16344 16359, 2022. Giannis Daras, Hyungjin Chung, Chieh-Hsin Lai, Yuki Mitsufuji, Jong Chul Ye, Peyman Milanfar, Alexandros G. Dimakis, and Mauricio Delbracio. A survey on diffusion models for inverse problems, 2024. URL https://arxiv.org/abs/2410.00083. Dale Decatur, Itai Lang, Kfir Aberman, and Rana Hanocka. 3d paintbrush: Local stylization of 3d shapes with cascaded score distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4473 4483, 2024. Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, and Luke Zettlemoyer. Qlora: Efficient finetuning of quantized llms. Advances in Neural Information Processing Systems, 36, 2024. Hangliang Ding, Dacheng Li, Runlong Su, Peiyuan Zhang, Zhijie Deng, Ion Stoica, and Hao Zhang. Efficientvdit: Efficient video diffusion transformers with attention tile. ar Xiv preprint ar Xiv:2502.06155, 2025. Tim Dockhorn, Arash Vahdat, and Karsten Kreis. Score-based generative modeling with critically-damped langevin diffusion. ar Xiv preprint ar Xiv:2112.07068, 2021. Bi an Du, Wei Hu, and Renjie Liao. Multi-scale latent point consistency models for 3d shape generation. ar Xiv preprint ar Xiv:2412.19413, 2024. Steven K Esser, Jeffrey L Mc Kinstry, Deepika Bablani, Rathinakumar Appuswamy, and Dharmendra S Modha. Learned step size quantization. ar Xiv preprint ar Xiv:1902.08153, 2019. Gongfan Fang, Xinyin Ma, and Xinchao Wang. Structural pruning for diffusion models. In Advances in Neural Information Processing Systems, 2023. Jiarui Fang and Shangchun Zhao. A unified sequence parallelism approach for long context generative ai. ar Xiv preprint ar Xiv:2405.07719, 2024. Jiarui Fang, Jinzhe Pan, Xibo Sun, Aoyu Li, and Jiannan Wang. xdit: an inference engine for diffusion transformers (dits) with massive parallelism. ar Xiv preprint ar Xiv:2411.01738, 2024a. Jiarui Fang, Jinzhe Pan, Jiannan Wang, Aoyu Li, and Xibo Sun. Pipefusion: Patch-level pipeline parallelism for diffusion transformers inference. ar Xiv preprint ar Xiv:2405.14430, 2024b. Yuxiang Fu, Qi Yan, Lele Wang, Ke Li, and Renjie Liao. Moflow: One-step flow matching for human trajectory forecasting via implicit maximum likelihood estimation based distillation. ar Xiv preprint ar Xiv:2503.09950, 2025. Rohit Gandikota, Joanna Materzynska, Tingrui Zhou, Antonio Torralba, and David Bau. Concept sliders: Lora adaptors for precise control in diffusion models. ar Xiv preprint ar Xiv:2311.12092, 2023. Ruiqi Gao, Emiel Hoogeboom, Jonathan Heek, Valentin De Bortoli, Kevin P. Murphy, and Tim Salimans. Diffusion meets flow matching: Two sides of the same coin. 2024. URL https://diffusionflow.github .io/. Shansan Gong, Mukai Li, Jiangtao Feng, Zhiyong Wu, and Ling Peng Kong. Diffuseq: Sequence to sequence text generation with diffusion models. ar Xiv preprint ar Xiv:2210.08933, 2022. Published in Transactions on Machine Learning Research (05/2025) Yuchao Gu, Xintao Wang, Jay Zhangjie Wu, Yujun Shi, Yunpeng Chen, Zihan Fan, Wuyou Xiao, Rui Zhao, Shuning Chang, Weijia Wu, et al. Mix-of-show: Decentralized low-rank adaptation for multi-concept customization of diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Jiaqi Guan, Xiangxin Zhou, Yuwei Yang, Yu Bao, Jian Peng, Jianzhu Ma, Qiang Liu, Liang Wang, and Quanquan Gu. Decompdiff: Diffusion models with decomposed priors for structure-based drug design. Proceedings of Machine Learning Research, 202:11827 11846, 2023. Hanzhong Guo, Cheng Lu, Fan Bao, Tianyu Pang, Shuicheng Yan, Chao Du, and Chongxuan Li. Gaussian mixture solvers for diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Tiankai Hang and Shuyang Gu. Improved noise schedule for diffusion training. ar Xiv preprint ar Xiv:2407.03297, 2024. Yefei He, Jing Liu, Weijia Wu, Hong Zhou, and Bohan Zhuang. Efficientdm: Efficient quantization-aware fine-tuning of low-bit diffusion models. ar Xiv preprint ar Xiv:2310.03270, 2023. Yefei He, Luping Liu, Jing Liu, Weijia Wu, Hong Zhou, and Bohan Zhuang. Ptqd: Accurate post-training quantization for diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Yingqing He, Tianyu Yang, Yong Zhang, Ying Shan, and Qifeng Chen. Latent video diffusion models for high-fidelity long video generation. ar Xiv preprint ar Xiv:2211.13221, 2022. Shashank Hegde, Sumeet Batra, KR Zentner, and Gaurav Sukhatme. Generating behaviorally diverse policies with latent diffusion models. Advances in Neural Information Processing Systems, 36:7541 7554, 2023. Eric Heitz, Laurent Belcour, and Thomas Chambon. Iterative α-(de) blending: A minimalist deterministic diffusion model. In ACM SIGGRAPH 2023 Conference Proceedings, pp. 1 8, 2023. Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. ar Xiv preprint ar Xiv:1503.02531, 2015. Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840 6851, 2020. Jonathan Ho, Chitwan Saharia, William Chan, David J Fleet, Mohammad Norouzi, and Tim Salimans. Cascaded diffusion models for high fidelity image generation. Journal of Machine Learning Research, 23 (47):1 33, 2022a. Jonathan Ho, Tim Salimans, Alexey Gritsenko, William Chan, Mohammad Norouzi, and David J Fleet. Video diffusion models. Advances in Neural Information Processing Systems, 35:8633 8646, 2022b. Marlis Hochbruck and Alexander Ostermann. Exponential integrators. Acta Numerica, 19:209 286, 2010. Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. Lora: Low-rank adaptation of large language models. ar Xiv preprint ar Xiv:2106.09685, 2021. Kaixuan Huang, Yukang Yang, Kaidi Fu, Yanyi Chu, Le Cong, and Mengdi Wang. Latent diffusion models for controllable rna sequence generation. ar Xiv preprint ar Xiv:2409.09828, 2024. Aapo Hyvärinen and Peter Dayan. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005. Max Jiang, Yijing Bai, Andre Cornman, Christopher Davis, Xiukun Huang, Hong Jeon, Sakshum Kulshrestha, John Lambert, Shuangyu Li, Xuanyu Zhou, et al. Scenediffuser: Efficient and controllable driving simulation initialization and rollout. Advances in Neural Information Processing Systems, 37: 55729 55760, 2024. Alexia Jolicoeur-Martineau, Ke Li, Rémi Piché-Taillefer, Tal Kachman, and Ioannis Mitliagkas. Gotta go fast when generating data with score-based models. ar Xiv preprint ar Xiv:2105.14080, 2021. Published in Transactions on Machine Learning Research (05/2025) Tero Karras, Miika Aittala, Timo Aila, and Samuli Laine. Elucidating the design space of diffusion-based generative models. Advances in neural information processing systems, 35:26565 26577, 2022. Amirhossein Kazerouni, Ehsan Khodapanah Aghdam, Moein Heidari, Reza Azad, Mohsen Fayyaz, Ilker Hacihaliloglu, and Dorit Merhof. Diffusion models in medical imaging: A comprehensive survey. Medical Image Analysis, 88:102846, 2023. ISSN 1361-8415. doi: https://doi.org/10.1016/j.media.2023.102846. URL https://www.sciencedirect.com/science/article/pii/S1361841523001068. Beomsu Kim and Jong Chul Ye. Denoising mcmc for accelerating diffusion-based generative models. ar Xiv preprint ar Xiv:2209.14593, 2022. Jinuk Kim, Marwa El Halabi, Mingi Ji, and Hyun Oh Song. Layermerge: Neural network depth compression through layer pruning and merging. ar Xiv preprint ar Xiv:2406.12837, 2024a. Yulhwa Kim, Dongwon Jo, Hyesung Jeon, Taesu Kim, Daehyun Ahn, Hyungjun Kim, et al. Leveraging earlystage robustness in diffusion models for efficient and high-quality image synthesis. Advances in Neural Information Processing Systems, 36, 2024b. Diederik P Kingma, Max Welling, et al. Auto-encoding variational bayes, 2013. Akio Kodaira, Chenfeng Xu, Toshiki Hazama, Takanori Yoshimoto, Kohei Ohno, Shogo Mitsuhori, Soichi Sugano, Hanying Cho, Zhijian Liu, and Kurt Keutzer. Streamdiffusion: A pipeline-level solution for real-time interactive generation. ar Xiv preprint ar Xiv:2312.12491, 2023. Zhifeng Kong and Wei Ping. On fast sampling of diffusion probabilistic models. ar Xiv preprint ar Xiv:2106.00132, 2021. Zhifeng Kong, Wei Ping, Jiaji Huang, Kexin Zhao, and Bryan Catanzaro. Diffwave: A versatile diffusion model for audio synthesis. ar Xiv preprint ar Xiv:2009.09761, 2020. Sang-gil Lee, Heeseung Kim, Chaehun Shin, Xu Tan, Chang Liu, Qi Meng, Tao Qin, Wei Chen, Sungroh Yoon, and Tie-Yan Liu. Priorgrad: Improving conditional denoising diffusion models with data-dependent adaptive prior. In International Conference on Learning Representations, 2021. Sangyun Lee, Zinan Lin, and Giulia Fanti. Improving the training of rectified flows. ar Xiv preprint ar Xiv:2405.20320, 2024a. Taegyeong Lee, Soyeong Kwon, and Taehwan Kim. Grid diffusion models for text-to-video generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8734 8743, 2024b. Benjamin Lefaudeux, Francisco Massa, Diana Liskovich, Wenhan Xiong, Vittorio Caggiano, Sean Naren, Min Xu, Jieru Hu, Marta Tintore, Susan Zhang, Patrick Labatut, Daniel Haziza, Luca Wehrstedt, Jeremy Reizenstein, and Grigory Sizov. xformers: A modular and hackable transformer modelling library. https: //github.com/facebookresearch/xformers, 2022. Ming Li, Taojiannan Yang, Huafeng Kuang, Jie Wu, Zhaoning Wang, Xuefeng Xiao, and Chen Chen. Controlnet++: Improving conditional controls with efficient consistency feedback. In European Conference on Computer Vision, pp. 129 147. Springer, 2025. Muyang Li, Tianle Cai, Jiaxin Cao, Qinsheng Zhang, Han Cai, Junjie Bai, Yangqing Jia, Kai Li, and Song Han. Distrifusion: Distributed parallel inference for high-resolution diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7183 7193, 2024a. Suyi Li, Lingyun Yang, Xiaoxiao Jiang, Hanfeng Lu, Zhipeng Di, Weiyi Lu, Jiawei Chen, Kan Liu, Yinghao Yu, Tao Lan, Guodong Yang, Lin Qu, Liping Zhang, and Wei Wang. Swiftdiffusion: Efficient diffusion model serving with add-on modules, 2024b. URL https://arxiv.org/abs/2407.02031. Published in Transactions on Machine Learning Research (05/2025) Xiang Li, John Thickstun, Ishaan Gulrajani, Percy S Liang, and Tatsunori B Hashimoto. Diffusion-lm improves controllable text generation. Advances in Neural Information Processing Systems, 35:4328 4343, 2022. Xinyang Li, Zhangyu Lai, Linning Xu, Jianfei Guo, Liujuan Cao, Shengchuan Zhang, Bo Dai, and Rongrong Ji. Dual3d: Efficient and consistent text-to-3d generation with dual-mode multi-view latent diffusion. ar Xiv preprint ar Xiv:2405.09874, 2024c. Xiuyu Li, Yijiang Liu, Long Lian, Huanrui Yang, Zhen Dong, Daniel Kang, Shanghang Zhang, and Kurt Keutzer. Q-diffusion: Quantizing diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 17535 17545, 2023. Yiheng Li, Heyang Jiang, Akio Kodaira, Masayoshi Tomizuka, Kurt Keutzer, and Chenfeng Xu. Immiscible diffusion: Accelerating diffusion training with noise assignment. ar Xiv preprint ar Xiv:2406.12303, 2024d. Yuhang Li, Ruihao Gong, Xu Tan, Yang Yang, Peng Hu, Qi Zhang, Fengwei Yu, Wei Wang, and Shi Gu. Brecq: Pushing the limit of post-training quantization by block reconstruction. In International Conference on Learning Representations, 2021. Han Lin, Jaemin Cho, Abhay Zala, and Mohit Bansal. Ctrl-adapter: An efficient and versatile framework for adapting diverse controls to any diffusion model. ar Xiv preprint ar Xiv:2404.09967, 2024a. Jianghao Lin, Jiaqi Liu, Jiachen Zhu, Yunjia Xi, Chengkai Liu, Yangtian Zhang, Yong Yu, and Weinan Zhang. A survey on diffusion models for recommender systems. ar Xiv preprint ar Xiv:2409.05033, 2024b. Lequan Lin, Zhengkun Li, Ruikun Li, Xuliang Li, and Junbin Gao. Diffusion models for time-series applications: a survey. Frontiers of Information Technology & Electronic Engineering, 25(1):19 41, 2024c. Yaron Lipman, Ricky TQ Chen, Heli Ben-Hamu, Maximilian Nickel, and Matt Le. Flow matching for generative modeling. ar Xiv preprint ar Xiv:2210.02747, 2022. Haohe Liu, Zehua Chen, Yi Yuan, Xinhao Mei, Xubo Liu, Danilo Mandic, Wenwu Wang, and Mark D Plumbley. Audioldm: Text-to-audio generation with latent diffusion models. ar Xiv preprint ar Xiv:2301.12503, 2023a. Haozhe Liu, Wentian Zhang, Jinheng Xie, Francesco Faccio, Mengmeng Xu, Tao Xiang, Mike Zheng Shou, Juan-Manuel Perez-Rua, and Jürgen Schmidhuber. Faster diffusion through temporal attention decomposition. Transactions on Machine Learning Research. Jiawei Liu, Weining Wang, Wei Liu, Qian He, and Jing Liu. Ed-t2v: An efficient training framework for diffusion-based text-to-video generation. In 2023 International Joint Conference on Neural Networks (IJCNN), pp. 1 8. IEEE, 2023b. Luping Liu, Yi Ren, Zhijie Lin, and Zhou Zhao. Pseudo numerical methods for diffusion models on manifolds. ar Xiv preprint ar Xiv:2202.09778, 2022a. Qiang Liu. Rectified flow: A marginal preserving approach to optimal transport. ar Xiv preprint ar Xiv:2209.14577, 2022. Xingchao Liu, Chengyue Gong, and Qiang Liu. Flow straight and fast: Learning to generate and transfer data with rectified flow. ar Xiv preprint ar Xiv:2209.03003, 2022b. Xingchao Liu, Xiwen Zhang, Jianzhu Ma, Jian Peng, et al. Instaflow: One step is enough for high-quality diffusion-based text-to-image generation. In The Twelfth International Conference on Learning Representations, 2023c. Justin Lovelace, Varsha Kishore, Chao Wan, Eliot Shekhtman, and Kilian Q Weinberger. Latent diffusion for language generation. Advances in Neural Information Processing Systems, 36, 2024. Published in Transactions on Machine Learning Research (05/2025) Cheng Lu and Yang Song. Simplifying, stabilizing and scaling continuous-time consistency models. ar Xiv preprint ar Xiv:2410.11081, 2024. Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. Dpm-solver++: Fast solver for guided sampling of diffusion probabilistic models. ar Xiv preprint ar Xiv:2211.01095, 2022. Eric Luhman and Troy Luhman. Knowledge distillation in iterative generative models for improved sampling speed. ar Xiv preprint ar Xiv:2101.02388, 2021. Simian Luo, Yiqin Tan, Longbo Huang, Jian Li, and Hang Zhao. Latent consistency models: Synthesizing high-resolution images with few-step inference. ar Xiv preprint ar Xiv:2310.04378, 2023a. Simian Luo, Yiqin Tan, Suraj Patil, Daniel Gu, Patrick von Platen, Apolinário Passos, Longbo Huang, Jian Li, and Hang Zhao. Lcm-lora: A universal stable-diffusion acceleration module. ar Xiv preprint ar Xiv:2311.05556, 2023b. Weijian Luo. A comprehensive survey on knowledge distillation of diffusion models. ar Xiv preprint ar Xiv:2304.04262, 2023. Weijian Luo, Tianyang Hu, Shifeng Zhang, Jiacheng Sun, Zhenguo Li, and Zhihua Zhang. Diff-instruct: A universal approach for transferring knowledge from pre-trained diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Xinyin Ma, Gongfan Fang, Michael Bi Mi, and Xinchao Wang. Learning-to-cache: Accelerating diffusion transformer via layer caching. ar Xiv preprint ar Xiv:2406.01733, 2024a. Xinyin Ma, Gongfan Fang, and Xinchao Wang. Deepcache: Accelerating diffusion models for free. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15762 15772, 2024b. Zhiyuan Ma, Yuzhu Zhang, Guoli Jia, Liangliang Zhao, Yichao Ma, Mingjie Ma, Gaofeng Liu, Kaiyan Zhang, Jianjun Li, and Bowen Zhou. Efficient diffusion models: A comprehensive survey from principles to practices. ar Xiv preprint ar Xiv:2410.11795, 2024c. Weibo Mao, Chenxin Xu, Qi Zhu, Siheng Chen, and Yanfeng Wang. Leapfrog diffusion model for stochastic trajectory prediction. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 5517 5526, 2023. Andrew Melnik, Michal Ljubljanac, Cong Lu, Qi Yan, Weiming Ren, and Helge Ritter. Video diffusion models: A survey. ar Xiv preprint ar Xiv:2405.03150, 2024. Chenlin Meng, Robin Rombach, Ruiqi Gao, Diederik Kingma, Stefano Ermon, Jonathan Ho, and Tim Salimans. On distillation of guided diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14297 14306, 2023. Ben Mildenhall, Pratul P Srinivasan, Matthew Tancik, Jonathan T Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM, 65(1):99 106, 2021. Shentong Mo. Efficient 3d shape generation via diffusion mamba with bidirectional ssms. ar Xiv preprint ar Xiv:2406.05038, 2024. Taehong Moon, Moonseok Choi, Eung Gu Yun, Jongmin Yoon, Gayoung Lee, Jaewoong Cho, and Juho Lee. A simple early exiting framework for accelerated sampling in diffusion models. ar Xiv preprint ar Xiv:2408.05927, 2024a. Taehong Moon, Moonseok Choi, Eung Gu Yun, Jongmin Yoon, Gayoung Lee, Jaewoong Cho, and Juho Lee. A simple early exiting framework for accelerated sampling in diffusion models. In Forty-first International Conference on Machine Learning, 2024b. Published in Transactions on Machine Learning Research (05/2025) Chong Mou, Xintao Wang, Liangbin Xie, Yanze Wu, Jian Zhang, Zhongang Qi, and Ying Shan. T2i-adapter: Learning adapters to dig out more controllable ability for text-to-image diffusion models. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pp. 4296 4304, 2024. Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. In International conference on machine learning, pp. 8162 8171. PMLR, 2021. Shen Nie, Hanzhong Allan Guo, Cheng Lu, Yuhao Zhou, Chenyu Zheng, and Chongxuan Li. The blessing of randomness: Sde beats ode in general diffusion-based image editing. ar Xiv preprint ar Xiv:2311.01410, 2023. Jangho Park, Gihyun Kwon, and Jong Chul Ye. Ed-nerf: Efficient text-guided editing of 3d scene using latent space nerf. ar Xiv preprint ar Xiv:2310.02712, 2023. William Peebles and Saining Xie. Scalable diffusion models with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4195 4205, 2023. Stefano Peluchetti. Non-denoising forward-time diffusions. ar Xiv preprint ar Xiv:2312.14589, 2023. Bohao Peng, Jian Wang, Yuechen Zhang, Wenbo Li, Ming-Chang Yang, and Jiaya Jia. Controlnext: Powerful and efficient control for image and video generation. ar Xiv preprint ar Xiv:2408.06070, 2024a. Mingxing Peng, Kehua Chen, Xusen Guo, Qiming Zhang, Hongliang Lu, Hui Zhong, Di Chen, Meixin Zhu, and Hai Yang. Diffusion models for intelligent transportation systems: A survey. ar Xiv preprint ar Xiv:2409.15816, 2024b. Ashwini Pokle, Zhengyang Geng, and J Zico Kolter. Deep equilibrium approaches to diffusion models. Advances in Neural Information Processing Systems, 35:37975 37990, 2022. Ben Poole, Ajay Jain, Jonathan T. Barron, and Ben Mildenhall. Dreamfusion: Text-to-3d using 2d diffusion. ar Xiv, 2022. Can Qin, Shu Zhang, Ning Yu, Yihao Feng, Xinyi Yang, Yingbo Zhou, Huan Wang, Juan Carlos Niebles, Caiming Xiong, Silvio Savarese, et al. Unicontrol: A unified diffusion model for controllable visual generation in the wild. ar Xiv preprint ar Xiv:2305.11147, 2023. Julien Rabin, Gabriel Peyré, Julie Delon, and Marc Bernot. Wasserstein barycenter and its application to texture mixing. In Scale Space and Variational Methods in Computer Vision: Third International Conference, SSVM 2011, Ein-Gedi, Israel, May 29 June 2, 2011, Revised Selected Papers 3, pp. 435 446. Springer, 2012. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748 8763. Pm LR, 2021. Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with clip latents. ar Xiv preprint ar Xiv:2204.06125, 2022. Jeff Rasley, Samyam Rajbhandari, Olatunji Ruwase, and Yuxiong He. Deepspeed: System optimizations enable training deep learning models with over 100 billion parameters. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining, pp. 3505 3506, 2020. Stephen Robertson, Hugo Zaragoza, et al. The probabilistic relevance framework: Bm25 and beyond. Foundations and Trends in Information Retrieval, 3(4):333 389, 2009. Barbara Roessle, Norman Müller, Lorenzo Porzi, Samuel Rota Bulò, Peter Kontschieder, Angela Dai, and Matthias Nießner. L3dg: Latent 3d gaussian diffusion. In SIGGRAPH Asia 2024 Conference Papers, pp. 1 11, 2024. Published in Transactions on Machine Learning Research (05/2025) Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10684 10695, 2022. Amirmojtaba Sabour, Sanja Fidler, and Karsten Kreis. Align your steps: Optimizing sampling schedules in diffusion models. ar Xiv preprint ar Xiv:2404.14507, 2024. Tim Salimans and Jonathan Ho. Progressive distillation for fast sampling of diffusion models. In International Conference on Learning Representations, 2022. Tal Schuster, Adam Fisch, Tommi Jaakkola, and Regina Barzilay. Consistent accelerated inference via confident adaptive transformers. ar Xiv preprint ar Xiv:2104.08803, 2021. Pratheba Selvaraju, Tianyu Ding, Tianyi Chen, Ilya Zharkov, and Luming Liang. Fora: Fast-forward caching in diffusion transformer acceleration. ar Xiv preprint ar Xiv:2407.01425, 2024. Yuzhang Shang, Zhihang Yuan, Bin Xie, Bingzhe Wu, and Yan Yan. Post-training quantization on diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 1972 1981, 2023. Mingzhu Shen, Pengtao Chen, Peng Ye, Guoxuan Xia, Tao Chen, Christos-Savvas Bouganis, and Yiren Zhao. Md-dit: Step-aware mixture-of-depths for efficient diffusion transformers. In Adaptive Foundation Models: Evolving AI for Personalized and Efficient Learning. Shelly Sheynin, Oron Ashual, Adam Polyak, Uriel Singer, Oran Gafni, Eliya Nachmani, and Yaniv Taigman. knn-diffusion: Image generation via large-scale retrieval. In The Eleventh International Conference on Learning Representations, 2022. Andy Shih, Suneel Belkhale, Stefano Ermon, Dorsa Sadigh, and Nima Anari. Parallel sampling of diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Junhyuk So, Jungwon Lee, Daehyun Ahn, Hyungjun Kim, and Eunhyeok Park. Temporal dynamic quantization for diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. In International conference on machine learning, pp. 2256 2265. PMLR, 2015. Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. ar Xiv preprint ar Xiv:2010.02502, 2020a. Yang Song and Prafulla Dhariwal. Improved techniques for training consistency models. ar Xiv preprint ar Xiv:2310.14189, 2023. Yang Song and Stefano Ermon. Generative modeling by estimating gradients of the data distribution. Advances in neural information processing systems, 32, 2019. Yang Song, Sahaj Garg, Jiaxin Shi, and Stefano Ermon. Sliced score matching: A scalable approach to density and score estimation. In Uncertainty in Artificial Intelligence, pp. 574 584. PMLR, 2020b. Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. ar Xiv preprint ar Xiv:2011.13456, 2020c. Yang Song, Conor Durkan, Iain Murray, and Stefano Ermon. Maximum likelihood training of score-based diffusion models. Advances in neural information processing systems, 34:1415 1428, 2021. Yang Song, Prafulla Dhariwal, Mark Chen, and Ilya Sutskever. Consistency models. ar Xiv preprint ar Xiv:2303.01469, 2023. Published in Transactions on Machine Learning Research (05/2025) Xibo Sun, Jiarui Fang, Aoyu Li, and Jinzhe Pan. Unveiling redundancy in diffusion transformers (dits): A systematic study. ar Xiv preprint ar Xiv:2411.13588, 2024. Shengkun Tang, Yaqing Wang, Caiwen Ding, Yi Liang, Yao Li, and Dongkuan Xu. Deediff: Dynamic uncertainty-aware early exiting for accelerating diffusion model generation. 2023a. Zhicong Tang, Shuyang Gu, Chunyu Wang, Ting Zhang, Jianmin Bao, Dong Chen, and Baining Guo. Volumediffusion: Flexible text-to-3d generation with efficient volumetric encoder. ar Xiv preprint ar Xiv:2312.11459, 2023b. Zhicong Tang, Jianmin Bao, Dong Chen, and Baining Guo. Diffusion models without classifier-free guidance. ar Xiv preprint ar Xiv:2502.12154, 2025. Zhiwei Tang, Jiasheng Tang, Hao Luo, Fan Wang, and Tsung-Hui Chang. Accelerating parallel sampling of diffusion models. In Forty-first International Conference on Machine Learning, 2024. Ye Tian, Zhen Jia, Ziyue Luo, Yida Wang, and Chuan Wu. Diffusionpipe: Training large diffusion models with efficient pipelines, 2024. URL https://arxiv.org/abs/2405.01248. Anwaar Ulhaq, Naveed Akhtar, and Ganna Pogrebna. Efficient diffusion models for vision: A survey. ar Xiv preprint ar Xiv:2210.09292, 2022. Aaron Van Den Oord, Oriol Vinyals, et al. Neural discrete representation learning. Advances in neural information processing systems, 30, 2017. Clement Vignac, Igor Krawczuk, Antoine Siraudin, Bohan Wang, Volkan Cevher, and Pascal Frossard. Digress: Discrete denoising diffusion for graph generation. In The Eleventh International Conference on Learning Representations, 2022. Homer F Walker and Peng Ni. Anderson acceleration for fixed-point iterations. SIAM Journal on Numerical Analysis, 49(4):1715 1735, 2011. Jiannan Wang, Jiarui Fang, Aoyu Li, and Peng Cheng Yang. Pipefusion: Displaced patch pipeline parallelism for inference of diffusion transformer models. ar Xiv preprint ar Xiv:2405.14430, 2024a. Xiang Wang, Shiwei Zhang, Han Zhang, Yu Liu, Yingya Zhang, Changxin Gao, and Nong Sang. Videolcm: Video latent consistency model. ar Xiv preprint ar Xiv:2312.09109, 2023. Zhengyi Wang, Cheng Lu, Yikai Wang, Fan Bao, Chongxuan Li, Hang Su, and Jun Zhu. Prolificdreamer: High-fidelity and diverse text-to-3d generation with variational score distillation. Advances in Neural Information Processing Systems, 36, 2024b. Daniel Watson, Jonathan Ho, Mohammad Norouzi, and William Chan. Learning to efficiently sample from diffusion probabilistic models. ar Xiv preprint ar Xiv:2106.03802, 2021. Felix Wimbauer, Bichen Wu, Edgar Schoenfeld, Xiaoliang Dai, Ji Hou, Zijian He, Artsiom Sanakoyeu, Peizhao Zhang, Sam Tsai, Jonas Kohler, et al. Cache me if you can: Accelerating diffusion models through block caching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6211 6220, 2024. Zike Wu, Pan Zhou, Xuanyu Yi, Xiaoding Yuan, and Hanwang Zhang. Consistent3d: Towards consistent high-fidelity text-to-3d generation with deterministic sampling prior. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9892 9902, 2024. Yifei Xia, Suhan Ling, Fangcheng Fu, Yujie Wang, Huixia Li, Xuefeng Xiao, and Bin Cui. Training-free and adaptive sparse attention for efficient long video generation. ar Xiv preprint ar Xiv:2502.21079, 2025. Zhen Xing, Qijun Feng, Haoran Chen, Qi Dai, Han Hu, Hang Xu, Zuxuan Wu, and Yu-Gang Jiang. A survey on video diffusion models. ACM Computing Surveys, 2023. Published in Transactions on Machine Learning Research (05/2025) Zhen Xing, Qi Dai, Han Hu, Zuxuan Wu, and Yu-Gang Jiang. Simda: Simple diffusion adapter for efficient video generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7827 7839, 2024. Xingqian Xu, Zhangyang Wang, Gong Zhang, Kai Wang, and Humphrey Shi. Versatile diffusion: Text, images and variations all in one diffusion model. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 7754 7765, 2023a. Yilun Xu, Mingyang Deng, Xiang Cheng, Yonglong Tian, Ziming Liu, and Tommi Jaakkola. Restart sampling for improving generative processes. Advances in Neural Information Processing Systems, 36:76806 76838, 2023b. Kaiwen Xue, Yuhao Zhou, Shen Nie, Xu Min, Xiaolu Zhang, Jun Zhou, and Chongxuan Li. Unifying bayesian flow networks and diffusion models through stochastic differential equations. ar Xiv preprint ar Xiv:2404.15766, 2024a. Shuchen Xue, Mingyang Yi, Weijian Luo, Shifeng Zhang, Jiacheng Sun, Zhenguo Li, and Zhi-Ming Ma. Sa-solver: Stochastic adams solver for fast sampling of diffusion models. Advances in Neural Information Processing Systems, 36, 2024b. Hanshu Yan, Xingchao Liu, Jiachun Pan, Jun Hao Liew, Qiang Liu, and Jiashi Feng. Perflow: Piecewise rectified flow as universal plug-and-play accelerator. ar Xiv preprint ar Xiv:2405.07510, 2024. Geng Yang, Yanyue Xie, Zhong Jia Xue, Sung-En Chang, Yanyu Li, Peiyan Dong, Jie Lei, Weiying Xie, Yanzhi Wang, Xue Lin, et al. Sda: Low-bit stable diffusion acceleration on edge fpgas. 2023a. Ling Yang, Zhilong Zhang, Yang Song, Shenda Hong, Runsheng Xu, Yue Zhao, Wentao Zhang, Bin Cui, and Ming-Hsuan Yang. Diffusion models: A comprehensive survey of methods and applications. ACM Computing Surveys, 56(4):1 39, 2023b. Yang Yang, Wen Wang, Liang Peng, Chaotian Song, Yao Chen, Hengjia Li, Xiaolong Yang, Qinglin Lu, Deng Cai, Boxi Wu, et al. Lora-composer: Leveraging low-rank adaptation for multi-concept customization in training-free diffusion models. ar Xiv preprint ar Xiv:2403.11627, 2024. Jingfeng Yao, Bin Yang, and Xinggang Wang. Reconstruction vs. generation: Taming optimization dilemma in latent diffusion models. ar Xiv preprint ar Xiv:2501.01423, 2025. Hu Ye, Jun Zhang, Sibo Liu, Xiao Han, and Wei Yang. Ip-adapter: Text compatible image prompt adapter for text-to-image diffusion models. ar Xiv preprint ar Xiv:2308.06721, 2023. Shengming Yin, Chenfei Wu, Huan Yang, Jianfeng Wang, Xiaodong Wang, Minheng Ni, Zhengyuan Yang, Linjie Li, Shuguang Liu, Fan Yang, et al. Nuwa-xl: Diffusion over diffusion for extremely long video generation. ar Xiv preprint ar Xiv:2303.12346, 2023. Tianwei Yin, Michaël Gharbi, Richard Zhang, Eli Shechtman, Fredo Durand, William T Freeman, and Taesung Park. One-step diffusion with distribution matching distillation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 6613 6623, 2024. Yuning You, Ruida Zhou, Jiwoong Park, Haotian Xu, Chao Tian, Zhangyang Wang, and Yang Shen. Latent 3d graph diffusion. International Conference on Learning Representations (ICLR), 2024. Sihyun Yu, Kihyuk Sohn, Subin Kim, and Jinwoo Shin. Video probabilistic diffusion models in projected latent space. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 18456 18466, 2023. Sihyun Yu, Sangkyung Kwak, Huiwon Jang, Jongheon Jeong, Jonathan Huang, Jinwoo Shin, and Saining Xie. Representation alignment for generation: Training diffusion transformers is easier than you think. ar Xiv preprint ar Xiv:2410.06940, 2024a. Published in Transactions on Machine Learning Research (05/2025) Yonghao Yu, Shunan Zhu, Huai Qin, and Haorui Li. Boostdream: Efficient refining for high-quality text-to-3d generation from multi-view diffusion. ar Xiv preprint ar Xiv:2401.16764, 2024b. Hongyi Yuan, Zheng Yuan, Chuanqi Tan, Fei Huang, and Songfang Huang. Text diffusion model with encoder-decoder transformers for sequence-to-sequence generation. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pp. 22 39, 2024. Jinhui Yuan, Xinqi Li, Cheng Cheng, Juncheng Liu, Ran Guo, Shenghang Cai, Chi Yao, Fei Yang, Xiaodong Yi, Chuan Wu, et al. Oneflow: Redesign the distributed deep learning framework from scratch. ar Xiv preprint ar Xiv:2110.15032, 2021. Zongsheng Yue, Jianyi Wang, and Chen Change Loy. Resshift: Efficient diffusion model for image superresolution by residual shifting. Advances in Neural Information Processing Systems, 36, 2024. Mohsen Zand, Ali Etemad, and Michael Greenspan. Diffusion models with deterministic normalizing flow priors. ar Xiv preprint ar Xiv:2309.01274, 2023. Denis Zavadski, Johann-Friedrich Feiden, and Carsten Rother. Controlnet-xs: Rethinking the control of text-to-image diffusion models as feedback-control systems, 2023. Dingkun Zhang, Sijia Li, Chen Chen, Qingsong Xie, and Haonan Lu. Laptop-diff: Layer pruning and normalized distillation for compressing diffusion models. ar Xiv preprint ar Xiv:2404.11098, 2024a. Hengrui Zhang, Jiani Zhang, Balasubramaniam Srinivasan, Zhengyuan Shen, Xiao Qin, Christos Faloutsos, Huzefa Rangwala, and George Karypis. Mixed-type tabular data synthesis with score-based diffusion in latent space. ar Xiv preprint ar Xiv:2310.09656, 2023a. Hui Zhang, Zuxuan Wu, Zhen Xing, Jie Shao, and Yu-Gang Jiang. Adadiff: Adaptive step selection for fast diffusion. ar Xiv preprint ar Xiv:2311.14768, 2023b. Kexun Zhang, Xianjun Yang, William Yang Wang, and Lei Li. Redi: efficient learning-free diffusion inference via trajectory retrieval. In International Conference on Machine Learning, pp. 41770 41785. PMLR, 2023c. Lvmin Zhang, Anyi Rao, and Maneesh Agrawala. Adding conditional control to text-to-image diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3836 3847, 2023d. Qinsheng Zhang and Yongxin Chen. Diffusion normalizing flow. Advances in neural information processing systems, 34:16280 16291, 2021. Qinsheng Zhang and Yongxin Chen. Fast sampling of diffusion models with exponential integrator. ar Xiv preprint ar Xiv:2204.13902, 2022. Wentian Zhang, Haozhe Liu, Jinheng Xie, Francesco Faccio, Mike Zheng Shou, and Jürgen Schmidhuber. Cross-attention makes inference cumbersome in text-to-image diffusion models. ar Xiv e-prints, pp. ar Xiv 2404, 2024b. Shihao Zhao, Dongdong Chen, Yen-Chun Chen, Jianmin Bao, Shaozhe Hao, Lu Yuan, and Kwan-Yee K Wong. Uni-controlnet: All-in-one control to text-to-image diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Kaiwen Zheng, Cheng Lu, Jianfei Chen, and Jun Zhu. Improved techniques for maximum likelihood estimation for diffusion odes. In International Conference on Machine Learning, pp. 42363 42389. PMLR, 2023. Daquan Zhou, Weimin Wang, Hanshu Yan, Weiwei Lv, Yizhe Zhu, and Jiashi Feng. Magicvideo: Efficient video generation with latent diffusion models. ar Xiv preprint ar Xiv:2211.11018, 2022. Published in Transactions on Machine Learning Research (05/2025) Wenyang Zhou, Zhiyang Dou, Zeyu Cao, Zhouyingcheng Liao, Jingbo Wang, Wenjia Wang, Yuan Liu, Taku Komura, Wenping Wang, and Lingjie Liu. Emdm: Efficient motion diffusion model for fast, high-quality motion generation. Yuanzhi Zhu, Xingchao Liu, and Qiang Liu. Slimflow: Training smaller one-step diffusion models with rectified flow. ar Xiv preprint ar Xiv:2407.12718, 2024.