# onestep_diffusion_distillation_via_deep_equilibrium_models__029f4191.pdf One-Step Diffusion Distillation via Deep Equilibrium Models Zhengyang Geng Carnegie Mellon University zgeng2@cs.cmu.edu Ashwini Pokle Carnegie Mellon University apokle@cs.cmu.edu J. Zico Kolter Carnegie Mellon University Bosch Center for AI zkolter@cs.cmu.edu Diffusion models excel at producing high-quality samples but naively require hundreds of iterations, prompting multiple attempts to distill the generation process into a faster network. However, many existing approaches suffer from a variety of challenges: the process for distillation training can be complex, often requiring multiple training stages, and the resulting models perform poorly when utilized in single-step generative applications. In this paper, we introduce a simple yet effective means of distilling diffusion models directly from initial noise to the resulting image. Of particular importance to our approach is to leverage a new Deep Equilibrium (DEQ) model as the distilled architecture: the Generative Equilibrium Transformer (GET). Our method enables fully offline training with just noise/image pairs from the diffusion model while achieving superior performance compared to existing one-step methods on comparable training budgets. We demonstrate that the DEQ architecture is crucial to this capability, as GET matches a 5 larger Vi T in terms of FID scores while striking a critical balance of computational cost and image quality. Code, checkpoints, and datasets are available here. 1 Introduction Diffusion models [35, 93, 95, 96] have demonstrated remarkable performance on a wide range of generative tasks such as high-quality image generation [71, 82, 85, 87] and manipulation [19, 67, 71, 85, 86], audio synthesis [39, 40, 51, 58], video [36, 92], 3D shape [44, 78], text [30, 56], and molecule generation [18, 38, 107]. These models are trained with a denoising objective derived from score matching [42, 95, 96], variational inference [35, 49, 93], or optimal transport [57, 60], enabling them to generate clean data samples by progressively denoising the initial Gaussian noise during the inference process. Unlike adversarial training, the denoising objective leads to a more stable training procedure, which in turn allows diffusion models to scale up effectively [82, 85, 87]. Despite the promising results, one major drawback of diffusion models is their slow generative process, which often necessitates hundreds to thousands of model evaluations [22, 35, 96]. This computational complexity limits the applicability of diffusion models in real-time or resource-constrained scenarios. In an effort to speed up the slow generative process of diffusion models, researchers have proposed distillation methods [11, 68, 88, 97, 111] aimed at distilling the multi-step sampling process into a more efficient few-step or single-step process. However, these techniques often come with their Equal Contribution. Correspondence to Zhengyang Geng and Ashwini Pokle 37th Conference on Neural Information Processing Systems (Neur IPS 2023). own set of challenges. The distillation targets must be carefully designed to successfully transfer knowledge from the larger model to the smaller one. Further, distilling a long sampling process into a few-step process often calls for multiple training passes. Most of the prevalent techniques for online distillation require maintaining dual copies of the model, leading to increased memory and computing requirements. As a result, there is a clear need for simpler and more efficient approaches that address the computational demands of distilling diffusion models without sacrificing the generative capabilities. In this work, our objective is to streamline the distillation of diffusion models while retaining the perceptual quality of the images generated by the original model. To this end, we introduce a simple and effective technique that distills a multi-step diffusion process into a single-step generative model, using solely noise/image pairs. At the heart of our technique is the Generative Equilibrium Transformer (GET), a novel Deep Equilibrium (DEQ) model [5] inspired by the Vision Transformer (Vi T) [25, 75]. GET can be interpreted as an infinite depth network using weight-tied transformer layers, which solve for a fixed point in the forward pass. This architectural choice allows for the adaptive application of these layers in the forward pass, striking a balance between inference speed and sample quality. Furthermore, we incorporate an almost parameter-free class conditioning mechanism in the architecture, expanding its utility to class-conditional image generation. Our direct approach for distillation via noise/image pairs generated by a diffusion model, can, in fact, be applied to both Vi T and GET architectures. Yet, in our experiments, we show that the GET architecture, in particular, is able to achieve substantially better quality results with smaller models. Indeed, GET delivers perceptual image quality on par with or superior to other complex distillation techniques, such as progressive distillation [68, 88], in the context of both conditional and unconditional image generation. This leads us to explore the potential of GETs further. We preliminarily investigate the scaling law of GETs how its performance evolves as model complexity, in terms of parameters and computations, increases. Notably, GET exhibits significantly better parameter and data efficiency compared to architectures like Vi T, as GET matches the FID scores of a 5 larger Vi T, underscoring the transformative potential of GET in enhancing the efficiency of generative models. To summarize, we make the following key contributions: We propose Generative Equilibrium Transformer (GET), a deep equilibrium vision transformer that is well-suited for single-step generative models. We streamline the diffusion distillation by training GET directly on noise/image pairs sampled from diffusion models, which turns out to be a simple yet effective strategy for producing one-step generative models in both unconditional and class-conditional cases. For the first time, we show that implicit models for generative tasks can outperform classic networks in terms of performance, model size, model compute, training memory, and speed. 2 Preliminaries Deep Equilibrium Models. Deep equilibrium models [5] compute internal representations by solving for a fixed point in their forward pass. Specifically, consider a deep feedforward model with L layers: z[i+1] = f [i] θ (z[i]; x) for i = 0, ..., L 1 (1) where x Rnx is the input injection, z[i] Rnz is the hidden state of ith layer, and f [i] θ : Rnx nz 7 Rnz is the feature transformation of ith layer, parametrized by θ. If the above model is weight-tied, i.e., f [i] θ = fθ, i, then in the limit of infinite depth, the output z[i] of this network approaches a fixed point z : lim i fθ(z[i]; x) = fθ(z ; x) = z (2) Deep equilibrium (DEQ) models [5] directly solve for this fixed point z using black-box root finding algorithms like Broyden s method [14], or Anderson acceleration [1] in the forward pass. DEQs utilize implicit differentiation to differentiate through the fixed point analytically. Let gθ(z ; x) = fθ(z ; x) z , then the Jacobian of z with respect to the model weights θ is given by θ = gθ(z , x) 1 fθ(z ; x) Equilibrium Transformer Injection Transformer Figure 1: Generative Equilibrium Transformer (GET). (Left) GET consists of two major components: Injection transformer and Equilibrium transformer. The Injection transformer transforms noise embeddings into an input injection for the Equilibrium transformer. The Equilibrium transformer is the equilibrium layer that takes in noise input injection and an optional class embedding and solves for the fixed point. (Right) Details of transformer blocks in the Injection transformer (Inj) and Equilibrium transformer (DEQ), respectively. Blue dotted boxes denote optional class label inputs. Computing the inverse of Jacobian can quickly become intractable as we deal with high-dimensional feature maps. One can replace the inverse-Jacobian term with cheap approximations [27 29] without sacrificing the final performance. Diffusion Models. Diffusion models [22, 35, 93, 94] or score-based generative models [95, 96] progressively perturb images with an increasing amount of Gaussian noise and then reverse this process through sequential denoising to generate images. Specifically, consider a dataset of i.i.d. samples pdata, then the diffusion process {x(t)}T t=0 for t [0, T] is given by an Itô SDE [96]: dx = f(x, t)dt + g(t)dw (4) where w is the standard Wiener process, f( , t) : Rd Rd is the drifting coefficient, g( ) : R R is the diffusion coefficient, and x(0) pdata and x(T) N(0, I). All diffusion processes have a corresponding deterministic process known as the probability flow ODE (PF-ODE) [96] whose trajectories share the same marginal probability densities as the SDE. This ODE can be written as: dx = σ(t)σ(t) x log p(x, σ(t))dt (5) where σ(t) is the noise schedule of diffusion process, and x log p(x, σ(t)) represents the score function. Karras et al. [48] show that the optimal choice of σ(t) in Eq. (5) is σ(t) = t. Thus, the PF-ODE can be simplified to dx/dt = t x log p(x, σ(t)) = (x Dθ(x; t))/t, where Dθ( , t) is a denoiser function parametrized with a neural network that minimizes the expected L2 denoising error for samples drawn from pdata. Samples can be efficiently generated from this ODE through numerical methods like Euler s method, Runge-Kutta method, and Heun s second-order solver [3]. 3 Generative Equilibrium Transformer We introduce the Generative Equilibrium Transformer (GET), a Deep Equilibrium (DEQ) vision transformer designed to distill diffusion models into generative models that are capable of rapidly sampling images using only a single model evaluation. Our approach builds upon the key components and best practices of the classic transformer [99], the Vision transformer (Vi T) [25], and the Diffusion transformer (Di T) [75]. We will now describe each component of the GET in detail. GET. Generative Equilibrium Transformer (GET) directly maps Gaussian noises e and optional class labels y to images x. The major components of GET include an injection transformer (Injection T, Eq. (7)) and an equilibrium transformer (Equilibrium T, Eq. (8)). The Injection T transforms tokenized noise embedding h to an intermediate representation n that serves as the input injection for the equilibrium transformer. The Equilibrium T, which is the equilibrium layer, solves for the fixed point z by taking in the noise injection n and an optional class embedding c. Finally, this fixed point z is decoded and rearranged to generate an image sample x (Eq. (9)). Figure 1 provides an overview of the GET architecture. Note that because we are directly distilling the entire generative process, there is no need for a time embedding t as is common in standard diffusion models. h, c = Emb (e) , Emb (y) ; if y / (6) n = Injection T (h, c) (7) z = Equilibrium T (z , n, c) (8) x = Decoder (z ) (9) Noise Embedding. GET first converts an input noise e RH W C into a sequence of 2D patches p RN (P 2 C), where C is the number of channels, P is the size of patch, H and W denotes height and width of the original image, and N = HW/P 2 is the resulting number of patches. Let D = P 2 C denote the width of the network. We follow Vi T to use a linear layer to project the N patches to D dimensional embedding. We add standard sinusoidal position encoding [99] to produce the noise embedding h. Position encoding plays a crucial role in capturing the spatial structure of patches by encoding their relative positional information. Injection T & Equilibrium T. Both Injection T and Equilibrium T are composed of a sequence of Transformer blocks. Injection T is called only once to produce the noise injection n, while Equilibrium T defines the function fθ of the implicit layer z = fθ(z , n, c) that is called multiple times creating a weight-tied computational graph until convergence. A linear layer is added at the end of Injection T to compute the noise injection nl RN 3D, l [Le], for each of the Le GET blocks in Equilibrium T. For convenience, we overload the notation nl and n, in the subsequent paragraphs. Transformer Block. GET utilizes a near-identical block design for the noise injection (Injection T) and the equilibrium layer (Equilibrium T), differing only at the injection interface. Specifically, the transformer block is built upon the standard Pre-LN transformer block [25, 75, 106], as shown below: z = z + Attention (LN (z) , u) z = z + FFN (LN (z)) Here, z RN D represents the latent token, u RN 3D is the input injection, LN, FFN, and Attention stand for Layer Normalization [4], a 2-layer Feed-Forward Network with a hidden dimension of size D E, and an attention [99] layer with an injection interface, respectively. For blocks in the injection transformer, u is equal to the class embedding token c R1 3D for conditional image generation, i.e., u = c for conditional models, and u = 0 otherwise. In contrast, for blocks in the equilibrium transformer, u is the broadcast sum of noise injection n RN 3D and class embedding token c R1 3D, i.e., u = n + c for conditional models and u = n otherwise. We modify the standard transformer attention layer to incorporate an additive injection interface before the query q RN D, key k RN D, and value v RN D, q, k, v = z Wi + u z = MHA (q, k, v) z = z Wo where Wi RD 3D, Wo RD D. The injection interface enables interactions between the latent tokens and the input injection in the multi-head dot-product attention (MHA) operation, qk = (z Wq + uq)(z Wk + uk) = z Wq W k z + z Wqu k + uq W k z + u q uk, (10) where Wq, Wk RD D are slices from Wi, and uq, uk RN D are slices from u. This scheme adds no more computational cost compared to the standard MHA operation, yet it achieves a similar effect as cross-attention and offers good stability during training. 200 400 600 800 Iterations (1k) FID vs Iterations Progressive Distillation (PD) PD-EDM-4 BS Vi T-B-1 BS (85M) Vi T-L-1 BS (302M) GET-B-1 BS (62M) 80 160 240 Parameters (M) Parameter Scaling GET-200k GET-400k GET-800k Vi T-200k Vi T-400k Vi T-800k Figure 2: Data and Parameter Efficiency of GET:(a) (Left) GET outperforms PD and a 5 larger Vi T in fewer iterations, yielding better FID scores. Additionally, longer training times lead to improved FID scores. (b) (Right) Smaller GETs can achieve better FID scores than larger Vi Ts, demonstrating DEQ s parameter efficiency. Each curve in this plot connects models of different sizes within the same model family at identical training iterations, as indicated by the numbers after the model names in the legend. Image Decoder. The output of the GET-DEQ is first normalized with Layer Normalization [4]. The normalized output is then passed through another linear layer to generate patches p RN D. The resulting patches p are rearranged back to the resolution of the input noise e to produce the image sample x RH W C. Thus, the decoder maps the features back to the image space. 4 Experiments We evaluate the effectiveness of our proposed Generative Equilibrium Transformer (GET) in offline distillation of diffusion models through a series of experiments on single-step class-conditional and unconditional image generation. Here, we use single-step to refer to the use of a single model evaluation while generating samples. We train and evaluate Vi Ts and GETs of varying scales on these tasks. GETs exhibit substantial data and parameter efficiency in offline distillation compared to the strong Vi T baseline. Note that owing to the computational resources required to fully evaluate models, we report all our results on CIFAR-10 [52]; extensions to the Image Net-scale [20] are possible, but would require substantially larger GPU resources. 4.1 Experiment setup We will first outline our data collection process, followed by an in-depth discussion of our offline distillation procedure. We also include a brief summary of training details and evaluation metrics. For detailed network configs and training specifics, please refer to the Appendix. Data Collection. For unconditional image generation on CIFAR-10 [52], we generate 1M noise/image pairs from the pretrained unconditional EDM Karras et al. [48]. This dataset is denoted as EDM-Uncond-1M. As in EDM, we sample 1M images using Heun s second-order deterministic solver [3]. Generating a batch of images takes 18 steps or 35 NFEs (Number of Function Evaluations). Overall, this dataset takes up around 29 GB of disk space. The entire process of data generation takes about 4 hours on 4 NVIDIA A6000 GPUs using Pytorch [74] Distributed Data Parallel (DDP) and a batch size of 128 per GPU. In addition to unconditional image generation, we sample 1M noise-label/image pairs from the conditional VP-EDM Karras et al. [48] using the same settings. This dataset is denoted as EDM-Cond-1M. Both the datasets will be released for future studies. Offline Distillation. We distill a pretrained EDM [48] into Vi Ts and GETs by training on a dataset D with noise/image pairs sampled from the teacher diffusion model using a reconstruction loss: L(θ) = Ee,x D x Gθ(e) 1 where x is the desired ground truth image, Gθ( ) is unconditional Vi T/GET with parameters θ, and e is the initial Gaussian noise. To train a class-conditional GET, we also use class labels y in addition to noise/image pairs: L(θ) = Ee,y,x D x Gc θ(e, y) 1 0 20 40 60 Sampling Speed (sec/50k) FID vs Sampling Speed GET-Tiny GET-Mini GET-Small VIT-B VIT-L Training Compute (GFLOPs) FID vs GFLOPs Vi T-B Vi T-L GET-Tiny GET-Mini GET-Small GET-Base+ Figure 3: (a) (Left) Sampling speed of GET: GET can sample faster than large Vi Ts, while achieving better FID scores. The size of each individual circle is proportional to the model size. For GETs, we vary the number of iterations in the Equilibrium transformer (2 to 6 iterations). The trends indicate that GETs can improve their FID scores by using more compute. (b) (Right) Compute efficiency of GET: Larger GET models use training compute more efficiently. For a given GET, the training budget is calculated from training iterations. Refer to Table 2 for the exact size of GET models. Figure 4: Uncurated CIFAR-10 image samples generated by (Left) (a) unconditional GET and (Right) (b) class-conditional GET. Each row corresponds to a class in CIFAR-10. where Gc θ( ) is class-conditional Vi T/GET with parameters θ. As is the standard practice, we also maintain an exponential moving average (EMA) of weights of the model, which in turn is used at inference time for sampling. Training Details and Evaluation Metrics. We use Adam W [63] optimizer with a learning rate of 1e-4, a batch size of 128 (denoted as 1 BS), and 800k training iterations, which are identical to Progressive Distillation (PD) [88]. For conditional models, we adopt a batch size of 256 (2 BS). No warm-up, weight decay, or learning rate decay is applied. We convert input noise to patches of size 2 2. We use 6 steps of fixed point iterations in the forward pass of GET-DEQ and differentiate through it. For the O(1) memory mode, we utilize gradient checkpoint [17] for DEQ s computational graph. We measure image sample quality for all our experiments via Frechet inception distance (FID) [34] of 50k samples. We also report Inception Score (IS) [89] computed on 50k images. We include other relevant metrics such as FLOPs, training speed, memory, sampling speed, and the Number of Function Evaluations (NFEs), wherever necessary. 4.2 Experiment Results We aim to answer the following questions through extensive experiments: 1) Can offline distillation match online distillation for diffusion models using GETs? 2) What is the scaling behavior of GET as the model size and training compute increase? 3) How does GET compare to existing one-step generative models in terms of image quality and diversity? Efficiency. Models trained with offline distillation require high data efficiency to make optimal use of limited training data sampled from pretrained diffusion models. DEQs have a natural regularization mechanism due to weight-tying, which allows us to efficiently fit significantly compact data-efficient models even in limited data settings. In Figure 2(a), we observe that even with a fixed and limited offline data budget of 1M samples, GET achieves parity with online distilled EDM [48, 88, 97] while using only half the number of training iterations. For comparison, PD, TRACT, and CM use a much larger data budget of 96M, 256M, and 409.6M samples, respectively. Moreover, GET is able to match the FID score of a 5 large Vi T, suggesting substantial parameter efficiency. Sampling Speed. Figure 3(a) illustrates the sampling speed of both Vi T and GET. A smaller GET (37.2M) can achieve faster sampling than a larger Vi T (302.6M) while achieving lower FID scores. GET can also improve its FID score by increasing its test-time iterations in the Equilibrium transformer at the cost of speed. Note that despite this trade-off, GET still outperforms larger VIT in terms of both sampling speed and sample quality. Why Scaling Laws for Implicit Models? As a prospective study, we preliminarily investigate the scaling properties of Deep Equilibrium models using GET. The scaling law is an attractive property, as it enables us to predict models performance at extremely large compute based on the performance of tiny models. This predictive capability allows us to select the most efficient model given the constraints of available training budget [13, 37, 72]. While the scaling law for explicit networks has been extensively studied, its counterpart for implicit models remains largely unexplored. Implicit models are different from explicit models as they utilize more computation through weight-tying under similar parameters and model designs. Therefore, it is natural to question whether their scaling laws align with those of their explicit counterparts. Scaling Model Size. We conduct extensive experiments to understand the trends of sample quality as we scale the model size of GET. Table 2 provides a summary of our findings on single-step unconditional image generation. We find that even small GET models with 10-20M parameters can generate images with sample quality on par with NAS-derived Auto GAN [31]. In general, sample quality improves with the increase in model size. Scaling Training Compute. Our experimental results support the findings of Peebles and Xie [75] for explicit models (Di T) and extend them to implicit models. Specifically, for both implicit and explicit models, larger models are better at exploiting training FLOPs. Figure 3 shows that larger models eventually outperform smaller models when the training compute increases. For implicit models, there also exists a sweet spot in terms of model size under a fixed training budget, e.g., GET-Small outperforms both smaller and larger GETs at 231 training GFLOPs. Furthermore, because of the internal dynamics of implicit models, they can match a much larger explicit model in terms of performance while using fewer parameters. This underscores the potential of implicit models as candidates for compute-optimal models [37] with substantially better parameter efficiency. For example, at 231 training GFLOPs, Figure 3(b) suggests that we should choose GET-Small (31.2M) among implicit models for the best performance, which is much more parameter efficient and faster in sampling than the best-performing explicit model, Vi T-L (302M), at this training budget. Benchmarking GET against Vi T. Table 4 summarizes key metrics for unconditional image generation for Vi T and GET. Our experiments indicate that a smaller GET (19.2M) can generate higher-quality images faster than a much larger Vi T (85.2M) while utilizing less training memory and fewer FLOPs. GET also demonstrates substantial parameter efficiency over Vi Ts as shown in Figure 2(b) where smaller GETs achieve better FID scores than larger Vi Ts. Comparizon of NFEs of teacher model. Offline distillation requires significantly fewer number of function evaluations (NFEs) for the teacher network compared to other online distillation methods. In the experimental setup used in this paper, GET requires 35M overall NFEs for the teacher model, as we train on 1M data samples, and use 35 NFEs to generate each data sample with EDM. In contrast, progressive distillation requires 179M NFEs to get 1-step distilled student model. Using the hyperparameters reported in Salimans and Ho [88], PD with DDIM model needs 13 passes of Table 1: Generative performance on unconditional CIFAR-10. Method NFE FID IS Diffusion Models DDPM [35] 1000 3.17 9.46 Score SDE [96] 2000 2.2 9.89 DDIM [94] 10 13.36 - EDM [48] 35 2.04 9.84 Style GAN2 [47] 1 8.32 9.18 Style GAN-XL [91] 1 1.85 - Diffusion Distillation KD [66] 1 9.36 8.36 PD [88] 1 9.12 - DFNO [111] 1 4.12 - TRACT-EDM [11] 1 4.17 - PD-EDM [88, 97] 1 8.34 8.69 CD-EDM (LPIPS) [97] 1 3.55 9.48 Consistency Models CT [97] 1 8.70 8.49 CT [97] 2 5.83 8.85 GET-Base 1 6.91 9.16 Table 2: Generative performance of GETs on unconditional CIFAR-10. Models Params NFE FID IS GET-Tiny 8.6M 1 15.19 8.37 GET-Mini 19.2M 1 10.72 8.69 GET-Small 37.2M 1 8.00 9.03 GET-Base 62.2M 1 7.42 9.16 GET-Base+ 83.5M 1 7.19 9.09 More Training GET-Tiny-4 Iters 8.9M 1 11.47 8.64 GET-Base-2 BS 62.2M 1 6.91 9.16 Table 3: Generative performance on classconditional CIFAR-10. w indicates the level of classifier guidance. Method NFE FID IS Big GAN [12] 1 14.73 9.22 Style GAN2-ADA [46] 1 2.42 10.14 Diffusion Distillation Guided Distillation (w = 0) [68] 1 8.34 8.63 Guided Distillation (w = 0.3) [68] 1 7.34 8.90 Guided Distillation (w = 1) [68] 1 8.62 9.21 Guided Distillation (w = 2) [68] 1 13.23 9.23 GET-Base 1 6.25 9.40 Table 4: Benchmarking GET against Vi T on unconditional image generation on CIFAR-10. For the first time, implicit models for generative tasks strictly surpass explicit models in all metrics. Results are benchmarked on 4 A6000 GPUs using a batch size of 128, 800k iterations, and Py Torch [74] distributed training protocol. Training Mem stands for training memory consumed per GPU. O(1) symbolizes the O(1) training memory mode, which differs only in training memory and speed. Model FID IS Params FLOPs Training Mem Training Speed Vi T-Base 11.49 8.61 85.2M 23.0G 10.1GB 4.83 iter/sec GET-Mini 10.72 8.69 19.2M 15.2G 9.2GB 5.79 iter/sec GET-Mini-O(1) - - - - 5.0GB 4.53 iter/sec distillation. The initial 12 passes use 50K iterations, and the last pass uses 100K iterations. Each step of PD uses 2 teacher model NFEs. Thus, the overall number of teacher model NFEs can be evaluated as 2 128 (batch size) (12 passes 50K + 100K) = 179M samples. The number of NFEs of the teacher model increases to 1.433B if we assume that each of 8 TPUs use a batch size of 128. Consistency distillation [97] needs 409.6M teacher model NFEs (512 batch size 800K iterations = 409.6M). In addition, the perceptual loss requires double NFEs as the teacher model. One-Step Image Generation. We provide results for unconditional and class-conditional image generation on CIFAR-10 in Table 1 and Table 3, respectively. GET outperforms a much more complex distillation procedure PD with classifier-free guidance in class-conditional image generation. GET also outperforms PD and KD in terms of FID score for unconditional image generation. This effectiveness is intriguing, given that our approach for offline distillation is relatively simpler when compared to other state-of-the-art distillation techniques. We have outlined key differences in the experimental setup between our approach and other distillation techniques in Table 5. We also visualize random CIFAR-10 [52] samples generated by GET for both unconditional and class-conditional cases in Figure 4. GET can learn rich semantics and world knowledge from the dataset, as depicted in the images. For instance, GET has learned the symmetric layout of dog faces solely using reconstruction loss in the pixel space, as shown in Figure 4(b). Table 5: Comparison of relevant training and hyperparameter settings for common distillation techniques. GET requires neither multiple training phases nor any trajectory information. We only count the number of models involved in the forward pass and exclude EMA in #Models. indicates offline distillation techniques. For CD, we count the VGG network used in the perceptual loss [109]. Model FID IS BS Training Phases #Models Trajectory Teacher KD [66] 9.36 - 4 1 1 DDIM PD [88] 9.12 - 1 log2(T) 2 DDIM DFNO [111] 4.12 - 2 1 1 DDIM TRACT [11] 14.40 - 2 1 1 DDIM TRACT [11] 4.17 - 2 2 1 EDM PD-EDM [88, 97] 8.34 8.69 4 log2(T) 2 EDM CD [97] 3.55 9.48 4 1 3 EDM Ours 7.42 9.16 1 1 1 EDM Ours 6.91 9.16 2 1 1 EDM Guided Distillation [68] 7.34 8.90 4 log2(T) + 1 3 DDIM Ours 6.25 9.40 2 1 1 EDM 5 Related Work Distillation techniques for diffusion models. Knowledge distillation (KD) [66] proposed to distill a multi-step DDIM [94] sampler into the pretrained UNet by training the student model on synthetic image samples. There are several key differences from this work: Our approach does not rely on temporal embeddings or generative pretrained weights and predicts images instead of noises. Further, GET is built upon Vi T [25], unlike the UNet in KD. Additionally, we demonstrate the effectiveness of our approach on both unconditional and class-conditional image generation. Progressive distillation (PD) [88] proposes a strategy for online distillation to distill a T-step teacher DDIM [94] diffusion model into a new T/2 step student DDIM model, repeating this process until one-step models are achieved. Transitive closure time-distillation (TRACT) [11] generalizes PD to distill N > 2 steps together at once, reducing the overall number of training phases. Consistency models [97] achieve online distillation in a single pass by taking advantage of a carefully designed teacher and distillation loss objective. Diffusion Fourier neural operator (DFNO) [111] maps the initial Gaussian distribution to the solution trajectory of the reverse diffusion process by inserting the temporal Fourier integral operators in the pretrained U-Net backbone. Meng et al. [68] propose a two-stage approach to distill classifier-free guided diffusion models into few-step generative models by first distilling a combined conditional and unconditional model, and then progressively distilling the resulting model for faster generation. Fast sampler for diffusion models. While distillation is a predominant approach to speed up the sampling speed of existing diffusion models, there are alternate lines of work to reduce the length of sampling chains by considering alternate formulations of diffusion model [48, 50, 94, 96, 103], correcting bias and truncation errors in the denoising process [9, 10, 90], and through training-free fast samplers at inference [24, 43, 50, 59, 65, 108]. Several works like Improved DDPM [70], SGMCLD [23], EDM [48] modify or optimize the forward diffusion process so that the reverse denoising process can be made more efficient. Diffusion Exponential Integrator Sampler (DEIS) [108] uses an exponential integrator over the Euler method to minimize discretization error while solving SDE. DPM-Solver [65], and GENIE [24] are higher-order ODE solvers that generate samples in a few steps. 6 Limitations Our method for offline distillation relies on deterministic samplers to ensure a unique mapping between initial noise e and image x. As a result, it cannot be directly applied to stochastic samplers which do not satisfy this requirement. However, this limitation also applies to many other distillation techniques, as they cannot maintain their fidelity under stochastic trajectories [11, 66, 88, 97]. 7 Conclusion We propose a simple yet effective approach to distill diffusion models into generative models capable of sampling with just a single model evaluation. Our method involves training a Generative Equilibrium Transformer (GET) architecture directly on noise/image pairs generated from a pretrained diffusion model, eliminating the need for trajectory information and temporal embedding. GET demonstrates superior performance over more complex online distillation techniques such as progressive distillation [68, 88] in both class-conditional and unconditional settings. In addition, a small GET can generate higher quality images than a 5 larger Vi T, sampling faster while using less training memory and fewer compute FLOPs, demonstrating its effectiveness. 8 Acknowledgements Zhengyang Geng and Ashwini Pokle are supported by grants from the Bosch Center for Artificial Intelligence. [1] Donald G. Anderson. Iterative procedures for nonlinear integral equations. Journal of the ACM (JACM), 12(4):547 560, October 1965. 2 [2] Cem Anil, Ashwini Pokle, Kaiqu Liang, Johannes Treutlein, Yuhuai Wu, Shaojie Bai, J Zico Kolter, and Roger B Grosse. Path independent equilibrium models can better exploit test-time computation. Advances in Neural Information Processing Systems, 35:7796 7809, 2022. 18 [3] Uri M Ascher and Linda R Petzold. Computer methods for ordinary differential equations and differential-algebraic equations, volume 61. Siam, 1998. 3, 5 [4] Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. ar Xiv preprint ar Xiv:1607.06450, 2016. 4, 5 [5] Shaojie Bai, J. Zico Kolter, and Vladlen Koltun. Deep equilibrium models. In Neural Information Processing Systems (Neur IPS), 2019. 2, 18 [6] Shaojie Bai, Vladlen Koltun, and J Zico Kolter. Multiscale deep equilibrium models. Advances in Neural Information Processing Systems, 33:5238 5250, 2020. 18 [7] Shaojie Bai, Vladlen Koltun, and J. Zico Kolter. Stabilizing Equilibrium Models by Jacobian Regularization. In International Conference on Machine Learning (ICML), 2021. 18 [8] Shaojie Bai, Zhengyang Geng, Yash Savani, and J Zico Kolter. Deep equilibrium optical flow estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 620 630, 2022. 18 [9] Fan Bao, Chongxuan Li, Jiacheng Sun, Jun Zhu, and Bo Zhang. Estimating the optimal covariance with imperfect mean in diffusion probabilistic models. ar Xiv preprint ar Xiv:2206.07309, 2022. 9 [10] Fan Bao, Chongxuan Li, Jun Zhu, and Bo Zhang. Analytic-dpm: an analytic estimate of the optimal reverse variance in diffusion probabilistic models. ar Xiv preprint ar Xiv:2201.06503, 2022. 9 [11] David Berthelot, Arnaud Autef, Jierui Lin, Dian Ang Yap, Shuangfei Zhai, Siyuan Hu, Daniel Zheng, Walter Talbott, and Eric Gu. Tract: Denoising diffusion models with transitive closure time-distillation, 2023. 1, 8, 9 [12] Andrew Brock, Jeff Donahue, and Karen Simonyan. Large scale gan training for high fidelity natural image synthesis. ar Xiv preprint ar Xiv:1809.11096, 2018. 8, 18 [13] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877 1901, 2020. 7, 18 [14] Charles G Broyden. A Class of Methods for Solving Nonlinear Simultaneous Equations. Mathematics of computation, 19(92):577 593, 1965. 2 [15] Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging properties in self-supervised vision transformers. In IEEE International Conference on Computer Vision (ICCV), 2021. 17 [16] Lili Chen, Kevin Lu, Aravind Rajeswaran, Kimin Lee, Aditya Grover, Misha Laskin, Pieter Abbeel, Aravind Srinivas, and Igor Mordatch. Decision transformer: Reinforcement learning via sequence modeling. Advances in neural information processing systems, 34:15084 15097, 2021. 17 [17] Tianqi Chen, Bing Xu, Chiyuan Zhang, and Carlos Guestrin. Training deep nets with sublinear memory cost. ar Xiv preprint ar Xiv:1604.06174, 2016. 6 [18] Gabriele Corso, Hannes Stärk, Bowen Jing, Regina Barzilay, and Tommi Jaakkola. Diffdock: Diffusion steps, twists, and turns for molecular docking. ar Xiv preprint ar Xiv:2210.01776, 2022. 1 [19] Guillaume Couairon, Jakob Verbeek, Holger Schwenk, and Matthieu Cord. Diffedit: Diffusionbased semantic image editing with mask guidance. ar Xiv preprint ar Xiv:2210.11427, 2022. 1 [20] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A largescale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248 255. Ieee, 2009. 5 [21] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL), 2019. 17, 18 [22] Prafulla Dhariwal and Alexander Nichol. Diffusion models beat gans on image synthesis. Neural Information Processing Systems (Neur IPS), 2021. 1, 3, 18 [23] Tim Dockhorn, Arash Vahdat, and Karsten Kreis. Score-based generative modeling with critically-damped langevin diffusion. ar Xiv preprint ar Xiv:2112.07068, 2021. 9 [24] Tim Dockhorn, Arash Vahdat, and Karsten Kreis. GENIE: Higher-Order Denoising Diffusion Solvers. In Advances in Neural Information Processing Systems, 2022. 9 [25] Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations (ICLR), 2021. 2, 3, 4, 9, 17, 18 [26] Patrick Esser, Robin Rombach, and Bjorn Ommer. Taming transformers for high-resolution image synthesis. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 12873 12883, 2021. 18 [27] Samy Wu Fung, Howard Heaton, Qiuwei Li, Daniel Mc Kenzie, Stanley J. Osher, and Wotao Yin. Fixed Point Networks: Implicit Depth Models with Jacobian-Free Backprop. ar Xiv preprint ar Xiv:2103.12803, 2021. 3 [28] Zhengyang Geng, Meng-Hao Guo, Hongxu Chen, Xia Li, Ke Wei, and Zhouchen Lin. Is Attention Better Than Matrix Decomposition? In International Conference on Learning Representations (ICLR), 2021. [29] Zhengyang Geng, Xin-Yu Zhang, Shaojie Bai, Yisen Wang, and Zhouchen Lin. On training implicit models. In Neural Information Processing Systems (Neur IPS), 2021. 3 [30] Shansan Gong, Mukai Li, Jiangtao Feng, Zhiyong Wu, and Ling Peng Kong. Diffuseq: Sequence to sequence text generation with diffusion models. ar Xiv preprint ar Xiv:2210.08933, 2022. 1 [31] Xinyu Gong, Shiyu Chang, Yifan Jiang, and Zhangyang Wang. Autogan: Neural architecture search for generative adversarial networks. In IEEE International Conference on Computer Vision (ICCV), 2019. 7 [32] Priya Goyal, Piotr Dollár, Ross Girshick, Pieter Noordhuis, Lukasz Wesolowski, Aapo Kyrola, Andrew Tulloch, Yangqing Jia, and Kaiming He. Accurate, large minibatch sgd: Training imagenet in 1 hour. ar Xiv preprint ar Xiv:1706.02677, 2017. 18 [33] Swaminathan Gurumurthy, Shaojie Bai, Zachary Manchester, and J Zico Kolter. Joint inference and input optimization in equilibrium networks. Neural Information Processing Systems (Neur IPS), 2021. 18 [34] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017. 6 [35] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Neural Information Processing Systems (Neur IPS), 2020. 1, 3, 8 [36] Jonathan Ho, Tim Salimans, Alexey A Gritsenko, William Chan, Mohammad Norouzi, and David J Fleet. Video diffusion models. In Neural Information Processing Systems (Neur IPS), 2022. 1 [37] Jordan Hoffmann, Sebastian Borgeaud, Arthur Mensch, Elena Buchatskaya, Trevor Cai, Eliza Rutherford, Diego de Las Casas, Lisa Anne Hendricks, Johannes Welbl, Aidan Clark, et al. Training compute-optimal large language models. ar Xiv preprint ar Xiv:2203.15556, 2022. 7 [38] Emiel Hoogeboom, Vıctor Garcia Satorras, Clément Vignac, and Max Welling. Equivariant diffusion for molecule generation in 3d. In International Conference on Machine Learning, pages 8867 8887. PMLR, 2022. 1 [39] Qingqing Huang, Daniel S Park, Tao Wang, Timo I Denk, Andy Ly, Nanxin Chen, Zhengdong Zhang, Zhishuai Zhang, Jiahui Yu, Christian Frank, et al. Noise2music: Text-conditioned music generation with diffusion models. ar Xiv preprint ar Xiv:2302.03917, 2023. 1 [40] Rongjie Huang, Max WY Lam, Jun Wang, Dan Su, Dong Yu, Yi Ren, and Zhou Zhao. Fastdiff: A fast conditional diffusion model for high-quality speech synthesis. ar Xiv preprint ar Xiv:2204.09934, 2022. 1 [41] Drew A Hudson and Larry Zitnick. Generative adversarial transformers. In International Conference on Machine Learning (ICML), 2021. 18 [42] Aapo Hyvärinen and Peter Dayan. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005. 1 [43] Alexia Jolicoeur-Martineau, Ke Li, Rémi Piché-Taillefer, Tal Kachman, and Ioannis Mitliagkas. Gotta go fast when generating data with score-based models. ar Xiv preprint ar Xiv:2105.14080, 2021. 9 [44] Heewoo Jun and Alex Nichol. Shap-e: Generating conditional 3d implicit functions, 2023. 1 [45] Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4401 4410, 2019. 18 [46] Tero Karras, Miika Aittala, Janne Hellsten, Samuli Laine, Jaakko Lehtinen, and Timo Aila. Training generative adversarial networks with limited data. Advances in neural information processing systems, 33:12104 12114, 2020. 8 [47] Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 8110 8119, 2020. 8 [48] Tero Karras, Miika Aittala, Timo Aila, and Samuli Laine. Elucidating the design space of diffusion-based generative models. In Neural Information Processing Systems (Neur IPS), 2022. 3, 5, 7, 8, 9 [49] Diederik Kingma, Tim Salimans, Ben Poole, and Jonathan Ho. Variational diffusion models. Neural Information Processing Systems (Neur IPS), 2021. 1 [50] Zhifeng Kong and Wei Ping. On fast sampling of diffusion probabilistic models. ar Xiv preprint ar Xiv:2106.00132, 2021. 9 [51] Zhifeng Kong, Wei Ping, Jiaji Huang, Kexin Zhao, and Bryan Catanzaro. Diffwave: A versatile diffusion model for audio synthesis. ar Xiv preprint ar Xiv:2009.09761, 2020. 1 [52] Alex Krizhevsky. Learning multiple layers of features from tiny images. 2009. 5, 8 [53] Guillaume Lample and Alexis Conneau. Cross-lingual language model pretraining. ar Xiv preprint ar Xiv:1901.07291, 2019. 18 [54] Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. Albert: A lite bert for self-supervised learning of language representations. ar Xiv preprint ar Xiv:1909.11942, 2019. 18 [55] Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Ves Stoyanov, and Luke Zettlemoyer. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. ar Xiv preprint ar Xiv:1910.13461, 2019. 17 [56] Xiang Li, John Thickstun, Ishaan Gulrajani, Percy S Liang, and Tatsunori B Hashimoto. Diffusion-lm improves controllable text generation. Advances in Neural Information Processing Systems, 35:4328 4343, 2022. 1 [57] Yaron Lipman, Ricky T. Q. Chen, Heli Ben-Hamu, Maximilian Nickel, and Matthew Le. Flow matching for generative modeling. In International Conference on Learning Representations (ICLR), 2023. 1 [58] Jinglin Liu, Chengxi Li, Yi Ren, Feiyang Chen, and Zhou Zhao. Diffsinger: Singing voice synthesis via shallow diffusion mechanism. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 11020 11028, 2022. 1 [59] Luping Liu, Yi Ren, Zhijie Lin, and Zhou Zhao. Pseudo numerical methods for diffusion models on manifolds. ar Xiv preprint ar Xiv:2202.09778, 2022. 9 [60] Xingchao Liu, Chengyue Gong, and Qiang Liu. Flow straight and fast: Learning to generate and transfer data with rectified flow. In International Conference on Learning Representations (ICLR), 2023. 1 [61] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized bert pretraining approach. ar Xiv preprint ar Xiv:1907.11692, 2019. 18 [62] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In IEEE International Conference on Computer Vision (ICCV), 2021. 17 [63] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. 6 [64] Cheng Lu, Jianfei Chen, Chongxuan Li, Qiuhao Wang, and Jun Zhu. Implicit normalizing flows. ar Xiv preprint ar Xiv:2103.09527, 2021. 18 [65] Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. Dpm-solver: A fast ode solver for diffusion probabilistic model sampling in around 10 steps. ar Xiv preprint ar Xiv:2206.00927, 2022. 9 [66] Eric Luhman and Troy Luhman. Knowledge distillation in iterative generative models for improved sampling speed. ar Xiv preprint ar Xiv:2101.02388, 2021. 8, 9 [67] Chenlin Meng, Yang Song, Jiaming Song, Jiajun Wu, Jun-Yan Zhu, and Stefano Ermon. Sdedit: Image synthesis and editing with stochastic differential equations. ar Xiv preprint ar Xiv:2108.01073, 2021. 1 [68] Chenlin Meng, Ruiqi Gao, Diederik P Kingma, Stefano Ermon, Jonathan Ho, and Tim Salimans. On distillation of guided diffusion models. ar Xiv preprint ar Xiv:2210.03142, 2022. 1, 2, 8, 9, 10 [69] Paul Micaelli, Arash Vahdat, Hongxu Yin, Jan Kautz, and Pavlo Molchanov. Recurrence without recurrence: Stable video landmark detection with deep equilibrium models. ar Xiv preprint ar Xiv:2304.00600, 2023. 18 [70] Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. In International Conference on Machine Learning (ICML), 2021. 9 [71] Alexander Quinn Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob Mcgrew, Ilya Sutskever, and Mark Chen. Glide: Towards photorealistic image generation and editing with text-guided diffusion models. In International Conference on Machine Learning (ICML), 2022. 1 [72] Open AI. Gpt-4 technical report. Ar Xiv, abs/2303.08774, 2023. 7 [73] Emilio Parisotto, Francis Song, Jack Rae, Razvan Pascanu, Caglar Gulcehre, Siddhant Jayakumar, Max Jaderberg, Raphael Lopez Kaufman, Aidan Clark, Seb Noury, et al. Stabilizing transformers for reinforcement learning. In International Conference on Machine Learning (ICML), 2020. 17 [74] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Py Torch: An Imperative Style, High-performance Deep Learning Library. In Neural Information Processing Systems (Neur IPS), 2019. 5, 8 [75] William Peebles and Saining Xie. Scalable diffusion models with transformers. ar Xiv preprint ar Xiv:2212.09748, 2022. 2, 3, 4, 7, 17, 18 [76] Ethan Perez, Florian Strub, Harm De Vries, Vincent Dumoulin, and Aaron Courville. Film: Visual reasoning with a general conditioning layer. In Association for the Advancement of Artificial Intelligence (AAAI), 2018. 18 [77] Ashwini Pokle, Zhengyang Geng, and J Zico Kolter. Deep equilibrium approaches to diffusion models. Advances in Neural Information Processing Systems, 35:37975 37990, 2022. 18 [78] Ben Poole, Ajay Jain, Jonathan T. Barron, and Ben Mildenhall. Dreamfusion: Text-to-3d using 2d diffusion. ar Xiv, 2022. 1 [79] Alec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever, et al. Improving language understanding by generative pre-training. 2018. 17, 18 [80] Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. Language models are unsupervised multitask learners. Open AI blog, 1(8):9, 2019. 18 [81] Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485 5551, 2020. 18 [82] Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with clip latents. ar Xiv preprint ar Xiv:2204.06125, 2022. 1, 18 [83] Max Revay, Ruigang Wang, and Ian R Manchester. Lipschitz bounded equilibrium networks. ar Xiv preprint ar Xiv:2010.01732, 2020. 18 [84] Adam Roberts, Colin Raffel, Katherine Lee, Michael Matena, Noam Shazeer, Peter J Liu, Sharan Narang, Wei Li, and Yanqi Zhou. Exploring the limits of transfer learning with a unified text-to-text transformer. 2019. 17 [85] Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10684 10695, 2022. 1, 18 [86] Chitwan Saharia, William Chan, Huiwen Chang, Chris Lee, Jonathan Ho, Tim Salimans, David Fleet, and Mohammad Norouzi. Palette: Image-to-image diffusion models. In ACM SIGGRAPH 2022 Conference Proceedings, pages 1 10, 2022. 1 [87] Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily L Denton, Kamyar Ghasemipour, Raphael Gontijo Lopes, Burcu Karagol Ayan, Tim Salimans, et al. Photorealistic text-to-image diffusion models with deep language understanding. Neural Information Processing Systems (Neur IPS), 2022. 1 [88] Tim Salimans and Jonathan Ho. Progressive distillation for fast sampling of diffusion models. ar Xiv preprint ar Xiv:2202.00512, 2022. 1, 2, 6, 7, 8, 9, 10 [89] Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, and Xi Chen. Improved techniques for training gans. Advances in neural information processing systems, 29, 2016. 6 [90] Robin San-Roman, Eliya Nachmani, and Lior Wolf. Noise estimation for generative diffusion models. ar Xiv preprint ar Xiv:2104.02600, 2021. 9 [91] Axel Sauer, Katja Schwarz, and Andreas Geiger. Stylegan-xl: Scaling stylegan to large diverse datasets. In ACM SIGGRAPH 2022 conference proceedings, pages 1 10, 2022. 8 [92] Uriel Singer, Adam Polyak, Thomas Hayes, Xi Yin, Jie An, Songyang Zhang, Qiyuan Hu, Harry Yang, Oron Ashual, Oran Gafni, et al. Make-a-video: Text-to-video generation without text-video data. ar Xiv preprint ar Xiv:2209.14792, 2022. 1 [93] Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning (ICML), 2015. 1, 3 [94] Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. In International Conference on Learning Representations (ICLR), 2021. 3, 8, 9 [95] Yang Song and Stefano Ermon. Generative modeling by estimating gradients of the data distribution. Neural Information Processing Systems (Neur IPS), 2019. 1, 3 [96] Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations (ICLR), 2021. 1, 3, 8, 9 [97] Yang Song, Prafulla Dhariwal, Mark Chen, and Ilya Sutskever. Consistency models. ar Xiv preprint ar Xiv:2303.01469, 2023. 1, 7, 8, 9 [98] Russell Tsuchida and Cheng Soon Ong. Deep equilibrium models as estimators for continuous latent variables. In International Conference on Artificial Intelligence and Statistics, pages 1646 1671. PMLR, 2023. 18 [99] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. Neural Information Processing Systems (Neur IPS), 2017. 3, 4, 17, 18 [100] Shuai Wang, Yao Teng, and Limin Wang. Deep equilibrium object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6296 6306, 2023. 18 [101] Thomas Wang, Adam Roberts, Daniel Hesslow, Teven Le Scao, Hyung Won Chung, Iz Beltagy, Julien Launay, and Colin Raffel. What language model architecture and pretraining objective works best for zero-shot generalization? In International Conference on Machine Learning, pages 22964 22984. PMLR, 2022. 18 [102] Tiancai Wang, Xiangyu Zhang, and Jian Sun. Implicit Feature Pyramid Network for Object Detection. ar Xiv preprint ar Xiv:2012.13563, 2020. 18 [103] Daniel Watson, Jonathan Ho, Mohammad Norouzi, and William Chan. Learning to efficiently sample from diffusion probabilistic models. ar Xiv preprint ar Xiv:2106.03802, 2021. 9 [104] Jason Wei, Maarten Bosma, Vincent Y Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M Dai, and Quoc V Le. Finetuned language models are zero-shot learners. ar Xiv preprint ar Xiv:2109.01652, 2021. 18 [105] Ezra Winston and J Zico Kolter. Monotone operator equilibrium networks. Advances in neural information processing systems, 33:10718 10728, 2020. 18 [106] Ruibin Xiong, Yunchang Yang, Di He, Kai Zheng, Shuxin Zheng, Chen Xing, Huishuai Zhang, Yanyan Lan, Liwei Wang, and Tieyan Liu. On layer normalization in the transformer architecture. In International Conference on Machine Learning (ICML), 2020. 4 [107] Minkai Xu, Lantao Yu, Yang Song, Chence Shi, Stefano Ermon, and Jian Tang. Geodiff: A geometric diffusion model for molecular conformation generation. ar Xiv preprint ar Xiv:2203.02923, 2022. 1 [108] Qinsheng Zhang and Yongxin Chen. Fast sampling of diffusion models with exponential integrator. In The Eleventh International Conference on Learning Representations, 2023. 9 [109] Richard Zhang, Phillip Isola, Alexei A Efros, Eli Shechtman, and Oliver Wang. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 586 595, 2018. 9 [110] Sai Zhang, Liangjia Zhu, and Yi Gao. An efficient deep equilibrium model for medical image segmentation. Computers in Biology and Medicine, 148:105831, 2022. 18 [111] Hongkai Zheng, Weili Nie, Arash Vahdat, Kamyar Azizzadenesheli, and Anima Anandkumar. Fast sampling of diffusion models via operator learning. ar Xiv preprint ar Xiv:2211.13449, 2022. 1, 8, 9 A Additional Experiments Class Conditioning. As both GET and Vi T share the same class injection interface, we perform an ablation study on Vi T. We consider two types of input injection schemes for class labels: 1) additive injection scheme 2) injection with adaptive layer normalization (Ada LN-Zero) as used in Di T [75]. Despite using almost the same parameters as unconditional Vi T-B, the class-conditional Vi T-B using additive injection interface has an FID of 12.43 at 200k, while the Vi T-B w/ Ada LN-Zero class embedding [75] set up an FID of 17.19 at 200k iterations. Another surprising observation is that Vi T-B w/ Ada LN-Zero class embedding performs worse than unconditional Vi T in terms of FID score. Therefore, it seems that adaptive layer normalization might not be useful when used only with class embedding. Table 6: Ablation on class conditioning. Model FID IS Params Vi T-Uncond 15.20 8.27 85.2M Vi T-Ada LN-Zero 17.19 8.38 128.9M Vi T-Inj-Interface 12.43 8.69 85.2M B Model Configuration We set the EMA momentum to 0.9999 for all the models. The configuration of different GET architectures are listed in Table 7. Here, Li and Le denote the number of transformer blocks in the Injection transformer and Equilibrium transformer, respectively. D denotes the width of the network. E corresponds to the expanding factor of the FFN layer in the Equilibrium transformer, which results in the hidden dimension of E D. For the injection transformer, we always adopt an expanding factor of 4. Table 7: Details of configuration for GET architectures. Model Params Li Le D E GET-Tiny 8.9M 6 3 256 6 GET-Mini 19.2M 6 3 384 6 GET-Small 37.2M 6 3 512 6 GET-Base 62.2M 1 3 768 12 GET-Base+ 83.5M 6 3 768 8 We have listed relevant model configuration details of Vi T in Table 8. The model configurations are adopted from Di T [75], whose effectiveness was tested for learning diffusion models. In this table, L denotes the number of transformer blocks in Vi T. D stands for the width of the network. We always adopt an expanding factor of 4 following the common practice [25, 75, 99]. Table 8: Details of configuration for Vi T architectures. Model Params L D Vi T-B 85.2M 12 768 Vi T-L 302.6M 24 1024 C Related Work Transformers. Transformers were first proposed by Vaswani et al. [99] for machine translation and since then have been widely applied in many domains like natural language processing [21, 55, 79, 84], reinforcement learning [16, 73], self-supervised learning [15], vision [25, 62], and generative modeling [26, 41, 75, 82]. Many design paradigms for transformer architectures have emerged over the years. Notable ones include encoder-only [21, 54, 61], decoder-only [13, 79, 80, 101, 104], and encoder-decoder architectures [53, 81, 99]. We are interested in scalable transformer architectures for generative modeling. Most relevant to this work are two encoder-only transformer architectures: Vision Transformer (Vi T) [25] and Diffusion Transformer (Di T) [75]. Vision Transformer (Vi T) closely follows the original transformer architecture. It first converts 2D images into patches that are flattened and projected into an embedding space. 2D Positional encoding is added to the patch embedding to retain positional information. This sequence of embedding vectors is fed into the standard transformer architecture. Diffusion Transformers (Di T) are based on Vi T architecture and operate on sequences of patches of an image that are projected into a latent space through an image encoder [85]. In addition, Di Ts adapt several architectural modifications that enable their use as a backbone for diffusion models and help them scale better with increasing model size, including adaptive Layer Normalization (Ada LN-Zero) [12, 22, 45, 76] for time and class embedding, and zero-initialization for the final convolution layer [32]. Deep equilibrium models. Deep Equilibrium models (DEQs) [5] solve for a fixed point in the forward pass. Specifically, given the input x and the equilibrium function fθ, DEQ models approach the infinite-depth representation of fθ by solving for its fixed point z : z = fθ(z , x). For the backward pass, one can differentiate analytically through z by the implicit function theorem. The training dynamics of DEQ models can be unstable for certain model designs [7]. As a result, recent efforts focus on addressing these issues by designing variants of DEQs with provable guarantees [83, 105], or through optimization techniques such as Jacobian regularization [7], and fixed-point correction [8]. DEQs have been successfully applied on a wide range of tasks such as image classification [6], semantic segmentation [6, 110], optical flow estimation [8], object detection [100, 102], landmark detection [69], out-of-distribution generalization [2], language modelling [5], input optimization [33], unsupervised learning [98], and generative models [64, 77].