# reward_guided_latent_consistency_distillation__4e43c615.pdf Published in Transactions on Machine Learning Research (10/2024) Reward Guided Latent Consistency Distillation Jiachen Li jiachen_li@cs.ucsb.edu University of California, Santa Barbara Weixi Feng weixifeng@cs.ucsb.edu University of California, Santa Barbara Wenhu Chen wenhu.chen@uwaterloo.ca University of Waterloo William Yang Wang william@cs.ucsb.edu University of California, Santa Barbara Reviewed on Open Review: https: // openreview. net/ forum? id= z116TO4LDT Latent Consistency Distillation (LCD) has emerged as a promising paradigm for efficient text-to-image synthesis. By distilling a latent consistency model (LCM) from a pre-trained teacher latent diffusion model (LDM), LCD facilitates the generation of high-fidelity images within merely 2 to 4 inference steps. However, the LCM s efficient inference is obtained at the cost of the sample quality. In this paper, we propose compensating the quality loss by aligning LCM s output with human preference during training. Specifically, we introduce Reward Guided LCD (RG-LCD), which integrates feedback from a reward model (RM) into the LCD process by augmenting the original LCD loss with the objective of maximizing the reward associated with LCM s single-step generation. As validated through human evaluation, when trained with the feedback of a good RM, the 2-step generations from our RG-LCM are favored by humans over the 50-step DDIM (Song et al., 2020a) samples from the teacher LDM, representing a 25-time inference acceleration without quality loss. As directly optimizing towards differentiable RMs can suffer from over-optimization, we take the initial step to overcome this difficulty by proposing the use of a latent proxy RM (LRM). This novel component serves as an intermediary, connecting our LCM with the RM. Empirically, we demonstrate that incorporating the LRM into our RG-LCD successfully avoids high-frequency noise in the generated images, contributing to both improved Fréchet Inception Distance (FID) on MS-COCO (Lin et al., 2014) and a higher HPSv2.1 score on HPSv2 (Wu et al., 2023a) s test set, surpassing those achieved by the baseline LCM. Project Page: https://rg-lcd.github.io/ 1 Introduction In the realm of modern generative AI (Gen AI) models, computational resources are typically allocated across three key areas: pretraining (Brown et al., 2020; Achiam et al., 2023a; Li et al., 2022b; Radford et al., 2021; Rombach et al., 2022; Saharia et al., 2022; Betker et al., 2023a), alignment (Ziegler et al., 2019; Stiennon et al., 2020; Ouyang et al., 2022; Hu et al., 2024; Clark et al., 2023; Rafailov et al., 2024), and inference (Zhang & Chen, 2022; Feng et al., 2023; Vijayakumar et al., 2016; Shih et al., 2024). Normally, increasing the computational budget across these areas leads to improvements in sample quality. For instance, the most advanced text-to-image (T2I) models, such as DALLE-3 (Betker et al., 2023a), Imagen (Saharia et al., 2022), and Stable Diffusion (Rombach et al., 2022) are built from diffusion models (DMs) (Sohl-Dickstein et al., 2015; Ho et al., 2020; Song & Ermon, 2019). These models are pretrained on massive web-scale Published in Transactions on Machine Learning Research (10/2024) Figure 1: Even with merely 2-4 sampling steps, our RG-LCMs that learned from the CLIP Score and HPSv2.1 can produce high-quality images. Published in Transactions on Machine Learning Research (10/2024) datasets (Schuhmann et al., 2022; Changpinyo et al., 2021), aligned with human preference on curated highquality images (Dai et al., 2023; Rombach et al., 2022), and benefit from DMs iterative sampling process. However, DM s iterative sampling requires performing 10 - 2000 sequential function evaluations (FEs) (Ho et al., 2020; Song et al., 2020a), thus impeding rapid inference. While there have been many works proposed to address this issue (Lu et al., 2022a;b; Zhang et al., 2023; Sauer et al., 2023; Geng et al., 2024; Nguyen & Tran, 2023; Song et al., 2023), consistency model (CM) (Song et al., 2023) emerges as a new family of Gen AI model to facilitate fast sampling. Specifically, a CM is trained to perform single-step generation while supporting multi-step sampling to trade compute for sample quality. We can distill a CM from a pretrained DM, a process known as consistency distillation (CD). For instance, Luo et al. (2023a) distill a Latent CM (LCM) from a pretrained Stable Diffusion (Rombach et al., 2022), achieving high-fidelity image generation in just 2 to 4 FE steps. However, the sample quality of LCM is inherently constrained by the pretrained LDM s capabilities (Song & Dhariwal, 2023). Additionally, the reduced inference computational resources stemming from the limited number of FE steps compromise LCM s sample quality. In this paper, we aim to offset LCM s sample quality by dedicating additional computational resources to the training process. Recent advancements in large language models (Achiam et al., 2023b; Team et al., 2023) have shown that aligning a Gen AI model with a reward model (RM) that mirrors human preferences can substantially improve sample quality by reducing undesirable outputs (Ouyang et al., 2022; Rafailov et al., 2024). Thus, we are motivated to align the learned LCM with human preferences by optimizing towards off-the-shelf text-image RMs. Instead of designing a separate alignment phase, we leverage the single-step generation that naturally arises from computing the LCD loss and implement a training objective to maximize its associated rewards given by a differentiable RM through gradient descent. Notably, our approach obviates the need for backpropagating gradients through the complicated denoising procedures, which is typically required by previous methods when optimizing a DM (Clark et al., 2023; Xu et al., 2024; Prabhudesai et al., 2023; Yang et al., 2024). We dub our method Reward Guided Latent Consistency Distillation (RG-LCD). Human evaluation shows that our RG-LCM significantly outperforms the LCM derived from standard LCD. Remarkably, our 2-step generation is favored by humans over the 50-step generation from the teacher LDM, representing a 25-fold inference acceleration without compromising image quality. While our RG-LCD is conceptually simple and already achieves impressive results, it can suffer from reward overestimation (Kim et al., 2023b; Zhang et al., 2024) due to direct optimization with the gradient from the RM. As shown in the top row of Fig. 3, performing RG-LCD with Image Reward (Xu et al., 2024) causes high-frequency noise in the generated images. In this paper, we take an initial step to tackle this challenge. We propose learning a latent proxy RM to serve as the intermediary that connects our LCM with the RM. Instead of directly optimizing towards the RM, we optimize the LCM towards the LRM while finetuning the LRM to match the preference of the expert RM in each RG-LCD iteration. This novel strategy allows us to optimize the expert RM indirectly, even allowing for learning from non-differentiable RMs. We empirically verify that incorporating the LRM into our RG-LCD successfully eliminates the high-frequency noise in the generated image, contributing to improved FID on MS-COCO (Lin et al., 2014) and a higher HPSv2.1 score on HPSv2 s test set (Wu et al., 2023a), outperforming the baseline LCM. In summary, our contributions are threefold: Introduction of RG-LCD framework, which incorporates feedback from an RM that mirrors human preference into the LCD process. Introduction of the LRM, which enables indirect optimization towards the RM, mitigating the issue of reward over-optimization. A 25 times inference acceleration over teacher LDM (Stable Diffusion v2.1) without compromising sample quality. 2 Related Work Accelerating DM inference. Centering around DM s SDE formulation (Song et al., 2020b), various methods have been proposed to accelerate the sampling process of a DM. For example, faster numerical Published in Transactions on Machine Learning Research (10/2024) ODE solvers (Song et al., 2020a; Lu et al., 2022a;b; Zheng et al., 2022; Dockhorn et al., 2022; Jolicoeur Martineau et al., 2021) and distillation techniques (Luhman & Luhman, 2021; Salimans & Ho, 2022; Meng et al., 2023; Zheng et al., 2023). Recent advances explore enhancing the single-step generation quality by incorporating an adversarial loss (Sauer et al., 2023) or by distillation (Nguyen & Tran, 2023). Consistency Model (Song et al., 2023) is also trained for single-step generation. We leverage this property and directly maximize the reward of this single-step generation given by a differentiable RM, avoiding the complexities of backpropagating gradients through the iterative sampling process of a DM. Consistency Model has emerged as a new family of Gen AI model (Song et al., 2023) that facilitates fast inference. While it is trained to perform single-step generation by mapping arbitrary points in the PF-ODE trajectory to the origin, CM also supports multi-step sampling, allowing for trading compute for better sample quality. On the one hand, a CM can be trained as a standalone Gen AI model (consistency training). Recently, Song & Dhariwal (2023) proposed improved techniques to support better consistency training. On the other hand, a CM can also be distilled from a pretrained DM (Kim et al., 2023a). For instance, Luo et al. (2023a) learn an LCM by distilling from a pretrained Stable Diffusion (Rombach et al., 2022). We defer more technical details to Sec. 3. Vision-and-language reward models. Motivated by the significant success of reinforcement from human feedback (RLHF) in training the LLMs, there have been many works delving into training an RM to mirror human preferences on a pair of text and image, including HPSv1 (Wu et al., 2023b), HPSv2 (Wu et al., 2023a), Image Reward (Xu et al., 2024), and Pick Score (Kirstain et al., 2024). These RMs are normally derived by finetuning a vision-and-language foundation model, e.g., CLIP (Radford et al., 2021) and BLIP (Li et al., 2022a), on human preference data. Since these RMs are differentiable, our RG-LCD augments the standard LCD with the objective of maximizing the differentiable reward associated with its single-step generation during training. Aligning DMs to Human preference has been extensively studied recently, including RL based methods (Fan et al., 2024; Prabhudesai et al., 2023; Zhang et al., 2024) and reward finetuning methods (Clark et al., 2023; Xu et al., 2024). Recently, Diffusion-DPO (Wallace et al., 2023a) is proposed by extending DPO (Rafailov et al., 2024) to train DMs on preference data. Moreover, other works focus on modifying the training data distribution (Wu et al., 2023b; Lee et al., 2023; Dong et al., 2023; Sun et al., 2023; Dai et al., 2023) to finetune DMs on visually appealing and textually cohered data. Additionally, alternative techniques (Betker et al., 2023b; Segalis et al., 2023) re-caption pre-collected web images to enhance text accuracy. On the other hand, DOODL (Wallace et al., 2023b) is proposed to optimize the RM during inference time. However, its improvement is made at the cost of inference speed. While we also propose to directly finetune our model with the gradient given by an RM, finetuning an LCM during LCD is much simpler than finetuning a DM, as we only tackle the single-step generation, circumventing the need to pass gradients through the complicated iterative sampling process of a DM. 3 Background 3.1 Diffusion Model Diffusion models (DMs) (Sohl-Dickstein et al., 2015; Song & Ermon, 2019; Ho et al., 2020; Nichol & Dhariwal, 2021) progressively inject Gaussian noise into data in the forward process and sequentially denoise the data to create samples in the reverse denoising process. The forward process perturbs the original data distribution pdata(x) p0(x0) to the marginal distributional pt(xt). From a continuous-time perspective, we can represent the forward process with a stochastic differential equation (SDE) (Song et al., 2020b; Karras et al., 2022) dxt = µ(t)xtdt + σ(t)dwt, x0 pdata (x0) , (1) where µ( ) and σ( ) are the drift and diffusion coefficients respectively, and wt denotes the standard Wiener process. The reverse time SDE above corresponds to an ordinary differential equation (ODE) (Song et al., 2020b), named Probability Flow (PF-ODE), which is given by dxt = µ (t) xt 1 2σ(t)2 log pt (xt) dt, x T p T (x T ). (2) Published in Transactions on Machine Learning Research (10/2024) PF-ODE s solution trajectories sampled at t are distributed the same as pt(xt). Empirically, we learn a denoising model ϵθ(xt, t) to fit log pt(xt) (score function) via score matching (Hyvärinen & Dayan, 2005; Song & Ermon, 2019; Ho et al., 2020). During sampling, we start from the sample x T N(0, σ2I) and follow the empirical PF-ODE below dxt = µ (t) xt + 1 2σ(t)2ϵθ(xt, t) dt, x T N(0, σ2I). (3) In this paper, we focus on conditional LDM that operates on the image latent space Z and includes a text prompt c passed to the denoising model ϵθ(zt, c, t), where zt = E(xt) Z is encoded by a VAE (Kingma et al., 2021) encoder E. Moreover, we utilize Classifier-Free Guidance (CFG) (Ho & Salimans, 2022) to improve the quality of conditional sampling by replacing the noise prediction with a linear combination of conditional and unconditional noise prediction for denoising, i.e., ϵθ (zt, ω, c, t) = (1 + ω)ϵθ (zt, c, t) ωϵθ(z, , t), where ω is the CFG scale. 3.2 Consistency Model Consistency model (CM) (Song et al., 2023) is proposed to facilitate efficient generation. At its core, CM learns a consistency function f : (xt, t) 7 xϵ that can map any point xt on the same PF-ODE trajectory to the trajectory s origin, where ϵ is a fixed small positive number. Learning the consistency function involves enforcing the self-consistency property f(xt, t) = f(x t, t ), t, t [ϵ, T], (4) where xt and x t belong to the same PF-ODE. The consistency function f is modeled with a CM fθ. To ensure fθ(x, ϵ) = x, fθ is parameteried as fθ(x, t) = cskip(t)x + cout(t)Fθ(x, t), (5) where cskip(t) and cout(t) are differentiable functions with cskip(ϵ) = 1 and cout(ϵ) = 0, and Fθ is a neural network. We can learn a CM fθ by distilling from a pretrained DM, known as consistency distillation (CD) (Song et al., 2023). The CD loss is given by LCD θ, θ ; Φ = Ex,t h d fθ xtn+1, tn+1 , fθ ˆxϕ tn, tn i . (6) where d( , ) measures the distance between two samples. θ is the parameter of the target CM fθ , updated by the exponential moving average (EMA) of θ, i.e., θ stop_grad (µθ + (1 µ)θ ). ˆxϕ tn is an estimation of xtn from xtn+1 using the one-step ODE solver Φ: ˆxϕ tn xtn+1 + (tn tn+1) Φ xtn+1, tn+1; ϕ . (7) Note that the parameter ϕ corresponds to the parameter of the pretrained DM, which is used to construct the ODE solver Φ. We includes the algorithm for sampling from a learned CM in Appendix B. 3.3 Latent Consistency Model Luo et al. (Luo et al., 2023a) extends CM to work on latent space Z and focuses on conditional generation. Specifically, a Latent CM (LCM) fθ : (zt, ω, c, t) 7 z0 is trained to minimized the LCD loss LLCD θ, θ ; Ψ = Ez,c,ω,n h d fθ ztn+k, ω, c, tn+k , fθ ˆzΨ,ω tn , ω, c, tn i , (8) where ˆzΨ,ω tn is an estimate of ztn obtained by the numerical augmented PF-ODE solver Ψ and k is skipping interval. ˆzΨ,ω tn ztn+k + (1 + ω)Ψ(ztn+k, tn+k, tn, c; ψ) ωΨ(ztn+k, tn+k, tn, ; ψ). (9) In this paper, we use DDIM (Song et al., 2020a) as the ODE solver Ψ to distill from a pretrained Stable Diffusion (Rombach et al., 2022) and refer interested readers to the original LCM paper for formula of the DDIM solver. We use huber loss as our distance function d( , ). Published in Transactions on Machine Learning Research (10/2024) Forward diffusion Distillation Loss Photo of on a mountain Gradient from the RM Gradient from the LCD Figure 2: Overview of our RG-LCD. We integrate feedback from a differentiable RM into the standard LCD procedures by training the LCM to maximize the reward associated with its single-step generation.. 4 Reward Guided Latent Consistency Distillation In this section, we start by presenting the core components of our RG-LCD framework, which augments the standard LCD loss equation 8 with an objective towards maximizing a differentiable RM, as shown in Fig. 2 (Sec. 4.1). We then motivate the development of a latent proxy RM (LRM) to support indirect RM optimization by illustrating the risk of suffering from reward over-optimization when directly optimizing towards the RM with a gradient-based method. Following this, we then detail the procedure to pretrain and finetune the LRM to match the preference of the RGB-based RM during RG-LCD (Sec. 4.2). 4.1 RG-LCD with Differentiable RMs Recall that each LCD iteration samples a timestep tn+k, and construct the noisy latent ztn+k by perturb the image latent z = E(x) with a Gaussian noise, given a sampled CFG scale ω and text prompt c. As the LCM fθ maps the ztn+k to the PF-ODE origin ˆz0 = fθ ztn+k, ω, c, tn+k , we construct the following objective to maximize the reward associated with D( ˆz0) J(θ) = Ez,c,ω,n R D fθ ztn+k, ω, c, tn+k , c , (10) where R is a differentiable RM that calculates the rewards associated with a pair of text and image. We define the training loss of our RG-LCD by a linear combination of the LCD loss in equation 8 and J(θ) with a weighting parameter β LRG-LCD θ, θ ; Ψ = LLCD θ, θ ; Ψ βJ(θ) (11) Appendix B includes pseudo-codes for our RG-LCD training. 4.2 RG-LCD with a Latent Proxy RM When training the LCM fθ towards J(θ) with a gradient-based method, we may suffer from the issue of reward over-optimization. As shown in the top row of Fig. 3, performing RG-LCD with Image Reward (Xu et al., 2024) causes high-frequency noise in the generated images. To mitigate this issue, we propose learning a latent proxy RM RL σ to serve as an intermediary to connect fθ and the expert RGB-based RM RE, where the E stands for Expert . Specifically, we train fθ to optimize the reward given by RL σ while simultaneously finetuning the RL σ to matches the preference given by the expert RM RE that process RGB images. Ideally, the LRM RL σ should be capable of accessing the text-image pair even at the beginning of RG-LCD. We thus initialize RL σ with a pretrained CLIP (Radford et al., 2021) text encoder, complemented by pretraining its latent encoder from scratch. This latent encoder is pretrained following the same methodology used for CLIP visual encoders, ensuring it aligns effectively with the text encoder s representation. After pretraining, we finetune RL σ to match the preference of RE. Note that we do not need to assume a differentiable RE anymore, allowing us to learn from the feedback from a wider range of RGB-based Published in Transactions on Machine Learning Research (10/2024) Figure 3: (Top) Optimizing the RG-LCM with the gradient from Image Reward (Xu et al., 2024) results in high-frequency noise in the generated images. (Bottom) Indirectly optimizing the Image Reward through the latent proxy RM eliminates the high-frequency noise, avoiding reward over-optimization. RM, e.g., LLMScore (Lu et al., 2024), VIEScore (Ku et al., 2023) and DA-score (Singh & Zheng, 2024). Next, we will derive the finetuning loss for our LRM(σ). Given z0 = z, z1 = fθ ztn+k, ω, c, tn+k , and z2 = fθ (ztn, ω, c, tn) in each RG-LCD iteration, we can group them into three pairs: (z0, z1), (z0, z2) and (z1, z2). We then use RL σ and RE to compute the rewards for each latent. For each latent pair (zi, zj), the probability of RL σ preferring zi over zj is modeled as: P σ i,j(i) = exp RL σ (zi, c) /τL exp (RLσ (zi, c) /τL) + exp (RLσ (zj, c) /τL) τL is the temperature parameter. Similarly, with the temperature τE, the probability of RE preferring zi over zj can be modeled as: Qi,j(i) = exp RE (D (zi) , c) /τE exp (RE (D (zi) , c) /τE) + exp (RE (D (zj) , c) /τE) And thus, we have P σ i,j(m) exp RL σ (zm, c) /τL , Qi,j(m) exp RE (D (zm) , c) /τE , m {i, j}. We can construct the KL divergence between the distribution P σ i,j and Qi,j(i) for each (zi, zj) pair. Our LRM(σ) is derived by summing the KL divergence for all three latent pairs as below LRM(σ) = Ez,c,ω,n j=i+1 DKL P σ i,j||stop_grad (Qi,j) Appendix B.3 includes pseudo-codes for training our RG-LCM with an LRM in Algorithm 4. LRM(σ) also supports matching a RE that only output preference over two images. In this case, we can set τE to a small Published in Transactions on Machine Learning Research (10/2024) 0 20 40 60 80 100 RG-LCM (CLIP) RG-LCM (CLIP) RG-LCM (CLIP) Parti Prompts: General Preference 0 20 40 60 80 100 Parti Prompts: Visual Appeal 0 20 40 60 80 100 Parti Prompts: Prompt Alignment 0 20 40 60 80 100 Preference [%] RG-LCM (HPS) RG-LCM (HPS) RG-LCM (HPS) 0 20 40 60 80 100 Preference [%] 0 20 40 60 80 100 Preference [%] LCM 4 steps SD 50 steps SD 50 steps LCM 4 steps SD 50 steps SD 50 steps Figure 4: Human evaluation results on the Parti Prompt (1632 prompts) across three evaluation questions. Top row evaluates the RG-LCM (CLIP). Bottom row evaluates the RG-LCM (HPS). positive number and only give a non-zero positive reward to the sample favored by the expert. Moreover, since z0 = z corresponds to the latent of a real image, we can increase likelihood for Q0,j to prefer k = 0. While calculating RE(D(z), c) still requires decoding the latent, the application of the stop_grad operation eliminates the need for gradient transmission through D, leading to a substantial reduction in memory usage. Moreover, optimizing RL σ with LRM(σ) is independent from optimizing fθ with LRG-LCD. Therefore, we can use a smaller batch size to optimize RL σ without affecting the batch size used to optimize fθ. In essence, our LRM acts as a proxy connecting the LCM fθ and the expert RM RE. As we will show Sec. 5.2, using this indirect feedback from the expert mitigates the issue of reward over-optimization, avoiding high-frequency noise in the generated images. 5 Experiment We perform thorough experiments to demonstrate the effectiveness of our RG-LCD. Sec. 5.1 conducts human evaluation to compare the performance of our methods with baselines. Sec. 5.2 further increases the experiment scales to experiment with a wider array of RMs with automatic metrics. By connecting both evaluation results, we identify problems with the current RMs. Finally, Sec. 5.3 conducts ablation studies on critical design choices. Settings Our training are conducted on the CC12M datasets (Changpinyo et al., 2021), as the LAIONAesthetics datasets (Schuhmann et al., 2022) used by the original LCM (Luo et al., 2023a) are no longer accessible1.We distill our LCM from the Stable Diffusion-v2.1 (Rombach et al., 2022) by training for 10K iterations on 8 NVIDIA A100 GPUs without gradient accumulation and set the batch size to reach the maximum capacity of our GPUs. We follow the hyperparameter settings listed in the diffusers (von Platen et al., 2022) library by setting learning rate 1e 6, EMA rate µ = 0.95 and the guidance scale range [ωmin, ωmax] = [5, 15]. As mentioned in Sec. 3.3, we use DDIM (Song et al., 2020a) as our ODE solver Ψ with a skipping step k = 20. We include more training details in Appendix A. Appendix D further includes experiment results with diverse teacher T2I models, including Stable Diffusion 1.5 and Stable Diffusion XL. 5.1 Evaluating RG-LCD with Human We train RG-LCM (HPS) and RG-LCM (CLIP) utilizing feedback from HPSv2.1 (Wu et al., 2023a) and CLIPScore (Radford et al., 2021), respectively. CLIPScore evaluates the relevance between text and images, whereas HPSv2.1, derived by fine-tuning CLIPScore with human preference data, is expected to mirror 1https://laion.ai/notes/laion-maintanence Published in Transactions on Machine Learning Research (10/2024) 0 20 40 60 80 100 RG-LCM (CLIP) RG-LCM (CLIP) RG-LCM (CLIP) HPSv2 Test: General Preference 0 20 40 60 80 100 HPSv2 Test: Visual Appeal 0 20 40 60 80 100 HPSv2 Test: Prompt Alignment 0 20 40 60 80 100 Preference [%] RG-LCM (HPS) RG-LCM (HPS) RG-LCM (HPS) 0 20 40 60 80 100 Preference [%] 0 20 40 60 80 100 Preference [%] LCM 4 steps SD 50 steps SD 50 steps LCM 4 steps SD 50 steps SD 50 steps Figure 5: Human evaluation results on the HPSv2 test set (3200 prompts) across three evaluation questions. Top row evaluates the RG-LCM (CLIP). Bottom row evaluates the RG-LCM (HPS). human preferences more accurately. We choose the teacher LDM (Stable Diffusion v2.1) and a standard LCM distilled from the same teacher LDM as the baseline methods. To demonstrate the efficacy of our methods, we compare the performance of our RG-LCMs over 2-step and 4-step generations against the 50step generations from the teacher LDM and evaluate the 4-step generation quality of our RG-LCMs against the standard LCM. We follow a similar evaluation protocol as in (Wallace et al., 2023a) to generate images by conditioning on prompts from Partiprompt (Yu et al., 2022) (1632 prompts) and of HPSv2 s test set (Wu et al., 2023a) (3200 prompts). We hire labelers from Amazon Mechanical Turk for a head-to-head comparison of images based on three criteria: Q1 General Preference (Which image do you prefer given the prompt?), Q2 Visual Appeal (Which image is more visually appealing, irrespective of the prompt?), and Q3 Prompt Alignment (Which image better matches the text description?). The full human evaluation results in Fig. 4 and 5 show that the 2-step generations from RG-LCM (CLIP) are generally preferred (Q1) over the 50-step generations of the teacher LDM in both prompt sets, representing a 25-fold acceleration in inference speed. Even with CLIPScore feedback, the 4-step generations from our RG-LCM are generally preferred (Q1) over the baseline methods. This indicates a noteworthy achievement, given that CLIPScore does not train on human preference data. Surprisingly, on the HPSv2 prompt set, the 4-step generations from the RG-LCM (CLIP) are more preferred (59.4% against 50-step DDIM samples from SD and 81.7% against 4-step LCM samples) compared to the 4-step generations of the RG-LCM (CLIP) (57.1% against 50-step DDIM samples from SD, and 69.0% against 4-step LCM samples). To investigate this phenomenon, we observe that both RG-LCMs score similarly in General Preference (Q1) and Prompt Alignment (Q3). However, the RG-LCM (CLIP) is rated slightly lower in Visual Appeal (Q2) than in the other criteria, whereas the RG-LCM (HPS) is rated significantly higher for Q2 compared to Q1 and Q3. This distinction highlights that CLIPScore s primary contribution is enhancing text-image alignment, whereas an RM like HPSv2.1 particularly focuses on improving visual quality. Thus, when overoptimizing towards HPSv2.1, the RG-LCM (HPS) can be biased in generating visually appealing samples by sacrificing prompt alignment. 5.2 Evaluating RG-LCD with Automatic Metrics In this section, we further train RG-LCD (Img Rwd) and RG-LCD (Pick) by leveraging feedback from Image Reward (Xu et al., 2024) and Pick Score (Kirstain et al., 2024). Both of these RMs are trained on human preference data. We will use automatic metrics to perform a large-scale evaluation of the performance of different models. As we have human evaluation results for RG-LCD (HPS) and RG-LCD (CLIP), we can Published in Transactions on Machine Learning Research (10/2024) Models NFEs Human Preference Score v2.1 FID-30K Anime Photo Concept-Art Paintings MS-COCO LCM 4 22.40 19.17 18.86 20.55 19.05 Stable Diffusion v2.1 50 25.66 24.37 24.58 25.72 12.66 RG-LCM (CLIP) 2 26.32 25.01 25.27 26.71 18.06 RG-LCM (CLIP) 4 27.80 26.92 27.04 28.11 19.22 RG-LCM (Pick) 2 26.44 28.26 28.24 29.04 22.84 RG-LCM (Pick) 4 27.33 29.42 29.29 30.26 22.02 RG-LCM (Pick) + LRM 2 23.82 21.31 21.90 22.99 15.91 RG-LCM (Pick) + LRM 4 25.17 23.06 22.90 24.87 16.27 RG-LCM (Img Rwd) 2 29.65 31.03 31.15 32.00 32.12 RG-LCM (Img Rwd) 4 30.26 31.83 31.88 32.73 42.69 RG-LCM (Img Rwd) + LRM 2 25.64 25.61 25.82 25.75 17.57 RG-LCM (Img Rwd) + LRM 4 26.84 26.72 26.72 27.30 17.20 RG-LCM (HPS) 2 30.85 33.66 33.35 33.66 24.04 RG-LCM (HPS) 4 31.83 34.84 34.43 34.75 25.11 RG-LCM (HPS) + LRM 2 27.58 25.94 26.77 27.24 16.71 RG-LCM (HPS) + LRM 4 28.53 27.49 27.94 28.87 17.52 Table 1: Evaluation of our RG-LCMs on the HPSv2 test prompts and MS-COCO datasets. NFEs denote the number of function evaluations during inference. We train RG-LCMs with CLIPScore, Pick Score, Image Reward (Img Rwd) and HPSv2.1. We employ the HPSv2.1 to evaluate the generations on the HPSv2 Benchmark s test set. We calculate the FID of the generations on the MS-COCO. Except trained with CLIPScore, our RG-LCMs achieve better HPSv2.1 scores on HPSv2 test prompts at the expense of higher FIDs on MS-COCO. Integrating a LRM into our RG-LCD process allows for simultaneous improvement on HPSv2.1 scores on HPSV2 test prompts and FID on MS-COCO against the baseline LCM. also evaluate the quality of the automatic metrics. For each RG-LCD, we collect their 2-step and 4-step generations by conditioning on prompts from HPSv2 s test set and measuring the HPSv2.1 score associated with the samples. To comprehensively understand the sample quality from different models, we further generate images conditioned on the prompts of MS-COCO (Lin et al., 2014) and measure their Fréchet Inception Distance (FID) to the ground truth images. Table 1 presents the full evaluation results with the automatic metrics. Except for RG-LCM (CLIP), all the other RG-LCMs achieve higher HPSv2.1 scores than the baseline LCM but at the expense of higher FID values on the MS-COCO dataset. Specifically, the RG-LCM (Img Rwd) model exhibits a notably high FID value, yet it still secures an impressive HPSv2.1 score when evaluated on HPSv2 test prompts. The elevated FID value aligns with expectations, as Figure 3 illustrates that optimization directed towards Image Reward tends to introduce a significant amount of high-frequency noise into the generated images. Surprisingly, these high-frequency noises do not adversely affect the HPSv2.1 scores. Furthermore, the HPSv2.1 scores do not capture the human preference for the 4-step samples from RG-LCM (CLIP) by giving the highest score to RG-LCM (HPS) s 4-step samples, contrary to what is depicted in human evaluation shown in Fig. 5. These observations suggest that the HPSv2.1 score, as a metric, has limitations and requires further refinement. We conjecture that the Resize operation, which happens during the preprocessing phase, causes the HPSv2.1 model to overlook the high-frequency noise during reward calculation. As illustrated in Fig. 3, the high-frequency noise becomes less perceptible when images are reduced in size. Although resizing operations enhance efficiency in tasks such as image classification (Lu & Weng, 2007; Deng et al., 2009; He et al., 2016) and facilitate high-level text-image understanding (Radford et al., 2021), they prevent the model from capturing critical visual nuances that are vital for accurately reflecting human preferences. Consequently, we advocate for future RMs to exclude the Resize operation. One potential approach could involve training an LRM, as in our paper, to learn human preferences in the latent space without resizing input images. Published in Transactions on Machine Learning Research (10/2024) Figure 6: Ablating the reward scale β for different reward functions. All samples are generated with 4 steps. We observe that over-optimizing RMs trained with preference data prioritizes visual appeal over text alignment, whereas over-optimizing CLIPScore compromises visual attractiveness in favor of text alignment. Connecting Table 1 with the human evaluation results in Fig. 5 suggests that images that achieve a high HPSv2.1 score and a low FID on MS-COCO are more aligned with human preferences. Moreover, this desirable outcome can be accomplished by integrating an LRM into our RG-LCD. Although these correlations do not imply causality, they underscore the potential benefits of utilizing an LRM in the RG-LCD process. As depicted in the bottom row in Fig. 3, the images generated by RG-LCM (Img Rwd) that integrates an LRM do not suffer from high-frequency noise, contributing to their improved FID on MS-COCO. In Appendix C, we include additional samples for each RG-LCM in Table 1. 5.3 Ablation Study Ablation on the reward scale β. We use the hyperparameter β to determine the optimization strength towards the RM. We are especially interested in the impact of an extremely large β, which can lead to reward over-optimization (Kim et al., 2023b). We already know that over-optimizing the Image Reward can lead to the introduction of high-frequency noise in the generated images. To expand our understanding, we conduct experiments a wider array of RMs including HPSv2.1, Pick Score and CLIPScore and evaluate whether over-optimizing these RMs will also leads to similar high-frequency noise. The results in Figure 6 reveal that an extremely large β value does not introduce the high-frequency noise when using HPSv2.1, Pick Score, and CLIPScore, even though all these metrics resize input images to 224x224 pixels as in Image Reward. Notably, over-optimization of HPSv2.1 leads to generating images with repetitive objects as described in the text prompts and increases color saturation. Conversely, over-optimization of Pick Score tends to result in images with more muted colors. On the other hand, excessive optimization of CLIPScore results in images where the text prompts are visibly incorporated into the imagery. These findings align with the discussions in Sec. 5.1, suggesting that optimizing towards a preference-trained RM generally prioritizes visual appeal over text alignment. In contrast, over-optimizing CLIPScore compromises visual attractiveness in favor of text alignment. We include additional image samples in Appendix C. Ablation on the training iterations. In total, we train each model for 10K iterations. We take checkpoints from 1K, 2K, 4K, and 10K iterations and sample images with the same prompts and seeds. We can observe performing RG-LCD with RMs that facilitate the visual appeal of the generated images also results Published in Transactions on Machine Learning Research (10/2024) Figure 7: Ablation study on the number of training iterations. We generate all samples with 4 steps. We observe that RG-LCM, which learns from an RM that prioritizes visual appeal, can generate high-quality images with fewer training iterations. in fast training, as the checkpoint at the 2K iterations can already produce high-quality images. In contrast, the images generated by RG-LCM (CLIP) still generate blurry images after training for 2K iterations. 6 Conclusion In this paper, we introduce RG-LCD, a novel strategy that integrates feedback from an RM into the LCD process. The RG-LCM learned via our method enjoys better sample quality while facilitating fast inference, benefiting from additional computational resources allocated to align with human preferences. By evaluating using prompts from the HPSv2 (Wu et al., 2023a) test set and Parti Prompt (Yu et al., 2022), we empirically show that humans favor the 2-step generations of our RG-LCD (HPS) over the 50-step DDIM generations of the teacher LDM. This represents a 25-fold increase in terms of inference speed without a loss in quality. Moreover, even when using CLIPScore a model not fine-tuned on human preferences our method s 4-step generations still surpass the 50-step DDIM generations from the teacher LDM. We also identify that directly optimizing towards an imperfect RM, e.g., Image Reward, can cause highfrequency noise in generated images. To reconcile the issue, we propose integrating an LRM into the RGLCD framework. Notably, our methods not only prevents reward over-optimization but also avoids passing gradients through the VAE decoder and facilitates learning from non-differentiable RMs. 7 Limitation and Impact Statement While our RG-LCD marks a critical advancement in the realm of efficient text-to-image synthesis, introducing an acceleration in the generation process without compromising on image quality, it is important to recognize certain limitations. The approach relies on employing a reward model that reflects human preference, which, while effective in improving image quality metrics, may introduce additional costs in the training pipeline and necessitate fine-tuning to adapt to various domains or datasets. Despite these challenges, the impact of RG-LCD is profound, offering a scalable solution that significantly enhances the accessibility and practicality of generating high-fidelity images at a remarkable speed. This innovation not only broadens the potential applications in fields ranging from digital art to visual content creation but also sets a new benchmark for future research in text-to-image synthesis, emphasizing the importance of human-centric design in the development of generative AI technologies. Published in Transactions on Machine Learning Research (10/2024) Acknowledgement The work was funded by an unrestricted gift from Google. We would like to thank Google for their generous sponsorship. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the sponsors official policy, expressed or inferred. Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. Gpt-4 technical report. ar Xiv preprint ar Xiv:2303.08774, 2023a. Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. Gpt-4 technical report. ar Xiv preprint ar Xiv:2303.08774, 2023b. James Betker, Gabriel Goh, Li Jing, Tim Brooks, Jianfeng Wang, Linjie Li, Long Ouyang, Juntang Zhuang, Joyce Lee, Yufei Guo, et al. Improving image generation with better captions. Computer Science. https://cdn. openai. com/papers/dall-e-3. pdf, 2(3):8, 2023a. James Betker, Gabriel Goh, Li Jing, Tim Brooks, Jianfeng Wang, Linjie Li, Long Ouyang, Juntang Zhuang, Joyce Lee, Yufei Guo, Wesam Manassra, Prafulla Dhariwal, Casey Chu, Yunxin Jiao, and Aditya Ramesh. Improving image generation with better captions. 2023b. URL https://api.semanticscholar.org/ Corpus ID:264403242. Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877 1901, 2020. Soravit Changpinyo, Piyush Sharma, Nan Ding, and Radu Soricut. Conceptual 12m: Pushing web-scale image-text pre-training to recognize long-tail visual concepts. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3558 3568, 2021. Kevin Clark, Paul Vicol, Kevin Swersky, and David J Fleet. Directly fine-tuning diffusion models on differentiable rewards. ar Xiv preprint ar Xiv:2309.17400, 2023. Xiaoliang Dai, Ji Hou, Chih-Yao Ma, Sam Tsai, Jialiang Wang, Rui Wang, Peizhao Zhang, Simon Vandenhende, Xiaofang Wang, Abhimanyu Dubey, et al. Emu: Enhancing image generation models using photogenic needles in a haystack. ar Xiv preprint ar Xiv:2309.15807, 2023. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248 255. Ieee, 2009. Tim Dockhorn, Arash Vahdat, and Karsten Kreis. Genie: Higher-order denoising diffusion solvers. Advances in Neural Information Processing Systems, 35:30150 30166, 2022. Hanze Dong, Wei Xiong, Deepanshu Goyal, Rui Pan, Shizhe Diao, Jipeng Zhang, Kashun Shum, and Tong Zhang. Raft: Reward ranked finetuning for generative foundation model alignment. ar Xiv preprint ar Xiv:2304.06767, 2023. Ying Fan, Olivia Watkins, Yuqing Du, Hao Liu, Moonkyung Ryu, Craig Boutilier, Pieter Abbeel, Mohammad Ghavamzadeh, Kangwook Lee, and Kimin Lee. Reinforcement learning for fine-tuning text-to-image diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Xidong Feng, Ziyu Wan, Muning Wen, Ying Wen, Weinan Zhang, and Jun Wang. Alphazero-like tree-search can guide large language model decoding and training. ar Xiv preprint ar Xiv:2309.17179, 2023. Published in Transactions on Machine Learning Research (10/2024) Zhengyang Geng, Ashwini Pokle, and J Zico Kolter. One-step diffusion distillation via deep equilibrium models. Advances in Neural Information Processing Systems, 36, 2024. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Jonathan Ho and Tim Salimans. Classifier-free diffusion guidance. ar Xiv preprint ar Xiv:2207.12598, 2022. Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840 6851, 2020. Hexiang Hu, Kelvin CK Chan, Yu-Chuan Su, Wenhu Chen, Yandong Li, Kihyuk Sohn, Yang Zhao, Xue Ben, Boqing Gong, William Cohen, et al. Instruct-imagen: Image generation with multi-modal instruction. ar Xiv preprint ar Xiv:2401.01952, 2024. Aapo Hyvärinen and Peter Dayan. Estimation of non-normalized statistical models by score matching. Journal of Machine Learning Research, 6(4), 2005. Alexia Jolicoeur-Martineau, Ke Li, Rémi Piché-Taillefer, Tal Kachman, and Ioannis Mitliagkas. Gotta go fast when generating data with score-based models. ar Xiv preprint ar Xiv:2105.14080, 2021. Tero Karras, Miika Aittala, Timo Aila, and Samuli Laine. Elucidating the design space of diffusion-based generative models. Advances in Neural Information Processing Systems, 35:26565 26577, 2022. Dongjun Kim, Chieh-Hsin Lai, Wei-Hsiang Liao, Naoki Murata, Yuhta Takida, Toshimitsu Uesaka, Yutong He, Yuki Mitsufuji, and Stefano Ermon. Consistency trajectory models: Learning probability flow ode trajectory of diffusion. ar Xiv preprint ar Xiv:2310.02279, 2023a. Kyuyoung Kim, Jongheon Jeong, Minyong An, Mohammad Ghavamzadeh, Krishnamurthy Dj Dvijotham, Jinwoo Shin, and Kimin Lee. Confidence-aware reward optimization for fine-tuning text-to-image models. In The Twelfth International Conference on Learning Representations, 2023b. Diederik Kingma, Tim Salimans, Ben Poole, and Jonathan Ho. Variational diffusion models. Advances in neural information processing systems, 34:21696 21707, 2021. Yuval Kirstain, Adam Polyak, Uriel Singer, Shahbuland Matiana, Joe Penna, and Omer Levy. Pick-a-pic: An open dataset of user preferences for text-to-image generation. Advances in Neural Information Processing Systems, 36, 2024. Max Ku, Dongfu Jiang, Cong Wei, Xiang Yue, and Wenhu Chen. Viescore: Towards explainable metrics for conditional image synthesis evaluation. ar Xiv preprint ar Xiv:2312.14867, 2023. Kimin Lee, Hao Liu, Moonkyung Ryu, Olivia Watkins, Yuqing Du, Craig Boutilier, Pieter Abbeel, Mohammad Ghavamzadeh, and Shixiang Shane Gu. Aligning text-to-image models using human feedback. ar Xiv preprint ar Xiv:2302.12192, 2023. Junnan Li, Dongxu Li, Caiming Xiong, and Steven Hoi. Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation. In ICML, 2022a. Junyi Li, Tianyi Tang, Wayne Xin Zhao, Jian-Yun Nie, and Ji-Rong Wen. Pretrained language models for text generation: A survey. ar Xiv preprint ar Xiv:2201.05273, 2022b. Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In Computer Vision ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pp. 740 755. Springer, 2014. Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. Dpm-solver: A fast ode solver for diffusion probabilistic model sampling in around 10 steps. Advances in Neural Information Processing Systems, 35:5775 5787, 2022a. Published in Transactions on Machine Learning Research (10/2024) Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. Dpm-solver++: Fast solver for guided sampling of diffusion probabilistic models. ar Xiv preprint ar Xiv:2211.01095, 2022b. Dengsheng Lu and Qihao Weng. A survey of image classification methods and techniques for improving classification performance. International journal of Remote sensing, 28(5):823 870, 2007. Yujie Lu, Xianjun Yang, Xiujun Li, Xin Eric Wang, and William Yang Wang. Llmscore: Unveiling the power of large language models in text-to-image synthesis evaluation. Advances in Neural Information Processing Systems, 36, 2024. Eric Luhman and Troy Luhman. Knowledge distillation in iterative generative models for improved sampling speed. ar Xiv preprint ar Xiv:2101.02388, 2021. Simian Luo, Yiqin Tan, Longbo Huang, Jian Li, and Hang Zhao. Latent consistency models: Synthesizing high-resolution images with few-step inference. ar Xiv preprint ar Xiv:2310.04378, 2023a. Simian Luo, Yiqin Tan, Suraj Patil, Daniel Gu, Patrick von Platen, Apolinário Passos, Longbo Huang, Jian Li, and Hang Zhao. Lcm-lora: A universal stable-diffusion acceleration module. ar Xiv preprint ar Xiv:2311.05556, 2023b. Chenlin Meng, Robin Rombach, Ruiqi Gao, Diederik Kingma, Stefano Ermon, Jonathan Ho, and Tim Salimans. On distillation of guided diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14297 14306, 2023. Thuan Hoang Nguyen and Anh Tran. Swiftbrush: One-step text-to-image diffusion model with variational score distillation. ar Xiv preprint ar Xiv:2312.05239, 2023. Alexander Quinn Nichol and Prafulla Dhariwal. Improved denoising diffusion probabilistic models. In International Conference on Machine Learning, pp. 8162 8171. PMLR, 2021. Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730 27744, 2022. Mihir Prabhudesai, Anirudh Goyal, Deepak Pathak, and Katerina Fragkiadaki. Aligning text-to-image diffusion models with reward backpropagation. ar Xiv preprint ar Xiv:2310.03739, 2023. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748 8763. PMLR, 2021. Rafael Rafailov, Archit Sharma, Eric Mitchell, Christopher D Manning, Stefano Ermon, and Chelsea Finn. Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36, 2024. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10684 10695, 2022. Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily L Denton, Kamyar Ghasemipour, Raphael Gontijo Lopes, Burcu Karagol Ayan, Tim Salimans, et al. Photorealistic textto-image diffusion models with deep language understanding. Advances in Neural Information Processing Systems, 35:36479 36494, 2022. Tim Salimans and Jonathan Ho. Progressive distillation for fast sampling of diffusion models. ar Xiv preprint ar Xiv:2202.00512, 2022. Axel Sauer, Dominik Lorenz, Andreas Blattmann, and Robin Rombach. Adversarial diffusion distillation. ar Xiv preprint ar Xiv:2311.17042, 2023. Published in Transactions on Machine Learning Research (10/2024) Christoph Schuhmann, Romain Beaumont, Richard Vencu, Cade Gordon, Ross Wightman, Mehdi Cherti, Theo Coombes, Aarush Katta, Clayton Mullis, Mitchell Wortsman, et al. Laion-5b: An open largescale dataset for training next generation image-text models. Advances in Neural Information Processing Systems, 35:25278 25294, 2022. Eyal Segalis, Dani Valevski, Danny Lumen, Yossi Matias, and Yaniv Leviathan. A picture is worth a thousand words: Principled recaptioning improves image generation. ar Xiv preprint ar Xiv:2310.16656, 2023. Andy Shih, Suneel Belkhale, Stefano Ermon, Dorsa Sadigh, and Nima Anari. Parallel sampling of diffusion models. Advances in Neural Information Processing Systems, 36, 2024. Jaskirat Singh and Liang Zheng. Divide, evaluate, and refine: Evaluating and improving text-to-image alignment with iterative vqa feedback. Advances in Neural Information Processing Systems, 36, 2024. Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. In International conference on machine learning, pp. 2256 2265. PMLR, 2015. Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models. ar Xiv preprint ar Xiv:2010.02502, 2020a. Yang Song and Prafulla Dhariwal. Improved techniques for training consistency models. ar Xiv preprint ar Xiv:2310.14189, 2023. Yang Song and Stefano Ermon. Generative modeling by estimating gradients of the data distribution. Advances in neural information processing systems, 32, 2019. Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. ar Xiv preprint ar Xiv:2011.13456, 2020b. Yang Song, Prafulla Dhariwal, Mark Chen, and Ilya Sutskever. Consistency models. International conference on machine learning, 2023. Nisan Stiennon, Long Ouyang, Jeffrey Wu, Daniel Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, and Paul F Christiano. Learning to summarize with human feedback. Advances in Neural Information Processing Systems, 33:3008 3021, 2020. Jiao Sun, Deqing Fu, Yushi Hu, Su Wang, Royi Rassin, Da-Cheng Juan, Dana Alon, Charles Herrmann, Sjoerd van Steenkiste, Ranjay Krishna, et al. Dreamsync: Aligning text-to-image generation with image understanding feedback. ar Xiv preprint ar Xiv:2311.17946, 2023. Gemini Team, Rohan Anil, Sebastian Borgeaud, Yonghui Wu, Jean-Baptiste Alayrac, Jiahui Yu, Radu Soricut, Johan Schalkwyk, Andrew M Dai, Anja Hauth, et al. Gemini: a family of highly capable multimodal models. ar Xiv preprint ar Xiv:2312.11805, 2023. Ashwin K Vijayakumar, Michael Cogswell, Ramprasath R Selvaraju, Qing Sun, Stefan Lee, David Crandall, and Dhruv Batra. Diverse beam search: Decoding diverse solutions from neural sequence models. ar Xiv preprint ar Xiv:1610.02424, 2016. Patrick von Platen, Suraj Patil, Anton Lozhkov, Pedro Cuenca, Nathan Lambert, Kashif Rasul, Mishig Davaadorj, and Thomas Wolf. Diffusers: State-of-the-art diffusion models. https://github.com/ huggingface/diffusers, 2022. Bram Wallace, Meihua Dang, Rafael Rafailov, Linqi Zhou, Aaron Lou, Senthil Purushwalkam, Stefano Ermon, Caiming Xiong, Shafiq Joty, and Nikhil Naik. Diffusion model alignment using direct preference optimization. ar Xiv preprint ar Xiv:2311.12908, 2023a. Published in Transactions on Machine Learning Research (10/2024) Bram Wallace, Meihua Dang, Rafael Rafailov, Linqi Zhou, Aaron Lou, Senthil Purushwalkam, Stefano Ermon, Caiming Xiong, Shafiq Joty, and Nikhil Naik. Diffusion model alignment using direct preference optimization. ar Xiv preprint ar Xiv:2311.12908, 2023b. Xiaoshi Wu, Yiming Hao, Keqiang Sun, Yixiong Chen, Feng Zhu, Rui Zhao, and Hongsheng Li. Human preference score v2: A solid benchmark for evaluating human preferences of text-to-image synthesis. ar Xiv preprint ar Xiv:2306.09341, 2023a. Xiaoshi Wu, Keqiang Sun, Feng Zhu, Rui Zhao, and Hongsheng Li. Human preference score: Better aligning text-to-image models with human preference. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 2096 2105, 2023b. Jiazheng Xu, Xiao Liu, Yuchen Wu, Yuxuan Tong, Qinkai Li, Ming Ding, Jie Tang, and Yuxiao Dong. Imagereward: Learning and evaluating human preferences for text-to-image generation. Advances in Neural Information Processing Systems, 36, 2024. Shentao Yang, Tianqi Chen, and Mingyuan Zhou. A dense reward view on aligning text-to-image diffusion with preference. ar Xiv preprint ar Xiv:2402.08265, 2024. Jiahui Yu, Yuanzhong Xu, Jing Yu Koh, Thang Luong, Gunjan Baid, Zirui Wang, Vijay Vasudevan, Alexander Ku, Yinfei Yang, Burcu Karagol Ayan, et al. Scaling autoregressive models for content-rich text-toimage generation. ar Xiv preprint ar Xiv:2206.10789, 2(3):5, 2022. Kexun Zhang, Xianjun Yang, William Yang Wang, and Lei Li. Redi: Efficient learning-free diffusion inference via trajectory retrieval. ar Xiv preprint ar Xiv:2302.02285, 2023. Qinsheng Zhang and Yongxin Chen. Fast sampling of diffusion models with exponential integrator. ar Xiv preprint ar Xiv:2204.13902, 2022. Yinan Zhang, Eric Tzeng, Yilun Du, and Dmitry Kislyuk. Large-scale reinforcement learning for diffusion models. ar Xiv preprint ar Xiv:2401.12244, 2024. Hongkai Zheng, Weili Nie, Arash Vahdat, Kamyar Azizzadenesheli, and Anima Anandkumar. Fast sampling of diffusion models via operator learning. In International Conference on Machine Learning, pp. 42390 42402. PMLR, 2023. Huangjie Zheng, Pengcheng He, Weizhu Chen, and Mingyuan Zhou. Truncated diffusion probabilistic models and diffusion-based adversarial auto-encoders. ar Xiv preprint ar Xiv:2202.09671, 2022. Daniel M Ziegler, Nisan Stiennon, Jeffrey Wu, Tom B Brown, Alec Radford, Dario Amodei, Paul Christiano, and Geoffrey Irving. Fine-tuning language models from human preferences. ar Xiv preprint ar Xiv:1909.08593, 2019. Published in Transactions on Machine Learning Research (10/2024) In the main paper, we distill our RG-LCD from the Stable Diffusion-v2.1 (768 x 768) (Rombach et al., 2022). Fig. 8 further shows the samples from our RG-LCM (HPSv) distilled from Stable Diffusion-v2.1-base (512 x 512) (Rombach et al., 2022). The rest of the appendix is structured as below Appendix A details the experimental setup and hyperparameter configurations. Appendix B elaborates on the training processes and sampling techniques from a (latent) CM. Appendix C shows extra samples generated by various models. Published in Transactions on Machine Learning Research (10/2024) Figure 8: Samples from our RG-LCM (HPSv2.1) with the teacher Stable Diffusion v2.1-base. The resolution is 512 x 512. Published in Transactions on Machine Learning Research (10/2024) CLIPScore Pick Score Image Reward HPSv2.1 β 5.0 5.0 1.0 1.0 Table 2: β for different RG-LCMs when training with different RMs. A Additional Experimental Details and Hyperparameters (HPs) For qualitative evaluation, we ensure consistency across all methods by using the same random seed for head-to-head image comparisons. As mentioned in Sec. 5, our training are conducted on the CC12M datasets (Changpinyo et al., 2021), as the LAION-Aesthetics datasets (Schuhmann et al., 2022) used in the original LCM paper (Luo et al., 2023a) are not accessible2. We train all LCMs (including RG-LCMs and the standard LCM) by distilling from the teacher LDM Stable Diffusion-v2.1 (Rombach et al., 2022) for 10K gradient steps on 8 NVIDIA A100 GPUs. When learning the standard LCM, we use a batch size 32 on each GPU (256 effective batch size). For RG-LCMs, we use a batch size 5 on each GPU (40 effective batch size). Interestingly, we observe that different batch sizes do not impact the final performance too much. We use the same set of hyperparameters (HP) for training RG-LCM and the standard LCM by following the settings listed in the diffusers (von Platen et al., 2022) library, except that RG-LCM has a unique HP β. Specifically, we set the learning rate 1e 6, EMA rate µ = 0.95 and the guidance scale range [ωmin, ωmax] = [5, 15]. We include more training details in Appendix A. Following the practice in (Luo et al., 2023a), we initialize fθ with the same parameters as the teacher LDM. We further encode the CFG scale ω by applying the Fourier embedding to ω and integrate it into the LCM backbone by adding the projected ω-embedding into the original embedding as in (Meng et al., 2023). As mentioned in Sec. 3.3, we use DDIM (Song et al., 2020a) as our ODE solver Ψ with a skipping step k = 20, the formula of the DDIM ODE solver ΨDDIM from tn+k to tn is given below (Luo et al., 2023a) ΨDDIM ztn+k, tn+k, tn, c = αtn αtn+k ztn+k βtn αtn+k βtn 1 ˆϵψ ztn+k, c, tn+k | {z } DDIM Estimated ztn ztn+k, (13) where ˆϵψ denotes the noise prediction model from the teacher LDM. αtn and βtn specify the noise schedule. For the forward process SDE defined in equation 1, we have µ(t) = d log α(t) dt , σ2(t) = dβ2(t) dt 2d log α(t) dt β2(t). (14) As a result, we have pt(xt) = N(xt|α(t), β2(t)I). We refer interested readers to the original LCM paper (Luo et al., 2023a) for further details. Reward scale β for different RG-LCMs with different RMs. In Sec. 5.1 and 5.2, we train our RG-LCMs with different RMs, including CLIPScore (Radford et al., 2021), Pick Score (Kirstain et al., 2024), Image Reward (Xu et al., 2024) and HPSv2.1 (Wu et al., 2023a). Table 2 shows the β we used for different RMs when obtaining the results in Fig. 4, 5 and Table 1. Details for integrating an LRM into RG-LCM As discussed in Sec. 4.2, the LRM admits a similar architecture as the CLIP (Radford et al., 2021) model, with the distinction of replacing the visual encoder with a latent encoder. We retain the original pretrained text encoder and focus on pretraining the latent encoder from scratch. This process mirrors the CLIP s pretraining approach, minimizing the same contrastive loss on the CC12M datasets (Changpinyo et al., 2021). The image latent is extracted with the same VAE encoder used in the teacher LDM Stable Diffusion-v2.1. Upon completing the pretraining phase, the LRM demonstrates promising initial results, achieving a zero-shot Top-1 Accuracy of 38.8% and Top-5 Accuracy of 2https://laion.ai/notes/laion-maintanence Published in Transactions on Machine Learning Research (10/2024) 66.47% on the Image Net validation set (Deng et al., 2009). These metrics underscore the model s fundamental capability in understanding text-image alignments. During the RG-LCD process, we finetune the LRM to match the preference of an expert RM. We train the last 2 layer of the latent encoder and the last 5 layers of the text encoder. We set the learning rate to 0.0000033 following (Wu et al., 2023a). Note that we do not perform heavy HP searches to determine their optimal values. As we are finetuning our LRM, there is a potential risk of overfitting the model to the training datasets, which could degrade the quality of generated outputs if training continues indefinitely. We emphasize that this is unlikely to pose a problem for our method. In practice, we use a large and diverse text-image dataset, such as CC12M. We also fixed the training to 10K iterations and observed stable performance without encountering any training instability. We hypothesize that performance degradation would only occur if training exceeds one full epoch of the dataset. However, even with a large batch size of 256, one epoch would require 12M / 256 = 47.9K iterations, which is far beyond the 10K iterations we used. Therefore, early stopping is not a critical concern for our approach. Nonetheless, we could still implement a stopping criterion by monitoring the average rewards of training batches. We can stop the LRM training when the average rewards converge to a specific value, ensuring the LRM is not overtrained. Published in Transactions on Machine Learning Research (10/2024) B Training and Sampling from (Latent) CM B.1 Multistep sampling from a learned CM and LCM We provide the pseudo-codes for multistep consistency sampling (Song et al., 2023) and multistep latent consistency sampling (Luo et al., 2023a) procedures in Algorithm 1 and Algorithm 2, respectively. The multistep sampling procedures alternate between the consistency mapping and noise-injection steps, trading additional computation resources for better sample quality. In the n-th iteration, we first perturb the predicted sample x (or z) with Gaussian noise to obtain ˆxτn (or ˆzτn). We then map the noisy sample ˆxτn (or ˆzτn) to obtain a new x (or z). Algorithm 1 Multistep Consistency Sampling Require: CM fθ, steps N, timestep sequence τ1 > τ2 > > τN 1, noise schedule α(t), β(t). Sample initial noise ˆx T N(0, I) x fθ(ˆx T , T) for n = 1, . . . , N 1 do Sample ˆxτn N(α(τn)x, β2(τn)I) x f(ˆxτn, τn) end for Return x Algorithm 2 Multistep Latent Consistency Sampling Require: LCM fθ, steps N, timestep sequence τ1 > τ2 > > τN 1, noise schedule α(t), β(t), text prompt c, CFG scale ω, VAE decoder D. Sample initial noise ˆz T N(0, I) z fθ(ˆz T , ω, c, T) for n = 1, . . . , N 1 do Sample ˆzτn N(α(τn)z, β2(τn)I) z fθ(ˆzτn, ω, c, T) end for Return D(z) B.2 Training procedures of RG-LCD Algorithm 3 Reward Guided Latent Consistency Distillation Require: dataset D, initial model parameter θ, learning rate η, ODE solver Ψ, distance metric d, EMA rate µ, learning rate η, noise schedule α(t), β(t), guidance scale [ωmin, ωmax], skipping interval k, VAE encoder E, decoder D, reward model R, reward scale β Encoding training data into latent space: Dz = {(z, c) | z = E(x), (x, c) D} θ θ repeat Sample (z, c) Dz, n U[1, N k] and ω [ωmin, ωmax] Sample ztn+k N α (tn+k) z; σ2 (tn+k) I ˆzΨ,ω tn ztn+k + (1 + ω)Ψ ztn+k, tn+k, tn, c ωΨ ztn+k, tn+k, tn, L (θ, θ ; Ψ) d fθ ztn+k, ω, c, tn+k , fθ ˆzΨ,ω tn , ω, c, tn - β R D fθ ztn+k, ω, c, tn+k , c θ θ η θL (θ, θ ) θ stop_grad (µθ + (1 µ)θ) until convergence Algorithm 3 presents the pseudo-codes for our RG-LCD. We use the red color to highlight the difference between our RG-LCD and the standard LCD (Luo et al., 2023a). B.3 Training procedures of RG-LCD with a Latent Proxy RM Algorithm 4 presents the training codes for our RG-LCD with a Latent Proxy RM. Published in Transactions on Machine Learning Research (10/2024) Algorithm 4 Reward Guided Latent Consistency Distillation with a Latent Proxy RM Require: dataset D, initial model parameter θ, learning rate η, ODE solver Ψ, distance metric d, EMA rate µ, learning rates η1, η2, noise schedule α(t), β(t), guidance scale [ωmin, ωmax], skipping interval k, VAE encoder E, decoder D, reward scale β, expert RM RE, LRM RL σ, temperature τE and τL Encoding training data into latent space: Dz = {(z, c) | z = E(x), (x, c) D} θ θ repeat # Calculate the Training loss for fθ Sample (z, c) Dz, n U[1, N k] and ω [ωmin, ωmax] Sample ztn+k N α (tn+k) z; σ2 (tn+k) I ˆzΨ,ω tn ztn+k + (1 + ω)Ψ ztn+k, tn+k, tn, c ωΨ ztn+k, tn+k, tn, Detach the parameter σ of RL σ L (θ, θ ; Ψ, σ) d fθ ztn+k, ω, c, tn+k , fθ ˆzΨ,ω tn , ω, c, tn - β RL σ ztn+k, c # Calculate the Training loss for RE σ z0 z, z1 fθ ztn+k, ω, c, tn+k , z2 fθ ˆzΨ,ω tn , ω, c, tn for i {0, 1} do for j {i, 2} do Derive the preference distribution: P σ i,j(m) exp RL σ (zm, c) /τL , m {i, j} Derive the preference distribution: Qi,j(m) exp RE (D (zm) , c) /τE , m {i, j} end for end for LRM(σ) P1 i=0 P2 j=i+1 DKL P σ i,j||stop_grad (Qi,j) # Update the learnable parameters via gradient descent θ θ η1 θL (θ, θ ) θ stop_grad (µθ + (1 µ)θ) σ θ η2 σLRM(σ) until convergence Published in Transactions on Machine Learning Research (10/2024) C Additional Qualitative Results We provide additional samples from our RG-LCMs with different RMs compared with the baseline LCM and teacher Stable Diffusion v2.1 in Fig. 9 and 10. The prompts for images in Fig. 9 in the left-to-right order are given below Van Gogh painting of a teacup on the desk Impressionist painting of a cat, textured, hypermodern photo of a kid playing , snow filling the air A fluffy owl sits atop a stack of antique books in a detailed and moody illustration. a deer reading a book a photo of a monkey wearing glasses in a suit The prompts for images in Fig. 10 in the left-to-right order are given below ornate archway inset with matching fireplace in room Poster of a mechanical cat, techical Schematics viewed from front portrait of a person with Cthulhu features, painted by Bouguereau. a serene nighttime cityscape with lake reflections, fruit trees Teddy bears working on new AI research on the moon in the 1980s. A cinematic shot of robot with colorful feathers Fig. 11 presents image samples when integrating a latent proxy RM (LRM) into our RG-LCD procedures. The prompts in the left-to-right order are given below a man in a brown blazer standing in front of smoke, backlit, in the style of gritty hollywood glamour, light brown and emerald, movie still, emphasis on facial expression, robert bevan, violent, dappled a cute pokemon resembling a blue duck wearing a puffy coat Highly detailed photograph of a meal with many dishes. Fig. 12 further qualitatively compares RG-LCM (HPS) + LRM and the standard LCM. The prompts in the top-to-bottom order are given below (Masterpiece:1. 5), RAW photo, film grain, (best quality:1. 5), (photorealistic), realistic, real picture, intricate details, photo of full body a cute cat in a medieval warrior costume, ((wastelands background)), diamond crown on head, (((dark background))) back view of a woman walking at Shibuya Tokyo, shot on Afga Vista 400, night with neon side lighting Fall And Autumn Wallpaper Daniel Wall Rainy Day In Autumn Painting Oil Artwork A bird-eye shot photograph of New York City, shot on Lomography Color Negative 800 painting of forest and pond Published in Transactions on Machine Learning Research (10/2024) Figure 9: Samples from our RG-LCMs with different RMs compared with the baseline LCM and teacher Stable Diffusion v2.1. Published in Transactions on Machine Learning Research (10/2024) Figure 10: More samples from our RG-LCMs with different RMs compared with the baseline LCM and teacher Stable Diffusion v2.1. Published in Transactions on Machine Learning Research (10/2024) Figure 11: Effect of the Latent proxy RM (LRM). Integrating the LRM into our RG-LCD procedures makes the generated images natural, corresponding to the lower FID in Table 1. Moreover, it helps eliminate the high-frequency noise in the generated images. 27 Published in Transactions on Machine Learning Research (10/2024) Figure 12: Comparison between RG-LCM (HPS) + LRM with LCM. The samples from RG-LCM (HPS) + LRM are visually appealing while remaining natural, corresponding to the high HPSv2.1 score and low FID in Table 1. Published in Transactions on Machine Learning Research (10/2024) Figure 13: Additional images to study the impact of the reward scale β. We generate all samples with 4 steps. Fig. 13 includes additional samples for the ablation on the reward scale β. The prompts in the top-to-bottom order are given below Ultra realistic photo of a single light bulb, dramatic lighting Pirate ship trapped in a cosmic maelstrom nebula A golden retriever wearing VR goggles. Highly detailed portrait of a woman with long hairs, stephen bliss, unreal engine, fantasy art by greg rutkowski. A stunning beautiful oil painting of a lion, cinematic lighting, golden hour light. Published in Transactions on Machine Learning Research (10/2024) A raccoon wearing a tophat and suit, holding a briefcase, standing in front of a city skyline. Fig. 14 and 15 provide additional qualitative comparisons for RG-LCM (CLIP) with β {5, 100}, RG-LCM (HPS) with β = 1, and RG-LCM (HPS) + LRM with β = 100. The prompts for Fig. 14 in the top-to-bottom order are given below A game screenshot featuring Woolie Madden with dreadlocks in Mass Effect. Two girls holding hands while watching the world burn A full-body portrait of a female cybered shadowrunner with a dark and cyberpunk atmosphere created by Echo Chernik in the style of Shadowrun Returns PC game. A portrait of a skeleton possessed by a spirit with green smoke exiting its empty eyes. A counter in a coffee house with choices of coffee and syrup flavors. The prompts for Fig. 15 in the top-to-bottom order are given below Black and white portrait of Thabo Mbeki with highly detailed ink lines and a cyberpunk flair, created for the Inktober challenge as part of the Cyberpunk 2020 manual coloring pages. The image features a big white cliff, a cargo favela, a wall fortress, a neon pub, and some plants, with vivid and colorful style depicted in hyperrealistic CGI. An albino lion wearing a Mafia hat, digitally painted by multiple artists, trending on Artstation. A galaxy-colored Dn D dice is shown against a sunset over a sea, in artwork by Greg Rutkowski and Thomas Kinkade that is trending on Artstation. A pirate skeleton. Fig. 16 additional quantitative results for RG-LCM (Img Rwd) with images resized to a low resolution of 224x224. Please use Adobe Acrobat Reader and set the zoom to 100% (actual size). At this resolution, high-frequency noise becomes less noticeable. The prompts, listed in order from top to bottom and left to right, are provided below: A creepy cartoon rabbit wearing pants and a shirt, with dramatic lightning and a cinematic atmosphere. A beaver in formal attire stands beside books in a library. A pencil sketch of Victoria Justice drawn in the Disney style by Milt Kahl. Portrait of a male furry Black Reindeer anthro wearing black and rainbow galaxy clothes, with wings and tail, in an outerspace city at night while it rains. A galaxy-colored Dn D dice is shown against a sunset over a sea, in artwork by Greg Rutkowski and Thomas Kinkade that is trending on Artstation. Architecture render with pleasing aesthetics. An empty road with buildings on each side. A computer monitor glows on a wooden desk that has a black computer chair near it. A man standing by his motorcycle is looking out to take in the view. A koala bear dressed as a ninja in a kayak. Baby Yoda depicted in the style of Assassination Classroom anime. A puppy is driving a car in a film still. Published in Transactions on Machine Learning Research (10/2024) Figure 14: We present additional qualitative comparisons between RG-LCM (CLIP) with β {5, 100}, RG-LCM (HPS) with β = 1, and RG-LCM (HPS) + LRM with β = 100. Published in Transactions on Machine Learning Research (10/2024) Figure 15: We present additional qualitative comparisons between RG-LCM (CLIP) with β {5, 100}, RG-LCM (HPS) with β = 1, and RG-LCM (HPS) + LRM with β = 100. Published in Transactions on Machine Learning Research (10/2024) Figure 16: Additional quantitative results for RG-LCM (Img Rwd) with images resized to a low resolution of 224x224. Please use Adobe Acrobat Reader and set the zoom to 100% (actual size). At this resolution, high-frequency noise becomes less noticeable. Published in Transactions on Machine Learning Research (10/2024) D Experiments with Additional Teacher T2I Models In this section, we conduct additional experiments using different teacher T2I models, including Stable Diffusion 1.5 (SD 1.5) and Stable Diffusion XL (SDXL). For each teacher model, we train both the baseline LCM and our RG-LCM (HPS) by learning from HPSv2.1, using DDIM as defined in equation 13 as our ODE solver Ψ. The CC12M (Changpinyo et al., 2021) dataset serves as our training dataset. For SDXL, due to GPU memory constraints, we apply LCM-Lo RA (Luo et al., 2023b) on top of SDXL to construct our RG-LCM and baseline LCM instead of performing full-model training. Additionally, we fix the weighting parameter β in equation 11 to 1, consistent with the settings for Stable Diffusion 2.1. Empirically, we evaluate different methods using 3,200 HPSv2 test prompts and employ VIEScore (Ku et al., 2023) as our evaluation metric with the GPT4o backbone. VIEScore achieves a high Spearman correlation of 0.4 with human evaluations, close to the human-to-human correlation of 0.45. Given a text-image pair, VIEScore provides Semantic Score, Quality Score, and Overall Score, reflecting text-image alignment, visual quality, and overall human preference, respectively. We compare the 4-step generation of our RG-LCM with the 4-step generation of the baseline LCM, as well as with the 4-step and 25-step generations from the teacher T2I model using DPM-Solver++ (Lu et al., 2022b) with CFG guidance (Ho & Salimans, 2022) and negative prompts. DPM-Solver++ is a high-order fast ODE solver that accelerates inference from diffusion models. It is important to note that we can also integrate DPM-Solver++ and negative prompts into our RG-LCM training. We leave it for future work. Table 3 and 4 present the evaluation results. In both cases, the 4-step generation of our RG-LCM (HPS) outperforms other 4-step baselines. When using SD 1.5 as the teacher model, our 4-step generation even surpasses the 25-step generation achieved using DPM-Solver++ with CFG and negative prompts. With SDXL as the teacher model, our 4-step generation slightly underperforms but still matches the 25-step generation from the teacher. We believe this performance drop may be due to 1) the Lo RA training and 2) the absence of high-quality image datasets. Therefore, we expect our RG-LCM to perform even better with full-model training and access to datasets with high image aesthetics, e.g., LAION-Aesthetics V2 6.5+ (Schuhmann et al., 2022). SD-1.5 as the Teacher Model NFEs Semantic Score Quality Score Overall Score DPM-Solver++, Negative Prompt 4 6.06 4.98 5.23 DPM-Solver++, Negative Prompt 25 6.77 6.63 6.45 LCM 4 6.75 5.88 6.02 RG-LCM (HPS) 4 7.55 7.02 7.11 Table 3: Evaluation of different methods using Stable Diffusion 1.5 as the teacher model on the HPSv2 test prompts. NFEs denote the number of function evaluations during inference. We employ VIEScore as the evaluation metric. By learning from the feedback of HPSv2.1, the 4-step generation of our RG-LCM (HPS) not only outperforms other 4-step baselines but also surpasses the 25-step generation achieved using DPM-Solver++ when sampling from the teacher model with CFG and negative prompts. SDXL as the Teacher Model NFEs Semantic Score Quality Score Overall Score DPM-Solver++, Negative Prompt 4 6.63 4.99 5.52 DPM-Solver++, Negative Prompt 25 8.23 7.73 7.83 LCM 4 7.26 6.43 6.65 RG-LCM (HPS) 4 8.1 7.46 7.64 Table 4: Evaluation of different methods using Stable Diffusion XL as the teacher model on the HPSv2 test prompts. NFEs denote the number of function evaluations during inference. We employ VIEScore as the evaluation metric. By learning from the feedback of HPSv2.1, the 4-step generation of our RG-LCM (HPS) not only outperforms other 4-step baselines but also matches the performance of the 25-step generation when using DPM-Solver++ to sample from the teacher model with CFG and negative prompts.