# heavytailed_diffusion_models__d194d0b9.pdf Published as a conference paper at ICLR 2025 HEAVY-TAILED DIFFUSION MODELS Kushagra Pandey1,2 , Jaideep Pathak1, Yilun Xu1, Stephan Mandt2, Michael Pritchard1,2, Arash Vahdat1:, Morteza Mardani1: NVIDIA1, University of California, Irvine 2 {pandeyk1,mandt}@uci.edu, {jpathak,yilunx,mpritchard,avahdat,mmardani}@nvidia.com Diffusion models achieve state-of-the-art generation quality across many applications, but their ability to capture rare or extreme events in heavy-tailed distributions remains unclear. In this work, we show that traditional diffusion and flow-matching models with standard Gaussian priors fail to capture heavy-tailed behavior. We address this by repurposing the diffusion framework for heavy-tail estimation using multivariate Student-t distributions. We develop a tailored perturbation kernel and derive the denoising posterior based on the conditional Student-t distribution for the backward process. Inspired by γ-divergence for heavy-tailed distributions, we derive a training objective for heavy-tailed denoisers. The resulting framework introduces controllable tail generation using only a single scalar hyperparameter, making it easily tunable for diverse real-world distributions. As specific instantiations of our framework, we introduce t-EDM and t-Flow, extensions of existing diffusion and flow models that employ a Student-t prior. Remarkably, our approach is readily compatible with standard Gaussian diffusion models and requires only minimal code changes. Empirically, we show that our t-EDM and t-Flow outperform standard diffusion models in heavy-tail estimation on high-resolution weather datasets in which generating rare and extreme events is crucial. 1 INTRODUCTION In many real-world applications, such as weather forecasting, rare or extreme events like hurricanes or heatwaves can have disproportionately larger impacts than more common occurrences. Therefore, building generative models capable of accurately capturing these extreme events is critically important (Gründemann et al., 2022). However, learning the distribution of such data from finite samples is particularly challenging, as the number of empirically observed tail events is typically small, making accurate estimation difficult. One promising approach is to use heavy-tailed distributions, which allocate more density to the tails than light-tailed alternatives. In popular generative models like Normalizing Flows (Rezende & Mohamed, 2016) and Variational Autoencoders (VAEs) (Kingma & Welling, 2022), recent works address heavy-tail estimation by learning a mapping from a heavy-tailed prior to the target distribution (Jaini et al., 2020; Kim et al., 2024). While these works advocate for heavy-tailed base distributions, their application to real-world, highdimensional datasets remains limited, with empirical results focused on small-scale or toy datasets. In contrast, diffusion models (Ho et al., 2020; Song et al., 2020; Lipman et al., 2023) have demonstrated excellent synthesis quality in large-scale applications. However, it is unclear whether diffusion models with Gaussian priors can effectively model heavy-tailed distributions without significant modifications. Work during an internship at NVIDIA :Equal Advising Published as a conference paper at ICLR 2025 Figure 1: Toy Illustration. Our proposed diffusion model (t-Diffusion) captures heavy-tailed behavior more accurately than standard Gaussian diffusion (see histogram comparisons in the top panel, x-axis). The framework allows for controllable tail estimation using a hyperparameter ν, which can be adjusted for each dimension. Lower ν values model heavier tails, while higher values approach Gaussian diffusion (Best viewed when zoomed in; see App. C.3 for details). Brighter colors indicate high-density regions In this work, we first demonstrate through extensive experiments that traditional diffusion models even with proper normalization, preconditioning, and noise schedule design (see Section 4) fail to accurately capture the heavy-tailed behavior in target distributions (see Fig. 1 for a toy example). We hypothesize that, in highdimensional spaces, the Gaussian distribution in standard diffusion models tends to concentrate on a spherical narrow shell, thereby neglecting the tails. To address this, we adopt the multivariate Student-t distribution as the base noise distribution, with its degrees of freedom providing controllability over tail estimation. Consequently, we reformulate the denoising diffusion framework using multivariate Student-t distributions by designing a tailored perturbation kernel and deriving the corresponding denoiser. Moreover, we draw inspiration from the γ-power Divergences (Eguchi, 2021; Kim et al., 2024) for heavy-tailed distributions to formulate the learning problem for our heavy-tailed denoiser. We extend widely adopted diffusion models, such as EDM (Karras et al., 2022) and straight-line flows (Lipman et al., 2023; Liu et al., 2022), by introducing their Student-t counterparts: t-EDM and t-Flow. We derive the corresponding SDEs and ODEs for modeling heavy-tailed distributions. Through extensive experiments on the HRRR dataset (Dowell et al., 2022), we train both unconditional and conditional versions of these models. The results show that standard EDM struggles to capture tails and extreme events, whereas t-EDM performs significantly better in modeling such phenomena. To summarize, we present, Heavy-tailed Diffusion Models. We repurpose the diffusion model framework for heavy-tail estimation by formulating both the forward and reverse processes using multivariate Student-t distributions. The denoiser is learned by minimizing the γ-power divergence (Kim et al., 2024) between the forward and reverse posteriors. Continuous Counterparts. We derive continuous formulations for heavy-tailed diffusion models and provide a principled approach to constructing ODE and SDE samplers. This enables the instantiation of t-EDM and t-Flow as heavy-tailed alternatives to standard diffusion and flow models. Empirical Results. Experiments on the HRRR dataset (Dowell et al., 2022), a high-resolution dataset for weather modeling, show that t-EDM significantly outperforms EDM in capturing tail distributions for both unconditional and conditional tasks. Theoretical Connections. To theoretically justify the effectiveness of our approach, we present several theoretical connections between our framework and existing work in diffusion models and robust statistical estimators (Futami et al., 2018). Published as a conference paper at ICLR 2025 2 BACKGROUND As prerequisites underlying our method, we briefly summarize Gaussian diffusion models (as introduced in (Ho et al., 2020; Sohl-Dickstein et al., 2015)) and multivariate Student-t distributions. 2.1 DIFFUSION MODELS Diffusion models define a forward process (usually with an affine drift and no learnable parameters) to convert data x0 ppx0q, x0 P Rd to noise. A learnable reverse process is then trained to generate data from noise. In the discrete-time setting, the training objective for diffusion models can be specified as, DKLpqpx T |x0q } ppx T qq loooooooooooooomoooooooooooooon tą t DKLpqpxt t|xt, x0q } pθpxt t|xtqq looooooooooooooooooooooomooooooooooooooooooooooon log pθpx0|x tq looooooooomooooooooon where T denotes the trajectory length while t denotes the time increment between two consecutive time points. DKL denotes the Kullback-Leibler (KL) divergence defined as, DKLpq } pq ş qpxq log qpxq ppxqdx. In the objective in Eq. 1, the trajectory length T is chosen to match the generative prior ppx T q and the forward marginal qpx T |x0q. The second term in Eqn. 1 proposes to minimize the KL divergence between the forward posterior qpxt t|xt, x0q and the learnable posterior pθpxt t|xtq which corresponds to learning the denoiser (i.e. predicting a less noisy state from noise). The forward marginals, posterior, and reverse posterior are modeled using Gaussian distributions, which exhibit an analytical form of the KL divergence. The discrete-time diffusion framework can also be extended to the continuous time setting (Song et al., 2020; 2021; Karras et al., 2022). Recently, Lipman et al. (2023); Albergo et al. (2023) proposed stochastic interpolants (or flows), which allow flexible transport between two arbitrary distributions. 2.2 STUDENT-T DISTRIBUTIONS The multivariate Student-t distribution tdpµ, Σ, νq with dimensionality d, location µ, scale matrix Σ and degrees of freedom ν is defined as, tdpµ, Σ, νq Cν,d 1 1 ν px µq JΣ 1px µq ı ν d where Cν,d is the normalizing factor. Since the multivariate Student-t distribution has polynomially decaying density, it can model heavy-tailed distributions. Interestingly, for ν 1, the Student-t distribution is analogous to the Cauchy distribution. As ν Ñ 8, the Student-t distribution converges to the Gaussian distribution. A Student-t distributed random variable x can be reparameterized as (Andrews & Mallows, 1974), x µ Σ1{2z{?κ, with z Np0, Idq, κ χ2pνq{ν where χ2 denotes the Chi-squared distribution (See Fig. 8 for an illustration of the pdf of the χ2 distribution). 3 HEAVY-TAILED DIFFUSION MODELS We now repurpose standard diffusion models using multivariate Student-t distributions. The main idea behind our design is the use of heavy-tailed generative priors (Jaini et al., 2020; Kim et al., 2024) for learning a transport map towards a potentially heavy-tailed target distribution. From Eqn. 1 we note three key requirements for training diffusion models: choice of the perturbation kernel qpxt|x0q, form of the target denoising posterior qpxt t|xt, x0q and the parameterization of the learnable reverse posterior pθpxt 1|xtq. Therefore, we begin our discussion in the context of discrete-time diffusion models and later extend to the continuous regime. This has several advantages in terms of highlighting these three key design choices, which Published as a conference paper at ICLR 2025 might be obscured by the continuous-time framework of defining a forward and a reverse SDE (Karras et al., 2022) while at the same time leading to a simpler construction. Lastly, without loss of generality, we assume a scalar ν for subsequent analysis due to mathematical convenience. 3.1 NOISING PROCESS DESIGN WITH STUDENT-T DISTRIBUTIONS. Our construction of the noising process involves three key steps. 1. Firstly, given two consecutive noisy states xt and xt t, we specify a joint distribution qpxt, xt t|x0q. 2. Secondly, given qpxt, xt t|x0q, we construct the perturbation kernel qpxt|x0q ş qpxt, xt t|x0qdxt t which can be used as a noising process during training. 3. Lastly, from Steps 1 and 2, we construct the forward denoising posterior qpxt t|xt, x0q qpxt,xt t|x0q qpxt|x0q . We will later utilize the form of qpxt t|xt, x0q to parameterize the reverse posterior. It is worth noting that our construction of the noising process bypasses the specification of the forward transition kernel qpxt|xt tq. This has the advantage that we can directly specify the form of the perturbation kernel parameters µt and σt as in Karras et al. (2022) unlike Song et al. (2020); Ho et al. (2020). We next highlight the noising process construction in more detail. Specifiying the joint distribution qpxt, xt t|x0q. We parameterize the joint distribution qpxt, xt t|x0q as a multivariate Student-t distribution with the following form, qpxt, xt t|x0q t2dpµ, Σ, νq, µ rµt; µt tsx0, Σ ˆ σ2 t σ2 12ptq σ2 21ptq σ2 t t where µt, σt, σ12ptq, σ21ptq are time-dependent scalar design parameters. While the choice of the parameters µt and σt determines the perturbation kernel used during training, the choice of σ12ptq and σ21ptq can affect the ODE/SDE formulation for the denoising process and will be clarified when discussing sampling. Constructing the perturbation kernel qpxt|x0q: Given the joint distribution qpxt, xt t|x0q specified as a multivariate Student-t distribution, it follows that the perturbation kernel distribution qpxt|x0q is also a Student-t distribution (Ding, 2016) parameterized as, qpxt|x0q tdpµtx0, σ2 t Id, νq (proof in App. A.1). We choose the scalar coefficients µt and σt such that the perturbation kernel at time t T converges to a standard Student-t distribution. Later, we will set our generative prior ppx T q qpx T |x0q tdp0, Id, νq to instantiate sample generation. We discuss practical choices of µt and σt in Section 3.5. Estimating the reference denoising posterior. Given the joint distribution qpxt, xt t|x0q and the perturbation kernel qpxt|x0q, the denoising posterior can be specified as (see Ding (2016)), qpxt t|xt, x0q tdp µt, ν d1 ν d σ2 t Id, ν dq, (4) µt µt tx0 σ2 21ptq σ2 t pxt µtx0q, σ2 t σ2 t t σ2 21ptqσ2 12ptq σ2 t where d1 1 σ2 t }xt µtx0}2. Next, we formulate the training objective for heavy-tailed diffusions. 3.2 PARAMETERIZATION OF THE REVERSE POSTERIOR Following Eqn. 4, we parameterize the reverse (or the denoising) posterior distribution as: pθpxt t|xtq tdpµθpxt, tq, σ2 t Id, ν dq, (6) Published as a conference paper at ICLR 2025 where the denoiser mean µθpxt, tq is further parameterized as follows: µθpxt, tq σ2 21ptq σ2 t xt µt t σ2 21ptq σ2 t µt ı Dθpxt, σtq. (7) We discuss alternative posterior parameterizations in App. A.2. It is worth noting that while the noising process defined in Eq. 3 is non-markovian, our parameterization of the posterior is still Markovian. However, similar to Song et al. (2022a), this choice works well empirically (see Section 4). Moreover, when parameterizing the reverse posterior scale, we drop the data-dependent coefficient pν d1q{pν dq. This choice is primarily inspired by simplicity in deriving preconditioners (Sec. 3.5) and developing continuous-time sampling methods (Sec. 3.4) for heavy-tailed diffusions, resulting in models that require minimal implementation overhead over standard diffusion models during training and sampling (see Fig. 2). However, heteroskedastic modeling of the denoiser is possible in our framework and could be an interesting direction for future work. Next, we reformulate the training objective in Eqn. 1 for heavy-tailed diffusions. 3.3 TRAINING WITH POWER DIVERGENCES The optimization objective in Eqn. 1 primarily minimizes the KL-Divergence between a given pair of distributions. However, since we parameterize the distributions in Eqn. 1 using multivariate Student-t distributions, using the KL-Divergence might not be a suitable choice of divergence. This is because computing the KL divergence for Student-t distributions does not exhibit a closed-form expression. An alternative is the γ-Power Divergence (Eguchi, 2021; Kim et al., 2024) defined as, Dγpq } pq 1 Cγpq, pq Hγpqq , γ P p 1, 0q Y p0, 8q Hγppq }p}1 γ ˆ ż ppxq1 γdx 1 1 γ Cγpq, pq ż qpxq ˆ ppxq where, like Kim et al. (2024), we set γ 2 ν d for the remainder of our discussion. Moreover, Hγ and Cγ represent the γ-power entropy and cross-entropy, respectively. Interestingly, the γ-Power divergence between two multivariate Student-t distributions, qν tdpµ0, Σ0, νq and pν tdpµ1, Σ1, νq, can be tractably computed in closed form and is defined as (see Kim et al. (2024) for a proof), Dγrqν||pνs 1 γ 1 γ ν,d 1 d ν 2 γ 1 γ |Σ0| γ 2p1 γq 1 d ν 2 |Σ1| γ 2p1 γq 1 1 ν 2tr Σ 1 1 Σ0 1 ν pµ0 µ1q JΣ 1 1 pµ0 µ1q ı . (8) Therefore, analogous to Eqn. 1, we minimize the following optimization objective, Dγpqpx T |x0q } ppx T qq ÿ tą1 Dγpqpxt t|xt, x0q } pθpxt t|xtqq log pθpx0|x1q ȷ . (9) Here, we note a couple of caveats. Firstly, while replacing the KL-Divergence with the γ-Power Divergence in the objective in Eqn. 1 might appear to be due to computational convenience, the γ-power divergence has several connections with robust estimators (Futami et al., 2018) in statistics and provides a tunable parameter γ which can be used to control the model density assigned at the tail (see Section 5). Secondly, while the objective in Eqn. 1 is a valid ELBO, the objective in Eq. 9 is not. However, the following result provides a connection between the two objectives (see proof in App. A.3), Proposition 1. For arbitrary distributions q and p, in the limit of γ Ñ 0, Dγpq } pq converges to DKLpq } pq. Consequently, for a finite-dimensional dataset with x0 P Rd and γ 2 ν d, under the limit of γ Ñ 0, the objective in Eqn. 9 converges to the objective in Eqn. 1. Published as a conference paper at ICLR 2025 Component Gaussian Diffusion (Ours) t-Diffusion Perturbation Kernel (qpxt|x0q) Npµtx0, σ2 t Idq tdpµtx0, σ2 t Id, νq Forward Posterior (qpxt t|xt, x0q) Np µt, σ2 t Idq tdp µt, ν d1 ν d σ2 t Id, ν dq Reverse Posterior (pθpxt t|xtq) Npµθpxt, tq, σ2 t Idq tdpµθpxt, tq, σ2 t Id, ν dq Divergence Measure DKLpq } pq Dγpq } pq Generative Prior (ppx T q) Np0, Idq tdp0, Id, νq Table 1: Comparison between different modeling components for constructing Gaussian vs Heavy-Tailed diffusion models. Under the limit of ν Ñ 8, our proposed t-Diffusion framework converges to Gaussian diffusion models. Therefore, under the limit γ Ñ 0, the standard diffusion model framework becomes a special case of our proposed framework. Moreover, for γ 2{pν dq, this also explains the tail estimation moving towards Gaussian diffusion for an increasing ν (See Fig. 1 for an illustration). Lastly, the divergence-based interpretation of the ELBO loss in Eq. 1 has also been considered in prior work in generative models (Xiao et al.; Kim et al., 2024) and is also commonplace in M-estimators (Futami et al., 2018) used in robust statistics. Therefore, our choice of the training objective in Eq. 9 is quite principled. Simplifying the Training Objective. Plugging the form of the forward posterior qpxt t|xt, x0q in Eqn. 4, the reverse posterior pθpxt t|xtq in the optimization objective in Eqn. 9, we obtain the following simplified training loss (proof in App. A.4), Lpθq Ex0 ppx0q Et pptq Eϵ Np0,Idq Eκ 1 ν χ2pνq Dθ µtx0 σt ϵ ?κ, σt x0 2 Intuitively, the form of our training objective is similar to existing diffusion models (Ho et al., 2020; Karras et al., 2022). However, the only difference lies in sampling the noisy state xt from a Student-t distribution based perturbation kernel instead of a Gaussian distribution. Next, we discuss sampling from our proposed framework under discrete and continuous-time settings. 3.4 SAMPLING Discrete-time Sampling. For discrete-time settings, we can simply perform ancestral sampling from the learned reverse posterior distribution pθpxt t|xtq. Therefore, following simple re-parameterization, an ancestral sampling update can be specified as, xt t µθpxt, tq σtzt{?κt, z Np0, Idq, κ χ2pν dq{pν dq, Continuous-time Sampling. Due to recent advances in accelerating sampling in continuous-time diffusion processes (Pandey et al., 2024a; Zhang & Chen, 2023; Lu et al., 2022; Song et al., 2022a; Xu et al., 2023a), we reformulate discrete-time dynamics in heavy-tailed diffusions to the continuous time regime. More specifically, we present a family of continuous-time processes in the following result (Proof in App. A.5). Proposition 2. The posterior parameterization pθpxt t|xtq tdpµθpxt, tq, σ2 t Id, ν dq induces the following continuous-time dynamics, 9µt µt xt fpσt, 9σtq 9µt ȷ pxt µt Dθpxt, σtqq βptqgpσt, 9σtqd St, (11) where f : R ˆ R Ñ R and g : R ˆ R Ñ R are user-specified functions, βt P R is a scaling coefficient such that 1 σ2 12ptq σ2 t t βptqgpσt, 9σtq t 1 fpσt, 9σtq t,, where 9µt, 9σt denote the firstorder time-derivatives of the perturbation kernel parameters µt and σt respectively and the differential d St tdp0, dt, ν dq. Published as a conference paper at ICLR 2025 Algorithm 1: Training (t-EDM) 1: repeat 2: x0 ppx0q 3: σ Log Normalpπmean, πstdq 4: x x0 n, n tdp0, σ2Id, νq 5: σ σ a ν{pν 2q 6: Dθpx, σq cskippσqx coutpσq Fθpcinpσqx, cnoisepσqq 7: λpσq c 2 out pσq 8: Take gradient descent step on 9: θ λpσq }Dθpx, σq x0}2 ı 10: until converged Algorithm 2: Sampling (t-EDM) (µt 1, σt t) 1: sample x0 td 0, t2 0Id, ν 2: for i P t0, . . . , N 1u do 3: di Ð xi Dθpxi; tiq {ti 4: xi 1 Ð xi pti 1 tiqdi 5: if ti 1 0 then 6: di Ð xi 1 Dθpxi 1; ti 1q {ti 1 7: xi 1 Ð xi pti 1 tiq 1 8: end if 9: end for 10: return x N Figure 2: Training and Sampling algorithms for t-EDM (ν ą 2). Our proposed method requires minimal code updates (indicated with blue) over traditional Gaussian diffusion models and converges to the latter as ν Ñ 8. Based on the result in Proposition 2, it is possible to construct deterministic/stochastic samplers for heavytailed diffusions. It is worth noting that the SDE in Eqn. 11 implies adding Student-t stochastic noise during inference (Bollerslev, 1987). This is intuitive since the denoising distribution pθpxt 1|xtq is modeled as a Student-t distribution. Next, we provide specific instantiations of the generic sampler in Eq. 11. Sampler Instantiations. We instantiate the continuous-time SDE in Eqn. 11 by setting gpσt, 9σtq 0 and σ12ptq σtσt t. Consequently, fpσt, 9σtq 9σt σt . In this case, the SDE in Eqn. 11 reduces to an ODE, ȷ pxt µt Dθpxt, σtqq. (12) Summary. Overall, we compare between Gaussian diffusion and heavy-tailed diffusion models in Table 1. 3.5 SPECIFIC INSTANTIATIONS: T-EDM Karras et al. (2022) highlight several design choices during training and sampling, which significantly improve sample quality while reducing sampling budget for image datasets like CIFAR-10 (Krizhevsky, 2009) and Image Net (Deng et al., 2009). With a similar motivation, we reformulate the perturbation kernel as qpxt|x0q tdpsptqx0, sptq2σptq2Id, νq and denote the resulting diffusion model as t-EDM. Training. During training, we set the perturbation kernel, qpxt|x0q, parameters sptq 1, σptq σ Log Normalp Pmean, Pstdq. We parameterize the denoiser Dθpxt, σtq similar to Karras et al. (2022) with the difference that coefficients like cout additionally depend on ν. We include full derivations in Appendix A.6. Consequently, our denoising loss can be specified as follows: Lpθq 9 Ex0 ppx0q EσEn tdp0,σ2Id,νq λpσ, νq}Dθpx0 n, σq x0}2 2 , (13) where λpσ, νq is a weighting function set to λpσ, νq 1{coutpσ, νq2. Sampling. Interestingly, it can be shown that the ODE in Eqn. 12 is equivalent to the deterministic dynamics presented in Karras et al. (2022) (See Appendix A.7 for proof). Consequently, we choose sptq 1 and σptq t during sampling, further simplifying the dynamics in Eqn. 12 to dxt{dt pxt Dθpxt, tqq{t. We adopt the timestep discretization schedule and the choice of the numerical ODE solver (Heun s method (Ascher & Petzold, 1998)) directly from EDM. Figure 2 illustrates the ease of transitioning from a Gaussian diffusion framework (EDM) to t-EDM. Under standard settings, transitioning to t-EDM requires as few as two lines of code change, making our method readily compatible with existing diffusion implementations. Similarly, we can also construct heavy-tailed flows (Albergo et al., 2023; Lipman et al., 2023). We denote the resulting model as t-Flow (see App. B for details). Published as a conference paper at ICLR 2025 EDM (+INC) EDM (+PCP) t-EDM Figure 3: Sample 1-d histogram comparison between EDM and t-EDM on the test set for the Vertically Integrated Liquid (VIL) channel. t-EDM captures heavy-tailed behavior more accurately than other baselines. INC: Inverse CDF Normalization, PCP: Per-Channel Preconditioning 4 EXPERIMENTS Next, we demonstrate the effectiveness of the proposed heavy-tailed diffusion models on real-world weather data for both unconditional and conditional generation tasks. We include full experimental details in App. C. Datasets. We adopt the High-Resolution Rapid Refresh (HRRR) (Dowell et al., 2022) dataset, which is an operational archive of the US km-scale forecasting model. Among multiple dynamical variables in the dataset that exhibit heavy-tailed behavior, based on their dynamic range, we only consider the Vertically Integrated Liquid (VIL) and Vertical Wind Velocity at level 20 (denoted as w20) channels (see App. C.1 for more details). It is worth noting that the VIL and w20 channels have heavier right and left tails, respectively (See Fig. 6). Tasks and Metrics. We consider both unconditional and conditional generative tasks relevant to weather and climate science. For unconditional modeling, we aim to generate the VIL and w20 physical variables in the HRRR dataset. For conditional modeling, we aim to generatively predict the hourly evolution of the target variable for the next lead-time (τ 1) hour-ahead evolution of VIL and w20 based on information only at the current state time τ; see more details in the appendix and see Pathak et al. (2024) for discussion of why hour-ahead, km-scale atmospheric prediction is a stochastic physical task appropriate for conditional generative models. To quantify the empirical performance of unconditional modeling, we rely on comparing 1-d statistics of generated and train/test set samples. More specifically, for quantitative analysis, we report the Kurtosis Ratio (KR), the Skewness Ratio (SR), and the Kolmogorov-Smirnov (KS)-2 sample statistic (at the tails) between the generated and train/test set samples. For qualitative analysis, we compare 1-d histograms between generated and train/test set samples. For the conditional task, we adopt standard probabilistic prediction score metrics such as the Continuous Ranked Probability Score (CRPS), the Root-Mean Squared Error (RMSE), and the skill-spread ratio (SSR); see, e.g., Mardani et al. (2024); Srivastava et al. (2023). A more detailed explanation of our evaluation protocol is provided in App. C. Methods and Baselines. In addition to standard diffusion (EDM (Karras et al., 2022)) and flow models (Albergo et al., 2023) based on Gaussian priors, we introduce two additional baselines that are variants of EDM. To account for the high dynamic-range often exhibited by heavy-tailed distributions, we include Inverse CDF Normalization (INC) as an alternative data preprocessing step to z-score normalization. Using the former reduces dynamic range significantly and can make the data distribution closer to Gaussian. We denote this preprocessing scheme combined with standard EDM training as EDM + INC. Alternatively, we could instead modulate the noise levels used during EDM training as a function of the dynamic range of the input channel while keeping the data preprocessing unchanged. The main intuition is to use more heavy-tailed noise for large values. We denote this modulating scheme as Per-Channel Preconditioning (PCP) and denote the resulting baseline as EDM + PCP. We elaborate on these baselines in more detail in App. C.1.2 Published as a conference paper at ICLR 2025 VIL (Train) VIL (Test) w20 (Train) w20 (Test) Method ν KR Ó SR Ó KS Ó KR Ó SR Ó KS Ó ν KR Ó SR Ó KS Ó KR Ó SR Ó KS Ó Baselines EDM 8 210.11 10.79 0.997 45.35 5.23 0.991 8 12.59 0.89 0.991 5.01 0.38 0.978 +INC 8 11.33 2.29 0.987 1.70 0.74 0.95 8 1.80 0.18 0.909 0.23 0.13 0.763 +PCP 8 2.12 0.72 0.800 0.31 0.09 0.522 8 2.17 0.70 0.838 0.40 0.24 0.648 Ours t-EDM 3 1.06 0.43 0.431 0.54 0.23 0.114 3 2.44 0.65 0.683 0.52 0.21 0.286 t-EDM 5 29.66 4.07 0.955 5.73 1.68 0.888 5 8.55 1.77 0.895 3.22 1.03 0.774 t-EDM 7 24.35 4.14 0.959 4.57 1.72 0.908 7 7.03 1.58 0.82 2.55 0.89 0.622 Table 2: t-EDM outperforms standard diffusion models for unconditional generation on the HRRR dataset. For all metrics, lower is better. Values in bold indicate the best results in a column. VIL (Train) VIL (Test) w20 (Train) w20 (Test) Method ν KR Ó SRÓ KSÓ KR Ó SR Ó KSÓ ν KR Ó SR Ó KS Ó KR Ó SR Ó KS Ó Baselines Gaussian Flow 8 0.46 0.09 0.897 0.67 0.52 0.704 8 2.03 0.36 0.294 0.34 0.01 0.384 Ours t-Flow 3 1.39 0.37 0.711 0.47 0.27 0.275 5 1.08 0.21 0.333 0.07 0.42 0.512 t-Flow 5 3.30 0.75 0.857 0.05 0.07 0.633 7 3.24 0.36 0.259 0.87 0.01 0.300 t-Flow 7 3.36 0.84 0.844 0.04 0.02 0.603 9 5.47 0.41 0.478 1.86 0.034 0.289 Table 3: t-Flow outperforms standard Gaussian flows for unconditional generation on the HRRR dataset. For all metrics, lower is better. Values in bold indicate the best results in a column. 4.1 UNCONDITIONAL GENERATION We assess the effectiveness of different methods on unconditional modeling for the VIL and w20 channels in the HRRR dataset. Fig. 3 qualitatively compares 1-d histograms of sample intensities between different methods for the VIL channel. We make the following key observations. Firstly, though EDM (with additional tricks like noise conditioning) can improve tail coverage, t-EDM covers a broader range of extreme values in the test set. Secondly, in addition to better dynamic range coverage, t-EDM qualitatively performs much better in capturing the density assigned to intermediate intensity levels under the model. We note similar observations from our quantitative results in Table 2, where t-EDM outperforms other baselines on the KS metric, implying our model exhibits better tail estimation over competing baselines for both the VIL and w20 channels. More importantly, unlike traditional Gaussian diffusion models like EDM, t-EDM enables controllable tail estimation by varying ν, which could be useful when modeling a combination of channels with diverse statistical properties. Lastly, we note that while EDM combined with improved preprocessing (INC) or preconditioning (PCP) outperforms t-EDM in some cases, these techniques can also be readily integrated with t-EDM. Therefore, the main point of comparison in Table 2 should be primarily between t-EDM and the standard EDM baseline (row 1). We present similar quantitative results for t-Flow in Table 3 where the advantages of t-Flow over Gaussian Flows for heavy-tailed modeling are quite apparent. We also present additional results for unconditional modeling in App. C.1.7. 4.2 CONDITIONAL GENERATION Next, we consider the task of conditional modeling, where we aim to predict the hourly evolution of a target variable for the next lead time (τ δτ) based on the current state at time τ with δτ 1hr . Table 4 illustrates the performance of EDM and t-EDM on this task for the VIL and w20 channels. We make the following key observations. Firstly, for both channels, t-EDM exhibits better CRPS and SSR scores, implying better probabilistic forecast skills and ensemble than EDM. Moreover, while t-EDM exhibits under-dispersion for VIL, while it is well-calibrated for w20, with its SSR close to an ideal score of 1. On the contrary, the baseline EDM model exhibits under-dispersion for both channels, thus implying overconfident predictions. Secondly, in addition to better calibration, t-EDM is better at tail estimation (as measured by the KS statistic) for the underlying conditional distribution. Lastly, we notice that different values of the parameter ν are optimal for Published as a conference paper at ICLR 2025 VIL (Test) w20 (Test) Method ν CRPS Ó RMSE Ó SSR (Ñ 1) KS Ó ν CRPS Ó RMSE Ó SSR (Ñ 1) KS Ó Baselines EDM 8 1.696 4.473 0.203 0.715 8 0.304 0.664 0.865 0.345 Ours t-EDM 3 1.649 4.526 0.255 0.419 3 0.295 0.734 1.045 0.111 t-EDM 5 1.609 4.361 0.305 0.665 5 0.301 0.674 0.901 0.323 Table 4: t-EDM outperforms EDM for conditional next frame prediction for the HRRR dataset. Values in bold indicate the best results in each column. We note that VIL has a higher dynamic range over w20, and thus, the gains for VIL are more apparent (see hist plots in Fig. 6). (Ñ 1) indicates values near 1 are better. different channels, which suggests a more data-driven approach to learning the optimal ν directly. We present additional results for conditional modeling in App. C.2. 5 DISCUSSION AND THEORETICAL INSIGHTS To conclude, we propose a framework for constructing heavy-tailed diffusion models and demonstrate their effectiveness over traditional diffusion models on unconditional and conditional generation tasks for a highresolution weather dataset. Here, we highlight some theoretical connections that help gain insights into the effectiveness of our proposed framework while establishing connections with prior work. Figure 4: Variation of the mean and standard deviation of the ratio pν d1q{pν dq with ν across diffusion sampling trajectory for the toy dataset. As ν decreases, the mean ratio and its standard deviation increase, leading to large score multiplier weights. Exploring distribution tails during sampling. The ODE in Eq. 12 can re-formulated as, µt xt σ2 t ν d1 1 ν d ȷ x log ppxt, tq, (14) where d1 1 p1{σ2 t q}xt µt Dθpxt, σtq}2 2. By formulating the ODE in terms of the score function, we can gain some intuition into the effectiveness of our model in modeling heavy-tailed distributions. Figure 4 illustrates the variation of the mean and variance of the multiplier pν d1 1q{pν dq along the diffusion trajectory across 1M samples generated from our toy models. Interestingly, as the value of ν decreases, the mean and variance of this multiplier increase significantly, which leads to large score multiplier weights. We hypothesize that this behavior allows our proposed model to explore more diverse regions during sampling (more details in App. A.10). Enabling efficient tail coverage during training. The optimization objective in Eq. 9 has several connections with robust statistical estimators. More specifically, it can be shown that (proof in App. A.11), θDγpq } pθq ż qpxq ˆ pθpxq γ θ log pθpxq E pθpxqr θ log pθpxqs dx, where q and pθ denote the forward (qpxt t|xt, x0q) and reverse diffusion posteriors (pθpxt t|xtq), respectively. Intuitively, the coefficient γ weighs the likelihood gradient, θ log pθpxq, and can be set accordingly to ignore or consider outliers when modeling the data distribution. Specifically, when γ ą 1, the model would learn to ignore outliers (Futami et al., 2018; Fujisawa & Eguchi, 2008; Basu et al., 1998) since data points on the tails would be assigned low likelihood. On the contrary, a negative value of γ (as is the case in this work since we set γ 2{pν dq), the model can assign more weights to capture these extreme values. We discuss some other connections to prior work in heavy-tailed generative modeling and more recent work in diffusion models in App. F.1 and some limitations of our approach in App. F.2. Published as a conference paper at ICLR 2025 ACKNOWLEDGEMENT Stephan Mandt acknowledges support from the National Science Foundation (NSF) under an NSF CAREER Award IIS-2047418 and IIS-2007719, the NSF LEAP Center, by the Department of Energy under grant DESC0022331, the IARPA WRIVA program, the Hasso Plattner Research Center at UCI, the Chan Zuckerberg Initiative, and gifts from Qualcomm and Disney. REPRODUCIBILITY STATEMENT We include proofs for all theoretical results introduced in the main text in Appendix A. We describe our complete experimental setup (including data processing steps, model specification for training and inference, description of evaluation metrics, and extended experimental results) in Appendix C. ETHICS STATEMENT We develop a generative framework for modeling heavy-tailed distributions and demonstrate its effectiveness for scientific applications. In this context, we do not think our model poses a risk of misinformation or other ethical biases associated with large-scale image synthesis models. However, we would like to point out that similar to other generative models, our model can sometimes hallucinate predictions for certain channels, which could impact downstream applications like weather forecasting. Michael S Albergo, Nicholas M Boffi, and Eric Vanden-Eijnden. Stochastic interpolants: A unifying framework for flows and diffusions. ar Xiv preprint ar Xiv:2303.08797, 2023. D. F. Andrews and C. L. Mallows. Scale mixtures of normal distributions. Journal of the Royal Statistical Society. Series B (Methodological), 36(1):99 102, 1974. ISSN 00359246. URL http://www.jstor. org/stable/2984774. Uri M. Ascher and Linda R. Petzold. Computer Methods for Ordinary Differential Equations and Differential-Algebraic Equations. Society for Industrial and Applied Mathematics, Philadelphia, PA, 1998. doi: 10.1137/1.9781611971392. URL https://epubs.siam.org/doi/abs/10.1137/1. 9781611971392. Ayanendranath Basu, Ian R Harris, Nils L Hjort, and MC Jones. Robust and efficient estimation by minimising a density power divergence. Biometrika, 85(3):549 559, 1998. Tim Bollerslev. A conditionally heteroskedastic time series model for speculative prices and rates of return. The Review of Economics and Statistics, 69(3):542 547, 1987. ISSN 00346535, 15309142. URL http://www.jstor.org/stable/1925546. Steve Brooks, Andrew Gelman, Galin Jones, and Xiao-Li Meng. Handbook of Markov Chain Monte Carlo. Chapman and Hall/CRC, May 2011. ISBN 9780429138508. doi: 10.1201/b10905. URL http://dx.doi.org/10.1201/b10905. Tianfeng Chai and Roland R Draxler. Root mean square error (rmse) or mean absolute error (mae)? arguments against avoiding rmse in the literature. Geoscientific Model Development, 7(3):1247 1250, 2014. Published as a conference paper at ICLR 2025 Hyungjin Chung, Jeongsol Kim, Michael Thompson Mccann, Marc Louis Klasky, and Jong Chul Ye. Diffusion posterior sampling for general noisy inverse problems. In The Eleventh International Conference on Learning Representations, 2022. Jacob Deasy, Nikola Simidjievski, and Pietro Liò. Heavy-tailed denoising score matching. ar Xiv preprint ar Xiv:2112.09788, 2021. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248 255, 2009. doi: 10.1109/CVPR.2009.5206848. Peng Ding. On the conditional distribution of the multivariate t distribution, 2016. URL https://arxiv. org/abs/1604.00561. Tim Dockhorn, Arash Vahdat, and Karsten Kreis. Score-based generative modeling with critically-damped langevin diffusion, 2022. URL https://arxiv.org/abs/2112.07068. Alexey Dosovitskiy and Thomas Brox. Generating images with perceptual similarity metrics based on deep networks, 2016. URL https://arxiv.org/abs/1602.02644. David C. Dowell, Curtis R. Alexander, Eric P. James, Stephen S. Weygandt, Stanley G. Benjamin, Geoffrey S. Manikin, Benjamin T. Blake, John M. Brown, Joseph B. Olson, Ming Hu, Tatiana G. Smirnova, Terra Ladwig, Jaymes S. Kenyon, Ravan Ahmadov, David D. Turner, Jeffrey D. Duda, and Trevor I. Alcott. The high-resolution rapid refresh (hrrr): An hourly updating convection-allowing forecast model. part i: Motivation and system description. Weather and Forecasting, 37(8):1371 1395, 2022. doi: 10.1175/ WAF-D-21-0151.1. URL https://journals.ametsoc.org/view/journals/wefo/37/8/ WAF-D-21-0151.1.xml. Shinto Eguchi. Chapter 2 - pythagoras theorem in information geometry and applications to generalized linear models. In Angelo Plastino, Arni S.R. Srinivasa Rao, and C.R. Rao (eds.), Information Geometry, volume 45 of Handbook of Statistics, pp. 15 42. Elsevier, 2021. doi: https://doi.org/10.1016/bs.host.2021.06.001. URL https://www.sciencedirect.com/science/article/pii/S0169716121000225. Patrick Esser, Sumith Kulal, Andreas Blattmann, Rahim Entezari, Jonas Müller, Harry Saini, Yam Levi, Dominik Lorenz, Axel Sauer, Frederic Boesel, Dustin Podell, Tim Dockhorn, Zion English, Kyle Lacey, Alex Goodwin, Yannik Marek, and Robin Rombach. Scaling rectified flow transformers for high-resolution image synthesis, 2024. URL https://arxiv.org/abs/2403.03206. Hironori Fujisawa and Shinto Eguchi. Robust parameter estimation with a small bias against heavy contamination. Journal of Multivariate Analysis, 99(9):2053 2081, 2008. ISSN 0047-259X. doi: https://doi.org/ 10.1016/j.jmva.2008.02.004. URL https://www.sciencedirect.com/science/article/ pii/S0047259X08000456. Futoshi Futami, Issei Sato, and Masashi Sugiyama. Variational inference based on robust divergences, 2018. URL https://arxiv.org/abs/1710.06595. Gaby Joanne Gründemann, Nick van de Giesen, Lukas Brunner, and Ruud van der Ent. Rarest rainfall events will see the greatest relative increase in magnitude under future climate change. Communications Earth &; Environment, 3(1), October 2022. ISSN 2662-4435. doi: 10.1038/s43247-022-00558-8. URL http://dx.doi.org/10.1038/s43247-022-00558-8. H. Guo, J.-C. Golaz, L. J. Donner, B. Wyman, M. Zhao, and P. Ginoux. Clubb as a unified cloud parameterization: Opportunities and challenges. Geophysical Research Letters, 42(11):4540 4547, 2015. doi: https://doi.org/10.1002/2015GL063672. URL https://agupubs.onlinelibrary.wiley. com/doi/abs/10.1002/2015GL063672. Published as a conference paper at ICLR 2025 Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium, 2018. URL https: //arxiv.org/abs/1706.08500. Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. Advances in Neural Information Processing Systems, 33:6840 6851, 2020. M.F. Hutchinson. A stochastic estimator of the trace of the influence matrix for laplacian smoothing splines. Communications in Statistics - Simulation and Computation, 19(2):433 450, 1990. doi: 10.1080/ 03610919008812866. URL https://doi.org/10.1080/03610919008812866. Priyank Jaini, Ivan Kobyzev, Yaoliang Yu, and Marcus Brubaker. Tails of lipschitz triangular flows, 2020. URL https://arxiv.org/abs/1907.04481. Tero Karras, Miika Aittala, Timo Aila, and Samuli Laine. Elucidating the design space of diffusion-based generative models, 2022. URL https://arxiv.org/abs/2206.00364. Juno Kim, Jaehyuk Kwon, Mincheol Cho, Hyunjong Lee, and Joong-Ho Won. $t 3$-variational autoencoder: Learning heavy-tailed data with student s t and power divergence. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=Rz Nl ECeo OB. Diederik P Kingma and Max Welling. Auto-encoding variational bayes, 2022. URL https://arxiv. org/abs/1312.6114. Alex Krizhevsky. Learning multiple layers of features from tiny images. pp. 32 33, 2009. URL https: //www.cs.toronto.edu/~kriz/learning-features-2009-TR.pdf. Mike Laszkiewicz, Johannes Lederer, and Asja Fischer. Marginal tail-adaptive normalizing flows, 2022. URL https://arxiv.org/abs/2206.10311. Yaron Lipman, Ricky T. Q. Chen, Heli Ben-Hamu, Maximilian Nickel, and Matthew Le. Flow matching for generative modeling. In International Conference on Learning Representations, 2023. URL https: //openreview.net/forum?id=Pqv MRDCJT9t. Xingchao Liu, Chengyue Gong, and Qiang Liu. Flow straight and fast: Learning to generate and transfer data with rectified flow, 2022. URL https://arxiv.org/abs/2209.03003. Cheng Lu, Yuhao Zhou, Fan Bao, Jianfei Chen, Chongxuan Li, and Jun Zhu. Dpm-solver: A fast ode solver for diffusion probabilistic model sampling in around 10 steps, 2022. URL https://arxiv.org/ abs/2206.00927. Morteza Mardani, Jiaming Song, Jan Kautz, and Arash Vahdat. A variational perspective on solving inverse problems with diffusion models. In The Twelfth International Conference on Learning Representations, 2023. Morteza Mardani, Noah Brenowitz, Yair Cohen, Jaideep Pathak, Chieh-Yu Chen, Cheng-Chin Liu, Arash Vahdat, Mohammad Amin Nabian, Tao Ge, Akshay Subramaniam, Karthik Kashinath, Jan Kautz, and Mike Pritchard. Residual corrective diffusion modeling for km-scale atmospheric downscaling, 2024. URL https://arxiv.org/abs/2309.15214. Frank J. Massey. The kolmogorov-smirnov test for goodness of fit. Journal of the American Statistical Association, 46(253):68 78, 1951. ISSN 01621459, 1537274X. URL http://www.jstor.org/ stable/2280095. Published as a conference paper at ICLR 2025 Radford M. Neal. Slice sampling. The Annals of Statistics, 31(3):705 767, 2003. doi: 10.1214/aos/ 1056562461. URL https://doi.org/10.1214/aos/1056562461. Kushagra Pandey and Stephan Mandt. A complete recipe for diffusion generative models, 2023. URL https://arxiv.org/abs/2303.01748. Kushagra Pandey, Avideep Mukherjee, Piyush Rai, and Abhishek Kumar. Diffusevae: Efficient, controllable and high-fidelity generation from low-dimensional latents, 2022. URL https://arxiv.org/abs/ 2201.00308. Kushagra Pandey, Maja Rudolph, and Stephan Mandt. Efficient integrators for diffusion generative models. In The Twelfth International Conference on Learning Representations, 2024a. URL https: //openreview.net/forum?id=q A4fox O5Gf. Kushagra Pandey, Ruihan Yang, and Stephan Mandt. Fast samplers for inverse problems in iterative refinement models, 2024b. URL https://arxiv.org/abs/2405.17673. Jaideep Pathak, Yair Cohen, Piyush Garg, Peter Harrington, Noah Brenowitz, Dale Durran, Morteza Mardani, Arash Vahdat, Shaoming Xu, Karthik Kashinath, and Michael Pritchard. Kilometer-scale convection allowing model emulation using generative diffusion modeling, 2024. URL https://arxiv.org/ abs/2408.10958. Dustin Podell, Zion English, Kyle Lacey, Andreas Blattmann, Tim Dockhorn, Jonas Müller, Joe Penna, and Robin Rombach. Sdxl: Improving latent diffusion models for high-resolution image synthesis, 2023. URL https://arxiv.org/abs/2307.01952. Danilo Jimenez Rezende and Shakir Mohamed. Variational inference with normalizing flows, 2016. URL https://arxiv.org/abs/1505.05770. Robin Rombach, Andreas Blattmann, Dominik Lorenz, Patrick Esser, and Björn Ommer. High-resolution image synthesis with latent diffusion models, 2022. URL https://arxiv.org/abs/2112.10752. Francisco R Ruiz, Titsias RC AUEB, David Blei, et al. The generalized reparameterization gradient. Advances in neural information processing systems, 29, 2016. Mehdi SM Sajjadi, Olivier Bachem, Mario Lucic, Olivier Bousquet, and Sylvain Gelly. Assessing generative models via precision and recall. Advances in neural information processing systems, 31, 2018. Dario Shariatian, Umut Simsekli, and Alain Durmus. Denoising lévy probabilistic models, 2024. URL https://arxiv.org/abs/2407.18609. Raghav Singhal, Mark Goldstein, and Rajesh Ranganath. Where to diffuse, how to diffuse, and how to get back: Automated learning for multivariate diffusions, 2023. URL https://arxiv.org/abs/2302. 07261. John Skilling. The Eigenvalues of Mega-dimensional Matrices, pp. 455 466. Springer Netherlands, Dordrecht, 1989. ISBN 978-94-015-7860-8. doi: 10.1007/978-94-015-7860-8_48. URL https://doi.org/10. 1007/978-94-015-7860-8_48. Jascha Sohl-Dickstein, Eric Weiss, Niru Maheswaranathan, and Surya Ganguli. Deep unsupervised learning using nonequilibrium thermodynamics. In International Conference on Machine Learning, pp. 2256 2265. PMLR, 2015. Jiaming Song, Chenlin Meng, and Stefano Ermon. Denoising diffusion implicit models, 2022a. URL https://arxiv.org/abs/2010.02502. Published as a conference paper at ICLR 2025 Jiaming Song, Arash Vahdat, Morteza Mardani, and Jan Kautz. Pseudoinverse-guided diffusion models for inverse problems. In International Conference on Learning Representations, 2022b. Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2020. Yang Song, Conor Durkan, Iain Murray, and Stefano Ermon. Maximum likelihood training of score-based diffusion models, 2021. URL https://arxiv.org/abs/2101.09258. Prakhar Srivastava, Ruihan Yang, Gavin Kerrigan, Gideon Dresdner, Jeremy Mc Gibbon, Christopher Bretherton, and Stephan Mandt. Precipitation downscaling with spatiotemporal video diffusion. ar Xiv preprint ar Xiv:2312.06071, 2023. Pascal Vincent. A connection between score matching and denoising autoencoders. Neural Computation, 23 (7):1661 1674, 2011. doi: 10.1162/NECO_a_00142. Daniel S Wilks. Statistical methods in the atmospheric sciences, volume 100. Academic press, 2011. Zhisheng Xiao, Karsten Kreis, and Arash Vahdat. Tackling the generative learning trilemma with denoising diffusion gans. In International Conference on Learning Representations. Yilun Xu, Mingyang Deng, Xiang Cheng, Yonglong Tian, Ziming Liu, and Tommi S. Jaakkola. Restart sampling for improving generative processes. In Thirty-seventh Conference on Neural Information Processing Systems, 2023a. URL https://openreview.net/forum?id=w Fuemocy HZ. Yilun Xu, Ziming Liu, Yonglong Tian, Shangyuan Tong, Max Tegmark, and Tommi Jaakkola. Pfgm++: Unlocking the potential of physics-inspired generative models. In International Conference on Machine Learning, pp. 38566 38591. PMLR, 2023b. Eunbi Yoon, Keehun Park, Sungwoong Kim, and Sungbin Lim. Score-based generative models with lévy processes. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=0Wp3VHX0Gm. Qinsheng Zhang and Yongxin Chen. Fast sampling of diffusion models with exponential integrator, 2023. URL https://arxiv.org/abs/2204.13902. 1 Introduction 1 2 Background 3 2.1 Diffusion Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 2.2 Student-t Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 3 Heavy-Tailed Diffusion Models 3 3.1 Noising Process Design with Student-t Distributions. . . . . . . . . . . . . . . . . . . . . . 4 3.2 Parameterization of the Reverse Posterior . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 Published as a conference paper at ICLR 2025 3.3 Training with Power Divergences . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 3.4 Sampling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 3.5 Specific Instantiations: t-EDM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 4 Experiments 8 4.1 Unconditional Generation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 4.2 Conditional Generation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 5 Discussion and Theoretical Insights 10 A Proofs 17 A.1 Derivation of the Perturbation Kernel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 A.2 On the Posterior Parameterization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 A.3 Proof of Proposition 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 A.4 Derivation of the simplified denoising loss . . . . . . . . . . . . . . . . . . . . . . . . . . . 21 A.5 Proof of Proposition 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 A.6 Deriving the Denoiser Preconditioner for t-EDM . . . . . . . . . . . . . . . . . . . . . . . 24 A.7 Equivalence with the EDM ODE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 A.8 Conditional Vector Field for t-Flow . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 A.9 Connection to Denoising Score Matching . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 A.10 ODE Reformulation and Connections to Inverse Problems. . . . . . . . . . . . . . . . . . . 27 A.11 Connections to Robust Statistical Estimators . . . . . . . . . . . . . . . . . . . . . . . . . . 28 B Extension to Flows 30 C Experimental Setup 31 C.1 Unconditional Modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 C.1.1 HRRR Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 C.1.2 Baselines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 C.1.3 Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 C.1.4 Denoiser Architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 C.1.5 Training . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 C.1.6 Sampling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 C.1.7 Extended Results on Unconditional Modeling . . . . . . . . . . . . . . . . . . . . . 35 C.2 Conditional Modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 Published as a conference paper at ICLR 2025 C.2.1 HRRR Dataset for Conditional Modeling . . . . . . . . . . . . . . . . . . . . . . . 35 C.2.2 Baselines . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 C.2.3 Denoiser Architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 C.2.4 Training . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 C.2.5 Sampling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 C.2.6 Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37 C.2.7 Extended Results on Conditional Modeling . . . . . . . . . . . . . . . . . . . . . . 37 C.3 Toy Experiments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 D Optimal Noise Schedule Design 39 D.1 Design for EDM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 D.2 Extension to t-EDM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 D.3 Extension to Correlated Gaussian Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 E Log-Likelihood for t-EDM 42 F Discussion and Limitations 43 F.1 Related Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 F.2 Limitations and Future Work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45 A.1 DERIVATION OF THE PERTURBATION KERNEL Proof. The proof is an adaptation from Ding (2016) and included here for completeness. We define the joint state x rxt; xt ts. Consequently, qpx|x0q t2dpµ, Σ, νq, µ rµt; µt tsx0, Σ ˆ σ2 t σ2 12ptq σ2 21ptq σ2 t t By re-parameterization of the joint Student-t distribution qpx|x0q, we have, x µ Σ 1 2 z{?κ, z Np0, Idq and κ χ2pνq{ν (15) This implies that the conditional distribution qpx|x0, κq Npµ, Σ{κq. Therefore, following properties of Gaussian distributions, qpxt|x0, κq Npµtx0, σ2 t {κIdq. Therefore, from reparameterization, xt|κ µtx0 σtz{?κ (16) which implies that qpxt|x0q tdpµtx0, σ2 t Id, νq. This completes the proof. A.2 ON THE POSTERIOR PARAMETERIZATION The perturbation kernel qpxt|x0q for Student-t diffusions is parameterized as, qpxt|x0q tdpµtx0, σ2 t Id, νq (17) Published as a conference paper at ICLR 2025 Using re-parameterization, xt µtx0 σt ϵ ?κ, ϵ Np0, Idq, κ χ2pνq{ν (18) During inference, given a noisy state xt, we have the following estimation problem, xt µt Erx0|xts σt E ϵ ?κ|xt (19) Therefore, the task of denoising can be posed as either estimating Erx0|xts or E ϵ ?κ|xt . With this motivation, the posterior pθpxt t|xtq can be parameterized appropriately. Recall the form of the forward posterior qpxt t|xt, x0q tdp µt, ν d1 ν d σ2 t Id, ν dq (20) µt µt tx0 σ2 21ptq σ2 t pxt µtx0q, σ2 t σ2 t t σ2 21ptqσ2 12ptq σ2 t where d1 1 σ2 t }xt µtx0}2. Further simplifying the mean µt, µtpxt, x0, tq µt tx0 σ2 21ptq σ2 t pxt µtx0q (22) µt tx0 σ2 21ptq σ2 t xt σ2 21ptq σ2 t µtx0 (23) µt t µt σ2 21ptq x0 σ2 21ptq σ2 t xt (24) Therefore, the mean µθpxt, tq of the reverse posterior pθpxt t|xtq can be parameterized as, µθpxt, tq µt t µt σ2 21ptq Erx0|xts σ2 21ptq σ2 t xt (25) µt t µt σ2 21ptq Dθpxt, σtq σ2 21ptq σ2 t xt (26) where Erx0|xts is learned using a parametric estimator Dθpxt, σtq. This corresponds to the x0-prediction parameterization presented in Eq. 7 in the main text. Alternatively, From Eqn. 19, µθpxt, tq µt t µt σ2 21ptq Erx0|xts σ2 21ptq σ2 t xt (27) µt t µt σ2 21ptq xt σt E ϵ ?κ|xt σ2 21ptq σ2 t xt (28) µt t µt σ2 21ptq E ϵ ?κ|xt (29) µt t µt σ2 21ptq ϵθpxt, σtq (30) where E ϵ ?κ|xt is learned using a parametric estimator ϵθpxt, σtq. This corresponds to the ϵ-prediction parameterization (Ho et al., 2020). Published as a conference paper at ICLR 2025 A.3 PROOF OF PROPOSITION 1 We restate the proposition here for convenience, Proposition 1. For arbitrary distributions q and p, in the limit of γ Ñ 0, Dγpq } pq converges to DKLpq } pq. Consequently, the objective in Eqn. 9 converges to the DDPM objective stated in Eqn. 1. Proof. We present our proof in two parts: 1. Firstly, we establish the following relation between the γ-Power Divergence and the KL-Divergence between two distributions q and p. Dγpq } pq DKLpq } pq Opγq (31) 2. Next, for the choice of γ 2 ν d, we show that in the limit of γ Ñ 0, the optimization objective in Eqn. 9 converges to the optimization objective in Eqn. 1 Relation between DKLpq } pq and Dγpq } pq. The γ-Power Divergence as stated in (Kim et al., 2024) assumes the following form: Dγpq } pq 1 Cγpq, pq Hγpqq (32) where γ P p 1, 0q Y p0, 8q, and, Hγppq }p}1 γ ˆ ż ppxq1 γdx 1 1 γ Cγpq, pq ż qpxq ˆ ppxq For subsequent analysis, we assume γ Ñ 0. Under this assumption, we simplify Hγpqq as follows. By definition, Hγpqq }q}1 γ ˆ ż qpxq1 γdx 1 1 γ (34) ˆ ż qpxqqpxqγdx 1 1 γ (35) ˆ ż qpxq exppγ log qpxqqdx 1 1 γ (36) ˆ ż qpxq 1 γ log qpxq Opγ2q dx 1 1 γ (37) ˆ ż qpxqdx γ ż qpxq log qpxqdx Opγ2q 1 1 γ (38) ˆ 1 γ ż qpxq log qpxqdx Opγ2q 1 1 γ (39) Using the approximation p1 δxqα 1 αδx for a small δ in the above equation, we have, Hγpqq }q}1 γ ˆ 1 γ 1 γ ż qpxq log qpxqdx Opγq ı (40) ˆ 1 γp1 γq ż qpxq log qpxqdx Opγq ı (41) Published as a conference paper at ICLR 2025 where we have used the approximation 1 1 γ 1 γ in the above equation. This is justified since γ is assumed to be small enough. Therefore, we have, Hγpqq }q}1 γ ˆ 1 γ ż qpxq log qpxqdx Opγ2q (42) Similarly, we now obtain an approximation for the power-cross entropy as follows. By definition, Cγpq, pq ż qpxq ˆ ppxq ż qpxq ˆ 1 γ log ˆ ppxq Opγ2q dx (44) ˆ ż qpxqdx γ ż qpxq log ˆ ppxq ˆ 1 γ ż qpxq log ppxqdx γ ż qpxq log }p}1 γ dx Opγ2q (46) From Eqn. 42, it follows that, }p}1 γ ˆ 1 γ ż ppxq log ppxqdx Opγ2q (47) log }p}1 γ log ˆ 1 γ ż ppxq log ppxqdx Opγ2q (48) γ ż ppxq log ppxqdx (49) where the above result follows from the logarithmic series and ignores the terms of order Opγ2q or higher. Plugging the approximation in Eqn. 49 in Eqn. 46, we have, Cγpq, pq ˆ 1 γ ż qpxq log ppxqdx γ ż qpxq log }p}1 γ dx Opγ2q (50) ˆ 1 γ ż qpxq log ppxqdx Opγ2q (51) Dγpq } pq 1 Cγpq, pq Hγpqq (52) γ ˆ ż qpxq log qpxqdx ż qpxq log ppxqdx Opγ2q (53) ż qpxq log qpxq ppxqdx Opγq (54) DKLpq } pq Opγq (55) This establishes the relationship between the KL and γ-Power divergence between two distributions. Therefore, for two distributions q and p, the difference in the magnitude of DKLpq } pq and Dγpq } pq is of the order of Opγq. In the limit of γ Ñ 0, the Dγpq } pq Ñ DKLpq } pq. This concludes the first part of our proof. Published as a conference paper at ICLR 2025 Equivalence between the objectives under γ Ñ 0. For γ 2 γ d and a finite-dimensional dataset with x0 P Rd, it follows that γ Ñ 0 implies ν Ñ 8. Moreover, in the limit of ν Ñ 8, the multivariate Student-t distribution converges to a Gaussian distribution. As already shown in the previous part, under this limit, Dγpq } pq converges to DKLpq } pq. Therefore, under this limit, the optimization objective in Eqn. 9 converges to the standard DDPM objective in Eqn. 1. This completes the proof. A.4 DERIVATION OF THE SIMPLIFIED DENOISING LOSS Here, we derive the simplified denoising loss presented in Eq. 10 in the main text. We specifically consider the term Dγpqpxt t|xt, x0q } pθpxt t|xtqq in Eq. 9. The γ-power divergence between two Student-t distributions is given by, Dγrqν||pνs 1 γ 1 γ ν,d 1 d ν 2 γ 1 γ |Σ0| γ 2p1 γq 1 d ν 2 |Σ1| γ 2p1 γq 1 1 ν 2tr Σ 1 1 Σ0 1 ν pµ0 µ1q JΣ 1 1 pµ0 µ1q ı . (56) where γ 2 ν d. Recall, the definitions of the forward denoising posterior, qpxt t|xt, x0q tdp µt, ν d1 ν d σ2 t Id, ν dq (57) µt µt tx0 σ2 21ptq σ2 t pxt µtx0q, σ2 t σ2 t t σ2 21ptqσ2 12ptq σ2 t and the reverse denoising posterior, pθpxt t|xtq tdpµθpxt, tq, σ2 t Id, ν dq (59) where the denoiser mean µθpxt, tq is further parameterized as follows: µθpxt, tq σ2 21ptq σ2 t xt µt t σ2 21ptq σ2 t µt ı Dθpxt, σtq (60) Since we only parameterize the mean of the reverse posterior, the majority of the terms in the γ-power divergence are independent of θ and can be ignored (or treated as scalar coefficients). Therefore, Dγpqpxt t|xt, x0q } pθpxt t|xtqq9p µt µθpxt, tqq Jp µt µθpxt, tqq (61) 9} µt µθpxt, tq}2 2 (62) 9 µt t σ2 21ptq σ2 t µt ı2 }x0 Dθpxt, σtq}2 2 (63) For better sample quality, it is common to ignore the scalar multiple in prior works (Ho et al., 2020; Song et al., 2020). Therefore, ignoring the time-dependent scalar multiple, Dγpqpxt t|xt, x0q } pθpxt t|xtqq 9}x0 Dθpxt, σtq}2 2 (64) Therefore, the final loss function Lθ can be stated as, Lpθq Ex0 ppx0q Et pptq Eϵ Np0,Idq Eκ 1 ν χ2pνq Dθ µtx0 σt ϵ ?κ, σt x0 2 Published as a conference paper at ICLR 2025 A.5 PROOF OF PROPOSITION 2 We restate the proposition here for convenience, Proposition 2. The posterior parameterization in Eqn. 6 induces the following continuous-time dynamics, 9µt µt xt fpσt, 9σtq 9µt ȷ pxt µt Dθpxt, tqq βptqgpσt, 9σtqd St (66) where f : R ˆ R Ñ R and g : R ˆ R Ñ R are scalar-valued functions, βt P R is a scaling coefficient such that the following condition holds, 1 σ2 12ptq σ2 t t βptqgpσt, 9σtq t 1 fpσt, 9σtq t (67) where 9µt, 9σt denote the first-order time-derivatives of the perturbation kernel parameters µt and σt respectively and the differential d St tdp0, dt, ν dq. Proof. We start by writing a single sampling step from our learned posterior distribution. Recall pθpxt t|xtq tdpµθpxt, tq, σ2 t Id, ν dq (68) where (using the ϵ-prediction parameterization in App. A.2), µθpxt, tq µt t σ2 21ptq µt t ȷ ϵθpxt, σtq (69) From re-parameterization, we have, xt t µθpxt, tq σt ?κt zt zt Np0, Idq, κt χ2pν dq{pν dq (70) σ2 21ptq µt t ȷ ϵθpxt, σtq σt ?κt zt (71) Moreover, we choose the posterior scale to be the same as the forward posterior qpxt 1|xt, x0q i.e. σ2 t σ2 t t σ2 21ptqσ2 12ptq σ2 t This implies, σ2 21ptq σ2 t σ2 12ptq σ2 t t σ2 t (73) Substituting this form of σ2 21ptq into Eqn. 71, we have, σ2 t σ2 12ptq σ2 t t σ2 t µt t ȷ ϵθpxt, σtq σt ?κt zt (74) σ2 t t σ2 t µt t ȷ ϵθpxt, σtq σt ?κt zt (75) σ2 t t σ2 t 1 1 µt t ȷ ϵθpxt, σtq σt ?κt zt (76) σ2 t t σ2 t 1 9µt µt t ȷ ϵθpxt, σtq σt ?κt zt (77) where in the above equation we use the first-order approximation 9µt µt µt t t . Next, we make the following design choices: Published as a conference paper at ICLR 2025 1. Firstly, we assume the following form of the reverse posterior variance σ2 t : σ2 t βptqgpσt, 9σtq t (78) where g : R ˆ R Ñ R and βptq P R represents a time-varying scaling factor chosen empirically which can be used to vary the noise injected at each sampling step. It is worth noting that a positive 9σt (as indicated in the definition of g) is a consequence of a monotonically increasing noise schedule σt in diffusion model design. 2. Secondly, we make the following design choice: σ2 t t σ2 t 1 fpσt, 9σtq t (79) where f : R ˆ R Ñ R With these two design choices, Eqn. 77 simplifies as: σ2 t t σ2 t 1 9µt µt t ȷ ϵθpxt, σtq σt ?κt z (80) fpσt, 9σtq t 9µt µt t ȷ ϵθpxt, σtq a βptqgpσt, 9σtq ? t zt ?κt (81) fpσt, 9σtq 9µt ȷ ϵθpxt, σtq t a βptqgpσt, 9σtq ? t zt ?κt (82) xt t xt µt t fpσt, 9σtq 9µt ȷ ϵθpxt, σtq t a βptqgpσt, 9σtq ? t zt ?κt (83) 9µt µt xt σt fpσt, 9σtq 9µt ȷ ϵθpxt, σtq βptqgpσt, 9σtq ? t zt ?κt (84) In the limit of t Ñ 0: 9µt µt xt σt fpσt, 9σtq 9µt ȷ ϵθpxt, σtq βptqgpσt, 9σtqd Wt ?κt (85) 9µt µt xt fpσt, 9σtq 9µt ȷ pxt µt Dθpxt, σtqq βptqgpσt, 9σtqd Wt ?κt (86) Moreover, since d Wt Np0, dtq and κt χ2pν dq{pν dq, the term d St d Wt{?κt is distributed as a Student-t random variable with d St tdp0, dt, ν dq. Therefore, 9µt µt xt σt fpσt, 9σtq 9µt ȷ ϵθpxt, σtq βptqgpσt, 9σtqd St (87) which gives the required SDE formulation for the diffusion posterior sampling. Next, we discuss specific choices of fpσt, 9σtq and gpσt, 9σtq. On the choice of fpσt, 9σtq and gpσt, 9σtq: Recall our design choices: σ2 t βptqgpσt, 9σtq t (88) σ2 t t σ2 t 1 fpσt, 9σtq t (89) Published as a conference paper at ICLR 2025 Substituting the value of σ2 t from the first design choice to the second yields the following condition: σ2 t t βptqgpσt, 9σtq t 1 fpσt, 9σtq t (90) This concludes the proof. It is worth noting that the above equation provides two degrees of freedom, i.e., we can choose two variables among σ12ptq, g, f and automatically determine the third. However, it is more convenient to choose σ12ptq and g, since both these quantities should be positive. Different choices of these quantities yield different instantiations of the SDE in Eqn. 87 as illustrated in the main text. A.6 DERIVING THE DENOISER PRECONDITIONER FOR T-EDM Recall our denoiser parameterization, Dθpx, σq cskippσ, νqx coutpσ, νq Fθpcinpσ, νqx, cnoisepσqq (91) Karras et al. (2022) highlight the following design choices, which we adopt directly. 1. Derive cin based on constraining the input variance to 1 2. Derive cskip and cout to constrain output variance to 1 and additionally minimizing cout to bound scaling errors in Fθpx, σq. The coefficient cin can be derived by setting the network inputs to have unit variance. Therefore, Varx0,n cinpσqpx0 nq 1 (92) cinpσ, νq2 Varx0,n x0 n 1 (93) cinpσ, νq2 σ2 data ν ν 2σ2 1 (94) cinpσ, νq 1 Lc ν ν 2σ2 σ2 data. (95) The coefficients cskip and cout can be derived by setting the training target to have unit variance. Similar to Karras et al. (2022) our training target can be specified as: Ftarget 1 coutpσ, νq x0 cskippσ, νqpx0 nq (96) Varx0,n Ftargetpx0, n; σq 1 (97) Varx0,n 1 coutpσ,νq x0 cskippσ, νqpx0 nq ı 1 (98) 1 coutpσ,νq2 Varx0,n x0 cskippσ, νqpx0 nq 1 (99) coutpσ, νq2 Varx0,n x0 cskippσ, νqpx0 nq coutpσ, νq2 Varx0,n 1 cskippσ, νq x0 cskippσ, νq n ı coutpσ, νq2 1 cskippσ, νq 2 σ2 data cskippσ, νq2 ν ν 2σ2.(102) Lastly, setting cskippσ, νq to minimize coutpσ, νq, we obtain, cskippσ, νq σ2 data{ ˆ ν ν 2σ2 σ2 data Published as a conference paper at ICLR 2025 Consequently coutpσ, νq can be specified as: coutpσ, νq c ν ν 2σ σdata Lc ν ν 2σ2 σ2 data. (104) A.7 EQUIVALENCE WITH THE EDM ODE Similar to Karras et al. (2022), we start by deriving the optimal denoiser for the denoising loss function. Moreover, we reformulate the form of the perturbation kernel as qpxt|x0q tdpµtx0, σ2 t Id, νq tdpsptqx0, sptq2σptq2Id, νq by setting µt sptq and σt sptqσptq. The denoiser loss can then be specified as follows, Lp D, σq Ex0 ppx0q En tdp0,σ2Id,νq λpσ, νq}Dpsptqrx0 ns, σq x0}2 2 (105) Ex0 ppx0q Ex tdpsptqx0,sptq2σ2Id,νq λpσ, νq}Dpx, σq x0}2 2 (106) Ex0 ppx0q ż tdpx; sptqxi, sptq2σ2Id, νq λpσ, νq}Dpx, σq x0}2 2 dx ı (107) ż tdpx; sptqxi, sptq2σ2Id, νq λpσ, νq}Dpx, σq xi}2 2 dx (108) where the last result follows from assuming ppx0q as the empirical data distribution. Thus, the optimal denoiser can be specified by setting DLp D, σq 0. Therefore, DLp D, σq 0 (109) Consequently, ż tdpx; sptqxi, sptq2σ2Id, νq λpσ, νq}Dpx, σq xi}2 2 dx 0 (110) ż tdpx; sptqxi, sptq2σ2Id, νq λpσ, νqp Dpx, σq xiq dx 0 (111) i tdpx; sptqxi, sptq2σ2Id, νqp Dpx, σq xiqdx 0 (112) The optimal denoiser Dpx, σq, can be obtained from Eq. 112 as, i tdpx; sptqxi, sptq2σ2Id, νqxi ř i tdpx; sptqxi, sptq2σ2Id, νq (113) We can further simplify the optimal denoiser as, i tdpx; sptqxi, sptq2σ2Id, νqxi ř i tdpx; sptqxi, sptq2σ2Id, νq (114) sptq; xi, σ2Id, νqxi ř sptq; xi, σ2Id, νq (115) sptq, σ (116) Next, recall the ODE dynamics (Eqn. 12) in our formulation, ȷ pxt µt Dθpxt, σptqqq (117) Published as a conference paper at ICLR 2025 Reparameterizing the above ODE by setting µt sptq and σt sptqσptq, ȷ pxt sptq Dθpxt, σptqqq (118) sptqxt 9sptq sptq 9σptqsptq σptq 9sptq ȷ pxt sptq Dθpxt, σptqqq (119) sptqxt 9σptq ȷ pxt sptq Dθpxt, σptqqq (120) sptqxt 9σptq σptqpxt sptq Dθpxt, σptqqq (121) Lastly, since Karras et al. (2022) propose to train the denoiser Dθ using unscaled noisy state, from Eqn. 116, we can re-write the above ODE as, ı xt 9σptqsptq sptq, σptq (122) The form of the ODE in Eqn. 122 is equivalent to the ODE presented in Karras et al. (2022) (Algorithm 1 line 4 in their paper). This concludes the proof. A.8 CONDITIONAL VECTOR FIELD FOR T-FLOW Here, we derive the conditional vector field for the t-Flow interpolant. Recall, the interpolant in t-Flow is derived as follows, xt tx0 p1 tq ϵ ?κ, ϵ Np0, Idq, κ χ2pνq{ν (123) It follows that, xt t Erx0|xts p1 tq E ϵ ?κ|xt (124) Moreover, following Eq. 2.10 in Albergo et al. (2023), the conditional vector field bpxt, tq can be defined as, bpxt, tq Er 9xt|xts Erx0|xts E ϵ ?κ|xt (125) From Eqns. 124 and 125, the conditional vector field can be simplified as, bpxt, tq xt E ϵ ?κ|xt This concludes the proof. A.9 CONNECTION TO DENOISING SCORE MATCHING We start by defining the score for the perturbation kernel qpxt|x0q. The pdf for the perturbation kernel qpxt|x0q tdpµtx0, σ2 t Id, νq can be expressed as (ignoring the normalization constant): qpxt|x0q 9 1 1 νσ2 t xt µtx0 J xt µtx0 ȷ pν dq Published as a conference paper at ICLR 2025 xt log qpxt|x0q ν d 2 xt log 1 1 νσ2 t }xt µtx0}2 2 2 1 1 1 νσ2 t }xt µtx0}2 2 Denoting d1 1 σ2 t }xt µtx0}2 2 for convenience, we have, xt log qpxt|x0q ˆ ν d In Denoising Score Matching (DSM) (Vincent, 2011), the following objective is minimized, LDSM Et pptq,x0 ppx0q,xt qpxt|x0q γptq} xt log qpxt|x0q sθpxt, tq}2 2 for some loss weighting function γptq. Parameterizing the score estimator sθpxt, tq as: sθpxt, tq ˆ ν d ˆ xt µt Dθpxt, σtq (132) With this parameterization of sθpxt, tq and some choice of γptq, the DSM objective can be further simplified as follows, LDSM Ex0 ppx0q Et Eϵ Np0,Iq,κ χ2pνq{ν ν d 2 x0 Dθpµtx0 σt ϵ ?κ, σtq 2 Ex0 ppx0q Et Eϵ Np0,Iq,κ χ2pνq{ν λpxt, ν, tq x0 Dθpµtx0 σt ϵ ?κ, σtq 2 where λpxt, ν, tq ν d ν d1 2 , which is equivalent to a scaled version of the simplified denoising loss Eq. 10. This concludes the proof. As an additional caveat, the score parameterization in Eq. 132 depends on d1 1 σ2 t }xt µtx0}2 2, which can be approximated during inference as, d1 1 σ2 t }xt µt Dθpxt, σtq}2 2 A.10 ODE REFORMULATION AND CONNECTIONS TO INVERSE PROBLEMS. Recall the ODE dynamics in terms of the denoiser can be specified as, ȷ pxt µt Dθpxt, σtqq (135) From Eq. 132, we have, xt µt Dθpxt, σtq σ2 t xt log ppxt, tq (136) where d1 1 σ2 t }xt µtx0}2 2. Substituting the above result in the ODE dynamics, we obtain the reformulated ODE in Eq. 14. dxt µt xt σ2 t ν d1 ȷ x log ppxt, tq (137) Published as a conference paper at ICLR 2025 Since the term d1 is data-dependent, we can estimate it during inference as d1 d1 1 1 σ2 t }xt µt Dθpxt, σtq}2 2. Thus, the above ODE can be reformulated as, µt xt σ2 t ν d1 1 ν d ȷ x log ppxt, tq (138) Tweedie s Estimate and Inverse Problems. Given the perturbation kernel qpxt|x0q tdpµtx0, σ2 t Id, νq, we have, xt µtx0 σt ϵ ?κ, ϵ Np0, Idq, κ χ2pνq{ν (139) It follows that, xt µt Erx0|xts σt E ϵ ?κ|xt ı (140) xt σt E ϵ ?κ|xt ı xt σ2 t ν d1 1 ν d x log ppxt, tq which gives us an estimate of the predicted x0 at any time t. Moreover, the framework can also be extended for solving inverse problems using diffusion models. More specifically, given a conditional signal y, the goal is to simulate the ODE, µt xt σ2 t ν d1 1 ν d ȷ x log ppxt|yq (143) µt xt σ2 t ν d1 1 ν d ȷ wt x log ppy|xtq x log ppxtq ı (144) where the above decomposition follows from ppxt|yq9ppy|xtqwtppxtq and the weight wt represents the guidance weight of the distribution ppy|xtq. The term log ppy|xtq can now be approximated using existing posterior approximation methods in inverse problems (Chung et al., 2022; Song et al., 2022b; Mardani et al., 2023; Pandey et al., 2024b) A.11 CONNECTIONS TO ROBUST STATISTICAL ESTIMATORS Here, we derive the expression for the gradient of the γ-power divergence between the forward and the reverse posteriors (denoted by q and pθ, respectively for notational convenience) i.e., θDγpq } pθq. By the definition of the γ-power divergence, Dγpq } pθq 1 Cγpq, pθq Hγpqq , γ P p 1, 0q Y p0, 8q (145) Hγppq }p}1 γ ˆ ż ppxq1 γdx 1 1 γ Cγpq, pq ż qpxq ˆ ppxq θDγpq } pθq 1 γ θCγpq, pθq (147) Published as a conference paper at ICLR 2025 Consequently, we next derive an expression for θCγpq, pθq. θCγpq, pθq θ ż qpxq ˆ pθpxq γ ż qpxq ˆ pθpxq pθpxq }pθ}1 γ γ ż qpxq ˆ pθpxq γ 1 }pθ}1 γ θpθpxq pθpxq θ }pθ}1 γ }pθ}2 1 γ dx (150) From the definition of }pθ}1 γ, }pθ}1 γ ˆ ż pθpxq1 γdx 1 1 γ (151) θ }pθ}1 γ θ ˆ ż pθpxq1 γdx 1 1 γ (152) ˆ ż pθpxq1 γdx γ 1 γ ż θ pθpxq1 γ dx (153) ˆ ż pθpxq1 γdx γ 1 γ p1 γq ż pθpxqγ θpθpxqdx (154) θ }pθ}1 γ ˆ ż pθpxq1 γdx γ 1 γ ż pθpxqγ θpθpxqdx (155) }pθ}1 γ ˆ ş pθpxq1 γdx ż pθpxqγ θpθpxqdx (156) }pθ}1 γ ˆ ş pθpxq1 γdx ż pθpxq1 γ θ log pθpxqdx (157) ż pθpxq1 γ ˆ ş pθpxq1 γdx looooooooomooooooooon θ log pθpxqdx (158) }pθ}1 γ E pθpxqr θ log pθpxqs (159) where we denote pθpxq pθpxq1 γ ˆ ş pθpxq1 γdx for notational convenience. Substituting the above result in Eq. 150, we have the following simplification, Published as a conference paper at ICLR 2025 Algorithm 3: Training (t-Flow) 1: repeat 2: x1 ppx1q 3: t Uniformpt1, . . . , Tuq 4: µt t, σt 1 t 5: xt µtx1 σtn, n tdp0, Id, νq 6: Take gradient descent step on 7: θ }n ϵθpxt, σtq}2 8: until converged Algorithm 4: Sampling (t-Flow) 1: sample x0 td 0, Id, ν 2: for i P t0, . . . , N 1u do 3: di Ð xi ϵθpxi; σtiq {ti 4: xi 1 Ð xi pti 1 tiqdi 5: if ti 1 0 then 6: di Ð xi 1 ϵθpxi 1; σti 1q {ti 1 7: xi 1 Ð xi pti 1 tiq 1 8: end if 9: end for 10: return x N Figure 5: Training and Sampling algorithms for t-Flow. Our proposed method requires minimal code updates (indicated with blue) over traditional Gaussian flow models and converges to the latter as ν Ñ 8. θCγpq, pθq γ ż qpxq ˆ pθpxq γ 1 }pθ}1 γ θpθpxq pθpxq θ }pθ}1 γ }pθ}2 1 γ dx (160) γ ż qpxq ˆ pθpxq γ 1 }pθ}1 γ θpθpxq pθpxq }pθ}1 γ E pθpxqr θ log pθpxqs }pθ}2 1 γ dx γ ż qpxq ˆ pθpxq γ 1 θpθpxq pθpxq E pθpxqr θ log pθpxqs }pθ}1 γ dx (162) γ ż qpxq ˆ pθpxq γ 1 pθpxq θ log pθpxq pθpxq E pθpxqr θ log pθpxqs }pθ}1 γ dx (163) θCγpq, pθq γ ż qpxq ˆ pθpxq γ θ log pθpxq E pθpxqr θ log pθpxqs dx (165) Plugging this result in Eq. 147, we have the following result, θDγpq } pθq 1 γ θCγpq, pθq ż qpxq ˆ pθpxq γ θ log pθpxq E pθpxqr θ log pθpxqs dx (166) This completes the proof. Intuitively, the second term inside the integral in Eq. 166 ensures unbiasedness of the gradients. Therefore, the scalar coefficient γ controls the weighting on the likelihood gradient and can be set accordingly to ignore or model outliers when modeling the data distribution. B EXTENSION TO FLOWS Here, we discuss an extension of our framework to flow matching models (Albergo et al., 2023; Lipman et al., 2023) with a Student-t base distribution. More specifically, we define a straight-line flow of the form, xt tx1 p1 tq ϵ ?κ, ϵ Np0, Idq, κ χ2pνq{ν (167) Published as a conference paper at ICLR 2025 where x1 ppx1q. Intuitively, at a given time t, the flow defined in Eqn. 167 linearly interpolates between data and Student-t noise. Following Albergo et al. (2023), the conditional vector field which induces this interpolant can be specified as (proof in App. A.8) dt bpxt, tq xt E ϵ ?κ|xt We estimate E ϵ ?κ|xt by minimizing the objective Lpθq Ex0 ppx0q Et Ur0,1s Eϵ Np0,Idq Eκ χ2pνq{ν ϵθptx0 p1 tq ϵ ?κ, tq ϵ ?κ We refer to this flow setup as t-Flow. To generate samples from our model, we simulate the ODE in Eq. 168 using Heun s solver. Figure 5 illustrates the ease of transitioning from a Gaussian flow to t-Flow. Similar to Gaussian diffusion, transitioning to t-Flow requires very few lines of code change, making our method readily compatible with existing implementations of flow models. C EXPERIMENTAL SETUP C.1 UNCONDITIONAL MODELING C.1.1 HRRR DATASET We adopt the High-Resolution Rapid Refresh (HRRR) (Dowell et al., 2022) dataset, which is an operational archive of the US km-scale forecasting model. Among multiple dynamical variables in the dataset that exhibit heavy-tailed behavior, based on their dynamic range, we only consider the Vertically Integrated Liquid (VIL) and Vertical Wind Velocity at level 20 (denoted as w20) channels. How to cope with the especially non-Gaussian nature of such physical variables on km-scales, represents an entire subfield of climate model subgrid-scale parameterization (e.g., Guo et al. (2015)). We only use data for the years 2019 2020 for training (17.4k samples) and the data for 2021 (8.7k samples) for testing; data before 2019 are avoided owing to non-stationaries associated with periodic version changes of the HRRR. Lastly, while the HRRR dataset spans the entire US, for simplicity, we work with regional crops of size 128 ˆ 128 (corresponding to 384 ˆ 384 km over the Central US). Unless specified otherwise, we perform z-score normalization using precomputed statistics as a preprocessing step. We do not perform any additional data augmentation. C.1.2 BASELINES Baseline 1: EDM. For standard Gaussian diffusion models, we use the recently proposed EDM (Karras et al., 2022) model, which shows strong empirical performance on various image synthesis benchmarks and has also been employed in recent work in weather forecasting and downscaling (Pathak et al., 2024; Mardani et al., 2024). To summarize, EDM employs the following denoising loss during training, Lpθq 9 Ex0 ppx0q EσEn Np0,σ2Idq λpσq}Dθpx0 n, σq x0}2 2 (170) where the noise levels σ are usually sampled from a Log Normal distribution, ppσq Log Normalpπmean, π2 stdq Baseline 2: EDM + Inverse CDF Normalization (INC). It is commonplace to perform z-score normalization as a data pre-processing step during training. However, since heavy-tailed channels usually exhibit a high dynamic range, using z-score normalization for such channels cannot fully compensate for this range, especially when working with diverse channels in downstream tasks in weather modeling. An alternative could be to use a more stronger normalization scheme like Inverse CDF normalization, which essentially involves the following key steps: Published as a conference paper at ICLR 2025 Original Samples Samples (after INC) Denormalized Samples Channel: VIL Channel: w20 Figure 6: Inverse CDF Normalization. Using Inverse CDF Normalization (INC) can help reduce channel dynamic range during training while providing accurate denormalization. (Top Panel) INC applied to the Vertically Integrated Liquid (VIL) channel in the HRRR dataset. (Bottom Panel) INC applied to the Vertical Wind Velocity (w20) channel in the HRRR dataset. 1. Compute channel-wise 1-d histograms of the training data. 2. Compute channel-wise empirical CDFs from the 1-d histograms computed in Step 1. 3. Use the empirical CDFs from Step 2 to compute the CDF at each spatial location. 4. For each spatial location with a CDF value p, replace its value by the value obtained by applying the Inverse CDF operation under the standard Normal distribution. Fig. 6 illustrates the effect of performing normalization under this scheme. As can be observed, using such a normalization scheme can greatly reduce the dynamic range of a given channel while offering reliable denormalization. Moreover, since our normalization scheme only affects data preprocessing, we leave the standard EDM model parameters unchanged for this baseline. Baseline 3: EDM + Per Channel-Preconditioning (PCP). Another alternative to account for extreme values in the data (or high dynamic range) could be to instead add more heavy-tailed noise during training. This can be controlled by modulating the πmean and πstd parameters based on the dynamic range of the channel under consideration. Recall that these parameters control the domain of noise levels σ used during EDM model training. In this work, we use a simple heuristic to modulate these parameters based on the normalized channel dynamic range (denoted as d). More specifically, we set, πmean 1.2 α RBFpd, 1.0, βq (171) Published as a conference paper at ICLR 2025 where RBF denotes the Radial Basis Function kernel with radius=1.0, parameter β and a magnitude scaling factor α. We keep πstd 1.2 fixed for all channels. Intuitively, this implies that a higher normalized dynamic range (near 1.0) corresponds to sampling the noise levels σ from a more heavy-tailed distribution during training. This is natural since a signal with a larger magnitude would need more amount of noise to convert it to noise during the forward diffusion process. In this work, we set α 3.0, β 2.0, which yields πvil mean 1.8 and πw20 mean 0.453 for the VIL and w20 channels, respectively. It is worth noting that, unlike the previous baseline, we use z-score normalization as a preprocessing step for this baseline. Lastly, we keep other EDM parameters during training and sampling fixed. Baseline 4. Gaussian Flow. Since we also extend our framework to flow matching models (Albergo et al., 2023; Lipman et al., 2023), we also compare with a linear one-sided interpolant with a Gaussian base distribution. More specifically, xt tx1 p1 tqϵ, ϵ Np0, Idq (172) Similar to t-Flow (Section B), we train the Gaussian flow with the following objective, Lpθq Ex0 ppx0q Et Ur0,1s Eϵ Np0,Idq ϵθptx0 p1 tqϵ, tq ϵ 2 C.1.3 EVALUATION Here, we describe our scoring protocol used in Tables 2 and 3 in more detail. Kurtosis Ratio (KR). Intuitively, sample kurtosis characterizes the heavy-tailed behavior of a distribution and represents the fourth-order moment. Higher kurtosis represents greater deviations from the central tendency, such as from outliers in the data. In this work, given samples from the underlying train/test set, we generate 20k samples from our model. We then flatten all the samples and compute empirical kurtosis for both the underlying samples from the train/test set (denoted as kdata) and our model (denoted as ksim). The Kurtosis ratio is then computed as, KR ˇˇˇ1 ksim Lower values of this ratio imply a better estimation of the underlying sample kurtosis. Skewness Ratio (SR). Intuitively, sample skewness represents the asymmetry of a tailed distribution and represents the third-order moment. In this work, given samples from the underlying train/test set, we generate 20k samples from our model. We then flatten all the samples and compute empirical skewness for both the underlying samples from the train/test set (denoted as sdata) and our model (denoted as ssim). The Skewness ratio is then computed as, SR ˇˇˇ1 ssim Lower values of this ratio imply a better estimation of the underlying sample skewness. Kolmogorov-Smirnov 2-Sample Test (KS). The KS (Massey, 1951) statistic measures the maximum difference between the CDFs of two distributions. For heavy-tailed distributions, evaluating the KS statistic at the tails could provide a useful measure of the efficacy of our model in estimating tails reliably. To evaluate the KS statistic at the tails, similar to prior metrics, we generate 20k samples from our model. We then flatten all samples in the generated and train/test sets, followed by retaining samples lying above the 99.9th percentile (quantifying right tails/extreme region) or below the 0.1th percentile (quantifying the left tail/extreme region). Lastly, we compute the KS statistic between the retained samples from the generated and the train/test sets individually for each tail and average the KS statistic values for both tails to obtain an average KS score. The final score estimates how well the model might capture both tails. As an exception,. for the VIL channel, we report KS scores only for the right tail due to the absence of a left tail in the underlying samples for this Published as a conference paper at ICLR 2025 Parameters EDM (+INC, +PCP) t-EDM Flow/t-Flow Preconditioner cskip σ2 data{ ˆ σ2 σ2 data σ2 data{ ˆ ν ν 2 σ2 σ2 data cout σ σ2 data Lb σ2 σ2 data b ν ν 2 σ σ2 data Lb ν ν 2 σ2 σ2 data 1 σ2 σ2 data 1 Lb ν ν 2 σ2 σ2 data 1 cnoise 1 4 log σ 1 4 log σ σ σ log σ Npπmean, π2 stdq log σ Npπmean, π2 stdq σ 1 t, t Up0, 1q µt 1 1 t λpσq 1{c2 outpσq 1{c2 outpσ, νq 1 Loss Eq. 170 Eq. 13 t-Flow - Eq. 169 Gaussian Flow - Eq. 173 Solver Heun s (2nd order) Heun s (2nd order) Heun s (2nd order) ODE Eq. 12, x T Np0, Idq Eq. 12, x T tdp0, Id, νq t-Flow: Eq. 168, x0 tdp0, Id, νq Flow: Eq. 168, x0 Np0, Idq Discretization σ 1 ρ max i N 1 σ 1 ρ max i N 1 σ 1 ρ max i N 1 σ Scaling: µt 1 1 t Schedule: σt t t 1-t Hyperparameters σdata 1.0 1.0 N/A ν 8 VIL: t3, 5, 7u w20: t3, 5, 7u Flow: 8 t-Flow (Ó) VIL=t3, 5, 7u w20=t5, 7, 9u πmean, πstd EDM: -1.2, 1.2 EDM (+INC) : -1.2, 1.2 EDM (+PCP) (Ó) VIL: 1.8, 1.2 w20: 0.453, 1.2 -1.2, 1.2 N/A σmax, σmin 80, 0.002 80, 0.002 1.0, 0.01 NFE 18 18 25 ρ 7 7 7 Table 5: Comparison between design choices and specific hyperparameters between EDM (Karras et al., 2022) (+ related baselines) and t-EDM (Ours, Section 3.5) for unconditional modeling (Section 4.1). INC: Inverse CDF Normalization baselines, PCP: Per-Channel Preconditioning baseline, VIL: Vertically Integrated Liquid channel in the HRRR dataset, w20: Vertical Wind Velocity at level 20 channel in the HRRR dataset, NFE: Number of Function Evaluations channel (see Fig. 6 (first column) for 1-d intensity histograms for this channel). Lower values of the KS statistic imply better density assignment at the tails by the model. Histogram Comparisons. As a qualitative metric, comparing 1-d intensity histograms between the generated and the original samples from the train/test set can serve as a reliable proxy to assess the tail estimation capabilities of all models. C.1.4 DENOISER ARCHITECTURE We use the DDPM++ architecture from (Karras et al., 2022; Song et al., 2020). We set the base channel multiplier to 32 and the per-resolution channel multiplier to [1,2,2,4,4] with self-attention at resolution 16. The rest of the hyperparameters remain unchanged from Karras et al. (2022), which results in a model size of around 12M parameters. C.1.5 TRAINING We adopt the same training hyperparameters from Karras et al. (2022) for training all models. Model training is distributed across 4 DGX nodes, each with 8 A100 GPUs, with a total batch size of 512. We train all Published as a conference paper at ICLR 2025 Parameter Model Levels Height Levels (m) Zonal Wind (u) 1,2,3,4,5,6,7,8,9,10,11,13,15,20 10 Meridonal Wind (v) 1,2,3,4,5,6,7,8,9,10,11,13,15,20 10 Geopotential Height (z) 1,2,3,4,5,6,7,8,9,10,11,13,15,20 None Humidity (q) 1,2,3,4,5,6,7,8,9,10,11,13,15,20 None Pressure (p) 1,2,3,4,5,6,7,8,9,10,11,13,15,20 None Temperature (t) 1,2,3,4,5,6,7,8,9,10,11,13,15,20 2 Radar Reflectivity (refc) N/A Integrated Table 6: Parameters in the HRRR dataset used for conditional modeling tasks. models for a maximum budget of 60Mimg and select the best-performing model in terms of qualitative 1-D histogram comparisons. C.1.6 SAMPLING For the EDM and related baselines (INC, PCP), we use the ODE solver presented in Karras et al. (2022). For the t-EDM models, as presented in Section 3.5, our sampler is the same as EDM with the only difference in the sampling of initial latents from a Student-t distribution instead (See Fig. 2 (Right)). For Flow and t-Flow, we numerically simulate the ODE in Eq. 168 using the 2nd order Heun s solver with the timestep discretization proposed in Karras et al. (2022). For evaluation, we generate 20k samples from each model. We summarize our experimental setup in more detail for unconditional modeling in Table 5. C.1.7 EXTENDED RESULTS ON UNCONDITIONAL MODELING Sample Visualization. We visualize samples generated from the t-EDM and t-Flow models for the VIL and w20 channels in Figs. 9-12 Visualization of 1-d histograms. Similar to Fig. 3, we present additional results on histogram comparisons between different baselines and our proposed methods for the VIL and w20 channels in Figs. 13 and 14. Is PCP enough for capturing heavy tails? In Table 2, we observe that the EDM + PCP baseline outperforms t-EDM in some scenarios. A natural question is whether a dynamic preconditioning scheme based on the dynamic range of a channel is enough for modeling heavy-tailed data. To present some insight into this observation, we train an unconditional EDM + PCP model on 10 channels in the HRRR dataset selected on the basis of the top 10 channels with the highest dynamic range and heavier tails. Our EDM training setup is the same as other experiments presented in Section 4 with the exception that the input has 10 channels instead of a single channel. The updated πmean for all 10 channels is computed based on the formulation in Eqn. 171. Consequently, this is a more difficult modeling task than presented in the main text in Table 2. We compare the generated histograms between the 10-channel run with PCP and the single-channel run with PCP for the VIL and w20 channels in Fig. 7. As can be observed, when scaling in terms of the number of channels, the performance of EDM + PCP takes a big hit. This suggests that merely adjusting the noise schedule during training using heuristics like PCP is not sufficient to resolve the fundamental limitation of Gaussian diffusion models to estimate heavy tails accurately. C.2 CONDITIONAL MODELING C.2.1 HRRR DATASET FOR CONDITIONAL MODELING Similar to unconditional modeling (See App. C.1.1), we use the HRRR dataset for conditional modeling at the 128 x 128 resolution. We train the model on HRRR forecast fields 1-hour post-analysis to allow for Published as a conference paper at ICLR 2025 Figure 7: Sample 1-d histogram comparison between EDM + PCP with 10 channels and single-channel training on the test set for the Vertically Integrated Liquid (VIL) channel and Vertical Wind Velocity (w20) channel. Tail estimation suffers when scaling up training in terms of the number of channels in the input. PCP: Per-Channel Preconditioning some model spin-up following data assimilation. More specifically, for a lead time of 1hr, we sample (input, output) pairs at time τ and τ 1, respectively. For the input, at time τ, we use a state vector consisting of a combination of 86 atmospheric channels (including the channel to be predicted at time τ), which are summarized in Table 6. For the output, at time τ 1, we use either the Vertically Integrated Liquid (VIL) or Vertical Wind Velocity at level 20 (w20) channels, depending on the prediction task. Unless specified otherwise, we perform z-score normalization using precomputed statistics as a preprocessing step without any additional data augmentation. C.2.2 BASELINES We adopt the standard EDM (Karras et al., 2022) for conditional modeling as our baseline. C.2.3 DENOISER ARCHITECTURE We use the DDPM++ architecture from (Karras et al., 2022; Song et al., 2020). We set the base channel multiplier to 32 and the per-resolution channel multiplier to [1,2,2,4,4] with self-attention at resolution 16. Additionally, our noisy state x is channel-wise concatenated with an 86-channel conditioning signal, increasing the total number of input channels in the denoiser to 87. The number of output channels remains 1 since we are predicting only a single VIL/w20 channel. However, the increase in the number of parameters is minimal since only the first convolutional layer in the denoiser is affected. Therefore, our denoiser is around 12M parameters. The rest of the hyperparameters remain unchanged from Karras et al. (2022). C.2.4 TRAINING We adopt the same training hyperparameters from Karras et al. (2022) for training all conditional models. Model training is distributed across 4 DGX nodes, each with 8 A100 GPUs, with a total batch size of 512. We train all models for a maximum budget of 60Mimg. C.2.5 SAMPLING For both EDM and t-EDM models, we use the ODE solver presented in Karras et al. (2022). For the t-EDM models, as presented in Section 3.5, our sampler is the same as EDM with the only difference in the sampling Published as a conference paper at ICLR 2025 of initial latents from a Student-t distribution instead (See Fig. 2 (Right)). For a given input conditioning state, we generate an ensemble of predictions of size 16 by randomly initializing our ODE solver with different random seeds. All other sampling parameters remain unchanged from our unconditional modeling setup (see App. C.1.6). C.2.6 EVALUATION Root Mean Square Error (RMSE). is a standard evaluation metric used to measure the difference between the predicted values and the true values Chai & Draxler (2014). In the context of our problem, let x be the true target and ˆx be the predicted value. The RMSE is defined as: E r}x ˆx}2s. This metric captures the average magnitude of the residuals, i.e., the difference between the predicted and true values. A lower RMSE indicates better model performance, as it suggests the predicted values are closer to the true values on average. RMSE is sensitive to large errors, making it an ideal choice for evaluating models where minimizing large deviations is critical. Continuous Ranked Probability Score (CRPS). is a measure used to evaluate probabilistic predictions Wilks (2011). It compares the entire predicted distribution Fpˆxq with the observed data point x. For a probabilistic forecast with cumulative distribution function (CDF) F, and the true value x, the CRPS can be formulated as follows: CRPSp F, xq ż 8 8 p Fpyq Ipy ě xqq2 dy, where Ip q is the indicator function. Unlike RMSE, CRPS provides a more comprehensive evaluation of both the location and spread of the predicted distribution. A lower CRPS indicates a better match between the forecast distribution and the observed data. It is especially useful for probabilistic models that output a distribution rather than a single-point prediction. Spread-Skill Ratio (SSR). is used to assess over/under-dispersion in probabilistic forecasts. Spread measures the uncertainty in the ensemble forecasts and can be represented by computing the standard deviation of the ensemble members. Skill represents the accuracy of the mean of the ensemble forecasts and can be represented by computing the RMSE between the ensemble mean and the observations. Scoring Criterion. Since CRPS and SSR metrics are based on predicting ensemble forecasts for a given input state, we predict an ensemble of size 16 for 4000 samples from the VIL/w20 test set. We then enumerate window sizes of 16 x 16 across the spatial resolution of the generated sample (128 x 128). Since the VIL channel is quite sparse, we filter out windows with a maximum value of less than a threshold (1.0 for VIL) and compute the CRPS, SSR, and RMSE metrics for all remaining windows. As an additional caveat, we note that while it is common to roll out trajectories for weather forecasting, in this work, we only predict the target at the immediate next time step. C.2.7 EXTENDED RESULTS ON CONDITIONAL MODELING Sample Visualization. We visualize samples generated from the t-EDM and t-Flow models for the VIL and w20 channels in Figs. 9 and 10 On t-EDM stability for Autoregressive Rollouts. While we demonstrate the effectiveness of t-EDM for the conditional task of predicting the next frame, in practice, it is more useful to perform autoregressive Published as a conference paper at ICLR 2025 Parameters EDM t-EDM Preconditioner cskip σ2 data{ ˆ σ2 σ2 data ν ν 2σ2 σ2 data cout σ σ2 data La σ2 σ2 data b ν ν 2σ σ2 data Lb ν ν 2σ2 σ2 data σ2 σ2 data 1 Lb ν ν 2σ2 σ2 data cnoise 1 4 log σ 1 4 log σ σ log σ Npπmean, π2 stdq log σ Npπmean, π2 stdq sptq 1 1 λpσq 1{c2 outpσq 1{c2 outpσ, νq Loss Eq. 2 in Karras et al. (2022) Eq. 13 Solver Heun s (2nd order) Heun s (2nd order) ODE Eq. 12, x T Np0, Idq Eq. 12, x T tdp0, Id, νq Discretization σ 1 ρ max i N 1 σ 1 ρ max i N 1 σ ρ Scaling: sptq 1 1 Schedule: σptq t t Hyperparameters σdata 1.0 1.0 ν 8 x1 20, x2 P t4, 7, 10u πmean, πstd -1.2, 1.2 -1.2, 1.2 σmax, σmin 80, 0.002 80, 0.002 NFE 18 18 ρ 7 7 Table 7: Comparison between design choices and specific hyperparameters between EDM (Karras et al., 2022) and t-EDM (Ours, Section 3.5) for the Toy dataset analysis in Fig. 1. NFE: Number of Function Evaluations rollouts for downstream applications like forecasting. Therefore, we include qualitative results to showcase the stability of t-EDM when performing autoregressive rollouts in Fig. 15 C.3 TOY EXPERIMENTS Dataset. For the toy illustration in Fig. 1, we work with the Neals Funnel dataset (Neal, 2003), which is commonly used in the MCMC literature (Brooks et al., 2011) due to its challenging geometry. The underlying generative process for Neal s funnel can be specified as follows: ppx1, x2q Npx1; 0, 3q Npx2; 0, exp x1{2q. (176) For training, we randomly generate 1M samples from the generative process in Eq. 176 and perform z-score normalization as a pre-processing step. Baselines and Models. For the standard Gaussian diffusion model baseline (2nd column in Fig. 1), we use EDM with standard hyperparameters as presented in Karras et al. (2022). Consequently, for heavy-tailed diffusion models (columns 3-5 in Fig. 1), we use the t-EDM instantiation of our framework as presented in Section 3.5. Since the hyperparameter ν is key in our framework, we tune ν for each individual dimension in our toy experiments. We fix ν to 20 for the x1 dimension and vary it between ν P t4, 7, 10u to illustrate controllable tail estimation along the x2 dimension. Denoiser Architecture. For modeling the underlying denoiser Dθpx, σq, we use a simple MLP for all toy models. At the input, we concatenate the 2-dimensional noisy state vector x with the noise level σ. We use two hidden layers of size 64, followed by a linear output layer. This results in around 8.5k parameters for the denoiser. We share the same denoiser architecture across all toy models. Published as a conference paper at ICLR 2025 Toy Dataset (Neal s Funnel) Method ν KR Ó SR Ó KS Ó Baselines EDM 8 0.909 2.334 0.234 Ours t-EDM 4 0.393 1.584 0.051 t-EDM 7 0.795 3.671 0.194 t-EDM 10 0.863 2.198 0.261 Table 8: Quantitative comparison between EDM and t-EDM on the Neals Funnel 2-d toy dataset. For all metrics, lower is better. Values in bold indicate the best results in a column. Training. We optimize the training objective in Eq. 13 for both t-EDM and EDM (See Fig. 2 (Left)). Our training hyperparameters are the same as proposed in Karras et al. (2022). We train all toy models for a fixed duration of 30M samples and choose the last checkpoint for evaluation. Sampling. For the EDM baseline, we use the ODE solver presented in Karras et al. (2022). For the t-EDM models, as presented in Section 3.5, our sampler is the same as EDM with the only difference in the sampling of initial latents from a Student-t distribution instead (See Fig. 2 (Right)). For visualization of the generated samples in Fig. 1, we generate 1M samples for each model. Overall, our experimental setup for the toy dataset analysis in Fig. 1 is summarized in Table 7. Quantitative Results on Neals Funnel. We present quantitative results on the Neals funnel dataset in Table 8. t-EDM outperforms EDM on all three metrics, indicating better tail estimation, which also supports our qualitative findings in Fig. 1. D OPTIMAL NOISE SCHEDULE DESIGN In this section we discuss a strategy for choosing the parameter σmax (denoted by σ in this section for notational brevity) in a more principled manner as compared to EDM (Karras et al., 2022). More specifically, our approach involves directly estimating σ from the empirically observed samples which circumvents the need to rely on ad-hoc choices of this parameter which can affect downstream sampler performance. The main idea behind our approach is minimizing the statistical mutual information between datapoints from the underlying data distribution, x0 pdata, and their noisy counterparts xσ ppxσq. While a trivial (and non-practical) way to achieve this objective could be to set a large enough σ i.e. σ Ñ 8, we instead minimize the mutual information Ipx0, xσq while ensuring the magnitude of σ to be as small as possible. Formally, our objective can be defined as, min σ2 σ2 subject to Ipx0, xσq 0 (177) As we will discuss later, minimizing this constrained objective provides a more principled way to obtain σ from the underlying data statistics for a specific level of mutual information desired by the user. Next, we first simplify the form of Ipx0, xσq, followed by a discussion on the estimation of σ in the context of EDM and t-EDM. We also extend to the case of non-i.i.d noise. Published as a conference paper at ICLR 2025 Simplification of Ipx0, xσq. The mutual information Ipx0, xσq can be stated and simplified as follows, Ipx0, xσq DKLpppx0, xσq } ppx0qppxσqq (178) ż ppx0, xσq log ppx0, xσq ppx0qppxσqdx0dxσ (179) ż ppx0, xσq log ppxσ|x0q ppxσq dx0dxσ (180) ż ppxσ|x0qppx0q log ppxσ|x0q ppxσq dx0dxσ (181) ż ppxσ|x0q log ppxσ|x0q Ex0 ppx0q DKLpppxσ|x0q } ppxσqq ı (183) D.1 DESIGN FOR EDM Given the simplification in Eqn. 183, the optimization problem in Eqn. 177 reduces to the following, min σ2 σ2 subject to Ex0 ppx0q DKLpppxσ|x0q } ppxσqq ı 0 (184) Since at σmax we expect the marginal distribution ppxσq to converge to the generative prior (i.e. completely destory the structure of the data), we approximate ppxσq Np0, σ2Idq. With this simplification, the Lagrangian for the optimization problem in Eqn. 177 can be specified as, σ2 arg min σ2 σ2 λEx0 ppx0q DKL Npxσ; x0, σ2q } Npxσ; 0, σ2q ı (185) Setting the gradient w.r.t σ2 = 0, σ4 Ex0rx J 0 x0s 0 (186) which implies, λEx0rx J 0 x0s (187) For an empirical dataset, i 1 }xi}2 2 (188) This allows us to choose a σmax from the underlying data statistics during training or sampling. It is worth noting that the choice of the multiplier λ impacts the magnitude of σmax. However, this parameter can be chosen in a principled manner. At σ2 max σ2 , the estimate of the mutual information is given by: I px0, xσq 1 σ2 Ex0rx J 0 x0s Ex0rx J 0 x0s λ (189) which implies, λ Ex0rx J 0 x0s I px0, xσq2 (190) The above result provides a way to choose λ. Given the dataset stastistics, Ex0rx J 0 x0s, the user can specify an acceptable level of mutual information Ipx0, xσq to compute the correponding λ, which can then be used to find the corresponding minimum σmax required to achieve that level of mutual information. Next, we extend this analysis to t-EDM. Published as a conference paper at ICLR 2025 D.2 EXTENSION TO T-EDM In the case of t-EDM, we pose the optimization problem as follows, σ arg min σ2 σ2 λEx0 ppx0q Dγ tdpx0, σ2Id, νq } tdp0, σ2Id, νq ı (191) where Dγpq } pq is the Gamma-power Divergence between two distributions q and p. From the definition of the γ-Power Divergence in Eqn. 8, we have, Dγ tdpx0, σ2Id, νq } tdp0, σ2Id, νq 1 γ 1 γ ν,d p1 d ν 2q γ 1 γ looooooooooooooooomooooooooooooooooon pσ2q γd 2p1 γq 1Ex0rx J 0 x0s (192) where Cν,d Γp ν d 2 qpνπq d 2 and γ 2 ν d Solving the optimization problem yields the following optimal σ, σ2 λfpν, dq ν 2 ν 2 d Erx J 0 x0s ı ν 2 d 2pν 2q d (193) For an empirical dataset, we have the following simplification, σ2 λfpν, dq ν 2 ν 2 d i }xi}2 2 ı ν 2 d 2pν 2q d (194) D.3 EXTENSION TO CORRELATED GAUSSIAN NOISE We now extend our formulation for optimal noise schedule design to the case of correlated noise in the diffusion perturbation kernel. This is useful, especially for scientific applications where the data energy is distributed quite non-uniformly across the (data) spectrum. Let R Ex0 ppx0qrx0x0Js P Rdˆd denote the data correlation matrix. Let also consider the perturbation kernel Np0, Σq for the postitve-defintie covaraince matrix Σ P Rdˆd. Following the steps in equation 185, the Lagrangian for noise covariance estimation can be formulated as follows: min Σ tracepΣq λEx0 ppx0q DKLp Npx0, Σq } Np0, Σqq ı (195) min Σ tracepΣq λEx0 ppx0q x J 0 Σ 1x0 ı (196) min Σ tracepΣq λEx0 ppx0q tracepΣ 1x0x J 0 q ı (197) min Σ tracepΣq λ tracepΣ 1Ex0 ppx0qrx0x J 0 sq (198) min Σ tracepΣq λ tracepΣ 1Rq (199) It can be shown that the optimal solution to this minimization problem is given by, Σ ? λR1{2, where R1{2 denotes the matrix square root of R. This implies that the noise energy must be distributed along the singular vectors of the correlaiton matrix, where the energy is proportional to the noise singular values. We include the proof below. Proof. We define the objective: fpΣq tracepΣq λ tracepΣ 1Rq. (200) Published as a conference paper at ICLR 2025 We compute the gradient of fpΣq with respect to Σ: ΣfpΣq B BΣtracepΣq λ B BΣtracepΣ 1Rq. (201) The gradient of the first term is straightforward: B BΣtracepΣq I. (202) For the second term, using the matrix calculus identity, the gradient is: λ tracepΣ 1Rq λΣ 1RΣ 1. (203) Combining these results, the total gradient is: ΣfpΣq I λΣ 1RΣ 1. (204) Setting the gradient to zero to find the critical point: I λΣ 1RΣ 1 0. (205) which implies, R λΣ2. (207) λR1{2. (208) which completes the proof. E LOG-LIKELIHOOD FOR T-EDM Here, we present a method to estimate the log-likelihood for the generated samples using the ODE solver for t-EDM (see Section 3.5). Our analysis is based on the likelihood computation in continuous-time diffusion models as discussed in Song et al. (2020) (Appendix D.2). More specifically, given a probability-flow ODE, dt fθpxt, tq, (209) with a vector field fθpxt, tq, Song et al. (2020) propose to estimate the log-likelihood of the model as follows, log ppx0q log ppx T q ż T 0 fθpxt, tqdt. (210) The divergence of the vector field fθpxt, tq is further estimated using the Skilling-Hutchinson trace estimator (Skilling, 1989; Hutchinson, 1990) as follows, fθpxt, tq EppϵqrϵJ fθpxt, tqϵs (211) Published as a conference paper at ICLR 2025 where usually, ϵ Np0, Idq. For t-EDM ODE, this estimate can be further simplified as follows, fθpxt, tq EppϵqrϵJ fθpxt, tqϵs (212) Eppϵq ϵJ xt Dθpxt, tq Eppϵq ϵJ Id xt Dθpxt, tq t Eppϵq ϵJϵ ϵJ xt Dθpxt, tqϵ ı (215) t Eppϵq ϵJϵ ϵJ xt Dθpxt, tqϵ ı (216) d Eppϵq ϵJ xt Dθpxt, tqϵ ı (217) where d is the data dimensionality. Thus, the log-likelihood can be specified as, log ppx0q log ppx T q ż T d Eppϵq ϵJ Dθpxt, tqϵ ı dt. (218) When ϵ Np0, σ2Idq, the above result can be re-formulated as, log ppx0q log ppx T q ż T σ2 Eppϵq ϵJ Dθpxt, tqϵ ı dt. (219) Moreover, using the first-order taylor series expansion Dθpx ϵq Dθpxq Dθϵ Op}ϵ}2q (220) For a sufficiently small σ, higher-order terms in Op}ϵ}2q can be ignored since Er}ϵ2}s σ2d. Therefore, Dθpx ϵq Dθpxq Dθϵ (221) ϵJ Dθϵ ϵJr Dθpx ϵq Dθpxqs (222) EϵrϵJ Dθϵs EϵrϵJp Dθpx ϵq Dθpxqqs (223) EϵrϵJ Dθϵs EϵrϵJDθpx ϵqs (224) Therefore, the log-likelihood expression can be further simplified as, log ppx0q log ppx T q ż T σ2 Eppϵq ϵJ Dθpxt, tqϵ ı dt (225) log ppx0q log ppx T q ż T σ2 Eppϵq ϵJDθpxt ϵ, tq ı dt (226) The advantage of this simplification is that we don t need to rely on expensive jacobian-vector products in Eq. 218. However, since the denoiser now depends on ϵ, monte-carlo approximation of the expectation in the above equation could be computationally expensive for many samples ϵ F DISCUSSION AND LIMITATIONS F.1 RELATED WORK Connections with Denoising Score Matching. For the perturbation kernel qpxt|x0q tdpµtx0, σ2 t Id, νq, the denoising score matching (Vincent, 2011; Song et al., 2020) loss, LDSM, can be formulated as, LDSMpθq 9 Ex0 ppx0q Et Eϵ Np0,Idq Eκ χ2pνq{ν λpxt, ν, tq Dθpµtx0 σt ϵ ?κ, σtq x0 2 Published as a conference paper at ICLR 2025 with the scaling factor λpxt, ν, tq rpν dq{pν d1qs2 where d1 p1{σ2 t q}xt µtx0}2 2 (proof in App. A.9). Therefore, the denoising score matching loss in our framework is equivalent to the simplified training objective in Eq. 10 scaled by a data-dependent coefficient. However, in this work, we do not explore this loss formulation and leave further exploration to future work. Prior work in Heavy-Tailed Generative Modeling. The idea of exploring heavy-tailed priors for modeling heavy-tailed distributions has been explored in several works in the past. More specifically, Jaini et al. (2020) argue that a Lipschitz flow map cannot change the tails of the base distribution significantly. Consequently, they use a heavy-tailed prior (modeled using a Student-t distribution) as the base distribution to learn Tail Adaptive flows (TAFs), which can model the tails more accurately. In this work, we make similar observations where standard diffusion models fail to accurately model the tails of real-world distributions. Consequently, Laszkiewicz et al. (2022) assess the tailedness of each marginal dimension and set the prior accordingly. On a similar note, we note that learning the tail parameter ν spatially and across channels can provide greater modeling flexibility for downstream tasks and will be an important direction for future work on this problem. More recently, Kim et al. (2024) introduce heavy-tailed VAEs (Kingma & Welling, 2022; Rezende & Mohamed, 2016) based on minimizing γ-power divergences (Eguchi, 2021). This is perhaps the closest connection of our method with prior work since we rely on γ-power divergences to minimize the divergence between heavy-tailed forward and reverse diffusion posteriors. However, VAEs often have scalability issues and tend to produce blurry artifacts (Dosovitskiy & Brox, 2016; Pandey et al., 2022). On the other hand, we work with diffusion models, which are known to scale well to large-scale modeling applications (Pathak et al., 2024; Mardani et al., 2024; Esser et al., 2024; Podell et al., 2023). Within diffusion models, Deasy et al. (2021) show that denoising score matching can be extended to Generalized Normal distributions and refer to the resulting training as heavy-tailed denoising score matching (HTDSM). Moreover, Deasy et al. (2021) rely on annealed Langevin dynamics for sampling. However, an exact formulation of the reverse process is non-trivial and would involve Levy formulations of Kolmogorov s forward and reverse equations, which can be quite complex. In contrast, due to the conditional properties of multivariate Student-t distributions, our method circumvents such complex formulations and enables a simple framework for designing samplers for Student-t based diffusion models (see Proposition 2 in the main text). Similarly, Yoon et al. (2023) present a framework for modeling heavy-tailed distributions using α-stable Levy processes while Shariatian et al. (2024) simplify the framework proposed in Yoon et al. (2023) and instantiate it for more practical diffusion models like DDPM. In contrast, our work deals with Student-t noise, which in general (with the exceptions of Cauchy and the Gaussian distribution) is not α-stable and, therefore, a distinct category of diffusion models for modeling heavy-tailed distributions. Moreover, prior works like Yoon et al. (2023); Shariatian et al. (2024) rely on empirical evidence from light-tailed variants of small-scale datasets like CIFAR-10 (Krizhevsky, 2009) and their efficacy on actual large-scale scientific datasets like weather datasets remains to be seen. Prior work in Diffusion Models. Our work is a direct extension of standard diffusion models in the literature (Karras et al., 2022; Ho et al., 2020; Song et al., 2020). Moreover, since it only requires a few lines of code change to transition from standard diffusion models to our framework, our work is directly compatible with popular families of latent diffusion models (Pandey et al., 2022; Rombach et al., 2022) and augmented diffusion models (Dockhorn et al., 2022; Pandey & Mandt, 2023; Singhal et al., 2023). Our work is also related to prior work in diffusion models on a more theoretical level. More specifically, PFGM++ (Xu et al., 2023b) is a unique type of generative flow model inspired by electrostatic theory. It treats d-dimensional data as electrical charges in a D d-dimensional space, where the electric field lines define a bijection between a heavy-tailed prior and the data distribution. D is a hyperparameter controlling the shape of the electric fields that define the generative mapping. In essence, their method can be seen as utilizing a perturbation kernel: ppxt|x0q9p||xt x0||2 2 σ2 t Dqq D d 2 tdpx0, σ2 t Id, Dq When setting ν D, the perturbation kernel becomes equivalent to that of t-EDM, indicating the Student-t perturbation kernel can be interpreted from another physical perspective that of electrostatic fields and Published as a conference paper at ICLR 2025 Figure 8: Illustration of the pdf for a χ2 distribution with varying degrees of freedom. charges. The authors demonstrated that using an intermediate value for D (or ν) leads to improved robustness compared to diffusion models (where D Ñ 8), due to the heavy-tailed perturbation kernel. F.2 LIMITATIONS AND FUTURE WORK While our proposed framework works well for modeling heavy-tailed data, it is not without its limitations. On tuning ν. Firstly, while the parameter ν offers controllability for tail estimation using diffusion models, it also increases the tuning budget by introducing an extra hyperparameter. Moreover, for diverse data channels, tuning ν per channel could be key to good estimation at the tails. This could result in a combinatorial explosion with manual tuning. We think there might be several ways to learn ν from the data. Firstly, it is worth noting that the parameter ν directly influences training in heavy-tailed diffusions. This is because the noising process xt µtx0 z ?κ where κ χ2pνq{ν. Since the χ2-squared distribution is a special case of a Gamma distribution, we can perhaps use reparameterization (Ruiz et al., 2016) to estimate the parameters (i.e. ν) of this distribution using a neural network to learn this parameter end-to-end. One important caveat in learning ν end-to-end is that the optimization process could just select a divergence which reduces the loss while not fitting ν on the data. Therefore, exploring these caveats for learning ν are an important direction for further research. Alternatively, one can also estimate ν from the data in a separate stage and use it as a perturbation kernel parameter. We highlight one simple way to do this in Appendix D where we attempt to derive the optimal noise schedule for EDM and t-EDM by minimizing the mutual information between the x0 and the noisy data point xσ w.r.t σ. In principle, for t-EDM, we can also solve this minimization problem jointly over ν and σ to learn an initial ν from the data. On Evaluation metrics. Secondly, our evaluation protocol relies primarily on comparing the statistical properties of samples obtained by flattening the generated or train/test set samples. One disadvantage of this approach is that our current evaluation metrics ignore the structure of the generated samples. On this Published as a conference paper at ICLR 2025 note, it may be tempting to use metrics like FID (Heusel et al., 2018), precision/recall (Sajjadi et al., 2018) to assess the performance between the generated and dataset samples. However, a key caveat in using these metrics is the reliance on a pre-trained feature extractor, which is commonly trained on natural images (like Image Net). However, in this work, we focus on weather data, which has a much different structure than natural images. Due to this distribution shift, it is unclear if a feature extractor pre-trained on natural images can extract meaningful structural features from weather datasets. Therefore, computing metrics like FID using a pre-trained network trained on natural images for weather modeling is dubious at best. In this spirit, training these feature extractors on large-scale weather data using classification or self-supervised losses so that downstream metrics like FID can be reported reliably can be an interesting direction for future work. Applications in other domains. In this work, while we explore the application of heavy-tailed diffusion models like t-EDM in the context of weather forecasting, the methods developed in this work are quite generic, and it will be interesting to apply these in the context of other domains like finance, which rely on heavy-tailed modeling of assets. Lastly, our conditional synthesis experiments are limited to the next step prediction and do not perform autoregressive rollouts like in the forecasting setup. Extending t-EDM with improved preconditioning and automatic ν tuning for forecasting would be an interesting research direction. Published as a conference paper at ICLR 2025 t-EDM (KS: 0.431) EDM (KS: 0.997) Train Set Samples Figure 9: Random samples generated from t-EDM (Top Panel) and EDM (Bottom Panel) for the Vertically Integrated Liquid (VIL) channel. KS: Kolmogorov-Smirnov 2-sample statistic. Samples have been scaled logarithmically for better visualization Published as a conference paper at ICLR 2025 t-Flow (KS: 0.711) Gaussian Flow (KS: 0.897) Train Set Samples Figure 10: Random samples generated from t-Flow (Top Panel) and Gaussian Flow (Bottom Panel) for the Vertically Integrated Liquid (VIL) channel. KS: Kolmogorov-Smirnov 2-sample statistic. Samples have been scaled logarithmically for better visualization Published as a conference paper at ICLR 2025 t-EDM (KS: 0.683) EDM (KS: 0.991) Train Set Samples Figure 11: Random samples generated from t-EDM (Top Panel) and EDM (Bottom Panel) for the Vertical Wind Velocity (w20) channel. KS: Kolmogorov-Smirnov 2-sample statistic Published as a conference paper at ICLR 2025 (KS: 0.259) Gaussian Flow (KS: 0.294) Train Set Samples Figure 12: Random samples generated from t-Flow (Top Panel) and Gaussian Flow (Bottom Panel) for the Vertical Wind Velocity (w20) channel. KS: Kolmogorov-Smirnov 2-sample statistic Published as a conference paper at ICLR 2025 (Channel: VIL) (Channel: w20) t-EDM Figure 13: 1-d Histogram Comparisons between samples from the generated and the Train/Test set for the Vertically Integrated Liquid (VIL, see Top Pandel) and Vertical Wind Velocity (w20, see Bottom Panel) channels using t-EDM (with varying ν). Published as a conference paper at ICLR 2025 (Channel: VIL) t-Flow (Channel: w20) t-Flow Figure 14: 1-d Histogram Comparisons between samples from the generated and the Train/Test set for the Vertically Integrated Liquid (VIL, see Top Pandel) and Vertical Wind Velocity (w20, see Bottom Panel) channels using t-Flow (with varying ν). Published as a conference paper at ICLR 2025 Ground Truth Ensemble Mean Prediction 1 Prediction 2 Ensemble (gif) Ground Truth Ensemble Mean Prediction 1 Prediction 2 Ensemble (gif) Figure 9: Qualitative visualization of samples generated from our conditional modeling for predicting the next state for the Vertically Integrated Liquid (VIL) channel. The ensemble mean represents the mean of ensemble predictions (16 in our case). Columns 2-3 represent two samples from the ensemble. The last column visualizes an animation of all ensemble members (Best viewed in a dedicated PDF reader). For each sample, the rows correspond to predictions from EDM, t-EDM (ν 3), and t-EDM (ν 5) from top to bottom, respectively. Samples have been scaled logarithmically for better visualization Published as a conference paper at ICLR 2025 Ground Truth Ensemble Mean Prediction 1 Prediction 2 Ensemble (gif) Ground Truth Ensemble Mean Prediction 1 Prediction 2 Ensemble (gif) Figure 10: Qualitative visualization of samples generated from our conditional modeling for predicting the next state for the Vertical Wind Velocity (w20) channel. The ensemble mean represents the mean of ensemble predictions (16 in our case). Columns 2-3 represent two samples from the ensemble. The last column visualizes an animation of all ensemble members (Best viewed in a dedicated PDF reader). For each sample, the rows correspond to predictions from EDM, t-EDM (ν 3), and t-EDM (ν 5) from top to bottom, respectively. Published as a conference paper at ICLR 2025 Figure 15: Qualitative illustration of autoregressive rollouts using t-EDM (ν 3) and EDM on the w20 channel. The conditional diffusion models were initialized using HRRR validation data and predictions were made with an interval of 1 hour up to a lead time of 10 hours. We do not observe any instabilities when generating trajectories using t-EDM. τ represents lead-time with τ 0 denoting initial state.