# interactionforce_transport_gradient_flows__9c777cb8.pdf Interaction-Force Transport Gradient Flows Egor Gladin Humboldt University of Berlin Berlin, Germany & HSE University egorgladin@yandex.ru Pavel Dvurechensky Weierstrass Institute for Applied Analysis and Stochastics Berlin, Germany pavel.dvurechensky@wias-berlin.de Alexander Mielke Humboldt University of Berlin & WIAS Berlin, Germany alexander.mielke@wias-berlin.de Jia-Jie Zhu Weierstrass Institute for Applied Analysis and Stochastics Berlin, Germany jia-jie.zhu@wias-berlin.de This paper presents a new gradient flow dissipation geometry over non-negative and probability measures. This is motivated by a principled construction that combines the unbalanced optimal transport and interaction forces modeled by reproducing kernels. Using a precise connection between the Hellinger geometry and the maximum mean discrepancy (MMD), we propose the interaction-force transport (IFT) gradient flows and its spherical variant via an infimal convolution of the Wasserstein and spherical MMD tensors. We then develop a particle-based optimization algorithm based on the JKO-splitting scheme of the mass-preserving spherical IFT gradient flows. Finally, we provide both theoretical global exponential convergence guarantees and improved empirical simulation results for applying the IFT gradient flows to the sampling task of MMD-minimization. Furthermore, we prove that the spherical IFT gradient flow enjoys the best of both worlds by providing the global exponential convergence guarantee for both the MMD and KL energy. 1 Introduction Optimal transport (OT) distances between probability measures, including the earth mover s distance [Werman et al., 1985, Rubner et al., 2000] and Monge-Kantorovich or Wasserstein distance [Villani, 2008], are one of the cornerstones of modern machine learning as they allow performing a variety of machine learning tasks, e.g., unsupervised learning [Arjovsky et al., 2017, Bigot et al., 2017], semi-supervised learning [Solomon et al., 2014], clustering [Ho et al., 2017], text classification [Kusner et al., 2015], image retrieval, clustering and classification [Rubner et al., 2000, Cuturi, 2013, Sandler and Lindenbaum, 2011], and distributionally robust optimization [Sinha et al., 2020, Mohajerin Esfahani and Kuhn, 2018]. Many recent works in machine learning adopted the techniques from PDE gradient flows over optimal transport geometries and interacting particle systems for inference and sampling tasks. Those tools not only add new interpretations to the existing algorithms, but also provide a new perspective on designing new algorithms. Corresponding author: Jia-Jie Zhu 38th Conference on Neural Information Processing Systems (Neur IPS 2024). For example, the classical Bayesian inference framework minimizes the Kulback-Leibler divergence towards a target distribution π. From the optimization perspective, this can be viewed as solving min µ A P F(µ) := DKL(µ|π) , (1) Figure 1: (Left) Wasserstein flow of the MMD energy [Arbel et al., 2019]. Some particles get stuck at points away from the target. (Right) IFT gradient flow (this paper) of the MMD energy. Particle mass is teleported to close to the target, avoiding local minima. Hollow circles indicate particles with zero mass. The red dots are the initial particles, and the green dots are the target distribution. See 4 for more details. where A is a subset of the space of probability measures P, e.g., the Gaussian family. The Wasserstein gradient flow of the KL gives the Fokker Planck equation, which can be simulated using the Langevin SDE for MCMC. Beyond the KL, many researchers following Arbel et al. [2019] advocated using the squared MMD instead as the driving energy for the Wasserstein gradient flows for sampling. However, in contrast to the KL setting, there is little sound convergence analysis for the MMDminimization scheme like the celebrated BakryÉmery Theorem. Furthermore, it was shown, e.g., in [Korba et al., 2021], that Arbel et al. [2019] s algorithm suffers a few practical drawbacks. For example, their particles tend to collapse around the mode or get stuck at local minima, and the algorithm requires a heuristic noise injection strategy that is tuned over the iterations; see Figure 1 and 4 for illustrations. Subsequently, many such as Carrillo et al. [2019], Chewi et al. [2020], Korba et al. [2021], Glaser et al. [2021], Craig et al. [2023], Hertrich et al. [2023], Neumayer et al. [2024] proposed modified energies to be used in the Wasserstein gradient flows. In contrast, this paper does not propose new energy objectives. Instead, we propose a new gradient flow geometry the IFT gradient flows. To summarize, our main contributions are: 1. We propose the interaction-force transport (IFT) gradient flow geometry over non-negative measures and spherical IFT over probability measures, constructed from the first principles of the reaction-diffusion type equations, previously studied in the context of the Hellinger Kantorovich (Wasserstein-Fisher-Rao) distance and gradient flows. It was first studied by three groups including Chizat et al. [2018, 2019], Liero et al. [2018], Kondratyev et al. [2016], Gallouët and Monsaingeon [2017]. Our IFT gradient flow is based on the inf-convolution of the Wasserstein and the newly constructed spherical MMD Riemannian metric tensors. This new unbalanced gradient flow geometry allows teleporting particle mass in addition to transportation, which avoids the flow getting stuck at local minima; see Figure 1 for an illustration. 2. We provide theoretical analysis such as the global exponential decay of energy functionals via the Polyak-Łojasiewicz type functional inequalities. As an application, we provide the first global exponential convergence analysis of IFT for both the MMD and KL energy functionals. That is, the IFT gradient flow enjoys the best of both worlds. 3. We provide a new algorithm for the implementation of the IFT gradient flow. We then empirically demonstrate the use of the IFT gradient flow for the MMD inference task. Compared to the original MMD-energy-flow algorithm of Arbel et al. [2019], IFT flow does not suffer issues such as the collapsing-to-mode issue. Leveraging the first-principled spherical IFT gradient flow, our method does not require a heuristic noise injection that is commonly tuned over the iterations in practice; see [Korba et al., 2021] for a discussion. Our method can also be viewed as addressing a long-standing issue of the kernel-mean embedding methods [Smola et al., 2007, Muandet et al., 2017, Lacoste-Julien et al., 2015] for optimizing the support of distributions. Notation We use the notation P(X), M+(X) to denote the space of probability and non-negative measures on the closed, bounded, (convex) set X Rd. The base space symbol X is often dropped if there is no ambiguity in the context. We note also that many of our results hold for X = Rd. In this paper, the first variation of a functional F at µ M+ is defined as a function δF δµ [µ] such that d dϵF(µ + ϵ v)|ϵ=0 = R δF δµ [µ](x) dv(x) for any valid perturbation in measure v such that µ + ϵ v M+ when working with gradient flows over M+ and µ + ϵ v P over P. We often omit the time index t to lessen the notational burden, e.g., the measure at time t, µ(t, ), is written as µ. The infimal convolution (inf-convolution) of two functions f, g on Banach spaces is defined as (f g)(x) = infy {f(y) + g(x y)}. In formal calculation, we often use measures and their density interchangeably, i.e., R f µ means the integral w.r.t. the measure µ. For a rigorous generalization of flows over continuous measures to discrete measures, see [Ambrosio et al., 2005]. 2k( , ) denotes the gradient w.r.t. the second argument of the kernel. 2 Background 2.1 Gradient flows of probability measures for learning and inference Gradient flows are powerful tools originated from the field of PDE. The intuition can be easily seen from the perspective of optimization as solving the variational problem min µ A M+(X) F(µ) using a continuous-time version of gradient descent, over a suitable metric space and, in particular, Riemannian manifold. Since the seminal works by Otto [1996] and colleagues, one can view many PDEs as gradient flows over the aforementioned Wasserstein metric space, canonically denoted as (P2(X), W2); see [Villani, 2008, Santambrogio, 2015] for a comprehensive introduction. Different from a standard OT problem, a gradient flow solution traverses along the path of the fastest dissipation of the energy F allowed by the corresponding geometry. In this paper, we are only concerned with the geometries with a (pseudo-)Riemannian structure, such as the Wasserstein, (spherical) Hellinger or Fisher-Rao geometries. In such cases, a formal Otto calculus can be developed to greatly simplify the calculations. For example, the Wasserstein Onsager operator (which is the inverse of the Riemannian metric tensor) KW (ρ) : T ρ M+ TρM+, ξ 7 div(ρ ξ), where TρM+ is the tangent plane of M+ at ρ and T ρ M+ the cotangent plane. Using this notation, a Wasserstein gradient flow equation of some energy F can be written as µ = KW (µ)δF δµ = div(µ δF In essence, many machine learning applications are about making different choices of the energy F in (2), e.g., the KL, χ2-divergence, or MMD. However, Wasserstein and its flow equation (2) are by no means the only meaningful geometry for gradient flows. One major development in the field is the Hellinger-Kantorovich a.k.a. the Wasserstein-Fisher-Rao (WFR) gradient flow. The WFR gradient flow equation is given by the reaction-diffusion equation, for some scaling coefficients α, β > 0, u = α div(u δF A few recent works have applied WFR to sampling and inference [Yan et al., 2024, Lu et al., 2019] by choosing the energy functional to be the KL divergence. 2.2 Reproducing kernel Hilbert space and MMD In this paper, we refer to a bi-variate function k : X X R as a symmetric positive definite kernel if k is symmetric and, for all n N, α1, . . . , αn R and all x1, . . . , xn X, we have Pn i=1 Pn j=1 αiαjk (xj, xi) 0. k is a reproducing kernel if it satisfies the reproducing property, i.e., for all x X and all functions in a Hilbert space f H, we have f(x) = f, k( , x) H. Furthermore, the space H is an RKHS if the Dirac functional δx : H 7 R, δx(f) := f(x) is continuous. It can be shown that there is a one-to-one correspondence between the RKHS H and the reproducing kernel k. Suppose the kernel is square-integrable k 2 L2ρ := R k(x, x)dρ(x) < w.r.t. ρ P. The integral operator Tk,ρ : L2 ρ H is defined by Tk,ρg(x) := R k (x, x ) g (x ) dρ (x ) for g L2 ρ. With an abuse of terminology, we refer to the following composition also as the integral operator Kρ := Id Tk,ρ, L2(ρ) L2(ρ). Kρ is compact, positive, self-adjoint, and nuclear; cf. [Steinwart and Christmann, 2008]. To simplify the notation, we simply write K when ρ is the Lebesgue measure. The kernel maximum mean discrepancy (MMD) [Gretton et al., 2012] emerged as an easy-tocompute alternative to optimal transport for computing the distance between probability measures, i.e., MMD2(µ, ν) := K (µ ν) 2 H = Z Z k(x, x ) d(µ ν)(x) d(µ ν)(x ) , where H is the RKHS associated with the (positive-definite) kernel k. While the MMD enjoys many favorable properties, such as a closed-form estimator and favorable statistical properties [Tolstikhin et al., 2017, 2016], its mathematical theory is less developed compared to the Wasserstein space especially in the geodesic structure and gradient flow geometries. It has been shown by Zhu and Mielke [2024] that MMD is a (de-)kernelized Hellinger or Fisher-Rao distance by using a dynamic formulation MMD2(µ, ν) = min Z 1 0 ξt 2 H dt u = K 1ξt, u(0) = µ, u(1) = ν, ξt H . (4) Mathematically, we can obtain the MMD geodesic structure if we kernelize the Hellinger (Fisher-Rao) Riemannian metric tensor, GMMD = Kµ GHe(µ), KMMD = KHe(µ) K 1 µ , (5) noting that the Onsager operator K is the inverse of the Riemannian metric tensor K = G 1. The MMD suffers from some shortcomings in practice, such as the vanishing gradients and kernel choices that require careful tuning; see e.g., [Feydy et al., 2019]. Furthermore, a theoretical downside of the MMD as a tool for optimizing distributions, and kernel-mean embedding [Smola et al., 2007, Muandet et al., 2017] in general, is that they do not allow transport dynamics. This limitation is manifested in practice, e.g., it is intractable to optimize the location of particle distributions; see e.g. [Lacoste-Julien et al., 2015]. In this paper, we address all those issues. 3 IFT gradient flows over non-negative and probability measures Figure 2: Illustration of the IFT gradient flow. Atoms are subject to both the transport (Kantorovich) potential and the interaction (repulsive) force from other atoms. In this section, we propose the IFT gradient flows over non-negative and probability measures. Note that our methodology is fundamentally different from a few related works in kernel methods and gradient flows such as [Arbel et al., 2019, Korba et al., 2021, Hertrich et al., 2023, Glaser et al., 2021, Neumayer et al., 2024] in that we are not concerned with the Wasserstein flows of a different energy, but a new gradient flow dissipation geometry. 3.1 (Spherical) IFT gradient flow equations over non-negative and probability measures The construction of the Wasserstein-Fisher Rao gradient flows crucially relies on the inf-convolution from convex analysis [Liero et al., 2018, Chizat, 2022]. There, the WFR metric tensor is defined using an inf-convolution of the Wasserstein tensor and the Hellinger (Fisher-Rao) tensor GWFR(µ) = GW (µ) GHe(µ). By Legendre transform, its inverse, the Onsager operator, is given by the sum KWFR(µ) = KW (µ) + KHe(µ). Therefore, we construct the IFT gradient flow by replacing the Hellinger (Fisher-Rao) tensor with the MMD tensor, as in (5). GIFT(µ) = GW (µ) GMMD(µ), KIFT(µ) = KW (µ) + KMMD(µ). (6) The MMD gradient flow equation is derived by Zhu and Mielke [2024] using the Onsager operator (5), µ = KMMD(µ)δF δµ [µ] = K 1 δF δµ [µ] . (7) Hence, we obtained the desired IFT gradient flow equation using (6). µ = αKW (µ)δF δµ [µ] βKMMD(µ)δF δµ [µ] = α div(µ δF δµ [µ]) β K 1 δF δµ [µ] . (8) Formally, the IFT gradient flow equation can also be viewed as a kernel-approximation to the Wasserstein-Fisher-Rao gradient flow equation, i.e., the reaction-diffusion equation (3). Corollary 3.1. Suppose R kσ(x, ) dµ = 1 and the kernel-weighted-measure converges to the Dirac measure kσ(x, ) dµ dδx as the bandwidth σ 0. Then, the IFT gradient flow equation (8) tends towards the WFR gradient flow equation as σ 0, i.e., the reaction-diffusion equation (3). Like the WFR gradient flow over non-negative measures, the gradient flow equation (8) and (7) are not guaranteed to stay within the probability measure space, i.e., total mass 1. This is useful in many applications such as chemical reaction systems. However, probability measures are often required for machine learning applications. We now provide a mass-preserving gradient flow equation that we term the spherical IFT gradient flow. The term spherical is used to emphasize that the flow stays within the probability measure, as in the spherical Hellinger distance [Laschos and Mielke, 2019]. To this end, we must first study spherical MMD flows over probability measures. Recall that (7) is a Hilbert space gradient flow (see [Ambrosio et al., 2005]) and does not stay within the probability space. Closely related, many works using kernel-mean embedding [Smola et al., 2007, Muandet et al., 2017] also suffer from this issue of not respecting the probability space. To produce a restricted (or projected) flow in P, our starting point is the MMD minimizing-movement scheme restricted to the probability space µk+1 argmin µ P F(µ) + 1 2η MMD2(µ, µk). (9) We now derive the following mass-preserving spherical gradient flows for the MMD and IFT. Proposition 3.2 (Spherical MMD and spherical IFT gradient flow equations). The spherical MMD gradient flow equation is given by (where 1 denotes the constant scalar) δµ [µ] R K 11 Consequently, the spherical IFT gradient flow equation is given by µ = α div(µ δF δµ [µ]) β K 1 δF δµ [µ] R K 11 Furthermore, those equations are mass-preserving, i.e., R µ = 0. So far, we have identified the gradient flow equations of interest. Now, we are ready to present our main theoretical results on the convergence of the IFT gradient flow via functional inequalities. For example, the logarithmic Sobolev inequality (LSI) log dµ L2(µ) c LSI DKL(µ|π) for some c LSI > 0 (LSI) is sufficient to guarantee the convergence of the pure Wasserstein gradient flow of the KL divergence energy, which governs the same dynamics as the Langevin equation. The celebrated Bakry-Émery Theorem [Bakry and Émery, 1985], is a cornerstone of convergence analysis for dynamical systems as it provides an explicit sufficient condition: suppose the target measure π is λ-log concave for some λ > 0, then the global convergence is guaranteed, i.e., π = e V dx and 2V λ Id = (LSI) with c LSI = 2λ = glob. exp. convergence. The question we answer below is whether the IFT gradient flow enjoys such favorable properties. Our starting point is the (Polyak-)Łojasiewicz type functional inequality. Theorem 3.3. Suppose the following Łojasiewicz type inequality holds for some c > 0, H c F(µ(t)) inf µ F(µ) (IFT-Łoj) for the IFT gradient flow, or δµ [µ] R K 11 H c F(µ(t)) inf µ F(µ) (SIFT-Łoj) for the spherical IFT gradient flow. Then, the energy F decays exponentially along the corresponding gradient flow, i.e., F(µ(t)) infµ F(µ) e ct (F(µ(0)) infµ F(µ)). To understand a specific gradient flow, one must delve into the detailed analysis of the conditions under which the functional inequalities hold instead of assuming them to hold by default. We provide such analysis for the IFT gradient flows next. 3.2 Global exponential convergence analysis MMD energy functional As discussed in the introduction, the MMD energy has been proposed as an alternative to the KL divergence energy for sampling by Arbel et al. [2019]2, where they assume the access to samples from the target measure yi π. However, the theoretical convergence guarantees under the MMD energy is a less-exploited topic. Those authors characterized a local decay behavior under the assumption that the µt must already be close to the target measure π for all t > 0. The assumptions they made are not only restrictive, but also difficult to check. There has also been no global convergence analysis. For example, Arbel et al. [2019] s Proposition 2 states that the MMD is non-increasing, which is not equivalent to convergence and is easily satisfied by other flows. The mathematical limitation is that the MMD is in general not guaranteed to be convex along the Wasserstein geodesics. In addition, our analysis also does not require the heuristic noise injection step as was required in their implementation. Mroueh and Rigotti [2020] also used the MMD energy but with a different gradient flow, which has been shown by Zhu and Mielke [2024] to be a kernel-regularized inf-convolution of the Allen-Cahn and Cahn-Hilliard type of dissipation. However, Mroueh and Rigotti [2020] s convergence analysis is not sufficient for establishing (global) exponential convergence as no functional inequality has been established there. In contrast, we now provide full global exponential convergence guarantees. We first provide an interesting property that will become useful for our analysis. Theorem 3.4. Suppose the driving energy is the squared MMD, F(µ) = 1 2 MMD2(µ, π) and initial datum µ0 P is a probability measure. Then, the spherical MMD gradient flow equation (10) coincides with the MMD gradient flow equation µ = (µ π) , (MMD-MMD-GF) whose solution is a linear interpolation between the initial measure µ0 and the target measure π, i.e., µt = e tµ0 + (1 e t)π. Furthermore, same coincidence holds for the spherical IFT and IFT gradient flow equation µ = α div µ Z 2k(x, ) d (µ π) (x) β (µ π) . (MMD-IFT-GF) The explicit solution to the ODE (MMD-MMD-GF) shows an exponential convergence to the target measure π along the (spherical) MMD gradient flow. The (spherical) IFT gradient flow equation (MMD-IFT-GF) differs from the Wasserstein gradient flow equation of [Arbel et al., 2019] by a linear term. This explains the intuition of why we can expect good convergence properties for the IFT gradient flow of the squared MMD energy. We exploit this feature of the IFT gradient flow to show global convergence guarantees for inference with the MMD energy. This has not been possible previously when confined to the pure Wasserstein gradient flow. 2We do not use the terminology MMD gradient flow from [Arbel et al., 2019] since it is inconsistent with the naming convention of Wasserstein gradient flow as Wasserstein refers to the geometry, not the functional. Theorem 3.5 (Global exponential convergence of the IFT flow of the MMD energy). Suppose the energy F is the squared MMD energy F(µ) = 1 2 MMD2(µ, ν). Then, the (IFT-Łoj) holds globally with a constant c 2β > 0. Consequently, for any initialization within the non-negative measure cone µ0 M+, the squared MMD energy decays exponentially along the IFT gradient flow of non-negative measures, i.e., 1 2 MMD2(µt, ν) e 2βt 1 2 MMD2(µ0, ν). (12) Furthermore, if the initial datum µ0 and the target measure π are probability measures µ0, π P, then the squared MMD energy decays exponentially globally along the spherical IFT gradient flow, i.e., the decay estimate (12) holds along the spherical IFT gradient flow of probability measures. We emphasize that no Bakry-Émery type or kernel conditions are required the Łojasiewicz inequality holds globally when using the IFT flow. In contrast, the Wasserstein-Fisher-Rao gradient flow µ = α div µ Z 2k(x, ) d (µ π) (x) β Z k(x, ) d (µ π) (x) (MMD-WFR-GF) does not enjoy such global convergence guarantees. Global exponential convergence under the KL divergence energy For variational inference and MCMC, a common choice of the energy is the KL divergence energy, i.e., F(µ) = DKL(µ|π). This has already been studied by a large body of literature, including the case of Wasserstein-Fisher Rao [Liero et al., 2023, Lu et al., 2019]. Not surprisingly, (LSI) is still sufficient for the exponential convergence of the WFR type of gradient flows since the dissipation of the Wasserstein part alone is sufficient for driving the system to equilibrium. For the IFT gradient flows under the KL divergence energy functional, the convergence can still be established. This showcases the strength of the IFT geometry it enjoys the best of both worlds. The IFT gradient flow equation of the KL divergence energy reads µ = α div(µ log dµ dπ ) β K 1 log dµ dπ . Unlike the MMD-energy flow case, the spherical IFT gradient flow of the KL over probability measures P no longer coincides with that of the (non-spherical) IFT and is given by µ = α div(µ log dµ R K 1 log dµ Proposition 3.6 (Exponential convergence of the SIFT gradient flow of the KL divergence energy). Suppose the (LSI) holds with c LSI = 2λ or the target measure π is λ-log concave for some λ > 0. Then, the KL divergence energy decays exponentially globally along the spherical IFT gradient flow (13) ,i.e., DKL(µt|π) e 2αλt DKL(µ0|π). The intuition behind the above result is that the SIFT gradient flow converges whenever the pure Wasserstein gradient flow, i.e., its convergence is at least as fast as the Wasserstein gradient flow. However, we emphasize that the decay estimate of the KL divergence energy only holds along the spherical IFT flow over probability measures P, but not the full IFT flow over non-negative measures M+. 3.3 Minimizing movement, JKO-splitting, and a practical particle-based algorithm In applications to machine learning and computation, continuous-time flow can be discretized via the JKO scheme [Jordan et al., 1998], which is based on the minimizing movement scheme (MMS) [De Giorgi, 1993]. For the reaction-diffusion type gradient flow equation in the Wasserstein Fisher-Rao setting, the JKO-splitting a.k.a. time-splitting scheme has been studied by Gallouët and Monsaingeon [2017], Mielke et al. [2023]. This amounts to splitting the diffusion (Wasserstein) and reaction (MMD) step in (8), i.e., at time step ℓ 1 2 argmin µ P F(µ) + 1 2τ W 2 2 (µ, µℓ), (Wasserstein step) µℓ+1 argmin µ P F(µ) + 1 2η MMD2(µ, µℓ+ 1 2 ). (MMD step) (14) A similar JKO-splitting scheme can also be constructed via the WFR gradient flow, which amounts to replacing the MMD step in (14) with a proximal step in the KL (as an approximation to the Hellinger), i.e., µℓ+1 argminµ P F(µ)+ 1 ηDKL(µ|µℓ+ 1 2 ), which is well-studied in the optimization literature as the entropic mirror descent [Nemirovskij and Yudin, 1983]. Our MMD step can also be viewed as a mirror descent step with the mirror map 1 2 K 2 H. However, for the task of MMD inference of [Arbel et al., 2019], WFR flow does not possess convergence guarantees such as our Theorem 3.4. The MMD step can also be easily implemented as in our simulation. We summarize the resulting overall IFT particle gradient descent from the JKO splitting scheme in Algorithm 1 in the appendix. We now look at those two steps respectively. For concreteness, we consider a flexible particle approximation to the probability measures, with possibly non-uniform weights allocated to the particles, i.e., µ = Pn i=1 αiδxi, α n, xi X. Wasserstein step: particle position update. (14) is a standard JKO scheme; see, e.g., [Santambrogio, 2015]. The optimality condition of the Wasserstein proximal step can be implemented using a particle gradient descent algorithm xℓ+1 i = xℓ i τ δF δµ [µℓ](xℓ i), i = 1, ..., n, (15) which is essentially the algorithm proposed by Arbel et al. [2019] when F(µ) = 1 2 MMD2(µ, π). MMD step: particle weight update. The MMD step in (14) is a discretization step of the spherical MMD gradient flow, as shown in (9) and Proposition 3.2. We propose to use the updated particle location xℓ+1 i from the Wasserstein step (15) and update the weights βi by solving i=1 βiδxℓ+1 i ) + 1 i=1 βiδxℓ+1 i , i=1 αℓ iδxℓ+1 i ), i.e., the MMD step only updates the weights. Alternatively, as in the classical mirror descent optimization, one can use a linear approximation F(µ) F(µℓ) + δF δµ µℓ , µ µℓ L2. We also provide a specialized discussion on the MMD-energy minimization task of [Arbel et al., 2019]. Let the energy objective be the squared MMD F(µ) := 1 2 MMD(µ, π)2. In this setting, we are given the particles sampled from the target measure yi π. For the MMD step in (14), the computation is drastically simplified to an MMD barycenter problem, which was also studied in [Cohen et al., 2021]. This amounts to solving a convex quadratic program with a simplex constraint; see the appendix for the detailed expression. 4 Numerical Example The overall goal of the numerical experiments is to approximate the target measure π by minimizing the squared MMD energy, i.e., min µ A P MMD2(µ, π). In all the experiments, we have access to the target measure π in the form of samples yi π. This setting was studied in [Arbel et al., 2019] as well as in many deep generative model applications. In the following experiments, we compare the performance of our proposed algorithm of IFT gradient flow, which implements the JKO-splitting scheme in (14) and is detailed in Algorithm 1, to that of (1) Arbel et al. [2019] s the MMD flow (see our discussion in 2), we used their algorithm both with and without a heuristic noise injection suggested by those authors; (2) the Wasserstein-Fisher-Rao flow of the MMD (MMD-WFR-GF). The WFR flow was also used by Yan et al. [2024], Lu et al. [2023] but for minimizing the KL divergence function. As discussed in 3.2, the MMD flow of [Arbel et al., 2019] does not possess global convergence guarantees while IFT does. Furthermore, in the Gaussian mixture target experiment, the target measure π is not log-concave. We emphasize that our convergence guarantee still holds for the IFT flow while there is no decay guarantee for the WFR flow. We provide the code for the implementation at https://github.com/egorgladin/ift_flow. Gaussian target in 2D experiment Figures 3(a) and 4 showcase the performance of the algorithms in a setting where µ0 and π are both Gaussians in 2D. Specifically, µ0 N(5 1, I) and π N 0, 1 1/2 1/2 2 . The number of samples drawn from µ0 and π was set to n = 100. A Gaussian 0 1000 2000 3000 4000 5000 6000 Steps MMD flow MMD flow + noise IFT GD (a) Gaussian target experiment 0 1000 2000 3000 4000 5000 6000 Steps MMD flow MMD flow + noise IFT GD (b) Gaussian mixture target experiment Figure 3: Mean loss and standard deviation computed over 50 runs 2 0 2 4 6 8 10 12 Step #0 Step #4000 (a) Vanilla MMD flow 2 0 2 4 6 8 10 12 Step #0 Step #4000 (b) MMD flow + noise 2 0 2 4 6 8 10 12 Step #0 Step #4000 (c) IFT (Algorithm 1) Figure 4: Trajectory of a randomly selected subsample produced by different algorithms in the Gaussian target experiment. Color intensity indicates points weights. The hollow dots indicate the particles that have already vanished. kernel with bandwidth σ = 10 was used. For all three algorithms, we chose the largest stepsize that didn t cause unstable behavior, τ = 50. The parameter η in (23) was set to 0.1. As can be observed from the trajectories produced by MMD flow (Figure 4(a)), most points collapse into small clusters near the target mode. Some points drift far away from the target distribution and get stuck; the resulting samples represent the target distribution poorly, which is a sign of suboptimal solution. MMD flow with the heuristic noise injection produces much better results. We suspect the noise 2 0 2 4 6 8 10 Step #0 Step #5000 (a) Vanilla MMD flow 2 0 2 4 6 8 10 Step #0 Step #5000 (b) MMD flow + noise 2 0 2 4 6 8 10 Step #0 Step #5000 (c) IFT (Algorithm 1) Figure 5: Trajectory of a randomly selected subsample produced by different algorithms in the Gaussian mixture experiment. Color intensity indicates points weights. The hollow dots indicate the particles that have already vanished. helps to escape local minima; however, the injection needs to be heuristically tuned. However, it takes a large number iterations for points to get close to locations with high density of the target distribution. Similarly to the previous research on noisy MMD flow, we use a relatively large noise level (10) in the beginning and turn off the noise after a sufficient number of iterations (3000 in our case). A drawback of this approach is that the right time for noise deactivation depends on the particular problem instance, which makes the algorithm behavior less predictable. 0 2000 4000 6000 8000 Steps MMD flow MMD flow + noise IFT GD WFR GD Figure 6: Comparison with the WFR flow of the MMD in 100 dimensions Algorithm 1 achieves a similar accuracy to that of the noise-injected MMD flow, but much faster already after 1000 steps without any heuristic noise injection. For the few particles that did not make it close to the target Gaussian s mean, their mass is teleported to those particles that are close to the target. Hence, the resulting performance of the IFT algorithm does not deteriorate. The hollow dots in the trajectory plot indicate the particles whose mass has been teleported and hence their weights are zero. This is a major advantage of unbalanced transport for dealing with local minima. For a faster implementation, in the implementation of the MMD step in (14), we only perform a single step of projected gradient descent instead of computing a solution to the auxiliary optimization problem (23). To be fair in comparison, we count each iteration as two steps. Thus, 6000 steps of the algorithm in Figure 3(a) correspond to only 3000 iterations, i.e., the results for IFT algorithm have already been handicapped by a factor of 2. We would also like to note that Algorithm 1 was executed with constant hyperparameters without further tuning over the iterations, in contrast to the noisy MMD flow. In practical implementations, it is possible to further improve the performance by sampling new locations (particle rejuvenation) in the MMD step (14) as done similarly in [Dai et al., 2016]. Since this paper is a theoretical one and not about competitive benchmarking, we leave this for future work. Gaussian mixture in 2D experiment The second experiment has a similar setup. However, this time the target is a mixture of equally weighted Gaussian distributions, N 0, 1 1/2 1/2 2 , N 3 1 , I , N 1 4 , 3 1/2 1/2 1 . Figures 3(b) and 5 showcase loss curves and trajectories produced by the considered algorithms. WFR flow for Gaussian mixture target in 100D We conducted an experiment in dimension d = 100, comparing the IFT flow with the WFR flow of the MMD energy. The initial distribution is N(0, I), and the target π is a mixture of 3 distributions N(mi, Σi), i = 1, 2, 3, where mi and Σi are randomly generated such that mi 2 = 20 and the smallest eigenvalue of Σi is greater than 0.5. For fairness, each iteration of IFT particle GD (as well as its version with KL step) counts as two steps, i.e., these methods only performed 4500 iterations. In the noisy MMD flow, the noise is disabled after 4000 steps. All methods are used with equal stepsize. 5 Discussion In summary, the (spherical) IFT gradient flows are a suitable choice for energy minimization of both the MMD and KL divergence energies with sound global exponential convergence guarantees. There is also an orthogonal line of works studying the Stein gradient flow and descent [Liu and Wang, 2019, Duncan et al., 2019], which also has the mechanics interpretation of repulsive forces. It can be related to our work in that the IFT gradient flow has a (de-)kernelized reaction part, while the Stein flow has a kernelized diffusion part. Furthermore, there is a work [Manupriya et al., 2024] that proposes a static MMD-regularized Wasserstein distance, which should not be confused with our IFT gradient flow geometry. Another future direction is sampling and inference when we do not have access to the samples from the target distribution π, but can only evaluate its score function log π. Acknowledgments and Disclosure of Funding We thank Gabriel Peyré for the helpful comments regarding the practical algorithms for the JKOsplitting scheme. This project has received funding from the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany s Excellence Strategy The Berlin Mathematics Research Center MATH+ (EXC-2046/1, project ID: 390685689) and from the priority programme "Theoretical Foundations of Deep Learning" (SPP 2298, project number: 543963649). During part of the project, the research of E. Gladin was prepared within the framework of the HSE University Basic Research Program. L. Ambrosio, N. Gigli, and G. Savare. Gradient Flows: In Metric Spaces and in the Space of Probability Measures. Springer Science & Business Media, 2005. M. Arbel, A. Korba, A. SALIM, and A. Gretton. Maximum mean discrepancy gradient flow. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips.cc/paper_files/paper/2019/file/ 944a5ae3483ed5c1e10bbccb7942a279-Paper.pdf. M. Arjovsky, S. Chintala, and L. Bottou. Wasserstein generative adversarial networks. In International conference on machine learning, pages 214 223. PMLR, 2017. D. Bakry and M. Émery. Diffusions hypercontractives. In J. Azéma and M. Yor, editors, Séminaire de Probabilités XIX 1983/84, volume 1123, pages 177 206. Springer Berlin Heidelberg, Berlin, Heidelberg, 1985. ISBN 978-3-540-15230-9 978-3-540-39397-9. doi: 10.1007/BFb0075847. J. Bigot, R. Gouet, T. Klein, and A. López. Geodesic PCA in the Wasserstein space by convex PCA. Ann. Inst. H. Poincaré Probab. Statist., 53(1):1 26, 02 2017. doi: 10.1214/15-AIHP706. URL https://doi.org/10.1214/15-AIHP706. J. A. Carrillo, K. Craig, and F. S. Patacchini. A blob method for diffusion. Calculus of Variations and Partial Differential Equations, 58:1 53, 2019. S. Chewi, T. Le Gouic, C. Lu, T. Maunu, and P. Rigollet. SVGD as a kernelized Wasserstein gradient flow of the chi-squared divergence. Advances in Neural Information Processing Systems, 33: 2098 2109, 2020. L. Chizat. Sparse optimization on measures with over-parameterized gradient descent. Mathematical Programming, 194(1-2):487 532, 2022. L. Chizat, G. Peyré, B. Schmitzer, and F.-X. Vialard. An interpolating distance between optimal transport and Fisher Rao metrics. Foundations of Computational Mathematics, 18(1):1 44, Feb. 2018. ISSN 1615-3375, 1615-3383. doi: 10.1007/s10208-016-9331-y. L. Chizat, G. Peyré, B. Schmitzer, and F.-X. Vialard. Unbalanced optimal transport: Dynamic and Kantorovich formulation. ar Xiv:1508.05216 [math], Feb. 2019. URL http://arxiv.org/abs/ 1508.05216. ar Xiv: 1508.05216. S. Cohen, M. Arbel, and M. P. Deisenroth. Estimating barycenters of measures in high dimensions. ar Xiv:2007.07105 [cs, stat], Feb. 2021. K. Craig, K. Elamvazhuthi, M. Haberland, and O. Turanova. A blob method for inhomogeneous diffusion with applications to multi-agent control and sampling. Mathematics of Computation, 92 (344):2575 2654, Nov. 2023. ISSN 0025-5718, 1088-6842. doi: 10.1090/mcom/3841. M. Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. Advances in neural information processing systems, 26, 2013. B. Dai, N. He, H. Dai, and L. Song. Provable Bayesian inference via particle mirror descent. In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics, pages 985 994. PMLR, May 2016. E. De Giorgi. New problems on minimizing movements. Ennio de Giorgi: Selected Papers, pages 699 713, 1993. A. Duncan, N. Nüsken, and L. Szpruch. On the geometry of Stein variational gradient descent. ar Xiv preprint ar Xiv:1912.00894, 2019. M. A. Efendiev and A. Mielke. On the rate-independent limit of systems with dry friction and small viscosity. Journal of Convex Analysis, 13(1):151, 2006. J. Feydy, T. Séjourné, F.-X. Vialard, S.-i. Amari, A. Trouve, and G. Peyré. Interpolating between optimal transport and MMD using Sinkhorn divergences. In Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, pages 2681 2690. PMLR, Apr. 2019. T. O. Gallouët and L. Monsaingeon. A JKO splitting scheme for Kantorovich Fisher Rao gradient flows. SIAM Journal on Mathematical Analysis, 49(2):1100 1130, 2017. P. Glaser, M. Arbel, and A. Gretton. KALE flow: A relaxed KL gradient flow for probabilities with disjoint support. In Neural Information Processing Systems, June 2021. A. Gretton, K. M. Borgwardt, M. J. Rasch, B. Schölkopf, and A. Smola. A kernel two-sample test. The Journal of Machine Learning Research, 13(1):723 773, 2012. J. Hertrich, C. Wald, F. Altekrüger, and P. Hagemann. Generative sliced MMD flows with Riesz kernels. ar Xiv preprint ar Xiv:2305.11463, 2023. N. Ho, X. Nguyen, M. Yurochkin, H. H. Bui, V. Huynh, and D. Phung. Multilevel clustering via Wasserstein means. In D. Precup and Y. W. Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 1501 1509, International Convention Centre, Sydney, Australia, 06 11 Aug 2017. PMLR. URL http://proceedings.mlr.press/v70/ho17a.html. R. Jordan, D. Kinderlehrer, and F. Otto. The variational formulation of the Fokker Planck equation. SIAM journal on mathematical analysis, 29(1):1 17, 1998. Publisher: SIAM. S. Kondratyev, L. Monsaingeon, and D. Vorotnikov. A new optimal transport distance on the space of finite Radon measures. Advances in Differential Equations, 21(11/12):1117 1164, 2016. doi: 10.57262/ade/1476369298. URL https://doi.org/10.57262/ade/1476369298. A. Korba, P.-C. Aubin-Frankowski, S. Majewski, and P. Ablin. Kernel Stein discrepancy descent. In Proceedings of the 38th International Conference on Machine Learning, pages 5719 5730. PMLR, July 2021. M. J. Kusner, Y. Sun, N. I. Kolkin, and K. Q. Weinberger. From word embeddings to document distances. In Proceedings of the 32nd International Conference on International Conference on Machine Learning - Volume 37, ICML 15, pages 957 966. JMLR.org, 2015. URL http: //dl.acm.org/citation.cfm?id=3045118.3045221. S. Lacoste-Julien, F. Lindsten, and F. Bach. Sequential kernel herding: Frank-Wolfe optimization for particle filtering. In G. Lebanon and S. V. N. Vishwanathan, editors, Proceedings of the Eighteenth International Conference on Artificial Intelligence and Statistics, volume 38 of Proceedings of Machine Learning Research, pages 544 552, San Diego, California, USA, 09 12 May 2015. PMLR. URL https://proceedings.mlr.press/v38/lacoste-julien15.html. V. Laschos and A. Mielke. Geometric properties of cones with applications on the Hellinger Kantorovich space, and a new distance on the space of probability measures. J. Funct. Analysis, 276(11):3529 3576, 2019. doi: 10.1016/j.jfa.2018.12.013. M. Liero, A. Mielke, and G. Savaré. Optimal entropy-transport problems and a new Hellinger Kantorovich distance between positive measures. Inventiones mathematicae, 211 (3):969 1117, Mar. 2018. ISSN 0020-9910, 1432-1297. doi: 10.1007/s00222-017-0759-8. URL http://link.springer.com/10.1007/s00222-017-0759-8. M. Liero, A. Mielke, and G. Savaré. Fine properties of geodesics and geodesic λ-convexity for the Hellinger Kantorovich distance. Arch. Rat. Mech. Analysis, 247(112):1 73, 2023. doi: 10.1007/s00205-023-01941-1. Q. Liu and D. Wang. Stein variational gradient descent: A general purpose Bayesian inference algorithm. ar Xiv:1608.04471 [cs, stat], Sept. 2019. URL http://arxiv.org/abs/1608.04471. ar Xiv: 1608.04471. Y. Lu, J. Lu, and J. Nolen. Accelerating Langevin sampling with birth-death. Ar Xiv, May 2019. Y. Lu, D. Slepˇcev, and L. Wang. Birth death dynamics for sampling: global convergence, approximations and their asymptotics. Nonlinearity, 36(11):5731, 2023. P. Manupriya, S. N. Jagarlapudi, and P. Jawanpuria. MMD-regularized unbalanced optimal transport. Transactions on Machine Learning Research, 2024. ISSN 2835-8856. URL https://openreview.net/forum?id=e N9Cj U3h1b. A. Mielke, R. Rossi, and A. Stephan. On time-splitting methods for gradient flows with two dissipation mechanisms. ar Xiv preprint ar Xiv:2307.16137, 2023. P. Mohajerin Esfahani and D. Kuhn. Data-driven distributionally robust optimization using the Wasserstein metric: Performance guarantees and tractable reformulations. Mathematical Programming, 171(1-2):115 166, Sept. 2018. ISSN 0025-5610, 1436-4646. doi: 10.1007/s10107-017-1172-1. Y. Mroueh and M. Rigotti. Unbalanced Sobolev descent, Sept. 2020. URL http://arxiv.org/ abs/2009.14148. ar Xiv:2009.14148 [cs, stat]. K. Muandet, K. Fukumizu, B. Sriperumbudur, and B. Schölkopf. Kernel mean embedding of distributions: A review and beyond. Foundations and Trends in Machine Learning, 10(1-2): 1 141, 2017. ISSN 1935-8237, 1935-8245. doi: 10.1561/2200000060. A. S. Nemirovskij and D. B. Yudin. Problem Complexity and Method Efficiency in Optimization. John Wiley, New York, 1983. S. Neumayer, V. Stein, and G. Steidl. Wasserstein gradient flows for Moreau envelopes of fdivergences in reproducing kernel Hilbert spaces. ar Xiv preprint ar Xiv:2402.04613, 2024. F. Otto. Double degenerate diffusion equations as steepest descent. Bonn University, 1996. Preprint. Y. Rubner, C. Tomasi, and L. J. Guibas. The earth mover s distance as a metric for image retrieval. International journal of computer vision, 40(2):99 121, 2000. R. Sandler and M. Lindenbaum. Nonnegative matrix factorization with earth mover s distance metric for image analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 33(8): 1590 1602, 2011. ISSN 0162-8828. doi: 10.1109/TPAMI.2011.18. F. Santambrogio. Optimal Transport for Applied Mathematicians: Calculus of Variations, PDEs, and Modeling. Springer International Publishing, 2015. ISBN 9783319208282. doi: 10.1007/ 978-3-319-20828-2. URL http://dx.doi.org/10.1007/978-3-319-20828-2. A. Sinha, H. Namkoong, R. Volpi, and J. Duchi. Certifying some distributional robustness with principled adversarial training. ar Xiv:1710.10571 [cs, stat], May 2020. A. Smola, A. Gretton, L. Song, and B. Schölkopf. A Hilbert space embedding for distributions. In M. Hutter, R. A. Servedio, and E. Takimoto, editors, Algorithmic Learning Theory, pages 13 31, Berlin, Heidelberg, 2007. Springer. ISBN 978-3-540-75225-7. doi: 10.1007/978-3-540-75225-7_ 5. J. Solomon, R. M. Rustamov, L. Guibas, and A. Butscher. Wasserstein propagation for semisupervised learning. In Proceedings of the 31st International Conference on International Conference on Machine Learning - Volume 32, ICML 14, pages I 306 I 314. JMLR.org, 2014. URL http://dl.acm.org/citation.cfm?id=3044805.3044841. I. Steinwart and A. Christmann. Support vector machines. Springer Science & Business Media, 2008. I. Tolstikhin, B. Sriperumbudur, and K. Muandet. Minimax estimation of kernel mean embeddings. ar Xiv:1602.04361 [math, stat], July 2017. I. O. Tolstikhin, B. K. Sriperumbudur, and B. Schölkopf. Minimax estimation of maximum mean discrepancy with radial kernels. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. URL https://proceedings.neurips.cc/paper_files/paper/2016/ file/5055cbf43fac3f7e2336b27310f0b9ef-Paper.pdf. A. B. Tsybakov. Introduction to Nonparametric Estimation. Springer Series in Statistics. Springer, New York, NY, 2009. ISBN 978-0-387-79051-0 978-0-387-79052-7. doi: 10.1007/b13794. URL https://link.springer.com/10.1007/b13794. C. Villani. Optimal transport: old and new, volume 338. Springer Science & Business Media, 2008. M. Werman, S. Peleg, and A. Rosenfeld. A distance metric for multidimensional histograms. Computer Vision, Graphics, and Image Processing, 32(3):328 336, 1985. ISSN 0734-189X. doi: https://doi.org/10.1016/0734-189X(85)90055-6. URL http://www.sciencedirect.com/ science/article/pii/0734189X85900556. Y. Yan, K. Wang, and P. Rigollet. Learning Gaussian mixtures using the Wasserstein-Fisher-Rao gradient flow. The Annals of Statistics, 52(4):1774 1795, 2024. J.-J. Zhu and A. Mielke. Kernel approximation of Fisher-Rao gradient flows, 2024. URL https: //arxiv.org/abs/2410.20622. A Appendix: proofs and additional technical details Proof of Proposition 3.2. We first consider the MMS step (9). The Lagrangian of the MMS step is, for λ R, L(µ, λ) = F(µ) + 1 2η MMD2(µ, µℓ) + λ( Z µ 1). The Euler-Lagrange equation gives δF η K(µ µℓ) + λ = 0, (16) Z µ = 1. (17) Rewriting the first equation, µ = µℓ ηK 1 δF δµ µℓ ηλK 1. Integrating both sides, we obtain 1 = 1 η Z K 1 δF δµ µℓ ηλ Z K 11 = λ = Let the time step η 0 in the first equation in the Euler-Lagrange equation (17), we obtain δµ [µ] + λ) = K 1 δF δµ [µ] R K 11 K 11, (18) which is the desired spherical MMD gradient flow equation. Spherical IFT gradient flow equation is obtained by an inf-convolution [Gallouët and Monsaingeon, 2017, Liero et al., 2018, Chizat et al., 2019] of the above spherical MMD and Wasserstein part. The verification of the mass-preserving property is by a straightforward integration of (18) 0 = Z µ = Z K 1 δF δµ [µ] R K 11 Z K 11 = 0. Hence, the theorem is proved. Proof of Theorem 3.3. The proof amounts to identifying the correct left-hand side of the Łojasiewicz type inequality. We take the time derivative of the energy d dt F(µ) = δF δµ [µ] , α div(µ δF δµ [µ]) β K 1 δF δµ [µ] 2 L2µ β δF δµ [µ] 2 H, which is the desired left-hand side of the Łojasiewicz type inequality. For the spherical IFT gradient flow, we have d dt F(µ) = δF δµ [µ] , α div(µ δF δµ [µ]) β K 1 δF δµ [µ] R K 11 δµ [µ] , α div(µ δF δµ [µ]) L2+ δF δµ [µ] R K 11 , β K 1 δF δµ [µ] R K 11 δµ [µ] R K 11 where the second equality follows from the fact that the spherical IFT gradient flow is masspreserving. Hence, the left-hand side of the Łojasiewicz type inequality is obtained. Proof of Corollary 3.1. The formal proof is by using well-known results for kernel smoothing in non-parametric statistics [Tsybakov, 2009]. Note that the gradient flow equation can be rewritten as u = α div(µ δF δµ [µ]) + β K 1 δF δµ [µ] = α div(µ δF δµ [µ]) + β µK 1 µ δF Recall the definition of the integral operator Kf = Z kσ( , y)f(y) dy, Kµf = Z kσ( , y)f(y) dµ(y). (19) Formally, as kσ(x, ) dµ dδx, we have Kµξ ξ for any ξ L2(µ). Then, the IFT gradient flow equation tends towards the PDE µ = α div(µ δF δµ [µ]) + βµ δF δµ [µ] . (20) Furthermore, we also have δF δµ [µ] 2 H δF δµ [µ] 2 L2(µ). Hence, the conclusion follows. Proof of Theorem 3.4. We first recall that the first variation of the squared MMD energy is given by δ δµ 2 MMD2(µ, ν) [µ] = Z k(x, )(µ ν)( dx). (21) Plugging the first variation of the squared MMD energy into the gradient flow equation (7), (10), (8), and (11), we obtain the desired flow equations in the theorem. The ODE solution is obtained by an elementary argument. Proof of Theorem 3.5. We take the time derivative of the energy and apply the chain rule formally and noting the gradient flow equation (8), d dt F(µ) = Z δF δµ [µ] , α div(µ δF δµ [µ]) β K 1 δF δµ [µ] 2 L2 µ β δF δµ [µ] 2 H β δF δµ [µ] 2 H. Plugging in F(µ) = 1 2 MMD2(µ, ν), an elementary calculation shows that 2 MMD2(µ, ν) β Z k(x, )(µ ν)( dx) 2 H = 2β 1 2 MMD2(µ, ν), (22) which establishes the desired Łojasiewicz inequality specialized to the squared MMD energy, which reads α Z k(x, )(µ ν)( dx) Z k(x, )(µ ν)( dx) 2 MMD2(µ, ν). (Łoj) By Grönwall s lemma, exponential decay is established. Furthermore, plugging the first variation of the squared MMD energy into the gradient flow equation (10), the extra term in the spherical flow equation becomes R K 1 δF δµ [µ] R K 11 = R K 1K(µ π) R K 11 = 0. Hence, the coincidence is proved. Proof of Proposition 3.6. By the Bakry-Émery Theorem, we have the LSI (LSI) hold with c LSI = 2λ. Taking the time derivative of the KL divergence energy along the SIFT gradient flow, we have d dt DKL(µt|π) = log dµt = α log dµt R K 1 log dµt H (LSI) 2αλ DKL(µt|π) + 0. By Grönwall s lemma, exponential convergence is established. Note that this result does not hold for the full IFT flows over non-negative measures as there exists no LSI globally on M+. Remark A.1 (Regularized inverse of the integral operator). Strictly speaking, the integral operator K is compact and hence its inverse is unbounded. Using a viscosity-regularization techniques by Efendiev and Mielke [2006], we can obtain the flow equation where the inverse is always well-defined, i.e., µ = α div(µ δF δµ [µ]) β (K + ϵ I) 1 δF δµ [µ]. This corresponds to an additive regularization of the kernel Gram matrix in practical computation. A particle gradient descent algorithm for IFT gradient flows We use the notation KXX to denote the kernel Gram matrix KXX = [k(xℓ+1 i , xℓ+1 j )]n i,j=1, KX X for the cross kernel matrix KX X = [k(xℓ+1 i , xℓ j)]n i,j=1, etc. Algorithm 1 A JKO-splitting for IFT particle gradient descent 1: for ℓ= 1 to T 1 do 2: Compute the first variation of the energy F at µℓ: gℓ= δF δµ µℓ . Then, xℓ+1 i xℓ i τ ℓ gℓ(xℓ i), i = 1, ..., n (Wasserstein step) αℓ+1 argmin α n F( i=1 αiδxℓ+1 i ) + 1 2ηℓ KXX KX X KX X K X X 3: end for 4: Output the particle measure bµT = Pn i=1 αT i δx T i . Implementing the MMD step The MMD step in the JKO-splitting scheme is a convex quadratic program with a simplex constraint, which can be formulated as inf β n 1 2 MMD2( i=1 βiδxℓ+1 i , π) + 1 i=1 βiδxℓ+1 i , i=1 αℓ iδxℓ+1 i ). (23) We further expand the optimization objective (multiplied by a factor of 2τ for convenience) i=1 βiϕ(xk+1 i ) 1 i=1 βiϕ(xk+1 i ) i=1 αk i ϕ(xk i ) = τ β KXXβ 2 mβ KXY 1 + 1 m2 1 KY Y 1 + β KXXβ 2β KX Xα + α K X Xα = (1 + τ) β KXXβ 2τ m β KXY 1 2β KX Xα + τ m2 1 KY Y 1 + α K X Xα. (24) Therefore, the MMD step in Algorithm 1 can be implemented as a convex quadratic program with a simplex constraint. A particle gradient descent algorithm for the WFR flow of the MMD energy We provide the implementation details of the WFR flow of the MMD energy. The goal is to simulate the PDE (MMD-WFR-GF). To the best of our knowledge, there has been no prior implementation of this flow. Nor is there a convergence guarantee. Similar to the JKO-splitting scheme of the IFT flow in (14), (MMD-WFR-GF) can be discretized using the two-step scheme 2 arg min µ P F(µ) + 1 2τ W 2 2 (µ, µℓ) (Wasserstein step) µℓ+1 argmin µ P F(µ) + 1 η KL(µ, µℓ+ 1 2 ) (KL step) where the energy function F is the squared MMD energy, F(µ) = 1 2 MMD2(µ, π). Use the explicit Euler scheme, the KL step amounts to the entropic mirror descent. In the optimization literature, this step can be implemented as multiplicative update of the weights (or density), i.e., suppose xℓ+1 i is the new particle location after the Wasserstein step, then we update the weights vector α via αℓ+1 i αℓ i exp η δF δµ [µℓ](xℓ+1 i ) . Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: Yes, our claims in the abstract and introduction do accurately reflect the paper s contributions and scope. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: In various places in the paper, we discuss the limitations of our work. Especially in the discussion/conclusion section, we discuss the limitations of our approach. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: Yes. For example, one focus of our work is precisely to study under what assumptions, functional inequalities such as the Łojasiewicz inequality and the log-Sobolev inequality hold. We do not avoid discussing such assumptions. We provide a complete proof for each theoretical result. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: Our implementation is standard and also similar to that of [Arbel et al., 2019] and [Korba et al., 2021]. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: Similar to the answer to the previous question. We have submitted the code and added to the text a link to a github repo with the details of the experiments. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: Ours is a theory paper with small-scale experiments. We provide all the necessary details in the main text, see Sect. 4. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: Yes, we have reported error bars in our experiments. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: Our experiments are small-scale and can be reproduced on a standard laptop. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: Our work is more on the theoretical side and, thus, we do not see any direct societal impact or potential harmful consequences. The research process did not use any personal data and did not involve human subjects or participants. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: Our work is more on the theoretical side and, thus, we do not see any direct societal impact. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: We do not release any data or models that have a risk for misuse. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: No specific assets are used in our work. We have cited the related work with code and data. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: We do not introduce any new assets. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: We did not use crowdsourcing and did not make research with Human Subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: We did not make research with Human Subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.