# constrained_sampling_with_primaldual_langevin_monte_carlo__f6003d02.pdf Constrained Sampling with Primal-Dual Langevin Monte Carlo Luiz F. O. Chamon University of Stuttgart luiz.chamon@simtech.uni-stuttgart.de Mohammad Reza Karimi ETH Zürich mkarimi@inf.ethz.ch Anna Korba CREST, ENSAE, IP Paris anna.korba@ensae.fr This work considers the problem of sampling from a probability distribution known up to a normalization constant while satisfying a set of statistical constraints specified by the expected values of general nonlinear functions. This problem finds applications in, e.g., Bayesian inference, where it can constrain moments to evaluate counterfactual scenarios or enforce desiderata such as prediction fairness. Methods developed to handle support constraints, such as those based on mirror maps, barriers, and penalties, are not suited for this task. This work therefore relies on gradient descent-ascent dynamics in Wasserstein space to put forward a discretetime primal-dual Langevin Monte Carlo algorithm (PD-LMC) that simultaneously constrains the target distribution and samples from it. We analyze the convergence of PD-LMC under standard assumptions on the target distribution and constraints, namely (strong) convexity and log-Sobolev inequalities. To do so, we bring classical optimization arguments for saddle-point algorithms to the geometry of Wasserstein space. We illustrate the relevance and effectiveness of PD-LMC in several applications. 1 Introduction Sampling is a fundamental task in statistics, with applications to estimation and decision making, and of growing interest in machine learning (ML), motivated by the need for uncertainty quantification and its success in generative tasks [1 3]. In these settings, the distribution we wish to sample from (target distribution) is often known only up to its normalization constant. This is the case, for instance, of score functions learned from data or posterior distributions of complex Bayesian models. Markov Chain Monte Carlo (MCMC) algorithms can be used to tackle this problem [4, 5] and Langevin Monte Carlo (LMC), in particular, has attracted considerable attention due to its simplicity, theoretical grounding, and effectiveness in practice [3, 6 9]. These sampling algorithms, however, do not naturally incorporate requirements on the samples they generate. Specifically, standard MCMC methods do not enforce restrictions on the target distributions, such as support (e.g., truncated Gaussian), conditional probabilities (e.g., fairness), or moments (e.g., portfolio return) constraints. This limitation is often addressed by post-processing, transforming variables, or by introducing penalties in the target distribution. Though successful in specific settings, these approaches have considerable downsides. Post-processing techniques such as rejection sampling (see, e.g., [10, 11]) may substantially reduce the effective number of samples (number of samples generated per iteration of the algorithm). Variable transformations based on link functions, projections, or mirror/proximal maps (see, e.g., [12 18]) only accommodate (deterministic) support constraints and are not suited 38th Conference on Neural Information Processing Systems (Neur IPS 2024). Table 1: Type and target of sampling constraints. Soft constraint Hard constraint Sample (x µ) Mirror/proximal LMC [12, 14, 15, 17], projected LMC [13], barriers [16, 18] Distribution (µ) Penalized LMC [23] PD-LMC for statistical requirements such as robustness or fairness [19 22]. Though modifying the target distribution directly offers more flexibility (see, e.g., [23]), it does not guarantee that constraints are satisfied (Table 1). We refer the reader to Appendix A for a more detailed literature review. This paper overcomes these issues by directly tackling the constrained sampling problem. Explicitly, it seeks to sample not from a target distribution π on Rd, but from the distribution µ that solves P min µ P2(Rd) KL(µ π) subject to Ex µ gi(x) 0, i = 1, . . . , I, Ex µ hj(x) = 0, j = 1, . . . , J, where P2(Rd) denotes the set of probability measures on Rd with bounded second moments and the functions gi, hj represent the requirements. Note that (PI) only considers measures µ against which gi, hj are integrable. Otherwise, the expectations are taken to be + , making the corresponding measure infeasible. Observe that (PI) is more general than the support-constrained sampling problem considered in, e.g., [12 18]. Indeed, it constrains the distribution µ rather than its samples x (Table 1). Algorithms based on projections, barrier, or mirror maps are not suited for this type of constraints (see Section 2.2 for more details). To tackle (PI), this paper instead derives and analyzes a primal-dual LMC algorithm (PD-LMC) that is the sampling counterpart of gradient descent-ascent (GDA) methods from (Euclidean) optimization. A dual ascent algorithm was previously proposed to tackle (PI), but it requires the exact computation of expectations with respect to intractable distributions [24]. This paper not only overcomes this limitation, but also provides convergence guarantees for a broader class of constraint functions. The main contributions of this work include: a discrete-time constrained sampling algorithm (PD-LMC, Algorithm 1) for solving (PI) that precludes any explicit integration (Section 3); an analysis of PD-LMC proving that it converges sublinearly (in expectation) with respect to the Kullback-Leibler divergence (convex case) or Wasserstein distance (strongly convex case). The analysis is performed directly on the discrete-time iterations and requires only local Lipschitz continuity and bounded variance assumptions (Section 3.1); an extension of these result for target distributions satisfying a log-Sobolev inequality (LSI) for a variant of PD-LMC (Algorithm 2, Section 3.2); numerical examples illustrating applications of (PI) and the effectiveness of PD-LMC (Section 4). 2 Problem formulation 2.1 Background on Langevin Monte Carlo Consider a target distribution π P2(Rd) absolutely continuous with respect to Lebesgue measure whose density (also denoted π) can be expressed as π(x) = e f(x)/Z for some normalization constant Z. Define the Kullback-Leibler (KL) divergence of µ with respect to π as KL(µ π) = Z log dµ dµ = Z fdµ + Z log(µ)dµ log(Z) V(µ) + H(µ) log(Z), (1) dπ is the Radon Nikodym derivative, V is the potential energy, and H is the negative entropy if µ is absolutely continuous with respect to π; and + otherwise. For a wide class of functions f (e.g., smooth and strongly convex), samples from π can be obtained from the path of the Langevin diffusion process, whose instantaneous values x(t) have distributions µ(t) evolving according to the Fokker Planck equation [25]. Explicitly, dx(t) = f(x(t))dt + 2d W(t) and µ(t) t = µ(t) W2KL(µ(t) π) , (2) for a d-dimensional Brownian motion W(t), where q denotes the divergence of q and W2 KL(µ π) denotes the Wasserstein-2 gradient of KL( |π) at µ [26, Theorem 10.4.17] (see Appendix B for more details). Indeed, the Langevin diffusion (2) brings the distribution µ(t) of x(t) progressively closer to the target π. In fact, the Fokker-Planck equation can be interpreted as a gradient flow of the KL divergence with respect to the Wasserstein-2 distance [7, 25]. However, computing the path of the stochastic differential equation in (2) is not practical and discretizations are used instead. Chief among them is the (forward) Euler Maruyama scheme, which leads to the celebrated Langevin Monte Carlo (LMC) algorithm [6] xk+1 = xk γk f(xk) + p 2γkβk, βk iid N(0, Id), (3) for a step size γk > 0, where Id denotes the d-dimensional identity matrix. Notice that it is not necessary to know Z in order to evaluate (3). This has made LMC and its variants widely popular in practice and the subject of extensive research. Despite (3) being a biased time-discretizations of the Langevin diffusion in (2) [7], rates of convergence of LMC have been obtained for smooth and strongly convex [8, 27] or convex [28] potentials or when the target distribution π verifies an LSI [29]. 2.2 Constrained sampling Our goal, however, is not to sample from π itself, but from a distribution close to π that satisfies a set of statistical requirements. Explicitly, we wish to sample from a distribution µ that solves (PI). Since (PI) constrains the distribution µ rather than its samples x, it can accommodate more general requirements than the support constraints typically considered in constrained sampling (e.g., [12 18]). Next, we illustrate the wide range of practical problems that can be formulated as (PI). These examples are further explored in Section 4 and more details on their formulations are provided in Appendix E. 1. Sampling from convex sets: though we have stressed that (PI) accommodates other types of requirements, it can also be used to constrain the support of π, i.e., to sample from µ argmin µ P2(Rd) KL(µ π) subject to Px µ[x C] = 1, (PII) for a closed convex set C Rd. Indeed, let C be the intersection of the 0-sublevel sets of convex functions {si}i=1,...,I. Such a description always exists (see Appendix E). Then, (PII) can be cast as (PI) using gi(x) = [si(x)]+ for [z]+ = max(0, z). Notice that the gi are convex and that although they are not everywhere differentiable, I(si(x) > 0) si(x) is a subgradient of gi, where I(E) = 1 on the event E and 0 otherwise. Observe that support constraints can also be imposed using projections, mirror/proximal maps, and barriers as in [12 18]. These methods, however, constrain the samples x rather than their distribution µ as in (PII). 2. Rate-constrained Bayesian models: rate constraints have garnered attention in ML due to their central role in fairness [20, 21]. Consider data pairs (x, y), where x X are features and y {0, 1} labels, and a protected (measurable) subgroup G X. Let π be a Bayesian posterior of the parameters θ of a model q( ; θ) denoting the probability of a positive outcome (based, e.g., on a binomial model). We wish to enforce statistical parity, i.e., we wish the prevalence of positive outcomes within the protected group G to be close to or higher than in the whole population. We cast this problem as (PI) by constraining the average probability of positive outcome as in P = min µ P2(Rd) KL(µ π) subject to Ex,θ µ q(x; θ) | G Ex,θ µ q(x; θ) δ. (PIII) where δ > 0 denotes our tolerance [22]. Naturally, multiple protected groups can be accommodated by incorporating additional constraints. Hence, constrained sampling provides a natural way to encode fairness in Bayesian inference. 3. Counterfactual sampling: rather than imposing requirements on probabilistic models, constrained sampling can also be used to probe them by evaluating counterfactual statements. Indeed, let π denote a reference probabilistic model such that sampling from π yields realizations of the real world. Consider the counterfactual statement how would the world have been if E[g(x)] 0? Constrained sampling not only gives realizations of this alternative world, but it also indicates its compatibility with the reference model, namely the value P of (PI). More concretely, consider a Bayesian stock market model. Here, π is a posterior model for the (log-)returns of I assets, e.g., distributed as Gaussians N(ρ, Σ). Here, the vector ρ describes the mean return of each stock and Σ their covariance. We can investigate what the market would look like if, e.g., the mean and variance of each stocks were to change by solving P = min µ P2(Rd) KL(µ π) subject to E(ρ,Σ) µ ρi = ρi, i = 1, . . . , I E(ρ,Σ) µ Σii σ2 i Due to correlations in the market, certain choices of ρi or σ2 i may be more unrealistic than others. Additionally, it could be that some of these conditions are vacuous conditioned on the others. As we show next, our approach to tackling (PI) effectively isolates the contribution of each requirement in the solution µ , thus enabling us to identify which are (conditionally) vacuous and which are most at odds with the reference model π. 2.3 Lagrangian duality and dual ascent algorithms Although directly sampling from µ does not appear straightforward, it admits a convenient characterization based on convex duality that is amenable to be sampled using the LMC algorithm (3). Indeed, let g : Rd RI and h : Rd RJ be vector-valued functions collecting the constraint functions gi and hj respectively. The Lagrangian of (PI) is then defined as L(µ, λ, ν) KL(µ π) + λ Eµ[g] + ν Eµ[h] = KL(µ µλν) + log Z for λ RI + and ν RJ, where µλν(x) = e U(x,λ,ν) Zλν for U(x, λ, ν) = f(x) + λ g(x) + ν h(x) (5) and a normalization constant Zλν. Notice that P = minµ maxλ 0, ν L(µ, λ, ν), which is why (PI) is referred to as the primal problem. To obtain the dual problem of (PI), define the dual function d(λ, ν) min µ P2(Rd) L(µ, λ, ν). (6) Notice from (4) that the minimum in (6) is achieved for µλν from (5), the Lagrangian minimizer, so that d(λ, ν) = log(Z/Zλν). The solution of (6) is therefore a tilted version of π, whose tilt is controlled by the dual variables (λ, ν). Since (6) is a relaxation of (PI), it yields a lower bound on the primal value, i.e., d(λ, ν) P for all (λ, ν) RI + RJ. The dual problem seeks the tilts (λ , ν ) that yield the best lower bound, i.e., D max λ RI +, ν RJ d(λ, ν). (DI) The set Φ = argmaxλ 0, ν d(λ, ν) of solutions of (DI) is called the set of Lagrange multipliers. Note from (6) that (DI) depends on the distributions µ and π through its objective d. The dual problem (DI) has several advantageous properties. Indeed, while the primal problem (PI) is an infinite dimensional, smooth optimization problem in probability space, the dual problem (DI) is a finite dimensional, non-smooth optimization problem in Euclidean space. What is more, it is a concave problem regardless of the functions f, g, h, since the dual function (6) is the minimum of a set of affine functions in (λ, ν) [30, Prop. 4.1.1]. These properties are all the more attractive given that, under mild conditions stated below, (DI) can be used to solve (PI). Assumption 2.1. There exists µ P2(Rd) with KL(µ π) C < such that Eµ [gi] δ < 0 and Eµ [hj] = 0 for all i, j. Proposition 2.2. Under Assumption 2.1, the following holds: (i) P = D ; (ii) there exists a finite pair (λ , ν ) Φ ; (iii) for any solution µ of (PI) and (λ , ν ) of (DI), it holds that L(µ , λ, ν) L(µ , λ , ν ) L(µ, λ , ν ), for all (µ, λ, ν) P2(Rd) RI + RJ; (7) (iv) the solution of (PI) is µ = µλ ν for (λ , ν ) Φ ; (v) consider the perturbation of (PI) P (u, v) min µ P2(Rd) KL(µ π) subject to Ex µ gi(x) ui, Ex µ hj(x) = vj. (PV) Then, (λ , ν ) are subgradients of P (0, 0) = P , i.e., P (u, v) P λ u ν v, and if P (u, v) is differentiable at (0, 0), then u P (u, v) = λ and v P (u, v) = ν at (0, 0). Proof. In finite dimensional settings, (i) (v) are well-known duality results (see, e.g., [30]). While they also hold for infinite dimensional optimization problems, their proofs are slightly more scattered. We collect their reference below. The objective of (PI) is a convex function and its constraints are linear functions of µ. Hence, (PI) is a convex program. Under Slater s condition (Assumption 2.1), it is (i) strongly dual (P = D ) and (ii) there exists at least one solution (λ , ν ) of (DI) (see [31, Sec. 8.6, Thm. 1] or [32, Cor. 4.1]). This implies (iii) the existence of the saddle-point (7) [33, Prop. 2.156], (iv) that µ argminµ L(µ, λ , ν ) = {µλ ν }, since the KL divergence is strongly convex and its minimizer is unique [34, Thm. 7.3.7], and (v) that (λ , ν ) are subgradients of the perturbation function P (u, v) [33, Prop. 4.27]. Proposition 2.2 shows that given solutions (λ , ν ) of (DI), the constrained sampling problem (PI) reduces to sampling from µλ ν e U( ,λ ,ν ) (see Appendix F for an explicit example of this result). It is important to note that this results only relies on the KL divergence being (strongly) convex in the standard L2 geometry, i.e., along mixtures of the form tµ0 + (1 t)µ1 for t [0, 1]. This does not imply that it is (geodesically) convex in the Wasserstein sense [35, Section 9.1.2]. This would require U in (5) to be convex for all λ 0 and ν RJ, i.e., for f, g to be convex and h to be linear. Hence, Proposition 2.2 reduces the constrained sampling problem (PI) to that of finding the Lagrange multipliers (λ , ν ). Despite their finite dimensionality, however, computing these parameters is intricate. Indeed, since (DI) is a concave program, we could obtain (λ , ν ) Φ using λk+1 = λk + ηk Eµλkνk [g] + and νk+1 = νk + ηk Eµλkνk [h], (8) for ηk > 0, where we used the fact that Eµλν[g] and Eµλν[h] are (sub)gradients of the dual function (6) at (λ, ν) [36, Thm. 2.87]. This procedure is known in optimization and game theory as dual ascent or best response [37, 38]. Notice, however, that (8) is not a practical algorithm as it requires explicit integration with respect to the intractable distribution µλν from (5). This issue was partially addressed in [24] (in continuous time and without equality constraints, i.e., J = 0) by replacing the Lagrangian minimizer µλkνk in (8) by the distribution of LMC samples, as in xk+1 = xk γk U(xk, λk) + p 2γkβk, βk iid N(0, Id) λk+1 = h λk + ηk Eµk[g] i +, µk = Law(xk). (9) Note that since J = 0, we omit the argument ν of U for clarity. Nevertheless, the updates in (9) still require an explicit integration. While it is now possible to sample from µk (namely, using the xk), empirical approximations of Eµk[g] may not only require an exponential (in the dimension d) number of samples (e.g., [39, Thm. 1.2]), but it introduces errors that are not taken into account in the analysis of [24]. In the sequel, we address these drawbacks by replacing these dual ascent algorithms by a saddle-point one. Algorithm 1 Primal-dual LMC 1: Inputs: ηk > 0 (step size), x0 µ0, and (λ0, ν0) = (0, 0). 2: for k = 0, . . . , K 1 3: xk+1 = xk ηk x U(xk, λk, νk) + p 2ηk βk, for βk N(0, Id) 4: λk+1 = λk + ηkg(xk) + 5: νk+1 = νk + ηkh(xk) 6: end 3 Primal-dual Langevin Monte Carlo Consider the GDA dynamics for the saddle-point problem (DI) in Wasserstein space. Explicitly, µ(t) t = µ(t) W2L µ(t), λ(t), ν(t) (10a) t = λi L µ(t), λ(t), ν(t) λi(t),+ (10b) t = νj L µ(t), λ(t), ν(t) (10c) for the Lagrangian L defined in (4), where [z]λ,+ = z for λ > 0 and [z]λ,+ = max(a, 0) otherwise [40, Sec. 2.2]. Observe that λi L(µ, λ, ν) = Eµ[gi] and νj L(µ, λ, ν) = Eµ[hj]. Hence, the algorithm from [24] described in (9) involves a deterministic implementation of (10b) that fully integrates over µ(t). In contrast, we consider a stochastic, single-particle implementation of (10) that leads to the practical procedure in Algorithm 1. Explicitly, we also use an Euler-Maruyama time-discretization of the Langevin diffusion associated to (10a) (step 3), but replace the expectations in (10b) (10c) by single-sample approximations (steps 4 5). Algorithm 1 can therefore be seen as a particle implementation of the deterministic Wasserstein GDA algorithm (10). As such, it resembles a primal-dual counterpart of the LMC algorithm in (3), which is why we dub it primal-dual LMC (PD-LMC). Alternatively, Algorithm 1 can be interpreted as a stochastic approximation of the dual ascent method in (9). This suggests that the gradient approximations in steps 4 5 could be improved using mini-batches, which is in fact how [24] approximates the expectation in (9). Our theoretical analysis and experiments show that these mini-batches are neither necessary nor always worth the additional computational cost (see Section 3.1 and Section 4). Note that the stochastic approximations in Algorithm 1 refer to the dual updates (steps 4 5) rather than the LMC update (step 3) as in stochastic gradient Langevin [41]. Though these methods could be combined, it is beyond the scope of this work. The remainder of this section is dedicated to analyzing the convergence properties of PD-LMC for both stochastic dual gradients (as in Algorithm 1) and exact dual gradient (as in (9)). For the latter, we obtain guarantees for the discrete implementation (9) under weaker assumptions than the continuous-time analysis of [24]. We consider strongly log-concave target distributions in Section 3.1 and those satisfying an LSI in Section 3.2. 3.1 PD-LMC with (strongly) convex potentials As opposed to the traditional LMC algorithm (3) or the deterministic updates in (9), Algorithm 1 involves three coupled random variables, namely, (xk, λk, νk). Hence, the LMC update (step 3) is based on a stochastic potential U and the distribution µk of xk is now a random measure. Our analysis sidesteps this obstacle by using techniques from stochastic optimization. We also leverage techniques from primal-dual algorithms in the Wasserstein space, in the spirit of works such as [7, 8, 42] that studied the LMC (3) or alternative time-discretizations of gradient flows of the KL divergence as splitting schemes. First, define the potential energy E and the (negative) entropy H for (µ, λ, ν) P2(Rd) RI + RJ as E(µ, λ, ν) = Z U(x, λ, ν)dµ(x) and H(µ) = Z log(µ)dµ, (11) for U as in (5). Notice from (4) that L(µ, λ, ν) = E(µ, λ, ν) + H(µ) log(Z), where we used the KL divergence decomposition in (1). To proceed, consider the following assumptions: Assumption 3.1. The potential energy E(µ, λ, ν) in (11) is m-strongly convex with respect to µ along Wasserstein-2 geodesics for m 0 and all (λ, ν) RI + RJ. Explicitly, E(µ, λ, ν) E(µ0, λ, ν) + Z W2E(µ0, λ, ν), x y ds(x, y) + m 2 W 2 2 (µ, µ0), where s is an optimal coupling achieving W 2 2 (µ, µ0) (see Appendix B). Assumption 3.2. The gradients and variances of f, g, h are bounded along iterations {µk}k 0, where µk is the distribution of xk, i.e., there exists G2 such that max f 2 L2(µk), gi 2 L2(µk), hj 2 L2(µk) G2 and max Eµk[ g 2], Eµk[ h 2] G2. Assumption 3.1 holds with m = 0 if f, g are convex and h is linear. If f is additionally strongly convex, then it holds with m > 0 [26, Prop. 9.3.2]. Assumption 3.2 is typical in (stochastic) nonsmooth optimization analyses (see, e.g., [36, 43, 44]). Notice, however, that gradients are only required to be bounded along trajectories of Algorithm 1, a crucial distinction in the case of strongly convex functions whose gradients can only be bounded locally. Assumption 3.2 can be satisfied under mild conditions on f, g, h, such as local Lipschitz continuity or linear growth. The following theorem provides the first convergence analysis of the discrete-time PD-LMC. Theorem 3.3. Denote by µk the distribution of xk in Algorithm 1. Under Assumptions 2.1, 3.1, and 3.2, there exists R2 0 such that, for ηk η, h KL(µk µ )+ m 2 W 2 2 (µk, µ ) i 3ηG2 + R2 0 ηK + ηG2 E[ λk 2]+E[ νk 2] . (12) For ηk R0/(G k) and ηk = ηk/ PK k=1 ηk, we obtain k=1 ηk h KL(µk µ )+ m 2 W 2 2 (µk, µ ) i R0G(1 + log(K)) 3+max k E[ λk 2]+E[ νk 2] . (13) Additionally, there exists a sequence of step sizes ηk > 0 such that W 2 2 (µk, µ ) R2 0 and E[ λk 2]+ E[ νk 2] < for all k. The same results hold (without expectations) when using exact dual gradients, i.e., if the updates in steps 4 5 are replaced by λk+1 = λk+ηk Eµk[g] + and νk+1 = νk+ηk Eµk[h]. Theorem 3.3, whose proof is deferred to Appendix C, implies rates similar to those for GDA schemes in finite-dimensional Euclidean optimization (see, e.g., [43]). To recover those rates, however, we must bound the magnitudes of λk, νk. In [43], this is done by bounding the iterates in the algorithm itself, i.e., by projecting them onto the set Dr = {(λ, ν) RI + RJ | max( λ 2 , ν 2) r} and choosing r such that Φ Dr (Proposition 2.2(ii) ensures this is possible). We then incur a bias on the order of η in (12) that vanishes in the decreasing step size setting of (13). Though convenient, this is not necessary since there exists a sequence of step sizes such that both E[ λk 2] and E[ νk 2] are bounded for all k 0. In the interest of generality, Theorem 3.3 holds without these hypotheses. It is worth noting that though faster rates and last iterates guarantees can be obtained for Euclidean saddle-point problems, they rely on more complex schemes than the GDA in Algorithm 1 involving acceleration or proximal methods [45 48]. The results in Theorem 3.3 are stated for the stochastic scheme in Algorithm 1. However, Theorem 3.3 yields the same rates (without expectations) for exact dual gradients, i.e., for the dual ascent scheme (9). In this case, the second condition in Assumption 3.2 simplifies to maxk( Eµk[g] 2, Eµk[h] 2) G2. Not only are these milder assumptions than [24, Eq. (16)], but the guarantees hold for discreterather than continuous-time dynamics. Finally, (12) (13) imply convergence with respect to the KL divergence for convex potentials (m = 0) with stronger guarantees in Wasserstein metric for strongly convex ones (m > 0). The convergence rates for distributions µk from Theorem 3.3 also imply convergence rates for empirical averages across iterates xk of Algorithm 1. This corollary is obtained by combining (12) (13) with the following proposition. By taking φ to be the constraint functions g or h from (PI) yields feasibility guarantees for PD-LMC . Algorithm 2 (Stochastic) dual LMC 1: Inputs: N 0 k > 0 (burn-in), γk, ηk > 0 (step sizes), and λ0 = 0. 2: for k = 0, . . . , K 1 3: x0 µ0 4: xn+1 = xn γk x U(xn, λk) + p 2γk βn, βn N(0, Id), for n = 0, . . . , N 0 k 1 5: λk+1 = λk + ηkg(x N0 k) Proposition 3.4. Consider samples xk distributed according to µk and ck 0 with PK k=1 ck = 1 such that PK k=1 ck h KL(µk µ ) + m 2 W 2 2 (µk, µ ) i K for m 0. Then, it holds that k=1 ckφ(xk) 2 K, if φ is bounded by 1, r m , if φ is 1-Lipschitz and m > 0. Proof. See Appendix C. 3.2 PD-LMC with LSI potentials In this section, we replace Assumption 3.1 on the convexity of the potential by an LSI common in the sampling literature. We consider only inequality constraints (J = 0) here and omit the function arguments ν, since accounting for equality constraints requires significant additional assumptions. Assumption 3.5. The distribution µλ satisfies the LSI for bounded λ, i.e., there exists σ > 0 such that 2σ KL(ζ µλ) log (dζ/dµλ) 2 L2(ζ) for all ζ P2(Rd). The LSI in Assumption 3.5 is often used in the analysis of the standard LMC algorithm [29, 49]. It holds, e.g., when f is strongly convex and g is a (possibly non-convex) bounded function due to the Holley-Stroock perturbation theorem [50]. In fact, if f is 1-strongly convex and |g| is bounded by 1, then Assumption 3.5 holds for σ e 2λ (see, e.g., [51, Prop. 5.1.6] or [52, Thm 1.1]). The LSI is akin to the Polyak-Łojasiewicz (PL) condition from Euclidean optimization [53], which supposes issues with GDA methods such as Algorithm 1. Indeed, it is not enough for the Lagrangian (4) to satisfy the PL condition in the primal variable to guarantee the convergence of GDA in Euclidean spaces. We must either modify Algorithm 1 using acceleration or proximal methods [47, 54 56] or impose the PL condition also on λ [57, 58]. Since the Lagrangian 4 is linear in λ, it is clear that Algorithm 1 will not suffice to provide theoretical guarantees in the LSI case. We therefore consider the variant in Algorithm 2, where N 0 k LMC iterations (step 3) are executed before updating the dual variables (step 4). This is akin to using different time-scales in continuoustime, a common technique for solving saddle-point problems [54, 58]. Since it resembles a dual ascent counterpart of the LMC algorithm (3), we refer to it as (stochastic) dual LMC (DLMC). As opposed to the dual ascent algorithm from [24] in (9), however, Algorithm 2 does not require any explicit evaluation of expected values. The following theorem provides an analysis of its convergence. Theorem 3.6. Assume that the functions f, g are M-smooth, i.e., have M-Lipschitz continuous gradients, satisfy Assumption 3.5, and that Eµ[ g 2] G2 for all µ P2(Rd). Let 0 < ηk η, 0 < ϵ ηG2 < 1, γk = γ σϵ 16d M 2 , and N 0 k 1 γσ log 2 KL(µ0 µλk) Under Assumption 2.1, there exists B < such that the distributions {µk} of the samples {x N0 k} generated by Algorithm 2 satisfy k=0 KL(µk µ ) ϵ + ηG2 Recall from (PI) that I is the number of inequality constraints. Additionally, E[ λk 1] is bounded for all k. 1.0 1.5 2.0 2.5 3.0 0 1.0 1.5 2.0 2.5 3.0 1.0 1.5 2.0 2.5 3.0 Figure 1: Sampling from a 1D truncated Gaussian (ground truth displayed as dashed lines). PD-LMC Proj. LMC Mirror LMC Rejection Figure 2: Sampling from a 2D truncated Gaussian (true mean in red and sample mean in orange). Theorem 3.6, whose proof is deferred to Appendix D, provides similar guarantees as (approximate) subgradient methods in finite-dimensional optimization (see, e.g., [22, 56]). This is not surprising seen as γk, N 0 k in Theorem 3.6 are chosen to ensure that step 4 yields a sample xk µk such that KL( µk µλk) ϵ using [29, Theorem 1]. At this point, g(x N 0 k) in step 5 is an approximate, stochastic subgradient of the dual function (6). Though it may appear from (12) and (14) that Algorithms 1 and 2 have the same convergence rates, an informal computation shows that the latter evaluates on the order of dκ2/η as many gradient per iteration, where κ = M/σ. Note that we can once again apply Theorem 3.6 to derive ergodic average and feasibility guarantees for Algorithm 2. 4 Experiments We now return to the applications described in Section 2.2 to showcase the behavior of PD-LMC. We defer implementation details and additional results to Appendix E. Code for these examples is publicly available at https://www.github.com/lfochamon/pdlmc. 1. Sampling from convex sets. We cast the problem of sampling from a Gaussian distribution N(0, 1) truncated to C = [1, 3] as (PI) by taking f(x) = x2/2 and g(x) = [(x 1)(x 3)]+ (see Section 2.2). Fig. 1 shows histograms for the samples obtained using PD-LMC, the projected LMC (Proj. LMC) from [13], and the mirror LMC from [59], all with the same step size. Both Proj. LMC and Mirror LMC generate an excess of samples close to the boundary (between 1.5 and 3 times more samples than expected). This leads to an underestimation of the mean (Proj. LMC: 1.488 / Mirror LMC: 1.470 vs. true mean: 1.510). In contrast, PD-LMC provides a more accurate estimate (1.508). Yet, since it constrains the distribution µ rather than its samples, it is not an interior-point method and can produce samples outside of C. Theorems 3.3 3.6 show that this becomes less frequent as the algorithm progresses (in Fig. 1, only 2% of the samples are not in C). This occurs even without using minibatches in steps 4 5 of Algorithm 1 as in [24]. In fact, our experiments show that mini-batches increase the computational complexity with no performance benefit (Appendix E). These issues are exacerbated in more challenging problems, such as sampling from a two-dimensional standard Gaussian centered at [2, 2] restricted to an unit ℓ2-norm ball (Fig. 2). In this case, Proj. LMC places almost 25% of its samples on the boundary (where only 0.14% of samples should be), while PD-LMC only places 1.8% of its samples outside of the support. Mirror LMC provides a better mean estimation in this setting, although a bit more asymmetric than PD-LMC [Mirror LMC: (0.312, 0.418) vs. PD-LMC: (0.446, 0.444) vs. true mean: (0.368, 0.368)]. 2. Rate-constrained Bayesian models. Here, we consider π to be the posterior of a Bayesian logistic regression model for the Adult dataset from [60], where the goal is to predict whether an individual 0 10 20 30 Prevalence of > $50k (%) Constrained (δ = 0.03) Constrained (δ = 0.01) Unconstrained Female Male Figure 3: Distribution of the probability of predicting > 50k under the Bayesian logistic model (black lines indicate the mean across genders). 0 1 2 3 4 5 6 Iteration 105 Dual variable (ν) Figure 4: Counterfactual sampling of the stock market: dual variables. makes more than $50k based on socioeconomic information (details on data pre-processing can be found in [61]). We consider a standard Gaussian prior on the parameters θ Rd+1 of the model, where d is the number of features. Using the LMC algorithm to sample from the posterior (i.e., no constraints), we find that while the average probability of positive predictions is 19.1% over the whole test set, it is 26.2% among males and 5% among females ( Unconstrained in Fig. 3). To overcome this disparity, we take gender to be the protected class in (PIII), constraining both Gmale and Gfemale with δ = 0.01. Using PD-LMC, we obtain a Bayesian model that leads to an average probability of positive outcomes of 18.1% and 15.1% for males and females respectively. In fact, we now observe a substantial overlap of the distributions of positive predictions across genders for the constrained posterior µ ( Constrained (δ = 0.01) in Fig. 3). This substantial reduction of prediction disparities comes at only a minor decline in accuracy (unconstrained: 84% vs constrained: 82%). 3. Counterfactual sampling. Though the distribution of positive predictions changes considerably for both male and female individuals, the final dual variables (λmale = 0 and λfemale 160) show that these changes are due uniquely to the female group [as per Prop. 2.2(iv)]. This implies that the reference model π is itself compatible with the requirement for the male group, but that reducing the disparity for females requires considerable deviations from it. By examining λfemale, we conclude without recalculating µ that even small changes in the tolerance δ for the female constraint would substantially change the distribution of outcomes [Prop. 2.2(v)]. This is confirmed by Constrained (δ = 0.03) in Fig. 3. Notice that this is only possible due to the primal-dual nature of PD-LMC. This type of counterfactual analysis is even more beneficial in the presence of multiple requirements. Indeed, let π be the posterior of a Bayesian model for the daily (log-)return of a set of assets (see Appendix E for more details). Using (PIV), we consider how the market would look like if the average (log-)return of each asset were to have been (exactly) 20% higher. Inspecting the dual variables (Fig. 4), we notice that this increased market return is essentially driven by two stocks: NVDA and LLY (ν < 0). In fact, the reference model π would be consistent with an even higher increase for JNJ and GOOG (ν > 0). We confirm these observations by constraining only NVDA and LLY, which yields essentially the same (log-)return distribution for all assets. 5 Conclusion We tackled the problem of sampling from a target distribution while satisfying a set of statistical constraints. Based on a GDA method in Wasserstein space, we put forward a fully stochastic, discretetime primal-dual LMC algorithm (PD-LMC) that precludes any explicit integration in its updates. We analyze the behavior of PD-LMC for (strongly) convex and log-Sobolev potentials, proving that the distribution of its samples converges to the optimal constrained distribution. We illustrated the use of PD-LMC for different constrained sampling applications. Future work include strengthening the convergence results to almost sure guarantees and improving the rates obtained using proximal and extra gradient methods, particularly in the LSI setting. Acknowledgments and Disclosure of Funding The work of L.F.O. Chamon is funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany s Excellence Strategy (EXC 2075-390740016). [1] M. F. Faulkner and S. Livingstone, Sampling algorithms in statistical physics: a guide for statistics and machine learning, ar Xiv preprint ar Xiv:2208.04751, 2022. [2] R. van de Schoot, S. Depaoli, R. King, B. Kramer, K. Märtens, M. G. Tadesse, M. Vannucci, A. Gelman, D. Veen, J. Willemsen et al., Bayesian statistics and modelling, Nature Reviews Methods Primers, vol. 1, no. 1, p. 1, 2021. [3] Y. Song and S. Ermon, Generative modeling by estimating gradients of the data distribution, Advances in neural information processing systems, vol. 32, 2019. [4] G. O. Roberts and J. S. Rosenthal, General state space Markov chains and MCMC algorithms, Probability Surveys, vol. 1, 2004. [5] C. Robert and G. Casella, Monte Carlo statistical methods. Springer Verlag, 2004. [6] G. O. Roberts and R. L. Tweedie, Exponential convergence of Langevin distributions and their discrete approximations, Bernoulli, pp. 341 363, 1996. [7] A. Wibisono, Sampling as optimization in the space of measures: The Langevin dynamics as a composite optimization problem, in Conference on Learning Theory. PMLR, 2018, pp. 2093 3027. [8] A. Durmus, S. Majewski, and B. Miasojedow, Analysis of Langevin Monte Carlo via convex optimization, The Journal of Machine Learning Research, vol. 20, no. 1, pp. 2666 2711, 2019. [9] Y. Wang and W. Li, Accelerated information gradient flow, J. Sci. Comput., vol. 90, no. 1, 2022. [10] L. Lang, W.-s. Chen, B. R. Bakshi, P. K. Goel, and S. Ungarala, Bayesian estimation via sequential Monte Carlo sampling Constrained dynamic systems, Automatica, vol. 43, no. 9, pp. 1615 1622, 2007. [11] Y. Li and S. K. Ghosh, Efficient sampling methods for truncated multivariate normal and student-t distributions subject to linear inequality constraints, Journal of Statistical Theory and Practice, vol. 9, pp. 712 732, 2015. [12] Y.-P. Hsieh, A. Kavis, P. Rolland, and V. Cevher, Mirrored Langevin dynamics, Advances in Neural Information Processing Systems, vol. 31, 2018. [13] S. Bubeck, R. Eldan, and J. Lehec, Sampling from a log-concave distribution with projected Langevin Monte Carlo, Discrete & Computational Geometry, vol. 59, no. 4, pp. 757 783, 2018. [14] A. Salim and P. Richtarik, Primal dual interpretation of the proximal stochastic gradient Langevin algorithm, in Advances in Neural Information Processing Systems, 2020, pp. 3786 3796. [15] K. Ahn and S. Chewi, Efficient constrained sampling via the mirror-Langevin algorithm, Advances in Neural Information Processing Systems, vol. 34, pp. 28 405 28 418, 2021. [16] Y. Kook, Y. Lee, R. Shen, and S. Vempala, Sampling with Riemannian Hamiltonian Monte Carlo in a constrained space, in Advances in Neural Information Processing Systems, A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho, Eds., 2022. [17] L. Sharrock, L. Mackey, and C. Nemeth, Learning rate free Bayesian inference in constrained domains, in Conference on Neural Information Processing Systems, 2023. [18] M. Noble, V. De Bortoli, and A. Durmus, Unbiased constrained sampling with self-concordant barrier Hamiltonian Monte Carlo, in Thirty-seventh Conference on Neural Information Processing Systems, 2023. [19] A. M adry, A. Makelov, L. Schmidt, D. Tsipras, and A. Vladu, Towards deep learning models resistant to adversarial attacks, in International Conference on Learning Representations, 2018. [20] M. Kearns, S. Neel, A. Roth, and Z. S. Wu, Preventing fairness Gerrymandering: Auditing and learning for subgroup fairness, in International Conference on Machine Learning, 2018, pp. 2564 2572. [21] A. Cotter, H. Jiang, M. Gupta, S. Wang, T. Narayan, S. You, and K. Sridharan, Optimization with non-differentiable constraints with applications to fairness, recall, churn, and other goals, Journal of Machine Learning Research, vol. 20, no. 172, pp. 1 59, 2019. [22] L. F. O. Chamon, S. Paternain, M. Calvo-Fullana, and A. Ribeiro, Constrained learning with non-convex losses, IEEE Trans. on Inf. Theory, vol. 69[3], pp. 1739 1760, 2023. [23] M. Gürbüzbalaban, Y. Hu, and L. Zhu, Penalized Langevin and Hamiltonian Monte Carlo Algorithms for Constrained Sampling, 2022. [24] X. Liu, X. Tong, and Q. Liu, Sampling with trusthworthy constraints: A variational gradient framework, Advances in Neural Information Processing Systems, vol. 34, pp. 23 557 23 568, 2021. [25] R. Jordan, D. Kinderlehrer, and F. Otto, The variational formulation of the Fokker Planck equation, SIAM Journal on Mathematical Analysis, vol. 29, no. 1, pp. 1 17, 1998. [26] L. Ambrosio, N. Gigli, and G. Savaré, Gradient flows: in metric spaces and in the space of probability measures. Springer Science & Business Media, 2005. [27] A. S. Dalalyan and A. Karagulyan, User-friendly guarantees for the Langevin Monte Carlo with inaccurate gradient, Stochastic Processes and their Applications, vol. 129, no. 12, pp. 5278 5311, 2019. [28] A. S. Dalalyan, A. Karagulyan, and L. Riou-Durand, Bounding the error of discretized Langevin algorithms for non-strongly log-concave targets, Journal of Machine Learning Research, vol. 23, no. 235, pp. 1 38, 2022. [29] S. Vempala and A. Wibisono, Rapid convergence of the unadjusted Langevin algorithm: Isoperimetry suffices, in Advances in Neural Information Processing Systems (Neur IPS), 2019, pp. 8092 8104. [30] D. P. Bertsekas, Convex Optimization Theory. Athena Scientific, 2009. [31] D. G. Luenberger, Optimization by Vector Space Methods. Wiley, 1968. [32] V. Jeyakumar and H. Wolkowicz, Generalizations of slater s constraint qualification for infinite convex programs, Math. Program., vol. 57, no. 1 3, pp. 85 101, 1992. [33] J. F. Bonnans and A. Shapiro, Perturbation Analysis of Optimization Problems. Springer, 2000. [34] A. J. Kurdila and M. Zabarankin, Convex Functional Analysis. Birkhäuser Basel, 2005. [35] C. Villani, Topics in optimal transportation. American Mathematical Soc., 2021, vol. 58. [36] A. P. Ruszczy nski, Nonlinear Optimization. Princeton University Press, 2006. [37] N. Nisan, T. Roughgarden, Éva Tardos, and V. V. Vazirani, Eds., Algorithmic Game Theory. Cambridge University, 2007. [38] D. P. Bertsekas, Constrained Optimization and Lagrange Multiplier Methods (Optimization and Neural Computation Series), 1st ed. Athena Scientific, 1996. [39] B. Kloeckner, Approximation by finitely supported measures, ESAIM: Control, Optimisation and Calculus of Variations, vol. 18, no. 2, pp. 343 359, 2012. [40] A. Nagurney and D. Zhang, Projected Dynamical Systems and Variational Inequalities with Applications. Springer, 1996. [41] M. Welling and Y. W. Teh, Bayesian learning via stochastic gradient Langevin dynamics, in Proceedings of the 28th International Conference on International Conference on Machine Learning, ser. ICML 11. Madison, WI, USA: Omnipress, 2011, p. 681 688. [42] A. Salim, A. Korba, and G. Luise, The Wasserstein proximal gradient algorithm, Advances in Neural Information Processing Systems, vol. 33, pp. 12 356 12 366, 2020. [43] A. Nedi c and A. Ozdaglar, Approximate primal solutions and rate analysis for dual subgradient methods, SIAM Journal on Optimization, vol. 19, no. 4, pp. 1757 1780, 2009. [44] A. Cherukuri, E. Mallada, and J. Cortés, Asymptotic convergence of constrained primal dual dynamics, Systems & Control Letters, vol. 87, pp. 10 15, 2016. [45] A. Nemirovski, Prox-method with rate of convergence o(1/t) for variational inequalities with Lipschitz continuous monotone operators and smooth convex-concave saddle point problems, SIAM Journal on Optimization, vol. 15, no. 1, pp. 229 251, 2004. [46] Y. Nesterov, Dual extrapolation and its applications to solving variational inequalities and related problems, Math. Program., vol. 109, no. 2 3, p. 319 344, 2007. [47] T. Lin, C. Jin, and M. I. Jordan, Near-optimal algorithms for minimax optimization, in Proceedings of Thirty Third Conference on Learning Theory, 2020, pp. 2738 2779. [48] A. Mokhtari, A. Ozdaglar, and S. Pattathil, A unified analysis of extra-gradient and optimistic gradient methods for saddle point problems: Proximal point approach, in Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, 2020, pp. 1497 1507. [49] S. Chewi, M. A. Erdogdu, M. Li, R. Shen, and S. Zhang, Analysis of Langevin Monte Carlo: from Poincaré to Log-Sobolev, in Proceedings of Thirty Fifth Conference on Learning Theory, ser. Proceedings of Machine Learning Research, P.-L. Loh and M. Raginsky, Eds., vol. 178. PMLR, 2022, pp. 1 2. [50] R. Holley and D. W. Stroock, Logarithmic Sobolev inequalities and stochastic Ising models, Journal of Statistical Physics, no. 5 6, 1986. [51] D. Bakry, I. Gentil, M. Ledoux et al., Analysis and geometry of Markov diffusion operators. Springer, 2014, vol. 103. [52] P. Cattiaux and A. Guillin, Functional inequalities for perturbed measures with applications to log-concave measures and to some Bayesian problems, Bernoulli, vol. 28, no. 4, 2022. [53] H. Karimi, J. Nutini, and M. Schmidt, Linear convergence of gradient and proximal-gradient methods under the Polyak-Łojasiewicz condition, in Machine Learning and Knowledge Discovery in Databases, P. Frasconi, N. Landwehr, G. Manco, and J. Vreeken, Eds. Cham: Springer International Publishing, 2016, pp. 795 811. [54] J. Yang, X. Li, and N. He, Nest your adaptive algorithm for parameter-agnostic nonconvex minimax optimization, in Advances in Neural Information Processing Systems, 2022. [55] M. Boroun, Z. Alizadeh, and A. Jalilzadeh, Accelerated primal-dual scheme for a class of stochastic nonconvex-concave saddle point problems, in American Control Conference, 2023, pp. 204 209. [56] M. Sanjabi, J. Ba, M. Razaviyayn, and J. D. Lee, On the convergence and robustness of training gans with regularized optimal transport, in Advances in Neural Information Processing Systems, 2018. [57] J. Yang, N. Kiyavash, and N. He, Global convergence and variance reduction for a class of nonconvex-nonconcave minimax problems, in Advances in Neural Information Processing Systems, 2020, pp. 1153 1165. [58] T. Fiez, L. Ratliff, E. Mazumdar, E. Faulkner, and A. Narang, Global convergence to local minmax equilibrium in classes of nonconvex zero-sum games, in Advances in Neural Information Processing Systems, 2021, pp. 29 049 29 063. [59] K. Ahn and S. Chewi, Efficient constrained sampling via the mirror-Langevin algorithm, in Advances in Neural Information Processing Systems, vol. 34, 2021, p. 26. [60] D. Dua and C. Graff, UCI machine learning repository, 2017. [Online]. Available: http://archive.ics.uci.edu/ml [61] L. F. O. Chamon and A. Ribeiro, Probably approximately correct constrained learning, in Advances in Neural Information Processing, 2020. [62] D. M. Blei, A. Y. Ng, and M. I. Jordan, Latent Dirichlet allocation, J. Mach. Learn. Res., vol. 3, pp. 993 1022, 2003. [63] G. Celeux, M. E. Anbari, J.-M. Marin, and C. P. Robert, Regularization in Regression: Comparing Bayesian and Frequentist Methods in a Poorly Informative Situation, Bayesian Analysis, vol. 7, no. 2, pp. 477 502, 2012. [64] A. Lamperski, Projected stochastic gradient Langevin algorithms for constrained sampling and non-convex learning, in Conference on Learning Theory. PMLR, 2021, pp. 2891 2937. [65] L. Li, Q. Liu, A. Korba, M. Yurochkin, and J. Solomon, Sampling with mollified interaction energy descent, ar Xiv preprint ar Xiv:2210.13400, 2022. [66] K. S. Zhang, G. Peyré, J. Fadili, and M. Pereyra, Wasserstein control of mirror Langevin Monte Carlo, in Conference on Learning Theory. PMLR, 2020, pp. 3814 3841. [67] Q. Jiang, Mirror Langevin Monte Carlo: the case under isoperimetry, Advances in Neural Information Processing Systems, vol. 34, pp. 715 725, 2021. [68] V. Srinivasan, A. Wibisono, and A. Wilson, Fast sampling from constrained spaces using the metropolis-adjusted mirror Langevin algorithm, ar Xiv preprint ar Xiv:2312.08823, 2023. [69] J. Shi, C. Liu, and L. Mackey, Sampling with mirrored Stein operators, International Conference of Learning Representations, 2022. [70] M. Girolami and B. Calderhead, Riemann manifold Langevin and Hamiltonian Monte Carlo methods, Journal of the Royal Statistical Society Series B: Statistical Methodology, vol. 73, no. 2, pp. 123 214, 2011. [71] M. Brubaker, M. Salzmann, and R. Urtasun, A family of MCMC methods on implicitly defined manifolds, in Artificial intelligence and statistics. PMLR, 2012, pp. 161 172. [72] P. Tseng, On linear convergence of iterative methods for the variational inequality problem, Journal of Computational and Applied Mathematics, vol. 60, no. 1, pp. 237 252, 1995. [73] N. Golowich, S. Pattathil, C. Daskalakis, and A. Ozdaglar, Last iterate is slower than averaged iterate in smooth convex-concave saddle point problems, in Proceedings of Thirty Third Conference on Learning Theory, 2020, pp. 1758 1784. [74] V. M. Panaretos and Y. Zemel, An invitation to statistics in Wasserstein space. Springer Nature, 2020. [75] F. Otto, The geometry of dissipative evolution equations: the porous medium equation, Communications in Partial Differential Equations, vol. 26, no. 1-2, pp. 101 174, 2001. [76] A. T. Schwarm and M. Nikolaou, Chance-constrained model predictive control, AICh E Journal, vol. 45[8], pp. 1743 1752, 1999. [77] P. Li, M. Wendt, and G. Wozny, Robust model predictive control under chance constraints, Computers & Chemical Engineering, vol. 24[2-7], pp. 829 834, 2000. [78] F. Borrelli, A. Bemporad, and M. Morari, Predictive Control for Linear and Hybrid Systems. Cambridge University Press, 2017. A Related work In constrained sampling, it is important to distinguish between two types of constraints: support constraints and statistical constraints. The former deals with sampling from a target distribution π that is supported on a proper subset X Rd, which arises in applications such as latent Dirichlet allocation [62] and regularized regression [63]. The latter is the problem tackled in the current work. A first family of constrained sampling methods relies on rejection sampling: it obtains samples via any (unconstrained) method, rejecting those that violate the constraint (see, e.g., [10, 11]). Though this approach can handle constraints of any nature, it is often inefficient in terms of number of samples generated per iteration of the method (effective number of samples), especially when confronted with intricate constraints and high dimensional problems. These drawbacks can be addressed for support constraints using techniques inspired by finitedimensional constrained optimization. Projected LMC, for instance, deals with the problem of sampling from a target distribution restricted to a convex set [13, 64]. Barrier methods have also been used to tackle the same problem [18, 65]. Similarly, mirror and proximal descent versions of popular sampling algorithms such as LMC [12, 14, 15, 66 68] and Stein Variational Gradient Descent (SVGD) [69] have been proposed. Mirror descent algorithms enforce constraints by mapping (mirroring) the primal variables to a space with a different geometry (induced by a Bregman divergence) over which the optimization is carried out. Alternatively, methods adapted to target distributions support on manifolds have also been put forward [18, 70, 71]. In practice, these methods require explicit expressions for the projections, barriers, and mirror maps describing the constraints and are therefore not adapted to handle statistical requirements such as those considered in (PI). Langevin dynamics with constraint violation penalties were considered in [23], although they cannot enforce exact constraint satisfaction. Statistical (moment) constraints such as those considered in (PI) were investigated in [24]. As we discussed at the end of Section 2.3, this paper considers the combination of LMC and approximate dual ascent shown in (9). It also introduces a similar version of SVGD as well as algorithms based on barriers. Aside from requiring exact integration against intractable measures (namely, µk), convergence guarantees for these methods hold under restrictive assumptions on the constraints gi. Additionally, guarantees are derived only for continuous-time (gradient flows) dynamics. This work is also closely related to saddle-point methods in finite-dimensional optimization. For the general problem of maxx miny f(x, y), the behavior of descent-ascent methods have been investigated under a myriad of scenarios, including for functions f that are (strongly) convex/(strongly) concave [43, 45 48, 72, 73] as well as non-convex/non-concave under, e.g., PL conditions [47, 54, 55, 57, 58]. In general, convergence holds for the ergodic average of iterates [43, 57, 73]. Last-iterate results often require different algorithms, involving proximal point or extra gradient methods and time scale separation [47, 48, 73]. In particular, guarantees for the GDA method used in Algorithm 1 often requires stringent conditions that are hard to enforce for dual problems such as (DI). B Wasserstein space and discrete-time flows In this subsection we give some background on the Wasserstein spaces and sampling as optimization. For µ, ν P2(Rd), we define the 2-Wasserstein distance as W 2 2 (µ, ν) = inf s S(µ,ν) Z x y 2ds(x, y), where S(µ, ν) is the set of couplings between µ and ν. The metric space (P2(Rd), W2) is referred to as the Wasserstein space [74]. It can be equipped with a Riemannian structure [75]. In this geometric interpretation, the tangent space to P2(Rd) at µ is included in L2(µ) and is equipped with a scalar product defined for f, g L2(µ) as f, g L2(µ) = Z f(x)g(x)dµ(x). We use its differential structure as introduced in [75] and [26, Chapter 10]. For the functionals at stake in this paper (e.g., potential energies and negative entropy), the set of subgradients of a functional F : P2(Rd) R at µ P(Rd) is non-empty if µ satisfies some Sobolev regularity [26, Theorem 10.4.13]. A Wasserstein gradient flow of F is a solution (µ(t))t (0,T ), T > 0, of the continuity equation t + (µ(t)v(t)) = 0 that holds in the distributional sense, where v(t) is a subgradient of F at µ(t). Among the possible processes (v(t))t, one has a minimal L2(µ(t)) norm and is called the velocity field of (µ(t))t. In a Riemannian interpretation of the Wasserstein space [75], this minimality condition can be characterized by v(t) belonging to the tangent space to P2(Rd) at µ(t) denoted Tµ(t)P2(Rd), which is a subset of L2(µ(t)). The Wasserstein gradient is defined as this unique element, and is denoted W2F(µ(t)). In particular, if µ P2(Rd) is absolutely continuous with respect to the Lebesgue measure, with density in C1(Rd) and such that F(µ) < , W2F(µ)(x) = F (µ)(x) for µ-a.e. x Rd, where F (µ) denotes the first variation of F evaluated at µ, i.e. (if it exists) the unique function F (µ) : Rd R s.t. lim ϵ 0 1 ϵ (F(µ + ϵξ) F(µ)) = Z F (µ)(x)dξ(x) (15) for all ξ = ν µ, where ν P2(Rd). Now, we denote by T#µ the pushforward measure of µ P2(Rd) by the measurable map T. We recall that the KL divergence of µ relative to π can be decomposed as (1). The distribution µk of xk in (3) is known to follow a forward-flow splitting scheme [7] of the Fokker-Planck equation in (2), namely µk+1/2 = I γk W2V(µk) #µk (forward discretization for V) µk+1 = µk+1/2 N(0, 2γk Id) (exact flow for H), (16) where I denotes the identity map in L2(µk) and W2V(µk) = xf(x). C Proofs from Section 3.1 Proof of Theorem 3.3. Consider the potential V (µ, λ, ν) W 2 2 (µ, µ ) + min (λ ,ν ) Φ λ λ 2 + ν ν 2, (17) where µ is the solution of the constrained sampling problem (PI) and Φ is the set of solutions of the dual problem (DI). Our goal is to show that V decreases (in some sense) along trajectories of Algorithm 1. We say in some sense because contrary to the standard LMC algorithm, the distribution µk of xk is now a random measure that depends on the random variables {λk, νk}. Explicitly, we consider the filtration Fk = σ(µ0, {λℓ, νℓ}ℓ k) and show that V decreases on average when conditioned on Fk. This turns out to be enough to prove Theorem 3.3. Indeed, notice that µk Fk. Hence, the potential energy E(µk, λk, νk) Fk and the conditional law µk = L(xk|Fk 1) evolves as in the regular LMC algorithm (3). That is to say, conditioned on the event Fk, step 5 of Algorithm 1 follows a splitting scheme as (16), i.e., µk+1/2 = I ηk W2E(µk, λk, νk) µk+1 = µk+1/2 N(0, 2γId) (18b) Notice that all distributions in (18) are now deterministic. The core of the proof is collected in the following lemma that shows that V is a non-negative supermartingale. Note that Lemma C.1 describes the gap between half-iterations. This will be inconsequential for our purposes. Lemma C.1. Under the conditions of Theorem 3.3, we have E V ( µk+1/2, λk+1, νk+1)|Fk V ( µk 1/2, λk, νk) 2ηk KL( µk µ ) + m 2 W 2 2 ( µk, µ ) + η2 k h W2E( µk, λk, νk) 2 L2( µk) + Ex µk g(x) 2 + Ex µk h(x) 2 i . (19) We defer the proof of Lemma C.1 to show how it implies the bounds in Theorem 3.3. To do so, take the expectation of (19) with respect to {λℓ, νℓ}ℓ k to obtain E[ k] E V ( µk+1/2, λk+1, νk+1) V ( µk 1/2, λk, νk) 2ηk E KL( µk µ ) + m 2 W 2 2 ( µk, µ ) + η2 k h W2E( µk, λk, νk) 2 L2(µk) + Ex µk g(x) 2 + Ex µk h(x) 2 i . (20) Using the bounds in Assumption 3.2 then yields E[ k] E V ( µk+1/2, λk+1, νk+1) V ( µk 1/2, λk, νk) 2ηk E KL( µk µ ) + m 2 W 2 2 ( µk, µ ) + η2 k 3 + E[ λk 2] + E[ νk 2] G2. (21) Then, summing the LHS of (21) over k and using the fact that V is non-negative yields k=1 E[ k] = E V ( µK+1/2, λk+1, νk+1) V ( µ1/2, λ1, ν1) E[V ( µ1/2, λ1, ν1)]. (22) Notice that the expectation here is taken only over µ0 given that (λ0, ν0) = (0, 0) are deterministic. To proceed, we use the following proposition, whose proof we defer. Lemma C.2. Under the hypothesis of Theorem 3.3 it holds that E[V ( µ1/2, λ1, ν1)] R2 0 for R2 0 = W 2 2 (µ0, µ ) + η2 0 h f 2 L2(µ0) + Eµ0[ g 2] + Eµ0[ h 2] i + λ 2 + ν 2 . Back in (21), we obtain that E[KL( µk µ )]+m 2 E[W 2 2 ( µk, µ )] R2 0+ k=1 η2 k3G2+G2 K X k=1 η2 k E[ λk 2]+E[ νk 2] . We conclude by using the convexity of the Wasserstein distance to write E W 2 2 ( µk+1/2, µ ) W 2 2 (E[ µk+1/2], µ ) = W 2 2 (µk+1/2, µ ). Similarly for the KL divergence. The bounds in Theorem 3.3 are then obtained for the specific choices of ηk by noticing that PK k=1 1/ K and PK k=1 1/k 1 + log(K). Notice that all inequalities in the proof continue to hold for deterministic {λk, νk}. The bounds in Theorem 3.3 therefore also hold (without expectations) when using exact gradients to update the dual variables. Finally, we show there exists a sequence of step sizes ηk such that E[V (µk 1/2, λk, νk)] R2 0 for all k 1, where the expectation is taken over the {λk, νk}. This immediately implies that W 2 2 (µk, µ ) R2 0 and both E[ λk 2] and E[ νk 2] are bounded for all k. We proceed by induction. The base case is covered by Lemma C.2. Suppose now that there exists a sequence {η0, . . . , ηk 1} such that E[V ( µk 1/2, λk, νk)] R2 0. From the definition of V in (17) and the fact that the (λ , ν ) are bounded (Prop. 2.2), we then obtain that E[ λk ], E[ νk ] are bounded. Consequently, there exists ηk > 0 such that η2 k(3 + E[ λk 2] + E[ νk 2])G2 2ηk E h KL( µk µ ) + m 2 W 2 2 ( µk, µ ) i . From (21), we obtain that E[ k] 0, which together with the induction assumption yields E V ( µk+1/2, λk+1, νk+1) R2 0. Proof of Lemma C.1. The proof proceeds by combining two inequalities bounding the primal and dual terms in (17). (i) W 2 2 ( µk+1/2, µ ). We proceed following a coupling argument. Let sk be an optimal coupling between the random variables Y µk and Z µ , i.e., a coupling that achieves W 2 2 ( µk, µ ). Consider now the random variable T = Y ηk W2E( µk, λk, νk) and observe from (18a) that it is distributed as µk+1/2. Naturally, the coupling sk is no longer optimal for (T, Z), so that by the definition of the Wasserstein distance it follows that W 2 2 ( µk+1/2, µ ) Z x ηk W2E( µk, λk, νk) y 2dsk(x, y). (23) Expanding the RHS of (23) and using the m-strong convexity of E (Assumption 3.1) yields W 2 2 ( µk+1/2, µ ) W 2 2 ( µk, µ ) ηkm W 2 2 ( µk, µ ) + 2ηk E(µ , λk, νk) E( µk, λk, νk) + η2 k W2E( µk, λk, νk) 2 L2( µk). We can then bound the effect of the diffusion step using [8, Lemma 5] as in W 2 2 ( µk, µ ) W 2 2 ( µk 1/2, µ ) 2ηk H(µ ) H( µk) , (24) which yields W 2 2 ( µk+1/2, µ ) W 2 2 ( µk 1/2, µ ) ηkm W 2 2 ( µk, µ ) + 2ηk E(µ , λk, νk) + H(µ ) E( µk, λk, νk) H( µk) + η2 k W2E( µk, λk, νk) 2 L2( µk). (ii) λk+1 λ 2 + νk+1 ν 2. Notice that since λ RI +, the projection x 7 [x]+ is a contraction, i.e., [λ]+ λ λ λ for all λ RI. Using the definition of λk+1, νk+1 from Algorithm 1, we then obtain that λk+1 λ 2 + νk+1 ν 2 λk λ 2 + 2ηk(λk λ ) g(xk) + η2 k g(xk) 2 + νk ν 2 + 2ηk(νk ν ) h(xk) + η2 k h(xk) 2 , (26) for all (λ , ν ) Φ . To proceed, consider the conditional expectation of (26) with respect to Fk, namely, E λk+1 λ 2|Fk + E νk+1 ν 2|Fk λk λ 2 + νk ν 2 + 2ηk h (λk λ ) Ex µk g(x) + (νk ν ) Ex µk h(x) i + η2 k h Ex µk g(x) 2 + Ex µk h(x) 2 i , (27) where we used the fact that µk, λk, νk Fk. We conclude by using the linearity of the potential energy E from (11) in both λ and ν to get (λk λ ) Ex µk g(x) + (νk ν ) Ex µk h(x) = E( µk, λk, νk) E( µk, λ , ν ). Back in (27), we obtain E λk+1 λ 2|Fk + E νk+1 ν 2|Fk λk λ 2 + νk ν 2 + 2ηk E( µk, λk, νk) E( µk, λ , ν ) + η2 k h Ex µk g(x) 2 + Ex µk h(x) 2 i , for all (λ , ν ) Φ . To proceed with the proof, combine (25) and (28) to get W 2 2 ( µk+1/2, µ ) + E λk+1 λ 2|Fk + E νk+1 ν 2|Fk W 2 2 ( µk 1/2, µ ) + λk λ 2 + νk ν 2 E(µ , λk, νk) + H(µ ) E( µk, λ , ν ) H( µk) ηkm W 2 2 ( µk, µ ) + η2 k h W2E( µk, λk, νk) 2 L2( µk) + Ex µk g(x) 2 + Ex µk h(x) 2 i . (29) To upper bound the term in brackets, we add and subtract log(Z) and use the decomposition of the Lagrangian in terms of (11) to obtain E(µ , λk, νk) + H(µ ) E( µk, λ , ν ) H( µk) = L(µ , λk, νk) L( µk, λ , ν ). Using the saddle-point property (7), we then get L(µ , λk, νk) L( µk, λ , ν ) L(µ , λ , ν ) L( µk, λ , ν ) KL( µk µ ), We therefore conclude that W 2 2 ( µk+1/2, µ ) + E λk+1 λ 2|Fk + E νk+1 ν 2|Fk W 2 2 ( µk 1/2, µ ) + λk λ 2 + νk ν 2 2ηk KL( µk µ ) + m 2 W 2 2 ( µk, µ ) + η2 k h W2E( µk, λk, νk) 2 L2( µk) + Ex µk g(x) 2 + Ex µk h(x) 2 i . (30) Since (30) holds for all (λ , ν ) Φ , it holds in particular for the minimizer of the RHS, for which we can then write V ( µk 1/2, λk, νk). By subsequently taking the minimum of the LHS, we obtain (19). Proof of Lemma C.2. From the updates in Algorithm 1, we obtain E[V ( µ1/2, λ1, ν1)] W 2 2 ( µ1/2, µ ) + E[ η0g(x0) + λ 2] + E[ η0h(x0) ν 2] for all (λ , ν ) Φ . Notice that since λ RI +, the projection x 7 [x]+ is a contraction, i.e., [λ]+ λ λ λ for all λ RI. Using the triangle inequality then yields E[V (µ1/2, λ1, ν1)] W 2 2 ( µ1/2, µ0) + W 2 2 (µ0, µ ) + E[ η0g(x0) λ 2] + E[ η0h(x0) ν 2] W 2 2 (µ0, µ ) + λ 2 + ν 2 + W 2 2 ( µ1/2, µ0) + η2 0 E[ g(x0) 2] + η2 0 E[ h(x0) 2]. To proceed, observe from (18a) that µ1/2 = Id η0 f #µ0, which implies that W 2 2 ( µ1/2, µ0) η2 0 f 2 L2(µ0). Using the bounds in Assumption 3.2, we obtain that E[V ( µ1/2, λ1, ν1)] R2 0. Proof of Proposition 3.4. Since ck 0 and P ck = 1, we can use Jensen s inequality to write E k=1 ckφ(xk) Z φdµk Z φdµ . Using the relation between the ℓ1and ℓ2-norm, we further obtain Z φdµk Z φdµ Z φdµk Z φdµ If φ is bounded by 1, then the summands on the right-hand side of (31) can be bounded by TV(µk, µ ). Indeed, Z φdµk Z φdµ Z φ|dµk dµ | Z |dµk dµ | = 2TV(µk, µ ). The total variation distance can in turn be bounded by the KL divergence using Pinsker s inequality. We therefore obtain E k=1 ckφ(xk) k=1 ck KL(µk µ ) p On the other hand, if φ is 1-Lipschitz, the summands on the right-hand side of (31) are bounded by Z φdµk Z φdµ 2 W 2 1 (µk, µ ) W 2 2 (µk, µ ), which implies that E k=1 ckφ(xk) as long as m > 0. D Proofs from Section 3.2 Proof of Theorem 3.6. The proof is based on the analysis of the stochastic dual ascent algorithm λk+1 = λk + ηkg(ξk) +, for ξk µk such that KL( µk µλk) ϵ, (32) where µλk is the Lagrangian minimizer from (4) and ϵ > 0. Observe that, once again, we analyze the conditional distribution µk = L(ξk|λk). We collect this result in the following proposition: Proposition D.1. Consider the iterations (32) and assume that Eµ[g2 i ] G2 for all µ P2(Rd). Then, for 0 < ηk η and ϵ ηG2, there exists B < such that k=0 E[KL( µk µ )] ϵ + ηG2 The expectations are taken over the samples ξk µk. We conclude by combining Prop. D.1 with [29, Theorem 1], which characterizes the convergence of the LMC algorithm (3) under Assumption 3.5. Indeed, using the γk, N 0 k from Theorem 3.6 in Algorithm 1 guarantees that the law µk of xk|λk is such that KL( µk µλk) ϵ, i.e., satisfies the conditions in Prop. D.1. We can then apply Jensen s inequality to get that E[KL( µk µ )] KL(E[ µk] µ ) = KL(µk µ ), where µk is the law of x N 0 k in Algorithm 2. Proof of Prop. D.1. The proof relies on the following lemmata: Lemma D.2. For all µ P2(Rd) such that KL(µ µλ) ϵ, the expected value Eµ[g] is an approximate subgradients of the dual function d in (6) at λ R+ I , i.e., d(λ) d(λ ) + (λ λ ) Eµ[g] ϵ, for all λ R+ I . (34) Lemma D.3. Under the conditions of Prop. D.1, it holds for all k that λ 1 B0 C D + ηG2 + ϵ δ and E[ λk λ 2] B2 2B2 0 + 3η2G2. Lemma D.4. The sequence (λk, g(ξk)) obtained from (32) is such that k=0 E λ k g(ξk) ηG2 where the expectation is taken over realizations of ξk. Before proving these results, let us show how they imply Prop. D.1. Start by noticing from (32) that λi,k+1 λi,k + ηgi(ξk). Solving the recursion and recalling that λ0 = 0 then yields k=0 gi(ξk). Taking the expected value over ξk and dividing by ηK, we obtain k=0 E µk[gi] E[λi,K] ηK E[|λi,K λ i |] + λ i ηK , (36) where the last bound stems from the triangle inequality. Since the upper bound is non-negative for all i, we use the fact that the maximum of a set of values is less than the sum of those values to write max i=1,...,I k=0 E µk[gi] E[ λK λ 1] + λ 1 I B ηK , (37) where we used Lemma D.3 together with (E[ z 1])2 I E[ z 2] and B0 < B. Observe that (37) bounds the infeasibility of the ergodic average 1 K PK 1 k=0 µk for µk as in (32). To proceed, use the relation between the normalization of µλ and the dual function value [see (6)] to decompose the KL divergence between µk and µ as KL( µk µ ) = KL( µk µλk) + (λ λk) E µk[g] + d(λk) d(λ ). Since KL( µk µλk) ϵ and d(λ) d(λ ) for all λ RI +, we get KL( µk µ ) ϵ + (λ λk) E µk[g] Averaging over k and using Hölder s inequality then yields k=0 KL( µk µ ) ϵ + λ 1 max i=1,...,I k=0 E µk[gi] k=0 λ k E µk[g], which from Lemma D.3 and (37) becomes k=0 KL( µk µ ) ϵ + k=0 λ k E µk[g], where we used B0 < B. Taking the expected value, applying Lemma D.4, and taking 1 + I 2I for I 1 yields (33). D.1 Proof of Lemmata D.2 D.4 Proof of Lemma D.2. From the definition of the Lagrangian (4) and the dual function (6) (with J = 0), we obtain L(µ, λ) = KL(µ µλ) + d(λ). Using the fact that KL(µ µλ) ϵ, we get 0 d(λ) L(µ, λ) + ϵ. (38) Again using the definition of the dual function (6), we also obtain that d(λ ) L(µ, λ ). Adding to (38) then gives d(λ ) +d(λ) + L(µ, λ ) L(µ, λ) + ϵ. (39) Notice from (4) that the first term of the Lagrangians in (39) cancel out, leading to (34). Proof of Lemma D.3. Start by combining the update in (32) and the fact that, since λ RI +, the projection x 7 [x]+ is a contraction, to get λk+1 λ 2 λk λ 2 + 2η(λk λ ) g(ξk) + η2 k g(ξk) 2 . Taking the conditional expectation given λk then yields E[ λk+1 λ 2 | λk] λk λ 2 + 2η(λk λ ) E[g(ξk) | λk] + η2G2. where we used the fact that Eµ[ g 2] G2 for any µ. Noticing from (32) that E[g(ξk) | λk] = E µk[g], we can use Lemma D.2 to get E[ λk+1 λ 2 | λk] λk λ 2 + 2η h d(λk) D + ηG2 + ϵ i , (40) where we used the fact that η/2 < η to simplify the relation. To proceed, consider the set of approximate maximizers of the dual function D λ RI + | d(λ) D ηG2 ϵ . (41) Notice that Φ D. Since there exists at least one λ that achieves D (Prop. 2.2), D is not empty. Notice that for λk / D, we have that d(λk) D + ηG2 + ϵ < 0. By the towering property, we therefore obtain from (40) that E[ λk+1 λ 2 | λk / D] < E[ λk λ 2 | λk / D], (42) since η > 0. To bound the case when λk D we use the strictly feasible candidate µ from Assumption 2.1. Indeed, recall that KL(µ π) C and Eµ [gi] δ < 0 for all i. From the definition of the dual function (6), we obtain that d(λ) L(µ , λ) C λ 1 δ, (43) where we used the fact that λ RI + to write P i λi = λ 1. Hence, it follows that λ 1 B0 C D + ηG2 + ϵ δ , for all λ D. (44) Using the fact that z 2 z 2 1 for all z and that Φ D, we immediately obtain that λ λ 2 2B2 0 for all λ D. Using the towering property and the fact that D d(λk) and ϵ ηG2 yields E[ λk+1 λ 2 | λk D] 2B2 0 + 2ηϵ + η2G2 B2. (45) To conclude, we write E[ λk+1 λ 2] = E[ λk+1 λ 2 | λk D] P[λk D] + E[ λk+1 λ 2 | λk / D] P[λk / D]. Using (42) and (45) then yields E[ λk+1 λ 2] B2 1 P[λk D] + E[ λk λ 2 | λk / D] P[λk / D] max(B2, E[ λk λ 2 | λk / D]). Since both (42) and (45) holds independently of λk+1, we can also write E[ λk λ 2 | λk / D] max(B2, E[ λk 1 λ 2 | λk 1 / D]) Applying these relations recursively, we obtain that E[ λk+1 λ 2] max(B2, λ0 λ 2) = max(B2 1, λ 2). Noticing from (44) that since λ D we have λ 2 B2 0 < B2 then concludes the proof. Proof of Lemma D.4. To bound (35), we once again use the non-expansiveness of the projection to obtain λk+1 2 λk 2 + η2 g(ξk) 2 + 2ηλ k g(ξk). Taking the expectation and using the fact that Eµ[ g 2] G2, we get E[ λk+1 2] E[ λk 2] + η2G2 + 2η E[λ k g(ξk)]. Applying this relation recursively from K and using the fact that λ0 = 0 (deterministic) yields E[ λK 2] Kη2G2 + 2η k=0 E[λ k g(ξk)]. Since E[ λK 2] 0, we can divide by 2ηK to obtain the desired result. E Applications In this section, we provide further details on the example applications described in Section 2.2 as well as additional results from the experiments in Section 4. In these experiments, we start all chains at zero (unless stated otherwise) and use different step-sizes for each of the updates in steps 3 5 from Algorithm 1. We refer to them as ηx, ηλ, and ην. In contrast, we do not use diminishing step-sizes. E.1 Sampling from convex sets We are interested in sampling from target distribution πo(x) e f(x) I(x C), (46) for some closed, convex set C Rd. Several methods have been developed to tackle this problem, based on projections [13, 64], mirror maps [12, 15, 16], and barriers [17, 18]. Here, we consider a constrained sampling approach based on (PI) instead. To do so, note that sampling from (46) is equivalent to sampling from µ argmin µ P2(C) KL(µ π) for the unconstrained π e f P2(Rd). Note that this is exactly (PII). Now let C be described by the intersection of the 0-sublevel sets of convex functions {si}i=1,...,I, i.e., i=1 {x : si(x) 0}. Such a description is always possible by, e.g., considering the distance function d(x, C) = infy C x y 2, which is convex (since C is convex) and for which {x : d(x, C) 0} = C (since C is closed). Immediately, we see that solving (PII) is equivalent to solving min µ P2(Rd) KL(µ π) subject to Ex µ [si(x)]+ 0, i = 1, . . . , I, (PVI) where [z]+ = max(0, z). Note that (PVI) has the same form as (PI). To see why this is the case, consider without loss of generality that i = 1. Since [si(x)]+ 0 for all x by definition, it immediately holds that C = {x : s(x) 0} = {x : [s(x)]+ 0} = {x : [s(x)]+ = 0}. By the monotonicity of Lebesgue integration, we obtain that the feasibility set of (PVI) is F = n µ P2(Rd) : Ex µ [s(x)]+ 0 o = n µ P2(Rd) : Ex µ [s(x)]+ = 0 o = n µ P2(Rd) : [s(x)]+ = 0, µ-a.e. o . In other words, the feasibility set of (PVI) is in fact F = {µ P2(Rd) : µ(C) = 1} = P2(C). To illustrate the use of (PVI), consider the one-dimensional truncated Gaussian sampling problem from Section 4. Namely, we wish to sample from a standard Gaussian distribution N(0, 1) truncated to C = [1, 3]. In the language of (PVI), we take f(x) = x2/2 (i.e., π e x2/2) and si(x) = Table 2: Mean and variance estimates True mean Proj. LMC Mirror LMC PD-LMC 1D truncated Gaussian 1.510 1.508 1.470 1.488 2D truncated Gaussian [0.368, 0.368] [0.611, 0.610] [0.312, 0.418] [0.446, 0.444] 0.0 0.2 0.4 0.6 0.8 1.0 Iteration 104 Ergodic constraint slack 0 1 2 3 4 5 Iteration 106 Dual variable (λ) Figure 5: One-dimensional truncated Gaussian sampling: (a) Ergodic average of the constraint function (slack) and (b) Evolution of the dual variable λ. 0 1 2 3 4 5 PD-LMC iterations 105 Sample average 0 1 2 3 4 5 LMC evaluations 105 Sample average Nb = 1 (13 s) Nb = 10 (44 s) Nb = 100 (334 s) Nb = 1000 (3232 s) True mean Figure 6: The effect of the mini-batch size Nb on PD-LMC for sampling from a 1D truncated Gaussian: Estimated mean vs. (a) iteration and (b) LMC evaluations. (x 1)(x 3). In order to satisfy the assumptions of our convergence guarantees (particularly 2.1), we leave some slack in the constraints by considering Eµ[[si(x)]+] 0.005. This also helps with the numerical stability of the algorithm. Fig. 1 shows histograms of the samples obtained using PD-LMC (Algorithm 1 with ηx = ηλ = 10 3), the projected LMC (Proj. LMC, η = 10 3) from [13], and the mirror LMC (η = 10 3) from [59]. In all cases, we take 5 106 samples and keep only the second half. Observe that, due to the projection step, Proj. LMC generates an excess of samples close to the boundaries. In fact, it generates over three times more than required. This leads to an underestimation of the distribution mean and variance (Table 2). A similar effect is observed for mirror LMC. In contrast, PD-LMC provides a more accurate estimate. Nevertheless, PD-LMC imposes constraints on the distribution µ rather than its samples. Indeed, note from (47) that its feasibility set is such that samples belong to C almost surely, which still allows for a (potentially infinite) number of realizations outside of C. Yet, though PD-LMC is not an interior-point method, Theorems 3.3 3.6 show that excursions of iterates outside of C become less frequent as the algorithm progresses. We can confirm this is indeed the case in Fig. 5a, which shows the ergodic average of [s(x)]+ along the samples of PD-LMC. Note that it almost vanishes by iteration 104 even though the dual variable λ only begins to stabilize later (Fig. 5b). This is not surprising given that it is guaranteed by Prop. 3.4. In fact, only roughly 2% of the samples displayed in Fig. 1 are not in C. Before proceeding, we examine whether the convergence of PD-LMC could be improved by averaging more than one LMC samples when updating the dual variables, i.e., using mini-batches in steps 4 5 of Algorithm 1. Mini-batches will reduce the variance of the dual updates, although at the cost of additional LMC steps per iteration. To compensate for this fact, Fig. 6b displays the evolution of the ergodic average of PD-LMC samples as a function of the number of LMC evaluations rather than the number of iterations (as in Fig. 6a). Notice that, in this application, increasing the number of LMC samples Nb does not lead to faster convergence. This illustrates that, though mini-batches could be useful in some applications (particularly when the constraints are not convex, as in Section 3.2), it Iter. = 1e4 Iter. = 1e5 Iter. = 1e6 Iter. = 5e6 Iter. = 1e4 Iter. = 1e5 Iter. = 1e6 Iter. = 5e6 Figure 7: Density estimate of two-dimensional truncated Gaussian using samples from (a) Proj. LMC and (b) PD-LMC. is not immediate that their benefits always outweigh the increased computational cost. Oftentimes, using a single LMC sample is more than enough. It is worth noting that using PD-LMC with a large mini-batch Nb was suggested in [24] to approximate the expectation needed by their continuous-time algorithm. As we see here, this is neither necessary nor always beneficial. We now turn to a more challenging, two-dimensional applications. We seek to sample from a Gaussian located at [2, 2] with covariance diag([1, 1]) restricted to an ℓ2-norm unit ball (Fig. 1). Specifically, we use f(x) = x 2 /2 (i.e., π e x 2/2) and si(x) = x 2 1. Once again, we leave some slack to the algorithm by taking the constraint in (PVI) to be Eµ[[si(x)]+] 0.001. For reference, we also display samples from the real distribution obtained using rejection sampling. This is indeed a challenging problem. The boundary of C is 2 standard deviations away from the mean of the target distribution, which means that samples from the target π are extremely scarce this region. Indeed, using the untruncated Gaussian as a proposal for rejection sampling yields an acceptance rate of approximately 1%. The strong push of the potential f towards the exterior of C leads Proj. LMC (η = 10 3; the last 106 samples are used after running 5 106 iterations) to be now even more concentrated around its boundary. In fact, almost 25% of its samples are in an annular region of radius [0.999, 1), where only 0.14% of the samples should be according to rejection sampling. Indeed, note from Fig. 7a, that even as iterations advance, the samples of Proj. LMC continue concentrate close to the boundary. In contrast, PD-LMC (ηx = 10 3 and ηλ = 2 10 1) only place 1.8% of its samples outside of C, mostly during the initial phase of the algorithm (Fig. 7a). Indeed, the average of the constraint function along samples from PD-LMC essentially vanishes around iteration 5 104. Achieving this requires larger values of λ (on the order of 250, Fig. 8b) compared to the one-dimensional case (Fig. 5b). This reflects the difficulty of constraining samples to C in this instance, a statement formalized in the perturbation results of Prop. 2.2(iv). Due to the more amenable numerical properties of the barrier function, mirror LMC (η = 10 3) performs well without concentrating samples on the boundary (0.15% of the samples on the annular region of radius [0.999, 1)). 0.0 0.2 0.4 0.6 0.8 1.0 Iteration 105 Ergodic constraint slack 0 1 2 3 4 5 Iteration 106 Dual variable (λ) Figure 8: Two-dimensional truncated Gaussian sampling: (a) Ergodic average of the constraint function (slack) and (b) Evolution of the dual variable λ. 0 10 20 30 40 Prevalence of > $50k (%) Constrained (Acc.: 82%) Unconstrained (Acc.: 84%) 0.0 0.5 1.0 1.5 2.0 Iteration 104 Dual variable (λ) Female Male Figure 9: Fair Bayesian logistic regression on the Adult dataset: (a) prevalence of positive outputs and (b) dual variables. E.2 Rate-constrained Bayesian models While rate constraints have become popular in ML due to their central role in fairness (see, e.g., [20]), they find applications in robust control [76 78] and to express other requirements on the confusion matrix, such as precision, recall, and false negatives [21]. For illustration, we consider here the problem of fairness in Bayesian classification. Let q(x; θ) = P[y = 1|x, θ] denote the probability of a positive outcome (y = 1) given the observed features x X and the parameters θ distributed according to the posterior π. This posterior is determined, e.g., by some arbitrary Bayesian model based on observations {(xn, yn)}n=1,...,N. Hence, Eθ π[q(x; θ)] denotes the likelihood of a positive outcome for x. Consider now a protected group, represented by a measurable subset G X, for which we wish to enforce statistical parity. In other words, we would like the prevalence of positive outcomes to be roughly the same as that of the whole population. Thus, we now want to sample not from the posterior π, but from a close-by distribution of parameters θ that ensures this parity. Explicitly, for some tolerance δ > 0, we want to sample from µ min µ P2(Rd) KL(µ π) subject to Ex,θ µ q(x; θ) | G Ex,θ µ q(x; θ) δ. (PVII) Naturally, we can account for more than one protected by incorporating additional constraints. In our experiments, we take π to be the posterior of a Bayesian logistic regression model for the Adult dataset from [60] (details on data pre-processing can be found in [61]). The N = 32561 data points in the training set are composed of d = 62 socio-economical features (x Rd, including the intercept) and the goal is to predict whether the individual makes more than US$ 50000 per year (y {0, 1}). The posterior is obtained by combining a binomial log-likelihood with independent zero-mean Gaussian (log)priors (σ2 = 3) on each parameter of the model, i.e., we consider the 10 20 30 Prevalence of > $50k (%) 100 (251 s) 1000 (2464 s) LMC iterations (Nb) Figure 10: The effect of the mini-batch size Nb on PD-LMC in fair Bayesian classification. n=1 log(1 + e (2yn 1)x n θ) + θ2 i 2σ2 . (48) We begin by using the LMC algorithm from (3) (i.e., we impose no constraints) to collect samples of the coefficients θ from this posterior (η = 10 4; the last 104 samples are used after running 2 104 iterations). We find that, while the probability of positive outputs is 19.1% across the whole test set, it is 26.2% among males and 0.05% among females. Looking at the distribution of this probability over the unconstrained posterior π (Fig. 9a), we see that this behavior goes beyond the mean. The model effectively amplifies the inequality already present in the test set, where the prevalence of positive outputs is 30.6% among males and 10.9% among females. To overcome this disparity, we consider gender to be the protected class in (PVII), constraining both Gfemale and Gmale. We formulate the constraint of (PVII) using an empirical distribution induced from the data. Explicitly, we consider constraints 1 |Gfemale| n Gfemale Eθ µ q(xn; θ) 1 n=1 Eθ µ q(xn; θ) δ n Gmale Eθ µ q(xn; θ) 1 n=1 Eθ µ q(xn; θ) δ where Gfemale, Gmale {1, . . . , N} partition the data set. For these experiments, we take δ = 0.01. Using PD-LMC (ηx = 10 4, ηλ = 5 10 3), we then obtain a new set of samples from the logistic regression parameters θ that lead to a prevalence of positive outcomes (in the test set) of 17.1% over the whole population, 18.1% for males, and 15.1% for females. In fact, we notice a substantial overlap between the distributions of this probability across the constrained posterior µ for male and female (Fig. 9a). Additionally, this substantial improvement over the previously observed disparity comes at only a minor reduction in accuracy. Though both distributions change considerably, notice from the value of the dual variables that these changes are completely guided by the female group. Indeed, λmale = 0 throughout the execution of PD-LMC (Fig. 9b). Before proceeding, we once again examine the effect of using multiple LMC samples to update the dual variables, i.e., using mini-batches in steps 4 5 of Algorithm 1. Fig. 10 shows the distribution of the prevalence of positive predictions (> $50k) for different mini-batch sizes Nb. In all cases, we collect 2 104 samples, which means that we evaluate 2Nb 104 LMC updates (step 3 in Algorithm 1). Same as in the 1D truncated Gaussian case, we notice no difference between the resulting distributions. This is to be expected given our results (Theorem 3.3). The computation time, on the other hand, increases considerably with the mini-batch size. Once again, we note that PD-LMC with a large mini-batch Nb was used in the experiments of [24] to overcome the challenge of computing an expectation in their dual variable updates. In turns out that this computationally intensive modification is not necessary. Adm/clerical African-Amer. Age (31.0, 37.0] Unconstrained Constrained 10 20 30 Prevalence of > $50k (%) Tolerance (δ) Figure 11: Counterfactual sampling in fair Bayesian logistic regression: (a) Selected (mean) coefficients and (b) different tolerances. E.3 Counterfactual sampling Previous applications were primarily interested in sampling from µ , the constrained version of the target distribution π. The goal of counterfactual sampling, on the other hand, is to probe the probabilistic model π by evaluating its compatibility with a set of moment conditions. It is therefore interested not only in µ , but in how each condition affects the value P = KL(µ π). We next describe how constrained sampling can be used to tackle this problem. Let π denote a reference probabilistic model, such as the posterior of the Bayesian logistic model in (48). Standard Bayesian hypothesis tests can be used to evaluate the validity of actual statements such as is it true that Ex π[g(x)] 0? or Ex π[h(x)] = 0? Hence, we could check is π more likely to yield a positive output for a male than a female individual? (from the distributions under Unconstrained in Fig. 9a, this is probably the case). In contrast, counterfactual sampling is concerned with counterfactual statements such as how would the world have been if E[g(x)] 0? In the case of fairness, how would the model have been if it predicted positive outcomes more equitably? Constrained sampling evaluates these counterfactual statements in two ways. First, by providing realizations of this alternative, counterfactual world (µ ). For instance, we can inspect the difference between realizations of π, obtained using the traditional LMC in (3), and µ , obtained using PD-LMC (Algorithm 1). In Fig. 11a, we show the mean of some coefficients of the Bayesian logistic models from Section E.2. Notice that it is not enough to normalize the Intercept and reduce the advantage given to males (Female is encoded as Male = 0). This alternative model also compensates for other correlated features, such as education (Bachelor), profession (Adm/clerical), and age. Second, constrained sampling evaluates the compatibility of each counterfactual condition (constraint) world with the reference model. While Algorithm 1 does not evaluate P explicitly, it provides measures of its sensitivity to perturbations of the constraints: the Lagrange multipliers (λ , ν ). Indeed, recall from Prop. 2.2 that i=1 e λ i gi ! j=1 e ν j hj Hence, (λ , ν ) describe the magnitude of tilts needed for the reference model π to satisfy the conditions E[g(x)] 0 or E[h(x)] = 0. This relation is made explicit in Prop. 2.2(iv). Concretely, observe that the dual variable relative to the constraint on the male subgroup is always zero (Fig. 9b). This implies that π is fully compatible with the condition n Gmale Eθ µ q(xn; θ) 1 n=1 Eθ µ q(xn; θ) δ, i.e., the statement the model predicts positive outcomes for males on average at least as much as for the whole population. In contrast, accommodating statistical parity for females requires considerable deviations from the reference model π (λ female 160). Without recalculating µ , we therefore know that even small changes in the tolerance δ for the female constraint would substantially change the TSLA NVDA JNJ AAPL GOOG BRK-B LLY (Log)return TSLA NVDA JNJ AAPL GOOG BRK-B LLY Figure 12: Counterfactual sampling of the stock market under a 20% average return increase on each stock: mean (ρ) and variance [diag(Σ)] distributions. 0 1 2 3 4 5 6 Iteration 105 Ergodic constraint slacks TSLA NVDA JNJ 0 1 2 3 4 5 6 Iteration 105 Dual variable (ν) Figure 13: Counterfactual sampling of the stock market under a 20% average return increase on each stock: (a) ergodic average of constraint functions (slacks) and (b) dual variables. distribution of outcomes. This statement is confirmed in Fig. 11b. Notice that this is only possible due to the primal-dual nature of PD-LMC. E.3.1 Stock market model Counterfactual analyses based on the dual variables become more powerful as the number of constraints grow. To see this is the case, consider the Bayesian stock market model introduced in Section 2.2. Here, π denotes the posterior model for the (log-)returns of 7 assets (TSLA, NVDA, JNJ, AAPL, GOOG, BRK-B, and LLY). The dataset is composed of 5 years of adjusted closing prices for a total of 1260 points per asset. The posterior is obtained by combining a Gaussian likelihood N(ρ, Σ) with Gaussian prior on the mean ρ [N(0, 3I)] and an inverse Wishart prior on the covariance Σ (with parameters Ψ = I and ν = 12). Using the LMC algorithm (η = 10 3; the last 3 105 samples are used after running 6 105 iterations), we collect samples from this posterior and estimate the mean and variance of the (log-)return for each stock (Table 3). In this case, Σ is initialized to 10 I. We might now be interested in understanding what the market would look like if all stocks were to incur a 20% increase in their average (log-)returns. To do so, we use PD-LMC (ηx = 10 3 and ην = 6 10 3) to solve the following constrained sampling problem minimize µ P2(Rd) KL(µ π) subject to E(ρ,Σ) µ ρi = 1.2 ρi, i = 1, . . . , 7, (PVIII) where ρi is the mean (log-)return of the i-th stock shown in Table 3. The distribution of ρ and the diagonal of Σ are compared to those from the unconstrained model in Fig. 12. Notice that, though we only impose constraints on the average returns ρ, we also see small changes in the stock volatilities. Inspecting the dual variables (Fig. 13b), we notice that three dual variables are essentially zero (TSLA, AAPL, and BRK-B). This means that their increased returns are completely dictated by those of other TSLA NVDA JNJ AAPL GOOG BRK-B LLY 0.4 (Log)return All stocks NVDA, LLY TSLA NVDA JNJ AAPL GOOG BRK-B LLY Figure 14: Distribution of mean (ρ) and variance [diag(Σ)] of the stock market constrained to a 20% average return increase on all stocks vs. only LLY and NVDA. Table 3: Mean standard deviation of mean (log-)returns (ρ). Reference model (π) +20% (all stocks) +20% (LLY and NVDA) TSLA 0.20 0.12 0.24 0.12 0.23 0.12 NVDA 0.25 0.09 0.31 0.09 0.31 0.09 JNJ 0.02 0.07 0.02 0.07 0.03 0.07 AAPL 0.11 0.06 0.14 0.06 0.14 0.06 GOOG 0.09 0.06 0.11 0.06 0.11 0.06 BRK-B 0.06 0.07 0.07 0.07 0.07 0.07 LLY 0.16 0.06 0.19 0.06 0.19 0.06 stocks. Said differently, their returns increasing 20% is consistent with the reference model π conditioned on the other returns increasing. Proceeding, two stocks have negative dual variables (LLY and NVDA). This implies that bringing their constraints down to ρi would yield a decrease in P (distance to the reference model π). This is in contrast to JNJ and GOOG, whose positive λ s imply that we should should increase their returns to reduce P . Indeed, by inspecting the ergodic slacks (Fig. 13a) we see that all stocks approach zero (i.e., feasibility), but that JNJ and GOOG do so from above. This behavior is expected according to Prop. 3.4. These observations show two things. First, that an increase in the average returns of LLY and NVDA is enough to drive up the returns of all other stocks. In fact, it leads to essentially the same distribution as if we had required the increase to affect all stocks (Fig. 14). Second, that the increase we would see in JNJ (and to a lesser extent GOOG) would actually be larger than 20%. Once again, we reach these conclusion without any additional computation. Their accuracy can be corroborated by the results in Table 3. F Example application of Prop. 2.2 In this section we illustrate the result in Prop. 2.2, i.e., we show that given solutions (λ , ν ) of (DI), the constrained sampling problem (PI) reduces to sampling from µλ ν e U( ,λ ,ν ). Indeed, consider a standard Gaussian target, i.e., π e x 2/2, and the linear moment constraint E[x] = b, for b Rd. This can be cast as (PI) with f(x) = x 2/2 and h(x) = b x (no inequality constraints, i.e., I = 0). Clearly, the solution of (PI) in this case is µ = N(b, I), i.e., a Gaussian distribution with mean b. What Prop 2.2 claims is that rather than directly solving (PI), we can solve (DI) to obtain a Lagrange multiplier ν such that µ = µν for µν defined as in (4). In this setting, we can see this is the case by doing the computations explicitly. Indeed, we have µν(x) π(x)e ν h(x) = exp x 2 2 ν (b x) . Completing the squares, we then obtain µν(x) exp x ν 2 2 ν b . (49) To compute the Lagrange multiplier ν , notice from the definition of the dual function in (6) that the dual problem (DI) is in fact a ratio of normalizing factors. Explicitly, ν = argmax ν Rd log R π(x)dx R µν(x)dx = argmax ν Rd log " R exp x 2 exp( ν 2/2 ν b) R exp x ν 2 Immediately, we obtain ν = argmax ν Rd ν 2/2 + ν b = b. (50) Note that the dual problem is a concave program, as is always the case [30]. To conclude, we can combine (49) and (50) to get µν (x) ν =b exp x b 2 µν = N(b, I) = µ . The main advantage of using Algorithms 1 and 2 is that we do not need to determine the Lagrange multipliers (λ , ν ) to then sample from µλ ν = µ . Indeed, Theorems 3.3 and 3.6 show that these stochastic primal-dual methods do both things simultaneously, without explicitly evaluating any expectations. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: the abstract and introduction list our contributions, i.e. the theoretical and experimental study of Primal-Dual Langevin Monte Carlo. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: we discuss the limitations of our theoretical results in the conclusion. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: For each theoretical result, we worked with justified assumptions, making all the dependencies of the problem clear. We provide clear and detailed proofs in the appendix. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and cross-referenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: For each experimental result, we provide the detailed setting. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: We will provide a link to a public github repository with a Python code. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/public/ guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https://nips.cc/public/ guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: the algorithms and settings of the experiments are rather simple, and detailed in the submission. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: We give all precisions relative to the the evaluation of our models, either on simulated data or real data, including precising train/set proportions for instance. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: Our experiments run on a standard laptop in a few minutes, and are illustrative. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: In our opinion, this paper does not address societal impact directly, and consider the generic problem of optimization over measures and sampling. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: In our opinion the paper does not have direct positive or negative social impact. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: Our paper does not present such risks. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [NA] Justification: our paper does not use existing assets Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [Yes] Justification: we only have experiments in Python that will be made public. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: our experiments do not involve crowdsourcing. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: our study do not involve risk for participants. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.