# functional_gradient_flows_for_constrained_sampling__72ba91bc.pdf Functional Gradient Flows for Constrained Sampling Shiyue Zhang School of Mathematical Sciences Peking University zhangshiyue@stu.pku.edu.cn Longlin Yu School of Mathematical Sciences Peking University llyu@pku.edu.cn Ziheng Cheng Department of Industrial Engineering and Operations Research University of California, Berkeley ziheng_cheng@berkeley.edu Cheng Zhang School of Mathematical Sciences and Center for Statistical Science Peking University chengzhang@math.pku.edu.cn Recently, through a unified gradient flow perspective of Markov chain Monte Carlo (MCMC) and variational inference (VI), particle-based variational inference methods (Par VIs) have been proposed that tend to combine the best of both worlds. While typical Par VIs such as Stein Variational Gradient Descent (SVGD) approximate the gradient flow within a reproducing kernel Hilbert space (RKHS), many attempts have been made recently to replace RKHS with more expressive function spaces, such as neural networks. While successful, these methods are mainly designed for sampling from unconstrained domains. In this paper, we offer a general solution to constrained sampling by introducing a boundary condition for the gradient flow which would confine the particles within the specific domain. This allows us to propose a new functional gradient Par VI method for constrained sampling, called constrained functional gradient flow (CFG), with provable continuous-time convergence in total variation (TV). We also present novel numerical strategies to handle the boundary integral term arising from the domain constraints. Our theory and experiments demonstrate the effectiveness of the proposed framework. 1 Introduction Efficiently approximating and sampling from unnormalized distributions is a fundamental and challenging task in probabilistic machine learning, especially in Bayesian inference. Various methods, including Markov Chain Monte Carlo (MCMC) and Variational Inference (VI), have been developed to address the intractability of the target distribution. In VI, the inference problem is reformulated as an optimization task that aims to find an approximation within a specific distribution family that minimizes the Kullback-Leibler (KL) divergence to the posterior (Jordan et al., 1999; Wainwright & Jordan, 2008; Blei et al., 2016). Leveraging efficient optimization algorithms, VI is often fast during training and more scalable to large datasets. However, its approximation power may be limited depending on the chosen family of variational distributions. In contrast, MCMC methods generate Equal contribution. Corresponding author. 38th Conference on Neural Information Processing Systems (Neur IPS 2024). samples from the posterior through a Markov chain that satisfies the detailed balance condition (Duane et al., 1987; Robert & Stramer, 2002; Neal, 2011; Welling & Teh, 2011; Chen et al., 2014). While MCMC is asymptotically unbiased, it may be slow to converge, and assessing convergence can be challenging. Recently, there has been a growing interest in the gradient flow formulation of both MCMC and VI, leading to the development of particle based variational inference methods (Par VIs) that tend to combine the best of both worlds (Liu & Wang, 2016; Chen et al., 2018; Liu et al., 2019; di Langosco et al., 2021; Fan et al., 2022; Alvarez-Melis et al., 2022). From the variational perspective, Par VIs take a non-parametric approach where the approximating distribution is represented as a set of particles. These particles are iteratively updated towards the steepest direction to reduce the KL divergence to the posterior, following the gradient flow in the space of distributions with certain geometries. This non-parametric nature of Par VIs significantly enhances its flexibility compared to classical parametric VIs, and the interaction between particles also makes Par VIs more particle-efficient than MCMCs. One prominent particle-based variational inference technique is Stein Variational Gradient Descent (SVGD) (Liu & Wang, 2016). It calculates the update directions for particles by approximating the gradient flows of the Kullback-Leibler (KL) divergence within a reproducing kernel Hilbert space (RKHS), where the approximation takes a tractable form (Liu, 2017; Chewi et al., 2020). However, the performance of SVGD heavily depends on the choice of the kernel function and the quadratic computational complexity of the kernel matrix also makes it impractical to use a large number of particles. As kernel methods are known to have limited expressive power, many attempts have been made recently to expand the function class for gradient flow approximation (Hu et al., 2018; Grathwohl et al., 2020; di Langosco et al., 2021; Dong et al., 2023; Cheng et al., 2023). By embracing a more expressive set of functions, such as neural networks, these functional gradient approaches have shown improved performance over vanilla SVGD while not requiring expensive kernel computation. While MCMC and VI methods have shown great success in sampling from unconstrained domains, they often struggle when dealing with target distributions supported on constrained domains. Sampling from constrained domains is an important and challenging problem that appears in various fields, such as topic modeling (Blei et al., 2003), computational statistics and biology (Morris, 2002; Lewis et al., 2012; Thiele et al., 2013). Recently, several attempts have been made to extend classical sampling methods like Hamiltonian Monte Carlo (HMC) or SVGD to constrained domains (Brubaker et al., 2012; Byrne & Girolami, 2013; Liu & Zhu, 2018; Shi et al., 2022). However, these extensions either involve computationally expensive numerical subroutines such as solving nonlinear systems of equations, or rely on intricate implicit and symplectic schemes or mirror maps that require a case-by-case design effort tailored to specific constraint domains. In this paper, we propose a functional gradient Par VI method for sampling from probability distributions subject to constrained domains with general shapes. We demonstrate that functional gradient approaches for Par VIs can be seamlessly adapted to constrained domains by learning the gradient flows with a vector field that adheres to a boundary condition. Intuitively, this boundary condition would confine particles within the specified domain, and it is indeed a sufficient condition for gradient flows of probability measures confined to this domain. Following previous works (di Langosco et al., 2021; Dong et al., 2023; Cheng et al., 2023), we employ the regularized Stein discrepancy objective for functional gradient estimation where we directly incorporate the boundary condition for the vector field into the design of its neural network approximation. Due to the domain constraints, integration by parts now would lead to an additional boundary integral term in the training objective. We, therefore, derived an effective approach for properly evaluating this term for general boundaries. Extensive numerical experiments across different constrained machine learning problems are conducted to demonstrate the effectiveness and efficiency of our method. 2 Related Work To better capture the gradient flow, a number of functional gradient methods for Par VIs have been proposed recently that employ a larger and more expressive class of functions than reproducing kernel Hilbert spaces (RKHS). In particular, di Langosco et al. (2021) used neural networks to learn the Stein Discrepancy with an L2 regularization term, and updated the particles based on the learned witness functions. Dong et al. (2023) and Cheng et al. (2023) provided extensions that accommodates a broader class of regularizers. Sampling from constrained domains is generally more challenging compared to the unconstrained ones. Brubaker et al. (2012) proposed a constrained version of HMC for sampling on implicit manifolds by applying Lagrangian mechanics to Hamiltonian dynamics, which requires solving a nonlinear system of equations for each numerical integration step. Lan et al. (2014) focused on constraints that can be transformed into hyper-spheres, which can be viewed a special case of Brubaker et al. (2012) that has close-form update formulas. Zhang et al. (2020); Ahn & Chewi (2021); Shi et al. (2022) extended Langevin algorithms and SVGD to constrained domains via mirror maps. However, these methods require explicit forms of transformations (e.g., spherical augmentation and mirror maps) that capture the constraints, which would limit their applications to simple domains. Bubeck et al. (2018); Brosse et al. (2017); Salim & Richtárik (2020) considered projected Langevin algorithms through projection oracle, which is of high computational cost for complex constrained domains. Zhang et al. (2022) proposed an orthogonal-space gradient flow approach for sampling in manifold domains with equality constraints, which employed a similar strategy to ours. Our method is different from theirs in that it is designed for domains with inequality constraints using Par VIs with neural network gradient flow approximations, and we also proposed novel numerical strategies to address boundary-related challenges. 3 Background Notations We use x to denote particles in Rd and Ω= {x|g(x) 0} to denote the constrained domain. Notation 1Ωis the indicator function of Ω. Let denote the standard Euclidean norm of a vector. Let P(Rd) denote the set of probability distributions on Rd that are absolute continuous with respect to the Lebesgue measure, and we do not distinguish the probability measure with its density function. Let DKL be the Kullback-Leibler divergence and TV be the total variation distance. We use C(X, Y) to denote the space of continuous mappings from X to Y, and use the shorthand C(X) for C(X, X). 3.1 Particle-based Variational Inference Let p P(Rd) be the target distribution we wish to sample from. We can frame the problem of sampling into a KL divergence minimization problem ˆp := arg min p Q DKL(p p ), (1) where Q P(Rd) is the space of probability measures. Particle-based variational inference methods (Par VIs) can be described within this framework where Q is represented as a set of particles. Starting from an initial distribution p0 and an initial particle x0 p0, we update the particle xt following dxt = vt(xt)dt where vt : Rd 7 Rd is the velocity field at time t. The density pt of xt follows the continuity equation dpt/dt = (vtpt), and the KL divergence decreases with the following rate: d dt DKL(pt p ) = Ept log p pt , vt . (2) Par VIs aim to find the optimal velocity field vt that minimizes (2) in a Hilbert space H with the squared norm regularizer as follows min vt H Ept log p pt , vt + 1 2 vt 2 H. (3) 3.2 Functional Wasserstein Gradient Different choices of the space H in (3) lead to different gradient flow algorithms. SVGD Liu & Wang (2016) chooses H to be a Reproducing Kernel Hilbert Space with kernel k( , ). This way, the optimal velocity field has a close form solution v t ( ) = Ept[k( , x) log p (x) + xk( , x)], (4) albeit the design of kernels can be restrictive for the flexibility of the method. When H = L2(pt), then v t = log p pt is also known as the Wasserstein gradient of KL divergence (Jordan et al., 1998). Since the score function of particle distribution log pt is generally inaccessible, recent works propose to parameterize vt as a neural network and train it through (3) di Langosco et al. (2021); Dong et al. (2023); Cheng et al. (2023). Due to Stein s identity, (3) has a tractable form v t = arg min v F Ept log p , v v + 1 2 v 2 , (5) where F is the neural network family. This method is called functional Wasserstein gradient in contrast to the kernelized version. 4 Main Method Consider sampling from a target distribution p (x) supported on Ω= {g(x) 0}, where g : Rd R is a continuously differentiable function. Throughout the work, we assume that p admits differentiable positive density function on Ω. We first reformulate the problem into a constrained optimization in the space of probability measures: min p P(Rd) DKL(p p ), s.t. p(Ω) = 1. (6) However, note that p is only supported on Ω, the problem is ill-posed if p(Ω) < 1, since in this case p will not be absolute continuous with respect to p and DKL(p p ) cannot be defined. Besides, note that the initial particle distribution p0 is generally assumed to be fully supported on Rd and thus does not meet the constraint. To fix this issue, a natural idea is to drive the particle distribution pt towards Ωto satisfy pt(Ω) = 1 and train the functional gradient flow on constrained domains by minimizing the regularized Stein discrepancy (RSD) min v F LRSD = Z 2 v 2 dx. (7) These stringent requirements make the construction of velocity field for constrained sampling far more volatile than conventional functional gradient for unconstrained domains. In the rest of this section, we first present the idea of velocity design in Section 4.1 and then derive the tractable training objective in Section 4.2. The practical algorithm is proposed in Section 4.3. 4.1 Necessity of Piece-wise Velocity Field We first show that a globally continuous velocity field may fail to achieve the exact minimum of RSD under the constraint. To ensure that particles inside Ωwill never escape, the velocity should satisfy the boundary condition vt n 0, on Ω. (8) In fact, if (8) holds, by Stoke s formula and the continuity equation tpt(x) = (pt(x)vt(x)), we have d dtpt(Ω) = d dt Ept1Ω= Z tpt(x)dx = Z Ω (pt(x)vt(x))dx = Z Ω pt(x)vt(x) nd S. (9) Thus we can conclude Proposition 4.1. If vt n 0 on Ω, then pt(Ω) will not decrease. However, it is possible that there does not exist an optimal solution of (7) that meets the boundary condition (8) and is continuous on Ω, as illustrated in the following 1D-example. Example 4.2. Let p exp( x2 2 ) 1{x2 1} and p exp( (x 1 2 ). We have log p Ωand LRSD = 1 2 R 1 1 p|v(x) + 1 8. The constraint on boundary is v(1) 0, v( 1) 0. Hence there is no optimal v in C(R). Therefore, it is necessary to design velocities for particles inside and outside separately. For particles located outside Ω(contributing to the probability p(g(x) > 0)), it is prudent to make the velocity drive the particles into the domain. On the other hand, for those already inside Ω, the velocity field should be able to fit the target distribution. Overall, it is reasonable to use the following piece-wise construction: vt = hnet 1{g<0} λ g g 1{g 0}. (10) Here λ > 0 is a constant. The gradient direction g will push the particles into Ω. And hnet is a continuous neural net to learn the optimal direction in Ω. We observed that the magnitudes of the gradients of different particles vary significantly in the numerical experiments and thus use the normalized gradient g g instead. 4.2 Training Objective Based on the velocity design in (10), we are now ready to derive a tractable training objective for hnet to learn the optimal direction in Ωthrough (7). Note that in principle, hnet should be trained after pt(Ω) = 1. However, in order to enhance efficiency in practice, we use inner particles to minimize RSD before all the particles are driven in the constrained domain. At this phase, we replace pt in (7) with the conditional measure ˆpt := pt( |Ω). For any v C(Ω, Rd), expand (7) and we get Ω ˆptv T log p Ω ˆpt v 2dx Ω ˆpt[v T log p + v]dx + 1 Ω ˆpt v 2dx + Z Ω ˆptv nd S. (11) Substitute v with the continuous extension of hnet on Ωand we obtain the training objective for hnet. Note that the first two terms in (11) are aligned with (5) by substituting in the integration domain, while the last term is a boundary integral specific to inequality constraints. This is one of the essential differences between training functional gradients on constrained and unconstrained domains. 4.3 Estimation of the Boundary Integral The boundary integral R Ωpv nd S is computationally intractable in general. Inspired by the co-area formula (Federer, 2014), we can estimate it using a heuristic band-wise approximation as follows Z Ω pv nd S 1 Sh pv ndx = 1 Sh p(x)v(x) np(x) p(x)dx = p(x Sh) Sh p(x)v(x) ndx, (12) where p(x) p(x) such that R Sh p(x)dx = 1, and Sh := {x Ω: d(x, Ω) h} is a band-like area near the boundary with a small bandwidth h > 0, which further has a proxy as g(x) 0, g x + h g(x) Please refer to Appendix A for more explanations. This way, we can use the particles in the band-like area to obtain a Monte Carlo estimate of the boundary integral. Suppose there are m particles in Ω, of which {xj}n j=1 are in the band-like area. Then p(x Sh) n m, and thus Ω pv nd S 1 mh j=1 v(xj)T g(xj)/ g(xj) . (14) We summarize all the techniques above and propose the Constrained Functional Gradient (CFG) in Algorithm 1. We use the neural network structure designed as in Appendix D for hnet. Algorithm 1 CFG: Constrained Functional Gradient Require: Unnormalized target distribution p , initial particles {xi 0}N i=1, initial neural network parameters w0 = {η0, ξ0}, weight parameter λ, iteration number L, L , particle step size α, parameter step size η, bandwidth h for k = 0, , L 1 do Assign w0 k = wk Identify {xij k }n j=1 := {xi k|xi k Sh}, {xlr k }m r=1 := {xl k|xl k Ω} for t = 0, , L 1 do Set hwt k = fηt k z2 ξt k g Compute b LRSD(w) = 1 r=1 [ log p (xlr k ) T hw(xlr k ) hw(xlr k ) 2 hw(xlr k ) 2] + 1 mh hw(xij k )T g(xij k ) Update wt+1 k = wt k η w b LRSD(wt k) end for Assign wk+1 = w L k Compute vwk+1 = (fηk+1 z2 ξk+1 g) 1{g<0} λ g g 1{g 0} Update particles xi k+1 = xi k + αvwk+1(xi k) for i = 1, , N end for return Particles {xi L}N i=1 5 Theoretical Analysis In this section, we present the convergence guarantee of CFG in terms of TV distance. Consider the continuous dynamic dxt = vt(xt)dt where vt is defined in (10). Let pt be the law of xt. We first investigate when particles can get into the domain. Assumption 5.1. g(x) C > 0 for x Ωc and M0 = max{g(x0) : x0 supp(p0)} < . For simplicity, we do not delve into the case that the particles outside Ωmay get stuck at local extremal point of g before getting in Ω. Based on this assumption, the particles will enter the constrained domain in finite time: Theorem 5.2. There exists a finite t0 M0 λC , such that g(xt) 0, t t0. The proof is deferred to Appendix B.1. Since domains with inequality constraints do not necessarily require the velocity field to gradually decay near the boundary as in Zhang et al. (2022), directly using λ g g 1{g 0} allows for a swift penetration of the particles into the constrained domain. Once the particles are in the constrained domain, an important technique is to extend the target distribution to be non-trivially supported on the whole space by convolution with a Gaussian kernel and use it as a proxy. Denote the convolution of the target distribution and the Gaussian distribution as ˆp = p N(0, σ2I), where σ > 0 can be arbitrarily small. Now the KL divergence DKL(pt ˆp ) is well defined on Rd. Moreover, by the convolution properties, TV(ˆp p ) 0 as σ 0 (Stein & Shakarchi, 2005). Assumption 5.3 (Poincaré inequality). The target distribution p satisfies varp [f] κEp [ f 2] for any smooth function f. This is a common assumption in convergence analysis that describes the degree of convexity of the target distribution (Chewi et al., 2022). In contrast to Zhang et al. (2022), our approach relies solely on the Poincaré property of the original target distribution, rather than requiring it for the conditional measure on a zero measure set. Following Cheng et al. (2023), we make the following assumption on the approximation error of neural networks. Assumption 5.4. For any t > t0, R Ωpt hnet(t) log p Now, we are ready to present the main result. Proposition 5.5. Suppose that p satisfies κ-PI and the data distribution pt is smooth. Denote F(pt, ˆp ) := Z Ω pt log ˆp then the following bound holds for any t > t0 TV(pt ˆp ) p 8(κ + σ2)(F(pt, ˆp )) + O(σ) (15) Relating the gradient of KL divergence and the Fisher information and letting σ 0, we have Theorem 5.6. Under assumption 5.4 and the assumptions in Proposition 5.5, suppose also that the KL divergence DKL(pt0 p ) < . The following bound holds min t T TV(pt p ) O(T 1 2 + ϵ 1 2 ). (16) Please refer to Appendix B.2 and B.3 for detailed proofs. 6 Experiments In this section, we compare CFG to other baseline methods for constrained domain sampling, including MSVGD (Shi et al., 2022), MIED (Li et al., 2022), Spherical HMC (Lan et al., 2014), PD-SVGD and Control SVGD (Liu et al., 2021). Throughout the section, we use neural networks with 2 hidden layers and initialize our particles with Gaussian distributions unless otherwise stated. We use the neural network structure designed as in Appendix D. All the experiments were implemented with Pytorch. More details can be found in Appendix D. The code is available at https://github. com/Shiyue Zhang66/Constrained-Functional-Gradient-Flow. 6.1 Toy Experiments We first conduct 2-D experiments to test the effectiveness of CFG on sampling from truncated Gaussian distributions within various constrained domains, including ring-shaped, cardioid-shaped and double-moon-shaped domains. We covered non-convex domains, including unconnected domains. We use 1000 particles for all domains. From the left of Figure 1 we can see that the particles quickly converge to the target distribution without escaping the constrained domain. Comparison to MIED (via various distributional metrics) on the first three domains are deferred to Appendix D.1, where CFG provides better approximation in most cases. To compare with MSVGD and MIED, we additionally conduct the experiment of sampling from truncated gaussian mixture distribution within the block-shaped domain. The initial distribution is the uniform distribution. The right of Figure 1 shows that our method outperformed MSVGD and achieved comparable results to MIED in terms of Wasserstein-2 distance and energy distance. 6.2 Bayesian Lasso The Lasso method is broadly used in model selection to avoid overfitting by imposing a penalty term on model parameters. Park & Casella (2008) introduced a Bayesian alternative, named Bayesian Lasso, that replaces the plenty term with the double exponential prior distribution pprior(β) exp( λ β 1). Lan et al. (2014) then proposed Spherical HMC method, which introduced more flexibility in choosing priors with explicit ℓq-norm constraints: pprior(β) p(β)1{ β q r}. For q = 1, it corresponds to the Lasso method. For more general case when q > 1, it is called Bayesian Bridge regression. Following Lan et al. (2014), we choose the prior to be the truncated Gaussian distribution, where p(β) = N(0, σ2I). This leads to the Bayesian regularized linear regression model (Lan et al., 2014): y|X, β, σ2 N(Xβ, σ2I), β|σ2 N(0, σ2I)1( β q r) (17) Ground Truth Ground Truth 4 2 0 2 4 4 Ground Truth Ground Truth 0 100 200 300 400 Iter Wasserstein-2 Distance (Sinkhorn) MSVGD CFG MIED 0 100 200 300 400 Iter Energy Distance MSVGD CFG MIED Figure 1: Left: CFG sampled particles at different numbers of iterations on constrained domains (ring, cardioid, double-moon, block). Right: The convergence curves of MSVGD, CFG and MIED on the block constraint. The posterior distribution for β is β|y, X, σ2 N(β , σ2(XT X + I) 1)1( β q r), where β = (XT X + I) 1XT y. This posterior has an inequality constraint and our method can apply. 6.2.1 Synthetic Dataset We first generate a 20-dimensional dataset (X, y), where X R1000 20, y R1000 and y = Xβtrue + ϵ, where ϵ N(0, 25I). We set the true regression coefficients to be βtrue = (10, . . . , 10, 0, . . . , 0), where the first half of components is tens and the second half is zeros. Let ˆβOLS denote the estimates obtained by ordinary least squares (OLS) regression. We follow Park & Casella (2008) and set r = ˆβOLS 1. We compare our method on Wasserstein-2 distance and energy distance with Spherical HMC and MIED using different number of particles. The ground truth is obtained via rejection sampling using 100,000 posterior samples. We use the h0(d N) 1 3 scheme to adapt the bandwidth to the number of particles N (see Appendix C), where we choose h0d 1 3 = 0.1. From Figure 2, we see that CFG performs the best in both metrics when N is small. This indicates the sample efficiency of particle-based variational inference for constrained sampling, which aligns with the findings in Liu & Wang (2016) for unconstrained sampling. All methods provide similar results when N is large. It is worth noting that as a kernel-based method, the time complexity of MIED scales quadratically as N increases, while functional gradient methods like CFG scale linearly. Meanwhile, we observed in our experiments that MIED tends to require more iterations to converge when the number of particles increases. These issues add to the overall time cost for MIED to achieve accurate approximation, especially when a large N is used (see Figure 9 in Appendix D.2). Similar issues of SVGD are also stated in Dong et al. (2023). Please refer to Appendix D.2 for detailed information. 6.2.2 Real Dataset Following Lan et al. (2014), we also evaluate our method using the diabetes dataset discussed in Park & Casella (2008). We compare the posterior median estimates given by Spherical HMC, CFG and MIED for the Lasso regression model (q = 1) and the Bridge regression model (q = 1.2), as the shrinkage factor s := r/ ˆβOLS 1 varies from 0 to 1. We use 5000 particles for CFG and MIED. 101 102 103 Sample Size (N) Wasserstein-2 Distance SPH CFG MIED 101 102 103 Sample Size (N) Energy Distance SPH CFG MIED Figure 2: Left: Wasserstein-2 distance of SPH, CFG and MIED versus the number of particles, Right: Energy distance of SPH, CFG and MIED versus the number of particles. Both on a synthetic dataset. 0.0 0.2 0.4 0.6 0.8 1.0 30 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Coefficients Shrinkage Factor s SPH CFG MIED Lasso Bridge 0.56 0.58 0.60 0.62 0.64 0.66 Test Accuracy Test Mono Loss SVGD PD SVGD Control SVGD MIED CFG threshold Figure 3: Left: Bayesian Lasso (q = 1) using Spherical HMC (upper left), CFG (upper middle) and MIED (upper right). Bayesian Bridge Regression (q = 1.2) using Spherical HMC (lower left) CFG (upper middle) and MIED (upper right). Right: Results of monotonic Bayesian neural network with ϵ = 0.01. Only the portion below 0.02 is shown on the y-axis to better display the performance of models satisfying constraint. From Figure 3 we can see that our method aligns well with Spherical HMC and MIED, which shows the effectiveness of our method. 6.3 Monotonic Bayesian Neural Network We use the COMPAS dataset following the setting in Liu et al. (2021). Given a Bayesian neural network ˆy( , θ), define the constraint function g(θ) = ℓmono(θ) ϵ with ℓmono(θ) = Ex D[ ( xmono ˆy(x; θ))+ 1]. Here, xmono denotes the subset of features to which the output should be monotonic. Note that in Liu et al. (2021), the constraint is only in the sense of expectation, i.e., Ep[g(θ)] 0. In the context of monotonic neural network (Karpf, 1991; Liu et al., 2020), however, the monotonicity constraint is g(θ) 0, which defines a domain with inequality constraints and thus our method can be applied. We use two-layer Re LU neural network with 50 hidden units. With different threshold ϵ {0.005, 0.01, 0.05}, we compare our CFG with MIED and two SVGD variants in Liu et al. (2021): PD SVGD and Control SVGD (C-SVGD). The number of particles is 200 and the results are shown in Table 1. We observe that in all the tasks, our method outperforms the other two baselines in terms of both test accuracy and test likelihood. Ratio Out" denotes the proportion of particles outside Ω= {g(θ) 0} to the total number of particles. All methods can successfully force almost all the particles into the domain. We further plot in the right of Figure 3 the averaged test monotonic loss ℓmono against test accuracy during the training process for ϵ = 0.01. For MIED, the test accuracy is higher when more particles are outside the domain, while lower when forcing the particles into the domain during training. Notably, Vanilla SVGD exhibits lower accuracy and higher monotonic loss, indicating the importance of constraint-sampling methods. Furthermore, our method can scale up to even higher dimensional real problems. We experimented on the COMPAS dataset using larger monotonic BNNs and on the 276-dimensional larger dataset Blog Feedback Liu et al. (2020). Please refer to Appendix D.3 for detailed information. Table 1: Results of monotonic Bayesian neural network under different monotonicity threshold. The results are averaged from the last 10 checkpoints for robustness. TEST ACC TEST NLL RATIO OUT (%) ε PD-SVGD C-SVGD MIED CFG PD-SVGD C-SVGD MIED CFG PD-SVGD C-SVGD MIED CFG 0.05 .647 .007 .649 .002 .596 .002 .661 .000 .634 .002 .633 .001 .684 .000 .632 .000 0.5 0.0 3.0 0.0 0.01 .645 .004 .650 .002 .590 .001 .660 .001 .635 .001 .634 .001 .678 .002 .632 .000 0.0 0.0 5.0 0.0 0.005 .645 .005 .650 .002 .586 .000 .659 .001 .635 .001 .633 .002 .676 .001 .632 .000 0.0 0.0 6.0 0.0 7 Conclusion We proposed a new functional gradient Par VI method for constrained domain sampling, named CFG, which uses neural networks to design a piece-wise velocity field that satisfies a non-escaping boundary condition. We presented novel numerical strategies to deal with boundary integrals arising from the domain constraints. We showed that our method has TV distance convergence guarantee. Empirically, we demonstrated the effectiveness of our method on various machine learning tasks with inequality constraints. Acknowledgments and Disclosure of Funding This work was supported by National Natural Science Foundation of China (grant no. 12201014 and grant no. 12292983). The research of Cheng Zhang was supported in part by National Engineering Laboratory for Big Data Analysis and Applications, the Key Laboratory of Mathematics and Its Applications (LMAM) and the Key Laboratory of Mathematical Economics and Quantitative Finance (LMEQF) of Peking University. The authors are grateful for the computational resources provided by the High-performance Computing Platform of Peking University. The authors appreciate the anonymous Neur IPS reviewers for their constructive feedback. Kwangjun Ahn and Sinho Chewi. Efficient constrained sampling via the mirror-langevin algorithm. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, 2021. URL https://openreview.net/forum?id=vh7q BSDZW3G. David Alvarez-Melis, Yair Schiff, and Youssef Mroueh. Optimizing functionals on the space of probabilities with input convex neural networks. Transactions on Machine Learning Research, 2022. ISSN 2835-8856. URL https://openreview.net/forum?id=dp OYN7o8Jm. D. M. Blei, A. Y. Ng, and M. I. Jordan. Latent dirichlet allocation. Journal of Machine Learning Research, 3(Jan):993 1022, 2003. URL https://jmlr.csail.mit.edu/papers/v3/blei03a. html. David M. Blei, Alp Kucukelbir, and Jon D. Mc Auliffe. Variational inference: A review for statisticians. Journal of the American Statistical Association, 112:859 877, 2016. Nicolas Brosse, Alain Durmus, Éric Moulines, and Marcelo Pereyra. Sampling from a log-concave distribution with compact support with proximal langevin monte carlo. In Conference on learning theory, pp. 319 342. PMLR, 2017. Marcus Brubaker, Mathieu Salzmann, and Raquel Urtasun. A family of mcmc methods on implicitly defined manifolds. In Artificial Intelligence and Statistics, pp. 161 172. PMLR, 2012. Sébastien Bubeck, Ronen Eldan, and Joseph Lehec. Sampling from a log-concave distribution with projected langevin monte carlo. Discrete & Computational Geometry, 59:757 783, 2018. Simon Byrne and Mark Girolami. Geodesic monte carlo on embedded manifolds. Scandinavian Journal of Statistics, 40(4):825 845, 2013. Changyou Chen, Ruiyi Zhang, Wenlin Wang, Bai Li, and Liqun Chen. A unified particle-optimization framework for scalable bayesian sampling. Ar Xiv, abs/1805.11659, 2018. Tianqi Chen, Emily Fox, and Carlos Guestrin. Stochastic gradient hamiltonian monte carlo. In International Conference on Machine Learning, pp. 1683 1691, 2014. Z. Cheng, S. Zhang, L. Yu, and C. Zhang. Particle-based variational inference with generalized wasserstein gradient flow. Advances in Neural Information Processing Systems, 2023. S. Chewi, M.A. Erdogdu, M. Li, R. Shen, and S. Zhang. Analysis of langevin monte carlo from poincaré to log-sobolev. Proceedings of Thirty Fifth Conference on Learning Theory, 178:1 2, 2022. Sinho Chewi, Thibaut Le Gouic, Chen Lu, Tyler Maunu, and Philippe Rigollet. Svgd as a kernelized wasserstein gradient flow of the chi-squared divergence. Advances in Neural Information Processing Systems, 33, 2020. Thomas A. Courtade. Bounds on the poincaré constant for convolution measures. Annales de l Institut Henri Poincaré - Probabilités et Statistiques, 56:566 579, 2020. Lauro Langosco di Langosco, Vincent Fortuin, and Heiko Strathmann. Neural variational gradient descent. ar Xiv preprint ar Xiv:2107.10731, 2021. Hanze Dong, Xi Wang, LIN Yong, and Tong Zhang. Particle-based variational inference with preconditioned functional gradient flow. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=6Oph WWAE3c S. S. Duane, A. D. Kennedy, B J. Pendleton, and D. Roweth. Hybrid Monte Carlo. Physics Letters B, 195(2):216 222, 1987. Jiaojiao Fan, Qinsheng Zhang, Amirhossein Taghvaei, and Yongxin Chen. Variational wasserstein gradient flow. In International Conference on Machine Learning, pp. 6185 6215. PMLR, 2022. H. Federer. Geometric measure theory. Springer, 2014. Will Grathwohl, Kuan-Chieh Wang, Jörn-Henrik Jacobsen, David Duvenaud, and Richard Zemel. Learning the stein discrepancy for training and evaluating energy-based models without sampling. In International Conference on Machine Learning, pp. 3732 3747. PMLR, 2020. Tianyang Hu, Zixiang Chen, Hanxi Sun, Jincheng Bai, Mao Ye, and Guang Cheng. Stein neural sampler. ar Xiv preprint ar Xiv:1810.03545, 2018. Michael I. Jordan, Zoubin Ghahramani, T. Jaakkola, and Lawrence K. Saul. An introduction to variational methods for graphical models. Machine Learning, 37:183 233, 1999. Richard Jordan, David Kinderlehrer, and Felix Otto. The variational formulation of the fokker planck equation. SIAM journal on mathematical analysis, 29(1):1 17, 1998. Jørgen Karpf. Inductive modelling in law: example based expert systems in administrative law. In Proceedings of the 3rd international conference on Artificial intelligence and law, pp. 297 306, 1991. S. Lan, B. Zhou, and B. Shahbaba. Spherical hamiltonian monte carlo for constrained target distributions. The 31st International Conference on Machine Learning, 32:629 637, 2014. N. E. Lewis, H. Nagarajan, and B. O. Palsson. Constraining the metabolic genotype-phenotype relationship using a phylogeny of in silico methods. Nature Reviews Microbiology, 10(4):291 305, 2012. Lingxiao Li, Qiang Liu, Anna Korba, Mikhail Yurochkin, and Justin Solomon. Sampling with mollified interaction energy descent. ar Xiv preprint ar Xiv:2210.13400, 2022. Chang Liu and Jun Zhu. Riemannian stein variational gradient descent for Bayesian inference. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018. Chang Liu, Jingwei Zhuo, Pengyu Cheng, Ruiyi Zhang, Jun Zhu, and Lawrence Carin. Understanding and accelerating particle-based variational inference. International Conference on Machine Learning. PMLR, pp. 4082-4092, 2019. Qiang Liu. Stein variational gradient descent as gradient flow. Advances in neural information processing systems, 30, 2017. Qiang Liu and Dilin Wang. Stein variational gradient descent: A general purpose bayesian inference algorithm. Advances in neural information processing systems, 29, 2016. Xingchao Liu, Xing Han, Na Zhang, and Qiang Liu. Certified monotonic neural networks. Advances in Neural Information Processing Systems, 33:15427 15438, 2020. Xingchao Liu, Xin Tong, and Qiang Liu. Sampling with trusthworthy constraints: A variational gradient framework. Advances in Neural Information Processing Systems, 34:23557 23568, 2021. B. J. Morris. Improved bounds for sampling contingency tables. Random Structures & Algorithms, 21(2):135 146, 2002. Radford Neal. MCMC using hamiltonian dynamics. In S Brooks, A Gelman, G Jones, and XL Meng (eds.), Handbook of Markov Chain Monte Carlo, Chapman & Hall/CRC Handbooks of Modern Statistical Methods. Taylor & Francis, 2011. ISBN 9781420079425. URL http://books. google.com/books?id=qf Rs AIKZ4r IC. T. Park and G. Casella. The bayesian lasso. Journal of the American Statistical Association, 103(482): 681 -686, 2008. Andrey Pilipenko. An introduction to stochastic differential equations with reflection, volume 1. Universitätsverlag Potsdam, 2014. G. O. Robert and O. Stramer. Langevin diffusions and metropolis-hastings algorithms. Methodology and Computing in Applied Probability, 4:337 357, 2002. Adil Salim and Peter Richtárik. Primal dual interpretation of the proximal stochastic gradient langevin algorithm. Advances in Neural Information Processing Systems, 33:3786 3796, 2020. Jiaxin Shi, Chang Liu, and Lester Mackey. Sampling with mirrored stein operators. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id= e Mudn Jsb1T5. Elias M. Stein and Rami Shakarchi. Real Analysis. Princeton University press, 2005. I. Thiele, N. Swainston, R. M. Fleming, A. Hoppe, S. Sahoo, M. K. Aurich, H. Haraldsdottir, M. L. Mo, O. Rolfsson, M. D. Stobbe, and et al. A community-driven global reconstruction of human metabolism. Nature Biotechnology, 31(5):419 425, 2013. M. J. Wainwright and M. I. Jordan. Graphical models, exponential families, and variational inference. Foundations and Trends in Maching Learning, 1(1-2):1 305, 2008. Max Welling and Yee W Teh. Bayesian learning via stochastic gradient langevin dynamics. In International Conference on Machine Learning, pp. 681 688, 2011. Kelvin Shuangjian Zhang, Gabriel Peyré, Jalal Fadili, and Marcelo Pereyra. Wasserstein control of mirror langevin monte carlo. In Conference on Learning Theory, pp. 3814 3841. PMLR, 2020. Ruqi Zhang, Qiang Liu, and Xin T. Tong. Sampling in constrained domains with orthogonal-space variational gradient descent. Advances in Neural Information Processing Systems, 2022. A Estimation of Boundary Integral The boundary integral R Ωpv nd S is a crucial factor that distinguishes Par VI on constrained domain from unconstrained case, due to its computational intractability in general. We derive a heuristic band-wise approximation (12) and thus the remaining issue is to identify which particle is in the band-wise area {d(x, Ω) h}. Proposition A.1. For any h > 0, the sufficient condition for {x Ω: d(x, Ω) h} is ( g(x) 0, g x + h g(x) g(x) 0. (18) Proof. For any x Ω, define f(r) = g x + r g(x) g(x) . Then f( ) is continuous with f(0) = g(x) 0. If f(h) 0, by intermediate value theorem, there exists h0 [0, h] such that f(h0) = 0. And we have either g(x) = 0 or g(x0) = 0 where x0 = x + h0 g(x) g(x) . Therefore x Ωor x0 Ω, implying d(x, Ω) h since x x0 = h0 h. In fact, if g(x) 0, then d(x, Ω) h is equivalent to max y x h g(y) 0. When h is small and g(x) = 0, the maximum point of LHS is approximately y = x + h g(x) g(x) since g(x) is the steepest ascent direction. Hence (18) is a reasonable proxy when h is small. B.1 Proof of Theorem 5.2 Theorem B.1. Let M0 = max{g(x0)} < , there exists a finite t0 M0 λC , such that g(xt) 0, t t0. Proof. It suffices to prove that for any z0, there exists a finite t0 g(z0) λC , such that g(zt) 0, t t0. Since d dtg(zt) = ( g)T v = ( g)T hnet 1{g<0} λ g 1{g 0} Using assumption g > C, we can state that there exists a finite t0 g(z0) λC , such that g(zt) 0, t t0. Firstly prove that there is a t0 g(z0) λC such that g(zt0) 0. By contradiction, if such t0 does not exist, we have g(zt) > 0 for all t [0, g(z0) λC ], then d λ g , integrate both sides yields g(zt) g(z0) = R g(z0) λC 0 λ g dt g(z0), then g(zt) 0, contradiction. Then prove that g(zt) 0, t t0. If not, there exists a minimum t > t0, such that g(zt) > 0, then d dtg(zt) = λ g < 0, which means for sufficiently small δ > 0, g(zt δ) > 0, contradiction. B.2 Proof of Proposition 5.5 Proposition B.2. Suppose that p satisfies κ-PI and the data distribution pt is smooth. Denote F(pt, ˆp ) := R Ωpt sˆp spt 2dx, then the following bound holds for t > t0 TV(pt ˆp ) p 8(κ + σ2)(F(pt, ˆp )) + O(σ) (19) Proof. By equivalent expression of TV distance, it suffices to prove that for any function f satisfying |f| 1, |Eˆp [f] Ept[f]| p 8(κ + σ2)F(pt, ˆp ) + O(σ) Note that for t > t0, pt(Ω) = 1 |Eˆp [f] Ept[f]| = Z 1{g 0}ˆp (x)f(x)dx Z 1{g 0}pt(x)f(x)dx + Z 1{g>0}ˆp (x)f(x)dx Z 1{g 0}(ˆp (x) pt(x))f(x)dx + Z 1{g>0}ˆp (x)f(x)dx Z 1{g 0}(ˆp (x) pt(x))f(x)dx + ϵ2. where ϵ2 = R 1{g>0}(ˆp (x) p (x))f(x)dx = O(σ) by the L1 convergence of Gaussian convolution Stein & Shakarchi (2005). The first term can be bounded by Z 1{g 0}(ˆp (x) pt(x))f(x)dx 2 = Z 1{g 0}ˆp (x)( pt(x) ˆp (x) 1)f(x)dx 2 Z 1{g 0}ˆp (x)( pt(x) ˆp (x) 1)2dx Z 1{g 0}ˆp (x)( pt(x) ˆp (x) + 1)2f(y)2dx 2 Z 1{g 0}ˆp (x)( pt(x) ˆp (x) 1)2dx Z 1{g 0}ˆp (x)( pt(x) ˆp (x) + 1)dx 4 Z 1{g 0}ˆp (x)( pt(x) ˆp (x) 1)2dx 8(1 Z 1{g 0} p pt(x)ˆp (x)dx) pt(x)ˆp (x)dx) = 8(1 b) with b = Z ˆp (x) pt(x) ˆp (x)dx 1. Then note that pt(x) ˆp (x) = Eˆp [ pt(x) ˆp (x)] b2 = 1 b2. From Courtade. (2020), we have the Poincaré coefficient of N(0, σ2I) is σ2, and the Poincaré coefficient of ˆp = p N(0, σ2I) is no larger than κ + σ2. Therefore, by the Poincaré inequality, we have 1 b (1 b2) = varˆp pt(x) ˆp (x) (κ + σ2) Z ˆp (x) pt(x) ˆp (x) = (κ + σ2) Z 1{g 0}pt(x) spt(x) sˆp (x) 2dx = (κ + σ2)F(pt, ˆp ) And |Eˆp [f] Ept[f]| p 8(κ + σ2)F(pt, ˆp ) + ϵ2 with ϵ2 = O(σ). B.3 Proof of Theorem 5.6 Theorem B.3. Following the same assumptions in Proposition 5.5. Suppose also that the KL divergence DKL(pt0, p ) < , the following bound holds as σ 0 min t T TV(pt p ) O(T 1 2 + ϵ 1 2 ). (20) Proof. By equivalent expression of TV distance, it suffices to prove that for any function f satisfying |f| 1, min t T |Ep [f] Ept[f]| O(T 1 2 + ϵ 1 2 ) We now only consider the time after t0, which means that pt(Ω) = 1. First we state the relation between ˆp = p N(0, σ2I) and p (x). When σ 0, by the pointwise convergence of the convolution Stein & Shakarchi (2005), we have ˆp (x) p (x) and ˆp (x) p (x). Note that for g(x) 0, ˆp (x) and p (x) are strictly positive, thus sˆp (x) sp (x), as σ 0 for any x Ω. Note that ˆp (x) is supported on Rd, we have d dt DKL(pt(x) ˆp (x)) = Z ptv T t (sˆp spt)dx = Z pt(hnet 1{g<0} λ g g 1{g 0})T (sˆp spt)dx = Z pt(hnet 1{g 0})T (sˆp spt)dx + Z pt((hnet + λ g g ) 1{g=0})T (sˆp spt)dx = Z pt 1{g 0} sˆp spt 2dx Z pt 1{g 0}(hnet (sˆp spt))T (sˆp spt)dx Z pt 1{g 0} sˆp spt 2dx + ϵ0 Z pt 1{g 0} sˆp spt 2dx + 1 Z pt 1{g 0} hnet (sˆp spt) 2dx (1 ϵ0) Z pt 1{g 0} sˆp spt 2dx + 1 Z pt 1{g 0} hnet (sp spt) 2dx Z pt 1{g 0} sˆp sp 2dx For ϵ0 < 1, the second term is O(ϵ), and the third term is O(σ). Using the notation F(pt, ˆp ) := R pt 1{g 0} sˆp spt 2dx, integrating both sides yields the following: t0 F(pt, ˆp )dt Z T 0 F(pt, ˆp )dt 1 1 ϵ0 DKL(p0, ˆp ) + T(O(ϵ) + O(σ)). min t0 t T F(pt, ˆp ) 2 (1 ϵ0)T DKL(p0, ˆp ) + O(ϵ) + O(σ) = O(T 1) + O(ϵ) + O(σ) (21) Plugging the result of F(pt, ˆp ) (21) in Proposition 5.5 and let σ 0, we have min t0 t T |Eˆp [f] Ept[f]| O(T 1 2 + ϵ 1 2 ). Finally, by triangular inequality: |Ep [f] Ept[f]| |Eˆp [f] Ept[f]| + |Eˆp [f] Ep [f]| Using the result that Gaussian convolution converge to original L1 function in L1 (Stein & Shakarchi, 2005), we have |Eˆp [f] Ep [f]| 0 when σ 0 for any p . This means mint0 t T |Ep [f] Ept[f]| O(T 1 2 + ϵ 1 2 ), the proof is complete. C Insights on Bandwidth Selection Error Analysis of Boundary Integral Estimation We first present the error analysis of band-wise approximation, based on which we derive a heuristic O((d N) 1 3 ) scheme for bandwidth selection when adjusting the number of particles. Denote the approximation error e = 1 mh Pn j=1 v(xj)T g(xj)/ g(xj) R Ωpv nd S. We claim that the mean square error E[e2] = O(h2 + 1 The idea is to leverage bias-variance decomposition. We estimate the bias of the band-wise approximation, using Lagrange s Mean Value Theorem Sh p(x)v ndx Z d(x, Ω)=l p(x)v nd S Z Ω p(x)v nd S Then we turn to bound variance. From (12) we have 1 h R Sh p(x)v ndx = p(x Sh) h R Sh p(x)v(x) ndx. And j=1 v(xj)T n(xj) p(x Sh) Sh p(x)v(x) ndx j=1 v(xj)T n(xj) n Sh p(x)v(x) ndx Sh p(x)v(x) ndx p(x Sh) Sh p(x)v(x) ndx The variance of Monte Carlo estimation is E[A2 3] = O( 1 By Central Limit Theorem, E[( n m p(x Sh))2] = O( 1 m), and thus E[A2 4] = O( 1 Finally, through bias-variance decomposition, we have E[e2] = E[A2 1 + A2 2] = O(E[A2 1 + A2 3 + A2 4]) = O(h2 + 1 m) = O(h2 + 1 102 103 104 105 106 Sample Size (N) v = v1 v = v2 v = v3 102 103 104 105 106 Sample Size (N) v = v1 v = v2 v = v3 102 103 104 105 106 Sample Size (N) v = v1 v = v2 v = v3 Figure 4: Left: MSE of boundary integral estimation of distribution p1. Middle: MSE of boundary integral estimation of distribution p2. Right: MSE of boundary integral estimation of distribution p3. However, n depends on h implicitly. We use a simple example to determine this relation. Consider the target distribution being the uniform distribution, and the constrained domain being the ball in Rd with radius R, then p(x Sh) Rd (R h)d Rd = O(dh). In this case, the total mean square error is of order O(h2 + 1 d Nh), which implies the optimal h is O((d N) 1 3 ), in this case the total mean square error is of order O((d N) 2 In practice, one can set h = h0(d N) 1 3 and tune h0 through sparse grid search. Simulation Verification of Boundary Integral Estimation To verify the above analysis, we conduct a toy 2D simulation study on estimating boundary integral of 3 different velocities at the boundary using 5 different numbers of samples from 3 different distributions. The constrained domain is the block area Ω= {x | |x1| 2, |x2| 2}. The number of particles are 102, 103, 104, 105, 106. The types of velocity and distribution are listed in Table 2. The corresponding true values of boundary integral R Ωpv nd S are listed in the Table 3. Table 2: The types of velocities and distributions. Types 1 2 3 v(x) n (x2, x1) (x2 2, x2 1) p(x) Unif(Ω) N((0, 0), I) on Ω N((0, 2), I) on Ω Table 3: The true values of boundary integral in the verification experiment. Distributions Velocities v1 v2 v3 p1 1 0 0 p2 0.226259 0 0 p3 0.911333 0 -0.617187 We estimate E[e2] by 10 trials for different number of particles and plot the curves as in Figure 4. We set h = h0(d N) 1 3 and h0d 1 3 = 0.5. The slope of the log-log plot is approximately 2 3 for all types of velocities and distributions, which supports the previous analysis. Additionally, if we do not follow the scheme of h = h0(d N) 1 3 , and for example, fixing h = 0.5 (102) 1 3 or h = 0.5 (106) 1 3 (the starting and ending edgewidth of the adaptive edgewidth scheme h = 0.5 N 1 3 ). Figure 5 shows that the estimated E[e2] will tend to go up in certain cases, which supports the rationality of the h = h0(d N) 1 102 103 104 105 106 Sample Size (N) h = 0.5 (102) h = 0.5 (106) Figure 5: MSE of boundary integral estimation of distribution p1 and velocity v1 using fixed edgewidths and adaptive edgewidth. D Additional Details of Experiments Structure of Neural Networks We parameterize hnet in (10) using two neural networks fnet and znet as follows hnet = fnet z2 net g, (25) where fnet : Rd Rd, znet : Rd R. Here, the z2 net g term relates to the reflection process in reflected stochastic differential equation Pilipenko (2014), and always points inward Ω. It can be regarded as a reflection term that forces the particle to stay inside Ωas it neutralizes the outward normal-pointing component. Heuristically, g serves as the reflection direction and z2 net contributes in learning the magnitude. Empirically, we observe that incorporating the gradient of the boundary g substantially alleviates the issue of particles adhering to the boundary during training, and may slightly improve the sampling quality. Please refer to D.1 below. D.1 Toy Experiments Setting Details On toy experiments, we conduct CFG on four 2-D constrained distributions. The experiments are implemented on Intel 2.30GHz CPU with RAM 16384MB and NVIDIA Ge Force RTX 3060 Laptop GPU with total memory 14066MB. The density and constrained domain are in Table 4. The truncated gaussian mixture is 9 equally weighted gaussian distribution with all standard variance equal to 0.2, the centers are (-1.7, -1.7), (-1.7, 0), (-1.7, 1.7), (0, -1.7), (0, 0), (0, 1.7), (1.7, -1.7), (1.7, 0), (1.7, 1.7). Table 4: Four 2-D constrained distributions implemented in the toy experiments. Name Density Constrained domain Ring p (x) N(0, I) on Ω Ω= {x | 1 x 2 4} Cardioid p (x) N(0, I) on Ω Ω= {x | x2 1 + ( 6 5x2 x2/3 1 )2 4} Double-moon p (x) q on Ω Ω= {x| log q(x) 2, where q(x) = e 2(x1 3)2+e 2(x1+3)2 e2( x 3)2 } Block Truncated gaussian mixture on Ω Ω= {x | |x1| 2 and |x2| 2} For CFG, fnet and znet are three-layer neural networks with Leaky Re LU activation (negative slope=0.1). The number of hidden units is 128, except for the RING on which 256 is used. Both neural nets are trained by Adam optimizer with learning rate 0.002, except for RING on which 0.005 is used. The number of inner-loop of gradient updates for fnet and znet is set to 10, except for the RING on which 3 is used. The total number of iterations is 2000 and the step size of particle is 0.005 except for the RING on which 0.01 is used. The band width is set to 0.05 except for BLOCK on which 0.001 is used. λ in the piece-wise construction of the velocity field is chosen to be 1. On the BLOCK experiment, the total number of iterations is 2000. The ground truth is obtained by rejection sampling using 104 posterior samples. We run the accuracy experiments on 3 random seeds. The average results and the standard errors of the means are represented in the figure using lines and shades. For MSVGD, we follow Shi et al. (2022) and set the learning rate to 0.05 (selected from {0.01, 0.05, 0.1}). The kernel is the IMQ kernel and the kernel width is 0.1 (selected from {0.05, 0.1, 0.2}). For MIED, we use the default hyperparameters in Li et al. (2022) and set the particle stepsize to be 0.01 (selected from {0.005, 0.01, 0.02}). Additional Comparison of MIED Table 5 shows the additional comparison of CFG and MIED on the first three constrained domains. CFG achieved better results in most of the cases. Table 5: Wasserstein-2 distance and energy distance between the target distribution and the variational approximation on the three toy datasets. Name Wasserstein-2 distance (Sinkhorn) Energy distance MIED CFG MIED CFG Ring 0.1074 0.1087 0.0004 0.0003 Cardioid 0.1240 0.1141 0.0016 0.0005 Double-moon 0.4724 0.1660 0.0629 0.0022 Additional Ablation Studies From Fig 6 and Table 6 we can see that, without estimating the boundary integral leads to failure in sampling in most cases. This indicates the essential difference between constrained and unconstrained domain sampling. Adding znet can achieve a slight improvement in avoiding clustering near the boundary and better KL convergence. Table 6: Ablation results of not estimating boundary integral, with and without znet. Wasserstein-2 distance (Sinkhorn) Energy distance Ring Cardioid Double-moon Block Ring Cardioid Double-moon Block w/o boundary integral 0.2138 0.2321 0.4866 0.2438 0.0097 0.1147 0.0068 0.0073 w/o znet 0.1248 0.2234 0.1217 0.2422 0.0013 0.0009 0.0049 0.0073 w/ znet 0.1087 0.1660 0.1141 0.2416 0.0003 0.0005 0.0022 0.0072 Geometric Generalization of Accommodating Multiple Constraints Our proposed method can generalize and accommodate multiple constraints (including more equality and inequality constraints) and more complicated geometries. For additional equality constraints, our proposed framework can still apply by adopting the idea of Zhang et al. (2022). We could use velocity fields like v (x) to guided the particles to satisfy the equality constraints, and use our method to propose velocity fields substituting v (x) to satisfy the inequality constraints. Adding these two types of velocity presents the desired velocity field. To illustrate the idea, we demonstrate the training process of a 3D toy ring distribution example. Suppose the coordinate of the particle is x = (x, y, z). The target distribution is a truncated standard gaussian located in the ring shaped domain in x Oy plane. This corresponds to the equality constraint z = 0 and the inequality constraint 1 x2 + y2 4. The initial distribution is the 3D standard gaussian distribution. We choose v (x) = sign(z)|z|1.5. From figure 7 we can see that the particles collapse to the x Oy plane and converge inside the ring domain. For additional inequality constraints, we need multiple velocity field outside the constrained domain for particles to enter the constrained domain, and the boundary integral term can be similarly estimated using band-wise approximation. D.2 Bayesian Lasso Setting Details The experiments are implemented on Intel 2.30GHz CPU with RAM 16384MB and NVIDIA Ge Force RTX 3060 Laptop GPU with total memory 14066MB. 3w/o boundary integral Ground Truth 2 0 2 2 0 2 2 0 2 Ground Truth 4 2 0 2 4 4 4 2 0 2 4 4 2 0 2 4 4 2 0 2 4 Ground Truth 2 0 2 2 0 2 2 0 2 Ground Truth Figure 6: Ablation sampling results of not estimating boundary integral (left), not using znet (middle left), using znet (middle right) and the ground truth (right). On synthetic dataset, we run the experiments on 5 random seeds. The average results and the standard errors of the means are represented in the figure using lines and shades. The number of particles N are chosen from {5, 15, 30, 70, 180, 400, 900, 2000}. For CFG, fnet and znet are three-layer neural networks with Leaky Re LU activation (negative slope=0.1). The number of hidden units is 256. Both neural nets are trained by Adam optimizer with learning rate 0.0005 for 10 iterations in the inner loop. The total number of iterations is 1000 and the stepsize of particle is 0.004. λ in the piece-wise construction of the velocity field is chosen to be 1. For Spherical HMC, we follow the same main setting as Lan et al. (2014). The number of burn-in epochs is 4000, and the number of leap-frog is selected from {50, 100}. For MIED, we use the default hyperparameters in Li et al. (2022) and set the particle stepsize to gradually decay when the particle number N goes up. We set particle stepsize to 0.5 for N {5, 15, 30, 70}, 0.35 for N {180, 400}, 0.3 for N = 900, and 0.2 for N = 2000. The number of iterations is 4000 for convergence except for N = 2000, which needs 10000 iterations to converge. 3 2 1 0 1 2 3 3 3 2 1 0 1 2 3 3 2 1 0 1 2 3 3 3 2 1 0 1 2 3 3 2 1 0 1 2 3 3 3 2 1 0 1 2 3 3 2 1 0 1 2 3 3 3 2 1 0 1 2 3 Ground Truth Figure 7: Illustration of generalizing our method to accommodating equality and inequality constraints. 101 102 103 Sample Size (N) log10(Wasserstein-2 Distance) SPH CFG MIED 101 102 103 Sample Size (N) log10(Wasserstein-2 Distance) SPH CFG MIED 101 102 103 Sample Size (N) log10(Wasserstein-2 Distance) SPH CFG MIED 101 102 103 Sample Size (N) log10(Wasserstein-2 Distance) SPH CFG MIED 101 102 103 Sample Size (N) Energy Distance SPH CFG MIED 101 102 103 Sample Size (N) Energy Distance SPH CFG MIED 101 102 103 Sample Size (N) Energy Distance SPH CFG MIED 101 102 103 Sample Size (N) Energy Distance SPH CFG MIED Figure 8: First row: Wasserstein-2 distance of SPH, CFG and MIED versus the number of particles on dimensions 5, 10, 15, 20, Second row: Energy distance of SPH, CFG and MIED versus the number of particles on dimensions 5, 10, 15, 20. Both on a synthetic dataset. On real diabetes dataset, for CFG, fnet and znet are three-layer neural networks with Leaky Re LU activation (negative slope=0.1). The number of hidden units is 50. Both neural nets are trained by Adam optimizer with learning rate 0.005 for 10 iterations in the inner loop. The total number of iterations is 300 and the bandwidth h = 1. The number of particles is 5000 and the stepsize of particle is selected from {0.9, 1.05, 1.2}. For Spherical HMC, we follow the same setting as Lan et al. (2014). The number of total epochs is 11000 and the number of burn-in epochs is 1000. For MIED, we use the default hyperparameters in Li et al. (2022) and the particle stepsize is selected from {0.1, 0.2, 0.5}. The number of iterations is 2000. Additional Comparisons on the Synthetic Dataset We further compare the sampling results of SPH, CFG and MIED on different dimensions. We choose dimensions 5, 10, 15 and 20. From the design of βtrue, the samples of first two dimensions are close to 10, and the last two dimensions are close to 0. Figure 8 shows that CFG achieves comparable results to MIED on Wasserstein-2 distances, while slightly better result on energy distance in most of the dimensions. SPH eventually catches up when the particle number grows up. This alignes with the results stated in Section 6.2.1. The Nonlinear Time Complexity of MIED Larger number of particles is important for higher sample qualities. Similar to Dong et al. (2023), CFG is more scalable than MIED in terms of particle numbers N. This is because CFG is obtained with iterative approximation, the complexity is O(N). On the other hand, MIED incurs complexity O(N 2) due to the calculation of the weight denominator P i,j e Ii,j. Figure 9 plot the Wasserstein-2 Distance versus the computation time using at most 1000 iterations for CFG and 10000 iterations for MIED. The number of particles are N {900, 2000, 4000}. It is clear that MIED entails more computation cost with the increase of particle numbers for larger N. 0 200 400 600 800 Time (s) Wasserstein-2 Distance CFG(N = 900) MIED(N = 900) CFG(N = 2000) MIED(N = 2000) CFG(N = 4000) MIED(N = 4000) Figure 9: Wasserstein-2 distance of CFG and MIED versus the training time using different number of particles. 101 102 103 Sample Size (N) log10(Energy Distance) CFG(h=0.1) CFG(h=0.01) Figure 10: Choosing the adaptive bandwidth (red) against fixed bandwidths for the Bayesian Lasso experiment on a synthetic dataset. From the ending points of each curves (denotes the computation time for 1000 iterations of CFG and 10000 iterations of MIED), the time cost of CFG only grows linearly (from 130s to 280s to 600s approximately), while the time cost of MIED grows quadratically (from 100s to 220s to 830s approximately), which supports the above claim of time complexity. Additionally, figure 9 indicates that for large N, the number of iterations for MIED should also increase. 4000 iterations is sufficient for N = 900, while 10000 iterations is needed for N = 2000, and more than 10000 iterations is needed for MIED to completely converge to achieve better sampling results. The Effect of Choosing the Adaptive Bandwidth Figure 10 compared the energy distance between the adaptive bandwidth scheme h = 0.1N 1 3 and fixed bandwidth h = 0.1, 0.01 of CFG. We can see that the adaptive scheme achieved slightly better energy distance results than the fixed bandwidth. D.3 Monotonic Bayesian Neural Network Setting Details The experiments are implemented on Nvidia Ge Force RTX 4080 Laptop GPU with memory 12 GB. We use 200 particles and 200 data batch size for stochastic gradient in all methods. For CFG, both fnet and znet are three layers with Leaky Re LU activation (negative slope = 0.1). fnet have 300 hidden units while znet have 200. Both neural nets are trained by Adam optimizer with learning rate 0.001 for 10 iterations in the inner loop. The bandwidth h = 0.02 and the stepsize of particle is 5e-5. The total number of iterations is 1200. λ in the piece-wise construction of the velocity field is chosen to be 100. For PD SVGD and Control SVGD, we run 3000 iterations with particle stepsize selected from {0.001, 0.002, 0.005} to ensure convergence. For MIED, we run 1500 iterations with default hyperparameters in Li et al. (2022) and particle stepsize selected from {2, 3, 5} 10 4. Additional Experiments on Scaling Up to Higher Dimensions To test our method on higher dimension, we first experiment on the COMPAS dataset using larger monotonic BNNs. The number of neurons in the layer of BNN increased from 50 to 100, making the particle dimension increase from 903 to 1502. From Table 7, our method still achieved the best accuracy results compared to other methods, while achieving competitive log-likelihood results. For even higher dimension, we additionally experiment on the 276-dimensional larger dataset Blog Feedback Liu et al. (2020) using monotonic BNN. The particle dimension is 13903. From Table 8, our method still achieved the best result. Table 7: Results of a larger monotonic Bayesian neural network on COMPAS dataset under different monotonicity threshold. The results are averaged from the last 10 checkpoints for robustness. For each monotonicity threshold, the best result is marked in black bold font and the second best result is marked in brown bold font. Positive proportion of particles outside the constrained domain is marked in red. TEST ACC TEST NLL RATIO OUT (%) ε PD-SVGD C-SVGD MIED CFG PD-SVGD C-SVGD MIED CFG PD-SVGD C-SVGD MIED CFG 0.05 .618 .006 .639 .004 .568 .000 .649 .001 .639 .001 .630 .001 .665 .000 .637 .000 0.0 0.0 0.0 0.0 0.01 .618 .006 .638 .004 .569 .000 .651 .002 .639 .002 .631 .001 .664 .000 .640 .000 13.0 0.0 0.0 0.0 0.005 .618 .006 .638 .004 .571 .001 .653 .003 .639 .001 .631 .001 .664 .000 .637 .000 24.0 0.0 0.0 0.0 Table 8: Results of monotonic BNN on a higher 276-dimensional dataset Blog Feedback. The particle dimension is 13903. The results are averaged from the last 10 checkpoints for robustness. TEST RMSE RATIO OUT (%) PD-SVGD C-SVGD MIED CFG PD-SVGD C-SVGD MIED CFG .205 .011 .217 .000 .212 .001 .204 .000 0.0 0.0 0.0 0.0 E Limitations and Future Work Extensions to SVGD. The main idea of our work is enforcing the particles into the constrained domain, and then using neural networks to capture the target distribution. The latter part can also be accomplished by using kernels like in SVGD. The boundary integral may be treated in the similar way. We believe that our framework can be adapted to a larger family of variational sampling methods. More Efficiently Choosing Bandwidth h. In this paper, we only consider the simple case of uniform target distribution in choosing a better adaptive scheme for bandwidth h. Proposing an adaptive scheme based on the current particles is an interesting topic, which will be left for future work. Extensions to GWG. Cheng et al. (2023) proposed a variational framework including general geometries by using lp norm regularization. In this case, the treatment of boundary integral is the same as in the RSD loss expansion. The effect of using other regularization terms is left for future study. Relations with the Reflected Stochastic Differential Equation. Contrast to SDE-based sampling methods, our work is based on ODE. Both SDE and ODE can simulate trajectories including reflection. We proposed a heuristic understanding of the relation between SDE and ODE reflection in this work, a more thorough theoretical analysis may be presented. We leave this to future research. F Broader Impact This paper presents work whose goal is to advance the field of machine learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here. As far as we are concerned, our paper has no potential negative societal impacts. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: The last four sentences of the abstract and the last four sentences of the introduction reflect the paper s contributions and scope. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: Please refer to the appendix E. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: Please refer to appendix B. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: Please refer to appendix D and the codes and toy datasets in the supplemental materials. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: Please refer to the codes and toy datasets in the supplemental materials. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: Please refer to appendix D and the codes and toy datasets in the supplementary materials. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: Please refer to the right of figure 1, figure 2 and figure 8. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: Please refer to appendix D and figure 9. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: We follow the Neur IPS Code of Ethics. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [Yes] Justification: Please refer to appendix F. As far as we are concerned, our paper has no potential negative societal impacts. We do not engage in negative fairness, privacy or security issues. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: Our paper poses no such risks for misuse. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: We cite the original papers in the experiments. Please refer to Section 6. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [Yes] Justification: Please refer to the supplemental materials. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: Our paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: Our paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.