# langevin_monte_carlo_beyond_lipschitz_gradient_continuity__a9839ed7.pdf Langevin Monte Carlo Beyond Lipschitz Gradient Continuity Matej Benko1, Iwona Chlebicka2, Jørgen Endal3, Bła zej Miasojedow2 1Institute of Mathematics, Faculty of Mechanical Engineering, Brno University of Technology Technick a 2896/2, 616 69 Brno, Czech Republic 2 Institute of Applied Mathematics and Mechanics, University of Warsaw ul. Banacha 2, 02-097 Warsaw, Poland 3Department of Mathematical Sciences, Norwegian University of Science and Technology (NTNU) N-7491 Trondheim, Norway matej.benko@vutbr.cz, i.chlebicka@mimuw.edu.pl, jorgen.endal@ntnu.no, b.miasojedow@mimuw.edu.pl We present a significant advancement in the field of Langevin Monte Carlo (LMC) methods by introducing the Inexact Proximal Langevin Algorithm (IPLA). This novel algorithm broadens the scope of problems that LMC can effectively address while maintaining controlled computational costs. IPLA extends LMC s applicability to potentials that are convex, strongly convex in the tails, and exhibit polynomial growth, beyond the conventional L-smoothness assumption. Moreover, we extend LMC s applicability to super-quadratic potentials and offer improved convergence rates over existing algorithms. Additionally, we provide bounds on all moments of the Markov chain generated by IPLA, enhancing its analytical robustness. Code https://github.com/192459/lmc-beyond-lipschitzgradient-continuity Introduction Langevin Monte Carlo methods are powerful tools for sampling and optimization in complex, high-dimensional problems across various fields, including machine learning, statistics, physics, and beyond. When the exact form of a distribution is unknown or difficult to compute, efficient sampling is essential for high-dimensional data, where traditional sampling methods often become computationally prohibitive. The ability of such algorithms to scale efficiently with dimensionality and manage non-smooth elements enhances their applicability in various machine learning tasks, including regression, classification, and clustering, providing a powerful tool for developing robust and accurate models. Given the potential V : Rd R and associated density µ (x) = exp( V (x)) R Rd exp( V (y)) dy our goal is to generate samples X1, . . . , Xn approximately from µ . The ability to perform effective sampling from a distribution with a density known up to a constant is crucial in Bayesian statistical inference (Gelman et al. 2014; Robert and Casella 2004), machine learning (Andrieu et al. 2003), ill-posed inverse problems (Stuart 2010), and computational physics (Krauth 2006). Copyright 2025, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. One of the most efficient solutions to the problem of generating a sample from µ is Langevin Monte Carlo (LMC), which was designed via discretization of Overdamped Langevin Equation d Yt = V (Yt)dt + where Bt is a d-dimensional Brownian motion. Classically LMC is derived by Euler Maruyama discretization of the above diffusion process. Namely, we iteratively define Markov Chains such that for a given X0 and all k 0 we have Xk+1 = Xk τk V (Xk) + 2τk Zk+1 , where {τk}k 0 is sequence of stepsizes and {Zk+1}k 0 are i.i.d standard d-dimensional Gaussian random variables which are independent of X0. In the recent years the LMC algorithm has been intensively investigated see (Dalalyan and Tsybakov 2012; Chewi et al. 2022; Dalalyan 2017a,b; Durmus, Majewski, and Miasojedow 2019; Dalalyan, Karagulyan, and Riou-Durand 2022; Durmus and Moulines 2017, 2019; Erdogdu and Hosseinzadeh 2021; Erdogdu, Hosseinzadeh, and Zhang 2022; Mousavi-Hosseini et al. 2023; Vempala and Wibisono 2019). Efficiently scaling with dimension ensures that machine learning models remain computationally feasible and accurate as the complexity of the data increases. To the best of our knowledge, the overall complexity of generating one sample from µ with given precision ε scales linearly with dimension, see (Durmus, Majewski, and Miasojedow 2019). More precisely, it is proven therein that in the case of a convex potential V an error in the Kullback Leibler divergence scales linearly with dimension. The same dependence on dimension is shown therein for error expressed in the Wasserstein distance when V is strongly convex. The abovementioned studies and related research are typically conducted under the assumption that V is L-smooth, i.e., V is L-Lipschitz. A particular consequence of such an assumption is that a quadratic function majorizes V . This narrows the scope of possible applications excluding many classical potentials behaving like polynomials with higher power than 2, such as the one generating the processes of Ginzburg Landau, cf. (Brosse et al. 2019; Goldenfeld 1992). Henceforth, aiming at capturing the natural potentials we relax the L-smoothness condition trading it with λ-convexity outside a ball. We stress that it is necessary to assume that V is L- The Thirty-Ninth AAAI Conference on Artificial Intelligence (AAAI-25) Lipschitz while considering Euler Maruyama scheme, because otherwise the generated chain is transient (Roberts and Tweedie 1996; Mattingly, Stuart, and Higham 2002). Therefore, to tackle non-L-smooth potentials one is forced to modify the scheme. The natural idea how to adjust the scheme comes from a variational formulation of (1) by Jordan, Kinderlehrer, and Otto (1998); Ambrosio, Gigli, and Savar e (2008): µ = arg min µ P2(Rd) F[µ] , (2) where the functional F is defined for µ P(Rd) (probabilistic measures) via the following formula F[µ] := FV [µ] + FE[µ] , where (denoting dµ(x) = µ(x) dx when a density exists) FV [µ] := Z Rd V (x)µ(dx) , Rd µ(x) log µ(x) dx , if µ Pac , + , otherwise . Note that µ is a minimizer of F. Indeed, due to (Durmus, Majewski, and Miasojedow 2019, Lemma 1) we know that if µ P2(Rd), then FV [µ ] < + and FE[µ ] < + . Moreover, for µ P2(Rd) satisfying FE[µ] < + , it holds F[µ] F[µ ] = KL(µ|µ ) ,where KL stands for the Kullback Leibler divergence between measures (it is also called the relative entropy). Since KL(µ|µ ) 0 and the equality holds only if µ = µ we get that µ is a minimizer of F. By this observation the LMC algorithm can be seen as an inexact optimization of F, see also (Bernton 2018; Durmus, Majewski, and Miasojedow 2019). The typical approach in order to find a minimizer of the variational functional F is going along the gradient flow in the Wasserstein space. We split the gradient flow of F into two gradient flows (of FE and of FV ). By Ambrosio, Gigli, and Savar e (2008, Theorem 11.1.4) and Øksendal (2010, Theorem 2.1.5) the gradient flow of FE is the standard Brownian motion. To describe the gradient flow of FV we introduce the proximal operator defined for the function V and the time step τ by proxτ V (x) := arg min y Rd V (y) + 1 2τ |y x|2 . By Ambrosio, Gigli, and Savar e (2008, Proposition 10.4.2(ii)) the gradient flow of FV starting from µ with time step τ is given by (proxτ V )#µ. Therefore, the splitting algorithm for minimizing F is the Proximal Langevin Monte Carlo (PLMC) given by Xk+1 = proxτ V (Xk) + The definition of proxτ V (x) implies V (proxτ V (x)) V (x), which prevents PLMC from exploding. PLMC was already studied for machine learning problems by Bernton (2018); Salim, Korba, and Luise (2020) and in the context of numerical solutions to aggregation-diffusion equations by Benko et al. (2024). The main drawback of PLMC is that it is typically not feasible to perform the proximal step exactly, so this step should be approximated. However, FV and FE are contractions in the Wasserstein space and an error introduced by an inexact proximal step does not accumulate significantly. In addition, the function to be optimized in the proximal step is strongly convex and there are several accurate and fast computational methods to approximate proxτ V . In the current paper, we propose and give theoretical guarantees that the Inexact Proximal Langevin Algorithm (IPLA for short) can be used successfully to generate from µ a sample beyond the assumption of global Lipschitz gradient continuity of V . The typical assumption of the strong convexity of V is also relaxed. Our Contribution We make significant progress in the field of Langevin Monte Carlo methods by introducing a novel algorithm handling essentially broader scope of problems, than treated up to now, and preserving controlled computational cost. Namely, we propose the Inexact Proximal Langevin Algorithm IPLA (Algorithm 1). We extend the current scope of LMC algorithms by effective handling a large and natural class of potentials beyond the typical assumption of L-smoothness. We assume that the potential V is convex, strongly convex in tails, and with polynomial growth with power q V +1 2. Under such hypothesis, when an error of the inexact proximal step is of order τ 2, then IPLA generates one sample from µ with accuracy ε in the KL-divergence with total complexity d 2 O(ε 2), see Theorem 5 and Corollary 6. Moreover, when V is strongly convex, we can allow an error of the inexact proximal step of order τ 3 2 . In this setting we obtain that the total number of iterations to get one sample with the error ε in the Wasserstein distance is of order d 2 O(ε 2), see Theorem 7 and Corollary 8. Let us note that a lower accuracy of approximation of the proximal step in the strongly convex regime is expected. Since the proximal step is a gradient flow of FE, which is contracting in the Wasserstein space, the introduced error accumulates linearly. However, for strongly convex V the gradient flow of FE is strongly contracting, and the error is progressively suppressed during the iterations. Our results provide better convergence rates than bounds under similar assumptions for the Tamed Unadjusted Langevin Algorithm, cf. Brosse et al. (2019). Moreover, we get the same complexity as the Unadjusted Barker Algorithm (Livingstone et al. 2024), but under weaker assumptions we do not assume one-sided Lipschitz conditions on V . We note that when q V = 1 we retrieve dependence on the dimension that is the best-known for LMC, see (Durmus, Majewski, and Miasojedow 2019). Let us recall that the known results for LMC cover only potentials with growth between linear and quadratic. Our contribution is complementary and embrace super-quadratic potentials. Henceforth, we complete the scope of an application of LMC to all (sufficiently smooth) convex potentials of power growth. Besides the convergence rates of IPLA, we show a bound on all moments of the Markov chain {Xk}k 0 generated by IPLA, see Theorem 1. Such bounds are not only crucial for analyzing IPLA, but are also of separate in- terest for analysis of Langevin Monte Carlo algorithms (Cattiaux, Guillin, and Malrieu 2008; Durmus and Moulines 2019; Livingstone et al. 2024). Description of Our Setting We consider an optimization problem on the space of probability measures over Rd for which we use the notation P(Rd). We define Pac(Rd) the subspace of P(Rd) consisting of measures absolutely continuous with respect to the Lebesgue measure. By Pm(Rd), for m 1, we denote the subspace of P(Rd) consisting of measures with finite m-th moment, i.e., Pm(Rd) := {µ P(Rd): Z Rd |x|mµ(dx) < } = {µ P(Rd): µ(| |m) < } . We denote by C1(Rd, R) the space of functions from Rd to R with continuous derivatives. A continuously differentiable function f : Rd R is called λ-convex (strong convex) for λ R on Rd, if for all x, y Rd it satisfies f(x) f(y) + f(y) (x y) + λ λ-convex for λ R outside of a ball B Rd, if for all x, y Rd it satisfies f(x) f(y) + f(y) (x y) + λ 2 1Rd\B(y)|x y|2 . If f is λ-convex, the function f λ 2 | |2 is 0-convex, so we immediately get that f(x) f(y) (x y) λ|x y|2 . We denote by Zk the standard Gaussian distribution on Rd. For the sake of clarity of the exposition, throughout the paper we denote by Zk the rescaled one, namely Zk N(0, 2τ Id), where τ will be clear from the context. We will analyze IPLA in terms of two distances between measures. The first of them is the Kullback-Leibler divergence, also known as relative entropy, given for any measure µ, ν P(Rd) by Rd dµ dν (x) log dµ dν (x) ν(dx) , if µ ν , + , otherwise . Let us note that due to the Jensen inequality for all measures, µ, ν P(Rd), KL(µ|ν) 0 and equality holds only if µ = ν. The second distance used in our analysis is the 2Wasserstein distance defined for all µ, ν P2(Rd) by W 2 2 (µ, ν) := inf γ Π(µ,ν) Rd |x y|2γ(dx, dy) , (3) where Π(µ, ν) denotes the coupling between µ and ν, that is, Π(µ, ν) is a set of probability measures γ P2(Rd Rd) such that for every measurable set A it holds γ(A, Rd) = µ(A) and γ(Rd, A) = ν(A). We observe that by Villani (2009, Theorem 4.1) there exists an optimal coupling γ , for which the infimum in (3) is obtained. By the pushforward of the function f : Rd Rd, we mean the measure f# defined as f#ν(U) := ν(f 1(U)) for any measure ν P(Rd) and any measurable set U Rd. Let us present our regime. (V) For the potential V we assume what follows: V C1(Rd, R) and is convex on Rd. V is λV -convex for λV > 0 outside a given ball BV (0, RV ) with the center in 0 and a radius RV 0. The minimum of V is attained at x = 0. For q V 1, CV > 0, and for all x, y Rd it holds V (y) V (x) + V (x) (y x) +CV 1 + |x|q V 1 + |y|q V 1 |y x|2 . (4) (ϱ0) The initial distribution ϱ0 Pm0(Rd) has a finite moment of order m0 = q V + 1. Let us comment on the assumptions (V) and (ϱ0): (i) Assumption (V) embraces convex potentials with tails of polynomial growth with power q V + 1 2. The common assumption in the literature for LMC allows for convex functions with growth bounded by square function. (ii) The assumption that x = 0 is imposed just to simplify the presentation. All results remain true when x = 0 is any vector. There is no need for x to be the unique minimizer of V . (iii) In assumption (V) condition (4) is satisfied for q V = 1 if V is L-Lipschitz with CV = L 2 . Taking q V > 1 allows for treating way broader class of potentials V . (iv) If V is locally Lipschitz with q V -power growth, i.e., if there exists Lq > 0 such that for all x, y Rd it holds | V (x) V (y)| Lq min{|x y|, 1}(1 + |x|q V 1 + |y|q V 1) , then (4) is satisfied for CV = 22q V 2Lq, cf. (Benko et al. 2024, Lemma A.2). (v) In assumption (ϱ0) we require finite (q V + 1)-th moment of the initial distribution. This is a very weak assumption, as typically the finiteness of higher moments is used to prove the existence and uniqueness of solutions to Langevin equation (1), see Cattiaux, Guillin, and Malrieu (2008). Moreover, such an assumption is not restrictive in the application, as the initial measure is chosen arbitrarily in practice. Inexact Proximal Langevin Algorithm Let us present the idea of our algorithm. By (2) we know that µ is a minimizer of F over a space of measures, so we design an algorithm optimizing this functional. The basic idea to reach the minimizer is the analog of the gradient descent algorithm. In our case the functional F lives on Algorithm 1: Inexact Proximal Langevin Algorithm (IPLA) Initialize: Sample initial distribution X0 ϱ0 for k = 0, . . . , n 1 do Step 1: Run routine for computing proxτ V (Xk) with an output Xk+ 2 3 = proxτ V (Xk) + Θk+ 2 Step 2: Add Gaussian noise gτ, i.e Xk+1 = Xk+ 2 3 + Zk+1, Zk+1 N(0, 2τId) . the Wasserstein space, so instead of the classical gradient we need to employ the gradient flow of the functional. For definition and basic properties of the gradient flows on Wasserstein spaces we refer to Ambrosio, Gigli, and Savar e (2008); Santambrogio (2017). We note that the gradient flow of F is not given by explicit formula, but we know that F = FV + FE where the gradient flows of both FV and FE are well-understood. As mentioned in the introduction, the gradient flow of FV is given by a pushforward of the proximal operator, i.e., (proxτ V )#. The gradient flow of FE is the standard Brownian motion. Therefore, we use the splitting algorithm of the form Xk+1 = proxτk V (Xk) + which can be called Proximal Langevin Algorithm. However, in practice, computing the exact proximal operator is challenging. Therefore, we perform its more feasible numerical approximation. Since under our assumptions, both flows are contracting the introduced error will not accumulate too much and can be controlled. The IPLA (Algorithm 1) is defined as follows: in each step we perform two half steps. Given the previous state Xk, first we move according to the inexact proximal step of a function V with stepsize τ, where the error Θk+ 2 3 (which could be stochastic or deterministic) is bounded by δ. Next we add an independent of history and centered Gaussian random variable with covariance 2τId. For the theoretical analysis, we will split the inexact proximal step into two substeps: exact proximal step and additive error. Further, we will use the following notation. (i) Xk+ 1 3 := proxτ V (Xk), (iii) Xk+1 := Xk+ 2 3 + Zk+1; Zk+1 gτ. In view of densities ϱk s.t. Xk ϱk it reads as follows (i) ϱk+ 1 3 := (proxτ V )#ϱk, 3 ξδ, (iii) ϱk+1 := ϱk+ 2 where ξδ is a probability measure supported on B(0, δ) and gτ denotes the density of Gaussian distribution given by Zk N(0, 2τ Id) . We will also assume that for some κ > 0 and α 0 the error bound δ and stepsize τ satisfy δ = κτ 1+α. Theoretical Results In this section, we provide theoretical guarantees for the accuracy of IPLA. We start with bounds for moments of the Markov chain {Xk}k generated by IPLA, see Theorem 1. Then we show error bounds of IPLA in KLdivergence and in Wasserstein distance, see Theorems 5 and 7, respectively. Now we show that provided that for m 0 the mth moment of initial measure ϱ0 is bounded, it holds supk E|Xk|m < . Such bounds are required since we allow super quadratic growth of V . Theorem 1 (Moment bound). Let m 0. Suppose that V satisfies (V) and ϱ0 satisfies (ϱ0), and, for k = 1, . . . , n, let Xk be as in Algorithm 1. Then there exists a constant Cm > 0 such that for all 0 < τ < 1/λV , we have supk E|Xk|m Cm min{1, λ m V }d m If the error and the stepsize satisfy δ = κτ 1+α for some κ > 0 and α 0, then Cm = Cm(m, ϱ0, κ, λV ). Moreover, Cm + when λV 0. The proof of the above theorem and the exact value of Cm are given in Appendix B. Let us observe that since V has super-quadratic growth in tails, µ is sub-Gaussian distribution with all moments finite. Remark 2. If the minimum of V is not at x = 0, then the result would have been supk E|Xk x |m Cm min{1, λ m V }d m In order to prove the main result on the error bound of IPLA in KL-divergence, we mimic the classical reasoning from the convex optimization based on the following observation. Let us consider {xk}k 0 as an output of the firstorder algorithm for minimizing convex function f over Rd. It is possible to obtain an inequality of the following form 2τ(f(xk+1) f(x )) |xk x |2 |xk+1 x |2 + Cτ 2 where x is minimizer of f and C > 0 is some constant, see for example Beck (2017). Such an inequality is a key step in the proofs of convergence of certain algorithms. We provide the counterpart of such a result in the form relevant to analyze IPLA on the Wasserstein space. It employs the quantity K(τ) := 24q V 3Lq V (1 + Cq V 1d Remark 3. There exists a constant Cq V < such that for τ 1 it holds K(τ) Cq V τd 2 . The mentioned auxiliary result reads as follows. Proposition 4. Assume that the function V : Rd R, d 1, satisfies assumption (V) with some q V 1, the initial measure ϱ0 satisfies (ϱ0), and ν Pq V +1(Rd) is arbitrary. Then there exist C(ν) > 0, such that 2τ(F[ϱk+1] F[ν]) W 2 2 (ϱk, ν) W 2 2 (ϱk+1, ν) + C(ν)δ + K(τ)τ , where C(ν) is given in Appendix B, while K in (5). The proof is postponed to Appendix B. We state our main results for a sequence of probability measures {νN n }n N, called average measures, defined for every n, N N, n 1 by νN n := 1 n PN+n k=N+1 ϱk , where N is a burn-in time. By the use of Proposition 4 we infer the following main theorem on convergence of IPLA. Theorem 5 (KL-error bound). Suppose that V satisfies (V) and ϱ0 satisfies (ϱ0). Let τ < 1/λV , κ, α > 0, and δ κτ 1+α. Then it holds: KL(νN n |µ ) 1 2nτ W 2 2 (ϱN, µ ) 1 2nτ W 2 2 (ϱN+n, µ ) + C(µ )κτ α + K(τ) , where K is defined by (5) and C(µ ) is the same as in Proposition 4 evaluated in ν = µ . Proof. We use the convexity of KL-divergence (Cover and Thomas 2012, Theorem 2.7.2.) and get KL(νN n |µ ) 1 k=N+1 KL(ϱk|µ ) . Applying Proposition 4 we are able to write KL(νN n |µ ) 1 2nτ W 2 2 (ϱk, ν) W 2 2 (ϱk+1, ν) + C(µ )κτ α + K(τ) . A direct consequence of Theorem 5 is the following result on the overall complexity of generating one sample from µ with accuracy ε. Corollary 6. Suppose that the assumptions of Theorem 5 are satisfied. Assume further that 0 < τε min ε 3C(µ )κ and K(τε) ε/3 . Let the number of iterations nε be such that nε 3W 2 2 (ϱ0, µ )/(2ετε) . Then KL(ν0 nε|µ ) ε . (6) Moreover, for computing one sample in terms of KL with precision ε, in a case of warm start (W 2 2 (ϱ0, µ ) C for some absolute constant C), we need 2 O(ε 2) iterations, if α 1; (ii) d 2 O(ε 1 α 1) iterations, if α < 1. Proof. The inequality (6) is a straightforward consequence of Theorem 5. To get estimates of computational complexity it is enough to observe that by Remark 3 we have that condition K(τε) ε is satisfied if 0 < τε min n εC 1 q V d q V +1 Finally, let us show that, when the potential V is globally λV -convex, we can refine our result by improving the convergence rates of IPLA. Based on results from Benko et al. (2024) we have the following estimate on precision of IPLA in terms of Wasserstein distance. Theorem 7 (Wasserstein error bound). Suppose that the potential V satisfies (V) for RV = 0 and λV > 0, and that the initial measure ϱ0 satisfies (ϱ0). Let τ < 1/λV , α 0, and δ κτ 1+α. Then for all k N it holds W 2 2 (ϱk, µ ) 2 1 τλV 2 k W 2 2 (ϱ0, µ ) + 4 λV K(τ) + 2κ2τ 2+2α 1 e λV τ(k 1) This leads to the following complexity bounds. Corollary 8. Suppose that the assumptions of Theorem 7 are satisfied. Assume further that τ 2α ε λ2 V ε 96κ2 log2(6W 2 2 (ϱ0, µ )ε 1) and K(τε) 1 12λV ε . Let the number of iterations nε be such that 2 log(6W 2 2 (ϱ0, µ )ε 1)τ 1 ε λ 1 V nε 4 log(6W 2 2 (ϱ0, µ )ε 1)τ 1 ε λ 1 V . Then W 2 2 (ϱnε, µ ) ε . Moreover, for computing one sample in terms of the Wasserstein distance with precision ε, in the case of warm start, up to logarithmic terms we need 2 O(ε 2) iterations, if α 1 2 O(ε α 1) iterations, if α < 1 Proof. Let us note that 1 τλV 2 nε exp( nε τλV 2 ) and 1 e λV τ(nε 1) 1 e λV τ nε . With the above bounds, the result follows directly from Theorem 7. Note that if in Corollary 8 the stepsize τε = τ is fixed, then W 2 2 (ϱn, µ ) ε for all sufficiently large n. Remark 9. The assumption of global λV -convexity of V decreases the computational cost of a single iteration of IPLA. We can approximate the proximal step with smaller precision and keep the complexity of the whole algorithm of the order ε 2. Indeed, as a consequence of λV -convexity of the functional FE we get (i) in Theorem 5 under assumption (V) with RV > 0 to achieve complexity of order ε 2 we need δ κτ 2; (ii) in Theorem 7 in the strongly convex case the order ε 2 we have for δ κτ 3/2. Remark 10. The complexity of IPLA, as shown in Corollaries 6 and 8, depends on the number of iterations requiring additional computations. Since the optimized function is strongly convex, the cost of one iteration with precision δ is O(d log(δ)), as detailed in Appendix E. Experiments In this section, we demonstrate the application of IPLA on 3 examples implemented in Python. We analyze the convergence rates and bias of the algorithm compared to the two known related LMC algorithms, namely TULA (Tamed Unadjusted Langevin Algorithm by Brosse et al. (2019)) and ULA (Unadjusted Langevin Algorithm by Durmus and Moulines (2019)). The codes of our simulations can be found at https://github.com/192459/lmc-beyond- -lipschitz-gradient-continuity. See Appendix D for additional results. We shall compare the relative error (RE) and the coefficient of variance (CV) of various methods. In this context, the RE of the estimator bθ of the true value θ is bθ θ CV is sd(bθ) θ , where sd stands for the standard deviation. Example 1: Distribution With Light Tails Let us start with a simple and natural case where the potential has non-Lipschitz gradient. Our goal is to sample from the density µ (x) exp |x|4 4 , which is a stationary distribution of the process d Yt = Y 3 t dt + 2 d Bt . We estimate the moments E|Y |2, E|Y |4, and E|Y |6 in dimension d = 103. For error analysis as true values of the estimated quantities, we use the Metropolis Hastings algorithm with 107 iterations. Note that V is not globally Lipschitz and V is 1-convex outside a ball, so the assumptions on the convergence of ULA are not satisfied, but the convergence of TULA and IPLA is ensured. We have run 105 samples with burn-in time 104. We considered two scenarios. An initial value in the first of them is x0 = 7 1d (start in tail), while in the second one considered we start at x0 = 0 (start in the minimizer of V ). Each experiment has been repeated 100 times. In Table 1 we compare the relative error and the coefficient of variance between the algorithms. We can see that IPLA has the smallest RE and its CV is comparable to ULA s. In Figure 1 we compare the trajectories of algorithms for the selected coordinate (i = 1). We see that ULA blows up in a few iterations, which results from the fact that V is not globally Lipschitz. To suppress this phenomenon, TULA adjusts stepsizes according to normalized V , namely to V 1+| V |. This, on the other hand, results in unnecessarily small stepsizes in the tails. This can be seen in Figure 1. As shown on Figure 2 IPLA is less sensitive to the choice of stepsizes τ than TULA. Example 2: Ginzburg Landau Model We pass to a more complicated, but physically relevant example. Ginzburg Landau model was introduced to describe phase transitions in physics, i.e., superconductivity, see Goldenfeld (1992, Chapter 5). It involves the potential 2 x2 ijk + υκ 2 |e xijk|2 + υς Mom. Method Start in tail Start in x0 = 0 RE CV RE CV E|Y |2 IPLA 0.0027 0.0019 0.0006 0.0018 TULA 0.0047 0.0016 0.0030 0.0019 ULA Na N Na N 0.0020 0.0018 E|Y |4 IPLA 0.0054 0.0039 0.0025 0.0036 TULA 0.0095 0.0032 0.0073 0.0039 ULA Na N Na N 0.0028 0.0035 E|Y |6 IPLA 0.0081 0.0058 0.0047 0.0054 TULA 0.0144 0.0047 0.0120 0.0058 ULA Na N Na N 0.0032 0.0053 Table 1: Estimation of the moments of light tails distribution from Example 1. 0 100 200 300 k IPLA TULA ULA 0 100 200 300 k Figure 1: Trajectory of the 1st coordinate from Example 1 starting in a tail. Both plots are based on the same data. 0.000 0.025 0.050 0.075 0.100 τ 0.000 0.002 0.004 τ Figure 2: Dependence RE of IPLA and TULA (for Example 1) on a stepsize τ starting in a tail. where e xijk = xi+jk xijk, xij+k xijk, xijk+ xijk , with the notation such that i := i 1 mod q and similarly for j, k. We consider κ = 0.1, ς = 0.5, υ = 2 and q = 5. The dimension in this example is d = q3 = 125. Note that V is not globally Lipschitz and V is λV -convex for λV > 0, (Brosse et al. 2019). We have run 2 104 samples with burn-in time 104. We consider two scenarios as in Example 1. As starting in tail, we consider the case x0 = (100, 0, . . . , 0) and starting in the minimizer we take x0 = 0. Each experiment has been repeated 100 times. From Table 2 we see that the results are analogous to those of Example 1. In the scenario starting in the minimizer, we have results similar to ULA s and better than TULA s. In the case of the start in tail, ULA blows up. (a) Original image (b) Blurred with additive noise (c) Processed (β = 0.03) (d) Processed (β = 0.0001) Figure 3: Result of the Bayesian Image Denoising from Example 3. The original photo by Zbyszko Siemaszko 1955-56. Mom. Method Start in tail Start in x0 = 0 RE CV RE CV E|Y |2 IPLA 0.0025 0.0244 0.0748 0.0786 TULA 0.0067 0.0213 0.0859 0.0739 ULA Na N Na N 0.0727 0.0697 E|Y |4 IPLA 0.0053 0.0491 0.1425 0.1558 TULA 0.0134 0.0425 0.1635 0.1473 ULA Na N Na N 0.1385 0.1399 E|Y |6 IPLA 0.0083 0.0742 0.2040 0.2323 TULA 0.0199 0.0638 0.2337 0.2208 ULA Na N Na N 0.1980 0.2117 Table 2: Estimation of moments of Ginzburg Landau model from Example 2. Example 3: Bayesian Image Deconvolution We show now an example that IPLA is effective in complex high-dimensional scenarios. We consider the Bayesian Image Deconvolution problem inspired by Experiment 1 in Durmus, Moulines, and Pereyra (2018, Subsection 4.1.2). Given a high-resolution grayscale image x Rd such that d = n n = 600 600 pixels from Figure 3 (a), the goal of the problem is to recover image x from a blurred image. Namely, we are to sample from posterior density p(x|y) of blurred image with additive noise (observation) such that y = Hx + W for W N(0, σ2Id) , where H is a circulant blur matrix representing 2D discrete convolution and W is an additive noise. We need to define 2D discrete isotropic total variation via |xin,j+1 xin,j|2 + |x(i+1)n,j xin,j|2 j=1 |xnd,j+1 xnd,j| + i=1 |x(i+1)n,n xin,n| , which was proposed in the context of image processing by Rudin, Osher, and Fatemi (1992). Note that TV is a discretization of the standard total variation norm, cf. Chambolle (2004, Section 2). We choose prior distribution p(x) exp( β TV (x)), where β > 0 is a regularizing factor. Henceforth, we provide sampling from the posterior distribution µ := p(x|y) satisfying µ exp 1 2σ2 |y Hx|2 β TV (x) . Even if TV is not smooth, and consequently the theoretical convergence rates are not known for IPLA, TULA, or ULA, we show the performance of our algorithm. It might be expected that the inexact proximal step could be non-trivial in this case. Nonetheless, it can be effectively solved with the use of the Adaptive Primal Dual Hybrid Algorithm by Goldstein, Li, and Yuan (2015, Algorithm 1). Our results are presented on Figure 3. On (b) we show blurred and noisy image y, where uniform circular blur matrix has depth 9 and additive noise is of the standard deviation σ = 0.5. On (c) and (d) we present posterior mean obtained by IPLA for 105 iterations and with the desired precision of the inexact proximal step δ = 10 1. Effective Implementation of IPLA Despite involving the inexact proximal step, the computational complexity of IPLA is controlled. In Examples 1 and 2, we used the Newton conjugate gradient method with the provided analytically computed gradient and the Hessian matrix, making use of the sparsity of the Hessian matrix. Using the standard Python library Sci Py (Virtanen et al. 2020), IPLA has comparable performance to ULA. We write the time of computations on standard Macbook Air M1 2020. In Example 1 a single run of 106 iterations took from 5 to 20 s for ULA or TULA and from 15 to 60 s, while in Example 2, all algorithms took from 3 to 5 min to run 104 iterations. In Example 3, the potential is not differentiable and we modify the existing implementation in the Py Proximal library (Ravasi et al. 2024). Setting the initial value in the minimizer of V reduces the cost of the inexact proximal step. Let us stress that in our simulation of Example 3, 105 samples in the dimension d = 360 000 are obtained in less than 1 min. Conclusion IPLA broadens the scope of problems that known LMC algorithms can address keeping low computational cost. Acknowledgements I.C. is supported by NCN grant 2019/34/E/ST1/00120, B.M. is supported by NCN grant 2018/31/B/ST1/00253. The authors are grateful to Adam Chlebicki-Miasojedow and Sonja Letnes for close assistance in the progress of this research. References Ambrosio, L.; Gigli, N.; and Savar e, G. 2008. Gradient flows in metric spaces and in the space of probability measures. Lectures in Mathematics ETH Z urich. Birkh auser Verlag, Basel, second edition. ISBN 978-3-7643-8721-1. Andrieu, C.; de Freitas, N.; Doucet, A.; and Jordan, M. I. 2003. An Introduction to MCMC for Machine Learning. Machine Learning, 50(1): 5 43. Beck, A. 2017. First-order methods in optimization, volume 25 of MOS-SIAM Series on Optimization. Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA; Mathematical Optimization Society, Philadelphia, PA. ISBN 978-1-611974-98-0. Benko, M.; Chlebicka, I.; Endal, J.; and Miasojedow, B. 2024. Convergence rates of particle approximation of forward-backward splitting algorithm for granular medium equations. ar Xiv:2405.18034. Bernton, E. 2018. Langevin Monte Carlo and JKO splitting. In Bubeck, S.; Perchet, V.; and Rigollet, P., eds., Proceedings of the 31st Conference On Learning Theory, volume 75 of Proceedings of Machine Learning Research, 1777 1798. PMLR. Brosse, N.; Durmus, A.; Moulines, E.; and Sabanis, S. 2019. The tamed unadjusted Langevin algorithm. Stochastic Process. Appl., 129(10): 3638 3663. Cattiaux, P.; Guillin, A.; and Malrieu, F. 2008. Probabilistic approach for granular media equations in the non-uniformly convex case. Probab. Theory Related Fields, 140(1-2): 19 40. Chambolle, A. 2004. An Algorithm for Total Variation Minimization and Applications. Journal of Mathematical Imaging and Vision, 20(1): 89 97. Chewi, S.; Erdogdu, M. A.; Li, M.; Shen, R.; and Zhang, S. 2022. Analysis of Langevin Monte Carlo from Poincare to Log-Sobolev. In Loh, P.-L.; and Raginsky, M., eds., Proceedings of Thirty Fifth Conference on Learning Theory, volume 178 of Proceedings of Machine Learning Research, 1 2. PMLR. Cover, T.; and Thomas, J. 2012. Elements of Information Theory. Wiley. ISBN 9781118585771. Dalalyan, A. 2017a. Further and stronger analogy between sampling and optimization: Langevin Monte Carlo and gradient descent. In Kale, S.; and Shamir, O., eds., Proceedings of the 2017 Conference on Learning Theory, volume 65 of Proceedings of Machine Learning Research, 678 689. PMLR. Dalalyan, A. S. 2017b. Theoretical guarantees for approximate sampling from smooth and log-concave densities. J. R. Stat. Soc. Ser. B. Stat. Methodol., 79(3): 651 676. Dalalyan, A. S.; Karagulyan, A.; and Riou-Durand, L. 2022. Bounding the error of discretized Langevin algorithms for non-strongly log-concave targets. J. Mach. Learn. Res., 23: Paper No. [235], 38. Dalalyan, A. S.; and Tsybakov, A. B. 2012. Sparse regression learning by aggregation and Langevin Monte-Carlo. J. Comput. System Sci., 78(5): 1423 1443. Durmus, A.; Majewski, S.; and Miasojedow, B. 2019. Analysis of Langevin Monte Carlo via convex optimization. J. Mach. Learn. Res., 20: Paper No. 73, 46. Durmus, A.; and Moulines, E. 2017. Nonasymptotic convergence analysis for the unadjusted Langevin algorithm. Ann. Appl. Probab., 27(3): 1551 1587. Durmus, A.; and Moulines, E. 2019. High-dimensional Bayesian inference via the unadjusted Langevin algorithm. Bernoulli, 25(4A): 2854 2882. Durmus, A.; Moulines, E.; and Pereyra, M. 2018. Efficient Bayesian Computation by Proximal Markov Chain Monte Carlo: When Langevin Meets Moreau. SIAM Journal on Imaging Sciences, 11(1): 473 506. Erdogdu, M. A.; and Hosseinzadeh, R. 2021. On the Convergence of Langevin Monte Carlo: The Interplay between Tail Growth and Smoothness. In Belkin, M.; and Kpotufe, S., eds., Proceedings of Thirty Fourth Conference on Learning Theory, volume 134 of Proceedings of Machine Learning Research, 1776 1822. PMLR. Erdogdu, M. A.; Hosseinzadeh, R.; and Zhang, S. 2022. Convergence of Langevin Monte Carlo in Chi-Squared and R enyi Divergence. In Camps-Valls, G.; Ruiz, F. J. R.; and Valera, I., eds., Proceedings of The 25th International Conference on Artificial Intelligence and Statistics, volume 151 of Proceedings of Machine Learning Research, 8151 8175. PMLR. Gelman, A.; Carlin, J. B.; Stern, H. S.; Dunson, D. B.; Vehtari, A.; and Rubin, D. B. 2014. Bayesian data analysis. Texts in Statistical Science Series. CRC Press, Boca Raton, FL, third edition. ISBN 978-1-4398-4095-5. Goldenfeld, N. 1992. Lectures On Phase Transitions And The Renormalization Group (1st ed.). CRC Press. Goldstein, T.; Li, M.; and Yuan, X. 2015. Adaptive Primal Dual Splitting Methods for Statistical Learning and Image Processing. In Cortes, C.; Lawrence, N.; Lee, D.; Sugiyama, M.; and Garnett, R., eds., Advances in Neural Information Processing Systems, volume 28. Curran Associates, Inc. Jordan, R.; Kinderlehrer, D.; and Otto, F. 1998. The variational formulation of the Fokker-Planck equation. SIAM J. Math. Anal., 29(1): 1 17. Krauth, W. 2006. Statistical mechanics, volume 13 of Oxford Master Series in Physics. Oxford University Press, Oxford. ISBN 978-0-19-851536-4; 0-19-851536-7. Algorithms and computations, Oxford Master Series in Statistical Computational, and Theoretical Physics. Livingstone, S.; N usken, N.; Vasdekis, G.; and Zhang, R.-Y. 2024. Skew-symmetric schemes for stochastic differential equations with non-Lipschitz drift: an unadjusted Barker algorithm. ar Xiv:2405.14373. Mattingly, J. C.; Stuart, A. M.; and Higham, D. J. 2002. Ergodicity for SDEs and approximations: locally Lipschitz vector fields and degenerate noise. Stochastic Process. Appl., 101(2): 185 232. Mousavi-Hosseini, A.; Farghly, T. K.; He, Y.; Balasubramanian, K.; and Erdogdu, M. A. 2023. Towards a Complete Analysis of Langevin Monte Carlo: Beyond Poincar e Inequality. In Neu, G.; and Rosasco, L., eds., Proceedings of Thirty Sixth Conference on Learning Theory, volume 195 of Proceedings of Machine Learning Research, 1 35. PMLR. Øksendal, B. 2010. Stochastic Differential Equations: An Introduction with Applications. Universitext. Springer Berlin Heidelberg. ISBN 9783642143946. Ravasi, M.; Ornhag, M. V.; Luiken, N.; Leblanc, O.; and Uru nuela, E. 2024. Py Proximal - scalable convex optimization in Python. Journal of Open Source Software, 9(95): 6326. Robert, C. P.; and Casella, G. 2004. Monte Carlo statistical methods. Springer Texts in Statistics. Springer-Verlag, New York, second edition. ISBN 0-387-21239-6. Roberts, G. O.; and Tweedie, R. L. 1996. Exponential convergence of Langevin distributions and their discrete approximations. Bernoulli, 2(4): 341 363. Rudin, L. I.; Osher, S.; and Fatemi, E. 1992. Nonlinear total variation based noise removal algorithms. Physica D: Nonlinear Phenomena, 60(1): 259 268. Salim, A.; Korba, A.; and Luise, G. 2020. The Wasserstein Proximal Gradient Algorithm. In Larochelle, H.; Ranzato, M.; Hadsell, R.; Balcan, M.; and Lin, H., eds., Advances in Neural Information Processing Systems, volume 33, 12356 12366. Curran Associates, Inc. Santambrogio, F. 2017. Euclidean, metric, and Wasserstein gradient flows: an overview. Bull. Math. Sci., 7(1): 87 154. Stuart, A. M. 2010. Inverse problems: a Bayesian perspective. Acta Numer., 19: 451 559. Vempala, S.; and Wibisono, A. 2019. Rapid Convergence of the Unadjusted Langevin Algorithm: Isoperimetry Suffices. In Wallach, H.; Larochelle, H.; Beygelzimer, A.; d'Alch e Buc, F.; Fox, E.; and Garnett, R., eds., Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc. Villani, C. 2009. Optimal transport. Old and new, volume 338 of Grundlehren der mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Springer-Verlag, Berlin. ISBN 978-3-540-71049-3. Virtanen, P.; Gommers, R.; Oliphant, T. E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; van der Walt, S. J.; Brett, M.; Wilson, J.; Millman, K. J.; Mayorov, N.; Nelson, A. R. J.; Jones, E.; Kern, R.; Larson, E.; Carey, C. J.; Polat, I.; Feng, Y.; Moore, E. W.; Vander Plas, J.; Laxalde, D.; Perktold, J.; Cimrman, R.; Henriksen, I.; Quintero, E. A.; Harris, C. R.; Archibald, A. M.; Ribeiro, A. H.; Pedregosa, F.; van Mulbregt, P.; and Sci Py 1.0 Contributors. 2020. Sci Py 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 17: 261 272.