# the_star_geometry_of_criticbased_regularizer_learning__76da9a02.pdf The Star Geometry of Critic-Based Regularizer Learning Oscar Leong Department of Statistics and Data Science University of California, Los Angeles oleong@stat.ucla.edu Eliza O Reilly Department of Applied Mathematics and Statistics Johns Hopkins University eoreill2@jh.edu Yong Sheng Soh Department of Mathematics National University of Singapore matsys@nus.edu.sg Variational regularization is a classical technique to solve statistical inference tasks and inverse problems, with modern data-driven approaches parameterizing regularizers via deep neural networks showcasing impressive empirical performance. Recent works along these lines learn task-dependent regularizers. This is done by integrating information about the measurements and ground-truth data in an unsupervised, critic-based loss function, where the regularizer attributes low values to likely data and high values to unlikely data. However, there is little theory about the structure of regularizers learned via this process and how it relates to the two data distributions. To make progress on this challenge, we initiate a study of optimizing critic-based loss functions to learn regularizers over a particular family of regularizers: gauges (or Minkowski functionals) of star-shaped bodies. This family contains regularizers that are commonly employed in practice and shares properties with regularizers parameterized by deep neural networks. We specifically investigate critic-based losses derived from variational representations of statistical distances between probability measures. By leveraging tools from star geometry and dual Brunn-Minkowski theory, we illustrate how these losses can be interpreted as dual mixed volumes that depend on the data distribution. This allows us to derive exact expressions for the optimal regularizer in certain cases. Finally, we identify which neural network architectures give rise to such star body gauges and when such regularizers have favorable properties for optimization. More broadly, this work highlights how the tools of star geometry can aid in understanding the geometry of unsupervised regularizer learning. 1 Introduction The choice and design of regularization functionals to promote structure in data has a long history in statistical inference and inverse problems, with roots dating back to classical Tikhonov regularization [89]. In such problems, one is tasked with solving the following: given measurements y Rm of the form y = A(x0) + η for some forward model A : Rd Rm and noise η Rm, recover an estimate of the ground-truth signal x0 Rd. Typically, the challenge in such problems is that they are ill-posed, meaning that either there are no solutions, infinitely many solutions, or the problem is discontinuous in the data y [7]. A pervasive and now classical technique is to identify regularization functionals R : Rd R such that, when minimized, promote structure that is present in the ground- 38th Conference on Neural Information Processing Systems (Neur IPS 2024). truth signal. Well-known examples of hand-crafted regularizers to promote structure include the ℓ1-norm to promote sparsity [23, 16, 26, 88], total variation [74], the nuclear norm [29, 15, 70], and more generally, atomic norms [18, 12, 87, 78, 64]. With the development of modern machine learning techniques, we have seen a surge of data-driven methods to directly learn regularizers instead of designing regularizers in a hand-crafted fashion. Early works along these lines include the now-mature field of dictionary learning or sparse coding (see [60, 61, 84, 3, 6, 1, 2, 75, 76, 10, 81] and the surveys [56, 27] for more). More recently, we have seen powerful nonconvex regularizers parameterized by deep neural networks. These come in a variety of flavors, including those based on plug-and-play [91, 71], generative models [13, 39, 40, 8, 20], and regularization functionals directly parameterized by deep neural networks [53, 59, 80, 36, 47]. Despite the widespread success of such learning-based regularizers, several outstanding questions remain. In particular, it is unclear what type of structure is learned by such regularizers. For example, what is the relationship between the underlying data geometry and the type of regularizer found by data-driven approaches? Is the found regularizer optimal in some meaningful sense? If not, what properties of the data distribution are lost due to structural constraints placed on the regularizer? In this work, we aim to tackle the above questions and others to further understand what types of regularizers are found via learning-based methods. We focus on a class of unsupervised methods that are task-dependent, i.e., those that learn a regularizer without paired training data, but still integrate information about the measurements in the learning process. This is done by identifying a loss inspired by variational representations of statistical distances, such as Integral Probability Metrics (IPMs) or divergences, where the regularizer plays the role of the critic or test function. As an example, the recent adversarial regularization framework used in [53, 59, 80] learns a regularizer that assigns high likelihood to clean data Dr and low likelihood to noisy data Dn via a loss derived from a dual formulation of the 1-Wasserstein distance. This framework has showcased impressive empirical performance in learning data-driven regularizers, but there is still a lack of an overarching understanding of the structure of regularizers learned. Moreover, while this Wasserstein distance interpretation has shown to be useful, it is natural to consider if other losses can be derived from this critic-based perspective to learn regularizers. 1.1 Our contributions In order to make progress on these challenges, we first fix a family of regularization functionals to analyze. We aim for this family to be (i) expressive (can describe both convex and nonconvex regularizers), (ii) exhibit properties akin to regularizers used in practice, and (iii) tractable to analyze. A family of functionals that satisfy such criteria are gauges of star bodies. In particular, we will consider regularization functionals of the form x K := inf{t > 0 : x t K} where K is a star body, i.e., a compact subset of Rd with 0 int(K) such that for each x Rd \{0}, the ray {tx : t > 0} intersects the boundary of K exactly once. Such regularizers are nonconvex for general star bodies K, but note that K is convex if and only if K is a convex body. Any norm is the gauge of a convex body, but this class also includes nonconvex quasinorms such as the ℓq-quasinorm for q (0, 1). Focusing on the above class of functionals, we aim to understand the structure of a regularizer K found by solving an optimization problem of the form inf K F H ( K; Dr, Dn) (1) where H( ; Dr, Dn) : F R compares the values K assigns to Dr and Dn and F is a class of functions on Rd. The adversarial regularization framework of [53] is recovered by setting H(f; Dr, Dn) = EDr[f(x)] EDn[f(x)] and F = Lip(1) := {f : Rd R : f is 1-Lipschitz}. Our contributions are as follows: 1. Using tools from star geometry [41, 33] and dual Brunn-Minkowski theory [54], we prove that under certain conditions, the solution to (1) under the adversarial regularization framework of [53] can be exactly characterized. In particular, we show that the objective is equivalent to a dual mixed volume between our star body K and a data-dependent star body Lr,n. This allows us to exploit known bounds on the dual mixed volume via (dual) Brunn-Minkowski theory. 2. We investigate new critic-based loss functions H( ; Dr, Dn) in (1) inspired by variational representations of divergences for probability measures. We specifically analyze α-divergences and show how they can give rise to loss functionals with dual mixed volume interpretations. We also experimentally show that such losses can be competitive for learning regularizers in a simple denoising setting. 3. We conclude with results showcasing when these star body regularizers exhibit useful optimization properties, such as weak convexity, as well as analyzing what types of neural network architectures give rise to star body gauges. This paper is organized as follows: Section 2 analyzes optimal adversarial regularizers using star geometry. We establish a general existence result in Section 2.1 and use dual mixed volumes to characterize the optimal star body regularizer under certain assumptions in Section 2.2. Visual examples are provided in Section 2.3. Section 3 introduces new critic-based losses for learning regularizers inspired by α-divergences. An empirical comparison between neural network-based regularizers learned using these losses and the adversarial loss is presented in Section 3.1. Section 4 examines computational properties of star body regularizers, such as beneficial properties for optimization and relevant neural network architectures. We conclude with a discussion in Section 5. 1.2 Related work We discuss the broader literature on learning-based regularization in Section A of the appendix and focus on optimal regularization here. To our knowledge, there are few works that analyze the optimality of a regularizer for a given dataset. The authors of the original adversarial regularization framework of [53] analyzed the case when Dr is supported on a manifold and Dn is related to Dr via the push-forward of the projection operator onto the manifold. They showed that in this case the distance function to the manifold is an optimal regularizer, but uniqueness does not hold (i.e., other regularizers could be optimal as well). Our work complements such results by analyzing the structure of the optimal regularizer for a variety of Dr and Dn which may not be related in this way. In [4] the authors analyze the optimal Tikhonov regularizer in the infinite-dimensional Hilbert space setting, and show that the optimal regularizer is independent of the forward model and only depends on the mean and covariance of the data distribution. The two most closely related papers to ours are [90] and [52]. In [90], the authors aim to characterize the optimal convex regularizer for linear inverse problems. Several notions of optimality for convex regularizers are introduced (dubbed compliance measures) and the authors establish that canonical low-dimensional models, such as the ℓ1-norm for sparsity, are optimal for sparse recovery under such compliance measures. In [52], similarly to this work, the authors analyze the optimal regularizer amongst the family of star body gauges for a given dataset. They also leverage dual Brunn-Minkowski theory to show that the optimal regularizer is induced by a data-dependent star body, whose radial function depends on the density of the data distribution. Interestingly, these results also characterize data distributions for which convex regularization is optimal and they provide several examples. The present work is novel in relation to [52] for several reasons. First, our work introduces novel theory and a new framework for understanding critic-based regularizer learning, addressing a significant gap in existing theoretical foundations for unsupervised regularizer learning. This new setting also brings novel technical challenges that were not present in [52]. For example, the losses in Section 3 exhibit more complicated dependencies on the star body of interest, leading to the need for new analysis. Such losses are also experimentally analyzed in Section C.1. Finally, we present new results that are relevant to downstream applications of star body regularizers in Section 4. 1.3 Notation In this section, we provide some brief background on star geometry and define notation used in the main body of the paper. For further details, please see Section B.1 in the appendix and the sources [77, 33, 41] for more. We say that a closed set K Rd is star-shaped (with respect to the origin) if for all x K, we have [0, x] K where, for two points x, y Rd, we define the line segment [x, y] := {(1 t)x + ty : t [0, 1]}. K is a star body if it is a compact star-shaped set such that for every x = 0, the ray Rx := {tx : t > 0} intersects the boundary of K exactly once. Equivalently, K is a star body if its radial function ρK is positive and continuous over the unit sphere Sd 1, where ρK is defined as ρK(x) := sup{t > 0 : t x K}. It follows that the gauge function of K satisfies x K = 1/ρK(x) for all x Rd such that x = 0. Let Sd to be the space of all star bodies on Rd. For K, L Sd, it is easy to see that K L if and only if ρK ρL. The kernel of a star body K is the set of all points for which K is star-shaped with respect to, i.e., ker(K) := {x K : [x, y] K, y K}. Note that ker(K) is a convex subset of K and ker(K) = K if and only if K is convex. For a parameter γ > 0, we define the following subset of Sd consisting of well-conditioned star bodies with nondegenerate kernels: Sd(γ) := {K Sd : γBd ker(K)} where Bd := {x Rd : x ℓ2 1}. For two distributions P and Q, let P Q denote their convolution, i.e., P Q is the distribution of X + Y where X P and Y Q. Finally, for a function f and a measure P, let f#(P) denote the push-forward measure of P under f, i.e., for all measurable subsets A, we have f#(P)(A) := P(f 1(A)). 2 Adversarial star body regularization To illustrate our tools and results, we first consider regularizer learning under the adversarial regularization framework of [53]. Given two distributions Dr and Dn on Rd and setting H(f; Dr, Dn) = EDr[f(x)] EDn[f(x)] and F := Lip(1) in (1), we aim to understand minimization of the functional F(K; Dr, Dn) := EDr[ x K] EDn[ x K] over all star bodies K such that x 7 x K is 1-Lipschitz. A result due to [73] shows that K is 1-Lipschitz if and only if the unit ball Bd ker(K). Here and throughout this paper, it will be useful to think of Dr as the distribution of ground-truth, clean data while Dn is a user-defined distribution describing noisy, undesired data. Using the notation from Section 1.3, we are interested in analyzing inf K Sd(1) F(K; Dr, Dn). (2) 2.1 Existence of minimizers As the problem (2) requires minimizing a functional over a structured subset of the infinitedimensional space of star bodies, even basic questions such as the existence of minimizers are unclear. Despite this, one can exploit tools in the star geometry literature to obtain guarantees regarding this problem. The proof of this result exploits Lipschitz continuity of the objective functional and local compactness properties of Sd(γ), proved in [52], akin to the celebrated Blaschke s Selection Theorem from convex geometry [77]. We defer the proof to the appendix in Section B.2. Theorem 2.1. For any two distributions Dr and Dn on Rd, we have that F(K; Dr, Dn) W1(Dr, Dn) for all K Sd(1) where W1( , ) is the 1-Wasserstein distance between two distributions. Moreover, if EDi[ x ℓ2] < for each i = r, n, then we always have that minimizers exist: arg min K Sd(1) F(K; Dr, Dn) = . 2.2 Minimization via dual Brunn-Minkowski theory We now aim to understand the structure of minimizers to the above problem. To do this, we first show that the objective in (2) can be interpreted as the dual mixed volume between K and a data-dependent star body. To begin, we start with a definition. Definition 2.2 (Definition 2* in [54]). Given two star bodies K, L Sd, the i-th dual mixed volume between L and K for i R is given by Vi(L, K) := 1 Sd 1 ρL(u)d iρK(u)idu. One can think of dual mixed volumes as functionals that measure the size of a star body K relative to another star body L. Note that for all i, Vi(K, K) = 1 Sd 1 ρK(u)ddu = vold(K) is the usual d-dimensional volume of K. Of particular interest to us will be the case i = 1. Figure 1: (Left) Contours of the Gaussian mixture model density p. (Right) The star body Kp induced by the radial function (3). To see how our main objective can be interpreted as a dual mixed volume, we need to be able to summarize our distributions Dr and Dn in such a way that can naturally be related to the gauge K. Since the gauge is positively homogenous, it is characterized by its behavior on the sphere Sd 1. Hence, given a distribution, we require a summary statistic that describes the distribution in each unit direction u Sd 1. Since star bodies are precisely defined by the distance of the origin to their boundary in each unit direction, we ideally would like to summarize the distribution in a similar fashion. For a distribution D with density p on Rd, define the map 0 tdp(tu)dt 1/(d+1) , u Sd 1 . (3) This function measures the average mass the distribution accrues in each unit direction and how far this mass lies from the origin. More precisely, one can show [52] that ρd+1 p is the density of the measure µD( ) := ED h x ℓ21{x/ x ℓ2 } i . If the map (3) is positive and continuous over the unit sphere, it defines a unique data-dependent star body. In Figure 1, we give an illustrative example showing how the geometry of the data distribution relates to the data-dependent star body. Here, the data distribution is given by a Gaussian mixture model with 3 Gaussians. We now aim to understand the structure of minimizers of the functional K 7 F(K; Dr, Dn). As we will show in our main result, the map (3) aids in interpreting the functional via a dual mixed volume. Extrema of such dual mixed volumes have been the object of study in convex and star geometry for some time. We cite a seminal result due to Lutwak. Theorem 2.3 (Special case of Theorem 2 in [54]). For star bodies K, L Sd, we have V 1(L, K)d vold(L)d+1 vold(K) 1, and equality holds if and only if K and L are dilates, i.e., there exists a λ > 0 such that K = λL. Armed with this inequality, we show that under certain conditions, we can guarantee the existence and uniqueness of minimizers of the above map. Note that optimal solutions are always defined up to scaling. To remove this degree of freedom, we impose an additional constraint on the volume of the solution. We chose a unit-volume constraint for simplicity, but changing the constraint would simply scale the optimal solution. Theorem 2.4. Suppose Dr and Dn are distributions on Rd that are absolutely continuous with respect to the Lebesgue measure with densities pr and pn, respectively. Suppose ρpr and ρpn as defined in (3) are continuous over the unit sphere and that ρpr(u) > ρpn(u) 0 for all u Sd 1. Then, there exists a star body Lr,n Sd such that the unique solution to the problem min K Sd:vold(K)=1 EDr[ x K] EDn[ x K] is given by K := vold(Lr,n) 1/d Lr,n. If Bd ker(K ), then K is, in fact, the unique solution to (2) with the additional constraint vold(K) = 1. Proof Sketch. One can prove that for each i = r, n, EDi[ x K] = R Sd 1 ρpi(u)d+1ρK(u)du. Then since ρpr > ρpn 0 on the unit sphere, we have that the map ρr,n(u) := (ρpr(u)d+1 ρpn(u)d+1)1/(d+1) is positive and continuous on Sd 1. We can then prove that it is the radial function of a new data-dependent star body Lr,n. Combining the previous two observations yields EDr[ x K] EDn[ x K] = d V 1(Lr,n, K). Applying Theorem 2.3 obtains the final result. We defer the full proof to Section B.2. Remark 2.5 (Uniqueness guarantees). We highlight that in our result we are able to obtain uniqueness of the solution everywhere. This is notably different than previous guarantees in the Optimal Transport literature, where the optimal transport potential for the 1-Wasserstein loss is not unique, but can be shown to be unique Dn-almost everywhere [58]. A significant reason for this is that our optimization problem is over the family of star body gauges, which have additional structure over the general class of 1-Lipschitz functions. In particular, star bodies are uniquely defined by their radial functions and, hence, dual mixed volumes can exactly specify them (see [54] for more details). Remark 2.6 (Distributional assumptions). Note that the conditions of this theorem require that the density-induced map ρp( ) must be a valid radial function (i.e., positive and continuous over the unit sphere). Many distributions satisfy these assumptions. For example, if the distribution s density is of the form p(x) = ψ( x q L) where R 0 tdψ(tq)dt < , q > 0, and L Sd, then the map ρp( ) is a valid radial function. See [52] for more examples of distributions that satisfy these assumptions. Remark 2.7 (Finite-data regime). While the conditions of this theorem would not apply to the case when Dr and Dn are empirical distributions, we prove in the appendix (Section D.1) that these results are stable in the sense that as the amount of available data goes to infinity, the solutions in the finite-data regime converge (up to a subsequence) to a population minimizer. The proof exploits tools from variational analysis, such as Γ-convergence [14]. Remark 2.8 (Implications for inverse problems). In the original framework of [53], the authors considered Dn = A #(Dy) where A is a linear forward operator in an inverse problem. Since inverse problems are typically under-constrained, such a distribution would be singular and not satisfy the assumptions of the theorem. One can solve this, however, by convolving Dn with a Gaussian to ensure it has full measure. Remark 2.9 (Scaling distributions). One can always reweight the objective so that the assumptions of the theorem are satisfied. This is due to positive homogeneity of the gauge, which guarantees ED[ λx K] = λ ED[ x K] for any distribution D. In particular, if ρpr > ρpn is not satisfied, one can choose a scaling λ so that ρpr > λ1/(d+1)ρpn. Inspecting the proof of Theorem 2.4, one can see that the result goes through when applied to the objective EDr[ x K] λ EDn[ x K]. 2.3 Examples We now provide examples of optimal regularizers via different choices of Dr and Dn. Further examples can be found in the appendix in Section D.3. Throughout these examples, the star body induced by Dr and Dn is denoted by Lr and Ln, respectively. Example 1 (Gibbs densities with ℓ1and ℓ2-norm energies). Suppose the distributions Dr and Dα n for some α 1 are given by Gibbs densities with ℓ1and ℓ2-norm energies, respectively: pr(x) = e x ℓ1/(cd vold(Bℓ1)) and pα n(x) = e α d x ℓ2/(cd vold(Bd/(α Here cd := Γ(d + 1) where Γ( ) is the usual Gamma function. The star bodies induced by pr and pα n are dilations of the ℓ1-ball and ℓ2-ball, respectively. Denote these star bodies by Lr and Lα n, respectively. Then, Lα r,n is defined by the difference of radial functions via ρLα r,n(u) := cr x (d+1) ℓ1 cn(α d x ℓ2) (d+1) 1/(d+1) where cr := R 0 td exp( t)/(cd vold(Bℓ1)) and cn := R 0 td exp( t)/(cd vold(Bd/(α d))). We visualize the geometry of Lα r,n for different values of α 1 in Figure 2 for d = 2. We see that for directions such that the boundaries of Lr and Lα n are far apart (namely, the ei directions), the optimal regularizer assigns small values. This is because such directions are considered highly likely under the distribution Dr and less likely under Dα n. However, for directions in which the boundaries are close (e.g., in the [0.5, 0.5]-direction), the regularizer assigns high values, since such directions are likely under the noise distribution Dα n. This aligns with the aim of the objective function namely, that a regularizer should assign low values to likely directions and high values to unlikely ones. Example 2 (Toy inverse problem). Consider the following example where x Dr := N(0, Σ) where Σ Rd d and we have measurements y = Ax where A Rm d with rank m d. We consider the case when Dn := A #(Dy) N(0, σ2Id) where y Dy if and only if y = Ax where x Dr. Let d = 2, Σ = UU T , and A = [1, 0] = e T 1 R1 2. Then A = e1. Note that y Dy if and only if y = e T 1 Uz = u T 1 z where u T 1 is the first row of U and z N(0, I2). By standard properties of Gaussians, Dy = N(0, u1 2 ℓ2). Hence Dn = N(0, Dσ) where Dσ := diag( u1 2 ℓ2 + σ2, σ2) R2 2. We visualize this example in Figure 3. We see that the regularizer induced by Lr,n penalizes directions in the row span of A, but does not for directions in the kernel of A. Figure 2: We plot Lα r,n from Example 1 for different values of α: (Left) α = 1.3, (Middle) α = 1.6, and (Right) α = 2.3. A full figure with Lr and Lα n can be found in Section D.3. Figure 3: We visualize the distributions from Example 2 when Σ = [0.5477, 0.2739; 0, 0.5477] R2 2, σ2 = 0.01, and we set Dσ := 0.01 diag( u1 2 ℓ2 + σ2, σ2): (Left) the boundaries of Lr and Ln, (Middle) the boundaries of Lr, Ln, and Lr,n and (Right) Lr,n. 3 Critic-based loss functions via f-divergences Inspired by the use of the variational representation of the Wasserstein distance to define loss functions to learn regularizers, we consider whether other divergences can give rise to valid loss functions and if they can be interpreted as dual mixed volumes. A general class of divergences of interest will be f-divergences: for a convex function f : (0, ) R with f(1) = 0, if P Q, Df(P||Q) := R Rd f d P d Q d Q. α-Divergences: For α ( , 0) (0, 1), there is a variational representation of the α-divergence where f = fα with fα(x) = xα αx (1 α) α(α 1) [69]: Dfα(P||Q) := Dα(P||Q) := 1 α(1 α) inf h:Rd (0, ) Note that the domain of functions in this variational representation consists of positive functions, which aligns with the class of gauges K. Many well-known divergences are α-divergences for specific α, such as the χ2-divergence (α = 1) and the (squared) Hellinger distance (α = 1/2) [69]. We discuss in the appendix (Section B.3.1) how one can derive a general loss based on α-divergences, but focus on a particular case here to illustrate ideas. The Hellinger Distance: Setting α = 1/2 and performing an additional change of variables in (4), we have the following useful representation of the (squared) Hellinger distance: H2(P||Q) := 2 inf h>0 EP [h(x)] + EQ[h(x) 1]. This motivates minimizing the following functional, which carries a similar intuition to (2) that the regularizer should be small on real data and large on unlikely data: K 7 EDr[ x K] + EDn[ x 1 K ]. (5) The following theorem, proved in Section B.3, shows that this loss functional is equal to single dual mixed volume between a data-dependent star body Lr and a star body that depends on Dr, Dn, and K. We can show that the equality cases of the lower bound only hold for dilations within a specific interval, with only two specific star bodies achieving equality. Theorem 3.1. Suppose Dr and Dn admit densities pr and pn, respectively, such that the following maps are positive and continuous over the unit sphere: u 7 ρDr(u)d+1 := Z 0 tdpr(tu)dt and u 7 ρDn(u)d 1 := Z 0 td 2pn(tu)dt. Let Lr and Ln denote the star body with radial function ρDr and ρDn, respectively. Then, for any K Sd, there exists a star body Kr,n that depends on Dr, Dn, and K such that the functional (5) is equal to d V 1(Lr, Kr,n). We also have the inequality V 1(Lr, Kr,n) vold(Kr,n) 1/d vold(Lr)(d+1)/d with equality if and only if Kr,n is a dilate of Lr. Moreover, there exists a λ := λ (Dr, Dn) > 0 such that for every λ (0, λ ], there are only two star bodies K+,λ and K ,λ that satisfy Kr,n = λLr: ρK ,λ(u) := ρLr(u)d 2λρ Ln(u)d 1 1 4λ2 ρ Ln(u) Remark 3.2. Among the two star bodies K+,λ, K ,λ that achieve Kr,n = λLr, we argue that K+,λ induces a better regularizer than K ,λ. As seen in previous examples, we would like our regularizer K to assign low values to likely data and high values to unlikely data. Equivalently, we would like ρK( ) to be large on likely data and small on unlikely data. This can be stated in terms of the geometry of Lr and Ln: we would like ρK to be large when ρLr is large and ρ Ln is small. Likewise, ρK should be small when ρ Ln is large and ρLr is small. Due to the small sign difference between K+,λ and K ,λ, K+,λ better captures this intuition. We show a visual example of this in Figure 4, where Dr and Dn are the same distributions as in Example 1. Figure 4: (Left) The star bodies Lr and Ln induced by the distributions Dr and Dn from from Theorem 3.1 and Example 1 with α = 0.5. Then we have (Middle) K+,λ and (Right) K ,λ as defined in Theorem 3.1. Note that K+,λ better captures the geometry of a regularizer that assigns higher likelihood to likely data and lower likelihood to unlikely data, while K ,λ does not. 3.1 Empirical comparison with adversarial regularization We now aim to understand whether the Hellinger-based loss (5) can provide a practical loss function to learn regularizers. To do this, we consider denoising on the MNIST dataset [50]. We take 10000 random samples from the MNIST training set (constituting our Dr distribution) and add Gaussian noise with variance σ2 = 0.05 (constituting our Dn distribution). The goal is then to reconstruct test samples from the MNIST dataset corrupted with Gaussian noise of the same variance seen during training. The regularizers were parameterized via a deep convolutional neural network with positively homogenous Leaky Re LU activation functions, giving rise to a star-shaped regularizer. They were trained using the adversarial loss and Hellinger-based loss (5). We also used the gradient penalty term from [53] for both losses. Once each regularizer Rθ has been trained, we then denoise a new noisy MNIST digit y by minimizing x 7 x y 2 ℓ2 + λRθ(x) via gradient descent. As a baseline, we compare both methods to Total Variation (TV) regularization [44]. We note that in [53], it was recommended to set λ := 2 λ where λ := EN(0,σ2I)[ z ℓ2], which was motivated by the fact that the regularizer will be 1-Lipschitz. We found our Hellinger-based regularizer performed best with λ := 5.1 λ2. We thus also compare to the adversarial regularizer with a further tuned parameter λ. Please see Section C.1 of the appendix for implementation details and a further discussion of regularization parameter choices. We show the average Peak Signal-to-Noise Ratio (PSNR) and mean squared error (MSE) over 100 test images in the table below. We see that the Hellinger-based loss gives competitive performance relative to the adversarial regularizer and TV regularization, while the tuned adversarial regularizer slightly outperforms all methods. These promising results suggests that these new α-divergence based losses are potentially worth exploring from a practical perspective as well. Hellinger 23.06 0.0052 Adversarial 22.32 0.0064 Adversarial (tuned) 24.14 0.0041 TV 20.52 0.0091 4 Computational considerations We now discuss various issues relating to employing such star body regularizers. In particular, we discuss optimization-based concerns due to the (potential) nonconvexity of x 7 x K and how one can use neural networks to learn star body regularizers in high-dimensions. Weak convexity: When such a regularizer is employed in inverse problems, one may require minimizing a cost of the form x 7 y A(x) 2 ℓ2 + λϕ( x K). Here ϕ( ) is a continuous monotonically increasing function (e.g., ϕ(t) = tq for q > 0). While the first term is convex for linear A( ), the second term can be highly nonconvex for general star bodies. One could enforce searching for a convex body as opposed to a star body, but convexity is strictly more limited in terms of expressibility. There is a large body of recent work [24, 25, 19, 36, 80] that has analyzed weak convexity in optimization, showing that while nonconvex, weakly convex functions can be provably optimized using simple first-order methods in some cases. Based on this, it is important to understand under what conditions is ϕ( K) a weakly convex function. Recall that a function f is ρ-weakly convex if f( ) + ρ 2 2 ℓ2 is convex. For certain functions ϕ( ), there is a natural way to determine if ϕ( K) is weakly convex. In particular, in considering the quantity 2 K + ρ 2 2 ℓ2, there is a natural way in which one can describe this functional as the gauge of a new star body. Definition 4.1 ([55]). For two star bodies K, L Sd and scalars α, β 0 (both not zero), the harmonic q-combination for q 1 is the star body α K ˆ+qβ L whose radial function is defined by ρα K ˆ+qβ L(u) q := αρK(u) q + βρL(u) q. Observe that the scaling operation α K differs from the usual αK := {αx : x K} since αρ q K = ρ q α q K so that α K = α q K. Notice that in our context, we have a similar combination given by the harmonic 2-combination of K and the unit ball Bd: M2,ρ := M2,ρ(K) := K ˆ+2 ρ 2Bd. By the definition of the harmonic 2-combination, the gauge of M2,ρ satisfies x 2 M2,ρ = x 2 K + ρ 2 x 2 ℓ2. We thus have the following Proposition, proven in Section B.4, showing the equivalence between the weak convexity of 2 K and the convexity of M2,ρ. We also show a visual example in Section D.4 of a data-dependent star body whose gauge squared is weakly convex. Proposition 4.2. Let K Sd. Then x 7 x 2 K is ρ-weakly convex if and only if M2,ρ is convex. Deep neural network-based parameterizations: Since deep neural networks allow for efficient learning in high-dimensions, it is natural to ask under what conditions does the regularizer Rθ(x) := f θL L f θL 1 L 1 f θ1 1 (x) define a star body regularizer, i.e., when is the set Kθ := {x Rd : Rθ(x) 1} a star body? As star bodies are in one-to-one correspondence with radial functions, one can answer this question by studying conditions under which the map u 7 1/Rθ(u) defines a radial function. This leads to the following Proposition, which is proven in Section B.4. Proposition 4.3. For L N, consider a regularizer R : Rd R of the form R(x) := G f L f1(x) where G : Rd L R and each fi : Rdi 1 Rdi for i [L] satisfy the following conditions: (i) d = d0 d1 d2 d L, (ii) G( ) is non-negative, positively homogenous, continuous, and only vanishes at the origin, (iii) fi( ) is injective, continuous, and positively homogenous. Then the set K := {x Rd : R(x) 1} is a star body in Rd with radial function ρK(u) := 1/R(u). Below we discuss the different types of commonly employed layers and activation functions that satisfy the conditions of the above Proposition. Example 3 (Feed-forward layers). Any intermediate layer fi : Rdi 1 Rdi can be parameterized by a single-layer feedforward MLP with no biases: fi(x) := σi(Wix) where Wi Rdi Rdi 1 has rank di 1. For the activation function σi( ), one can ensure injectivity and positive homogeneity by choosing the Leaky Re LU activation Leaky Re LUα(t) := Re LU(t)+αRe LU( t) where α (0, 1). Observe that this map is positively homogenous, continuous, and bijective. The layer fi thus satisfies condition (iii). For the final layer G( ), note that any norm on Rd satisfies condition (ii). Example 4 (Residual layers). If di = di 1, one could use residual layers of the form fi(x) := x + gi(x) where gi : Rdi Rdi is a Lipschitz continuous, positively homogenous function with Lip(gi) < 1. This can be achieved by having gi be a neural network with 1-Lipschitz positively homogenous activations (such as Re LU), no biases, and weight matrices with norm strictly less than 1. Then each fi is positively homogenous, continuous, and invertible [11], satisfying condition (iii). Remark 4.4 (Star-shaped regularizers). Note that if one does not require the set K := {x Rd : R(x) 1} to be a star body, but simply star-shaped, then there is more flexibility in the neural network architecture used. In particular, one simply needs the layers to be positively homogenous and continuous, and the output layer be non-negative, positively homogenous, and continuous. Invertibility would not be necessary, so that the dimensions di need not be increasing. This can easily be achieved via linear convolutional layers and positively homogenous activation functions. In fact, in the original work [53], experiments were done with a network precisely of this form. Remark 4.5 (Positive homogeneity of activations). While our theory for neural networks mainly applies to positively homogenous activation functions such as Re LU and Leaky Re LU, we show in Section C.2 of the appendix that focusing on such activations does not limit performance. Specifically, we compare networks with non-positively homogenous activation functions (such as Tanh or the Gaussian Error Linear Unit (GELU)) with a network exclusively using Leaky Re LU activations in denoising. We show that the Leaky Re LU-based regularizer outperforms regularizers with non-positively homogenous activations. This potentially highlights empirical benefits of positive homogeneity, which warrants further analysis. 5 Discussion In this work, we studied optimal task-dependent regularization over the class of regularizers given by star body gauges. Our analysis focused on learning approaches that optimize a critic-based loss function derived from variational representations of probability measures. Utilizing tools from dual Brunn-Minkowski theory and star geometry, we precisely characterized the optimal regularizer in specific cases and provided visual examples to illustrate our findings. Overall, our work underscores the utility of star geometry in enhancing our understanding of data-driven regularization. Limitations and future work: There are numerous exciting directions for future research. While our analysis concentrated on formulations based on the Wasserstein distance and α-divergence, it would be valuable to extend the analysis to other critic-based losses for learning regularizers. Additionally, it would be interesting to give a precise characterization of global minimizers for the Hellinger-inspired loss (5). Relaxing the assumptions in our results is another significant challenge. For instance, Theorem 2.4 hinges on a containment property of the distributions, and finding ways to relax this assumption could broaden the applicability of our theory. Moreover, many of our results assume a certain regularity in the densities of our distributions. Addressing this would involve extending dual mixed volume inequalities to operate on star-shaped sets rather than star bodies. Finally, it is crucial to assess the performance of these regularizers in solving inverse problems. This includes studying questions regarding sample complexity, algorithmic properties, robustness to noise, variations in the forward model, and scenarios with limited clean data [22, 45, 32, 51, 85, 86, 21, 31]. [1] Alekh Agarwal, Animashree Anandkumar, Prateek Jain, and Praneeth Netrapalli. Learning sparsely used overcomplete dictionaries via alternating minimization. SIAM Journal on Optimization, 26(4):2775 2799, 2016. [2] Alekh Agarwal, Animashree Anandkumar, and Praneeth Netrapalli. A clustering approach to learn sparsely-used overcomplete dictionaries. IEEE Transactions on Information Theory, 63(1):575 592, 2017. [3] Michal Aharon, Michael Elad, and Alfred Bruckstein. K-svd: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Transactions on Signal Processing, 54(11):4311 4322, 2006. [4] Giovanni S. Alberti, Ernesto De Vito, Matti Lassas, Luca Ratti, and Matteo Santacesaria. Learning the optimal tikhonov regularizer for inverse problems. Advances in Neural Information Processing Systems (Neur IPS), 2021. [5] Dennis Amelunxen, Martin Lotz, Michael B. Mc Coy, and Joel A. Tropp. Living on the edge: Phase transitions in convex programs with random data. Information and Inference: A Journal of the IMA, 3(3):224 294, 2014. [6] Sanjeev Arora, Rong Ge, Tengyu Ma, and Ankur Moitra. Simple, efficient, and neural algorithms for sparse coding. Conference on Learning Theory, 2015. [7] Simon Arridge, Peter Maass, Ozan Öktem, and Carola-Bibiane Schönlieb. Solving inverse problems using data-driven models. Acta Numerica, 28:1 174, 2019. [8] Muhammad Asim, Max Daniels, Oscar Leong, Ali Ahmed, and Paul Hand. Invertible generative models for inverse problems: mitigating representation error and dataset bias. Proceedings of the 37th International Conference on Machine Learning, 2020. [9] Muhammad Asim, Fahad Shamshad, and Ali Ahmed. Blind image deconvolution using deep generative priors. IEEE Transactions on Computational Imaging, 6:1493 1506, 2018. [10] Boaz Barak, Jonathan A. Kelner, and David Steurer. Dictionary learning and tensor decomposition via the sum-of-squares method. Proceedings of the Forty-seventh Annual ACM Symposium on Theory of Computing, page 143 151, 2015. [11] Jens Behrmann, Will Grathwohl, Ricky T. Q. Chen, David Duvenaud, and Jörn-Henrik Jacobsen. Invertible residual networks. International Conference on Machine Learning, 2018. [12] Badri Bhaskar and Benjamin Recht. Atomic norm denoising with applications to line spectral estimation. IEEE Transactions on Signal Processing, 61(23):5987 5999, 2013. [13] Ashish Bora, Ajil Jalal, Eric Price, and Alexandros Dimakis. Compressed sensing using generative models. International Conference on Machine Learning, 2017. [14] Andrea Braides. A handbook of Γ-convergence. Handbook of Differential Equations: Stationary Partial Differential Equations, Volume 3, 2007. [15] Emmanuel J. Candès and Benjamin Recht. Exact matrix completion via convex optimization. Foundations of Computational Mathematics, 9(6):717 772, 2009. [16] Emmanuel J. Candès, Justin K. Romberg, and Terence Tao. Stable signal recovery from incomplete and inaccurate measurements. Communications on Pure and Applied Mathematics, 59(8):1207 1223, 2006. [17] Venkat Chandrasekaran and Michael I. Jordan. Computational and statistical tradeoffs via convex relaxation. Proceedings of the National Academy of Sciences, 110(13):E1181 E1190, 2013. [18] Venkat Chandrasekaran, Benjamin Recht, Pablo A. Parrilo, and Alan S. Willsky. The convex geometry of linear inverse problems. Foundations of Computational Mathematics, 12:805 849, 2012. [19] Vasileios Charisopoulos, Yudong Chen, Damek Davis, Mateo Díaz, Lijun Ding, and Dmitriy Drusvyatskiy. Low-rank matrix recovery with composite optimization: good conditioning and rapid convergence. Foundations of Computational Mathematics, 21:1505 1593, 2021. [20] Hyungjin Chung, Jeongsol Kim, Michael T. Mccann, Marc L. Klasky, and Jong Chul Ye. Diffusion posterior sampling for general noisy inverse problems. International Conference on Learning Representations (ICLR), 2023. [21] Giannis Daras, Kulin Shah, Yuval Dagan, Aravind Gollakota, Alexandros G. Dimakis, and Adam Klivans. Ambient diffusion: Learning clean distributions from corrupted data. Advances in Neural Information Processing Systems (Neur IPS), 2023. [22] Mohammad Zalbagi Darestani, Akshay S. Chaudhari, and Reinhard Heckel. Measuring robustness in deep learning based compressive sensing. International Conference on Machine Learning (ICML), 2021. [23] Ingrid Daubechies, Michel Defrise, and Christine de Mol. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Communications on Pure and Applied Mathematics, 57(11):1413 1457, 2004. [24] Damek Davis and Dmitriy Drusvyatskiy. Stochastic model-based minimization of weakly convex functions. SIAM Journal on Optimization, 29(1):207 239, 2019. [25] Damek Davis and Dmitriy Drusvyatskiy. Proximal methods avoid active strict saddles of weakly convex functions. Foundations of Computational Mathematics, 22:561 606, 2022. [26] David Donoho. For most large underdetermined systems of linear equations the minimal l1norm solution is also the sparsest solution. Communications on Pure and Applied Mathematics, 59(6):797 829, 2006. [27] Michael Elad. Sparse and redundant representations: From theory to applications in signal and image processing. Springer, 2010. [28] Zhenghan Fan, Sam Buchanan, and Jeremias Sulam. What s in a prior? learned proximal networks for inverse problems. International Conference on Learning Representations (ICLR), 2024. [29] Maryam Fazel. Matrix rank minimization with applications. Ph.D. Thesis, Department of Electrical Engineering, Stanford University, 2002. [30] Berthy T. Feng, Jamie Smith, Michael Rubinstein, Huiwen Chang, Katherine L. Bouman, and William T. Freeman. Score-based diffusion models as principled priors for inverse imaging. Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 10520 10531, 2023. [31] Angela Gao, Jorge Castellanos, Yisong Yue, Zachary Ross, and Katherine Bouman. Deepgem: Generalized expectation-maximization for blind inversion. Advances in Neural Information Processing Systems (Neur IPS), 34:11592 11603, 2021. [32] Angela F. Gao, Oscar Leong, He Sun, and Katherine L. Bouman. Image reconstruction without explicit priors. ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1 5, 2023. [33] Richard J. Gardner. Geometric tomography. Cambridge: Cambridge University Press, 2006. [34] Richard J. Gardner, Markus Kiderlen, and Peyman Milanfar. Convergence of algorithms for reconstructing convex bodies and directional measures. The Annals of Statistics, 34(3):1331 1374, 2006. [35] Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. Advances in Neural Information Processing Systems, 2014. [36] Alexis Goujon, Sebastian Neumayer, Pakshal Bohra, Stanislas Ducotterd, and Michael Unser. A neural-network-based convex regularizer for inverse problems. IEEE Transactions on Computational Imaging, 9:781 795, 2023. [37] Alexis Goujon, Sebastian Neumayer, and Michael Unser. Learning weakly convex regularizers for convergent image-reconstruction algorithms. SIAM Journal on Imaging Sciences, 17(1):91 115, 2024. [38] Adityanand Guntuboyina. Optimal rates of convergence for convex set estimation from support functions. Annals of Statistics, 40(1):385 411, 2012. [39] Paul Hand, Oscar Leong, and Vlad Voroninski. Phase retrieval under a generative prior. Advances in Neural Information Processing Systems (Neur IPS), 31, 2018. [40] Paul Hand, Oscar Leong, and Vladislav Voroninski. Compressive phase retrieval: Optimal sample complexity with deep generative priors. Communications on Pure and Applied Mathematics, 77(2):1147 1223, 2024. [41] G. Hansen, I. Herburt, H. Martini, and M. Moszynska. Starshaped sets. Aequationes Mathematicae, 94:1001 1092, 2020. [42] Reinhard Heckel, Wen Huang, Paul Hand, and Vladislav Voroninski. Rate-optimal denoising with deep neural networks. Information and Inference: A Journal of the IMA, 10(4):1251 1285, 2021. [43] Rakib Hyder, Viraj Shah, Chinmay Hedge, and M. Salman Asif. Alternating phase projected gradient descent with generative priors for solving compressive phase retrieval. Proc. IEEE Int. Conf. Acoust., Speech, and Signal Processing (ICASSP), 2019. [44] Leonid I.Rudin, Stanley Osher, and Emad Fatemi. Nonlinear total variation based noise removal algorithms. Physica D: Nonlinear Phenomena, 60:259 268, 1992. [45] Ajil Jalal, Marius Arvinte, Giannis Daras, Eric Price, Alexandros G. Dimakis, and Jon Tamir. Robust compressed sensing mri with deep generative priors. Advances in Neural Information Processing Systems (Neur IPS), 2021. [46] Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. International Conference on Learning Representations (ICLR), 2014. [47] Erich Kobler, Alexander Effland, Karl Kunisch, and Thomas Pock. Total deep variation for linear inverse problems. IEEE Conference on Computer Vision and Pattern Recognition, 2020. [48] Frederic Koehler, Lijia Zhou, Danica J. Sutherland, and Nathan Srebro. Uniform convergence of interpolators: Gaussian width, norm bounds, and benign overfitting. Advances in Neural Information Processing Systems (Neur IPS), 2021. [49] Gil Kur, Alexander Rakhlin, and Adityanand Guntuboyina. On suboptimality of least squares with application to estimation of convex bodies. Conference on Learning Theory, pages 2406 2424, 2020. [50] Yann Le Cun, Leon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. [51] Oscar Leong, Angela F. Gao, He Sun, and Katherine L. Bouman. Discovering structure from corruption for unsupervised image reconstruction. IEEE Transactions on Computational Imaging, 9:992 1005, 2023. [52] Oscar Leong, Eliza O Reilly, Yong Sheng Soh, and Venkat Chandrasekaran. Optimal regularization for a data source. ar Xiv preprint ar Xiv:2212.13597, 2022. [53] Sebastian Lunz, Ozan Öktem, and Carola-Bibiane Schönlieb. Adversarial regularizers in inverse problems. Advances in Neural Information Processing Systems (Neur IPS), 31, 2018. [54] Erwin Lutwak. Dual mixed volumes. Pacific Journal of Mathematics, 58(2):531 538, 1975. [55] Erwin Lutwak. The brunn minkowski firey theory ii: Affine and geominimal surface areas. Advances in Mathematics, 118(2):244 294, 1996. [56] Julien Mairal, Francis Bach, and Jean Ponce. Sparse modeling for image and vision processing. Foundations and Trends in Computer Graphics and Vision, 8(2 3):85 283, 2014. [57] Morteza Mardani, Enhao Gong, Joseph Y Cheng, Shreyas S Vasanawala, Greg Zaharchuk, Lei Xing, and John M Pauly. Deep generative adversarial neural networks for compressive sensing mri. IEEE Transactions on Medical Imaging, 38(1):167 179, 2019. [58] Tristan Milne, Etienne Bilocq, and Adrian Nachman. A new method for determining wasserstein 1 optimal transport maps from kantorovich potentials, with deep learning applications. ar Xiv preprint ar Xiv:2211.00820, 2022. [59] Subhadip Mukherjee, Sören Dittmer, Zakhar Shumaylov, Sebastian Lunz, Ozan Öktem, and Carola-Bibiane Schönlieb. Learned convex regularizers for inverse problems. ar Xiv preprint ar Xiv:2008.02839, 2021. [60] Bruno A. Olshausen and David J. Field. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature, 381(6583):607 609, 1996. [61] Bruno A. Olshausen and David J. Field. Sparse coding with an overcomplete basis set: A strategy employed by v1? Vision in Research, 37(23):3311 3325, 1997. [62] Gregory Ongie, Ajil Jalal, Christopher A. Metzler, Richard G. Baraniuk, Alexandros G. Dimakis, and Rebecca Willett. Deep learning techniques for inverse problems in imaging. IEEE Journal on Selected Areas in Information Theory, 1(1):39 56, 2020. [63] Eliza O Reilly and Ngoc Mai Tran. Stochastic geometry to generalize the mondrian process. SIAM Journal on the Mathematics of Data Science, 4(2):531 552, 2022. [64] Samet Oymak and Babak Hassibi. Sharp mse bounds for proximal denoising. Foundations of Computational Mathematics, 16:965 1029, 2016. [65] Samet Oymak, Benjamin Recht, and Mahdi Soltanolkotabi. Sharp time-data tradeoffs for linear inverse problems. IEEE Transactions on Information Theory, 64(6):4129 4158, 2018. [66] Yaniv Plan and Roman Vershynin. The generalized lasso with non-linear observations. IEEE Transactions on Information Theory, 62:1528 1537, 2016. [67] Yaniv Plan, Roman Vershynin, and Elena Yudovina. High-dimensional estimation with geometric constraints. Information and Inference: A Journal of the IMA, 6(1):1 40, 2017. [68] David Pollard. Convergence of stochastic processes. Springer-Verlag, 1984. [69] Yury Polyanskiy and Yihong Wu. Information Theory: From Coding to Learning. Cambridge University Press, 2023. [70] Benjamin Recht, Maryam Fazel, and Pablo A. Parrilo. Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM Review, 52(3):471 501, 2010. [71] Yaniv Romano, Michael Elad, and Peyman Milanfar. The little engine that could: Regularization by denoising (red). SIAM Journal on Imaging Sciences, 10(4):1804 1844, 2017. [72] Litu Rout, Negin Raoof, Giannis Daras, Constantine Caramanis, Alexandros G. Dimakis, and Sanjay Shakkottai. Solving linear inverse problems provably via posterior sampling with latent diffusion models. Advances in Neural Information Processing Systems (Neur IPS), 2023. [73] A.M. Rubinov. Radiant sets and their gauges. In: Quasidifferentiability and Related Topics, 43:235 261, 2000. [74] Leonid I Rudin, Stanley Osher, and Emad Fatemi. Nonlinear total variation based noise removal algorithms. Physica D: nonlinear phenomena, 60(1-4):259 268, 1992. [75] Karin Schnass. On the identifiability of overcomplete dictionaries via the minimisation principle underlying k-svd. Applied and Computational Harmonic Analysis, 37(3):464 491, 2014. [76] Karin Schnass. Convergence radius and sample complexity of itkm algorithms for dictionary learning. Applied and Computational Harmonic Analysis, 45(1):22 58, 2016. [77] Rolf Schneider. Convex bodies: The brunn minkowski theory. Cambridge: Cambridge University Press., 2013. [78] Parikshit Shah, Badri Narayan Bhaskar, Gongguo Tang, and Benjamin Recht. Linear system identification via atomic norm regularization. Proceedings of the 51st Annual Conference on Decision and Control, 2012. [79] Fahad Shamshad and Ali Ahmed. Compressed sensing-based robust phase retrieval via deep generative priors. IEEE Sensors Journal, 21(2):2286 2298, 2017. [80] Zakhar Shumaylov, Jeremy Budd, Subhadip Mukherjee, and Carola-Bibiane Schönlieb. Weakly convex regularisers for inverse problems: Convergence of critical points and primal-dual optimisation. ar Xiv preprint ar Xiv:2402.01052, 2024. [81] Yong Sheng Soh and Venkat Chandrasekaran. Learning semidefinite regularizers. Foundations of Computational Mathematics, 19:375 434, 2019. [82] Yong Sheng Soh and Venkat Chandrasekaran. Fitting tractable convex sets to support function evaluations. Discrete and Computational Geometry, 66:510 551, 2021. [83] Mahdi Soltanolkotabi. Structured signal recovery from quadratic measurements: Breaking sample complexity barriers via nonconvex optimization. IEEE Transactions on Information Theory, 65(4):2374 2400, 2019. [84] Daniel A. Spielman, Huan Wang, and John Wright. Exact recovery of sparsely-used dictionaries. Conference on Learning Theory, 23(37):1 18, 2012. [85] Julián Tachella, Dongdong Chen, and Mike Davies. Unsupervised learning from incomplete measurements for inverse problems. Advances in Neural Information Processing Systems (Neur IPS), 2022. [86] Julián Tachella, Dongdong Chen, and Mike Davies. Sensing theorems for unsupervised learning in linear inverse problems. Journal of Machine Learning Research, 24(39):1 45, 2023. [87] Gongguo Tang, Badri Narayan Bhaskar, Parikshit Shah, and Benjamin Recht. Compressed sensing off the grid. IEEE Transactions on Information Theory, 59(11):7465 7490, 2013. [88] Ryan Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society, Series B, 58:267 288, 1994. [89] Andrey Nikolayevich Tikhonov. On stability of inverse problems. Dokl. Akad. Nauk SSSR, 39(5):176 179, 1943. [90] Yann Traonmilin, Rémi Gribonval, and Samuel Vaiter. A theory of optimal convex regularization for low-dimensional recovery. Information and Inference: A Journal of the IMA, 13(2):iaae013, 2024. [91] Singanallur V Venkatakrishnan, Charles A Bouman, and Brendt Wohlberg. Plug-and-play priors for model based reconstruction. In 2013 IEEE Global Conference on Signal and Information Processing, pages 945 948. IEEE, 2013. [92] Cédric Villani. Topics in optimal transport. Providence, RI: American Mathematical Society, 2003. 1 Introduction 1 1.1 Our contributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 1.2 Related work . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 1.3 Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 2 Adversarial star body regularization 4 2.1 Existence of minimizers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 2.2 Minimization via dual Brunn-Minkowski theory . . . . . . . . . . . . . . . . . . . 4 2.3 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 3 Critic-based loss functions via f-divergences 7 3.1 Empirical comparison with adversarial regularization . . . . . . . . . . . . . . . . 8 4 Computational considerations 9 5 Discussion 10 A Further discussion on related work 16 B Proofs 17 B.1 Preliminaries on star and convex geometry . . . . . . . . . . . . . . . . . . . . . . 17 B.2 Proofs for Section 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 B.3 Proofs for Section 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 B.3.1 A general result for α-divergences . . . . . . . . . . . . . . . . . . . . . . 21 B.4 Proofs for Section 4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 C Experimental details 24 C.1 Hellinger and adversarial loss comparison on MNIST . . . . . . . . . . . . . . . . 24 C.2 Analysis on positive homogeneity . . . . . . . . . . . . . . . . . . . . . . . . . . 25 D Additional results 25 D.1 Stability results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 D.2 Proof of Theorem D.2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 D.3 Further examples for Section 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 D.4 A star body with weakly convex squared gauge . . . . . . . . . . . . . . . . . . . 31 A Further discussion on related work We survey topics relating to data-driven regularization and convex geometry for inverse problems. For a more in-depth discussion on learning-based regularization methods, we refer the reader to the following survey articles [7, 62]. Regularization via generative modeling: Generative networks have demonstrated their effectiveness as signal priors in various inverse problems, including compressed sensing [45, 13, 57], denoising [42], phase retrieval [39, 40, 43, 79], and blind deconvolution [9]. Early approaches in this area leveraged push-forward generators such as Generative Adversarial Networks (GANs) [35] and Variational Autoencoders (VAEs) [46]. In some cases, generative models can serve as traditional regularizers by providing access to likelihoods, which can be incorporated as penalty functions into the objective [8]. Alternatively, they can constrain reconstructions to lie on the natural image manifold, ensuring that solutions are realistic and consistent with the underlying data distribution. More recently, there has been a surge of interest in techniques based on diffusion models [20, 30, 21, 72]. Explicit regularization: There have also been recent works that have directly learned regularizers as penalty functions in a data-driven fashion. Early works in this area include the now-mature field of dictionary learning (see the surveys [56, 27] for more). Regularization by Denoising [71] constructs an explicit regularization functional using an off-the-shelf deep neural network-based denoiser. The works [53, 59, 80] learn a regularizer via the critic-based adversarial loss inspired by the Wasserstein distance. The difference in these works lie in the way the neural network is parameterized to enforce certain properties, such as convexity or weak convexity. Recently, [28] developed an approach that can approximate the proximal operator of a potentially nonconvex data-driven regularizer. Other works along these lines include [36, 4, 37]. Convex geometry: Our work is also a part of a line of research showcasing the tools of convex and star geometry in machine learning. In statistical inference tasks and inverse problems, the utility of convex geometry cannot be understated. In particular, complexity measures of convex bodies, such their Gaussian width [18, 17] or statistical dimension [5], have proven useful in developing sample complexity guarantees for a variety of inverse problems [64, 66, 67, 48, 65, 83]. Additionally, convex geometry has been fruitful in problems such as function estimation from noisy data [38, 34, 63, 82, 49]. B.1 Preliminaries on star and convex geometry In this section, we provide relevant background concerning convex bodies and star bodies. For a more detailed treatment of each subject, we recommend [77] for convex geometry and [33, 41] for star geometry. We say that a closed set K Rd is star-shaped (with respect to the origin) if for all x K, the line segment [0, x] := {tx : t [0, 1]} K. In order for K to be considered a star body, we require K to be a compact star-shaped set such that for every x = 0, the ray Rx := {tx : t > 0} intersects the boundary of K exactly once. Note that every convex body (a compact, convex set with the origin its interior) is a star body. In an analogous fashion to the correspondence between convex bodies and their support functions, star bodies are in one-to-one correspondence with radial functions. For a star body K, its radial function is defined by ρK(x) := sup{t > 0 : t x K}. It follows that the gauge function of K satisfies x K = 1/ρK(x) for all x Rd such that x = 0. The radial function can also help capture important geometric information about K. For example, one can show that vold(K) = 1 Sd 1 ρd K(u)du. For K, L Sd, it is easy to see that K L if and only if ρK ρL. We say that L is a dilate of K if there exists λ > 0 such that L = λK; that is, ρL = λρK. In addition, given a linear transformation ϕ GL(Rd), one has ρϕK(x) = ρK(ϕ 1x) for all x Rd\{0}. An additional important aspect of star bodies is their kernels. Specifically, we define the kernel of a star body K as the set of all points for which K is star-shaped with respect to, i.e., ker(K) := {x K : [x, y] K, y K}. Note that ker(K) is a convex subset of K and ker(K) = K if and only if K is convex. For a parameter γ > 0, we define the following subset of Sd consisting of well-conditioned star bodies with nondegenerate kernels: Sd(γ) := {K Sd : γBd ker(K)}. Here Bd := {x Rd : x ℓ2 1} is the unit (Euclidean) ball in Rd. We finally discuss metric aspects of the space of star bodies. While the space of convex bodies enjoys the natural Hausdorff metric, there is a different, yet more natural metric over the space of star bodies dubbed the radial metric. Concretely, let K, L Rd be star bodies. We define the radial sum + between K and L as K +L := {x + y : x K, y L, x = λy}; that is, unlike the Minkowski sum, we restrict the pair of vectors to be parallel. The radial sum obeys the relationship ρK +L(u) := ρK(u) + ρL(u). We denote the radial metric between two star bodies K, L as δ(K, L) := inf{ε 0 : K L + εBd, L K + εBd}. In a similar fashion to how the support function induces the Haussdorf metric for convex bodies, the radial metric satisfies δ(K, L) := ρK ρL . A functional F : Sd R on the space of star bodies is continuous if it is continuous with respect to the radial metric, i.e., F( ) is continuous if for every K Sd and (Ki) such that δ(Ki, K) 0 as i , then F(Ki) F(K) as i . Moreover, a subset C Sd is closed if it is closed with respect to the topology induced by the radial metric. Finally, a subset C Sd is bounded if there exists an 0 < R < such that K RBd for every K C. B.2 Proofs for Section 2 We require some preliminary results for our proofs. The first is a simple extension of a result in [52] to show continuity of our objective functional. Lemma B.1. Let Dr and Dn be distributions over Rd such that EDi[ x ℓ2] < for i = r, n. Then for any K, L Sd(γ), we have |F(K; Dr, Dn) F(L; Dr, Dn)| EDr[ x ℓ2] + EDr[ x ℓ2] γ2 δ(K, L). Proof of Lemma B.1. We first note via the triangle inequality that |F(K; Dr, Dn) F(L; Dr, Dn)| | EDr[ x K] EDr[ x L]| + | EDn[ x K] EDn[ x L]|. Then the result follows from Corollary 2 in [52]. We additionally need the following local compactness result over the space of star bodies. Theorem B.2 (Theorem 5 in [52]). Fix 0 < γ < and let C be a bounded and closed subset of Sd(γ). Let (Ki) be a sequence of star bodies in C. Then (Ki) has a subsequence that converges in the radial and Hausdorff metric to a star body K C. Proof of Theorem 2.1. The condition EDr[ x ℓ2], EDn[ x ℓ2] < guarantees that the expectation of x K exists for any K Sd(1). To see this, note that since K Sd(1), we have that Bd ker(K) so that x K x ℓ2 for all x Rd. Thus EDi[ x K] EDi[ x ℓ2] < for i = r, n. To prove the existence of minimizers, we first make the following useful connection regarding the minimization problem and the Wasserstein distance. Use the shorthand notation F(K) := F(K; Dr, Dn). Recall the dual Kantorovich formulation of the 1-Wasserstein distance [92]: W1(Dr, Dn) = sup f Lip(1) EDn[f(x)] EDr[f(x)]. Since { K : K Sd(1)} Lip(1), we see that inf K Sd(1) F(K) = inf K Lip(1) EDr[ x K] EDn[ x K] inf f Lip(1) EDr[f(x)] EDn[f(x)] = sup f Lip(1) EDn[f(x)] EDr[f(x)] = W1(Dr, Dn). This will prove to be useful in the existence proof. To show the existence of a minimizer, we only consider the case when W1(Dr, Dn) > 0 since if W1(Dr, Dn) = 0, then any K Sd(1) is a minimizer of the objective since F(K) = 0 for all K Sd(1). Consider a minimizing sequence (Ki) Sd(1) such that F(Ki) inf K Sd(1) F(K) as i . We claim that this sequence must be bounded, i.e., there exists an R > 0 such that Ki RBd for all i 1. Suppose this is not true. Then, for any ℓ= 1, 2, . . . , there exists a subsequence (Kiℓ) (Ki) such that Kiℓ ℓBd for all ℓ. That is, Kiℓgrows arbitrarily large as ℓ . By Lemma B.1, we know that F( ) is continuous. Thus, we also must have that F(Kiℓ) inf K Sd(1) F(K) as ℓ . But since ℓBd Kiℓ, we must have that x Kiℓ 1 ℓ x ℓ2 for any x Rd. This gives 1 ℓEDr[ x ℓ2] EDr[ x Kiℓ] F(Kiℓ) EDr[ x Kiℓ] 1 ℓEDn[ x ℓ2]. Taking the limit as ℓ shows inf K Sd(1) F(K) = 0. But this is a contradiction since 0 = inf K Sd(1) F(K) W1(Dr, Dn) > 0 by assumption. Thus, we must have that the sequence (Ki) is bounded, i.e., there exists an R > 0 such that Ki RBd for all i 1. To complete the proof, observe that this minimizing sequence (Ki) C := {K Sd(1) : Ki RBd}, which is a closed and bounded subset of Sd(1). By Theorem B.2, this sequence must have a convergent subsequence (Kin) with limit K C Sd(1). By continuity of F( ), this means lim in F(Kin) = F(K ) = inf K Sd(1) F(K) as desired. Proof of Theorem 2.4. We first focus on proving the identity EDi[ x K] = Z Sd 1 ρpi(u)d+1ρK(u) 1du for i = r, n. We prove this for Dr, since the proof is identical for Dn. Observe that for any K Sd, since u K = ρK(u) 1, integrating in spherical coordinates gives EDr[ x K] = Z Rd x Kpr(x)dx = Z 0 td u Kpr(tu)dtdu 0 tdpr(tu)dt u Kdu Sd 1 ρpr(u)d+1ρK(u) 1du. Now, applying this identity to our objective yields EDr[ x K] EDn[ x K] = Z Sd 1(ρpr(u)d+1 ρpn(u)d+1)ρK(u) 1du. Since ρpr > ρpn 0 on the unit sphere, we have that the map ρr,n(u) := (ρpr(u)d+1 ρpn(u)d+1)1/(d+1) is positive and continuous over the unit sphere. It is also positively homogenous of degree 1, as both ρpr and ρpn are. Then, we claim that ρr,n is the radial function of the unique star body Lr,n := {x Rd : ρr,n(x) 1 1}. We first show that Lr,n is star-shaped with gauge x Lr,n = ρr,n(x) 1. Note that the set is starshaped since for each x Lr,n, we have for any t [0, 1], ρr,n(tx) 1 = (ρr,n(x)/t) 1 = tρr,n(x) 1 t 1 since ρr,n is positively homogenous of degree 1. Hence [0, x] Lr,n for any x Lr,n. Then, we have that the gauge of Lr,n satisfies x Lr,n = inf{t : x/t Lr,n} = inf{t : ρr,n(x/t) 1 1} = inf{t : ρr,n(x) 1 t} = ρr,n(x) 1. Hence the gauge of Lr,n is precisely ρr,n(x) 1. Moreover, by assumption, u 7 ρr,n(u) is positive and continuous over the unit sphere so Lr,n is a star body and is uniquely defined by ρr,n. Combining the above results, we have the dual mixed volume interpretation EDr[ x K] EDn[ x K] = Z Sd 1 ρr,n(u)d+1ρK(u) 1du = d V 1(Lr,n, K). Armed with this identity, we can apply Theorem 2.3 to obtain EDr[ x K] EDn[ x K] d vold(Lr,n)(d+1)/d vold(K) 1/d with equality if and only if Lr,n and K are dilates. Hence the objective is minimized over the collection of unit-volume star bodies when K := vold(Lr,n) 1/d Lr,n. B.3 Proofs for Section 3 Proof of Theorem 3.1. Let ρDr(u) denote the function induced by Dr via equation (3). We first understand EDn[ x 1 K ] : EDn[ x 1 K ] = Z Rd x 1 K pn(x)dx 0 td 2pn(tu)dt u 1 K du Sd 1 ρDn(u)d 1 u 1 K du Sd 1 ρDn(u)d 1ρK(u)du where here we have defined ρDn(u) := R 0 td 2pn(tu)dt 1/(d 1) . Then, we can further write EDr[ x K] + EDn[ x 1 K ] = Z Sd 1 ρK(u) 1ρDr(u)d+1du + Z Sd 1 ρK(u) ρDn(u)d 1du Sd 1 ρK(u) 1ρDr(u)d+1du + Z Sd 1 ρK(u) ρDn(u)d 1 ρDr(u)d+1 ρDr(u)d+1du ρK(u) 1 + ρK(u) ρDn(u)d 1 ρDr(u)d+1du =: d V 1(Lr, Kr,n) where the star body Kr,n is defined by its radial function u 7 ρK(u) 1 + ρK(u) ρDn(u)d 1 Note that Kr,n is the L1 harmonic radial combination of K and the star body with radial function u 7 ρDr (u)d+1 ρK(u) ρDn(u)d 1 . Let Lr denote the star body with ρDr as its radial function and Ln denote the star body with radial function ρDn. The equality in the lower bound on V 1(Kr,n, Lr) is achieved by a star body K such that for some λ > 0, ρK(u) 1 + ρK(u)ρ Ln(u)d 1 ρLr(u)d+1 = 1 λρLr(u). This is equivalent to λρ Ln(u)d 1 ρLr(u)d ρK(u)2 ρK(u) + λρLr(u) = 0. For notational simplicity, for u Sd 1, let a(u) = ρLr(u) and b(u) = ρ Ln(u). Then for λ > 0, Kr,n = λLr λb(u)d 1 a(u)d ρK(u)2 ρK(u) + λa(u) = 0, u Sd 1. To solve for ρK(u), we must analyze the roots of the polynomial f(x) := λ b(u)d 1 a(u)d x2 x + λa(u). But its roots are given by x := x (u) := 1 1 4λ2 b(u) a(u) d 1 Recall the requirement that ρK must be positive and continuous over the unit sphere. For each u Sd 1, f has 2 positive roots if the discriminant is positive and 1 positive root if it is zero. Otherwise, both roots are complex. Hence, in order to have positive solutions for each u, we must have that d 1 0, u Sd 1 0 < λ d 1 u Sd 1. λ := λ (Dr, Dn) := min u Sd 1 Observe that λ is positive and finite since ρLr and ρ Ln are positive and continuous over the unit sphere. Then for any λ (0, λ ], there are two possible star bodies K+,λ, K ,λ that satisfy the above equation: ρK+,λ(u) := ρLr(u)d 2λρ Ln(u)d 1 1 4λ2 ρ Ln(u) ρK ,λ(u) := ρLr(u)d 2λρ Ln(u)d 1 1 4λ2 ρ Ln(u) Notice that for any u that achieves the minimum in the definition of λ , we have that ρK+,λ (u ) = ρK ,λ (u ). B.3.1 A general result for α-divergences As discussed in the main body of the paper, we are interested in deriving critic-based loss functions inspired by variational representations of divergences between probability measures. For general convex functions f, there is a well-known variational representation of the f-divergence Df defined via its convex conjugate f (y) := supx R xy f(x). Proposition 1 (Theorem 7.24 in [69]). For f convex, let f denote its convex conjugate and Ω denote the effective domain of f . Then any f-divergence has the following variational representation Df(P||Q) = sup g:Rd Ω EP [g] EQ[f g]. When applying this result to α-divergences with f(x) = fα(x) := xα αx (1 α) α(α 1) , we saw that it admits the representation (4). In particular, for α ( , 0) (0, 1), we can consider analyzing the following loss: α 1 EDr[ x α K] + (1 α) 1 EDn[ x α 1 K ]. One can show that this loss can be written in terms of dual mixed volumes. When α (0, 1), the loss can be written as a single dual mixed volume. We will define the following notation, which will be useful throughout the proof: for a distribution D with density p D and β R, we set ρβ,D(u) := Z 0 td+β 1p D(tu)dt 1 d+β . We prove the following: Theorem 5. Fix α ( , 0) (0, 1). For two distributions Dr and Dn with densities pr and pn, suppose that ρα,Dr and ρα 1,Dn and are positive and continuous over the unit sphere. Let Lα r and Lα n denote the star bodies with radial functions ρα,Dr and ρα 1,Dn, respectively. Then there exists star bodies Lα r and Lα n that depend on Dr, Dn, and K such that α 1 EDr[ x α K] + (1 α) 1 EDn[ x α 1 K ] = α 1d V 1( Lα r , K) + (1 α) 1d V 1( Lα n, K). If α (0, 1), then there exists a star body Kα r,n that depends on Dr, Dn, and K such that α 1 EDr[ x α K] + (1 α) 1 EDn[ x α 1 K ] = d V α(Kα r,n, Lα r ) where V α(L, K) is the α-dual mixed volume between L and K. Moreover, we have the inequality V α(Lα r , Kα r,n) vold(Lα r )(d+α)/d vold(Kα r,n) α/d with equality if and only if Kα r,n is a dilate of Lα r . Any star body K such that Kα r,n is a dilate of Lα r must satisfy the following: there exists a λ > 0 such that α 1ρK(u) α + (1 α) 1 ρLα n(u)d+α 1 ρLα r (u)d+α ρK(u)1 α = λ 1/αρLr(u) 1/α for all u Sd 1. Proof of Proposition 5. Consider the loss α 1 EDr[ x α K] + (1 α) 1 EDn[ x α 1 K ]. We first focus on EDr[ x α K]. Observe that EDr[ x α K] = Z Rd x α Kpr(x)dx 0 td+α 1pr(tu)dt u α Kdu Sd 1 ρα,Dr(u)d+α u α 1 K u Kdu Sd 1 ρα,Dr(u)d+αρK(u)1 αρK(u) 1du. Let Lα r be the star body with radial function ρ Lα r (u) := ρα,Dr(u)d+αρK(u)1 α 1 d+1 . Indeed, this is a valid radial function since ρ Lα r is positively homogenous of degree 1 and is positive and continuous over the unit sphere. This shows that EDr[ x α K] = Z Sd 1 ρ Lα r (u)d+1ρK(u) 1du = d V 1( Lα r , K). Similarly, we see that EDn[ x α 1 K ] = Z Rd x α 1 K pn(x)dx 0 td+α 2pn(tu)dt u α 1 K du Sd 1 ρα 1,Dn(u)d+α 1 u α 2 K u Kdu Sd 1 ρα 1,Dn(u)d+α 1ρK(u)2 αρK(u) 1du. Let Lα n be the star body with radial function ρ Lα n(u) := ρα 1,Dn(u)d+α 1ρK(u)2 α 1 d+1 . This shows that EDn[ x α 1 K ] = Z Sd 1 ρ Lα n(u)d+1ρK(u) 1du = d V 1( Lα n, K). For the final result, note that if α (0, 1), then α 1 EDr[ x α K] + (1 α) 1 EDn[ x α 1 K ] Sd 1 α 1ρ Lα r (u)d+1ρ 1 K (u) + (1 α) 1ρ Lα n(u)d+1ρ 1 K (u)du α 1ρα,Dr(u)d+αρK(u)1 α + (1 α) 1ρα 1,Dn(u)d+α 1ρK(u)2 α ρK(u) 1du Sd 1 ρα,Dr(u)d+α α 1ρK(u) α + (1 α) 1 ρα 1,Dn(u)d+α 1 ρα,Dr(u)d+α ρK(u)1 α du =: d V α(Lα r , Kα r,n) where we have defined the star body Kα r,n via its radial function u 7 α 1ρK(u) α + (1 α) 1 ρα 1,Dn(u)d+α 1 ρα,Dr(u)d+α ρK(u)1 α α and we also use the α-dual mixed volume, defined by V α(L, K) := 1 Sd 1 ρL(u)d+αρK(u) αdu. The general dual mixed volume inequality of Lutwak (Theorem 2 in [55]) reads V k i j (L, K) V k j i (L, K) V j i k (L, K), i < j < k, i, j, k R, L, K Sd. Setting i = α, j = 0, k = d gives V α(L, K) vold(L)(d+α)/d vold(K) α/d with equality if and only if K is a dilate of L. Applying this in our scenario gives V α(Lα r , Kα r,n) vold(Lα r )(d+α)/d vold(Kα r,n) α/d with equality Kα r,n = λLα r . This only holds if there exists a K and a λ > 0 such that the radial function of K satisfies α 1ρK(u) α + (1 α) 1 ρLα n(u)d+α 1 ρLα r (u)d+α ρK(u)1 α ! α = λρLr(u) for all u Sd 1. B.4 Proofs for Section 4 Proof of Proposition 4.2. Suppose M2,ρ is convex. Then its gauge x M2,ρ is a convex function. One can show that for any non-decreasing convex function ϕ and non-negative convex function f, the composition ϕ f is convex. Setting f(x) := x M2,ρ and ϕ(t) := t2 shows that x 2 M2,ρ is convex. Recalling that x 2 M2,ρ = x 2 K + ρ 2 x 2 ℓ2 implies that x 2 K is ρ-weakly convex. On the other hand, suppose x 2 K is ρ-weakly convex. Then x 2 M2,ρ is convex. This implies that the level set S 1 := {x Rd : x 2 M2,ρ 1} is convex. But notice that x 2 M2,ρ 1 if and only if x M2,ρ 1 so S 1 = {x Rd : x 2 M2,ρ 1} = {x Rd : x M2,ρ 1} = M2,ρ which implies M2,ρ is convex. Proof of Proposition 4.3. Observe that R( ) is positively homogenous as the composition of positively homogenous functions. Hence x 7 1/R(x) is positively homogenous of degree 1. Then, we claim that min u Sd 1 R(u) > 0. Note that since each fi is injective, positively homogenous, and continuous, it only vanishes at 0 and is non-zero otherwise. Hence we have that for any u Sd 1, f L f1(u) = 0 so that R(u) = 0. Moreover, since R( ) is continuous and Sd 1 is compact, this minimum must be achieved by some u0 Sd 1. If R(u0) = 0, then this would contradict injectivity of each layer and positivity of G( ). Hence R(u0) > 0. We conclude by noting u 7 1/R(u) is continuous over Sd 1 since R(u) is continuous as the composition of continuous functions and it is positive over Sd 1. This guarantees K is a star body. Note that this proof also shows that R( ) is coercive, since for any sequence (xk) Rd: xk ℓ2 as k , we have by positive homogenity of R( ), R(xk) = xk ℓ2R(xk/ xk ℓ2) xk ℓ2 min u Sd 1 R(u) as k . C Experimental details C.1 Hellinger and adversarial loss comparison on MNIST We provide additional implementation details for the experiments presented in Section 3.1. For our training data, we take 10000 random samples from the MNIST training set, constituting our Dr distribution. We then add Gaussian noise of variance σ2 = 0.05 to each image, constituting our Dn distribution. The regularizers were parameterized via a deep convolutional neural network. Specifically, the network has 8 convolutional layers with Leaky Re LU activation functions and an additional 3-layer MLP with Leaky Re LU activations and no bias terms. The final layer is the Euclidean norm. This network implements a star-shaped regularizer, as outlined in the discussion of our Proposition 4.3. The regularizers were trained using the adversarial loss and Hellinger-based loss (5). We also used the gradient penalty term from [53] for both losses. We used the Adam optimizer for 20000 epochs and learning rate 10 3. The experiments were run on a single NVIDIA A100 GPU. After training the regularizer Rθ, we then use it to denoise noisy test samples y = x0 + z where x0 is a random test sample from the MNIST training set and z N(0, σ2I) with σ2 = 0.05. We do this by solving the following with gradient descent initialized at y: min x Rd x y 2 ℓ2 + λRθ(x). We ran gradient descent for 2000 iterations with a learning rate of 10 3. For the choice of regularization parameter λ, we note that in [53], the authors fix this value to be λ := 2 λ where λ := EN(0,σ2I) [ z ℓ2] as the regularizer that achieves a small gradient penalty will be (approximately) 1-Lipschitz. For the Hellinger-based network, we found that λ = 5.1 λ2 gave better performance, so we used this for recovery. This may potentially be due to the fact that the regularizer learned via the Hellinger-based loss is less Lipschitz than the one learned via adversarial regularization. We additionally hypothesize this may also result from the fact that the Hellinger loss and adversarial loss weight the distributions Dr and Dn differently, resulting in different regularity properties of the learned regularizer. As such, we decided to additionally tune the regularization strength for the adversarially trained regularizer and found λ = 0.75 λ performed better than the original fixed value. Further studying these hyperparameter choices and how they are influenced by the loss function is an interesting direction for future study. C.2 Analysis on positive homogeneity We conducted the same experiment as in Section 3.1. The main difference is that we changed the activation functions used in the network. Specifically, we use the same deep neural network as discussed in Section C.1, but considered the following four activation functions: Leaky Re LU, Exponential Linear Unit (ELU), Tanh, and the Gaussian Error Linear Unit (GELU) activation. In Table 1, we show the average PSNR and MSE of each regularizer over the same 100 test images used in Section 3.1. We see that the Leaky Re LU-based regularizer outperforms regularizers using non-positively homogenous activation functions. Leaky Re LU 22.32 0.0065 ELU 20.74 0.0090 Tanh 13.66 0.0431 GELU 18.90 0.0157 Table 1: We show the average PSNR and MSE in recovering 100 test MNIST digits using regularizers trained via the adversarial loss. Each regularizer was parameterized by a convolutional neural network utilizing a single activation function from the following options: Leaky Re LU, ELU, Tanh, or GELU. The Leaky Re LU-based regularizer achieves the highest PSNR and lowest MSE. D Additional results D.1 Stability results We now illustrate stability results showcasing that if one learns star bodies over a finite amount of data, the sequence will exhibit a convergent subsequence whose limit will be a solution to the idealized population risk. We prove this result when the constraint set corresponds to star bodies of unit-volume with γBd ker(K). Define Sd(γ, 1) := {K Sd : γBd ker(K), vold(K) = 1}. A result due to [52] shows that this subset of star bodies is uniformly bounded in the sense that there exists an R > 0 such that every K Sd(γ, 1) satisfies K RBd. Lemma D.1 (Lemma 1 in [52]). For any γ > 0, the collection Sd(γ, 1) is a bounded subset of Sd(γ). In particular, for Rγ := d+1 γd 1κd 1 where κd 1 := vold 1(Bd 1), Sd(γ, 1) {K Sd(γ) : K RγBd}. Now, to state the convergence result, for a distribution D, let DN denote the empirical distribution over N i.i.d. draws from D. The proof of this result uses tools from Γ-convergence [14] and local compactness properties of Sd(γ). Theorem D.2. Let Dr and Dn be distributions over Rd such that EDr x ℓ2, EDn x ℓ2 < and fix 0 < γ < . Then the sequence of minimizers (K N) Sd(γ, 1) of F(K; DN r , DN n ) over Sd(γ, 1) has the property that any convergent subsequence converges in the radial and Hausdorff metric to a minimizer of the population risk almost surely: any convergent (K Nℓ) satisfies K Nℓ K arg min K Sd(γ,1) F(K; Dr, Dn). Moreover, a convergent subsequence of (K N) exists. To prove this result, we first state the definition of Γ-convergence here and cite some useful results that will be needed in our proofs. Definition D.3. Let (Fi) be a sequence of functions Fi : X R on some topological space X. Then we say that Fi Γ-converges to a limit F and write Fi Γ F if the following conditions hold: For any x X and any sequence (xi) such that xi x, we have F(x) lim inf i Fi(xi). For any x X, we can find a sequence xi x such that F(x) lim sup i Fi(xi). In fact, if the first condition holds, then the second condition could be taken to be the following: for any x X, there exists a sequence xi x such that limi Fi(xi) = F(x). In addition to Γ-convergence, we also require the notion of equi-coercivity, which states that minimizers of a sequence of functions are attained over a compact domain. Definition D.4. A family of functions Fi : X R is equi-coercive if for all α, there exists a compact set Kα X such that {x X : Fi(x) α} Kα. Finally, this notion combined with Γ-convergence guarantees convergence of minimizers, which is known as the Fundamental Theorem of Γ-convergence [14]. Proposition D.5 (Fundamental Theorem of Γ-Convergence). If Fi Γ F and the family (Fi) is equicoercive, then the every limit point of the sequence of minimizers (xi) where xi arg minx X Fi(x) converges to some x arg minx X F(x). In order to show that our empirical risk functional Γ-converges to the population risk, we need to show that the empirical risk uniformly converges to the population risk as the amount of samples N . We exploit the following result, which shows that uniform convergence is possible when the hypothesis class of functions admits ε-coverings. Theorem D.6 (Theorem 3 in [68]). Let Q be a probability measure, and let Qm be the corresponding empirical measure. Let G be a collection of Q-integrable functions. Suppose that for every ε > 0 there exists a finite collection of functions Gε such that for every g G there exists g, g Gε satisfying (i) g g g, and (ii) EQ[g g] < ε. Then supg G |EQm[g] EQ[g]| 0 almost surely. We now state and prove the uniform convergence result. Theorem D.7. Fix 0 < γ < and let Dr and Dn be distributions on Rd such that EDi x ℓ2 < for i = r, n. Then, we have strong consistency in the sense that sup K Sd(γ,1) |F(K; DN r , DN n ) F(K; Dr, Dn)| 0 as N almost surely. Proof of Theorem D.7. By Lemma B.1, we have that the map K 7 F(K; Dr, Dn) is CMP - Lipschitz over Sd(γ, 1) where C = C(γ) := 1/γ2 and MP := maxi=r,n EDi x ℓ2 < . Now, consider the set of functions G := { K : K Sd(γ, 1)}. We will show that we can construct a finite set of functions that approximate K for any K Sd(γ, 1) via an ε-covering argument. This will allow us to apply Theorem D.6. First, note that Sd(γ, 1) is closed (with respect to both the radial and Hausdorff topology) and bounded as a subset of Sd(γ). Thus, by Theorem B.2, we have that the space (Sd(γ, 1), δ) is sequentially compact, as every sequence has a convergent subsequence with limit in Sd(γ, 1). On metric spaces, sequential compactness is equivalent to compactness. Since the space is compact, it is totally bounded, thus guaranteeing the existence of a finite ε-net for every ε > 0 as desired. For fixed ε > 0, construct a η-cover Sη of Sd(γ, 1) such that sup K Sd(γ,1) min L Sη δ(K, L) η where η ε/(2CMP ). Define the following sets of functions: Gη, := {( K CMP η)+ : K Sη} and Gη,+ := { K + CMP η : K Sη}. Let K G be arbitrary. Let K0 Sη be such that δ(K, K0) η. Define f = K0 +CMP η Gη,+ and f = ( K0 CMP η)+ Gη, . It follows that f f f. Moreover, by our choice of η, we have that EDi[f f] 2CMP η ε for each i = r, n. Thus, the conditions of Theorem D.6 are met and we get sup K Sd(γ,1) |F(K; DN r , DN n ) F(K; Dr, Dn)| sup K Sd(γ,1) |EDN r [ x K] EDr[ x K| + sup K Sd(γ,1) |EDN n [ x K] EDn[ x K| 0 as N a.s. D.2 Proof of Theorem D.2 We first establish the two requirements of Γ-convergence. For the first, fix K Sd(γ, 1) and consider a sequence KN K. Then we have that F(K; Dr, Dn) = F(K; Dr, Dn) F(K; DN r , DN n ) + F(K; DN r , DN n ) F(KN; DN r , DN n ) + F(KN; DN r , DN n ) |F(K; Dr, Dn) F(K; DN r , DN n )| + |F(K; DN r , DN n ) F(KN; DN r , DN n )| + F(KN; DN r , DN n ) sup K Sd(γ,1) |F(K; Dr, Dn) F(K; DN r , DN n )| + |F(K; DN r , DN n ) F(KN; DN r , DN n )| + F(KN; DN r , DN n ). (6) By Theorem D.7, we have that the first term goes to 0 as N goes to almost surely. We now show that |F(K; DN r , DN n ) F(KN; DN r , DN n )| 0 as N almost surely. To see this, observe that by Lemma B.1 we have |F(K; DN r , DN n ) F(KN; DN r , DN n )| maxi=r,n EDN i [ x ℓ2] r2 δ(K, KN). Since EDN i [ x ℓ2] EDi[ x ℓ2] < for each i = r, n and δ(K, KN) 0 as N almost surely, we attain |F(K; DN r , DN n ) F(KN; DN r , DN n )| 0. Thus, taking the limit inferior of both sides in equation (6) yields F(K; Dr, Dn) lim inf m F(KN; DN r , DN n ). For the second requirement, we exhibit a realizing sequence so let K Sd(γ, 1) be arbitrary. By the proof of Theorem D.7, for any N 1, there exists a finite 1 N -net S1/N of Sd(γ, 1) in the radial metric δ. Construct a sequence (KN) Sd(γ, 1) such that for each N, KN S1/N and satisfies δ(KN, K) 1/N. Hence this sequence satisfies KN K in the radial metric and KN Sd(γ, 1) for all N 1. Hence we can apply Theorem D.7 to get |F(KN; DN r , DN n ) F(K; P)| |F(KN; DN r , DN n ) F(K; DN r , DN n )| + |F(K; DN r , DN n ) F(K; P)| 0 as N a.s. so lim N F(KN; DN r , DN n ) = F(K; Dr, Dn). Now, we show that (F( ; DN r , DN n )) is equi-coercive on Sd(γ, 1). In fact, this follows directly from Theorem B.2, the variant of Blaschke s Selection Theorem we proved for the radial metric. Thus equi-coerciveness of the family (F( ; DN r , DN n )) trivially holds over Sd(γ, 1). As a result, applying Proposition D.5 to the family F( ; DN r , DN n ) : Sd(γ, 1) R, if we define the sequence of minimizers K N arg min K Sd(γ,1) F(K; DN r , DN n ) we have that any limit point of K N converges to some K arg min K Sd(γ,1) F(K; Dr, Dn) almost surely, as desired. The existence of a convergent subsequence of (K N) follows from Sd(γ, 1) being a closed and bounded subset of Sd(γ) and Theorem B.2. D.3 Further examples for Section 2 Example 6 (ℓ1and ℓ2-ball example). We additionally plot the underlying sets Lr and Lα n from the example in the main body in Figure 5. Example 7 (ℓ -ball and a Gaussian). We consider the following case for d = 2, when our distributions Dr and Dn are given by a Gibbs density of the ℓ -norm and a Gaussian with mean 0 and covariance Σ Rd d : pr(x) = 1 Γ(d + 1) vold(2Bℓ )e 1 2 x ℓ and pα n(x) = 1 p (2π)d det(Σ) e 1 Here, we will set Σ = [0.1, 0.3; 0, 0.1] R2 2. The star bodies induced by pr and pα n are dilations of the ℓ1-ball and the ellipsoid induced by Σ, respectively. Denote these star bodies by Lr and Ln, respectively. Then, the data-dependent star body Lr,n is defined by ρLα r,n(u) := cr x (d+1) ℓ1 cn Σ 1/2x (d+1) ℓ2 where the constants cr := R 0 td exp( t)/(Γ(d + 1) vold(2Bℓ ))dt and cn := 1 (2π)d det(Σ) R 0 td exp( 1 2t2)dt. We visualize the star bodies in Figure 6. Example 8 (Densities induced by star bodies). We give general examples of Lr,n when the distributions Dr and Dn have densities induced by star bodies. If we assume the conditions of Theorem 2.4, using the definition of ρpr and ρpn, we have that radial function ρr,n is of the form ρr,n(u)d+1 = Z 0 td(pr(tu) pn(tu))dt. Suppose for some functions ψr, ψn such that R 0 tdψi(t)dt < for i = r, n, we have pr(u) = ψr( x Kr) and pn(u) = ψn( x Kn) for two star bodies Kr and Kn. Then we have that Z 0 td(pr(tu) pn(tu))dt = c(ψr) x (d+1) Kr c(ψn) x (d+1) Kn where c(ψ) := R 0 tdψ(t)dt. Note then we must have c(ψr) x (d+1) Kr > c(ψn) x (d+1) Kn c(ψr) 1/(d+1) x Kr < c(ψn) 1/(d+1) x Kn. This effectively requires the containment property c(ψn)1/(d+1)Kn c(ψr)1/(d+1)Kr. Example 9 (Exponential densities). Suppose, for simplicity, that we have exponential densities. We need them to integrate to 1, so, for example, letting cd := Γ(d + 1), we know from [52] that R Rd e x Kdx = cd vold(K). Hence suppose our two distributions have densities pr(x) = e x Kr /(cd vold(Kr)) and pn(x) = e x Kn/(cd vold(Kn)). Then R Rd pr(x)dx = Figure 5: We plot the sets Lr, Lα n, and Lα r,n for different values of α: (Top) α = 1.3, (Middle) α = 1.6, and (Bottom) α = 2.3. In each row, the left figure shows the boundaries of Lr and Lα n, the middle figure additionally overlays the boundary of Lα r,n and the right figure shows Lα r,n := {x R2 : x Lα r,n 1}. Figure 6: We consider the example where Lr is induced by a Gibbs density with the ℓ -norm and Ln is induced by a mean 0 Gaussian distribution with covariance Σ = [0.1, 0.3; 0, 0.1]. (Left) We show the boundaries of Lr and Ln. (Middle) We additionally overlay the boundary of Lr,n. (Right) We show Lr,n := {x R2 : x Lr,n 1}. Rd pn(x)dx = 1. Here, ψr(t) = e t/(cd vold(Kr)) and ψn(t) = e t/(cd vold(Kn)). Hence, in order for the distributions Dr and Dn to satisfy the assumptions of Theorem 2.4, we require vold(Kr) > ρKn(u)d+1 vold(Kn) for all u Sd 1. In the end, our star body will have radial function of the form ρr,n(u)d+1 = cd vold(Kr) ρKn(u)d+1 for some constant cd that depends on the dimension. Example 10 (Scaling star bodies). Using the previous example, we can see how scaling a star body allows one to satisfy the conditions of the theorem. In particular, suppose we have exponential densities where Kr = αKn where α > 1. Then note that the above requirement for containment will be satisfied since for any u Sd 1, ρKr(u)d+1/ vold(Kr) = αd+1ρKn(u)d+1/(αd vold(Kn)) = αρKn(u)d+1/ vold(Kn) > ρKn(u)d+1/ vold(Kn). The above example shows that for certain distributions, we are always able to scale them in such a way that they will always satisfy the assumptions of our Theorem. Below we give a more general result along these lines. Proposition D.8. Suppose Kr and Kn are two star bodies in Rd. Then, there exists constants 0 < m < M < (depending on Kr and Kn) such that if we set we have that ρKr(u)d+1 vold(Kr) > ρα Kn(u)d+1 vold(α Kn) , u Sd 1. That is, for any two star bodies Kr, Kn Sd, the assumptions of Theorem 2 will be satisfied with the distributions Dr and Dn with densities pr(x) := e x Kr cd vold(Kr) and pn(x) := e x α Kn cd vold(α Kn), respectively. Figure 7: (First) The Gaussian mixture model in Section D.4 induces a star body LPε Here, we set ε = 0.1. Smaller values of ε induce a higher degree of nonconvexity in LPε. In the next three images, we show the resulting harmonic 2-combination M2,ρ := LPε ˆ+2 ρ 2Bd for (Second) ρ = 10, (Third) ρ = 50, and (Fourth) ρ = 100. We see that for some value of ρ (ε) := ρ > 10, M2,ρ becomes convex. By Proposition 4.2, x 7 x 2 LPε will be ρ -weakly convex. Proof of Proposition D.8. Because Kn and Kr are star bodies, they are bounded with the origin in their interior. Hence, there exists constants 0 < m < M < such that m Bd Kn, Kr MBd. Note that in order for us to have ρKr(u)d+1/ vold(Kr) > ραKn(u)d+1/ vold(αKn) = αρKn(u)d+1/ vold(Kn) u Sd 1 α < vold(Kn) d+1 u Sd 1. But by our assumptions on Kr and Kn, note that ρKr ρm Bd = mρBd = m and also ρKn ρMBd = MρBd on the sphere. This follows by the monotonicity property of the radial function for star bodies: K L ρK ρL. Thus, using these bounds achieves the desired result. D.4 A star body with weakly convex squared gauge We will consider an example of a data-dependent, nonconvex star body such that its squared gauge is weakly convex for a sufficiently large parameter ρ > 0. For a parameter ε (0, 1), consider the following 2-dimensional Gaussian mixture model Pε = 1 2N(0, Σε,1) + 1 2N(0, Σε,2) where Σε,1 := [1, 0; 0, ε] R2 2 and Σε,2 := [ε, 0; 0, 1] R2 2. Denote its density by pε(x). One can show that pε induces a valid radial function via equation (3) given by 0 r2pε(ru)du 1/3 = cε,1 Σ 1/2 ε,1 u 3 ℓ2 + cε,2 Σ 1/2 ε,2 u 3 ℓ2 where cε,i = 1 2 det(2πΣε,i) 1/2 R 0 t2e t2/2dt for i = 1, 2. Let LPε denote the star body with radial function ρPε. For a fixed ε (0, 1), we investigate the geometry of M2,ρ := LPε ˆ+2 ρ 2Bd for various values of ρ in Figure 7. In this example, we set ε = 0.1. As discussed in [52], LPε can be described as the harmonic Blaschke linear combination between the star bodies induced by the distributions N(0, Σε,1) and N(0, Σε,2). These distributions induce ellipsoids that are concentrated on one of the axes, and their resulting harmonic Blaschke linear combination is the nonconvex star body shown in the left panel in Figure 7. We see that as ρ increases, the resulting harmonic 2-combination M2,ρ eventually becomes convex for a large enough parameter ρ (ε) := ρ that depends on ε. By Proposition 4.2, this shows that the squared gauge 2 LPε is ρ -weakly convex. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: Yes, all claims made in the abstract and introduction reflect our contributions. We reiterate our main contributions in Section 1.1 and highlight in which sections each contribution is made. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: Yes, we have a section dedicated to limitations of the work and potential future directions. Please see Section 5. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: Each theoretical result provides the full set of assumptions made, and no informal results are given. A proof sketch of our main theorem (Theorem 2.4) is provided, but the full complete and correct proof is given in Section B.2. All proofs are also complete and correct. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: Yes, a full discussion regarding experimental settings (hyperparameter choices, neural network architecture, datasets, etc.) are provided in Section C.1 and in Section C.2. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [No] Justification: We provide a full discussion of experimental settings in Sections C.1 and C.2. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: These details can be found in Section C.1 and in Section C.2. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [No] Justification: Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: Yes, please see Section C.1. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: The current work is foundational theoretical research, and there are no direct societal impacts. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [NA] Justification: Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.