# elliptic_loss_regularization__05d7c3d2.pdf Published as a conference paper at ICLR 2025 ELLIPTIC LOSS REGULARIZATION Ali Hasan1,2, , Haoming Yang2, , Yuting Ng2, Vahid Tarokh2 1 Machine Learning Research, Morgan Stanley 2 Department of Electrical and Computer Engineering, Duke University {ali.hasan, haoming.yang, yuting.ng, vahid.tarokh}@duke.edu, Regularizing neural networks is important for anticipating model behavior in regions of the data space that are not well represented. In this work, we propose a regularization technique for enforcing a level of smoothness in the mapping between the data input space and the loss value. We specify the level of regularity by requiring that the loss of the network satisfies an elliptic operator over the data domain. To do this, we modify the usual empirical risk minimization objective such that we instead minimize a new objective that satisfies an elliptic operator over points within the domain. This allows us to use existing theory on elliptic operators to anticipate the behavior of the error for points outside the training set. We propose a tractable computational method that approximates the behavior of the elliptic operator while being computationally efficient. Finally, we analyze the properties of the proposed regularization to understand the performance on common problems of distribution shift and group imbalance. Numerical experiments confirm the promise of the proposed regularization technique. 1 INTRODUCTION Designing effective losses is a longstanding goal for training neural networks. Standard techniques such as Empirical Risk Minimization (ERM) are commonly used for training, but these may overfit to finite samples of data and do not provide any explicit regularization for regions outside the support of the training data. This can be an issue in the case of overparameterized neural networks, which can minimize the loss up to arbitrary accuracy for data points in the training set but have unknown behavior outside the points in the training set. For real-world deployment scenarios, this can be especially concerning where data shifts and imbalances may be present. In response to this limitation, various techniques have been proposed to regularize the loss landscape as a function of input samples of a classifier, e.g. (Hernández-García and König, 2018; Wang et al., 2021; Balestriero et al., 2022). ERM Elliptic ℓboundary < ℓinterior No Maximum Principle ℓboundary > ℓinterior Maximum Principle Figure 1: Loss surface of the two moons dataset. The scattered points illustrate the training (blue boundary) and testing (orange interior) samples. The surface represents the loss values of a welltrained classifier that classifies samples to their respective moon. Zooming in on the interior of the ERM loss surface, the training loss exceeds the loss of boundary (circled area) whereas elliptic regularization bounds the loss via the maximum principle of elliptic PDEs. In this paper, we consider a viewpoint for regularization that stems from the properties of elliptic differential operators where we impose that the loss satisfies an elliptic partial differential equation with boundary data given by the loss. By imposing a loss that satisfies this operator, we can interpret equal contribution Published as a conference paper at ICLR 2025 the expected loss through the lens of the powerful theory developed for elliptic operators and control the parameters of the operator to enforce desirable properties of the learned function approximator. Why the PDE Perspective? PDEs are useful for characterizing the joint rates in change of different variables. Imposing constraints on how the loss changes as a function of perturbations in the input space is helpful in understanding the behavior of the learned function approximator under different transformations. Additionally, minimizing the loss under the PDE constraints allows us to use the tools from PDEs to qualitatively describe properties of the loss function under the estimated function; an example is given in figure 1 where we show how the maximum principle of elliptic operators is used to bound the loss of the estimator. This provides valuable insights into the robustness of the model for regions outside of the training data. To that end, we are mainly interested in the qualitative behavior of the operator and its implications in two major problems related to generalization in machine learning: Data Shifts: How does the operator influence the loss under shifts in the data distribution? Data Imbalance: How does the operator influence the loss for underrepresented groups? More explicitly, suppose we have data pairings of (X, y) with features X and target variables y. Data shifts are data that are collected under different conditions but represent the same phenomena, i.e. (T(x), y) for different transformations T applied to X. Data imbalance is the existence of subgroups of y within the dataset that may be more or less frequently represented. By carefully choosing the parameters of the elliptic operator, we can show how behavior of the approximator can be anticipated under these regimes. To summarize, by grounding our regularization in the rich theory of elliptic PDEs, we provide a principled approach for understanding the generalization properties of neural networks, specifically: 1. We describe a new regularization that corresponds to the solution of an elliptic PDE; 2. We theoretically characterize the practical properties of this regularization through PDE theory; 3. We introduce an efficient computational approach that endows the properties of the elliptic regularization. 2 BACKGROUND Our goal is to investigate a loss that promotes functions that address the two above problems while minimally affecting performance related to our target tasks. We focus on problems of estimating a mapping between features and target variables, such as those in classification and regression tasks. To impose the desired regularity on our mapping, we can consider a few different techniques. On one hand, we can directly restrict the hypothesis class of functions from which we estimate our mapping to ensure those contain the relevant properties. On the other hand, it may be difficult to a priori know which class of functions is sufficient for the approximation task. We instead consider an approach that regularizes the loss over the data space without explicitly constraining the function class, which is implemented through a specific data augmentation. Related Work The connection between regularization and data augmentation has been studied in a number of settings (Balestriero et al., 2022; Geiping et al., 2022). A particular instance of augmentations is the mixup algorithm (Zhang et al., 2017). This algorithm has empirically shown to create more robust classifiers that are better suited to perturbations of the data, provide a level of uncertainty estimation with respect to new samples, and improve performance for classes with few samples (Zhang et al., 2017). These properties lead to many extensions of mixup; one of which is mixup E (Zou et al., 2023), which imposes an explicit directional derivative regularization to improve the mixup algorithm. The class of mixup algorithms has been theoretically analyzed from a regularization perspective (Carratino et al., 2022; Zhang et al., 2020). Various statements have been proved such as in Zhang et al. (2020) which showed that the loss at the interpolated points is an upper bound for ℓ2 adversarial training loss. The regularization and generalization effect of mixup led to development of mixup based robust optimization methods such as UMIX (Han et al., 2022), which combines mixup and weighting algorithm to achieve robust training in classification; Published as a conference paper at ICLR 2025 and c-mixup (Yao et al., 2022a), which samples based on the distribution of regression label distance for adapting mixup to the robust regression task. Motivated by the theoretical connection between regularization and data augmentation, we propose a PDE based regularization that can be implemented through a specific type of random data augmentation. This new regularization, which we refer to as elliptic regularization, is highly flexible and applies to arbitrary nonnegative loss functions in both regression and classification tasks. Elliptic regularization focuses on improving optimization under the data shifts and data imbalance scenarios which are widely studied. Algorithms such as DRO (Sagawa et al., 2019) and DORO (Zhai et al., 2021) were introduced to optimize over long-tail distributions and be robust against worst-case shifts in the data distribution. JTT (Liu et al., 2021) and UMIX (Han et al., 2022) are two popular multi-stage methods that weigh data during training based on the earlier results. While most of the above algorithms deal with classification tasks, c-mixup has emerged as a robust optimization method over regression (Yao et al., 2022a). These algorithms, including ours, are group-oblivious where the underlying subpopulation that causes data shifts and imbalance is unknown. We discuss group-informed methods in Appendix F.4. 3 PROBLEM SETUP Let s first define a region which we will consider our domain. Given observations {(X, y) X Y} with X Rd, Y Rk we will define D = DX Dy X Y as a subset of Rd Rk with nonzero Lebesgue measure. For purposes of analysis we will take D as the convex hull of X Y denoted as CX Y. Additionally, denote the empirical measure of points within a training set X Y as δX Y 1 |X Y| P|X Y| i=1 δ(X,y)i, where | | denotes the cardinality of a set. We want to minimize the risk ℓ: Dy Dy R+ over all the points in the training set for a mapping fθ : DX Dy with parameters θ. We will disregard any implicit regularization associated with the optimization procedure associated with the minimization problem (i.e. the associated regularization when minimizing such problems using gradient descent). With these in mind, we will consider different flavors of the risk minimization problem and analyze the corresponding regularization from imposing additional terms. Let us compare two frameworks for minimizing a risk ℓ, the first performs empirical risk minimization (ERM) and the second minimizes the risk after applying a class of transformations TX, Ty parameterized by ϕ which we will refer to as transformed empirical risk minimization (TRM): minθ EδX Y[ℓ(fθ(X), y)] (ERM) minθ EϕEδX Y[ℓ(fθ(T ϕ XX), T ϕ y y)] (TRM) From equation ERM we can see that this optimization, which takes place over a discrete set of points, relates to the issues we are trying to address. Since δX Y is a set of measure zero, its unclear what the interpolating behavior of f is for any subset of X or how ℓ(fθ(X), y) will behave for any subset of (X, y) D. Equation TRM on the other hand may cover more of D depending on how ϕ is chosen. Instead, we can consider a loss that is defined over all of D by modifying expectation in equation ERM in a number of different ways. One way is explicitly regularizing the class of functions fθ such that the behavior of ℓ(fθ(X), y) over all D can be anticipated (for example, making fθ Lipschitz would constrain the growth between two points). However, explicitly regularizing fθ can be difficult since constraining the function class when fθ is a neural network may require defining specific architectures. Alternatively, we can borrow techniques often used in computer vision problems (e.g. (Wang et al., 2021; Yang et al., 2023b)) by applying transformations or augmentations T ϕ X, T ϕ y with parameters ϕ to X, y such that a new objective in equation TRM is minimized for some distribution on ϕ. This also has the benefit that we now can sample transformations rather than directly constrain the class of functions. The question now remains: what class of transformations would be useful to regularize ℓ(fθ(X), y) over D? 4 ELLIPTIC LOSS LANDSCAPES Our answer to this question involves modifying the ERM problem to construct loss that satisfies the elliptic operator. We introduce all components of the operator in three different parts. To first develop intuition behind the loss, we describe an elliptic operator that has only Laplacian terms that we wish Published as a conference paper at ICLR 2025 to constrain. Using this example, we then describe how this operator can be imposed through its stochastic representation given by the Feynman-Kac theorem, providing both a connection to the data augmentation regularization initially described and a technique for implementation. Finally, using this stochastic representation, we describe the full method that includes low order terms through an importance sampling framework. We define the loss landscape as a function u(X, y) : D R+1. Our goal is to prescribe the function u(X, y) with a specified level of regularity over the data space in a way that also imbues ℓ(fθ(X), y) with regularity and thereby obtain the desired properties listed above. We propose doing this by solving a new minimization problem defined by the following equations: min θ u(X, y), (X, y) D 0 = σ 2u(X, y), (X, y) D (1) u(X, y) = ℓ(fθ(X), y), (X, y) D (2) where is taken with respect to (X, y), σ > 0 is a coefficient related to the Brownian diffusion which will be later discussed, and D represents the boundary of the domain. The key aspects of this formulation are: 1. the constraint in equation 1, corresponding to the elliptic operator, ensures a certain level of regularity within the domain beyond the training data; 2. the boundary condition in equation 2 connects the loss landscape u to the neural network loss ℓ(fθ(X), y). The regularity imposed by the PDE in equation 1 describes the behavior for regions of the space away from the observations. Specifically, equation 1 represents the steady state of the heat equation, which diffuses the boundary data from equation 2. The diffusive behavior of this equation defines the regularity of the loss based on the closest points in the observation set for regions undefined in the observation set. Additionally, the equation provides a direct connection to an Itô diffusion process which we will discuss in the next section. Interpreting the loss landscape The loss landscape u can be understood as the expected loss under a new data point (X, y) D. u satisfies an elliptic PDE with boundary data given by the points in the training set. Since the points in the training set provide all the information we have about the particular phenomena we are representing, the information at these points is diffused according to equation 1. 4.1 SOLVING THE REGULARIZED PROBLEM Computing equation 1 requires solving a PDE over a high dimensional space, which can be computationally infeasible. Fortunately, solutions to PDEs of the type in equation 1 have a Monte Carlo representation which allows scalable computation of solutions. This is formalized through the Feynman-Kac formula, which describes the relationship between expectations of stochastic processes and PDEs (Øksendal, 2003; Pardoux and Ra scanu, 2014). For self-containment, we provide additional descriptions of the Feynman-Kac formula as it relates to the problem we are solving in Appendix D. Using this form, the solution to equation 1 can be written as the following expectation u(x, y) := min θ E [ℓ(fθ(xτ), yτ) | x0 = x, y0 = y] (3) where xt, yt satisfies d(xt, yt) = σ d Wt, where τ = inf{t > 0 | (xt, yt) X Y}, σ > 0 is the diffusion coefficient, and Wt is standard Brownian motion, and sample path distribution q. Since δX Y Lebesgue measure zero, we will almost surely never hit a single point in the set, leading to a τ being infinite. To circumvent this, we place a ball of size ε > 0 around each point and define a new set (X Y)ε {x, y D | (x , y ) X Y | (x, y) (x , y ) < ε}. The hitting time is then defined under this new set as τ (ε) = inf{t > 0 | (xt, yt) (X Y)ε CX Y}, and we compute the expectation with respect to these hitting times. To relate this to equation TRM, sample paths of (xt, yt) are a particular type of transformation of the data within the domain. Using a finite number of samples, we write the empirical expectation as: j=1 ℓ fθ(x(i) τ (i) ε ), y(i) s.t. x(i) 0 , y(i) 0 = x(j), y(j). 1Note that this is different from the loss landscape definition which is given as a function of training iterations and considers perturbations in the parameter space. Published as a conference paper at ICLR 2025 where τ (i) ε = inf{t > 0 | (xt, yt)(i) (X Y)ε}. Sample paths of (xt, yt) are generated using a standard Euler-Maruyama method. Qualitatively, as ε increases, the behavior of the solution becomes piecewise constant at each point in the training set. As ε decreases, the solution becomes more diffusive where each data point becomes a point source. 4.2 GENERALIZATIONS OF THE OPERATOR The form of the PDE described in equation 1 is motivated by the diffusive properties of the second order derivative, which has the property that the loss values on the boundaries diffuse at a rate proportional to σ. Including only second order derivatives is not strictly necessary, however. The equation can be easily generalized to include lower order terms by including an advection term which acts as a deterministic term in the sense that information is being propagated according to an expected trajectory rather than diffused. The connection between the stochastic process (xt, yt) and the operator easily extends to this case. Through Girsanov s theorem, we can define the Radon-Nikodym derivative dp dq between two path measures p q to redefine the optimization criterion as up(x, y) = min θ Eq ℓ(fθ(xτ), yτ)dp dq | x0 = x, y0 = y (4) which has the effect of reweighting some of the sample paths of (xt, yt). This results in u satisfying a new PDE with lower order terms given by the structure of dp i=1 ai 2u z2 i + b u (5) which we assume is parameterized by a drift function b : D Rd+k. This reweighting term can then be interpreted with respect to the problems we are trying to solve. For example, regarding the problem of data imbalance, sample paths corresponding to the underrepresented data groups can be reweighed to greater influence the loss. This provides a connection to existing loss functions, such as the focal loss in Lin et al. (2017), where points within the training distribution are weighted according to their loss values if we choose b to be a function of ℓ(fθ(X), y). 5 PRACTICAL CONSIDERATIONS OF THE REGULARIZER Figure 2: Illustration of the loss values over a domain with 4 points on the boundary. The expected loss at point X is composed of losses at ε balls around X(i), i = 1 . . . 4. Black paths represent sample paths starting at X . There are a few important practical notes regarding applying elliptic regularization. We will first describe how to efficiently implement this loss and discuss a choice for the importance factor in equation 4 while maintaining the desired properties of the operator. We then theoretically illustrate the main properties of interest and how they are relevant to the problems presented above: The maximum principle which allows us to bound the error on the interior of the domain in Proposition 1; and, the qualitative behavior on the on the loss landscape in regression in cases of distribution shifts and class imbalance in Propositions 2 and 3. 5.1 NUMERICAL PROCEDURE In practice, it may be difficult to compute the first hitting time of the boundary within a reasonable amount of time. For example, computing the paths up to the first hitting time requires sampling for an undefined amount of time. For a practical implementation, we consider a sampling method that approximates the behavior of the first hitting time but does not require sampling unconstrained sample paths. We require that the approximation a) runs in a finite amount of time and b) maintains the important properties of the PDE. To do this, we first compute the pairwise distances between all points within a batch according Published as a conference paper at ICLR 2025 to some distance metric. Then, for each starting point (x, y), we sample an endpoint according to a discrete distribution P over endpoints with mass inversely proportional to the distances computed. Equation 3 would ordinarily be solved over all (X, y) D (e.g. by sampling uniformly over the space). We instead consider sampling over paths that connect data points using the minimum distance. Figure 2 illustrates an example of this where four points on the boundary are used to define the loss at X . The loss is minimized at all points along the paths between the source point X and the target points X(i), i = 1 . . . 4. This is done by sampling Brownian bridges between the pairings and optimize the following loss: min θ E(X ,y ) (X ,y ) P (Π)EXs,ys BBX ,y X ,y 0 ℓ(fθ(Xs), ys)ds (6) for all s, where we denote BBX ,y X ,y as a Brownian bridge sample path where denotes starting points and denotes end points and Π is the set of points in the support of δ(X Y) δ(X Y). This expectation involves solving the problem equation 3 for all points along the Brownian bridge between X , y and X , y . This minimizes the loss over all paths between the data points where the paths satisfy the distribution corresponding to the PDE between the endpoints. This loss, through an application of Dynkin s formula, corresponds to exactly solving equation 3 for points along the Brownian bridge paths, which we formally describe in Lemma 1 in Appendix B. Since the support of the bridges includes the domain, all points within the domain should eventually be sampled. Additionally, the starting point of the bridge can be an arbitrary point within the domain; we use the data points for convenience, but the theoretical properties still hold with this sampling technique. The connection to the transformation based regularization described in equation TRM is made more explicit T in equation TRM can be seen as the Brownian bridge transformation on the data points. Note that there exist some numerical error with this approximation, since it requires discretizing a continuous process, we refer to Graham and Talay (2013, Chapter 7.2) for additional details. We describe additional implementations of the bridge regularizer and how they relate to the original elliptic operator in Appendix E. 5.2 BOUNDING THE LOSS One property we would like to guarantee on our loss is, for any new point within D, what should we expect the loss to be? For example, under distribution shifts we may want to predict what the expected loss is. We do this through characterizing the loss landscape u as a function of the observed data. Since we impose that u must satisfy a PDE, we can use the properties of the solution of the PDE to bound the solution in the interior of the domain. In particular, since the PDE is elliptic, the maximum principle is satisfied. This allows us to bound the interior of the domain by the values on the boundary (which are our training points). We formalize this in the following proposition: Proposition 1 (Bounding the Loss for an Interior Point). Consider any point X, y D and suppose the function pairs u, fθ solves equation 1. Then, the expected loss u at X, y satisfies the following inequality: min X,y X Y ℓ(fθ(X), y) u(X, y) max X,y X Y ℓ(fθ(X), y). Proof. The proof follows from the fact that u satisfies an elliptic PDE that satisfies both the maximum and minimum principle. Therefore, all extreme points lie on the boundary of the domain which correspond to the points in the training set. This has an important implication insofar as we can anticipate the behavior of our loss landscape u for new, unseen points by guaranteeing the loss is bounded by the training set points. This condition holds for any point within the domain D. 5.3 EFFECTS ON DISTRIBUTION SHIFTS AND DATA IMBALANCE Finally, we discuss how to interpret the elliptic operator for the problems of interest involving distribution shift and class imbalance which motivate this work. We focus on describing the behavior for the regression case where the loss is given by the ℓ2 distance and the function f is a two-layer Re LU neural network and study the loss landscape u(X, y) satisfying the following conditions: σ 2u(X, y) = 0, X, y D (7) u(X, y) = (f(X) y)2, X, y D Published as a conference paper at ICLR 2025 with f(X) = W1Re LU(W0X) and where W0 RK d and W1 RK 1. Applied to data shifts Suppose we define a set of transformations Tφ : X X parameterized by φ Φ for some space of parameters Φ and require that these transformations only affect the input features X but do not affect the target variable y. As examples, we can think of medical imaging data collected at different hospitals under different operating conditions but of the same disease and of the same modality with the target variable being the class of disease in the image. Certain parameter values of φ may be sampled more in the dataset than others (e.g. data from larger hospitals) leading to greater uncertainty on samples from the underrepresented parameter values. We will denote an example subset with few sample points Φ . As such, the subset of X associated with Tφ(X), φ Φ will have small cardinality. Then, the expected hitting time of a point within the space is longer since fewer points are within the vicinity of a point within this subset. In the following proposition, we formalize this idea and study the changes in the loss landscape when an affine transformation is applied to the features: Proposition 2 (Expected Error Under Affine Transformations). Consider the loss landscape u satisfying equation 7 and consider the class of affine transformations T where T(X) = AT X +b T T for AT Rd d and b T Rd and (X, y) D = X Y. Suppose the true error is given by (f(X) y)2 and f is C-Lipschitz. Then, the loss landscape u satisfies u(T(X), y) 2W1W0(W1W0) τT + C| |(| | + 2ε), where ε := |f(X) y|, := AT X + b T X, and τT = inft>0{Xt, yt D | X0 = AT X + b T , y0 = y}. Proposition 2 allows us to anticipate the behavior of our loss under affine distribution shifts of our data. By applying this regularization, we can guarantee that the expected loss is no greater than a function of the parameters of the neural network used for regression. Applied to data imbalance With regards to imbalanced data in the classification setting, we consider a local structure on the points within each class. We will partition Y into two classes, Y and Y for the over-represented and under-represented classes respectively. We can show that when the loss landscape satisfying the operator in equation 7, the expected loss is greater for regions of low data density versus for regions of high data density. Following a similar argument as in Proposition 2, we note that the hitting times for Y is greater than the ones in Y with certain probability. The proposition is based on a probabilistic interpretation of how many components of the weight matrices in the neural network are greater than or equal to zero. Proposition 3 (Expected Error in Regions of Low Density). Consider the loss landscape u(X, y) satisfying equation 7. Let q(ϵ) := P( 2(f(X) y)2 2ϵW1W0(W1W0) ) be a continuous function, then an ϵ (0, 1) such that q(ϵ ) = 1 ϵ exists. Choose the smallest ϵ . Let (X , y ) and (X , y ) be two points where y is in a class that is well represented such that for τy = inft>0{(Xt, yt) D | X0 = X , y0 = y } and τy = inft>0{(Xt, yt) D | X0 = X , y0 = y } the relation τ τy ϵ holds. Suppose also that (f(X ) y )2 = (f(X ) y )2. Then, u(X , y ) u(X , y ) with probability 1 ϵ . Proposition 3 states that within regions of low data density, the loss landscape to be greater than in regions of high density, which is important in cases involving data imbalance. 6 EXPERIMENTS We present empirical results over a comprehensive set of tasks related to the proposed regularization scheme. We first empirically evaluate the bound derived from Proposition 1 and verify that the proposed regularization retains the necessary loss within the domain of interest. Next, we benchmark the elliptic regularization against another popular regularization scheme known as mixup (Zhang et al., 2017) and its variants on balanced classification and in-distribution regression. We then investigate the benefits of elliptic training on classification and regression with imbalanced group populations, imbalanced domains, and subdomain shifts. Finally, we experiment with elliptic training on imbalanced classification. Additional comparisons and ablation studies are provided in Published as a conference paper at ICLR 2025 Table 1: Classification error on CIFAR-10, CIFAR-100, best method is bolded while the second best is underlined CIFAR-10 CIFAR-100 Pre Act RN50 ERM 4.71 0.06 24.68 0.4 mixup 4.53 0.04 23.03 0.5 mixup E 3.53 0.05 20.23 0.5 Elliptic 4.15 0.43 18.95 0.3 Pre Act RN101 ERM 4.21 0.07 23.20 0.4 mixup 4.43 0.05 23.05 0.4 mixup E 3.35 0.05 18.86 0.4 Elliptic 3.56 0.2 19.03 0.6 ERM 4.24 0.1 22.20 0.1 mixup 3.03 0.09 19.38 0.1 mixup E 2.94 0.05 17.12 0.1 Elliptic 3.09 0.1 17.71 0.1 Table 2: Classification accuracy on Tiny Imagenet 200, best method is bolded while the second best is underlined Top-1 (%) Top-5(%) Pre Act RN18 ERM 54.97 0.5 72.71 0.5 mixup 54.65 0.4 72.53 0.5 mixup E 62.21 0.4 82.09 0.4 Elliptic 65.03 0.1 84.98 0.1 Pre Act RN34 ERM 57.25 0.5 72.58 0.5 mixup 57.79 0.4 76.15 0.4 mixup E 65.37 0.3 83.77 0.4 Elliptic 66.67 1.7 85.71 1.2 Pre Act RN50 ERM 55.91 0.6 73.50 0.6 mixup 54.86 0.5 73.11 0.4 mixup E 67.22 0.4 85.14 0.4 Elliptic 70.28 0.1 88.14 0.2 Appendix F with detailed data descriptions for all datasets in Appendix G. All hyperparameter settings are described in Appendix H. We note that although the implementation described in Section 5.1 specifies computing pairwise distances between data, we empirically show in Appendix F that this distance computation has a minor effect on performance of the method. 6.1 EMPIRICAL STUDY OF PROPOSITION 1 Using the two moons dataset as an example, we empirically illustrate that optimizing the neural network by solving the PDE through equation 6 enforces the bound in Proposition 1. Boundary {x, y } Interior {x, y } Figure 3: Comparison of loss for data on boundary vs within the boundary of the data space. The dashed line indicates the max loss of the boundary data. We first train a 2-layer, 4-hidden unit multilayer perceptron using the different optimization criteria of ERM, elliptic, and mixup with the generated boundary data (blue in Figure 3). We then compute the loss on the within-boundary data (orange in Figure 3). We sampled 10000 points on the boundary to train ERM and mixup while using 1000 boundary data with 10 discretized timesteps for the elliptic loss2 to sample Brownian bridges. We show in Figure 3 that only the neural network trained using equation 6 maintains the interior loss within the bounds of the boundary loss as specified in Proposition 1. 6.2 EXPERIMENTS ON REGRESSION AND CLASSIFICATION We now consider how the proposed elliptic regularization performs on standard regression and classification tasks in machine learning. While we are interested in understanding the regimes in which the regularization scheme improves the performance for the cases of group imbalance and distribution shift, we are also interested in seeing how the performance is in normal regimes without these issues. To do this, we evaluate elliptic training against suitable baselines of vanilla mixup (Zhang et al., 2017) and related state-of-the-art mixup algorithms for classification (Zou et al., 2023) and regression (Yao et al., 2022a). We then consider the performance when applying the Brownian bridge algorithm described in Section 5.1 without additional importance weighting. We showcase that using the elliptic regularization achieves comparable results to the state-of-the-art mixup algorithms for these balanced, in-distribution experiments. We benchmark on CIFAR-10 and CIFAR-100 with classification error rate in Table 1; 2Two endpoints and 8 diffusion steps. Published as a conference paper at ICLR 2025 we also evaluate on tiny-Imagenet 200 and report the top-1 and top-5 accuracy in Table 2. A number of different architectures are used in this experiment including Pre Act Res Net of various depths (He et al., 2016) and Wide-Resnet-28-10 (Zagoruyko and Komodakis, 2016). The experiment was repeated with 5 random seeds. The results suggest that elliptic training generally improves performance over the vanilla mixup algorithm and improves upon mixup E on the more complicated Tiny-Imagenet 200. The results also suggest that the elliptic regularization may be most helpful in datasets with a larger number of classes. This may be because the interpolation between classes will more likely lie on the interior of the simplex rather than on discrete corners, as is the case for datasets with fewer classes. Table 3: Average Regression (RMSE) on Regression Algorithm Air Foil NO2 C-mixup (Yao et al., 2022a) 2.88 0.187 0.524 0.006 Elliptic 2.85 0.174 0.523 0.008 For regression, we benchmarked the elliptic regularization against c-mixup algorithm on two real world tabular datasets: Airfoil Self-Noise (Airfoil) (Brooks et al., 2014) and NO2 (Kooperberg, 1997); Airfoil contains aerodynamic features of airfoil blade and related acoustic statistics and NO2 studies the concentration of NO2 particles given traffic volume and meteorological variables. Average RMSE is reported in Table 3, and the standard deviations are calculated over 10 repetitions. 6.3 ROBUSTNESS TO GROUP IMBALANCE AND DISTRIBUTION SHIFT We now explicitly apply the elliptic regularization to the loss function to evaluate the performance on data with imbalanced subpopulations and under distribution shifts. We continue to evaluate on both classification and regression tasks. For the robust classification task, we benchmark against two types of algorithms: 1) single-stage training algorithm including Focal Loss (Lin et al., 2017), CVa R-DRO and χ2-DRO (Levy et al., 2020), CVa R-DORO and χ2-DORO (Zhai et al., 2021); and 2) two-stage training algorithm including JTT (Liu et al., 2021) and UMIX (Han et al., 2022). These are groupoblivious algorithms that train without subpopulation/domain information; a full comparison with group-informed algorithms on classification tasks is provided in appendix F.4. We benchmark elliptic training with importance weighting drift b = ℓ(f(x), y) in equation 5 on Water Birds (Koh et al., 2021) and Celeb A (Sagawa et al., 2019) to evaluate performance on robustness for group imbalance. We also evaluate on the Camelyon17 (Bandi et al., 2018) dataset to examine robustness under domain shifts. To remain consistent with previous literature, we report average and worst group accuracy for Waterbirds and Celeb A across 3 random seeds and report 10 seed average accuracy for Camelyon17. The classification results are presented in Table 4. The elliptic regularization greatly improves upon the one-stage algorithms and is comparable to the two-stage algorithm Just-Train-Twice (Liu et al., 2021), which requires two separate training sessions of the model. Table 4: Robust classification accuracy under imbalance subpopulations and domain shift; best methods are bolded and best one-stage methods are underlined. Algorithm Water Birds Celeb A Camelyon17 Avg(%) Worst(%) Avg(%) Worst(%) Avg(%) Focal Loss (Lin et al., 2017) 87.0 0.5 73.1 1.0 88.4 0.3 72.1 3.8 68.1 4.8 CVa R-DRO (Levy et al., 2020) 90.3 1.2 77.2 2.2 86.8 0.7 76.9 3.1 70.5 5.1 CVa R-DORO (Zhai et al., 2021) 91.5 0.7 77.0 2.8 89.6 0.4 75.6 4.2 67.3 7.2 χ2-DRO (Levy et al., 2020) 88.3 1.5 74.0 1.8 87.7 0.3 78.4 3.4 68.0 6.7 χ2-DORO (Zhai et al., 2021) 89.5 1.0 76.0 3.1 87.0 0.6 75.6 3.4 68.0 7.5 Elliptic + IW 92.0 0.3 84.1 1.1 91.3 0.3 77.4 4.5 77.9 3.0 Two-stage: JTT (Liu et al., 2021) 93.6 NA 86.0 NA 88.0 NA 81.1 NA 69.1 6.4 Two-stage: UMIX (Han et al., 2022) 93.0 0.5 90.0 1.1 90.1 0.4 85.3 4.1 75.1 5.9 For the regression task, we compare against the c-mixup (Yao et al., 2022a) algorithm. We experiment on Skill Craft and Crime to examine domain shift and RCF-MNIST for sub-domain shift. Communities and Crime (Crime) (Redmond, 2009) and Skill Craft1 Master Table (Skill Craft) Blair et al. (2013) are real-world tabular datasets where domain shifts exist between the training and testing data. Detailed description of these dataset is provided in Appendix G. All experiments for regressions are repeated with 10 different random seeds; we report the average and worst RMSE in Table 5. Published as a conference paper at ICLR 2025 Table 5: Regression (RMSE) performance under (sub)domain shifts datasets; best method is bolded. Algorithm Skill Craft Crime RCF-MNIST Avg Worst Avg Worst Avg Worst C-mixup (Yao et al., 2022a) 6.27 0.537 8.83 1.010 0.132 0.003 0.167 0.010 0.165 0.001 0.180 0.001 Elliptic 5.97 0.283 9.17 1.150 0.132 0.003 0.164 0.01 0.162 0.002 0.178 0.002 6.4 CLASS IMBALANCE AND ROBUSTNESS TO NOISE We further evaluate the efficacy of elliptic regularization from two perspectives: imbalance classification (average and worst class) and robustness to noise. We focus on the real-world medical dataset Med-MNIST (Yang et al., 2023a) where class imbalance exists due to the prevalence of different diseases in the patient populations. We chose 4 sub-datasets of the Med-MNIST dataset to evaluate the effectiveness of the elliptic regularization. Detailed descriptions of these datasets are provided in Appendix G. To showcase the robustness induced by elliptic training, we include 50% label noise to the training data where half of the training data labels are randomly shuffled. Average and worst-class accuracy over 10 seeds are presented in Table 6, where we note the significant improvement of elliptic training and further improvement brought forth by importance weighting. The results in Table 6 indicate that elliptic training performs well in the class-imbalance tasks while still maintaining high performance under large label noise. Table 6: Average and worst class classification accuracy for selected Med-MNIST dataset, best method is bolded while the second best is underlined Breast Blood Path Organ C Method Avg(%) Worst(%) Avg(%) Worst(%) Avg(%) Worst(%) Avg(%) Worst(%) ERM 80.0 3.4 32.4 13.8 81.4 2.7 56.3 17.3 55.7 3.8 10.7 8.6 84.1 1.9 65.1 8.2 mixup 83.1 2.4 47.1 10.9 78.8 3.1 41.7 19.3 54.3 2.1 1.2 2.1 85.6 2.4 56.8 11.0 mixup E 76.5 2.8 16.7 11.3 74.2 3.6 31.9 16.9 52.5 2.0 2.7 7.6 70.4 7.6 35.6 16.8 Elliptic 87.6 1.8 67.6 4.4 85.5 1.6 60.2 14.9 62.3 2.5 26.4 10.2 87.7 0.9 65.9 3.4 + IW 87.3 0.9 67.6 3.0 84.1 4.0 68.0 13.7 62.7 1.8 20.9 6.8 88.0 0.7 66.2 4.8 7 DISCUSSION In this work, we proposed a regularization technique for the loss landscape over a defined domain where we borrow ideas from PDEs to impose properties on the landscape. We require that the loss landscape satisfies an elliptic operator and described computational tools for enforcing this. The operator allows us to bound the expected loss on the interior of the domain using the points that we observe while also providing theoretical behavior for the loss in distribution shifts and class imbalance. There exist many avenues for extending the work provided. A further study on the properties of Radon-Nikodym derivative corresponding to the parameters of the first order derivatives of the PDE should be undertaken. We chose the current parameters due to their ease in computation, but there may be more ways one can impose coefficients. Related to this, an important theoretical investigation involves studying the coefficients of the operator from a stochastic control perspective to understand the behavior under, for example, stochastic gradient descent. One can easily show that the gradient with respect to parameters satisfies another elliptic PDE over the space of inputs using similar tools of analysis, but the implications on the training behavior of the parameters is still not clear. It could be helpful to link to other theories, e.g. flatness of minima, to understand the effects on the parameter space; or other types of PDEs, such as the parabolic PDEs, to provide another avenue of study on other learning regimes (Yang et al., 2025). Finally, extensions to data supported on manifolds could provide an interesting avenue for regularization for data on complex geometries. Limitations There are a few limitations with the proposed framework. Without using the Brownian bridge technique, solving the PDE may have infinite time until hitting the first point on the boundary. It is therefore advisable to study what happens when including boundary conditions (e.g. reflecting boundaries) or imposing drift such that the diffusion is guaranteed to hit a point on the boundary. Additionally, the computation time is greater for sampling these stochastic processes than with ERM. Published as a conference paper at ICLR 2025 ACKNOWLEDGEMENT Ali Hasan and Vahid Tarokh were supported in part by the Air Force Office of Scientific Research (AFOSR) under award number FA955020-1-0397. Haoming Yang was supported in part by Air Force Office of Scientific Research (AFOSR) under award number FA9550-22-1-0315 Kartik Ahuja, Ethan Caballero, Dinghuai Zhang, Jean-Christophe Gagnon-Audet, Yoshua Bengio, Ioannis Mitliagkas, and Irina Rish. Invariance principle meets information bottleneck for out-ofdistribution generalization. Advances in Neural Information Processing Systems, 34:3438 3450, 2021. Martin Arjovsky, Léon Bottou, Ishaan Gulrajani, and David Lopez-Paz. Invariant risk minimization. ar Xiv preprint ar Xiv:1907.02893, 2019. Randall Balestriero, Leon Bottou, and Yann Le Cun. The effects of regularization and data augmentation are class dependent. Advances in Neural Information Processing Systems, 35:37878 37891, 2022. Peter Bandi, Oscar Geessink, Quirine Manson, Marcory Van Dijk, Maschenka Balkenhol, Meyke Hermsen, Babak Ehteshami Bejnordi, Byungjae Lee, Kyunghyun Paeng, Aoxiao Zhong, et al. From detection of individual metastases to classification of lymph node status at the patient level: the camelyon17 challenge. IEEE transactions on medical imaging, 38(2):550 560, 2018. Mark Blair, Joe Thompson, Andrew Henrey, and Bill Chen. Skill Craft1 master table dataset. UCI Machine Learning Repository, 2013. DOI: https://doi.org/10.24432/C5161N. Thomas Brooks, D. Pope, and Michael Marcolini. Airfoil self-noise. UCI Machine Learning Repository, 2014. DOI: https://doi.org/10.24432/C5VW2C. Luigi Carratino, Moustapha Cissé, Rodolphe Jenatton, and Jean-Philippe Vert. On mixup regularization. The Journal of Machine Learning Research, 23(1):14632 14662, 2022. Jonas Geiping, Micah Goldblum, Gowthami Somepalli, Ravid Shwartz-Ziv, Tom Goldstein, and Andrew Gordon Wilson. How much data are augmentations worth? an investigation into scaling laws, invariance, and implicit regularization. ar Xiv preprint ar Xiv:2210.06441, 2022. Carl Graham and Denis Talay. Stochastic simulation and Monte Carlo methods: mathematical foundations of stochastic simulation, volume 68. Springer Science & Business Media, 2013. Zongbo Han, Zhipeng Liang, Fan Yang, Liu Liu, Lanqing Li, Yatao Bian, Peilin Zhao, Bingzhe Wu, Changqing Zhang, and Jianhua Yao. Umix: Improving importance weighting for subpopulation shift via uncertainty-aware mixup. Advances in Neural Information Processing Systems, 35: 37704 37718, 2022. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. Alex Hernández-García and Peter König. Data augmentation instead of explicit regularization. ar Xiv preprint ar Xiv:1806.03852, 2018. Pang Wei Koh, Shiori Sagawa, Henrik Marklund, Sang Michael Xie, Marvin Zhang, Akshay Balsubramani, Weihua Hu, Michihiro Yasunaga, Richard Lanas Phillips, Irena Gao, Tony Lee, Etienne David, Ian Stavness, Wei Guo, Berton Earnshaw, Imran Haque, Sara M Beery, Jure Leskovec, Anshul Kundaje, Emma Pierson, Sergey Levine, Chelsea Finn, and Percy Liang. Wilds: A benchmark of in-the-wild distribution shifts. 139:5637 5664, 18 24 Jul 2021. Charles Kooperberg. Statlib: an archive for statistical software, datasets, and information. The American Statistician, 51(1):98, 1997. Published as a conference paper at ICLR 2025 David Krueger, Ethan Caballero, Joern-Henrik Jacobsen, Amy Zhang, Jonathan Binas, Dinghuai Zhang, Remi Le Priol, and Aaron Courville. Out-of-distribution generalization via risk extrapolation (rex). In International Conference on Machine Learning, pages 5815 5826. PMLR, 2021. Daniel Levy, Yair Carmon, John C Duchi, and Aaron Sidford. Large-scale methods for distributionally robust optimization. Advances in Neural Information Processing Systems, 33:8847 8860, 2020. Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Dollár. Focal loss for dense object detection. In Proceedings of the IEEE international conference on computer vision, pages 2980 2988, 2017. Evan Z Liu, Behzad Haghgoo, Annie S Chen, Aditi Raghunathan, Pang Wei Koh, Shiori Sagawa, Percy Liang, and Chelsea Finn. Just train twice: Improving group robustness without training group information. In International Conference on Machine Learning, pages 6781 6792. PMLR, 2021. Bernt Øksendal. Stochastic differential equations. Springer, 2003. E. Pardoux and A. Ra scanu. Stochastic differential equations, backward SDEs, partial differential equations. Stochastic Modelling and Applied Probability. Springer International Publishing, 2014. ISBN 9783319057132. Michael Redmond. Communities and crime. UCI Machine Learning Repository, 2009. DOI: https://doi.org/10.24432/C53W3X. Shiori Sagawa, Pang Wei Koh, Tatsunori B Hashimoto, and Percy Liang. Distributionally robust neural networks for group shifts: On the importance of regularization for worst-case generalization. ar Xiv preprint ar Xiv:1911.08731, 2019. Yuge Shi, Jeffrey Seely, Philip HS Torr, N Siddharth, Awni Hannun, Nicolas Usunier, and Gabriel Synnaeve. Gradient matching for domain generalization. ar Xiv preprint ar Xiv:2104.09937, 2021. Baochen Sun and Kate Saenko. Deep coral: Correlation alignment for deep domain adaptation. In Computer Vision ECCV 2016 Workshops: Amsterdam, The Netherlands, October 8-10 and 15-16, 2016, Proceedings, Part III 14, pages 443 450. Springer, 2016. Yulin Wang, Gao Huang, Shiji Song, Xuran Pan, Yitong Xia, and Cheng Wu. Regularizing deep networks with semantic data augmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(7):3733 3748, 2021. Minghao Xu, Jian Zhang, Bingbing Ni, Teng Li, Chengjie Wang, Qi Tian, and Wenjun Zhang. Adversarial domain adaptation with domain mixup. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 6502 6509, 2020. Haoming Yang, Ali Hasan, and Vahid Tarokh. Parabolic continual learning. In The 28th International Conference on Artificial Intelligence and Statistics, 2025. Jiancheng Yang, Rui Shi, Donglai Wei, Zequan Liu, Lin Zhao, Bilian Ke, Hanspeter Pfister, and Bingbing Ni. Medmnist v2-a large-scale lightweight benchmark for 2d and 3d biomedical image classification. Scientific Data, 10(1):41, 2023a. Shuo Yang, Yijun Dong, Rachel Ward, Inderjit S Dhillon, Sujay Sanghavi, and Qi Lei. Sample efficiency of data augmentation consistency regularization. In International Conference on Artificial Intelligence and Statistics, pages 3825 3853. PMLR, 2023b. Huaxiu Yao, Yiping Wang, Linjun Zhang, James Y Zou, and Chelsea Finn. C-mixup: Improving generalization in regression. Advances in Neural Information Processing Systems, 35:3361 3376, 2022a. Huaxiu Yao, Yu Wang, Sai Li, Linjun Zhang, Weixin Liang, James Zou, and Chelsea Finn. Improving out-of-distribution robustness via selective augmentation. In International Conference on Machine Learning, pages 25407 25437. PMLR, 2022b. Published as a conference paper at ICLR 2025 Sergey Zagoruyko and Nikos Komodakis. Wide residual networks. ar Xiv preprint ar Xiv:1605.07146, 2016. Runtian Zhai, Chen Dan, Zico Kolter, and Pradeep Ravikumar. Doro: Distributional and outlier robust optimization. In International Conference on Machine Learning, pages 12345 12355. PMLR, 2021. Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. ar Xiv preprint ar Xiv:1710.09412, 2017. Linjun Zhang, Zhun Deng, Kenji Kawaguchi, Amirata Ghorbani, and James Zou. How does mixup help with robustness and generalization? In International Conference on Learning Representations, 2020. Yingtian Zou, Vikas Verma, Sarthak Mittal, Wai Hoh Tang, Hieu Pham, Juho Kannala, Yoshua Bengio, Arno Solin, and Kenji Kawaguchi. Mixupe: Understanding and improving mixup from directional derivative perspective. In Uncertainty in Artificial Intelligence, pages 2597 2607. PMLR, 2023. Published as a conference paper at ICLR 2025 ELLIPTIC OPERATORS IN LOSS LANDSCAPES (SUPPLEMENTARY MATERIAL) A ALGORITHM DETAILS To fully supplement the algorithmic contributions in the main paper, we detail the elliptic training procedure that solves the PDE of equation 1 using Brownian bridges in Algorithm 1. There are several hyperparameters for the training algorithm, such as the number of bridges nb, the number of time discretization steps nt, the bridge diffusion constant σb, and the strength of importance weighting ξ. We will present some ablation studies exploring the effect of hyperparameter ξ and σb in Section F.1. We also note that depending on different tasks, there could be constraints imposed on the Brownian bridges of y. For example, for a classification task, y [0, 1], so we project the Brownian bridge of y onto the simplex by taking the absolute value of the path and normalizing it. Algorithm 1 Elliptic training algorithm Input: Data Xall, yall, related Brownian bridge hyperparameters. Initialize: neural network fθ. Compute pairwise l2-distance between all data in Xall (optional). for X, y in mini-batched Xall, yall do Obtain Brownian bridge pairs X , y by sampling inversely proportional to distance computed Sample Brownian bridges Xs, ys BBX ,y X,y for arbitrary timestep s (see Algorithm 2) Compute loss ℓas equation 6 with ℓ(f(Xs), ys) using the Euler s method over s [0, 1]. if use importance weight then Compute gradient ℓ, then compute ℓ= ℓ+ ξ ℓ end if Minimize ℓusing gradient-based optimizer end for To sample a brownian bridge, we used the Euler-Maruyama Scheme below Algorithm 2 Sampling a Brownian bridge Input: Initial condition X Rd, terminal condition X Rd, diffusion coefficient σ, number of time steps M Let t = T t0 M where t0 = 0, t M = 1, and ti+1 ti = t Sample M samples from standard normal N(0, I), denoted as Wt Integrate samples to form Brownian motion Wt = Pt s=0 σ Ws t Set W0 = 0; then compute W X t = Wt + X Output: Brownian bridge as BBt = W X t t(W X M X ) A.1 COMPUTATION COMPLEXITY The most computationally expensive part of our method is computing pairwise distance for every dataset, which is of time-complexity O(n2). However, this can be pre-computed and saved as a dictionary to be called for O(1) speed during training. We also show, through empirical experiment, the computation of distance may not be necessary. We also tested the computation time between UMIX and the proposed method on the waterbirds dataset. All hyperparameters of UMIX are taken from Han et al. (2022). Three run average training time with the proposed method is 8345 seconds while UMIX requires 11795 seconds (stage 1: 5918, stage 2: 5877). This result suggests that the additional computational cost associated with sampling bridges is marginal and the proposed elliptic regularization remains competitive in performance with a lower training time. Published as a conference paper at ICLR 2025 B.1 PROOF OF PROPOSITION 2 Proof. We will assume τ < τT , where τT is the stopping time for the Brownian motion when starting at the transformed coordinates. This is a natural assumption since the transformed data may be further from other points after transformation and result in a longer hitting time. From Dynkin s formula, we can write the solution in terms of the stopping time and the boundary condition as: E[(f(X) y) | X0 = T(X), y0 = y] = (f(T(X)) y)2 0 2(f(Xs) y)2ds | X0 = T(X), y0 = y . Since 2(f(Xs) y)2 2W1W0(W1W0) , we rewrite the solution as E[(f(X) y) | X0 = T(X), y0 = y] (f(T(X)) y)2 + 2W1W0(W1W0) τT . Compared with the original expected error given by (f(X) y)2 E[(f(X) y) | X0 = X, y0 = y] (f(X) y)2 + 2W1W0(W1W0) τ. Taking the difference, we get error 2W1W0(W1W0) τT + (f(T(X)) y)2 (f(X) y)2 2W1W0(W1W0) τT + (f(T(X)) f(X))(f(T(X)) + f(X) 2y) 2W1W0(W1W0) τT + |f(T(X)) f(X)||f(T(X)) + f(X) 2y| 2W1W0(W1W0) τT + C|AT X b T X||f(AT X + b T ) f(X) + f(X) + f(X) 2y| 2W1W0(W1W0) τT + C|AT X b T X|(|f(AT X + b T ) f(X)| + |f(X) + f(X) 2y|) 2W1W0(W1W0) τT + C|AT X b T X|(C|AT X b T X| + |f(X) + f(X) 2y|) 2W1W0(W1W0) τT + C|AT X b T X|(C|AT X b T X| + 2ε). B.2 PROOF OF PROPOSITION 3 Proof. We follow a similar proof strategy as in Proposition 2 where we use Dynkin s formula to compare the solution at different points. The first assumption follows from placing a uniform distribution over the Re LU activation functions being nonzero. Note that from the intermediate value theorem there exists ϵ (0, 1) such that q(ϵ ) = 1 ϵ . From Dynkin s formula, we can compare the expected loss at the two points by u(X , y ) = (f(X ) y )2 + E Z τy 0 2(f(Xs) ys)2ds | X0 = X , y0 = y and for the underrepresented class, u(X , y ) = (f(X ) y )2 + E Z τy 0 2(f(Xs) ys)2ds | X0 = X , y0 = y Since 2ϵ W1W0(W1W0) 2(f(Xs) ys)2 2W1W0(W1W0) , we let the lower bound correspond to the integrand of the solution at (X , y ) and the upper bound correspond to the solution at (X , y ). This gives us with probability 1 ϵ , (f(X ) y )2 + 2ϵ W1W0(W1W0) τy u(X , y ) (8) and with probability 1: u(X , y ) (f(X ) y )2 + 2W1W0(W1W0) τy . (9) Setting τy τy ϵ , the left hand side of equation 8 is greater than the right hand side of equation 9. This achieves the desired result. Published as a conference paper at ICLR 2025 B.3 PROOF OF APPROXIMATION We now illustrate how the approximation relates to the expected loss landscape through the following lemma: Lemma 1 (Approximation Error). Let ℓfθ(X) represent the loss and ℓfθ(X) represent the expected loss over the domain. Then, under the Brownian bridge loss approximation, 2 ℓfθ(X) 2ℓfθ(X) = EBB 0 ℓfθ(Xs)ds . That is, the difference between expected loss and the real is given by the value of the Brownian bridge objective. Proof. Recall that Dynkin s formula states E[ℓfθ(Xτ) | X0 = x] | {z } ℓfθ (x) = ℓfθ(x) + E Z τ 0 2ℓfθ(Xs)ds | X0 = x where the expectation is taken over Brownian motion sample paths Xt. The left hand side is equal to the expected loss landscape ℓ, the Brownian bridge loss minimizes the BB term. C LOWER ORDER DERIVATIVES As mentioned in the main text, we also have control over the parameters in the low order derivatives through an importance sampling technique. Here we will review how this is used to modify the Monte Carlo scheme such that certain points are given more weight in the loss. For additional details, please see Øksendal (2003) and Pardoux and Ra scanu (2014). For convenience, we will consider the stochastic process zt (xt, yt). C.1 EXPONENTIAL MARTINGALE Consider first the problem of the Laplace equation solved over a domain D Rd+k and its connection to Brownian motion. Specifically, we write the solution of 1 2 2u = 0, z D according to the following expectation u(z) := E [ℓ(zτ D) | z0 = z] where τ D = inf{t > 0 | zt D} and d Zt = d Wt. Now suppose we wish to write the solution for the case where there exist first order derivatives with coefficients µ(z) : D Rd+k corresponding to 1 2 2uµ + µ (z) uµ = 0, z D (10) uµ = ℓ, z D which involves the following expectation uµ(z) := E [ℓ(Zτ D) | Z0 = Z] with τ D defined as before and d Zt = µ(Zt)dt + d Wt. (11) Published as a conference paper at ICLR 2025 Instead of sampling paths of equation 11, we can instead compute the expectation with the original Brownian motions paths weighted by the exponential martingale given by d Q := exp Z τ D 0 µ (Zs)d Zs 1 0 µ µ(Zs)ds where Pµ denotes the measure associated with the sample paths of equation 11 and Q is the Wiener measure. This then gives us the solution to the new PDE in equation 10 as ℓ(Zτ D)d Pµ Since we want to minimize uµ for boundary conditions dependent on our learned mapping f, we can apply Jensen s inequality to get rid of the exp term in the exponential martingale which may be subject to numerical errors when approximated using an Euler scheme. C.2 CHOOSING THE RIGHT µ In our experiments, we choose µ to be ξ ℓ(f(x), y) for some ξ > 0 to get the following PDE 1 2 2uµ + ξ zℓ(z) (z) uµ = 0, z D. In practice, one can always make µ some function (e.g. a neural network) and separately optimize it achieve certain properties of the loss landscape. We consider the gradient of ℓdue to the connection between ℓand the uncertainty around a new point. Points with large uncertainty are then given a large weight based on the magnitude of the gradient and influence the loss more. D REVIEW OF THE FEYNMAN-KAC FORMULA The Feynman-Kac formula provides a correspondance between expectations of SDEs and PDEs. For full details, we refer to Øksendal (2003, Chapter 9) which describes the correspondence in the case of boundary value problems as used in this work. For self-containment, we provide a review of the formula here. Often, the Fenyman-Kac formula is presented in its form for solving parabolic PDEs. Let Xt satisfy the following Itô diffusion: d Xt = µ(t, Xt)dt + σ(t, Xt)d Wt and consider the following PDE: t (t, x) = u(t, x) µ(t, x) + 1 2Tr(σσ (t, x)(Hessxu)(t, x)) r(t, x)u(t, x) (12) with terminal condition u(x, T) = g(x). Then the solution to equation 12 can be represented by the following expectation: u(t, x) = E g(Xt) exp Z t 0 r(Xs)ds X0 = x . (13) In the elliptic case, consider the following PDE solved over a domain D: 0 = u(x) µ(x) + 1 2Tr(σσ (x)(Hessxu)(x)) r(x)u(x), x D. (14) with boundary condition u(x) = g(x) x D. Then the solution to equation 14 can be represented by the following expectation: u(x) = E g(Xτ) exp Z τ 0 r(Xs)ds X0 = x (15) where τ = inf{s > 0 | Xs D}. Published as a conference paper at ICLR 2025 E VARIATIONS OF THE BRIDGE COMPARED TO THE ORIGINAL PDE Here we describe a few more interpretations of the Brownian bridge augmentation as compared to the original PDE that we are solving. Recall that solving the full elliptic PDE by computing hitting times of Brownian motion is prohibitively expensive. We therefore need an algorithm that is more scalable for the typical problems in machine learning. We show now how the bridge can be interpreted through two different frameworks. To do this, we will rely on the following lemma which states that the hitting location of Brownian motion is most likely to be the location closest to the starting point for any stopping time. Lemma 2 (Hitting Location of Brownian Motion). Let {(X(i), y(I)}K i=1 be a set of boundary centroids ordered such that (X(i), y(i) (X0, y0) < (X(i+1), y(i+1) (X0, y0) for all i = 1. . . K 1 and let B(i) = {X, y| X(i) X + y(i) y ϵ} for some ϵ > 0. Then, P(Xτ B(i)) > P(Xτ B(i+1)) for all i = 1...K 1. Proof. Since Xt N(X0, y0, t), the variance of Xt increases with time. Then for any τ > 0, P(Xτ B(i)) exp( X(i) X0 + y(i) y0 /2τ). Since R ϵ 0 X(i) X0 + y(i) + ε y0 dε < R ϵ 0 X(i+1) X0 + y(i+1) y0 + ε dε, the probabilities of being in B(i) > B(i+1). Lemma 2 allows us to sample Brownian bridges with endpoints mapped to the nearest starting points rather than sample paths with random stopping times. Next we describe alternative implementations of the bridge algorithm in terms of Dynkin s theorem and an elliptic PDE with a source term. Next, we use a fixed τ for our implementation so we have a fixed time and thus a finite time algorithm for sampling paths for the expectation. Since the variance of the sample paths is given by σ, we note that there is an ambiguity between σ and τ. This is easy to show for two Brownian motions Wt, Bt with Wτ1 N(0, τ1) and Bτ2 N(0, στ2), then their distributions will be equal if σ = τ1 τ2 . We then fix τ in our experiments and optimize over σ in the implementation. E.1 INTEPRETATION VIA DYNKIN S FORMULA Dynkin s formula states the following: E[ℓ(f(Xτ), yτ) | X0 = x] := ℓ(f(X0), y0) + E Z τ 0 Aℓ(f(Xs), ys)ds | X0 = x which allows to rewrite the expectation of a stopping time with respect to the integral of the generator A of Xs, ys applied to the function ℓ. In our case, A is related to the PDE in equation 1 we are trying to impose. We can consider an approximation of this operator through the following E[ℓ(f(Xτ), yτ) | X0 = x] = E Z τ 0 lim t 0 1 t(E[ℓ(f(Xs+ t), ys+ t)] ℓ(f(Xs), ys))ds + ℓ(f(X0), y0) 1 t(E[ℓ(f(Xi+1), yi+1)] ℓ(f(Xi), yi)) + ℓ(f(X0), y0). These expectations can be computed over sample paths with a fixed τ. In general, τ can be sampled as some function of the distance between the endpoints. E.2 INTERPRETATION VIA AN ELLIPTIC PDE WITH A SOURCE TERM In the original formulation, all sources of variation arose from the boundary condition which was based on the loss value for a ball around each data point. We can instead think of extending the boundary to be at infinity and only consider instead source terms given by the loss function at different points. This corresponds to the following PDE: 2u(X, y) = ℓ(fθ(X), y), (X, y) D (16) Published as a conference paper at ICLR 2025 4 0 5 2 7 3 1 6 0.0 6 4 0 7 1 2 3 5 8 Class Index Sorted by Class Size 103601040112182 2 3 1 7 8 4 0 5 9 10 6 0.0 Figure 4: Test data uncertainty score (min-max normalized) of each class for different datasets. X-axis are class indexes sorted by data size; the errorbars present standard deviation; the number above each bar is the class size. with u(X, y) = ℓ(fθ(X), y) for (X, y) D. This corresponds to the stochastic solution given by u(X, y) := E ℓ(fθ(Xτ), yτ) + Z τ 0 ℓ(fθ(Xs), ys)ds | X0 = X, y0 = y . Using this representation, we can again sample Brownian bridges between endpoints but instead include an integral over the bridge points instead of computing the expectation over all points along the bridge path. For the new PDE in equation 16, u is no longer harmonic; rather it is subharmonic. Therefore, only the maximum principle is satisfied, not the minimum principle. This is formalized in the following proposition: Proposition 4. Using the interpretation in equation 16, the expected loss is bounded from above by the loss on the boundary of the domain. Proof. In equation 16, u is subharmonic and the maximum principle applies. E.3 MINIMIZING THE SOURCE TERM Putting the two together, we can minimize the source term to develop a computationally effective way to satisfy the PDE. We make the substitution in equation 16 such that we can guarantee the maximum principle holds. We then apply Dynkin s formula with A(ℓ(f(Xs), ys)) = ℓ(f(Xs), ys) to get min E[ℓ(f(Xτ), yτ) | X0 = x] = min ℓ(f(X0), y0) + E Z τ 0 ℓ(f(Xs), ys)ds | X0 = x . Minimizing this as described in the main text has the advantage of being computationally efficient as well as at the minimum corresponding to the original problem. We can also see how the harmonic property is being violated by checking the value of E R τ 0 ℓ(f(Xs), ys)ds | X0 = x . At the minimum, ℓ(f(Xs), ys) = 0 and we will recover the original operator given by Aℓ(f(Xs), y) = 0. F ADDITIONAL EXPERIMENTS In this section we present additional experiments including ablation studies on some parameters of elliptic training, further empirical results on the robustness of the model trained with elliptic regularization, as well as additional benchmarks for the data imbalance/distribution shift regime against group-informed algorithms. F.1 ABLATION STUDY We provide ablation studies on some hyperparameters used in elliptic training as well as the robust performance against artificial noise induced in test data. For the ablation studies, we mainly focus on the Waterbirds dataset. The hyperparameters we focus on include ξ and σb. Parameter ξ is related to the strength of the importance weighting, or also as the magnitude of the drift derived from Girsanov s Theorem; σb is Published as a conference paper at ICLR 2025 Table 7: Ablation Study on Waterbirds Avg(%) Worst(%) Baseline (ξ = 1, σ = 0.1) 92.0 0.25 84.1 1.1 No Drift (ξ = 0, σ = 0.1) 91.2 0.5 81.7 2.6 Moderate Drift (ξ = 2, σ = 0.1) 92.0 0.05 81.2 2.9 Large Drift (ξ = 10, σ = 0.1) 91.2 0.6 82.5 0.6 Large Diffusion (ξ = 1, σ = 1) 89.9 1.1 75.6 0.5 Small Diffusion (ξ = 1, σ = 0.01) 91.1 0.6 81.1 1.8 Table 8: Noisy evaluation on Waterbirds. Avg(%) Worst(%) Baseline 92.0 0.25 84.1 1.1 Noise Std = 0.1 92.1 1.6 79.8 3.0 Noise Std = 0.5 86.0 0.8 71.7 4.7 the diffusion coefficient for data during sampling of Brownian bridges. We show in Table 7 that the importance weighting is necessary to improve model performance in group-imbalanced classification setting, and a moderate diffusion coefficient should be selected to achieve the best possible result. We also show that models trained with elliptic regularization are robust against noise. We first train a model using elliptic regularization and clean data and then we add artificial Gaussian noise with different standard deviations to the test data. Table 8 showcases the robust result of average and worst-group accuracy. F.2 CONTROLLED COMPUTATION EXPENSE In this section we reduce the number of epoch for Elliptic training such that the number of function evaluation is similar to the benchmarks. Specifically we rerun the Med MNIST experiments: both ERM and mixup is trained with 500 epochs, while Elliptic training is experimented with 5 bridge timesteps and 100 epochs. We show in Table 9 that Elliptic training continues to outperform ERM and mixup with similar number of function evaluations. F.3 COMPUTING PAIRWISE DISTANCE In the main text, we noted that the pairwise distance computation does not affect the model performance. Using the med MNIST datasets, we empirically demonstratet this in Table 10 where the model s robust performance is not affected. We suspect that the randomness in the batch leads to a general diffusive behavior that enforces the elliptic operator. Table 9: Average and worst class classification accuracy for selected Med-MNIST dataset with a controlled number of function evaluations. Breast Blood Path Organ C Algorithm Avg(%) Worst(%) Avg(%) Worst(%) Avg(%) Worst(%) Avg(%) Worst(%) ERM 80.0 3.4 32.4 13.8 81.4 2.7 56.3 17.3 55.7 3.8 10.7 8.6 80.7 2.9 52.9 6.9 mixup 83.1 2.4 47.1 10.9 78.8 3.1 41.7 19.3 54.3 2.1 1.2 2.1 81.3 2.2 56.8 11.0 Elliptic 86.7 0.9 67.6 4.9 81.5 0.7 63.4 7.0 57.8 0.7 12.8 7.1 85.8 0.6 64.6 6.4 E + IW 86.9 1.6 67.6 3.8 81.5 1.0 62.9 9.7 56.3 0.3 7.1 3.5 86.0 1.0 60.4 9.2 Published as a conference paper at ICLR 2025 Table 10: Average and worst class classification accuracy for selected Med-MNIST dataset with pairwise distance computation (P-elliptic) and without (elliptic). Breast Blood Path Organ C Algorithm Avg(%) Worst(%) Avg(%) Worst(%) Avg(%) Worst(%) Avg(%) Worst(%) Elliptic 87.6 1.8 67.6 4.4 85.5 1.6 60.2 14.9 62.3 2.5 26.4 10.2 87.7 0.9 65.9 3.4 E +IW 87.3 0.9 67.6 3.0 84.1 4.0 68.0 13.7 62.7 1.8 20.9 6.8 88.0 0.7 66.2 4.8 P-Elliptic 87.3 1.8 68.1 3.9 85.3 1.8 61.9 13.2 62.7 2.3 24.7 9.7 87.9 1.0 65.6 3.3 P + IW 87.6 1.1 68.1 3.2 84.3 3.3 67.0 10.9 62.7 1.4 23.6 8.4 88.2 0.8 65.5 4.5 Table 11: Robust classification accuracy under imbalance subpopulations and domain shift (groupedinformed). Algorithm Water Birds Celeb A Camelyon17 Avg(%) Worst(%) Avg(%) Worst(%) Avg(%) IRM (Arjovsky et al., 2019) 87.5 0.7 75.6 3.1 94.0 0.4 77.8 3.9 64.2 8.1 IB-IRM (Ahuja et al., 2021) 88.5 0.6 76.5 1.2 93.6 0.3 85.0 1.8 68.9 6.1 V-Rex (Krueger et al., 2021) 88.0 1.0 73.6 0.2 92.2 0.1 86.7 1.0 71.5 8.3 CORAL (Sun and Saenko, 2016) 90.3 1.1 79.8 1.8 93.8 0.3 76.9 3.6 59.5 7.7 Group DRO (Sagawa et al., 2019) 91.8 0.3 90.6 1.1 92.1 0.4 87.2 1.6 68.4 7.3 Domain Mix (Xu et al., 2020) 76.4 0.3 53.0 1.3 93.4 0.1 65.6 1.7 69.7 5.5 Fish (Shi et al., 2021) 85.6 0.4 64.0 0.3 93.1 0.3 61.2 2.5 74.7 7.1 LISA (Yao et al., 2022b) 91.8 0.3 89.2 0.6 92.4 0.4 89.3 1.1 77.1 6.5 Elliptic + IW 92.0 0.25 84.1 1.1 91.3 0.3 77.4 4.5 77.9 3.0 F.4 ADDITIONAL COMPARISONS In this subsection we compare Elliptic regularization training with importance weighting with algorithms that leverage the group-labels during training. This set of algorithms includes IRM (Arjovsky et al., 2019), IB-IRM (Ahuja et al., 2021), V-Rex (Krueger et al., 2021), CORAL (Sun and Saenko, 2016), Group DRO (Sagawa et al., 2019), Domain Mix (Xu et al., 2020), Fish (Shi et al., 2021), LISA (Yao et al., 2022b). We show in Table 11 that our one-stage, group-oblivious training algorithm is comparable to these group-informed algorithms in all three classification datasets. Elliptic regularization achieved the best average and third-best worst group accuracy for waterbirds, results comparable to IRM and CORAL in Celeb A, and the best accuracy for Camelyon17. G DATASET DETAILS We briefly describe the contents and goals of each dataset used in our experiments, provide details about data preprocessing, and reference to related publically available codebases. G.1 BALANCED/IN-DISTRIBUTION EXPERIMENTS CIFAR10, CIFAR100, Tiny-Imagenet 200: For these well-known, small-scale image classification datasets, we followed preprocessing pipeline in Zou et al. (2023), which is publically available in the mixup E repository. Air Foil, NO2: These are publically available tabular datasets with continuous labels for regression tasks. We applied preprocessing to these two datasets following Yao et al. (2022a). The preprocessing is publically available in the c-mixup repository The Air Foil (Air Foil Self-Noise) dataset (Brooks et al., 2014) aims to predict acoustic testing results given the physical features of two and three-dimensional airfoil and the wind tunnel environment. Published as a conference paper at ICLR 2025 There are 1503 instances in Air Foil; a min-max normalization is applied to the data; there are 1003, 300, and 200 data instances for training, validation, and testing respectively. The NO2 dataset is included in Statlib (Kooperberg, 1997). The dataset studies air pollution and its relationship to traffic volume and meteorological variables. The dataset is collected by the Norwegian Public Roads Administration on Alnabru in Oslo between October 2001 and August 2003. The response variable column 1 consists of the hourly logged concentration of NO2 particles. There is no preprocessing applied; we split 200, 200, and 100 for training, validation, and testing respectively. G.2 GROUP IMABLANCE/DISTRIBUTIONAL ROBUST EXPERIMENTS G.2.1 CLASSIFICATION Waterbirds, Celeb A, and Camelyon: For these three datasets, we followed the preprocessing of Han et al. (2022). We applied the same data preprocessing, which is available in the UMIX repository. The Waterbirds dataset (Koh et al., 2021) aims to classify whether the bird is a waterbird or a landbird. This dataset has four predefined subpopulations including landbirds on land, landbirds on water, waterbirds on land, and waterbirds on water. Imbalanced group/subpopulation exist in the training set: the largest subpopulation is landbirds on land with 3,498 samples, while the smallest subpopulation is landbirds on water with only 56 samples. Celeb A (Sagawa et al., 2019) is a well-known large-scale face dataset. We predict the color of the human hair as blond or not blond. There are four imbalanced subpopulations based on gender and hair color including dark hair, female, dark hair, male, blond hair, female and blond hair, male with 71,629, 66,874, 22,880, and 1,387 training samples respectively. Camelyon17 (Bandi et al., 2018) is a pathological image dataset with over 450, 000 lymph node scans for predicting the existence of cancer tissue in a patch. The training data consists of scans from three hospitals, while the validation and test data are sampled from other hospitals. Distribution and domain shifts exist in this data as the classification requires generalization across different hospitals and coloring methods. Due to the complexity of the data, especially considering that different coloring methods are observed even in samples from the same hospital, there are no reliable subpopulation labels of Camelyon17. We applied the official split scheme of this dataset. G.2.2 REGRESSION Crime consists of demographic and economic statistics of different communities which For Skill Craft, RCF-MNIST is a simulated dataset derived from Fashion-MNIST (Yao et al., 2022a). Skill Craft, Crime: Both of these datasets are obtained from the UCI data repository. The preprocessing applied in our experiments is also publically available in the c-mixup repository Skill Craft (Skill Craft1 Master)(Blair et al., 2013) contains video game telemetry data from real-time strategy (RTS) games to explore the development of expertise. The goal is to predict input action latency based on 17 player-related parameters in the game, such as the Cognition-Action-cycle variables and the Hotkey Usage variables. Missing data are filled by mean padding on each attribute. Levels of competitors, identified through League Index" variable, are used as the domain information. We split 4,1,3 domains into training, validation, and testing subsets, which contain 1878, 806, 711 data instances, respectively. We train to predict the mean action latency of video game players in perception-action cycles; we treat League Index" as domain information and generalize predictions across different leagues. Crime (Communities And Crimes)(Redmond, 2009) is a tabular dataset that aims to predict violent crimes per capita. The 122 attributes of this dataset combine socio-economic data from the 1990 US Census, law enforcement data from the 1990 US LEMAS survey, and crime data from the 1995 FBI UCR. Following the description of Yao et al. (2022a), we applied a min-max to normalize all numeric features into [0,1]. The missing values are filled with the average values of the corresponding attributes. The 46 different State-IDs are used as the domain information, and we split the dataset into training, validation, and test sets into subsets containing 31, 6, and 9 disjoint domains. There are 1,390, 231, and 373 instances in the training, validation, and testing subsets respectively. We train to Published as a conference paper at ICLR 2025 predict the total number of violent crimes(per 100K population) of some states and aim to generalize to unseen states. RCF-MNIST The RCF-MNIST (Rotated-Colored Fashion-MNIST) is a simulated dataset based on Fashion-MNIST Yao et al. (2022a). The main goal of this simulation is to understand the model performance under spurious correlation in the training dataset. The simulation for the training set contains two steps: for a specific data sample, 1) choose a normalized angle of rotation g [0, 1], this angle of rotation is used as the label; then 2) color the normalized RGB vector of the original white pixel into [1 g, 0, g]. The color and the angle create a spurious correlation. For testing data, the coloring is reversed; hence the testing data would have the coloring of [g, 0, 1 g] for a chosen angle of rotation g. We used an 80-20 split to split the training and testing data. We train the model to predict the rotation angle of the object in the images. G.3 CLASS IMBALANCE MEDMNIST EXPERIMENTS The Med-MNIST dataset consists of medical image datasets of different scales. The details of these datasets can be found in Yang et al. (2023a). We are using the official training and testing split of these datasets. All of the images for this set of experiments are normalized to mean = 0.5 and std = 0.5. Imbalanced class exists in all of the selected datasets in our experiments including Breast MNIST, Blood MNIST, Path MNIST, and Organ CMNIST. Breast MNIST aims to use ultrasound images to identify the existence of malignant breast cancer. There are 780 images in total where a 70-10-20 data split for training, validation, and testing is applied. Among the training data, there are 147 malignant labels and 399. An official data split of 70-10-20 is applied for each dataset to construct the training, validation, and testing subsets respectively. Blood MNIST consists of microscopic images of blood cells. The 17092 images are obtained from individuals without infection, hematologic, or oncologic disease and free of any pharmacologic treatment during blood collection. The goal is to identify the 8 different cell types that exist in the dataset. An official data split of 70-10-20 is applied for each dataset to construct the training, validation, and testing subsets respectively. Path MNIST is collected from 100,000 non-overlapping image patches from hematoxylin & eosin stained histological images. The dataset aims to classify the 9 types of tissues. According to the description of the official data splits, the training and validation data is obtained from one clinical center, while the test data is curated from a different clinical center. Hence, to test performance under class imbalance and avoid additional distribution shifts, we only used the official training split as training data and validation data as testing data in our experiments. Organ CMNIST is derived from 3D computed tomography (CT) images from Liver Tumor Segmentation Benchmark (Li TS). These 3D images are then preprocessed into different views (Yang et al., 2023a). The Organ C images consist of the coronal view of these 3D images. The goal is to classify the 11 body organs from these CT scans. The official data split is applied for this dataset. There are over 10,000 samples in this dataset. G.4 PREPROCESSING Our data preprocessing follows the preprocessing pipeline of previous works such as mixup E, UMIX, and c-mixup. For CIFARs, standard normalization is applied; for tiny-imagenet, normalization, Random Crop (to height=width=64), and Random Horizontal Flip are applied; for celeb A and Camelyon17, normalization and Random Horizontal Flip is applied; waterbirds transformation includes normalization, Random Horizontal Flip and a Random Resized Crop to crop the resolution to 224*224. The normalizations applied to above image datasets use the recommended means and standard deviation for the respective datasets. Min-max scaling is applied to all features of the regression datasets. We utilize the recommended preprocessing pipeline for Med MNIST datasets, which normalizes the data to 0.5 mean and 0.5 standard deviation. Published as a conference paper at ICLR 2025 Table 12: Architecture and Hyperparameter settings for each dataset/experiments. Optim stands for optimizer; Mom stands for momentum; WD stands for weight decay; and LR stands for learning rate. Note, * means the learning rate is annealed by a factor of 10 on epoch 100 and 150. Dataset Epoch Batch Size Optim Mom WD LR nb nt σb CIFAR10/CIFAR100 200 100 SGD 0.9 10 4 1 1* 1 5 0.05 Tiny-Image Net 200 200 100 SGD 0.9 10 4 1 1* 1 5 0.05 Air Foil 100 16 Adam - - 1 2 20 5 0.05 NO2 100 32 Adam - - 1 2 10 5 0.01 Waterbirds 200 32 Adam - - 1 5 1 5 0.1 Celeb A 50 128 SGD 0.9 0.1 1 4 1 5 0.1 Camelyon17 5 32 SGD 0.9 0.01 1 5 1 5 0.1 Crime 100 16 Adam - - 1 3 10 5 0.05 Skill Craft 100 32 Adam - - 1 2 10 5 0.05 RCF-MNIST 30 64 Adam - - 1 5 10 5 0.05 Med MNIST 500 500 Adam - - 1 3 1 10 1 H HYPERPARAMETERS Now we describe the hyperparameters/architecture information for each set of experiments. For all of our bridge samples, we sampled according to a uniform discretization over the time range [0, 1], except for Med MNIST which we used time range [0, 0.01]. We use nb to denote the number of bridges, and nt to determine the number of discretization. We use σb All related training hyperparameters for all experiments are listed in Table 12. For all experiments, parameter ξ = 1 H.1 ARCHITECTURES FOR BALANCED CLASSIFICATION We followed mixup E s experiment format (Zou et al., 2023) and used Pre Act Res Net50, Pre Act Res Net101, and Wide-Res Net-28 for benchmarking on CIFAR10/CIFAR100. We used Pre Act Res Net18, Pre Act Res Net34, Pre Act Res Net50 for benchmarking on Tiny-imagenet200. These are implementations without pre-trained weights. H.2 ARCHITECTURES FOR REGRESSION We applied a 3-layer neural network with hidden layer dimension=128 and the Leaky Re LU activation with negative_slope=0.1, which is the same as c-mixup for comparing the regression result (Yao et al., 2022a). H.3 ARCHITECTURES FOR GROUP-IMBALANCE/DISTRIBUTION SHIFT CLASSIFICATION To be consistent with previous work Han et al. (2022), for waterbirds and Celeb A, we trained on Pytorch implementation of Res Net50 that is pre-trained on Image Net; for Camelyon17, we trained on Pytorch implementation of Dense Net121 without pre-trained weights. H.4 ARCHITECTURES FOR REGRESSION WITH DISTRIBUTION SHIFT For Crime and Skill Craft, we used a 3-layer neural network with hidden layer dimension of 128 and the Leaky Re LU activation function with negative_slope parameter of 0.1. For RCF-MNIST, we used Res Net18 pre-trained on Image Net, but only up to the second to last layer, as a feature extractor. Then we applied a linear layer to predict the angle. Published as a conference paper at ICLR 2025 H.5 ARCHITECTURES FOR MEDMNIST IMBALANCE CLASSIFICATION For all of the Med MNIST datasets, we used a 2-layer neural network with hidden layer dimension of 512. Batch normalization is applied to the output of the first layer.