# multiaccuracy_and_multicalibration_via_proxy_groups__1aa4e4b4.pdf Multiaccuracy and Multicalibration via Proxy Groups Beepul Bharti 1 2 Mary Versa Clemens-Sewall 3 Paul Yi 4 Jeremias Sulam 1 2 5 As the use of predictive machine learning algorithms increases in high-stakes decision-making, it is imperative that these algorithms are fair across sensitive groups. However, measuring and enforcing fairness in real-world applications can be challenging due to missing or incomplete sensitive group information. Proxy-sensitive attributes have been proposed as a practical and effective solution in these settings, but only for parity-based fairness notions. Knowing how to evaluate and control for fairness with missing sensitive group data for newer, different, and more flexible frameworks, such as multiaccuracy and multicalibration, remains unexplored. In this work, we address this gap by demonstrating that in the absence of sensitive group data, proxy-sensitive attributes can provably be used to derive actionable upper bounds on the true multiaccuracy and multicalibration violations, providing insights into a predictive model s potential worst-case fairness violations. Additionally, we show that adjusting models to satisfy multiaccuracy and multicalibration across proxy-sensitive attributes can significantly mitigate these violations for the true, but unknown, sensitive groups. Through several experiments on real-world datasets, we illustrate that approximate multiaccuracy and multicalibration can be achieved even when sensitive group data is incomplete or unavailable. 1Department of Biomedical Engineering, Johns Hopkins University, Baltimore, USA 2Mathematical Institute of Data Science, Johns Hopkins University, Baltimore, USA 3Department of Applied Mathematics & Statistics, Johns Hopkins University, Baltimore, USA 4St. Jude Children s Research Hospital, Arlington, USA 5Department of Computer Science, Johns Hopkins University, Baltimore, USA. Correspondence to: Beepul Bharti , Jeremias Sulam . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). 1. Introduction Predictive machine learning algorithms are increasingly being used in high-stakes decision-making contexts such as healthcare (Shailaja et al., 2018), employment (Freire & de Castro, 2021), credit scoring (Thomas et al., 2017), and criminal justice (Rudin et al., 2020). Although these predictive models demonstrate impressive overall performance, growing evidence indicates that they can often exhibit biases and discriminate against certain sensitive groups (Obermeyer & Mullainathan, 2019; Dastin, 2022; Li et al., 2023). For instance, Pro Publica s investigation (Angwin et al., 2022) revealed significant racial disparities in recidivism risk assessment algorithms, which disproportionately classified African Americans as high-risk for re-offending. As the deployment of these algorithms increases, regulatory bodies worldwide, including the US Office of Science and Technology Policy (OSTP) (of Science & Policy, 2022), European Union (Commission, 2021), and United Nations (United Nations Educational & , UNESCO), have emphasized the importance of ensuring that predictive algorithms avoid discrimination and uphold fairness. These concerns have led to the emergence of algorithmic fairness, a field dedicated to ensuring that predictive models do not inadvertently discriminate against sensitive groups defined by sensitive attributes such as race, age, or biological sex. Unfortunately, measuring and controlling a model s fairness can be challenging in many real-world settings, as sensitive group information is often incomplete or unavailable (Holstein et al., 2019; Garin et al.; Yi et al., 2025). In certain contexts, like healthcare, privacy and legal regulations such as the HIPAA Privacy Rule restrict access to sensitive data. In other cases, the information was not collected because it was considered unnecessary (Weissman & Hasnain-Wynia, 2011; Fremont et al., 2016; Zhang, 2018). Despite these obstacles, it remains crucial to evaluate a model s fairness before and during its deployment. This raises the question: how can we evaluate and promote fairness when sensitive group information is imperfect or missing altogether? One popular approach, widely applied in healthcare (Brown et al., 2016), finance (Zhang, 2018), and politics (Imai & Khanna, 2016) is to utilize proxy attributes in place of true attributes. Proxy methods have been immensely effective Multiaccuracy and Multicalibration via Proxy Groups in evaluating and controlling for traditional parity-based notions of fairness (Diana et al., 2022; Bharti et al., 2024; Awasthi et al., 2021; Prost et al., 2021; Zhao et al., 2022; Awasthi et al., 2020; Gupta et al., 2018; Kallus et al., 2022), such as demographic parity (Calders et al., 2009), equalized odds (Hardt et al., 2016), and disparate mistreatment (Zafar et al., 2017), which all aim to equalize model statistics across protected groups. While enforcing parity is desirable in some settings, it can also lead to undesirable trade-offs. For instance, in breast cancer screening, incidence rates vary by age, with older women generally at higher risk than younger women (Kim et al., 2025). Thus, equalizing a model s false positive rates across age groups might reduce the sensitivity of cancer detection in older women, who are more likely to have the disease. Conversely, equalizing false negative rates might lead to unnecessary biopsies by increasing false positive rates in certain groups. Instead, a more appropriate fairness criterion would be to ask that the model s risk predictions approximately reflect true probabilities within each age group. These types of domain-specific challenges have led to the development of two newer fairness notions: multiaccuracy and multicalibration (Kim et al., 2019; Gopalan et al., 2022; H ebert-Johnson et al., 2018). Instead of enforcing parity, these methods ensure that model predictions are unbiased and well-calibrated across groups. They can be applied to complex, overlapping sensitive groups such as those defined by race and gender while maintaining high predictive accuracy and ensuring the model remains useful in practice, offering significant advantages over traditional parity-based metrics. As a result, enforcing multiaccuracy or multicalibration is powerful and often preferable in many contexts (Kim et al., 2019; H ebert-Johnson et al., 2018). However, a key challenge remains: when sensitive group data is missing, how can we build provably multiaccurate and multicalibrated models leveraging proxies? Tackling this issue is essential for developing models that are fair across multiple complex groups without sacrificing accuracy or utility. In this work, we address this gap. We study how to estimate multiaccuracy and multicalibration fairness violations without access to true sensitive attributes. We show that proxy-sensitive attributes can be used to derive computable upper bounds on these violations, capturing the model s worst-case fairness. Additionally, we demonstrate that post-processing a model to satisfy multiaccuracy or multicalibration across proxies effectively reduces the worst-case fairness violations, offering practical insights. In conclusion, we demonstrate that even when sensitive information is incomplete or inaccessible, proxies can greatly help in providing approximate multiaccuracy and multicalibration protections in a useful and meaningful way. 1.1. Related Work Using proxy-sensitive attributes to measure and enforce model fairness has been extensively studied for various parity-based fairness notions. Measuring fairness. Measuring a model s true fairness through proxies has become an important area of research. Chen et al. (2019) were among the first to tackle this challenge by studying the error in measuring demographic parity using proxies derived from thresholding the Bayes optimal predictor for the sensitive attribute. Awasthi et al. (2021) focus on equalized odds, identifying key properties that proxies must satisfy to accurately estimate true equalized odds disparities. Kallus et al. (2022) further examine the ability to identify traditional parity-based fairness violations. They demonstrate that, under general assumptions about the distribution and classifiers, it is usually impossible to pinpoint fairness violations accurately using proxies. Additionally, by assuming access to the Bayes optimal predictor for the sensitive attribute, they provide tight upper and lower bounds on various fairness criteria, thereby characterizing the feasible regions for these violations. More generally, considering any proxy model instead of the Bayes optimal, Zhu et al. (2023) shows that estimating true parity-based fairness disparities using proxies results in errors proportional to the proxy error and the true fairness disparity. Most recently, Bharti et al. (2024) address a setting with more limited information compared to Zhu et al. (2023), providing computable and actionable upper bounds on true equalized odds disparities based on the proxy s misclassification error and proxy group-wise predictor statistics. Enforcing fairness. An equally important question is how to ensure fairness using proxies. Awasthi et al. (2020) examine the post-processing method for equalized odds (Hardt et al., 2016) when noisy proxies are used instead of true sensitive attributes. They show that, under conditional independence assumptions, using proxies in the post-processing method results in a predictor with reduced equalized odds disparity. Wang et al. (2020), working with a slightly different noise model, propose robust optimization approaches to train fair models using noisy sensitive features. Having proven that fairness violations are often unidentifiable, Kallus et al. (2022) take a different approach and focus on reducing the worst-case violations. Under additional smoothness assumptions they derive tighter feasible regions for fairness disparities, offering improved worst-case guarantees for fairness violations. More recently, Bharti et al. (2024) characterize the predictor that has optimal worst-case violations and provide a generalized version of Hardt et al. (2016) s method that returns such a predictor. Taking a different perspective, Lahoti et al. (2020) avoid relying on proxies altogether. Instead, they propose solving a minimax optimization problem over a vast set of subgroups, reasoning Multiaccuracy and Multicalibration via Proxy Groups that any good proxy for a sensitive feature would naturally be included in this set. Diana et al. (2022) address the problem of learning proxies that enable downstream model developers to train models that satisfy common parity-based fairness notions. They demonstrate that this entails constructing a multiaccurate proxy and introduce a general oracle-efficient algorithm to learn such proxies. 1.2. Our Contributions There exists a rich line of work that studies how to evaluate and enforce parity-based notions of fairness when sensitive attribute data is missing via proxies. These, however, do not extend to settings where multiaccuracy and multicalibration are more appropriate, limiting their applicability in datascare regimes. In this work, we address this issue. Our main contributions are the following: 1. We study the problem of estimating multiaccuracy and multicalibration violations of a predictive model without access to sensitive group information. 2. We derive computable upper bounds for multiaccuracy and multicalibration violations using proxy-sensitive attributes. 3. We show that post-processing a model to satisfy multiaccuracy and multicalibration across proxies reduces the worst-case violations, allowing us to provide meaningful fairness guarantees without access to sensitive group data. Organization. The remainder of the paper is structured as follows. In Section 2, we introduce the necessary notation and formalize the setting. Section 3 provides background on multiaccuracy and multicalibration. Our main theoretical results are presented in Section 4 and Section 5, where we establish computable upper bounds on the multiaccuracy and multicalibration violations and demonstrate how to minimize them. Experimental results are detailed in Section 6. Finally, in Section 7 we discus the implications of our work and provide closing remarks. 2. Preliminaires Notation. We consider a binary classification setting 1 with a data distribution D supported on X Z Y, where X Rd and Z RK represent a d-dimensional feature space and K-dimensional sensitive group space, respectively, and Y = {0, 1} denotes the binary label space. For an individual represented by the pair (X, Z), X is a vector of features and Z is a vector of sensitive features (e.g., race, biological, age). We denote G = {g : X Z 7 {0, 1}} as the set 1One can also extend this to a K-class problem using a one-vsall approach. of functions that define complex, potentially intersecting groups in X Z. For any g G, g(X, Z) = 1 indicates that the individual (X, Z) belongs to group g. For example, let X1 be an individual s credit score and let Z1 and Z2 represent the individual s age and membership in the African American group. Then, g(X, Z) = 1{X1 > 700 Z1 > 40 Z2 = 1} specifies the group of all African Americans over the age of 40 with a credit score over 700. In this way, it is easy to define arbitrary overlapping groups defined by the intersection of basic attributes and other features. Finally, in our setting, a model is a function f : X R that maps from the feature space to some discrete domain R [0, 1]. We denote its image as Im(f) = {f(X) : X X} and assume that |Im(f)| < . Problem Setting. In this work, the primary objective is to assess whether a model f is fair with respect to a set of sensitive groups G without having access to the functions in G to determine group membership. Formally, and similar to previous work (Awasthi et al., 2021; Kallus et al., 2022), we consider a setting where we do not have access to samples (X, Z, Y ) from the complete distribution D and thus we are unable to use the true set of grouping functions G since their domain is supported X Z. Instead, we assume access to a sufficient number of samples (X, Y ) from DXY, the marginal distribution over X Y, which allow us to reliably evaluate the overall performance of the predictor f via its mean-squared error MSE(f) = E (X,Y ) DXY[(Y f(X))2]. (1) Additionally, we assume a proxy developer that has access to samples (X, Z) from DXZ, the marginal distribution over X Z. They provide a set of learned proxy functions ˆG = {ˆg : X 7 {0, 1}} for G that only use features X, allowing us to determine proxy group membership. Moreover, via the proxy developer, we know how well any proxy ˆg ˆG approximates its associated true g G through its misclassification error, err(ˆg) = P (X,Z) DXZ[ˆg(X) = g(X, Z)]. (2) Through this setup, we are modeling real-world situations where we lack information about individuals basic sensitive attributes, such as sex and race, preventing us from accurately identifying individuals membership in complex intersecting groups (for example, white women over the age of 40). Instead, we rely on proxies to (approximately) represent all groups. Importantly, we do not make stringent, unverifiable assumptions about these proxy functions, unlike other studies (Prost et al., 2021; Awasthi et al., 2020; 2021); we only consider knowing their error rates, err(ˆg). With our setting fully described, we now turn to our main objective: assessing the fairness of the model f with re- Multiaccuracy and Multicalibration via Proxy Groups spect to G. In this work we focus on two fairness concepts multiaccuracy and multicalibration which we now formally define. 3. Multiaccuracy and Multicalibration Multiaccuracy. Multiaccuracy (MA) is a notion of fairness originally introduced by (Kim et al., 2019; H ebert-Johnson et al., 2018). For any sensitive group g G, MA evaluates the bias of a model f, conditional on membership in g via AED(f, g) = ED[g(X, Z)(f(X) Y )] (3) and requires that AED(f, g) be small for all groups g G. Definition 3.1 (Multiaccuracy (Kim et al., 2019)). Fix a distribution D and let G be a set of groups. A model f is (G, α)-multiaccurate if AEmax D (f, G) := max g G AED(f, g) α (4) (G, α)-MA requires that the predictions of f be approximately unbiased overall and on every group. Building on this, (H ebert-Johnson et al., 2018) introduced a stronger notion of group fairness known as multicalibration (MC), which demands unbiased and calibrated predictions. Central to evaluating MC is the expected calibration error (ECE) for a group g G ECED(f, g) = E v Df E[g(X, Z)(f(X) Y )|f(X) = v] where Df is the distribution of the predictions made by f under D. MC requires that ECED(f, g) be small for all groups g G. Definition 3.2 (Multicalibration (H ebert-Johnson et al., 2018)). Fix a distribution D and let G be a set of groups. A model f is (G, α)-multicalibrated if ECEmax D (f, G) := max g G ECED(f, g) α. (5) This is an l1 notion of MC as studied in (Gopalan et al., 2022). There also exist l2 (Globus-Harris et al., 2023b) and l (H ebert-Johnson et al., 2018) variants. (G, α)-MC requires that f s predictions be approximately calibrated on all groups defined by G. Note, MC is stronger than MA because intuitively MC requires MA on every level set of f. Having presented these definitions, the problem we face is now clear: Ideally, we would like to evaluate AEmax D and ECEmax D . However, this requires access to samples (X, Z, Y ) D and the functions G, neither of which we assume to have. As alluded to before, no method exists that can guarantee let alone, correct that a predictor is (G, α)- MA/MC in the absence of ground truth groups G. Fortunately, in the following sections, we demonstrate that with proxies, one can successfully circumvent these limitations and still provide meaningful guarantees. 4. Bounds on Multigroup Fairness Violations We will now demonstrate that is still possible to derive computable and useful upper bounds for the MA and MC violations of a model f across the true groups G, even in the absence of true group information. Our first result provides computable upper bounds on AED(f, g) and ECED(f, g) for any group g. Lemma 4.1. Fix a distribution D and model f. For any group g and its corresponding proxy ˆg, AED(f, g) F(f, ˆg) + AED(f, ˆg) (6) ECED(f, g) F(f, ˆg) + ECED(f, ˆg) (7) F(f, ˆg) = min err(ˆg), p MSE(f) err(ˆg) . (8) Furthermore the upper bounds are tight. A proof of this result is provided in Appendix B.1. We now make a few remarks. First, the upper bounds on the true MA/MC violations are tight: there exist non-trivial joint distributions over (f(X), g(X), ˆg(X), Y ) specifically, those for which err(ˆg) > 0 and MSE(f) > 0 such that the bounds are attained with equality. Please see Appendix B.1 for a comprehensive discussion regarding these statements. Second, in our setting, both bounds can be directly evaluated because (1) we know err(ˆg) for all g G via the proxy developer, who has access to G and a sufficient number of samples (X, Z) from DXZ to compute err(ˆg); and (2) we can reliably compute MSE(f), AED(f, ˆg), and ECED(f, ˆg) because we have access to a sufficient number of samples (X, Y ) from DXY. This result aligns with intuition, showing how the relationship between a proxy ˆg and the model f constrains the maximum possible values of AED(f, g) and ECED(f, g). If the proxy ˆg is highly accurate in that it predicts the true group g better than f predicts the true label Y , i.e. if err(ˆg) < MSE(f), (9) then F(f, ˆg) = err(ˆg), and the true violations AED(f, ˆg) and ECED(f, ˆg) are approximately bounded by their proxy estimates. Conversely, if f is highly accurate in predicting the label Y but the proxy is weaker, so that MSE(f) < err(ˆg), (10) Multiaccuracy and Multicalibration via Proxy Groups Algorithm 1 Multiaccuracy Regression 1: Input: Initial model f and set of groups G 2: Solve ˆf arg min λ R|G| MSE( ˆf) s.t. ˆf(X, Z) = f(X) + X g G λg g(X, Z) 3: Return ˆf then F(f, ˆg) = p MSE(f) err(ˆg), and one can attain a better bound by considering a factor of p MSE(f) err(ˆg) instead of err(ˆg). Naturally, as MSE(f) decreases, the maximum possible values of AED(f, ˆg) and ECED(f, ˆg) decreases as well. Most importantly, with this result, we can provide an upper bound for the true MA and MC violations of f across G. Theorem 4.2. Fix a distribution D and model f. Let G be a set of true groups and ˆG be its associated set of proxies. Then, f is (G, β(f, ˆG))-MA and (G, γ(f, ˆG))-MC where β(f, ˆG) = max ˆg ˆG F(f, ˆg) + AED(f, ˆg) (11) γ(f, ˆG) = max ˆg G F(f, ˆg) + ECED(f, ˆg) (12) This result directly follows from Lemma 4.1, with a complete proof provided in Appendix B.2. It is particularly valuable because, even without directly evaluating the quantities of interest, AEmax D (f, G) and ECEmax D (f, G), we can still evaluate these worst-case violations, which offers practical utility. For instance, if we need to ensure that f is (G, α)- MC before deployment, we can proceed confidently even without direct access to G, provided that γ(f, ˆG) < α. Conversely, if the worst-case violations are large, this suggests that f may potentially be significantly biased or uncalibrated for certain groups g. In a scenario where the worst-case violations are large, as model developers, we should pause deployment and ask: if it is the case that β(f, ˆG) or γ(f, ˆG) are large, can we reduce them such that we can provide better guarantees on the worst-case MA and MC violations of f? We show that it is possible in the following section. 5. Reducing Worst-Case Violations The results from the previous section allow us to upper bound the MA and MC violations using proxies. Now, we show that these violations can be provably reduced, yielding stronger worst-case guarantees. Recall that we have a fixed set of proxies ˆG, a model f, and access to samples (X, Y ). Algorithm 2 Multicalibration Boosting 1: Input: Initial model f, set of groups G, and α > 0 2: Let m = 1 α , t = 0, f0 := f 3: while max g G P[g(X, Z) = 1] E[ 2 v,g|g(X, Z) = 1] > α pg,v = P[g(X, Z) = 1, ft(X) = v] (vt, gt) = arg max m],g G pg,v 2 v,g α Svt,gt = {X X : ft(X) = v, gt(X, Z) = 1} vt = E[Y |ft(X) = vt, gt(X, Z) = 1] v t = arg min ( v t, if X Svt,gt ft(X) otherwise 5: end while 6: Let T := t, ˆf := f T 7: Return ˆf Within the bounds β(f, ˆG) and γ(f, ˆG), there are only two quantities we can modify by adjusting f: the mean squared error MSE(f) and either AED(f, ˆg) or ECED(f, ˆg). Since we lack access to the true grouping functions G and samples (X, Z) DXZ, we cannot reduce the proxy errors err(ˆg). Thus, the question is, what modifications can be made to f such that the updated ˆf has smaller bounds? We answer this question for the bound on MC in the following theorem. Theorem 5.1. Fix a distribution D, initial model f, and set of proxy groups ˆG. If a model ˆf satisfies ECEmax D ( ˆf, ˆG) < min ˆg ˆG ECED(f, ˆg) (13) MSE( ˆf) MSE(f) (14) then, it will have a smaller worst-case MC violation, i.e. γ( ˆf, ˆG) γ(f, ˆG). (15) An identical result for MA is given in Appendix A and proofs for both results are provided in Appendix B.3. This results simply states that if we can obtain a new model ˆf that 1) is ( ˆG, α)-MC at level α = minˆg ˆG ECED(f, ˆg) and 2) has smaller MSE, then it is guaranteed to have a smaller worst-case violation. Fortunately, both objectives can be nearly achieved using Algorithm 1, proposed by Gopalan et al. (2022), and Algorithm 2, introduced by Roth (2022), which produce MC and MC predictors, respectively. Multiaccuracy and Multicalibration via Proxy Groups Group err(ˆg) Black Women 0.027 White Women 0.122 Asian 0.060 Seniors 0.000 Women 0.000 Multiracial 0.047 Table 1. Proxy errors for ACSIncome Group err(ˆg) Black Women 0.005 White Women 0.046 Asian 0.000 Multiracial 0.000 Black Adults 0.000 Women 0.079 Table 2. Proxy errors for ACSPub Cov Group err(ˆg) Women 0.027 White 0.092 Asian 0.068 Black 0.039 Asian Men 0.039 Black Women 0.020 Table 3. Proxy errors for Che Xpert Algorithm 1 outlines a simple algorithm for MA. The following theorem establishes that the algorithm produces a model ˆf that satisfies MA while also guaranteeing an improvement or no deterioration in MSE. Theorem 5.2. Fix a distribution D, predictor f, and set of groups G. Algorithm 1 returns a model ˆf that is (0, G)-MA. Moreover, MSE( ˆf) MSE(f). (16) A proof of this result can be found in (Detommaso et al., 2024), with finite-sample guarantees discussed in (Roth, 2022). Algorithm 1 updates the model ˆf by solving a standard linear regression problem, where the features are the predictions of the initial model f and grouping functions g. While Algorithm 1 solves a optimization problem to generate a MA model in a single step, Algorithm 2 is an iterative method to ensure MC. Algorithm 2 starts by checking if f is α-MC via the group average squared calibration error E[ 2 v,g|g(X, Z) = 1] (17) where v,g = E[Y f(X)|f(X) = v, g(X, Z) = 1]. If this exceeds α, it identifies the conditioning event where the calibration error is the largest and refines f s predictions. It iterates like this until convergence and this process returns a new model ˆf that is α-MC and has an MSE that is close, potentially even lower, than the MSE of the initial model f. Theorem 5.3. Fix a distribution D, predictor f and set of groups G. Algorithm 2 stops after T < 4 α2 rounds and returns model ˆf that is ( α, G)-MC. Moreover, MSE( ˆf) MSE(f) + (1 T)α2 A proof, along with with finite-sample guarantees can be found in (Globus-Harris et al., 2023a; Roth, 2022). Having introduced these algorithms for obtaining multiaccurate and multicalibrated models ˆf, we now have a direct path to reducing worst-case violations. By applying Algorithm 1 or Algorithm 2 (at an appropriate level α) to our initial model f using the proxies ˆG that is, enforcing multiaccuracy or multicalibration with respect to the proxies we can systematically reduce worst-case violations on the true groups G. This simple yet effective approach ensures allows us to still provide meaningful fairness guarantees and reliable predictions across sensitive subpopulations. 6. Experimental Results We illustrate various aspects of our theoretical results on two tabular datasets, ACSIncome and ACSPublic Coverage (Ding et al., 2021), as well as on the Che Xpert medical imaging dataset (Irvin et al., 2019). For the ACS datasets, we use a fixed 10% of the samples as the evaluation set. The remaining 90% of the data is split into training and validation sets, with 60% used for training the model f and proxies ˆG and 30% for adjusting f. All reported results are averages over five train/validation splits on the evaluation set. For Che Xpert, we use the splits provided by (Glocker et al., 2023) for training, calibration, and evaluation. The results and metrics are computed on the evaluation set. We report results for MC in the main body of the paper and defer those for MA to Appendices C and E because MC is a stronger notion that implies MA. The code necessary to reproduce these experiments is available at https:// github.com/Sulam-Group/proxy_ma-mc. 6.1. ACS Experiments For the two following tabular data experiments we use the ACS dataset, a larger version of the UCI Adult dataset. In particular, we use the 2018 California data, which contains approximately 200,000 samples. We follow Hansen et al. (2024) and define multiple sensitive groups G using basic sensitive attributes Z (e.g., sex and race, which model developers aim not to discriminate towards), along with certain features X (e.g., age). Examples of groups g G include white women and black adults. For both experiments, we simulate missing sensitive attributes by excluding some Zi from the data we use to train our predictive model f. Instead, with an auxiliary dataset of samples (X, Z) and the true set of grouping functions G, we obtain a set of proxy functions ˆG to approximate G. For our initial model f, we report the worst-case violations, which we can evaluate in our setting. To demonstrate that Multiaccuracy and Multicalibration via Proxy Groups MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound (a) ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACS Income. Here, f is a decision tree. MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound (b) ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and updated model fadj on ACSPubcov. Here, f is a decision tree. enforcing MA and MC with respect to ˆG provably reduces our upper bounds, we apply Algorithm 1 and Algorithm 2 to obtain an adjusted predictor fadj and report its worstcase violations as well. Additionally, for both the initial model f and adjusted model fadj we report the AE and ECE, with respect to the true groups g G, along with their maximums, AEmax and ECEmax. Recall that in our setting, we cannot actually evaluate these quantities but we report them to showcase that they lie under our bounds, illustrating the validity of our theoretical results. For these tabular experiments, we model f with a logistic regression model, a decision tree, and Random Forest. We report the results for the decision tree in the following sections and defer the remainder to Appendices C and E. 6.1.1. ACSINCOME For this experiment, we consider the task of predicting whether working adults living in California have a yearly income that exceeds $50,000. Examples of features X include occupation and education. To simulate missing sensitive attributes, we exclude the race attribute. In Table 1, we report the errors of the learned proxies ˆG for specific groups. For groups that do not depend on race, such as seniors and women, their respective proxies ˆg are perfectly accurate, exhibiting zero misclassification error. However, for groups like multiracial and white women, the proxies exhibit some error, albeit small. This proves to be useful in providing meaningful guarantees on how multiaccurate and multicalibrated the model f may potentially be with respect to the true (but unobserved) groups G. Figure 1a showcases the utility of our bounds. Notably, the worst-case violation (dotted red line) allows us to certify that the initial model is approximately 0.21-multicalibrated with respect to the true groups G. This is indeed practically useful as it enables practitioners to obtain a certificate on the MC violation without having access to the true sensitive group information. Additionally, the right-hand graph in Figure 1a highlights the benefit of applying Algorithm 2 and multicalibrating the initial model f with respect to the proxy groups ˆG. After adjusting f, the upper bound decreases, allowing us to certify that the resulting model fadj is approximately 0.13-multicalibrated a substantial improvement of 38%. 6.1.2. ACSPUBLICCOVERAGE (ACSPUBCOV) In this experiment, we consider the task of predicting whether low-income individuals (< $30,000) , not eligible for Medicare, have coverage from public health insurance. Examples of features X include age, education, income, and more. To simulate missing sensitive attributes, we exclude the sex attribute. In Table 2, we report the errors of the learned proxies ˆG for specific groups. Notably, for groups independent of sex, such as asian and multiracial, the proxies ˆg are perfectly accurate, exhibiting zero misclassification error. However, for groups like black women and white women, the proxies exhibit some error, though they are small. This arises because, although the proxies functions ˆg are not explicit functions of sex attribute, there exists a feature, fertility, that indicates whether an individual has given birth within the past 12 months and serves as a good predictor of sex. This is a prime example of a real-world setting where, even though sensitive attributes may be missing, strong proxies can still enable us to determine true sensitive group membership with high accuracy. In Figure 1b, we show the results of applying Algorithm 2 to multicalibrate the initial model f with respect to the proxies ˆG. Notably, our approach allows us to certify that the initial model is approximately 0.25-MC with respect to the true groups G despite not having access to them. This result highlights the utility of our method, as it enables practitioners to obtain performance guarantees without needing the true group information. Furthermore, after applying Algorithm 2 to f, the resulting model fadj is certified to be approximately 0.09-MC, thereby providing a stronger guarantee. Multiaccuracy and Multicalibration via Proxy Groups MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound (a) ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on Che Xpert. Here f is a decision tree trained on embeddings of a Dense Net-121 model pretrained on Image Net. MC Violations and Bounds (Before Adjustment) True Violation Upper Bound MC Violations and Bounds (Before Adjustment) True Violation Upper Bound (b) ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original, unadjusted, logistic regression model (left) and end-to-end trained Dense Net-121 model (right). 6.2. Che Xpert Che Xpert is a large public dataset for chest radiograph interpretation, with labeled annotations for 14 observations (positive, negative, or unlabeled) including cardiomegaly, atelectasis, and several others. The dataset contains self-reported sensitive attributes including race, sex, and age. Following the set up of Glocker et al. (2023), we work with a sample containing a total of 127,118 chest X-ray scans and consider the task of predicting the presence of pleural effusion in the X-rays. We consider all 14 groups that can be made from conjunctions of sex and race. Examples of groups g G include black men, asian women, white women, etc. In this example, we assume that we do not have direct knowledge of patient s self-reported sex or race when training or evaluating our model f (as is common for privacy reasons). Instead, with an auxiliary dataset with samples (X, Z) we use the X-rays to learn proxy functions for sex and race. We then use them to construct proxies for all conjunctions as well. In Table 3, we report the proxy errors for specific groups. We consider three different models for f. The first is a decision tree classifier trained on features extracted from a Dense Net-121 model (Huang et al., 2017) pretrained on Image Net (Deng et al., 2009). The second is a linear model (Breiman, 2001) trained on the same features. The third is a Dense Net-121 trained end-to-end on the X-rays. Figure 2a illustrates the results for the decision tree model. Before any adjustments, our worst-case violation serves as an early warning that the model f may be significantly uncalibrated on certain groups, with a violation as large as α 0.42. In a medical setting like this, such a finding is crucial, as it indicates that our predictions could be either overly confident or underconfident on sensitive groups. On the other hand, the right-hand graph of Figure 2a demonstrates the practical benefit of applying Algorithm 2 to multicali- brate the initial model f with respect to our highly accurate proxies ˆG. After a straightforward adjustment, the upper bound on the worst-case violation decreases significantly, certifying that the adjusted model fadj is approximately 0.13MC with respect to the true groups. Figure 2b presents the results for the logistic regression and fully-trained Dense Net models. In these cases, the worstcase violations for both models indicate that they are guaranteed to be approximately 0.11 and 0.12-multicalibrated with respect to the true groups. Notably, both models are approximately 0.03-multicalibrated with respect to the proxies. Thus, further adjustments provide negligible improvements. 7. Conclusion In this work, we address the challenge of measuring multiaccuracy and multicalibration with respect to sensitive groups when sensitive group data is missing or unobserved. By leveraging proxy-sensitive attributes, we derive actionable upper bounds on true the multiaccuracy and multicalibration violations, offering a principled approach to assessing worstcase fairness violations. Furthermore, we demonstrate that adjusting models to be multiaccurate or multicalibrated with respect to proxy-sensitive attributes can significantly reduce these upper bounds, thereby providing useful guarantees on multiaccuracy and multicalibration violations for the true, but unknown, sensitive groups. Through empirical validation on real-world datasets, we show that multiaccuracy and multicalibration can be approximated even in the absence of complete sensitive group data. These findings highlight the practicality of using proxies to assess and enforce fairness in high-stakes decision-making contexts, where access to demographic information is often restricted. In particular, we illustrate the practical benefit of enforcing multiaccuracy and multicalibration with respect to proxies, providing practitioners with a simple and effective tool to improve fairness in their models. Multiaccuracy and Multicalibration via Proxy Groups Naturally, multiaccuracy and multicalibration may not be the most appropriate fairness metrics across all settings. Nonetheless, whenever these notions are relevant, our methods offer, for the first time, the possibility to provide certificates and strengthen them without requiring access to ground truth group data. Lastly, note that without our recommendations of correcting for worst-case fairness with proxies, models trained on this data can inadvertently learn these sensitive attributes indirectly and base decisions on them, leading to potential negative outcomes. Our results and methodology prevent this by providing a principled approach to adjust models and reduce worst-case multiaccuracy and multicalibration violations. Impact Statement In this work, we propose an effective solution to a technical problem: estimating and controlling for multiaccuracy and multicalibration using proxies. While proxies can be controversial and pose risks such as reinforcing discrimination or compromising privacy predictive models often learn sensitive attributes indirectly, leading to unintended harm. When used carefully, proxies can help mitigate these risks, as we demonstrate in this work. However, deploying proxies in real-world scenarios requires a careful evaluation of trade-offs through discussions with policymakers, domain experts, and other stakeholders. Ultimately, proxies should be employed responsibly and solely for assessing and promoting fairness. Acknowledgements This work was supported by NIH award R01CA287422. Angwin, J., Larson, J., Mattu, S., and Kirchner, L. Machine bias. In Ethics of data and analytics, pp. 254 264. Auerbach Publications, 2022. Awasthi, P., Kleindessner, M., and Morgenstern, J. Equalized odds postprocessing under imperfect group information. In International Conference on Artificial Intelligence and Statistics, pp. 1770 1780. PMLR, 2020. Awasthi, P., Beutel, A., Kleindessner, M., Morgenstern, J., and Wang, X. Evaluating fairness of machine learning models under uncertain and incomplete information. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp. 206 214, 2021. Bharti, B., Yi, P., and Sulam, J. Estimating and controlling for equalized odds via sensitive attribute predictors. Advances in neural information processing systems, 36, 2024. Breiman, L. Random forests. Machine learning, 45(1): 5 32, 2001. Brown, D. P., Knapp, C., Baker, K., and Kaufmann, M. Using bayesian imputation to assess racial and ethnic disparities in pediatric performance measures. Health services research, 51(3):1095 1108, 2016. Calders, T., Kamiran, F., and Pechenizkiy, M. Building classifiers with independency constraints. In 2009 IEEE international conference on data mining workshops, pp. 13 18. IEEE, 2009. Chen, J., Kallus, N., Mao, X., Svacha, G., and Udell, M. Fairness under unawareness: Assessing disparity when protected class is unobserved. In Proceedings of the conference on fairness, accountability, and transparency, pp. 339 348, 2019. Commission, E. Proposal for a regulation laying down harmonised rules on artificial intelligence (artificial intelligence act), 2021. URL https: //digital-strategy.ec.europa.eu/en/ policies/regulatory-framework-ai. Dastin, J. Amazon scraps secret ai recruiting tool that showed bias against women. In Ethics of data and analytics, pp. 296 299. Auerbach Publications, 2022. Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition (CVPR), pp. 248 255. IEEE, 2009. Detommaso, G., Bertran, M., Fogliato, R., and Roth, A. Multicalibration for confidence scoring in llms. In Proceedings of the 41st International Conference on Machine Learning, ICML 24, 2024. Diana, E., Gill, W., Kearns, M., Kenthapadi, K., Roth, A., and Sharifi-Malvajerdi, S. Multiaccurate proxies for downstream fairness. 2022 ACM Conference on Fairness, Accountability, and Transparency, 2022. Ding, F., Hardt, M., Miller, J., and Schmidt, L. Retiring adult: New datasets for fair machine learning. Advances in neural information processing systems, 34:6478 6490, 2021. Freire, M. N. and de Castro, L. N. e-recruitment recommender systems: a systematic review. Knowledge and Information Systems, 63:1 20, 2021. Fremont, A., Weissman, J. S., Hoch, E., and Elliott, M. N. When race/ethnicity data are lacking: using advanced indirect estimation methods to measure disparities. Rand health quarterly, 6(1), 2016. Multiaccuracy and Multicalibration via Proxy Groups Garin, S., Parekh, V., Sulam, J., et al. Medical imaging data science competitions should report dataset demographics and evaluate for bias [published online ahead of print april 3, 2023]. Nat Med. Globus-Harris, I., Gupta, V., Jung, C., Kearns, M., Morgenstern, J., and Roth, A. Multicalibrated regression for downstream fairness. In Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society, pp. 259 286, 2023a. Globus-Harris, I., Harrison, D., Kearns, M., Roth, A., and Sorrell, J. Multicalibration as boosting for regression. ar Xiv preprint ar Xiv:2301.13767, 2023b. Glocker, B., Jones, C., Bernhardt, M., and Winzeck, S. Algorithmic encoding of protected characteristics in chest x-ray disease detection models. EBio Medicine, 89, 2023. Gopalan, P., Kalai, A. T., Reingold, O., Sharan, V., and Wieder, U. Omnipredictors. In Braverman, M. (ed.), 13th Innovations in Theoretical Computer Science Conference (ITCS 2022), volume 215 of Leibniz International Proceedings in Informatics (LIPIcs). Schloss Dagstuhl Leibniz-Zentrum f ur Informatik, 2022. Gupta, M., Cotter, A., Fard, M. M., and Wang, S. Proxy fairness. ar Xiv preprint ar Xiv:1806.11212, 2018. Hansen, D., Devic, S., Nakkiran, P., and Sharan, V. When is multicalibration post-processing necessary? In Advances in Neural Information Processing Systems, 2024. Hardt, M., Price, E., and Srebro, N. Equality of opportunity in supervised learning. Advances in neural information processing systems, 29, 2016. H ebert-Johnson, U., Kim, M., Reingold, O., and Rothblum, G. Multicalibration: Calibration for the (computationallyidentifiable) masses. In International Conference on Machine Learning, pp. 1939 1948. PMLR, 2018. Holstein, K., Wortman Vaughan, J., Daum e III, H., Dudik, M., and Wallach, H. Improving fairness in machine learning systems: What do industry practitioners need? In Proceedings of the 2019 CHI conference on human factors in computing systems, pp. 1 16, 2019. Huang, G., Liu, Z., Van Der Maaten, L., and Weinberger, K. Q. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4700 4708, 2017. Imai, K. and Khanna, K. Improving ecological inference by predicting individual ethnicity from voter registration records. Political Analysis, 24(2):263 272, 2016. Irvin, J., Rajpurkar, P., Ko, M., Yu, Y., Ciurea-Ilcus, S., Chute, C., Marklund, H., Haghgoo, B., Ball, R., Shpanskaya, K., et al. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 590 597, 2019. Kallus, N., Mao, X., and Zhou, A. Assessing algorithmic fairness with unobserved protected class using data combination. Management Science, 68(3):1959 1981, 2022. Kim, J., Harper, A., Mc Cormack, V., Sung, H., Houssami, N., Morgan, E., Mutebi, M., Garvey, G., Soerjomataram, I., and Fidler-Benaoudia, M. M. Global patterns and trends in breast cancer incidence and mortality across 185 countries. Nature Medicine, pp. 1 9, 2025. Kim, M. P., Ghorbani, A., and Zou, J. Multiaccuracy: Blackbox post-processing for fairness in classification. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, pp. 247 254, 2019. Lahoti, P., Beutel, A., Chen, J., Lee, K., Prost, F., Thain, N., Wang, X., and Chi, E. Fairness without demographics through adversarially reweighted learning. Advances in neural information processing systems, 33:728 740, 2020. Li, D., Bharti, B., Wei, J., Sulam, J., and Yi, P. H. Sex imbalance produces biased deep learning models for knee osteoarthritis detection. Canadian Association of Radiologists Journal, 74(1):219 221, 2023. Obermeyer, Z. and Mullainathan, S. Dissecting racial bias in an algorithm that guides health decisions for 70 million people. In Proceedings of the conference on fairness, accountability, and transparency, pp. 89 89, 2019. of Science, W. H. O. and Policy, T. The blueprint for an ai bill of rights, 2022. URL https://www.whitehouse.gov/ostp/ ai-bill-of-rights/. Prost, F., Awasthi, P., Blumm, N., Kumthekar, A., Potter, T., Wei, L., Wang, X., Chi, E. H., Chen, J., and Beutel, A. Measuring model fairness under noisy covariates: A theoretical perspective. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, pp. 873 883, 2021. Roth, A. Uncertain: Modern Topics in Uncertainty Estimation. 2022. Rudin, C., Wang, C., and Coker, B. The age of secrecy and unfairness in recidivism prediction. Harvard Data Science Review, 2(1):1, 2020. Multiaccuracy and Multicalibration via Proxy Groups Shailaja, K., Seetharamulu, B., and Jabbar, M. Machine learning in healthcare: A review. In 2018 Second international conference on electronics, communication and aerospace technology (ICECA), pp. 910 914. IEEE, 2018. Thomas, L., Crook, J., and Edelman, D. Credit scoring and its applications. SIAM, 2017. United Nations Educational, S. and (UNESCO), C. O. Recommendation on the ethics of artificial intelligence, 2021. Wang, S., Guo, W., Narasimhan, H., Cotter, A., Gupta, M., and Jordan, M. Robust optimization for fairness with noisy protected groups. Advances in neural information processing systems, 33:5190 5203, 2020. Weissman, J. S. and Hasnain-Wynia, R. Advancing health care equity through improved data collection. New England Journal of Medicine, 364(24):2276 2277, 2011. Yi, P. H., Bachina, P., Bharti, B., Garin, S. P., Kanhere, A., Kulkarni, P., Li, D., Parekh, V. S., Santomartino, S. M., Moy, L., et al. Pitfalls and best practices in evaluation of ai algorithmic biases in radiology. Radiology, 315(2): e241674, 2025. Zafar, M. B., Valera, I., Gomez Rodriguez, M., and Gummadi, K. P. Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In Proceedings of the 26th international conference on world wide web, pp. 1171 1180, 2017. Zhang, Y. Assessing fair lending risks using race/ethnicity proxies. Management Science, 64(1):178 197, 2018. Zhao, T., Dai, E., Shu, K., and Wang, S. Towards fair classifiers without sensitive attributes: Exploring biases in related features. 2022. Zhu, Z., Yao, Y., Sun, J., Li, H., and Liu, Y. Weak proxies are sufficient and preferable for fairness with missing sensitive attributes. In International Conference on Machine Learning, pp. 43258 43288. PMLR, 2023. Multiaccuracy and Multicalibration via Proxy Groups A. Additional Theoretical Results Here we present a version of Theorem 5.1 for multiaccuracy. Theorem A.1. Fix a distribution D, initial model f, and set of proxy groups ˆG. If a model ˆf satisfies AEmax( ˆf, ˆG) < min ˆg ˆG AED(f, ˆg) (19) MSE( ˆf) MSE(f) (20) then, it will have a smaller worst-case MA violation, i.e. β( ˆf, ˆG) β(f, ˆG). A proof of this result is provided in Appendix B.3. B.1. Proof of Lemma 4.1 Proof. Throughout these proofs, denote µg i,j = P[g(X, Z) = i, ˆg(X) = j] and µg i,j(v) = P[g(X, Z) = i, ˆg(X) = j | f(X) = v]. (21) We begin with proving the result for multiaccuracy. Step 1: Establishing the Multiaccuracy Bound Fix a distribution D and predictor f. Consider any group g G and its corresponding proxy ˆg ˆG. Then, AED(f, g) = E[g(X, Z)(f(X) Y )] (22) = E[g(X, Z)(f(X) Y )] E[ˆg(X)(f(X) Y )] + E[ˆg(X)(f(X) Y )] (23) E[g(X, Z)(f(X) Y )] E[ˆg(X)(f(X) Y )] + E[ˆg(X)(f(X) Y )] (24) = E[g(X, Z)(f(X) Y ) ˆg(X)(f(X) Y )] + AED(f, ˆg) (25) = E[(g(X, Z) ˆg(X)) (f(X) Y )] + AED(f, ˆg) (26) E[|g(X, Z) ˆg(X)| |f(X) Y )|] + AED(f, ˆg) (27) E[|g(X, Z) ˆg(X)|2] E[|f(X) Y |2], E[|g(X, Z) ˆg(X)|] + AED(f, ˆg) (28) MSE(f) err(ˆg), err(ˆg) + AED(f, ˆg). (29) Here we applied the triangle inequality in line 24, Jensen s inequality in line 27, and Holder s inequality in line 28. Step 2: Tightness of the Multiaccuracy Bound We now show these bounds are tight. To be precise, we will prove that there exists a joint distribution over the random variables (f(X), Y, g(X, Z), ˆg(X)) for which these bounds hold with equality. Consider a group g G and its corresponding proxy ˆg ˆG. First, consider the scenario where MSE(f) err(ˆg) so that by the first result of Lemma 4.1 we have AED(f, g) AED(f, ˆg) + p MSE(f). (30) Consider the following data generating process: Conditioned on the event {g(X, Z) = ˆg(X)}, one has f(X) = Y , Multiaccuracy and Multicalibration via Proxy Groups Conditioned on the event {g(X, Z) = 1, ˆg(X) = 0}, one has that f(X) = err(ˆg) and Y = 0, µg 0,1 = 0 so that err(ˆg) = µg 1,0. E[g(X, Z)(f(X) Y )] = E[f(X) Y | g(X, Z) = 1, ˆg(X) = 1]µg 1,1 + E[f(X) Y | g(X, Z) = 1, ˆg(X) = 0]µg 1,0 (31) MSE(f). (32) E[ˆg(X)(f(X) Y )] = E[f(X) Y | g(X, Z) = 1, ˆg(X) = 1]µg 1,1 + E[f(X) Y | g(X, Z) = 0, ˆg(X) = 1]µg 0,1 (33) As a result, E[g(X, Z)(f(X) Y )] = E[ˆg(X)(f(X) Y )] + p MSE(f) (35) = E[ˆg(X)(f(X) Y )] + p MSE(f). (36) Now, consider the scenario where MSE(f) > err(ˆg) so that by the first result of Lemma 4.1 we have AED(f, g) AED(f, ˆg) + err(ˆg). (37) Consider the following data generating process: Conditional on the event {g(X, Z) = ˆg(X)}, f(X) Y Conditional on the event {g(X, Z) = 1, ˆg(X) = 0}, f(X) = 1 and Y = 0. µg 0,1 = 0 so that err(ˆg) = µg 1,0 E[g(X, Z)(f(X) Y )] = E[f(X) Y | g(X, Z) = 1, ˆg(X) = 1]µg 1,1 + E[f(X) Y | g(X, Z) = 1, ˆg(X) = 0]µg 1,0 (38) = E[f(X) Y | g(X, Z) = 1, ˆg(X) = 1]µg 1,1 + err(ˆg). (39) E[ˆg(X)(f(X) Y )] = E[f(X) Y | g(X, Z) = 1, ˆg(X) = 1]µg 1,1 + E[f(X) Y | g(X, Z) = 0, ˆg(X) = 1]µg 0,1 (40) = E[f(X) Y | g(X, Z) = 1, ˆg(X) = 1]µg 1,1 0. (41) In passing, note that requiring f(X) Y above is much more than needed, but we take this for simplicity. Moving on, and as a result, E[g(X, Z)(f(X) Y )] = E[ˆg(X)(f(X) Y )] + err(ˆg) = E[ˆg(X)(f(X) Y )] + err(ˆg). (42) Multiaccuracy and Multicalibration via Proxy Groups We now prove the result for multicalibration. Step 1: Establishing the Multicalibration Bound Fix a distribution D and predictor f. Consider any group g G and its corresponding proxy ˆg ˆG. Then, ECED(f, g) = E E[g(X, Z)(f(X) Y )|f(X) = v] = E E[g(X, Z)(f(X) Y )|f(X) = v] E[ˆg(X)(f(X) Y )|f(X) = v] (44) + E[ˆg(X)(f(X) Y )|f(X) = v] = E E[g(X, Z)(f(X) Y )|f(X) = v] E[ˆg(X)(f(X) Y )|f(X) = v] + ECED(f, ˆg) (46) E E[g(X, Z)(f(X) Y )|f(X) = v] E[ˆg(X)(f(X) Y )|f(X) = v] + ECED(f, ˆg) (47) = E E[g(X, Z)(f(X) Y ) ˆg(X)(f(X) Y )|f(X) = v] + ECED(f, ˆg) (48) E E g(X, Z)(f(X) Y ) ˆg(X)(f(X) Y ) |f(X) = v + ECED(f, ˆg) (49) = E g(X, Z)(f(X) Y ) ˆg(X)(f(X) Y ) + ECED(f, ˆg) (50) = E (g(X, Z) ˆg(X)) (f(X) Y ) + ECED(f, ˆg) (51) E[|g(X, Z) ˆg(X)|2] E[|f(X) Y |2], E[|g(X, Z) ˆg(X)|] + ECED(f, ˆg) (52) MSE(f) err(g, ˆg), err(ˆg) + ECED(f, ˆg) (53) Here we applied the reverse triangle inequality in (47), Jensen s inequality in line (49), and Holder s inequality in line (52). Step 2: Tightness of the Multicalibration Bound We now show these bounds are tight. To be precise, we will prove that there exists a joint distribution over the random variables (f(X), Y, g(X, Z), ˆg(X)) for which these bounds hold with equality. Consider a group g G and its corresponding proxy ˆg ˆG. First, consider the scenario where MSE(f) err(ˆg) so that by the first result of Lemma 4.1 we have ECED(f, g) ECED(f, ˆg) + p MSE(f). (54) Consider the same data generating process used to establish the MA bound, AED(f, g) AED(f, ˆg)+ p MSE(f). Then, E[g(X, Z)(f(X) Y )|f(X) = v] = E v Df E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 1] µg 1,1(v) (55) + E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 0] µg 1,0(v) (56) err(ˆg) µg 1,0(v) (57) err(ˆg) E v Df µg 1,0(v) (58) err(ˆg) µg 1,0 = p err(ˆg). (59) Multiaccuracy and Multicalibration via Proxy Groups E[ˆg(X)(f(X) Y )|f(X) = v] = E v Df E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 1] µg 1,1(v) (60) + E[f(X) Y | f(X) = v, g(X, Z) = 0, ˆg(X) = 1] µg 0,1(v) (61) As a result, E[g(X, Z)(f(X) Y )|f(X) = v] = E v Df E[ˆg(X)(f(X) Y )|f(X) = v] + p err(ˆg). (63) Now, consider the scenario where MSE(f) > err(ˆg) so that by the first result of Lemma 4.1 we have ECED(f, g) ECED(f, ˆg) + err(ˆg). (64) Consider the same data generating process used to establish the MA bound, AED(f, g) AED(f, ˆg) + err(ˆg). Then, E[g(X, Z)(f(X) Y )|f(X) = v] = E v Df E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 1] µg 1,1(v) (65) + E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 0] µg 1,0(v) (66) E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 1] µg 1,1(v) + µg 1,0(v) E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 1] µg 1,1(v) (68) + E v Df[µg 1,0(v)] (69) E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 1] µg 1,1(v) + µg 1,0. E[ˆg(X)(f(X) Y )|f(X) = v] = E v Df E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 1] µg 1,1(v) (71) + E[f(X) Y | f(X) = v, g(X, Z) = 0, ˆg(X) = 1] µg 0,1(v) (72) E[f(X) Y | f(X) = v, g(X, Z) = 1, ˆg(X) = 1] µg 1,1(v) (73) As a result, E[g(X, Z)(f(X) Y )|f(X) = v] = E v Df E[ˆg(X)(f(X) Y )|f(X) = v] + err(ˆg). (74) The following bounds are tight in that we show there exists a distribution for which the bound is attained. The bounds can nonetheless be improved by assuming complete access to the marginal distributions over (f(X), Y, ˆg(X)) and (g(X, Z), ˆg(X)). In this case, one could derive tight bounds that agree with this observed information with techniques used in(Kallus et al., 2022; Bharti et al., 2024) that rely on the Fr echet inequalities. Multiaccuracy and Multicalibration via Proxy Groups B.2. Proof of Theorem 4.2 Proof. We prove the result for multiaccuracy. Recall Lemma 4.1, which states that for any group g and its proxy ˆg AED(f, g) F(f, ˆg) + AED(f, ˆg) (75) AED(f, g) β(f, ˆG) := max ˆg ˆG F(f, ˆg) + AED(f, ˆg). (76) This proves that f is (G, β(f, ˆG))-multiaccurate. The proof for multicalibration follows an identical argument. B.3. Proof of Theorem 5.1 Proof. Fix a distribution D, model f, set of groups G and its corresponding proxies ˆG. Recall by Theorem 4.2 that f is (G, γ(f,ˆ)-multicalibrated where γ(f, ˆG) = max ˆg G min err(ˆg), p MSE(f) err(ˆg) + ECED(f, ˆg) (77) First, note that for MSE(f) > err(ˆg), the term min err(ˆg), p MSE(f) err(ˆg) (78) is constant with respect to MSE(f) and for MSE(f) err(ˆg) it increases as MSE(f) increases. Now, suppose another model ˆf satisfies the following ECEmax( ˆf, ˆG) < min ˆg ˆG ECED(f, ˆg) (79) MSE( ˆf) MSE(f). (80) γ( ˆf, ˆG) = max ˆg G min err(ˆg), q MSE( ˆf) err(ˆg) + ECED( ˆf, ˆg) (81) max ˆg G min err(ˆg), p MSE(f) err(ˆg) + ECED( ˆf, ˆg) (82) max ˆg G min err(ˆg), p MSE(f) err(ˆg) + ECEmax( ˆf, ˆG) (83) max ˆg G min err(ˆg), p MSE(f) err(ˆg) + min ˆg ˆG ECED(f, ˆg) (84) max ˆg G min err(ˆg), p MSE(f) err(ˆg) + ECED(f, ˆg) (85) = γ(f, ˆG). (86) The proof of the multiaccuracy result in Theorem A.1 follows an identical argument. Multiaccuracy and Multicalibration via Proxy Groups C. Additional Experiment Details C.1. ACS Experiments Models. In the ACSIncome and Pub Cov experiments, we use Random Forests as the proxy models ˆg, and train three types of models f: logistic regression, decision tree, and Random Forest. Results. The sensitive groups we use along with their proxy errors are reported in Tables 4 and 5. All MA related results for the different models f are presented in Appendices E.1 and E.2. Note, all of the models are MA with respect to G and ˆG. As a result, adjusting with respect to the proxies provides no benefit. All MC related results for the different models f are presented in Appendices E.4 and E.5. Note, the logistic regression and Random Forest models are highly multicalibrated with respect to G and ˆG. As a result, adjusting provides no benefit. On the other hand, the decision tree is grossly uncalibrated with respect to some proxies. As a result, we see a benefit in multicalibrating with respect to the proxies. C.2. Che Xpert Models. In the Che Xpert experiment, we follow Glocker et al. (2023) and train a Dense Net-121 model for the to predict race and sex. For the models f, we use three types. The first is a decision tree classifier trained on features extracted from a Dense Net-121 model (Huang et al., 2017) pretrained on Image Net (Deng et al., 2009). The second is a linear model (Breiman, 2001) trained on the same features. The third is a Dense Net-121 model trained end-to-end on the raw X-ray images. Results. The groups used along with proxy errors are reported in Table 6. All multiaccuracy related results for the different types of models f are presented in Appendix E.3. Note, all of the models are multiaccurate with respect to G and ˆG. As a result, adjusting provides no benefit. All multicalibration related results for the different types of models f are presented in Appendix E.6. Note, the logistic regression and fully trained Dense Net-121 models are highly multicalibrated with respect to G and ˆG. As a result, adjusting provides no benefit. On the other hand, the decision tree is grossly uncalibrated with respect to some proxies. As a result, we see a benefit in multicalibrating with respect to the proxies. D. Additional Tables Group err(ˆg) Black Adults 0.044 Black Women 0.027 Women 0.000 Never Married 0.000 American Indian 0.007 Seniors 0.000 White Women 0.123 Multiracial 0.047 White Children 0.002 Asian 0.060 Table 4. Sensitive groups and the proxy errors used in the ACSIncome experiment Group err(ˆg) Black Adults 0.044 Black Women 0.027 Women 0.000 Never Married 0.000 American Indian 0.007 White Women 0.123 Multiracial 0.047 White Children 0.002 Asian 0.060 Table 5. Sensitive groups and the proxy errors used in the ACSPub Cov experiment Group err(ˆg) Men 0.027 Women 0.027 White 0.920 Asian 0.068 Black 0.039 Asian Men 0.039 Asian Women 0.034 Black Men 0.021 Black Women 0.020 White Men 0.067 White Women 0.062 Table 6. Sensitive groups and the proxy errors used in the Che Xpert experiment Multiaccuracy and Multicalibration via Proxy Groups E. Additional Figures E.1. Multiaccuracy results for ACSIncome MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.1. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSIncome. Here, f is a logistic regression. MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.2. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSIncome. Here, f is a decision tree. MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.3. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSIncome. Here, f is a Random Forest. Multiaccuracy and Multicalibration via Proxy Groups E.2. Multiaccuracy results for ACSPub Cov MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.4. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSPub Cov. Here, f is a logistic regression. MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.5. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSPub Cov. Here, f is a decision tree. MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.6. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSPub Cov. Here, f is a Random Forest. Multiaccuracy and Multicalibration via Proxy Groups E.3. Multiaccuracy results for Che Xpert MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.7. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on Che Xpert. Here, f is a logistic regression. MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.8. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on Che Xpert. Here, f is a decision tree. MA Violations and Bounds (Before Adjustment) True Violation Upper Bound AE(fadj, g) MA Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.9. AE(f, g), AEmax(f, g) (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on Che Xpert. Here, f is a Dense Net-121 model. Multiaccuracy and Multicalibration via Proxy Groups E.4. Multicalibration results for ACSIncome MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.10. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSIncome. Here, f is a logistic regression. MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.11. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSIncome. Here, f is a decision tree. MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.12. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSIncome. Here, f is a Random Forest. Multiaccuracy and Multicalibration via Proxy Groups E.5. Multicalibration results for ACSPub Cov MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.13. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSPub Cov. Here, f is a logistic regression. MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.14. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSPub Cov. Here, f is a decision tree. MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.15. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on ACSPub Cov. Here, f is a Random Forest. Multiaccuracy and Multicalibration via Proxy Groups E.6. Multicalibration results for Che Xpert MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.16. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on Che Xpert. Here, f is a logistic regression. MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.17. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on Che Xpert. Here, f is a decision tree. MC Violations and Bounds (Before Adjustment) True Violation Upper Bound ECE(fadj, g) MC Violations and Bounds (After Adjustment) True Violation Upper Bound Figure B.18. ECE, ECEmax (dotted red line), and worst case violations (dotted blue line) of the original model f and adjusted model fadj on Che Xpert. Here, f is a Dense Net-121 model.