# robust_ml_auditing_using_prior_knowledge__b3561c90.pdf Robust ML Auditing using Prior Knowledge Jade Garcia Bourr ee * 1 2 3 Augustin Godinot * 1 2 3 4 Sayan Biswas 5 Anne-Marie Kermarrec 5 Erwan Le Merrer 2 Gilles Tredan 6 Martijn de Vos 5 Milos Vujasinovic 5 Abstract Among the many technical challenges to enforcing AI regulations, one crucial yet underexplored problem is the risk of audit manipulation. This manipulation occurs when a platform deliberately alters its answers to a regulator to pass an audit without modifying its answers to other users. In this paper, we introduce a novel approach to manipulation-proof auditing by taking into account the auditor s prior knowledge of the task solved by the platform. We first demonstrate that regulators must not rely on public priors (e.g., a public dataset), as platforms could easily fool the auditor in such cases. We then formally establish the conditions under which an auditor can prevent audit manipulations using prior knowledge about the ground truth. Finally, our experiments with two standard datasets illustrate the maximum level of unfairness a platform can hide before being detected as malicious. Our formalization and generalization of manipulation-proof auditing with a prior opens up new research directions for more robust fairness audits. 1. Introduction Machine learning (ML) models are becoming central to numerous businesses, industrial processes, and administrations. Such models are being employed in high-stakes domains where ML-driven decisions can have profound impacts on individuals and communities (Rudin, 2019). For instance, financial institutions have been leveraging ML-driven systems to evaluate loan applications based on attributes like income, credit score, and employment history *Equal contribution 1Universit e de Rennes, Rennes, France 2Inria, Rennes, France 3IRISA/CNRS, Rennes, France 4PERe N, Paris, France 5EPFL, Lausanne, Switzerland 6LAAS, CNRS, Toulouse, France. Correspondence to: Augustin Godinot , Jade Garcia Bourr ee . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). (West, 2000). Given the far-reaching consequences of these applications, ensuring the fairness (Mehrabi et al., 2021) and regulatory compliance (Petersen et al., 2022) of such models is paramount. Independent fairness audits serve as a critical tool for assessing the fairness of ML models and ensure that model providers remain accountable to the public (Birhane et al., 2024; Raji, 2024; Raji et al., 2022). As models are placed in production, auditors rely on black-box interactions, where queries are sent to the model, and the responses are analyzed to identify potential fairness violations (e.g., see (Kim et al., 2019)). However, this reliance on black-box audits leaves the process vulnerable to manipulations by the platform, also known as fairwashing. Regulatory practices currently require auditors to notify platforms in advance of an audit. Platforms can thus strategically alter the model or its responses during the audit to create the appearance of fairness, effectively concealing underlying biases and unfair practices from the auditor while maintaining operational efficiency for its users. Consider, for example, a social media platform that employs an ML model to moderate content, automatically removing posts deemed harmful or misleading. During a fairness audit, the platform could deploy a more lenient moderation model that appears unbiased, only to revert to a stricter, potentially biased version once the audit concludes, effectively concealing unfair treatment of certain user groups. In fact, such discrepancies between a platform s behavior during audits and its real-world operations have been observed. Initially created as part of the Social Science One project, a data sharing program by Meta encountered a major setback when consistency issues were discovered in the data provided to scientists (Timberg, 2021). Similarly, a collaboration between Meta and independent scientists, studying the polarization effects of Facebook s recommendation algorithm, recently faced criticism over discrepancies found between the algorithm s behavior before and during the audit (Ribeiro, 2024). Academic studies show that fairness audits are easily manipulatable, whether the platform is required to prove its fairness through the release of a public dataset (Fukuchi et al., 2020), through the explanation of decisions (A ıvodji et al., 2021; Shamsabadi et al., 2022; Le Merrer & Tr edan, 2020; A ıvodji et al., 2019), or Robust ML Auditing using Prior Knowledge through black-box query interactions (Yan & Zhang, 2022; Garcia Bourr ee et al., 2023; Godinot et al., 2024). The potential for manipulation underscores the need for more robust auditing strategies. This work presents a novel theoretical framework and a practical implementation for preventing manipulations by the platform. Our analysis starts from a simple observation: auditors can readily collect labeled data, reflecting the platform s service from independent sources a common practice whose theoretical and empirical implications remain unexplored. For example, in the moderation example discussed earlier, the auditor could have some undeniable evidence at hand, to confront the model under scrutiny, e.g., A post with this content must pass the moderation filter, otherwise there is some bias on a protected feature of the user profile . Thus, by incorporating this dataset, the auditor can independently verify the platform s responses, cross-referencing them against known ground truth labels. By combining black-box interactions with prior knowledge from the labeled dataset, our method enables more reliable detection of fairness violations while reducing the reliance on assumptions about the platform s behavior. Specifically, we aim to answer the following research question: Can the auditor s prior knowledge of the ground truth prevent fairwashing in fairness audits? Our paper makes the following three contributions: We introduce and analyze a new fairness auditing approach for black-box interactions where the auditor has access to prior knowledge about the platform and the ML task (Section 3). We theoretically analyze how much unfairness a platform can conceal given the auditor s prior knowledge. For any auditor priors, our results highlight the importance of keeping the auditor s prior knowledge private (Section 3). For the dataset prior we introduce, we establish bounds on the concealable unfairness when the auditor prior remains confidential (Section 4). By simulating fairness audits on multiple tabular and vision datasets, we provide a more nuanced understanding of how our framework should be implemented. Our experiments offer insights into setting the detection threshold used to identify manipulations (Section 5). 2. Background: Auditing ML Models This work studies fairness audits of ML decision-making systems under manipulation by the model-hosting platform. We first formalize the decision-making system and then introduce the dynamics of fairness auditing. ML decision-making systems From feature transforms to specific business rules, modern ML decision-making systems can be remarkably complex. We abstract all this complexity by modeling the entire system as a function h : X Y (e.g., h can be a ML model). The set of possible queries X is called the input space, and the set of possible answers is called the output space. We consider binary classification problems, which is in line with related work in the domain of ML fairness analysis (Yan & Zhang, 2022; Godinot et al., 2024). Each query is associated with a protected attribute a A, which the platform is legally required not to discriminate against. Examples of such attributes include gender, age, or race and are typically defined by law. The platform has access to the protected attribute a either as a feature of the input space X or by a proxy (e.g., looking at the name of the person to determine the gender). We define D as the data distribution on X A. For any subset S X A and protected feature value a A, we will write Sa = {x | (x, a ) S, a = a} and h(S) = {h(x) | (x, a) S}. Throughout the paper, when it is clear from the context we will abuse the S notation: S will either be a subset of X, X A or X A Y. This work analyses how the platform can manipulate its model to pass a fairness audit, we now define relevant notation for this. The space of models that the platform can implement is called the hypothesis space H. The loss function L : H (X Y) R measures the discrepancy between predictions and ground truth values. For a given hypothesis h H, its expected loss over the distribution D is L(h, D) = E(x,a) D [ℓ(h(x), x, a)] with ℓis a loss function that quantifies the error of h(x) given a single input x and its protected attribute a. ML auditing An ML audit is any independent assessment of an identified audit target via an evaluation of articulated expectations with the implicit or explicit objective of accountability (Birhane et al., 2024). A ML audit involves three entities. The platform is the entity hosting the ML decision-making system. The users are those using the service hosted by the platform. The auditor is the entity conducting the audit to verify whether the ML model is compliant for all users. The auditor could be a state regulator, a consulting firm, or even a group of users. Fairness metric In this work, we consider ML audits targeting the fairness of the studied system. Specifically, the auditor chooses a fairness metric and sends queries to the platform to determine whether the platform abides by their fairness criterion. Among all the (un-)fairness metrics, we study Demographic Parity (DP) (Calders et al., 2009), which is commonly used in the fairness evaluation literature thanks to its simplicity. DP is defined as follows: Robust ML Auditing using Prior Knowledge Platform Auditor Platform Users Figure 1. The auditing process as conducted by an auditor, which proceeds in three steps. The platform exposes a model hp to the users. To appear fair to the auditor while not deteriorating the utility for its users, the platform manipulates its answers on the audit set S. µ(h) = P(X,A) D (h(X) = 1|A = 1) P(X,A) D (h(X) = 1|A = 0) (1) For a platform, DP is the easiest metric to manipulate (Yan & Zhang, 2022; Ajarra et al., 2024) as it only depends on the outcome of the ML model and not on its performance on the different protected groups. Thus, a platform can artificially adjust outputs, e.g., providing more positive outcomes for an underrepresented group. To decide whether a platform passes the audit or not, the auditor builds an audit set S X A and evaluates the plugin DP estimator: ˆµ(h, S) = 1 |S1| P x S1 1 {h(x) = 1} x S0 1 {h(x) = 1}. Based on µ(h), we also define the set of fair models F = h YX : µ(h) = 0 . 3. Enhancing Black-box Auditing with a Prior Since a malicious platform can manipulate the DP metric with relative ease, the auditor has to find ways to prevent these manipulations (e.g., using a different metric) or to detect them. In this section, we explore the latter. To detect manipulations, the auditor must use prior knowledge about what constitutes a likely set of answers on its audit dataset S. Then, using this prior, they would be able to estimate the likelihood that the received set of answers hm(S) has been manipulated. 3.1. Modeling the Auditor Prior Previous work has demonstrated that prior knowledge is both a practical and an essential tool for auditing, yet the notion of an auditor prior has not been explicitly leveraged in the analysis of fairness audits. We define an auditor prior as follows. Definition 3.1 (Auditor prior). The auditor prior is a set of models Ha YX that the auditor can reasonably expect to observe given her knowledge of the decision task by the platform. For example, in (Tan et al., 2018), the authors study feature importance by training two models one on a public dataset and another via distillation of the audited ML model and comparing the resulting models. Using a more theoretical approach, Yan & Zhang and Godinot et al. explored the case of an auditor knowing the hypothesis class of the platform, i.e., Ha = H. Ajarra et al. proposed to use an assumption about the Boolean Fourier coefficients of H to construct Ha. Finally, Garcia Bourr ee et al. and Shamsabadi et al. used side-channel access (e.g., an additional API or explanations) to the ML model to define Ha and derive guarantees on the measured fairness. In Section 4, we introduce a labeled dataset Da that the auditor will leverage to define Ha. Definition 3.1 captures all of the situations above and allows to formulate general results about the problem of robust auditing. The auditing process The auditing process consists of three steps which we visualize in Figure 1. Here, hp refers to the model that the platform exposes to its users (the top part of Figure 1) and hm refers to the model exposed to the auditor (bottom part of Figure 1). First, the auditor builds an audit set S X and sends the queries in S to the platform (step 1 ). The platform receives S all at once and computes the answers using its model hp. To appear fair if it is not, the platform projects its labels hp(S) on the set F of fair models. This defines a manipulated model hm and the answers hm(S) the platform will send to the auditor (step 2 ). The auditor receives hm(S) and exploits these samples to evaluate whether the platform is fair (hm F) and honest (hp = hm), (step 3 ). Since the auditor does not have direct access to hp, they compare hm to their prior Ha to decide whether the platform is honest or malicious. Thus, the auditor tests the two following properties of hm: Is the platform fair? hm ? F (2) Is the platform honest? hm ? Ha (3) For dataset priors (i.e., when Ha is a ball, see Section 4), we draw F and Ha in Figure 2. Given a model hm, the fairness audit is equivalent to checking if hm belongs to the blue shaded area. In the example of Figure 2, the platform would be flagged as malicious as hm belongs to F but not to Ha. Online v.s. batch auditing Note that we assume that the platform receives all audit queries at once and that it is possible to detect all the audit queries. In practice, the queries Robust ML Auditing using Prior Knowledge are usually issued online (that is, one-by-one) by the auditor, through web-scraping or through an API. Compared to online auditing, it is easier for the platform to manipulate an audit if it knows all the audit queries before having to answer. On the other hand, because the auditor has to send all their queries at once, they cannot use the answers of the platform to actively guide the generation of the audit questions (e.g., as in (Yan & Zhang, 2022; Godinot et al., 2024)). Ultimately, our setting is built as a worst-case analysis of the auditing game for the auditor. Auditing axioms To avoid trivial audits, we add two modeling assumptions. The first assumption ensures that the auditors prior is correct so that a honest platform does not appear as lying. The second assumption asserts that an audit is necessary, otherwise the auditor could directly conclude from his prior that the platform is unfair. That is to say, the auditor should never flag a honest platform malicious. In particular, the auditor must have a prior that is close to the ground truth. Those assumptions are expressed as: hp Ha and Ha F = . (4) 3.2. On Public Auditor Priors A typical auditor proceeds in the following way. Upon examining a platform s model hm, the auditor must first understand the task addressed by hm and what constitutes a good-performing model on this task. In our moderation example, the auditor might try to look for public moderation datasets to test the performance of hm using a few examples. It might also look for publicly-available moderation models to compare their resulting input/output pairs with those of hm. Unfortunately, our first remark is that regardless of the prior the auditor might construct, if these models are public (or at least known by the platform), the platform will always be able to manipulate the audit: Theorem 3.2. Assume the platform knows Ha, it can then always pick hm {Ha F} to appear both fair and honest. Proof. First, recall that by definition the platform knows F. Assume that the dataset prior is public, the platform also knows Ha. Hence the platform can compute F Ha. As by assumption, F Ha = (Equation (4)), the platform can pick any model hm F Ha. In the case of (Shamsabadi et al., 2022), the platform perfectly knows Ha (because the Ha is coming from queries of its model) so the detector is subject to this manipulation (called Irreducibility in the paper). In Yan & Zhang s work, Ha is the hypothesis class H of the platform, communicated to the auditor before the audit. Theorem 3.2 provides a novel view on the impossibility results that were later proved in (Godinot et al., 2024). 4. Using Labeled Datasets for More Robust Audits Against Manipulations In an ideal, yet unrealistic audit scenario, the auditor would have access to non-manipulated answers from the original platform model hp. The prior Ha would then be the set of models that agree with these non-manipulated answers and would allow the auditor to detect inconsistencies between the original hp and manipulated hm models. Yet in general, the auditor does not have access to such non-manipulated answers. As an alternative, we propose to study the use of a private (because of Theorem 3.2) dataset Da, collected by the auditor to construct the auditor prior Ha. This idea (coupled with an assumption on the hypothesis class) has been studied experimentally (Tan et al., 2018) but the more recent theoretical works on robust auditing diverged towards studying priors on the model itself rather than on the data (Shamsabadi et al., 2022; Yan & Zhang, 2022; Ajarra et al., 2024). In the following, we define what a dataset prior is, and study the guarantees an auditor can achieve using this prior. Unless noted otherwise, in this section and in Section 5, Ha will denote the dataset prior. Definition 4.1 (Dataset prior). Let Da = (Xa, Aa, Ya) X n An Yn be a labeled dataset the auditor has access to. The dataset prior Ha is defined as the set of models that have a reasonable risk on Da. Ha = h YX : L(h, Da) < τ . (5) To test if the platform is honest, the auditor needs to verify whether hm Ha, i.e., whether L(hm, Da) < τ. The risk threshold τ thus plays a big role in the guarantees the auditor will be able to achieve. We discuss the impact of τ in Section 4.2 and guidelines to set its value in Section 4.3, but first, we need to discuss the definition of optimal manipulation in Section 4.1. 4.1. Optimal Manipulation Given the audit set S and its model hp, the objective of a manipulative platform is to create a set of answers hm(S) that appear fair to the auditor but also do not raise suspicions. Ideally, the platform would like to know the auditor prior Ha (see Theorem 3.2), but in the general case it cannot because it is not public information. As a consequence, the platform cannot directly optimize its answers to be expectable and fair. However, it still has cards up its sleeve; it already trained a model hp on a dataset D that is close to that of the auditor Da. Thus, instead of searching hm in Ha F, the platform can assume that its true model hp is expectable that is, hp Ha and try to find a fair model hm F while flipping as few labels as possible from hp. Therefore, the Robust ML Auditing using Prior Knowledge Figure 2. Representation of the auditor prior Ha, the honest platform model hp and a corresponding malicious model hm on the fair F plane. The red area represents the area where platforms optimal manipulations are detected as dishonest: they fall outside of the blue region of F optimal manipulation is the projection of hp on F: h m = proj F(hp) = arg min h F d(h, hp). (6) The distance d in Equation (6) is the value of risk L of h using the labels of hp as the ground truth. This scenario captures the fairwashing approach in (A ıvodji et al., 2021) in the context of explanation manipulations. 4.2. Achievable Guarantees Following the second auditing axiom formulated in Equation (4), the original model of the platform is always expectable, i.e., hp Ha. Thus, the manipulation detection test has no false positives, and the main quantity of interest to the auditor is the manipulation detection rate. Definition 4.2 (Detection rate). The probability Puf that the auditor correctly detects a manipulative platform with optimal manipulation is Puf = P(h m / Ha|hp Ha). Estimating or computing Puf requires the knowledge of the distribution of models in Ha. Unfortunately, unless they have access to the training pipeline of the platform, this model distribution is inaccessible to the auditor. To overcome this issue, we make the assumption of an uninformative prior: since the auditor does not know the model distribution in Ha, they must assume it is uniform. Theorem 4.3 (Prior-Uniform detection rate). Under the dataset prior of Definition 4.1 with L defined as the ℓ2 norm, and the uninformative prior assumption, the probability that the auditor correctly detects a malicious platform trying to be fair is Z arccos(δ/τ) 0 sinn(θ)dθ δ with δ = d(ha, F), the distance of ha to F and Wn is the n-term of Wallis integrals. To gain intuition about the proof, we represent the audit case for |S| = 3 in Figure 2. By definition of the dataset prior, Ha is a ball of radius τ, centered on Ya, the labels given in the audit dataset Da. The manipulation of a model hp can be detected only if the resulting model is outside of Ha, as shown in orange on Figure 2. The probability of detection is thus 1 minus the volume of original models hp whose projection on F lies outside on Ha. This volume is highlighted in red in Figure 2. The detailed proof of Theorem 4.3 is deferred to Appendix A. Theorem 4.3 highlights two key parameters to the auditor s success: the unfairness of the prior δ = d(ha, F) and the expectability threshold τ. If the dataset prior is perfectly fair (i.e., δ = 0), then the auditor has no chance to detect a manipulated model as non-expectable (Puf = 0, Corollary A.5). On the other hand, Corollary A.4 proves that, if τ = δ 1 then Puf = 1. Finally, in Corollary 4.4, we derive a lower bound on Puf for the case 0 < δ < τ. We provide the proof of Corollary 4.4 in Appendix A. Corollary 4.4 (Detection rate lower bound). If n is even, (n 1)/2 Puf 1. 4.3. Practical Considerations and Discussion In practice, τ is determined by the task difficulty, and the amount of data available to solve the task. One possibility to tune the value of τ is to use the error rate of current stateof-the-art models that solve the task at hand as a minimum value. We empirically explore this option in Section 5.4. An alternative, if the auditor has the resources, would be to train a set of models on the task and use them to calibrate τ. We leave further exploration of the calibration of τ to future work. On the other hand, the value of δ is determined by the audit set sampling procedure. In most cases, the audit set is sampled independently from a pre-specified audit distribution. In this case, the value of δ is fully determined. To regain some control over δ, the auditor has to allow other audit set sampling strategies, at the expense of potential statistical bias in the fairness and accuracy estimations. Takeaway. The auditor can always calculate a priori the probability to correctly detect a malicious platform trying to be fair. This probability depends on the ratio between unfairness δ of the auditor prior and the chosen risk threshold τ, and depends on the audit budget n = |S|. 1Per our first axiom in Equation (4), we have that δ τ. Robust ML Auditing using Prior Knowledge Concealable Unfairness Detection Score High Cheekbones Honest ROC Mitigation Label Transport Linear Relaxation Threshold Manipulation Log. Reg. Celeb A ACSEmployment Figure 3. The concealable unfairness by the platform for different detection scores and manipulation strategies. We highlight this for two features of the Celeb A dataset (left) and for two different ML models trained on the ACSEmployment dataset (right). The horizontal red line indicates the DP of the most unfair model without manipulation. 5. Empirical Evaluation We now empirically quantify the extent to which the platform can manipulate the unfairness of its ML model. To that end, we study the concealable unfairness: the maximum level of unfairness a platform can hope to hide before being detected as malicious. First, we evaluate the effectiveness of different manipulation strategies and determine the optimal one. Since any practical fairness repair method can be used as a manipulation methods, we ask (RQ1) What is the best manipulation strategy implementation (Section 5.3)? Then, we study the dynamics of the concealable unfairness when the audit budget |S| increases: (RQ2) Can the auditor always find an audit budget that prevents the platform from hiding any unfairness, i.e., that always allows to flag the platform if malicious (Section 5.4)? 5.1. Experimental Setup We conduct our experiments on tabular and vision modalities. The tabular dataset comes from the ACSEmployment task for the state of Minnesota in 2018, which is derived from US Census data and provided in folktables (Ding et al., 2021). The objective of this task is to predict whether an individual between the age of 16 and 90 is employed or not. As input features of the model hp, we consider several attributes of the individual, including gender, race, and age. The fairness of the models is evaluated along the race attribute given in the dataset: one group consists of individuals identified as white alone , while the other includes all remaining individuals. For the vision modality, we study Celeb A (Liu et al., 2015), which consists of images of celebrities along with several binary attributes associated with each image, such as whether the person in the photo is blond, smiling, or if the photo is blurry. As input to a vision model, we use the image to predict one of the associated attributes. The target attribute varies across experiments and will be specified accordingly. Demographic Parity is evaluated along the gender attribute given in the dataset. For the ACSEmployment dataset, we train Gradient Boosted Decision Tree (GBDT) and Logistic Regression (Log. Reg.) models, while for Celeb A, we train a Le Net convolutional neural network (Lecun et al., 1998). GBDT and Log. Reg. are trained using the default parameters of their respective implementations in SCIKIT-LEARN. Meanwhile, Le Net is trained irrespective of the target attribute using the Adam optimizer with a learning rate of γ = 0.001, a batch size of 32, and for two epochs, which is sufficient for the model to converge on all features. The code to run the experiments is available online.2 5.2. Implementing Optimal Audit Manipulations In practice, computing the optimal manipulation hm = proj F(hp) amounts to solving: hm(S) arg min L(h, {(x, hp(x)) : x S}) s.t. ˆµ(h, S) < τ (7) We note that this problem is the same problem solved by in-processing and post-processing fairness repair methods (Caton & Haas, 2024). Thus, ironically, computing the optimal manipulation is equivalent to choosing the optimal fairness repair method. The only difference being on which set the fairness constraints and accuracy objectives are defined: the audit set S instead of the training dataset. Thus, since any practical fairness repair method can be repurposed for manipulation, we adapted four classical fairness repair methods: ROC Mitigation (ROC) (Kamiran et al., 2012), 2See https://github.com/grodino/merlin. Robust ML Auditing using Prior Knowledge 0 2,000 4,000 Concealable Unfairness 0 2,000 4,000 Audit Budget High Cheekbones Honest ROC Mitigation Label Transport Linear Relaxation Threshold Manipulation 0 2,000 4,000 0 2,000 4,000 Log. Reg. Celeb A ACSEmployment Figure 4. The concealable unfairness for different audit budgets (i.e., data samples from the labeled dataset). We highlight this for two features of the Celeb A dataset (left) and for two different ML models trained on the ACSEmployment dataset (right). Optimal Label Transport (OT-L) (Jiang et al., 2019), Linear Relaxation (Lin R) (Lohaus et al., 2020) and Threshold Manipulation (Thresh Opt) (Hardt et al., 2016). 5.3. Quantifying the Concealable Unfairness For Different Detection Scores We next quantify the amount of fairness that a platform can hide for the different manipulation strategies above. This concealable unfairness µ(hp, hm) is defined as the Demographic Parity gap between the manipulated and honest models. µ(hp, hm) = |ˆµ(hm, S) ˆµ(hp, S)| (8) To decide whether the model observed during the audit is manipulated, the auditor has to decide whether hm Ha or not. To do so, the auditor estimates L(h, Da) by computing the detection score Detect(hm, S). Detect(hm, S) = X (x,y) S 1 {hm(x) = y} (9) To build (hp, hm) model pairs, we consider manipulation methods among ROC, OT-L, Lin R and Thresh Opt, varying hyperparameter values when applicable. In Figure 3, we plot the value of the concealable unfairness µ(hp, hm) against the detection score Detect(hm, S) computed by the auditor. We show the results of Le Net models trained on two Celeb A targets (first and second subplots), and GBDT and Log. Reg. models trained on ACSEmployment (third and fourth subplots). The horizontal red lines indicates the DP of the most unfair model without manipulation. First, we observe that for all the datasets, the platform can conceal significant amounts of unfairness: from 10 to 20 points differences between the two protected groups. Comparing the concealable unfairness values with the DP of the most unfair honest model (red horizontal line), we observe that the manipulation strategies almost all able to totally conceal the original model unfairness. Then, focusing on the x axis, the difference in Detect(hm, S) between the different honest models highlights the impact the performance of the platform s model should have on the detection threshold τ. In fact, depending on the dataset and on the model, Detect(honest, S) varies from 0.1 to 0.2. In Section 5.4, we explore a solution to setup the threshold. 5.4. Dynamics of the Concealable Unfairness as The Audit Budget Increases The probability of detecting manipulations (via the the detection score) should intuitively increase as the auditor gains access to a larger number of data samples (i.e., has a higher audit budget) since this allows for a more accurate comparison of hm with the data prior Ha. In this experiment, we explore how well this intuition holds in practice. For this purpose, we fix the hyperparameters for each manipulation method by selecting those that result in the highest concealable unfairness for a given base model, as discussed in Section 5.3. Then, for each base model target attribute pair, we determine the maximum concealable unfairness that a platform can achieve while ensuring that its detection score (see eq. 9) remains below the detection threshold. As proposed in Section 4 the threshold for each model is set to 1 x, where x represents the maximum accuracy achieved when training a set of models on the corresponding target. This process is repeated for audit budgets ranging from 100 to 5, 000. The results of this experiment are shown in Figure 4. The two plots on the left display the results for Celeb A using Robust ML Auditing using Prior Knowledge the same base model but different target attributes, while the two plots on the right show results for ACSEmployment using the same target attribute but different base models. These results reveal two distinct cases. In the first case (Celeb A Smiling in Figure 4), the concealable unfairness converges to zero as the audit budget increases. This is due to the low aleatoric uncertainty associated to the Smiling target. Since the task is easier, the accuracy range of models trained on Smiling is narrower, leading to a tighter detection threshold τ. In the second case (all the other facets of Figure 4), the concealable unfairness remains nonzero despite an increasing budget. Furthermore, in many cases, even with a high audit budget, some increase of unfairness remains undetectable by the auditor. Consequently, the platform retains some capacity to conceal unfairness even at high audit budgets. This stresses the hardness of the auditor s task in some configurations, and lead to a negative answer to (RQ2). In that light, we also observe that in response to (RQ1) , the Linear Relaxation and ROC Mitigation manipulation strategies are the most effective for a manipulative platform. 6. Related Work Fairwashing and rationalization Addressing fairness issues often requires compromising model performance for advantaged groups which can discourage companies from embracing fair training practices (Zietlow et al., 2022; Zhao & Gordon, 2022). Companies have two incentives to pay attention to the impact of their system on society. The first incentive comes from regulatory efforts such as the Algorithmic Accountability Act (AAA) (Congress, 2022) (US) and the Digital Markets Act (DMA) (Union, 2022) (EU) that impose fairness, transparency, and accountability constraints on large digital platforms. Yet, how to enforce these regulations is still an open problem (Cr emer et al., 2023). The second incentive is public image. Since fairness, transparency and accountability are laudable goal, audits, investigative journalism and certifications (Costanza-Chock et al., 2022) should force companies to pay attention to these objectives. However, both incentives are external: the platform just has to appear fair, transparent and accountable. This rationalization risk has been studied in the context of explanations fairwashing (A ıvodji et al., 2019; 2021; Shamsabadi et al., 2022). Fairness auditing Fairness auditing evaluates ML models to ensure fairness and accountability, often without access to proprietary model internals (Ng, 2021). This blackbox auditing approach relies on querying the model and analyzing its outputs against pre-defined fairness metrics (Birhane et al., 2024; de Vos et al., 2024). Current attempts to enhance fairness audits with tangible guarantees draw inspiration from hypothesis testing (Si et al., 2021; Taskesen et al., 2021; Di Ciccio et al., 2020; Cen & Alur, 2024; Cherian & Cand es, 2024; B enesse et al., 2024), online fairness auditing (Chugg et al., 2023; Maneriker et al., 2023), and formal methods for fairness certification (Albarghouthi et al., 2017; Ghosh et al., 2021; 2022; Borca-Tasciuc et al., 2022). Beyond statistical methods, the work of Yadav et al. explore the role of explanations in the auditing process (Yadav et al., 2022). Recent works also stress the importance of broadening the lens of algorithm auditing by incorporating user perspectives and sociotechnical factors (Lam et al., 2023; Deng et al., 2023). On another line of research, Confidential-PROFITT and Fair Proof propose to integrate cryptographic techniques in cooperation with the platforms, to ensure the faithfulness of platform responses during audits (Yadav et al., 2024; Shamsabadi et al., 2023; Waiwitlikhit et al., 2024); this is, however, more intrusive and technically restrictive, and thus awaits for adoption. Manipulating audits Manipulating fairness audits is an active area of research. Auditors can be fooled by biased sampling when the decision maker is allowed to publish a labeled dataset as proof of model fairness (Fukuchi et al., 2020). Adversarial attacks on explanation methods, such as LIME and SHAP, can be employed to produce misleading interpretations of model behavior (Fokkema et al., 2023; Shamsabadi et al., 2022; Laberge et al., 2023; Slack et al., 2020; Anders et al., 2020; A ıvodji et al., 2019; Le Merrer & Tr edan, 2020). Platforms can also modify the output of their models to create the appearance of fairness without addressing underlying biases (Yan & Zhang, 2022; Garcia Bourr ee et al., 2023; Godinot et al., 2024).. However, the challenge of designing audits that are robust to advanced manipulation strategies remains open. The idea of using auditor prior knowledge that we formalize in this work has been implicitly studied in different contexts. Based on active learning techniques work has studied how auditors could leverage knowledge about the hypothesis class (Yan & Zhang, 2022; Godinot et al., 2024). In a more practical setting, Tan et al. studied using model distillation methods (Tan et al., 2018) to use prior about the ground truth and hypothesis class (Tan et al., 2018). 7. Conclusion and Discussion We investigated, both theoretically and experimentally, the conditions under which an auditor can or cannot be manipulated when auditing with a prior. We introduced an empirical method for tuning the manipulation detection threshold to maximize the auditor s probability of detecting malicious platforms. While our work offers regulators a framework for defending against audit manipulations, the path to accountability Robust ML Auditing using Prior Knowledge extends much further. A significant gap remains between audit evaluations and the actual mitigation of identified issues (Raji et al., 2021; Mukobi, 2024). Moreover, one-time audits are inherently limited, as platforms can alter their models in harmful ways after the audit has concluded. Addressing these challenges in future work will require the development of continuous or adaptive auditing mechanisms, potentially incorporating auditor priors, to ensure sustained accountability and fairness. Impact Statement This work provides both theoretical and empirical analyses of fairness audits in ML decision-making systems, with a focus on their vulnerability to strategic manipulations by platforms aiming to evade regulatory scrutiny. By demonstrating how auditors access to prior knowledge can enhance the robustness of black-box audits, we offer actionable insights for mitigating potential audit-a manipulations. Our findings have important implications for policymakers, auditors, and ML practitioners, underscoring the urgent need for rigorous auditing frameworks resilient to adversarial behavior. The societal impact of this work is twofold. On the positive side, strengthening the robustness of fairness audits promotes greater accountability for platforms deploying ML models in high-stakes domains such as finance and healthcare. By mapping the risk landscape of audit manipulation, our approach advances the development of more trustworthy ML systems. However, we also draw attention to the limitations of current audit practices, showing that over-reliance on public priors can be exploited by strategic actors. Acknowledgements This work of Martijn de Vos, Milos Vujasinovic, Sayan Biswas, and Anne-Marie Kermarrec has been funded by the Swiss National Science Foundation, under the project FRIDAY: Frugal, Privacy-Aware and Practical Decentralized Learning , SNSF proposal No. 10.001.796. Jade Garcia Bourr ee, Augustin Godinot, Gilles Tr edan and Erwan Le Merrer acknowledge the support of the French Agence Nationale de la Recherche (ANR), under grant ANR-24-CE237787 (project PACMAM). A ıvodji, U., Arai, H., Fortineau, O., Gambs, S., Hara, S., and Tapp, A. Fairwashing: the risk of rationalization. In Chaudhuri, K. and Salakhutdinov, R. (eds.), Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pp. 161 170. PMLR, 2019. A ıvodji, U., Arai, H., Gambs, S., and Hara, S. Characterizing the risk of fairwashing. In Ranzato, M., Beygelzimer, A., Dauphin, Y. N., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, December 6-14, 2021, virtual, pp. 14822 14834, 2021. Ajarra, A., Ghosh, B., and Basu, D. Active Fourier Auditor for Estimating Distributional Properties of ML Models, 2024. Albarghouthi, A., D Antoni, L., Drews, S., and Nori, A. V. Fair Square: Probabilistic verification of program fairness. Proc. ACM Program. Lang., 1(OOPSLA):80:1 80:30, 2017. doi: 10.1145/3133904. Anders, C. J., Pasliev, P., Dombrowski, A., M uller, K., and Kessel, P. Fairwashing explanations with off-manifold detergent. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp. 314 323. PMLR, 2020. B enesse, C., Gamboa, F., Loubes, J.-M., and Boissin, T. Fairness seen as global sensitivity analysis. Machine Learning, 113(5):3205 3232, 2024. ISSN 1573-0565. doi: 10.1007/s10994-022-06202-y. Birhane, A., Steed, R., Ojewale, V., Vecchione, B., and Raji, I. D. AI auditing: The Broken Bus on the Road to AI Accountability. In 2024 IEEE Conference on Secure and Trustworthy Machine Learning (Sa TML), pp. 612 643, 2024. doi: 10.1109/Sa TML59370.2024.00037. Borca-Tasciuc, G., Guo, X., Bak, S., and Skiena, S. Provable Fairness for Neural Network Models using Formal Verification, 2022. Buyl, M. and Bie, T. D. Optimal transport of classifiers to fairness. In Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., and Oh, A. (eds.), Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, Neur IPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022, 2022. Calders, T., Kamiran, F., and Pechenizkiy, M. Building Classifiers with Independency Constraints. In 2009 IEEE International Conference on Data Mining Workshops, pp. 13 18, 2009. doi: 10.1109/ICDMW.2009.83. Caton, S. and Haas, C. Fairness in Machine Learning: A Survey. ACM Comput. Surv., 56(7):166:1 166:38, 2024. ISSN 0360-0300. doi: 10.1145/3616865. Cen, S. H. and Alur, R. From Transparency to Accountability and Back: A Discussion of Access and Evidence in AI Auditing, 2024. Robust ML Auditing using Prior Knowledge Cherian, J. J. and Cand es, E. J. Statistical Inference for Fairness Auditing. Journal of Machine Learning Research, 25(149):1 49, 2024. ISSN 1533-7928. Chugg, B., Cortes-Gomez, S., Wilder, B., and Ramdas, A. Auditing fairness by betting. In Oh, A., Naumann, T., Globerson, A., Saenko, K., Hardt, M., and Levine, S. (eds.), Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, Neur IPS 2023, New Orleans, LA, USA, December 10 - 16, 2023, 2023. Congress, U. Algorithmic accountability act of 2022. https://www.congress.gov/bill/ 117th-congress/house-bill/6580, 2022. Costanza-Chock, S., Raji, I. D., and Buolamwini, J. Who audits the auditors? recommendations from a field scan of the algorithmic auditing ecosystem. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, pp. 1571 1583, 2022. Cr emer, J., Dinielli, D., Heidhues, P., Kimmelman, G., Monti, G., Podszun, R., Schnitzer, M., Scott Morton, F., and De Streel, A. Enforcing the digital markets act: institutional choices, compliance, and antitrust. Journal of Antitrust Enforcement, 11(3):315 349, 2023. de Vos, M., Dhasade, A., Garcia Bourr ee, J., Kermarrec, A.-M., Le Merrer, E., Rottembourg, B., and Tredan, G. Fairness auditing with multi-agent collaboration. In ECAI 2024, pp. 1116 1123. IOS Press, 2024. Deng, W. H., Guo, B. B., De Vrio, A., Shen, H., Eslami, M., and Holstein, K. Understanding practices, challenges, and opportunities for user-engaged algorithm auditing in industry practice. In Schmidt, A., V a an anen, K., Goyal, T., Kristensson, P. O., Peters, A., Mueller, S., Williamson, J. R., and Wilson, M. L. (eds.), Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, CHI 2023, Hamburg, Germany, April 23-28, 2023, pp. 377:1 377:18. ACM, 2023. doi: 10.1145/3544548.3581026. Di Ciccio, C., Vasudevan, S., Basu, K., Kenthapadi, K., and Agarwal, D. Evaluating fairness using permutation tests. In Gupta, R., Liu, Y., Tang, J., and Prakash, B. A. (eds.), KDD 20: The 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, CA, USA, August 23-27, 2020, pp. 1467 1477. ACM, 2020. Ding, F., Hardt, M., Miller, J., and Schmidt, L. Retiring adult: New datasets for fair machine learning. In Ranzato, M., Beygelzimer, A., Dauphin, Y. N., Liang, P., and Vaughan, J. W. (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, December 6-14, 2021, virtual, pp. 6478 6490, 2021. Fokkema, H., de Heide, R., and van Erven, T. Attributionbased Explanations that Provide Recourse Cannot be Robust. Journal of Machine Learning Research, 24(360): 1 37, 2023. ISSN 1533-7928. Fukuchi, K., Hara, S., and Maehara, T. Faking fairness via stealthily biased sampling. In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7-12, 2020, pp. 412 419. AAAI Press, 2020. Garcia Bourr ee, J., Le Merrer, E., Tredan, G., and Rottembourg, B. On the relevance of APIs facing fairwashed audits, 2023. Ghosh, B., Basu, D., and Meel, K. S. Justicia: A stochastic SAT approach to formally verify fairness. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Thirty-Third Conference on Innovative Applications of Artificial Intelligence, IAAI 2021, The Eleventh Symposium on Educational Advances in Artificial Intelligence, EAAI 2021, Virtual Event, February 2-9, 2021, pp. 7554 7563. AAAI Press, 2021. Ghosh, B., Basu, D., and Meel, K. S. Algorithmic fairness verification with graphical models. In Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, pp. 9539 9548. AAAI Press, 2022. Godinot, A., Le Merrer, E., Tr edan, G., Penzo, C., and Ta ıani, F. Under manipulations, are some AI models harder to audit? In 2024 IEEE Conference on Secure and Trustworthy Machine Learning (Sa TML), pp. 644 664, 2024. doi: 10.1109/Sa TML59370.2024.00038. Hardt, M., Price, E., and Srebro, N. Equality of opportunity in supervised learning. In Lee, D. D., Sugiyama, M., von Luxburg, U., Guyon, I., and Garnett, R. (eds.), Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona, Spain, pp. 3315 3323, 2016. Jiang, R., Pacchiano, A., Stepleton, T., Jiang, H., and Chiappa, S. Wasserstein fair classification. In Globerson, A. and Silva, R. (eds.), Proceedings of the Thirty-Fifth Conference on Uncertainty in Artificial Intelligence, UAI Robust ML Auditing using Prior Knowledge 2019, Tel Aviv, Israel, July 22-25, 2019, volume 115 of Proceedings of Machine Learning Research, pp. 862 872. AUAI Press, 2019. Kamiran, F., Karim, A., and Zhang, X. Decision theory for discrimination-aware classification. In 2012 IEEE 12th international conference on data mining, pp. 924 929. IEEE, 2012. Kim, M. P., Ghorbani, A., and Zou, J. Multiaccuracy: Blackbox post-processing for fairness in classification. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, pp. 247 254, 2019. Laberge, G., A ıvodji, U., Hara, S., Marchand, M., and Khomh, F. Fooling SHAP with stealthily biased sampling. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. Open Review.net, 2023. Lam, M. S., Pandit, A., Kalicki, C. H., Gupta, R., Sahoo, P., and Metaxa, D. Sociotechnical Audits: Broadening the Algorithm Auditing Lens to Investigate Targeted Advertising. Proc. ACM Hum.-Comput. Interact., 7(CSCW2): 360:1 360:37, 2023. doi: 10.1145/3610209. Le Merrer, E. and Tr edan, G. Remote explainability faces the bouncer problem. Nature Machine Intelligence, 2 (9):529 539, 2020. ISSN 2522-5839. doi: 10.1038/ s42256-020-0216-z. Lecun, Y., Bottou, L., Bengio, Y., and Haffner, P. Gradientbased learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. doi: 10.1109/5.726791. Li, S. Concise formulas for the area and volume of a hyperspherical cap. Asian Journal of Mathematics & Statistics, 4(1):66 70, 2010. Liu, Z., Luo, P., Wang, X., and Tang, X. Deep learning face attributes in the wild. In 2015 IEEE International Conference on Computer Vision, ICCV 2015, Santiago, Chile, December 7-13, 2015, pp. 3730 3738. IEEE Computer Society, 2015. doi: 10.1109/ICCV.2015.425. Lohaus, M., Perrot, M., and von Luxburg, U. Too relaxed to be fair. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pp. 6360 6369. PMLR, 2020. Maneriker, P., Burley, C., and Parthasarathy, S. Online fairness auditing through iterative refinement. In Singh, A. K., Sun, Y., Akoglu, L., Gunopulos, D., Yan, X., Kumar, R., Ozcan, F., and Ye, J. (eds.), Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD 2023, Long Beach, CA, USA, August 6-10, 2023, pp. 1665 1676. ACM, 2023. doi: 10.1145/3580305.3599454. Mehrabi, N., Morstatter, F., Saxena, N., Lerman, K., and Galstyan, A. A survey on bias and fairness in machine learning. ACM computing surveys (CSUR), 54(6):1 35, 2021. Mukobi, G. Reasons to doubt the impact of ai risk evaluations. Ar Xiv preprint, abs/2408.02565, 2024. Ng, A. Can auditing eliminate bias from algorithms?, 2021. Accessed: 2025-01-07. NIST. Nist digital library of mathematical functions, 2013. Release 1.0.6 of 2013-05-06. Petersen, E., Potdevin, Y., Mohammadi, E., Zidowitz, S., Breyer, S., Nowotka, D., Henn, S., Pechmann, L., Leucker, M., Rostalski, P., et al. Responsible and regulatory conform machine learning for medicine: a survey of challenges and solutions. IEEE Access, 10:58375 58418, 2022. Raji, D., Denton, E., Bender, E. M., Hanna, A., and Paullada, A. Ai and the everything in the whole wide world benchmark. In Vanschoren, J. and Yeung, S. (eds.), Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks, volume 1, 2021. Raji, I. D. The Anatomy of AI Audits: Form, Process, and Consequences. In Bullock, J. B., Chen, Y.-C., Himmelreich, J., Hudson, V. M., Korinek, A., Young, M. M., and Zhang, B. (eds.), The Oxford Handbook of AI Governance, pp. 0. Oxford University Press, 2024. doi: 10.1093/oxfordhb/9780197579329.013.28. Raji, I. D., Xu, P., Honigsberg, C., and Ho, D. Outsider Oversight: Designing a Third Party Audit Ecosystem for AI Governance. In Proceedings of the 2022 AAAI/ACM Conference on AI, Ethics, and Society, AIES 22, pp. 557 571, New York, NY, USA, 2022. Association for Computing Machinery. ISBN 978-1-4503-9247-1. doi: 10.1145/3514094.3534181. Ribeiro, M. H. Is Facebook Standard Algorithm Polarizing? https://doomscrollingbabel.manoel.xyz/p/isfacebook-standard-algorithm-polarizing, 2024. Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature machine intelligence, 1(5):206 215, 2019. Shamsabadi, A. S., Yaghini, M., Dullerud, N., Wyllie, S. C., A ıvodji, U., Alaagib, A., Gambs, S., and Papernot, N. Washing the unwashable : On the (im)possibility of fairwashing detection. In Koyejo, S., Mohamed, S., Agarwal, Robust ML Auditing using Prior Knowledge A., Belgrave, D., Cho, K., and Oh, A. (eds.), Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, Neur IPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022, 2022. Shamsabadi, A. S., Wyllie, S. C., Franzese, N., Dullerud, N., Gambs, S., Papernot, N., Wang, X., and Weller, A. Confidential-profitt: Confidential proof of fair training of trees. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. Open Review.net, 2023. Si, N., Murthy, K., Blanchet, J. H., and Nguyen, V. A. Testing group fairness via optimal transport projections. In Meila, M. and Zhang, T. (eds.), Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 9649 9659. PMLR, 2021. Slack, D., Hilgard, S., Jia, E., Singh, S., and Lakkaraju, H. Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, AIES 20, pp. 180 186, New York, NY, USA, 2020. Association for Computing Machinery. ISBN 978-1-4503-7110-0. doi: 10.1145/3375627.3375830. Tan, S., Caruana, R., Hooker, G., and Lou, Y. Distilland-Compare: Auditing Black-Box Models Using Transparent Model Distillation. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, AIES 18, pp. 303 310, New York, NY, USA, 2018. Association for Computing Machinery. ISBN 978-1-4503-6012-8. doi: 10.1145/3278721.3278725. Taskesen, B., Blanchet, J., Kuhn, D., and Nguyen, V. A. A Statistical Test for Probabilistic Fairness. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAcc T 21, pp. 648 665, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 978-1-4503-8309-7. doi: 10.1145/3442188.3445927. Timberg, C. Facebook made big mistake in data it provided to researchers, undermining academic work. Washington Post, 2021. ISSN 0190-8286. Union, E. Regulation (eu) 2022/1925 of the european parliament and of the council of 14 september 2022 on contestable and fair markets in the digital sector (digital markets act). https://eur-lex.europa.eu/eli/ reg/2022/1925/oj/eng, 2022. Waiwitlikhit, S., Stoica, I., Sun, Y., Hashimoto, T., and Kang, D. Trustless audits without revealing data or models. In Forty-first International Conference on Machine Learning, ICML 2024, Vienna, Austria, July 21-27, 2024. Open Review.net, 2024. West, D. Neural network credit scoring models. Computers & operations research, 27(11-12):1131 1152, 2000. Yadav, C., Moshkovitz, M., and Chaudhuri, K. XAudit : A Theoretical Look at Auditing with Explanations, 2022. Yadav, C., Chowdhury, A. R., Boneh, D., and Chaudhuri, K. Fairproof : Confidential and certifiable fairness for neural networks. In Forty-first International Conference on Machine Learning, ICML 2024, Vienna, Austria, July 21-27, 2024. Open Review.net, 2024. Yan, T. and Zhang, C. Active fairness auditing. In Chaudhuri, K., Jegelka, S., Song, L., Szepesv ari, C., Niu, G., and Sabato, S. (eds.), International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pp. 24929 24962. PMLR, 2022. Zhao, H. and Gordon, G. J. Inherent tradeoffs in learning fair representations. J. Mach. Learn. Res., 23:57:1 57:26, 2022. Zietlow, D., Lohaus, M., Balakrishnan, G., Kleindessner, M., Locatello, F., Sch olkopf, B., and Russell, C. Leveling down in computer vision: Pareto inefficiencies in fair deep classifiers. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022, pp. 10400 10411. IEEE, 2022. doi: 10.1109/CVPR52688.2022.01016. Robust ML Auditing using Prior Knowledge Table 1. Notations H HYPOTHESIS CLASS F SET OF FAIR MODELS Ha SET OF EXPECTABLE MODELS ha GROUND TRUTH δ DISTANCE BETWEEN THE GROUNDTRUTH AND THE SET OF EXPECTABLE MODEL hp ORIGINAL MODEL OF THE PLATFORM hm MANIPULATED MODEL OF THE PLATFORM X INPUT SPACE D DATA DISTRIBUTION X SAMPLE FROM INPUT SPACE Y OUTPUT SPACE Y SAMPLE FROM OUTPUT SPACE A PROTECTED FEATURE Z SAMPLE SPACE Z SAMPLE n DIMENSION OF Z A. Proofs and additional theoretical results As in (Buyl & Bie, 2022), let Z X A {0, 1} denote the sample space, from which the auditor draws samples Z (X, A, Y ). The auditor sample the binary predictions ˆY {0, 1} from a probabilistic classifier h : X [0, 1] that assigns a score h(X) to the belief that a sample with features X belongs to the positive class. It is assumed that X Rd X and A = {[A0, A1]/A0, A1 {0, 1}} = {[1, 0], [0, 1]}(the one-hot encoding of the protected feature with two groups). We also assume that Ha is an open set of Z. We denote F the set of all score functions f : X {0, 1} that satisfy (PDP): F {f : X {0, 1} : EZ [g(Z)f(X)] = 0n} with k [2], gk = Ak EZ[Ak] 1, 0n a vector of n = d F zeros. Assuming that the predictions ˆY |X are randomly sampled from a probabilistic classifier h(X), then the traditional fairness notion of demographic parity (DP) is equivalent to PDP. But if ˆY is not sampled from h(X) but instead decided by a threshold, DPD is a relaxation of the actual DP notion. That is to say, F is the set of all score functions that are fair regarding the demographic parity on A. As F is the kernel of the linear transformation f : EZ [g(Z)f(X)], F is a hyperplane of Z. As F is a hyperplane of Z, it is dense or closed in Z. A.1. Cases where F is dense in Z. Lemma A.1. If F is dense in Z, the auditor has a probability to detects it as manipulated equals to zero. Proof. If F is dense in Z then for every function f Z, every open neighborhood of f intersects F. In particular, it always exists a model hm F that is in a neighborhood of hp and in Ha. In that case, hm is fair and expectable, so the auditor has a probability to detects it as manipulated equals to zero. This case is a pathological case where the platform can still appear fair and honest. For the next theoretical results, we are interested in the case where F is closed in Z. A.2. Cases where F is closed in Z. If F is a hyperplan closed in Z, it has an empty interior (i.e. F = ) as its codimension is 1. In the following, we can thus use F instead of F, as both are equals. Robust ML Auditing using Prior Knowledge Similarly, we can define the normal vector to F which is actually the vector that is used for all the projections we use in this paper. In Equation (6), we defined h m = proj F(hp) (i.e. h m is the orthographic projection of the expectable model hp in the set of fair models F). Having an hyperplan lead to the natural definition of (hyper)cylinder, that we use in the following theorem. Definition A.2. A right cylinder C(H, B) is the set of all points whose orthographic projection on a hyperplane H lies in a set B with B a subset of the boundary of H. B is called the base of the cylinder. Theorem A.3. The probability Puf that the auditor correctly detects a malicious platform trying to be fair is P(Ha\C(F, Ha F)|Ha). Proof. The auditor correctly detects a malicious platform trying to be fair if and only if the manipulated model is fair but not expectable. The manipulated model is fair but not expectable if and only if the orthographic projection h m of hp in F is not in Ha F. Thus, the manipulated model is fair but not expectable if and only if hp / C(F, Ha F) (following Definition A.2). As by assumption hp Ha (Equation (4)), it means that hp Ha\C(F, Ha F). The auditor correctly detects a malicious platform trying to be fair with probability P(Ha\C(F, Ha F)|Ha). Theorem 4.3 is a special case of Theorem A.3 with additional assumption. We now prove the main Theorem Theorem 4.3. Theorem 4.3 (Prior-Uniform detection rate). Under the dataset prior of Definition 4.1 with L defined as the ℓ2 norm, and the uninformative prior assumption, the probability that the auditor correctly detects a malicious platform trying to be fair is Z arccos(δ/τ) 0 sinn(θ)dθ δ with δ = d(ha, F), the distance of ha to F and Wn is the n-term of Wallis integrals. Proof. As established in Theorem A.3, Puf = P(Ha\C(F, Ha F)|Ha). The probability P(Ha\C(F, Ha F)|Ha) is the probability to be in the ball Ha without the probability to be in the intersection between the ball Ha and the cylinder C(F, Ha F). In the following, we denote V # n (τ, δ) this quantity. As Ha is a ball, its volume is: V ball n (τ) = πn/2τ n with Γ(z) = R 0 tz 1e tdt (NIST, 2013). The volume of the intersection between the cylinder and the ball is the sum of the three following volumes: the solid cylinder with height between δ and δ the spherical cap of Ha that is above the previous cylinder (i.e. the part of Ha with height between δ and τ) the spherical cap of Ha that is bellow the previous cylinder (i.e. the part of Ha with height between δ and τ) According to (Li, 2010), the volume of each spherical cap is V cap n (τ, δ) = π(n 1)/2τ n Z arccos(δ/τ) 0 sinn(θ)dθ And the volume of the cylinder of height 2δ is V cylinder n (τ, δ) = 2δV ball n 1( p V # n (τ, δ) = V ball n (τ) 2V cap n (τ, δ) V cylinder n (τ, δ) Robust ML Auditing using Prior Knowledge 2 ) 2π(n 1)/2τ n Z arccos(δ/τ) 0 sinn(θ)dθ 2δ π(n 1)/2( According to Theorem A.3, the probability that the auditor correctly detects a malicious platform trying to be fair is P(Ha\C(F, Ha F)|Ha). That is to say, it is the ratio of V # n (τ, δ) over V ball n (τ): Puf = P(Ha\C(F, Ha F)|Ha) = V # n (τ, δ) V ball n (τ) = 1 2Γ( n+2 2 ) π(n 1)/2 Z arccos(δ/τ) 0 sinn(θ)dθ 2δ (τ 2 δ2)(n 1)/2 2 ) π(n 1)/2 = 1 2 π Γ( n+2 Z arccos(δ/τ) 0 sinn(θ)dθ 2δ π (τ 2 δ2)(n 1)/2 = 1 2 π Γ( n+2 Z arccos(δ/τ) 0 sinn(θ)dθ δ (τ 2 δ2)(n 1)/2 The function Γ can be written with Wallis integrals as: Wn = π 2 ) with n, Wn = R π/2 0 sinn(θ)dθ. In the other hand, δ (τ 2 δ2)(n 1)/2 τ (τ 2 δ2)(n 1)/2 Thus, Puf = 1 1 Wn R arccos(δ/τ) 0 sinn(θ)dθ δ τ 2 (n 1)/2 . Before dealing with this complete expression, we propose some particular cases that are easily interpretable. Corollary A.4. If Ha is a ball centered in the ground-truth ha that is tangent to F, then the auditor has a probability one to correctly detect a malicious platform trying to be fair. Ha = B(ha, τ) τ = δ = Puf = 1. with δ = d(ha, F), the distance of ha to F. Proof. If Ha is tangent to F then δ = τ. Thus, arccos(δ/τ) = arccos(1) = 0 and R arccos(δ/τ) 0 sinn(θ)dθ = 0. Thanks to the formula of Theorem 4.3 with δ/τ = 1, Puf = 1 1 Wn (0 0) = 1. If Ha is tangent to F, Puf = 1. This corollary means that by reducing the threshold τ to the minimal value (δ), the auditor is sure to detect any manipulation of the platform. Robust ML Auditing using Prior Knowledge Corollary A.5. If Ha is a ball centered in the ground-truth ha that is fair, then the auditor has a probability zero to correctly detect a malicious platform trying to be fair. Ha = B(ha, τ) ha F = Puf = 0. Proof. If ha F then δ = 0 and arccos(δ/τ) = arccos(0) = π/2 in the formula of Theorem 4.3. Thus, Puf = 1 1 Wn (Wn 0) = 0. This last case is the case where the ha of the auditor is fair. Intuitively, if ha is fair, half of the model that the platform can construct are naturally fair and the other half are naturally unfair. Thus, it is very easy to change from an unfair model to a fair model without changing too much the honest model. Thus, detecting such manipulation is very hard for the auditor. Now, we study the general expression of Puf in Theorem 4.3. In particular, we study a lower bound of Puf to study when the probability is strictly positive. Corollary 4.4 (Detection rate lower bound). If n is even, (n 1)/2 Puf 1. Proof. Puf = 1 1 Wn R arccos(δ/τ) 0 sinn(θ)dθ δ τ 2 (n 1)/2 Wn = R arccos(δ/τ) 0 sinn(θ)dθ + R π/2 arccos(δ/τ) sinn(θ)dθ So, 1 Wn R arccos(δ/τ) 0 sinn(θ)dθ 1 (with n even). And Puf 1 Wn δ τ 1 δ2 τ 2 (n 1)/2 Lemma A.6. The lower bound according to δ/τ has two extremums that are for δ/τ = 1 or δ/τ = γ with γ = n+3 n 1 Remark. Note that γ only depend on the dimension n and leads to 0 when n leads to infinity. Proof. We define fn (the lower bound) s.t. fn(δ, τ) = δ Wnτ Change of variable x = δ τ , f(x) = x W n(1 x2)(n 1)/2. We are interested in cases where τ > δ, i.e. 0 < x < 1. Moreover, f has an extremum iff f = 0 somewhere in [0, 1]. x [0, 1], Wnf (x) = (1 x2)(n 1)/2 (n 1)x2(1 x2)(n 3)/2 = (1 x2)(n 3)/2(x2 + i.e. f (x) = 0 for the following elements: Robust ML Auditing using Prior Knowledge 2 > 1 (if n 2) So f has two local extremums in [0, 1], one for 1 and one for γ = n+3 n 1