# fairgrad_fairness_aware_gradient_descent__51096a08.pdf Published in Transactions on Machine Learning Research (08/2023) Fair Grad: Fairness Aware Gradient Descent Gaurav Maheshwari gaurav.maheshwari@inria.fr Univ. Lille, Inria, CNRS, Centrale Lille, UMR 9189 - CRISt AL, F-59000 Lille, France Michaël Perrot michael.perrot@inria.fr Univ. Lille, Inria, CNRS, Centrale Lille, UMR 9189 - CRISt AL, F-59000 Lille, France Reviewed on Open Review: https://openreview.net/forum?id=0f8t U3Qw WD We address the problem of group fairness in classification, where the objective is to learn models that do not unjustly discriminate against subgroups of the population. Most existing approaches are limited to simple binary tasks or involve difficult to implement training mechanisms which reduces their practical applicability. In this paper, we propose Fair Grad, a method to enforce fairness based on a re-weighting scheme that iteratively learns group specific weights based on whether they are advantaged or not. Fair Grad is easy to implement, accommodates various standard fairness definitions, and comes with minimal overhead. Furthermore, we show that it is competitive with standard baselines over various datasets including ones used in natural language processing and computer vision. Fair Grad is available as a Py PI package at - https://pypi.org/project/fairgrad 1 Introduction Fair Machine Learning addresses the problem of learning models that are free of any discriminatory behavior against a subset of the population. For instance, consider a company developing a model to predict whether a person would be a suitable hire based on their biography. A possible source of discrimination here can be if, in the data available to the company, individuals that are part of a subgroup formed based on their gender, ethnicity, or other sensitive attributes, are consistently labelled as unsuitable hires regardless of their true competency due to historical bias. This kind of discrimination can be measured by a fairness notion called Demographic Parity (Calders et al., 2009). If the data is unbiased, another source of discrimination may be the model itself that consistently mislabels the competent individuals of a subgroup as unsuitable hires. This can be measured by a fairness notion called Equality of Opportunity (Hardt et al., 2016). Several such fairness notions have been proposed in the literature as different problems call for different measures. They can be divided into two major paradigms, namely (i) Individual Fairness (Dwork et al., 2012; Kusner et al., 2017) where the idea is to treat similar individuals similarly regardless of the sensitive group they belong to, and (ii) Group Fairness (Calders et al., 2009; Hardt et al., 2016; Zafar et al., 2017a; Denis et al., 2021) where the underlying idea is that no sensitive group should be disadvantaged compared to the overall reference population. In this paper, we focus on group fairness in the context of classification where we only assume access to the sensitive attributes during the training phase. The existing approaches for group fairness in Machine Learning may be divided into three main paradigms. First, pre-processing methods aim at modifying a dataset to remove any intrinsic unfairness that may exist in the examples. The underlying idea is that a model learned on this modified data is more likely to be fair (Dwork et al., 2012; Kamiran & Calders, 2012; Zemel et al., 2013; Feldman et al., 2015; Calmon et al., 2017). Then, post-processing approaches modify the predictions of an accurate but unfair model so that it becomes fair (Kamiran et al., 2010; Hardt et al., 2016; Woodworth et al., 2017; Iosifidis et al., 2019; Chzhen et al., 2019). Finally, in-processing methods aim at learning a model that is fair and accurate in a single step (Calders & Verwer, 2010; Kamishima et al., 2012; Goh et al., 2016; Zafar et al., 2017a;b; Donini et al., Published in Transactions on Machine Learning Research (08/2023) # The library is available at https://pypi.org/project/fairgrad. from fairgrad.torch import Cross Entropy Loss # Same as Py Torch's loss with some additional meta data. # A fairness rate of 0.01 is a good rule of thumb for standardized data. criterion = Cross Entropy Loss(y_train, s_train, fairness_measure, fairness_rate=0.01) # The dataloader and model are defined and used in the standard way. for x, y, s in data_loader: optimizer.zero_grad() loss = criterion(model(x), y, s) loss.backward() optimizer.step() Figure 1: A standard training loop where the Py Torch s loss is replaced by Fair Grad s loss. 2018; Krasanakis et al., 2018; Agarwal et al., 2018; Wu et al., 2019; Cotter et al., 2019; Iosifidis & Ntoutsi, 2019; Jiang & Nachum, 2020; Lohaus et al., 2020; Roh et al., 2020; Ozdayi et al., 2021). In this paper, we propose a new in-processing group fairness approach based on a re-weighting scheme that may also be used as a kind of post-processing approach by fine-tuning existing classifiers. Motivation. In-processing approaches can be further divided into several sub-categories (Caton & Haas, 2020). Common amongst them are methods that cast the fairness task as a constrained optimization problem, and then relax the fairness constraints under consideration to simplify the learning process (Zafar et al., 2017a; Donini et al., 2018; Wu et al., 2019). Indeed, standard fairness notions are usually difficult to handle due to their non-convexity and non-differentiability. Unfortunately, these relaxations may be far from the actual fairness measures, leading to sub-optimal models (Lohaus et al., 2020). Similarly, several approaches address the fairness problem by designing specific algorithms and solvers. This is, for example, done by reducing the optimization procedure to a simpler problem (Agarwal et al., 2018), altering the underlying solver (Cotter et al., 2019), or using adversarial learning (Raff& Sylvester, 2018). However, these approaches are difficult to adapt to existing systems as they require special training procedures or changes in the model. They are also limited in the range of problems to which they can be applied. For example, the work of Agarwal et al. (2018) can only be applied in a binary classification setting, while the work of Ozdayi et al. (2021) is limited to two sensitive groups. Furthermore, they may come with several hyper-parameters that need to be carefully tuned to obtain fair models. For instance, the scaling parameter in adversarial learning (Raff& Sylvester, 2018; Li et al., 2018) or the number of iterations in inner optimization for bi-level optimization based mechanisms (Ozdayi et al., 2021). The complexity of the existing methods might hinder their deployment in practical settings. Hence, there is a need for simpler methods that are straightforward to integrate into existing training loops. Contributions. In this paper, we present Fair Grad, a general purpose approach to enforce fairness in empirical risk minimization solved using gradient descent. We propose to dynamically update the influence of the examples after each gradient descent update to precisely reflect the fairness level of the models obtained at each iteration and guide the optimization process in a relevant direction. Hence, the underlying idea is to use lower weights for examples from advantaged groups than those from disadvantaged groups. Our method is inspired by recent re-weighting approaches that also propose to change the importance of each group while learning a model (Krasanakis et al., 2018; Iosifidis & Ntoutsi, 2019; Jiang & Nachum, 2020; Roh et al., 2020; Ozdayi et al., 2021). We discuss these works in Appendix A. Interestingly, we also find that Fair Grad can be seen as solving a kind of constrained optimization problem. In Section 2.2, we expand upon this link and show how Fair Grad can be seen as a solution that connects these two kinds of methods. A key advantage of Fair Grad is that it is straightforward to incorporate into standard gradient based solvers that support examples re-weighting like Stochastic Gradient Descent. Hence, we developed a Python library (provided in the supplementary material) where we augmented standard Py Torch losses to accommodate our approach. From a practitioner point of view, it means that using Fair Grad is as simple as replacing their existing loss from Py Torch with our custom loss and passing along some meta data, while the rest of Published in Transactions on Machine Learning Research (08/2023) the training loop remains identical. This is illustrated in Figure 1. It is interesting to note that Fair Grad only brings one extra hyper-parameter, the fairness rate, besides the usual optimization ones (learning rates, batch size, . . . ). Moreover, Fair Grad incurs minimal computational overhead during training as it relies on objects that are already computed for standard gradient descent, namely the predictions on the current batch and the loss incurred by the model for each example. In particular, the overhead is independent of the number of parameters of the model. Furthermore, as many in-processing approaches in fairness (Cotter et al., 2019; Roh et al., 2020), Fair Grad does not introduce any overhead at test time. Overall, Fair Grad is a lightweight fairness solution that is compatible with various group fairness notions, including exact and approximate fairness, can handle both multiple sensitive groups and multiclass problems, and can fine tune existing unfair models. Through extensive experiments, we also show that, in addition to its versatility, Fair Grad is competitive with several standard baselines in fairness on both standard datasets as well as complex natural language processing and computer vision tasks. 2 Problem Setting, Notations, and Related Work In the remainder of this paper, we assume that we have access to a feature space X, a finite discrete label space Y, and a set S of values for the sensitive attribute. We further assume that there exists a distribution D DZ where DZ is the set of all distributions over Z = X Y S. Our goal is then to learn an accurate model hθ H, with learnable parameters θ Rd, such that hθ : X Y is fair with respect to a given fairness definition that depends on the sensitive attribute. In Section 2.1, we formally define the family of fairness measures that are compatible with our approach and provide several examples of popular notions encompassed by our fairness definition. As usual in machine learning, we will assume that D is unknown and that we only get to observe a finite dataset T = {(xi, yi, si)}n i=1 of n examples drawn i.i.d. from D. Let P (E(X, Y, S)) represent the probability that an event E happens with respect to (X, Y, S) D while b P (E(x, y, s)) = 1 n Pn i=1 IE(xi,yi,si) is an empirical estimate with respect to T where IP is the indicator function which is 1 when the property P is verified and 0 otherwise. In the remainder of this paper, all our derivations will be considered in the finite sample setting and we will assume that what was measured on our finite sample is sufficiently close to what would be obtained if one had access to the overall distribution. This seems reasonable in light of the previous work on generalization in standard machine learning (Shalev-Shwartz & Ben-David, 2014) and the recent work of Woodworth et al. (2017) or Mangold et al. (2022) which show that the kind of fairness measures we consider in this paper tend to generalize well when the hypothesis space is not too complex, as measured respectively by the VC or the Natarajan Dimension (Shalev-Shwartz & Ben-David, 2014). Since these generalization results only rely on a capacity measure of the hypothesis space and are otherwise algorithm agnostic, they are applicable to the models returned by Fair Grad when they have finite VC or Natarajan dimensions. This is for example the case for linear models. 2.1 Fairness Definition We assume that the data may be partitioned into K disjoint groups denoted T1, . . . , Tk, . . . , TK such that SK k=1 Tk = T and TK k=1 Tk = . These groups highly depend on the fairness notion under consideration. They might correspond to the usual sensitive groups, as is the case for Accuracy Parity (see Example 1), or might be subgroups of the usual sensitive groups, as in Equalized Odds where the subgroups are defined with respect to the true labels (see Example 2 in Appendix B). For each group, we assume that we have access to a function b Fk : Dn H R such that b Fk > 0 when the group k is advantaged by the given classifier and b Fk < 0 when the group k is disadvantaged. Furthermore, we assume that the magnitude of b Fk represents the degree to which the group is (dis)advantaged. Finally, we assume that each b Fk can be rewritten as follows: b Fk(T , hθ) = C0 k + k =1 Ck k b P (hθ(x) = y|Tk ) (1) where the constants C are group specific and independent of hθ. The probabilities b P (hθ(x) = y|Tk ) represent the error rates of hθ over each group Tk with a slight abuse of notation. Below, we show that Accuracy Published in Transactions on Machine Learning Research (08/2023) Parity (Zafar et al., 2017a) respects this definition. In Appendix B, we show that Equality of Opportunity (Hardt et al., 2016), Equalized Odds (Hardt et al., 2016), and Demographic Parity (Calders et al., 2009) also respect this definition. It means that using this generic formulation allows us to simultaneously reason about multiple fairness notions. Example 1 (Accuracy Parity (AP) (Zafar et al., 2017a)). A model hθ is fair for Accuracy Parity when the probability of being correct is independent of the sensitive attribute, that is, r S b P (hθ(x) = y | s = r) = b P (hθ(x) = y) . It means that we need to partition the space into K = |S| groups and, r S, we define b F(r) as the fairness level of group (r) b F(r)(T , hθ) = b P (hθ(x) = y) b P (hθ(x) = y | s = r) = (b P (s = r) 1)b P (hθ(x) = y | s = r) + X b P (s = r ) b P (hθ(x) = y | s = r ) where the law of total probability was used to obtain the last equality. Thus, Accuracy Parity satisfies all our assumptions with C(r) (r) = b P (s = r) 1, C(r ) (r) = b P (s = r ) with r = r, and C0 (r) = 0. It is worth noting that Fair Grad applies to any fairness measure that respects the definition above, even when there is a large number of groups. However, the performance of Fair Grad may degrade when there are only a few samples per group, as fairness estimations become unreliable. In this case, the risk is that the learned model is fair on the training set but does not generalize well to new examples. To circumvent some of these issues works such as Hebert-Johnson et al. (2018); Kearns et al. (2018) have extended fairness definitions to multi-group settings and proposed mechanisms to optimize them. In this work, we focus on classical fairness definitions and keep the line of research to extend Fairgrad to these alternative definitions for the future. 2.2 Related Work Various in-processing methods have been proposed in the fair machine learning literature. Amongst them, many methods rely on formulating the problem as either a constrained optimization, which is later relaxed to an unconstrained case or using re-weighting techniques where examples are dynamically re-weighed based on the fairness levels of the model (Caton & Haas, 2020). In this sub-section, we will provide a brief overview of these methods and explain the similarities and differences between Fair Grad and the corresponding approaches. Additionally, we will also demonstrate how Fair Grad can be seen as a solution that connects these two streams of work. For more details about very closely related works, please refer to Appendix A. Constrained Optimization The problem of fair machine learning can be seen as the following constrained optimization problem (Cotter et al., 2019; Agarwal et al., 2018): arg min hθ H b P (hθ(x) = y) s.t. k [K], b Fk(T , hθ) = 0. (2) This problem can then be reformulated as an unconstrained optimization problem using Lagrange multipliers. More specifically, with multipliers denoted by λ1, . . . , λK, the unconstrained objective that should be minimized for hθ H and maximized for λ1, . . . , λK R is: L (hθ, λ1, . . . , λK) = b P (hθ(x) = y) + k=1 λk b Fk(T , hθ) . (3) Several strategies may then be employed to find a saddle point for this objective1. Agarwal et al. (2018) first relax the problem by searching for a distribution over the models rather than a single optimal hypothesis. 1These min-max formulations are not new in the literature and was already used in the 1940 s (Wald, 1945). More recently, Madry et al. (2018) employed the formulation to make deep neural networks more robust against adversarial attacks. Similarly, Ben-Tal et al. (2012) modeled uncertainty in input via this formulation. Published in Transactions on Machine Learning Research (08/2023) Then, they alternate between using an exponentiated gradient step to find λ1, . . . , λK R and a procedure based on cost sensitive learning to find the next hθ to add to their distribution. Similarly, Cotter et al. (2019) also search for a distribution over the models using an alternating approach based on Lagrange multipliers where they relax objective (3) by replacing the error rate with a loss term. To update the λ multipliers, unlike Agarwal et al. (2018), they use projected gradient descent based on the original fairness terms. To search the next hθ to add to their distribution of models they use a projected gradient descent update over a relaxed overall objective function where the fairness measures are replaced with smooth upper bounds. In this work, we also use an alternating approach based on objective (3). However, we look for a single model rather than a distribution of models. To this end, at each iteration, we update λ using a projected gradient descent step similar to Cotter et al. (2019), that is using the original fairness measures. To solve for hθ, contrary to Cotter et al. (2019), we first show that Objective (3), with fixed λ, may be rewritten as a weighted sum of group-wise error rates. This is similar in spirit to the cost-sensitive learning method of Agarwal et al. (2018) but can be applied beyond simple binary classification. We then follow Cotter et al. (2019) and replace in our new objective the error rate terms with a loss function, albeit not necessarily an upper bound, to obtain meaningful gradient directions. Re-weighting Another way to learn fair models is to use a re-weighting approach where each example x is associated with a weight wx R so that minimizing the following objective for hθ outputs a fair model: W (hθ) = b E wx I{hθ(x) =y} . The underlying idea for the methods which posit the problem as above is to propose a cost function that outputs weights for each example. On the one hand, the weights can be determined in a pre-processing step (Kamiran & Calders, 2012), based on the statistics of the data under consideration. On the other hand, the weights may evolve with hθ, that is they are dynamically updated each time the model changes during the training process (Roh et al., 2020). In this work, to find hθ, we also use a dynamic re-weighting approach where the weights change at each iteration. To choose the weights, we initially give the same importance to each example. Then, we increase the weights of disadvantaged examples and decrease the weights of advantaged examples proportionally to the fairness level of the current model for their group. An important feature of our approach, unlike other re-weighting approaches, is that we do not constrain ourselves to positive weights but rather allow the use of negative weights. Indeed, we show in Lemma 1 that the latter are sometimes necessary to learn fair models. To summarize, we first frame the task as a constrained optimization problem, similar to Cotter et al. (2019) and Agarwal et al. (2018). We then propose an alternating approach, where we update λ at each iteration using a projected gradient descent step similar to Cotter et al. (2019). However, in order to learn the model hθ, we show that Objective (3), with fixed λ, can be rewritten as a weighted sum of group-wise error rates. This step can be interpreted as an instance of dynamic re-weighting where the weights change at each iteration. Thus our method can be seen as a connection between constrained optimization and re-weighting. 3 Fair Grad In the previous section, we argued that Fair Grad is connected to both constrained optimization and reweighting approaches. In this section, we provide details on our method and we present it starting from the constrained optimization point of view as we believe it makes it easier to understand how the weights are selected and updated. We begin by discussing Fair Grad for exact fairness and then extend it to ϵ-fairness. 3.1 Fair Grad for Exact Fairness To solve the fairness problem described in equation 3, we propose to use an alternating approach where the hypothesis and the multipliers are updated one after the other2. We begin by describing our method to update the multipliers and then the model. 2It is worth noting that, here, we do not have formal duality guarantees and that the problem is not even guaranteed to have a fair solution. Nevertheless, the approach seems to work well in practice as can be seen in the experiments. Published in Transactions on Machine Learning Research (08/2023) Updating the Multipliers. To update λ1, . . . , λK, we will use a standard gradient ascent procedure. Hence, given that the gradient of Problem (3) is λ1,...,λKL (hθ, λ1, . . . , λK) = b F1(T , hθ) ... b FK(T , hθ) we have the following update rule k [K]: λT +1 k = λT k + ηλ b Fk T , h T θ where ηλ is a rate that controls the importance of each update. In the experiments, we use a constant rate of 0.01 as our initial tests showed that it is a good rule of thumb when the data is properly standardized. Updating the Model. To update the parameters θ RD of the model hθ, we use a standard gradient descent. However, first, we notice that, given our fairness definition, Equation (3) can be written as L (hθ, λ1, . . . , λK) = k=1 b P (hθ(x) = y|Tk) " b P (Tk) + k =1 Ck k λk k=1 λk C0 k. (4) where PK k=1 λk C0 k is independent of hθ by definition. Hence, at iteration t, the update rule becomes θT +1 = θT ηθ " b P (Tk) + k =1 Ck k λk θb P (hθ(x) = y|Tk) where ηθ is the usual learning rate that controls the importance of each parameter update. Here, we obtain our group specific weights k, wk = h b P (Tk) + PK k =1 Ck k λk i , that depend on the current fairness level of the model through λ1, . . . , λK, the relative size of each group through b P (Tk), and the fairness notion under consideration through the constants C. The exact values of these constants are given in Section 2.1 and Appendix B for various group fairness notions. Overall, they are such that, at each iteration, the weights of the advantaged groups are reduced and the weights of the disadvantaged groups are increased. The main limitation of the above update rule is that one needs to compute the gradient of 0 1-losses since θb P (hθ(x) = y|Tk) = 1 nk P (x,y) Tk θI{hθ(x) =y}. Unfortunately, this usually does not provide meaningful optimization directions. To address this issue, we follow the usual trend in machine learning and replace the 0 1-loss with one of its continuous and differentiable surrogates that provides meaningful gradients. For instance, in our experiments, we use the cross entropy loss. 3.2 Computational Overhead of Fair Grad. We summarize our approach in Algorithm 1, where we have used italic font to highlight the steps inherent to Fair Grad that do not appear in classic gradient descent. We consider batch gradient descent rather than full gradient descent as it is a popular scheme. We empirically investigate the impact of the batch size in Section 4.7. The main difference is Step 5 (in italic font), that is the computation of the group-wise fairness levels. However, these can be cheaply obtained from the predictions of h(t) θ on the current batch which are always available since they are also needed to compute the gradient. Hence, the computational overhead of Fair Grad is very limited. 3.3 Importance of Negative Weights. A key property of Fair Grad is that we allow the use of negative weights, that is h b P (Tk) + PK k =1 Ck k λk i may become negative, while existing methods (Roh et al., 2020; Iosifidis & Ntoutsi, 2019; Jiang & Nachum, 2020) restrict themselves to positive weights. In this section, we show that these negative weights are important as they are sometimes necessary to learn fair models. Hence, in the next lemma, we provide sufficient conditions so that negative weights are mandatory if one wants to enforce Accuracy Parity. Published in Transactions on Machine Learning Research (08/2023) Algorithm 1 Fair Grad for Exact Fairness Input: Groups T1, . . . , TK, Functions b F1, . . . , b FK, Function class H of models hθ with parameters θ RD, Learning rates ηλ, ηθ, and Iterator iter that returns batches of examples. Output: A fair model h θ. 1: Initialize the group specific weights and the model. 2: for B in iter do 3: Compute the predictions of the current model on the batch B. 4: Compute the group-wise losses using the predictions. 5: Compute the current fairness level using the predictions and update the group-wise weights. 6: Compute the overall weighted loss using the group-wise weights. 7: Compute the gradients based on the loss and update the model. 8: end for 9: return the trained model h θ Lemma 1 (Negative weights are necessary.). Let the fairness notion be Accuracy Parity (Example 1). Let h θ be the most accurate and fair model. Then using negative weights is necessary as long as min hθ H hθunfair max Tk b P (hθ(x) = y|Tk) < b P (h θ(x) = y) . Proof. The proof is provided in Appendix C. The previous condition can sometimes be verified in practice. As a motivating example, assume a binary setting with only two sensitive groups T1 and T 1. Let h 1 θ be the model minimizing b P (hθ(x) = y|T 1) and assume that b P h 1 θ (x) = y < b P h 1 θ (x) = y|T 1 , that is group T 1 is disadvantaged for accuracy parity. Given h θ the most accurate and fair model, we have min hθ H hθunfair max Tk b P (hθ(x) = y|Tk) = b P h 1 θ (x) = y|T 1 < b P (h θ(x) = y) as otherwise we would have a contradiction since the fair model would also be the most accurate model for group T 1 since b P (h θ(x) = y) = b P (h θ(x) = y|T 1) by definition of Accuracy Parity. In other words, a dataset where the most accurate model for a given group still disadvantages it requires negative weights. This might be connected to the notion of leveling down (Zietlow et al., 2022; Mittelstadt et al., 2023), where fairness can only be achieved by harming all the groups or bringing advantaged groups closer to disadvantaged groups by harming them. It is generally an artifact of strictly egalitarian fairness measures. Investigating this negative effect is an important research direction that goes beyond the scope of this paper. Nevertheless, a potential solution to mitigate it is to use other kind of fairness definitions. As a first step in this direction, in the next section we extend Fair Grad to ϵ-fairness where strict equality is relaxed. 3.4 Fair Grad for ϵ-fairness In the previous section, we considered exact fairness and we showed that this could be achieved by using a re-weighting approach. Here, we extend this procedure to ϵ-fairness where the fairness constraints are relaxed and a controlled amount of violations is allowed. Usually, ϵ is a user defined parameter but it can also be set by the law, as it is the case with the 80% rule in the US (Biddle, 2006). The main difference with exact fairness is that each equality constraint in Problem (2) is replaced with two inequalities of the form k [K], b Fk(T , hθ) ϵ k [K], b Fk(T , hθ) ϵ. The main consequence is that we need to maintain twice as many Lagrange multipliers and that the groupwise weights are slightly different. Since the two procedures are similar, we omit the details here but provide them in Appendix D for the sake of completeness. Published in Transactions on Machine Learning Research (08/2023) 4 Experiments In this section, we present several experiments that demonstrate the competitiveness of Fair Grad as a procedure to learn fair models for classification. We begin by presenting results over standard fairness datasets and a Natural language Processing dataset in Section 4.4. We then study the behaviour of the ϵ-fairness variant of Fair Grad in Section 4.5. Next, we showcase the fine-tuning ability of Fair Grad on a Computer Vision dataset in Section 4.6. Finally, we investigate the impact of batch size on the learned model in Section 4.7 and present results related to the computational overhead incurred by Fair Grad in Section 4.8. 4.1 Datasets In the main paper, we consider 4 different datasets and postpone the results on another 6 datasets to Appendix E.3 as they follow similar trends. We also postpone the detailed descriptions of these datasets as well as the pre-processing steps to Appendix E.2. We consider commonly used fairness datasets, namely Adult Income (Kohavi, 1996) and Celeb A (Liu et al., 2015). Both are binary classification datasets with binary sensitive attributes (gender). We also consider a variant of the Adult Income dataset where we add a second binary sensitive attribute (race) to obtain a dataset with 4 disjoint sensitive groups. For both datasets, we use 20% of the data as a test set and the remaining 80% as a train set. We further divide the train set into two and keep 25% of the training examples as a validation set. For each repetition, we randomly shuffle the data before splitting it, and thus we have unique splits for each random seed. Lastly, we standardize each features independently by subtracting the mean and scaling to unit variance which were estimated on the training set. To showcase the wide applicability of Fair Grad, we consider the Twitter Sentiment3 (Blodgett et al., 2016) dataset from the Natural Language Processing community. It consists of 200k tweets with binary sensitive attribute (race) and binary sentiment score. We employ the same setup, splits, and the pre-processing as proposed by Han et al. (2021) and Elazar & Goldberg (2018) and create bias in the dataset by changing the proportion of each subgroup (race-sentiment) in the training set. Following the footsteps of Elazar & Goldberg (2018) we encode the tweets using the Deep Moji (Felbo et al., 2017) encoder with no fine-tuning, which has been pre-trained over millions of tweets to predict their emoji, thereby predicting the sentiment. We also employ the UTKFace dataset4 (Zhang et al., 2017) from the Computer Vision community. It consists of 23, 708 images tagged with race, age, and gender with pre-defined splits. 4.2 Performance Measures For fairness, we consider the four measures introduced in Section 2.1 and Appendix B, namely Equalized Odds (EOdds), Equality of Opportunity (EOpp), Accuracy Parity (AP), and Demographic Parity (DP). For each specific fairness notion, we report the average absolute fairness level of the different groups over the test set, that is 1 K PK k=1 b Fk(T , hθ) (lower is better). To assess the utility of the learned models, we use their accuracy levels over the test set, that is 1 n Pn i=1 Ihθ(xi)=yi (higher is better). All the results reported are averaged over 5 independent runs and standard deviations are provided. Note that, in the main paper, we graphically report a subset of the results over the aforementioned datasets. We provide detailed results in Appendix E.3, including the missing pictures as well as complete tables with accuracy levels, fairness levels, and fairness level of the most well-offand worst-offgroups for all the relevant methods. 4.3 Methods We compare Fair Grad to a wide variety of baselines, namely: Unconstrained, which is oblivious to any fairness measure and is trained using a standard batch gradient descent method. 3http://slanglab.cs.umass.edu/Twitter AAE/ 4https://susanqq.github.io/UTKFace/ Published in Transactions on Machine Learning Research (08/2023) Adversarial learning based method where we employ adversarial mechanism (Goodfellow et al., 2014) using a gradient reversal layer (Ganin & Lempitsky, 2015), similar to GRAD-Pred (Raff& Sylvester, 2018), where an adversary, with an objective to predict the sensitive attribute, is added to the unconstrained model Bi-level optimization based method implemented in the form of Bi Fair (Ozdayi et al., 2021) Re-weighting based methods in the form of Fair Batch (Roh et al., 2020). We also compare against a simpler baseline called Weighted ERM where each example is reweighed based on the size of the sensitive group the example belongs to in the beginning. Unlike Fair Batch these weights are not updated during training. Constrained optimization based method as proposed by Cotter et al. (2019). We refer to this method as Constraints in this article. Reduction implements the exponentiated gradient based fair classification approach as proposed by Agarwal et al. (2018). In all our experiments, we consider two different hypothesis classes. On the one hand, we use linear models implemented in the form of neural networks with no hidden layers. On the other hand, we use a more complex, non-linear architecture with three fully-connected hidden layers of respective sizes 128, 64, and 32. We use Re LU as our activation function with batch normalization and dropout. In both cases, we optimize the cross-entropy loss. In several experiments, we only consider subsets of the baselines due to the limitations of the methods. For instance, Bi Fair was designed to handle binary labels and binary sensitive attributes and thus is not considered for the datasets with more than two sensitive groups or two labels. Furthermore, we implemented it using the authors code that is freely available online but does not include AP as a fairness measure, thus we do not report results related to this measure for Bi Fair. Similarly, we also implemented Fair Batch from the authors code which does not support AP as a fairness measure, thus we also exclude it from the comparison for this measure. For Constraints, we based our implementation on the publicly available authors library but were only able to reliably handle linear models and thus we do not consider this baseline for non-linear models. Finally, for Adversarial, we used our custom made implementation. However, it is only applicable when learning non-linear models since it requires at least one hidden layer to propagate its reversed gradient. Apart from the common hyper-parameters such as dropout, several baselines come with their own set of hyper-parameters. For instance, Bi Fair has the inner loop length, which controls the number of iterations in its inner loop, while Adversarial has the scaling, which re-weights the adversarial branch loss and the task loss. We provide details of common and approach specific hyper-parameters with their range in Appendix E.1. With several hyper-parameters for each approach, selecting the best combination is often crucial to avoid undesirable behaviors such as over-fitting (Maheshwari et al., 2022). In this paper, we opt for the following procedure. First, for each method, we consider all the X possible hyper-parameter combinations and we run the training procedure for 50 epochs for each combination. Then, we retain all the models returned by the last 5 epochs, that is, for a given method, we have 5X models and the goal is to select the best one among them. Since we have access to two performance measures, we can select either the most accurate model, the most fair, or a trade-offbetween the two depending on the end goal. Here, we chose to focus on the third option and select the model with the lowest fairness score between certain accuracy intervals. More specifically, let α be the highest validation accuracy among the 5X models. We choose the model with the lowest validation fairness score amongst all models with a validation accuracy in the interval [α k, α ]. In this work, we fix k to 0.03. 4.4 Results for Exact Fairness We report the results over the Adult Income dataset using a linear model, the Adult Income dataset with multiple groups with a non-linear model, and the Twitter sentiment dataset using both linear and nonlinear models in Figures 2, 3, and 4 respectively. In these figures, the best methods are closer to the bottom right Published in Transactions on Machine Learning Research (08/2023) corner. If a method is closer to the bottom left corner, it has good fairness but reduced accuracy. Similarly, a method closer to the top right corner has good accuracy but poor fairness. The main take-away from these experiments is that there is no fairness enforcing method that is consistently better than the others in terms of both accuracy and fairness. All of them have strengths, that is datasets and fairness measures where they obtain good results, and weaknesses, that is datasets and fairness measures for which they are sub-optimal. Fair Batch induces better accuracy than the other approaches over Adult with linear model and EOdds and only pays a small price in fairness. However, it is significantly worse in terms of fairness over the Adult Multigroup dataset with a non-linear model. Similarly, Bi Fair is sub-optimal on Adult with EOpp, while being comparable to the other approaches on the Twitter Sentiment dataset. We observed similar trends on the other datasets, available in Appendix E.3, with different methods coming out on top for different datasets and fairness measures. Interestingly, Fair Grad generally outperforms other approaches in terms of fairness, albeit with a slight loss in accuracy. These observations are even more amplified in the Accuracy Parity and Equalized Odds settings. Moreover, it is generally more robust and tends to show a lower standard deviation in accuracy and fairness than the other approaches. Even in terms of accuracy, the largest difference is over the Crime dataset, where the difference between Fair Grad and Unconstrained is 0.04. However, in most cases, the difference is within 0.02. In terms of the multi-group setup, we find similar observations, that is Fair Grad outperforms other approaches in fairness, albeit with a drop in accuracy. In fact, for Equality of Opportunity Fair Grad almost outperforms all approaches in terms of fairness and accuracy. Overall, Fair Grad performs reasonably well in all the settings we considered with no obvious weaknesses, that is no datasets with the lowest accuracy and fairness compared to the baselines. 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp Figure 2: Results for the Adult dataset using Linear Models. 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (a) Non Linear - AP 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction (b) Non Linear - EOdds 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction (c) Non Linear - EOpp Figure 3: Results for the Adult Multigroup dataset using Non Linear models. 4.5 Accuracy Fairness Trade-off In this second set of experiments, we demonstrate the capability of Fair Grad to support approximate fairness (see Section 3.4). In Figure 5, we show the performance, as accuracy-fairness pairs, of several models learned Published in Transactions on Machine Learning Research (08/2023) 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 4: Results for the Twitter Sentiment dataset for Linear and Non Linear Models. 0.835 0.840 0.845 0.850 Accuracy Accuracy Parity eps: 0.0 eps: 0.01 eps: 0.02 eps: 0.03 eps: 0.04 eps: 0.05 eps: 0.06 eps: 0.07 eps: 0.08 eps: 0.09 eps: 0.1 eps: 0.2 eps: 0.3 eps: 0.5 eps: 1.0 (a) Celeb - AP 0.835 0.840 0.845 0.850 0.855 Accuracy eps: 0.0 eps: 0.01 eps: 0.02 eps: 0.03 eps: 0.04 eps: 0.05 eps: 0.06 eps: 0.07 eps: 0.08 eps: 0.09 eps: 0.1 eps: 0.2 eps: 0.3 eps: 0.5 eps: 1.0 (b) Celeb - EOdds 0.83 0.84 0.85 Accuracy Equal opportunity eps: 0.0 eps: 0.01 eps: 0.02 eps: 0.03 eps: 0.04 eps: 0.05 eps: 0.06 eps: 0.07 eps: 0.08 eps: 0.09 eps: 0.1 eps: 0.2 eps: 0.3 eps: 0.5 eps: 1.0 (c) Celeb - EOpp Figure 5: Results for Celeb A using Linear models. The Unconstrained Linear model achieves a test accuracy of 0.8532 with fairness level of 0.0499 for EOdds, 0.0204 for AP, and 0.0387 for EOpp. on the Celeb A dataset by varying the fairness level parameter ϵ. These results suggest that Fair Grad respects the constraints well. Indeed, the average absolute fairness level (across all the groups, see Section 4.2) achieved by Fair Grad is either the same or less than the given threshold. It is worth mentioning that Fair Grad is designed to enforce ϵ-fairness for each constraint individually which is slightly different from the summarized quantity displayed here. Finally, as the fairness constraint is relaxed, the accuracy of the model increases, reaching the same performance as Unconstrained when the fairness level of the latter is below ϵ. 4.6 Fair Grad as a Fine-Tuning Procedure While Fair Grad has primarily been designed to learn fair classifiers from scratch, it can also be used to finetune an existing classifier to achieve better fairness. To showcase this, we fine-tune the Res Net18 (He et al., 2016) model, developed for image recognition, over the UTKFace dataset (Zhang et al., 2017), consisting of human face images tagged with Gender, Age, and Race information. Following the same process as Roh et al. (2020), we use Race as the sensitive attribute and consider two scenarios. Either we consider Demographic Parity as the fairness measure and use the gender (binary) as the target label or we consider Equalized Odds Published in Transactions on Machine Learning Research (08/2023) Table 1: Results for the UTKFace dataset where a Res Net18 is fine-tuned using different strategies. Method s=Race ; y=Gender s=Race ; y=Age Accuracy DP Accuracy EOdds Unconstrained 0.8691 0.0075 0.0448 0.0066 0.6874 0.0080 0.0843 0.0089 Fair Grad 0.8397 0.0085 0.0111 0.0064 0.6491 0.0082 0.0506 0.0059 Table 2: Batch size effect on the Celeb A dataset with Linear Models and EOdds as the fairness measure. Batch Size 8 16 32 64 128 256 512 1024 2048 Accuracy 0.8186 0.8234 0.8215 0.8268 0.8273 0.8286 0.8292 0.8289 0.8303 Accuracy Std 0.0013 0.006 0.0028 0.0025 0.0031 0.0008 0.0027 0.0017 0.0031 Fairness 0.0031 0.0091 0.0045 0.0036 0.0051 0.0046 0.004 0.0038 0.0057 Fairness Std 0.0042 0.0062 0.0012 0.0014 0.0025 0.0032 0.0026 0.0019 0.0018 and predict the age (multi-valued). The results are displayed in Table 1. In both settings, Fair Grad learns models that are more fair than an Unconstrained fine-tuning procedure, albeit at the expense of accuracy. 4.7 Impact of the Batch-size In this section, we evaluate the impact of batch size on the fairness and accuracy level of the learned model. Indeed, at each iteration, in order to minimize the overhead associated with Fair Grad (see Section 3.1), we update the weights using the fairness level of the model estimated solely on the current batch. When these batches are small, these estimates are unreliable and might lead the model astray. In Table 2 we present the performances of several linear models learned with different batch sizes on the Celeb A dataset. Over this dataset, we observe that Fair Grad consistently learns a fair model across all batch sizes and obtains reasonable accuracy since Unconstrained has an accuracy of 0.8532 for this problem. Nevertheless, we still recommend the practitioners to use a larger batch size whenever possible as we observe a slight reduction in terms of fairness standard deviations. 4.8 Computational Overhead In this last experiment, we evaluate the overhead of Fair Grad, by reporting the wall clock time in seconds to train for an epoch with the Unconstrained approach and our method in various settings. We show the effect of model size by varying the number of hidden layers of the model over the Adult Income dataset, which consists of 45, 222 records. We used an Intel Xeon E5-2680 CPU to train. We consider a large convolutional neural network (Res Net18 (He et al., 2016)) fine tuned over the UTK-Face dataset consisting of 23, 708 images. We trained the model using a Tesla P100 GPU. We experiment with a large transformer (bert-base-uncased (Devlin et al., 2019)) fine tuned over the Twitter Sentiment Dataset consisting of 200k tweets. We trained it using a Tesla P100 GPU. We present results of the computation overhead of Fair Grad in Table 3. We find that the overhead is limited and should not be critical in most applications as it does not depend on the complexity of the model but, instead, on the number of examples and the batch size. Overall, these observations are in line with the arguments presented in Section 3.2. 5 Conclusion In this paper, we proposed Fair Grad, a fairness aware gradient descent approach based on a re-weighting scheme. We showed that it can be used to learn fair models for various group fairness definitions and is able Published in Transactions on Machine Learning Research (08/2023) Table 3: The computational overhead of Fair Grad in various settings. BS here refers to Batch Size, and the Unconstrained and Fair Grad columns refers to the average time in seconds taken by these approaches for an epoch, respectively. Delta refers to the difference in time between these two approaches. Setting Parameters BS Unconstrained Fair Grad Delta Linear model - Adult Dataset -CPU 106 512 0.277 0.031 0.307 0.01 0.03 2 layers -Adult Dataset -CPU 1762 512 0.315 0.036 0.316 0.029 0.01 5 layers -Adult Dataset -CPU 21346 512 0.370 0.042 0.394 0.025 0.02 10 layers -Adult Dataset -CPU 39042 512 0.483 0.021 0.499 0.034 0.02 20 layers -Adult Dataset -CPU 80642 512 0.672 0.034 0.689 0.026 0.02 Res Net18 trained -UTKFace -GPU 11177538 64 31.173 0.085 31.588 0.055 0.42 Bert Twitter Sentiment -GPU 109505310 32 2246.342 3.20 2294.382 4.01 48.04 to handle multiclass problems as well as settings where there is multiple sensitive groups. We empirically showed the competitiveness of our approach against several baselines on standard fairness datasets and on a Natural Language Processing task. We also showed that it can be used to fine-tune an existing model on a Computer Vision task. Finally, since it is based on gradient descent and has a small overhead, we believe that Fair Grad could be used for a wide range of applications, even beyond classification. Limitations and Societal Impact While appealing, Fair Grad also has limitations. It implicitly assumes that a set of weights that would lead to a fair model exists but this might be difficult to verify in practice. Thus, even if in our experiments Fair Grad seems to behave quite well, a practitioner using this approach should not trust it blindly. It remains important to always check the actual fairness level of the learned model. On the other hand, we believe that, due to its simplicity and its versatility, Fair Grad could be easily deployed in various practical contexts and, thus, could contribute to the dissemination of fair models. Acknowledgements This work has been supported by the Région Hauts de France (Projet STa RS Equité en apprentissage décentralisé respectueux de la vie privée) and Agence Nationale de la Recherche under grant number ANR19-CE23-0022. The authors would also like to thank Michael Lohaus and anonymous reviewers for helpful discussions and feedbacks. Alekh Agarwal, Alina Beygelzimer, Miroslav Dudík, John Langford, and Hanna Wallach. A reductions approach to fair classification. In International Conference on Machine Learning, pp. 60 69. PMLR, 2018. Aharon Ben-Tal, Sahely Bhadra, Chiranjib Bhattacharyya, and Arkadi Nemirovski. Efficient methods for robust classification under uncertainty in kernel matrices. J. Mach. Learn. Res., 13:2923 2954, 2012. doi: 10.5555/2503308.2503335. URL https://dl.acm.org/doi/10.5555/2503308.2503335. Dan Biddle. Adverse impact and test validation: A practitioner s guide to valid and defensible employment testing. Gower Publishing, Ltd., 2006. Su Lin Blodgett, Lisa Green, and Brendan O Connor. Demographic dialectal variation in social media: A case study of African-American English. 2016. Toon Calders and Sicco Verwer. Three naive bayes approaches for discrimination-free classification. Data mining and knowledge discovery, 21(2):277 292, 2010. Toon Calders, Faisal Kamiran, and Mykola Pechenizkiy. Building classifiers with independency constraints. In 2009 IEEE International Conference on Data Mining Workshops, pp. 13 18. IEEE, 2009. Published in Transactions on Machine Learning Research (08/2023) Flavio P Calmon, Dennis Wei, Bhanukiran Vinzamuri, Karthikeyan Natesan Ramamurthy, and Kush R Varshney. Optimized pre-processing for discrimination prevention. volume 30, 2017. Simon Caton and Christian Haas. Fairness in machine learning: A survey. ar Xiv preprint ar Xiv:2010.04053, 2020. Evgenii Chzhen, Christophe Denis, Mohamed Hebiri, Luca Oneto, and Massimiliano Pontil. Leveraging labeled and unlabeled data for consistent fair binary classification. ar Xiv preprint ar Xiv:1906.05082, 2019. Andrew Cotter, Heinrich Jiang, and Karthik Sridharan. Two-player games for efficient non-convex constrained optimization. In Algorithmic Learning Theory, pp. 300 332. PMLR, 2019. Christophe Denis, Romuald Elie, Mohamed Hebiri, and François Hu. Fairness guarantee in multi-class classification. ar Xiv preprint ar Xiv:2109.13642, 2021. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: pre-training of deep bidirectional transformers for language understanding. In Jill Burstein, Christy Doran, and Thamar Solorio (eds.), Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers), pp. 4171 4186. Association for Computational Linguistics, 2019. doi: 10.18653/v1/n19-1423. URL https://doi.org/10.18653/v1/n19-1423. Frances Ding, Moritz Hardt, John Miller, and Ludwig Schmidt. Retiring adult: New datasets for fair machine learning. In Marc Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan (eds.), Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, December 6-14, 2021, virtual, pp. 6478 6490, 2021. URL https://proceedings.neurips.cc/paper/2021/hash/ 32e54441e6382a7fbacbbbaf3c450059-Abstract.html. Michele Donini, Luca Oneto, Shai Ben-David, John Shawe-Taylor, and Massimiliano Pontil. Empirical risk minimization under fairness constraints. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp. 2796 2806, 2018. Dheeru Dua, Casey Graff, et al. Uci machine learning repository. 2017. Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference, pp. 214 226, 2012. Yanai Elazar and Yoav Goldberg. Adversarial removal of demographic attributes from text data. In Ellen Riloff, David Chiang, Julia Hockenmaier, and Jun ichi Tsujii (eds.), Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31 - November 4, 2018, pp. 11 21. Association for Computational Linguistics, 2018. Bjarke Felbo, Alan Mislove, Anders Søgaard, Iyad Rahwan, and Sune Lehmann. Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm. In Martha Palmer, Rebecca Hwa, and Sebastian Riedel (eds.), Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, EMNLP 2017, Copenhagen, Denmark, September 9-11, 2017, pp. 1615 1625. Association for Computational Linguistics, 2017. Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. Certifying and removing disparate impact. In proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, pp. 259 268, 2015. Yaroslav Ganin and Victor Lempitsky. Unsupervised domain adaptation by backpropagation. In Francis Bach and David Blei (eds.), Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pp. 1180 1189, Lille, France, 07 09 Jul 2015. PMLR. Published in Transactions on Machine Learning Research (08/2023) Gabriel Goh, Andrew Cotter, Maya Gupta, and Michael P Friedlander. Satisfying real-world goals with dataset constraints. In Advances in Neural Information Processing Systems, pp. 2415 2423, 2016. Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. Advances in neural information processing systems, 27, 2014. Xudong Han, Timothy Baldwin, and Trevor Cohn. Diverse adversaries for mitigating bias in training. In Paola Merlo, Jörg Tiedemann, and Reut Tsarfaty (eds.), Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, EACL 2021, Online, April 19 - 23, 2021, pp. 2760 2765. Association for Computational Linguistics, 2021. Moritz Hardt, Eric Price, and Nati Srebro. Equality of opportunity in supervised learning. Advances in neural information processing systems, 29:3315 3323, 2016. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Ursula Hebert-Johnson, Michael Kim, Omer Reingold, and Guy Rothblum. Multicalibration: Calibration for the (Computationally-identifiable) masses. In Jennifer Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pp. 1939 1948. PMLR, 10 15 Jul 2018. URL https://proceedings.mlr.press/v80/ hebert-johnson18a.html. Vasileios Iosifidis and Eirini Ntoutsi. Adafair: Cumulative fairness adaptive boosting. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, pp. 781 790, 2019. Vasileios Iosifidis, Besnik Fetahu, and Eirini Ntoutsi. Fae: A fairness-aware ensemble framework. In 2019 IEEE International Conference on Big Data (Big Data), pp. 1375 1380. IEEE, 2019. Heinrich Jiang and Ofir Nachum. Identifying and correcting label bias in machine learning. In International Conference on Artificial Intelligence and Statistics, pp. 702 712. PMLR, 2020. Faisal Kamiran and Toon Calders. Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems, 33(1):1 33, 2012. Faisal Kamiran, Toon Calders, and Mykola Pechenizkiy. Discrimination aware decision tree learning. In 2010 IEEE International Conference on Data Mining, pp. 869 874. IEEE, 2010. Toshihiro Kamishima, Shotaro Akaho, Hideki Asoh, and Jun Sakuma. Fairness-aware classifier with prejudice remover regularizer. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 35 50. Springer, 2012. Michael Kearns, Seth Neel, Aaron Roth, and Zhiwei Steven Wu. Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. In Jennifer Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pp. 2564 2572. PMLR, 10 15 Jul 2018. URL https://proceedings.mlr.press/v80/kearns18a.html. Ron Kohavi. Scaling up the accuracy of naive-bayes classifiers: A decision-tree hybrid. In Evangelos Simoudis, Jiawei Han, and Usama M. Fayyad (eds.), Kdd, pp. 202 207. AAAI Press, 1996. Emmanouil Krasanakis, Eleftherios Spyromitros-Xioufis, Symeon Papadopoulos, and Yiannis Kompatsiaris. Adaptive sensitive reweighting to mitigate bias in fairness-aware classification. In Proceedings of the 2018 World Wide Web Conference, pp. 853 862, 2018. Matt J Kusner, Joshua Loftus, Chris Russell, and Ricardo Silva. Counterfactual fairness. Advances in neural information processing systems, 30, 2017. JeffLarson, Surya Mattu, Lauren Kirchner, and Julia Angwin. How we analyzed the compas recidivism algorithm. Pro Publica (5 2016), 9(1):3 3, 2016. Published in Transactions on Machine Learning Research (08/2023) Yitong Li, Timothy Baldwin, and Trevor Cohn. Towards robust and privacy-preserving text representations. In Iryna Gurevych and Yusuke Miyao (eds.), Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Melbourne, Australia, July 15-20, 2018, Volume 2: Short Papers, pp. 25 30. Association for Computational Linguistics, 2018. doi: 10.18653/v1/P18-2005. URL https://aclanthology.org/P18-2005/. Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision, pp. 3730 3738, 2015. Michael Lohaus, Michaël Perrot, and Ulrike Von Luxburg. Too relaxed to be fair. In International Conference on Machine Learning, pp. 6360 6369. PMLR, 2020. Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. Open Review.net, 2018. URL https://openreview.net/forum?id=r Jz IBf ZAb. Gaurav Maheshwari, Pascal Denis, Mikaela Keller, and Aurélien Bellet. Fair NLP models with differentially private text encoders. In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang (eds.), Findings of the Association for Computational Linguistics: EMNLP 2022, Abu Dhabi, United Arab Emirates, December 7-11, 2022, pp. 6913 6930. Association for Computational Linguistics, 2022. URL https://aclanthology. org/2022.findings-emnlp.514. Paul Mangold, Michaël Perrot, Aurélien Bellet, and Marc Tommasi. Differential privacy has bounded impact on fairness in classification. ar Xiv preprint ar Xiv:2210.16242, 2022. Ninareh Mehrabi, Fred Morstatter, Nripsuta Saxena, Kristina Lerman, and Aram Galstyan. A survey on bias and fairness in machine learning. ACM Computing Surveys (CSUR), 54(6):1 35, 2021. Brent D. Mittelstadt, Sandra Wachter, and Chris Russell. The unfairness of fair machine learning: Levelling down and strict egalitarianism by default. Co RR, abs/2302.02404, 2023. doi: 10.48550/ar Xiv.2302.02404. URL https://doi.org/10.48550/ar Xiv.2302.02404. Mustafa Safa Ozdayi, Murat Kantarcioglu, and Rishabh Iyer. Bifair: Training fair models with bilevel optimization. ar Xiv preprint ar Xiv:2106.04757, 2021. Edward Raffand Jared Sylvester. Gradient reversal against discrimination: A fair neural network learning approach. In 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA), pp. 189 198. IEEE, 2018. Michael Redmond and Alok Baveja. A data-driven software tool for enabling cooperative information sharing among police departments. European Journal of Operational Research, 141(3):660 678, 2002. Yuji Roh, Kangwook Lee, Steven Euijong Whang, and Changho Suh. Fairbatch: Batch selection for model fairness. In International Conference on Learning Representations, 2020. Shai Shalev-Shwartz and Shai Ben-David. Understanding machine learning: From theory to algorithms. Cambridge university press, 2014. Abraham Wald. Statistical decision functions which minimize the maximum risk. Annals of Mathematics, pp. 265 280, 1945. Blake Woodworth, Suriya Gunasekar, Mesrob I Ohannessian, and Nathan Srebro. Learning nondiscriminatory predictors. In Conference on Learning Theory, pp. 1920 1953. PMLR, 2017. Yongkai Wu, Lu Zhang, and Xintao Wu. On convexity and bounds of fairness-aware classification. In The World Wide Web Conference, pp. 3356 3362, 2019. Published in Transactions on Machine Learning Research (08/2023) Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In Proceedings of the 26th international conference on world wide web, pp. 1171 1180, 2017a. Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rogriguez, and Krishna P Gummadi. Fairness constraints: Mechanisms for fair classification. In Artificial Intelligence and Statistics, pp. 962 970. PMLR, 2017b. Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. Learning fair representations. In International conference on machine learning, pp. 325 333. PMLR, 2013. Zhifei Zhang, Yang Song, and Hairong Qi. Age progression/regression by conditional adversarial autoencoder. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 5810 5818, 2017. Dominik Zietlow, Michael Lohaus, Guha Balakrishnan, Matthäus Kleindessner, Francesco Locatello, Bernhard Schölkopf, and Chris Russell. Leveling down in computer vision: Pareto inefficiencies in fair deep classifiers. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2022, New Orleans, LA, USA, June 18-24, 2022, pp. 10400 10411. IEEE, 2022. doi: 10.1109/CVPR52688.2022.01016. URL https://doi.org/10.1109/CVPR52688.2022.01016. Indre Žliobaite, Faisal Kamiran, and Toon Calders. Handling conditional discrimination. In 2011 IEEE 11th International Conference on Data Mining, pp. 992 1001. IEEE, 2011. In this appendix, we provide details that were omitted in the main paper. First, in Section A, we review several works closely related to ours. Then, in Section B, we show that several well known group fairness measures are compatible with Fair Grad. In Section C, we prove Lemma 1. Next, in Section D, we derive the update rules for Fair Grad with ϵ-fairness. Finally, in Section E, we provide additional experiments. A Related Work The fairness literature is extensive and we refer the interested reader to recent surveys (Caton & Haas, 2020; Mehrabi et al., 2021) to get an overview of the subject. Here, we focus on recent works that are more closely related to our approach. Bi Fair (Ozdayi et al., 2021). This paper proposes a bilevel optimization scheme for fairness. The idea is to use an outer optimization scheme that learns weights for each example so that the trade-offbetween fairness and accuracy is as favorable as possible while an inner optimization scheme learns a model that is as accurate as possible. One limitation of this approach is that it does not directly optimize the fairness level of the model but rather a relaxation that does not provide any guarantees on the goodness of the learned predictor. Furthermore, it is limited to binary classification with a binary sensitive attribute. In this paper, we also learn weights for the examples in an iterative way. However, we use a different update rule. Furthermore, we focus on exact fairness definitions rather than relaxations and our objective is to learn accurate models with given levels of fairness rather than a trade-offbetween the two. Finally, our approach is not limited to the binary setting. Fair Batch (Roh et al., 2020). This paper proposes a batch gradient descent approach to learn fair models. More precisely, the idea is to draw a batch of examples from a skewed distribution that favors the disadvantaged groups by oversampling them. In this paper, we propose to use a re-weighting approach which could also be interpreted as altering the distribution of the examples based on their fairness level if all the weights were positive. However, we allow the use of negative weights, and we prove that they are sometimes necessary to achieve fairness. Furthermore, we employ a different update rule for the weights. Published in Transactions on Machine Learning Research (08/2023) Ada Fair (Iosifidis & Ntoutsi, 2019). This paper proposes a boosting based framework to learn fair models. The underlying idea is to modify the weights of the examples depending on both the performances of the current strong classifier and the group memberships. Hence, examples that belong to the disadvantaged group and are incorrectly classified receive higher weights than the examples that belong to the advantaged group and are correctly classified. In this paper, we use a similar high level idea but we use different weights that do not depend on the accuracy of the model but solely on its fairness. Furthermore, rather than a boosting based approach, we consider problems that can be solved using gradient descent. Finally, while Ada Fair only focuses on Equalized Odds, we show that our approach works with several fairness notions. Identifying and Correcting Label Bias in Machine Learning (Jiang & Nachum, 2020). This paper tackles the fairness problem by assuming that the observed labels are biased compared to the true labels. The goal is then to learn a model with respect to the true labels using only the observed labels. To this end, it proposes to use an iterative re-weighting procedure where positive example-wise weights and the model are alternatively updated. In this paper, we also propose a re-weighting approach. However, we use different weights that are not necessarily positive. Furthermore, our approach is not limited to binary labels and can handle multiclass problems. B Reformulation of Various Group Fairness Notion In this section, we present several group fairness notions which respect our fairness definition presented in Section 2.1. Example 2 (Equalized Odds (EOdds) (Hardt et al., 2016)). A model hθ is fair for Equalized Odds when the probability of predicting the correct label is independent of the sensitive attribute, that is, l Y, r S b P (hθ(x) = l | s = r, y = l) = b P (hθ(x) = l | y = l) . It means that we need to partition the space into K = |Y S| groups and, l Y, r S, we define b F(l,r) as b F(l,r)(T , hθ) = b P (hθ(x) = l | y = l) b P (hθ(x) = l | s = r, y = l) (l,r ) =(l,r) b P (s = r |y = l) b P (hθ(x) = l | s = r , y = l) (1 b P (s = r|y = l))b P (hθ(x) = l | s = r, y = l) where the law of total probability was used to obtain the last equation. Thus, Equalized Odds satisfies all our assumptions with C(l,r) (l,r) = b P (s = r|y = l) 1, C(l,r ) (l,r) = b P (s = r |y = l), C(l ,r ) (l,r) = 0 with r = r and l = l, and C0 (l,r) = 0. Example 3 (Equality of Opportunity (EOpp) (Hardt et al., 2016)). A model hθ is fair for Equality of Opportunity when the probability of predicting the correct label is independent of the sensitive attribute for a given subset Y Y of labels called the desirable outcomes, that is, l Y , r S b P (hθ(x) = l | s = r, y = l) = b P (hθ(x) = l | y = l) . It means that we need to partition the space into K = |Y S| groups and, l Y, r S, we define b F(l,r) as b F(l,r)(T , hθ) = b P (hθ(x) = l | s = r, y = l) b P (hθ(x) = l | y = l) (l, r) Y S 0 (l, r) Y S \ Y S which can then be rewritten in the correct form in the same way as Equalized Odds, the only difference being that C (l,r) = 0, (l, r) Y S \ Y S. Published in Transactions on Machine Learning Research (08/2023) Example 4 (Demographic Parity (DP) (Calders et al., 2009)). A model hθ is fair for Demographic Parity when the probability of predicting a binary label is independent of the sensitive attribute, that is, l Y, r S b P (hθ(x) = l | s = r) = b P (hθ(x) = l) . It means that we need to partition the space into K = |Y S| groups and, l Y, r S, we define b F(l,r) as b F(l,r)(T , hθ) = b P (hθ(x) = l) b P (hθ(x) = l | s = r) = b P (y = l, s = r) b P (y = l | s = r) b P (hθ(x) = y | s = r, y = l) (l,r ) =(l,r) b P (y = l, s = r ) b P (hθ(x) = y | s = r , y = l) + b P y = l | s = r b P y = l, s = r b P hθ(x) = y | s = r, y = l ( l,r ) =( l,r) b P y = l, s = r b P hθ(x) = y | s = r , y = l b P y = l b P y = l | s = r where the law of total probability was used to obtain the last equation. Thus, Demographic Parity satisfies all our assumptions with C(l,r) (l,r) = b P (y = l, s = r) b P (y = l | s = r), C(l,r ) (l,r) = b P (y = l, s = r ) with r = r, C( l,r) (l,r) = b P y = l | s = r b P y = l, s = r , C( l,r ) (l,r) = b P y = l, s = r with r = r, and C0 (l,r) = b P y = l b P y = l | s = r . C Proof of Lemma 1 Lemma (Negative weights are necessary.). Assume that the fairness notion under consideration is Accuracy Parity. Let h θ be the most accurate and fair model. Then using negative weights is necessary as long as min hθ H hθunfair max Tk b P (hθ(x) = y|Tk) < b P (h θ(x) = y) . Proof. To prove this Lemma, one first need to notice that, for Accuracy Parity, since PK k=1 b P (Tk) = 1 we have that k =1 Ck k = (b P (Tk) 1) + b P (Tk ) = 0. This implies that " b P (Tk) + k =1 Ck k λk This implies that, whatever our choice of λ, the weights will always sum to one. In other words, since we also have that PK k=1 λk C0 k = 0 by definition, for a given hypothesis hθ, we have that max λ1,...,λK R k=1 b P (hθ(x) = y|Tk) " b P (Tk) + k =1 Ck k λk = max w1,...,w K R s.t.P k=1 b P (hθ(x) = y|Tk) wk (6) Published in Transactions on Machine Learning Research (08/2023) where, given w1, . . . , w K, the original values of lambda can be obtained by solving the linear system Cλ = w where C1 1 . . . C1 K ... ... CK 1 . . . CK K w1 b P (T1) ... w K b P (TK) which is guaranteed to have infinitely many solutions since the rank of the matrix C is K 1 and the rank of the augmented matrix (C|w) is also K 1. Here we are using the fact that b P (Tk) = 0, k since all the groups have to be represented to be taken into account. We will now assume that all the weights are positive, that is wk 0, k. Then, the best strategy to solve Problem (6) is to put all the weight on the worst offgroup k, that is set wk = 1 and wk = 0, k = k. It implies that max w1,...,w K R s.t.P k=1 b P (hθ(x) = y|Tk) wk = max k b P (hθ(x) = y|Tk) . Furthermore, notice that, for fair models with respect to Accuracy Parity, we have that b P (hθ(x) = y|Tk) = b P (hθ(x) = y) , k. Thus, if it holds that min hθ H hθunfair max Tk b P (hθ(x) = y|Tk) < b P (h θ(x) = y) where h θ is the most accurate and fair model, then the optimal solution of Problem (3) in the main paper will be unfair. It implies that, in this case, using positive weights is not sufficient and negative weights are necessary. D Fair Grad for ϵ-fairness To derive Fair Grad for ϵ-fairness we first consider the following standard optimization problem arg min hθ H b P (hθ(x) = y) s.t. k [K], b Fk(T , hθ) ϵ k [K], b Fk(T , hθ) ϵ. We, once again, use a standard multipliers approach to obtain the following unconstrained formulation: L (hθ, λ1, . . . , λK, δ1, . . . , δK) = b P (hθ(x) = y) + k=1 λk b Fk(T , hθ) ϵ δk b Fk(T , hθ) + ϵ (7) where λ1, . . . , λK and δ1, . . . , δK are the multipliers that belong to R+, that is the set of positive reals. Once again, to solve this problem, we will use an alternating approach where the hypothesis and the multipliers are updated one after the other. Updating the Multipliers. To update the values λ1, . . . , λK, we will use a standard gradient ascent procedure. Hence, noting that the gradient of the previous formulation is λ1,...,λKL (hθ, λ1, . . . , λK, δ1, . . . , δK) = b F1(T , hθ) ϵ ... b FK(T , hθ) ϵ Published in Transactions on Machine Learning Research (08/2023) δ1,...,δKL (hθ, λ1, . . . , λK, δ1, . . . , δK) = b F1(T , hθ) ϵ ... b FK(T , hθ) ϵ we have the following update rule k [K] λT +1 k = max 0, λT k + η b Fk T , h T θ ϵ δT +1 k = max 0, δT k η b Fk T , h T θ + ϵ where η is a fairness rate that controls the importance of each weight update. Updating the Model. To update the parameters θ RD of the model hθ, we proceed as before, using a gradient descent approach. However, first, we notice that given the fairness notions that we consider, Equation (7) is equivalent to L (hθ, λ1, . . . , λK, δ1, . . . , δK) = k=1 b P (hθ(x) = y|Tk) " b P (Tk) + k =1 Ck k (λk δk ) k=1 (λk + δk) ϵ + k=1 (λk δk)C0 k. Since the additional terms in the optimization problem do not depend on hθ, the main difference between exact and ϵ-fairness is the nature of the weights. More precisely, at iteration t, the update rule becomes θT +1 = θT ηθ " b P (Tk) + k =1 Ck k (λk δk ) θb P (hθ(x) = y|Tk) where ηθ is a learning rate. Once again, we obtain a simple re-weighting scheme where the weights depend on the current fairness level of the model through λ1, . . . , λK and δ1, . . . , δK, the relative size of each group through b P (Tk), and the fairness notion through the constants C. E Extended Experiments In this section, we provide additional details related to the baselines and the hyper-parameters tuning procedure. We then provide descriptions of the datasets and finally the results. E.1 Baselines Adversarial: One of the common ways of removing sensitive information from the model s representation is via adversarial learning. Broadly, it consists of three components, namely an encoder, a task classifier, and an adversary. On the one hand, the objective of the adversary is to predict sensitive information from the encoder. On the other hand, the encoder aims to create representations that are useful for the downstream task (task classifier) and, at the same time, fool the adversary. The adversary is generally connected to the encoder via a gradient reversal layer (Ganin & Lempitsky, 2015) which acts like an identity function during the forward pass and scales the loss with a parameter λ during the backward pass. In our setting, the encoder is a Multi-Layer Perceptron with two hidden layers of size 64 and 128 respectively, and the task classifier is another Multi-Layer Perceptron with a single hidden layer of size 32. The adversary is the same as the main task classifier. We use a Re LU as the activation function with the dropout set to 0.2 and employ batch normalization with default Py Torch parameters. As a part of the hyper-parameter tuning, we did a grid search over λ, varying it between 0.1 to 3.0 with an interval of 0.2. Published in Transactions on Machine Learning Research (08/2023) Bi Fair (Ozdayi et al., 2021): For this baseline, we fix the weight parameter to be of length 8 as suggested in the code released by the authors5. In this fixed setting, we perform a grid search over the following hyper-parameters: Batch Size: 128, 256, 512 Weight Decay: 0.0, 0.001 Fairness Loss Weight: 0.5, 1, 2, 4 Inner Loop Length: 5, 25, 50 Constraints: We use the implementation available in the Tensor Flow Constrained Optimization6 library with default hyper-parameters. Fair Batch: We use the implementation publicly released by the authors7. Weighted ERM: We reweigh each example in the dataset based on inverse of the proportion of the sensitive group it belongs to. Reduction: We use the implementation available in the Fairlearn8 with default hyper-parameters. In our initial experiments, we varied the batch size, and learning rates for both Constraints and Fair Batch. However, we found that the default hyper-parameters as specified by the authors result in the best performances. In the spirit of being comparable in terms of hyper-parameter search budget, we also fix all hyper-parameters of Fair Grad, apart from the batch size and weight decay. We experiment with two different batch sizes namely, 64 or 512 for the standard fairness dataset. Similarly, we also experiment with three weight decay values namely, 0.0, 0.001 and 0.01. Note that we also vary weight decay and batch sizes for Fair Batch, Adversarial, Unconstrained, and Bi Fair. For all our experiments, apart from Bi Fair, we use Batch Gradient Descent as the optimizer with a learning rate of 0.1 and a gradient clipping of 0.05 to avoid exploding gradients. For Bi Fair, we employ the Adam optimizer as suggested by the authors with a learning rate of 0.001. For Fair Grad, Fair Batch and Unconstrained, we considered 6 hyper-parameters combinations. For Bi Fair, we considered 72 such combinations, while for Adversarial, there were 90 combinations. E.2 Datasets Here, we provide additional details on the datasets used in our experiments. We begin by describing the standard fairness datasets for which we follow the pre-processing procedure described in Lohaus et al. (2020). Adult9: The dataset (Kohavi, 1996) is composed of 45222 instances, with 14 features each describing several attributes of a person. The objective is to predict the income of a person (below or above 50k) while remaining fair with respect to gender (binary in this case). Following the pre-processing step of Wu et al. (2019), only 9 features were used for training. Celeb A10: The dataset (Liu et al., 2015) consists of 202, 599 images, along with 40 binary attributes associated with each image. We use 38 of these as features while keeping gender as the sensitive attribute and Smiling as the class label. Dutch11: The dataset (Žliobaite et al., 2011) is composed of 60, 420 instances with each instance described by 12 features. We predict Low Income or High Income as dictated by the occupation as the main classification task and gender as the sensitive attribute. 5https://github.com/Tinfoil Hat0/Bi Fair 6https://github.com/google-research/tensorflow_constrained_optimization 7https://github.com/yuji-roh/fairbatch 8https://fairlearn.org/ 9https://archive.ics.uci.edu/ml/datasets/adult 10https://mmlab.ie.cuhk.edu.hk/projects/Celeb A.html 11https://sites.google.com/site/conditionaldiscrimination/ Published in Transactions on Machine Learning Research (08/2023) Compas12: The dataset (Larson et al., 2016) contains 6172 data points, where each data point has 53 features. The goal is to predict if the defendant will be arrested again within two years of the decision. The sensitive attribute is race, which has been merged into White and Non White categories. Communities and Crime13: The dataset (Redmond & Baveja, 2002) is composed of 1994 instances with 128 features, of which 29 have been dropped. The objective is to predict the number of violent crimes in the community, with race being the sensitive attribute. German Credit14: The dataset (Dua et al., 2017) consists of 1000 instances, with each having 20 attributes. The objective is to predict a person s creditworthiness (binary), with gender being the sensitive attribute. Gaussian15: It is a toy dataset with binary task label and binary sensitive attribute, introduced in Lohaus et al. (2020). It is constructed by drawing points from different Gaussian distributions. We follow the same mechanism as described in Lohaus et al. (2020), and sample 50000 data points for each class. Adult Folktables16: This dataset (Ding et al., 2021) is an updated version of the original Adult Income dataset. We use California census data with gender as the sensitive attribute. There are 195665 instances, with 9 features describing several attributes of a person. We use the same preprocessing step as recommended by the authors. For all these datasets, we use a 20% of the data as a test set and 80% as a train set. We further divide the train set into two and keep 25% of the training examples as a validation set. For each repetition, we randomly shuffle the data before splitting it, and thus we had unique splits for each random seed. We use the following seeds: 10, 20, 30, 40, 50 for all our experiments. As a last pre-processing step, we centered and scaled each feature independently by substracting the mean and dividing by the standard deviation both of which were estimated on the training set. Twitter Sentiment Analysis17: The dataset (Blodgett et al., 2016) consists of 200k tweets with binary sensitive attribute (race) and binary sentiment score. We follow the setup proposed by Han et al. (2021) and Elazar & Goldberg (2018) and create bias in the dataset by changing the proportion of each subgroup (race-sentiment) in the training set. With two sentiment classes being happy and sad, and two race classes being AAE and SAE, the training data consists of 40% AAE-happy, 10% AAE-sad, 10% SAE-happy, and 40% SAE-sad. The test set remains balanced. The tweets are encoded using the Deep Moji (Felbo et al., 2017) encoder with no fine-tuning, which has been pre-trained over millions of tweets to predict their emoji, thereby predicting the sentiment. Note that the train-test splits are pre-defined and thus do not change based on the random seed of the repetition. E.3 Detailed Results 12https://github.com/propublica/compas-analysis 13http://archive.ics.uci.edu/ml/datasets/communities+and+crime 14https://archive.ics.uci.edu/ml/datasets/Statlog+%28German+Credit+Data%29 15https://github.com/mlohaus/Search Fair/blob/master/examples/get_synthetic_data.py 16https://github.com/zykls/folktables 17https://slanglab.cs.umass.edu/Twitter AAE/ Published in Transactions on Machine Learning Research (08/2023) 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.67 0.71 0.75 0.80 0.84 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 6: Results for the Adult dataset with different fairness measures. Table 4: Results for the Adult dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8456 0.0033 AP 0.0571 0.0022 0.077 0.0029 -0.0373 0.0017 Constant 0.751 0.0 AP 0.102 0.0 0.138 0.0 0.067 0.0 Weighted ERM 0.8442 0.0016 AP 0.0581 0.0021 0.0783 0.0028 -0.0379 0.0014 Constrained 0.783 0.007 AP 0.005 0.003 0.007 0.005 0.004 0.002 Reduction 0.7064 0.0315 AP 0.0361 0.0158 0.0235 0.0103 -0.0487 0.0214 Fair Grad 0.8124 0.005 AP 0.0097 0.0029 0.0131 0.004 -0.0063 0.0019 Unconstrained 0.846 0.0028 Eodds 0.0453 0.0039 0.048 0.0043 -0.0878 0.01 Constant 0.748 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8475 0.0024 Eodds 0.044 0.0043 0.0477 0.0031 -0.0837 0.0124 Constrained 0.805 0.004 Eodds 0.007 0.005 0.019 0.017 0.002 0.001 Bi Fair 0.793 0.009 Eodds 0.036 0.008 0.085 0.027 -0.03 0.016 Fair Batch 0.8437 0.0013 Eodds 0.0228 0.0071 0.0411 0.0105 -0.0245 0.0183 Reduction 0.7059 0.0277 Eodds 0.0542 0.0158 0.0711 0.0189 -0.1055 0.022 Fair Grad 0.8284 0.004 Eodds 0.0051 0.0021 0.0078 0.0068 -0.0078 0.0054 Unconstrained 0.8457 0.0028 Eopp 0.0263 0.0024 0.0157 0.0011 -0.0893 0.0083 Constant 0.754 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8475 0.0024 Eopp 0.0246 0.0036 0.0148 0.002 -0.0837 0.0124 Constrained 0.846 0.002 Eopp 0.011 0.004 0.039 0.012 0.0 0.0 Bi Fair 0.8 0.009 Eopp 0.031 0.024 0.019 0.014 -0.107 0.083 Fair Batch 0.8457 0.0016 Eopp 0.0098 0.0068 0.0225 0.0174 -0.0166 0.0241 Reduction 0.8226 0.0149 Eopp 0.0341 0.0168 0.116 0.0575 -0.0204 0.0098 Fair Grad 0.8353 0.0106 Eopp 0.0053 0.006 0.0177 0.021 -0.0037 0.0033 Published in Transactions on Machine Learning Research (08/2023) Table 5: Results for the Adult dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8438 0.0025 AP 0.0575 0.0025 0.0776 0.0033 -0.0375 0.0018 Constant 0.751 0.0 AP 0.102 0.0 0.138 0.0 0.067 0.0 Weighted ERM 0.8469 0.0035 AP 0.0564 0.003 0.0761 0.0038 -0.0368 0.0021 Adversarial 0.8364 0.0063 AP 0.0526 0.0017 0.0709 0.0025 -0.0343 0.0009 Reduction 0.7015 0.0225 AP 0.0681 0.0184 0.0444 0.0122 -0.0917 0.0247 Fair Grad 0.8054 0.0051 AP 0.0034 0.0033 0.0033 0.0031 -0.0036 0.0042 Unconstrained 0.8299 0.0142 Eodds 0.0448 0.0109 0.0404 0.0136 -0.0977 0.0422 Constant 0.748 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8285 0.0085 Eodds 0.0102 0.0025 0.0196 0.0102 -0.0099 0.0047 Adversarial 0.8202 0.0068 Eodds 0.0145 0.0052 0.0288 0.0177 -0.0153 0.0067 Bi Fair 0.823 0.017 Eodds 0.038 0.009 0.09 0.034 -0.038 0.015 Fair Batch 0.8379 0.0009 Eodds 0.02 0.0088 0.0327 0.0153 -0.0244 0.0218 Reduction 0.729 0.0252 Eodds 0.0636 0.0176 0.0673 0.0203 -0.115 0.0334 Fair Grad 0.827 0.0071 Eodds 0.0118 0.0024 0.022 0.014 -0.0165 0.0135 Unconstrained 0.8382 0.0076 Eopp 0.0242 0.0031 0.0145 0.0017 -0.0822 0.0108 Constant 0.754 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8293 0.0091 Eopp 0.0051 0.0033 0.0141 0.0137 -0.0062 0.0038 Adversarial 0.8324 0.0058 Eopp 0.007 0.0044 0.0139 0.0159 -0.0144 0.0133 Bi Fair 0.815 0.014 Eopp 0.03 0.015 0.019 0.009 -0.103 0.053 Fair Batch 0.8415 0.0054 Eopp 0.0082 0.0073 0.0157 0.0121 -0.017 0.0271 Reduction 0.8343 0.0059 Eopp 0.0294 0.0164 0.0779 0.0662 -0.0396 0.0455 Fair Grad 0.8373 0.0043 Eopp 0.0053 0.0047 0.0099 0.0146 -0.0112 0.0127 Published in Transactions on Machine Learning Research (08/2023) 0.67 0.71 0.76 0.80 0.85 0.89 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.67 0.71 0.76 0.80 0.85 0.89 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.67 0.71 0.76 0.80 0.85 0.89 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.67 0.71 0.76 0.80 0.85 0.89 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.67 0.71 0.76 0.80 0.85 0.89 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.67 0.71 0.76 0.80 0.85 0.89 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 7: Results for the Celeb A dataset with different fairness measures. Table 6: Results for the Celeb A dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8532 0.0009 AP 0.0204 0.0022 0.017 0.0019 -0.0238 0.0025 Constant 0.516 0.0 AP 0.072 0.0 0.084 0.0 0.06 0.0 Weighted ERM 0.853 0.0008 AP 0.0193 0.0021 0.0161 0.0018 -0.0225 0.0023 Constrained 0.799 0.013 AP 0.01 0.001 0.012 0.002 0.009 0.001 Reduction 0.7734 0.011 AP 0.0242 0.006 0.0282 0.0071 -0.0201 0.005 Fair Grad 0.835 0.0028 AP 0.0012 0.0009 0.0011 0.0007 -0.0014 0.0011 Unconstrained 0.8532 0.0009 Eodds 0.0499 0.0019 0.0538 0.0024 -0.1011 0.0033 Constant 0.518 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.853 0.0009 Eodds 0.0504 0.0019 0.0532 0.0024 -0.1001 0.0032 Constrained 0.802 0.004 Eodds 0.006 0.001 0.01 0.003 0.002 0.001 Bi Fair 0.845 0.007 Eodds 0.021 0.005 0.02 0.003 -0.036 0.009 Fair Batch 0.8518 0.0009 Eodds 0.0226 0.0017 0.0218 0.0028 -0.0411 0.0053 Reduction 0.7268 0.011 Eodds 0.0312 0.0036 0.0628 0.0089 -0.0334 0.0047 Fair Grad 0.8274 0.002 Eodds 0.0025 0.0009 0.0038 0.0018 -0.0046 0.0026 Unconstrained 0.8532 0.0009 Eopp 0.0387 0.0014 0.0538 0.0024 -0.1011 0.0033 Constant 0.518 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.853 0.0008 Eopp 0.0383 0.0014 0.0531 0.0024 -0.0999 0.0032 Constrained 0.834 0.005 Eopp 0.002 0.001 0.005 0.002 0.0 0.0 Bi Fair 0.848 0.004 Eopp 0.014 0.006 0.02 0.009 -0.037 0.017 Fair Batch 0.8498 0.001 Eopp 0.0102 0.0016 0.0142 0.0022 -0.0268 0.0042 Reduction 0.7358 0.0159 Eopp 0.0698 0.0118 0.1824 0.0313 -0.0968 0.0158 Fair Grad 0.844 0.0022 Eopp 0.0013 0.0009 0.0025 0.0021 -0.0028 0.0018 Published in Transactions on Machine Learning Research (08/2023) Table 7: Results for the Celeb A dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8587 0.0015 AP 0.0184 0.0014 0.0154 0.0012 -0.0215 0.0016 Constant 0.516 0.0 AP 0.072 0.0 0.084 0.0 0.06 0.0 Weighted ERM 0.8593 0.0018 AP 0.018 0.0017 0.015 0.0014 -0.021 0.0019 Adversarial 0.8588 0.0012 AP 0.0178 0.0014 0.0148 0.0012 -0.0208 0.0015 Reduction 0.7802 0.0142 AP 0.0436 0.0108 0.0508 0.0123 -0.0364 0.0092 Fair Grad 0.8359 0.0033 AP 0.0023 0.0012 0.0025 0.0015 -0.0021 0.0009 Unconstrained 0.8583 0.0012 Eodds 0.0432 0.003 0.0475 0.0028 -0.0893 0.0049 Constant 0.518 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8589 0.0009 Eodds 0.0419 0.0021 0.0459 0.0025 -0.0864 0.0038 Adversarial 0.8567 0.0014 Eodds 0.0223 0.002 0.0272 0.0039 -0.0511 0.0073 Bi Fair 0.856 0.004 Eodds 0.023 0.002 0.028 0.005 -0.052 0.009 Fair Batch 0.8533 0.0037 Eodds 0.0217 0.0014 0.0197 0.0026 -0.0321 0.005 Reduction 0.7021 0.0323 Eodds 0.0813 0.0253 0.1777 0.0426 -0.0946 0.0238 Fair Grad 0.8304 0.0031 Eodds 0.0037 0.0017 0.0048 0.0018 -0.0055 0.0023 Unconstrained 0.8585 0.0016 Eopp 0.0341 0.002 0.0473 0.003 -0.0889 0.0052 Constant 0.518 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.859 0.0009 Eopp 0.0331 0.0014 0.046 0.0023 -0.0866 0.0035 Adversarial 0.8557 0.0019 Eopp 0.0161 0.002 0.0223 0.0029 -0.0419 0.0053 Bi Fair 0.854 0.004 Eopp 0.015 0.009 0.021 0.012 -0.039 0.022 Fair Batch 0.8475 0.0043 Eopp 0.0051 0.0024 0.007 0.0033 -0.0131 0.0063 Reduction 0.765 0.0149 Eopp 0.0533 0.0124 0.1393 0.033 -0.0738 0.0167 Fair Grad 0.8439 0.0063 Eopp 0.0009 0.0008 0.002 0.0022 -0.0016 0.0011 Published in Transactions on Machine Learning Research (08/2023) 0.63 0.68 0.74 0.79 0.85 0.90 Accuracy Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.63 0.68 0.74 0.79 0.85 0.90 Accuracy Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.63 0.68 0.74 0.79 0.85 0.90 Accuracy Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.63 0.68 0.74 0.79 0.85 0.90 Accuracy Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.63 0.68 0.74 0.79 0.85 0.90 Accuracy Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.63 0.68 0.74 0.79 0.85 0.90 Accuracy Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 8: Results for the Crime dataset with different fairness measures. Table 8: Results for the Crime dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8145 0.0136 AP 0.0329 0.0195 0.0258 0.0162 -0.0399 0.0229 Constant 0.734 0.0 AP 0.272 0.0 0.377 0.0 0.168 0.0 Weighted ERM 0.808 0.0246 AP 0.0361 0.0108 0.0284 0.0091 -0.0438 0.0129 Constrained 0.775 0.015 AP 0.025 0.019 0.031 0.025 0.019 0.014 Reduction 0.8521 0.0075 AP 0.055 0.0197 0.0426 0.0147 -0.0673 0.0253 Fair Grad 0.814 0.0102 AP 0.0403 0.0181 0.0316 0.0147 -0.049 0.0218 Unconstrained 0.8035 0.0212 Eodds 0.2152 0.0215 0.1038 0.0231 -0.396 0.0433 Constant 0.677 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8045 0.0271 Eodds 0.2086 0.0357 0.0974 0.0165 -0.3747 0.0679 Constrained 0.751 0.014 Eodds 0.036 0.012 0.088 0.043 0.007 0.004 Bi Fair 0.76 0.03 Eodds 0.082 0.048 0.048 0.03 -0.163 0.092 Fair Batch 0.8306 0.0237 Eodds 0.2015 0.035 0.1054 0.0333 -0.3704 0.067 Reduction 0.6842 0.0339 Eodds 0.0611 0.0281 0.0349 0.0111 -0.1291 0.047 Fair Grad 0.7634 0.03 Eodds 0.0938 0.0144 0.0491 0.016 -0.1927 0.0362 Unconstrained 0.804 0.0215 Eopp 0.1215 0.0183 0.1009 0.0238 -0.3852 0.0549 Constant 0.697 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8171 0.0213 Eopp 0.1209 0.0154 0.0985 0.0106 -0.3851 0.0599 Constrained 0.762 0.021 Eopp 0.044 0.021 0.138 0.066 0.0 0.0 Bi Fair 0.806 0.01 Eopp 0.085 0.038 0.073 0.042 -0.268 0.112 Fair Batch 0.8225 0.0252 Eopp 0.1126 0.0259 0.1002 0.0281 -0.3501 0.0821 Reduction 0.6747 0.0488 Eopp 0.0283 0.022 0.0413 0.0375 -0.0718 0.0829 Fair Grad 0.7755 0.0233 Eopp 0.0609 0.0149 0.0507 0.0166 -0.193 0.0456 Published in Transactions on Machine Learning Research (08/2023) Table 9: Results for the Crime dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8165 0.019 AP 0.0535 0.0199 0.0423 0.0155 -0.0648 0.0251 Constant 0.734 0.0 AP 0.272 0.0 0.377 0.0 0.168 0.0 Weighted ERM 0.8271 0.0114 AP 0.0483 0.0167 0.0382 0.0139 -0.0584 0.02 Adversarial 0.809 0.0175 AP 0.0592 0.0173 0.0464 0.0135 -0.0719 0.0223 Reduction 0.8501 0.0096 AP 0.0559 0.0215 0.0432 0.0166 -0.0685 0.0269 Fair Grad 0.822 0.0203 AP 0.0434 0.0206 0.0341 0.0162 -0.0526 0.0252 Unconstrained 0.8115 0.014 Eodds 0.1635 0.0395 0.0854 0.014 -0.3326 0.0649 Constant 0.677 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8135 0.0137 Eodds 0.1739 0.0394 0.0861 0.0212 -0.3309 0.0778 Adversarial 0.791 0.007 Eodds 0.1464 0.0168 0.0797 0.0192 -0.3001 0.0296 Bi Fair 0.793 0.022 Eodds 0.161 0.032 0.091 0.025 -0.339 0.048 Fair Batch 0.8391 0.0195 Eodds 0.189 0.0368 0.1106 0.0313 -0.3828 0.0671 Reduction 0.7258 0.0267 Eodds 0.0743 0.0409 0.0553 0.014 -0.1556 0.0976 Fair Grad 0.7734 0.0251 Eodds 0.0982 0.0513 0.0511 0.0179 -0.2016 0.0771 Unconstrained 0.817 0.0152 Eopp 0.1044 0.0133 0.0856 0.0123 -0.3321 0.0489 Constant 0.697 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8205 0.0184 Eopp 0.1159 0.0191 0.0955 0.019 -0.368 0.0642 Adversarial 0.795 0.0148 Eopp 0.0959 0.0153 0.0802 0.0227 -0.3036 0.042 Bi Fair 0.807 0.025 Eopp 0.11 0.031 0.091 0.031 -0.351 0.097 Fair Batch 0.8411 0.0177 Eopp 0.1217 0.0277 0.1083 0.0311 -0.3784 0.0891 Reduction 0.6887 0.0271 Eopp 0.0282 0.0159 0.034 0.0281 -0.0788 0.0619 Fair Grad 0.7799 0.0243 Eopp 0.0675 0.0179 0.0556 0.0147 -0.2143 0.0592 Published in Transactions on Machine Learning Research (08/2023) 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction (a) Linear - AP 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Fair Grad Fair Batch Unconstrained Weighted Erm Reduction (b) Linear - EOdds 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction (c) Linear - EOpp 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction (e) Non Linear - EOdds 0.61 0.66 0.72 0.77 0.83 0.88 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction (f) Non Linear - EOpp Figure 9: Results for the Adult with multiple groups dataset with different fairness measures. Table 10: Results for the Adult with multiple groups dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8451 0.0042 AP 0.0559 0.0047 0.0985 0.0111 -0.042 0.003 Constant 0.754 0.0 AP 0.097 0.0 0.159 0.0 0.024 0.0 Weighted ERM 0.8454 0.0032 AP 0.0562 0.0042 0.0993 0.0117 -0.0426 0.0018 Reduction 0.6436 0.0178 AP 0.049 0.01 0.0493 0.017 -0.0661 0.0113 Fair Grad 0.807 0.0022 AP 0.0148 0.0041 0.0256 0.0048 -0.0107 0.0045 Unconstrained 0.844 0.0011 Eodds 0.0558 0.0062 0.0578 0.0069 -0.1586 0.0621 Constant 0.75 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8448 0.0038 Eodds 0.0586 0.0097 0.0567 0.0048 -0.1702 0.0776 Fair Batch 0.8396 0.0034 Eodds 0.0308 0.0057 0.0565 0.0116 -0.0641 0.0234 Reduction 0.6932 0.0264 Eodds 0.0446 0.0048 0.0806 0.043 -0.0896 0.0278 Fair Grad 0.8162 0.0052 Eodds 0.0197 0.0118 0.0373 0.0233 -0.0493 0.0403 Unconstrained 0.8431 0.002 Eopp 0.0391 0.0052 0.0297 0.0131 -0.169 0.0565 Constant 0.762 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8443 0.0038 Eopp 0.0415 0.01 0.0316 0.0145 -0.1767 0.0797 Fair Batch 0.8392 0.004 Eopp 0.0219 0.0055 0.05 0.0133 -0.0749 0.0285 Reduction 0.7615 0.0357 Eopp 0.026 0.0189 0.0487 0.0378 -0.1115 0.0867 Fair Grad 0.834 0.0044 Eopp 0.0201 0.0099 0.0442 0.0415 -0.0679 0.0808 Published in Transactions on Machine Learning Research (08/2023) Table 11: Results for the Adult with multiple groups dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8427 0.0041 AP 0.0546 0.0026 0.0966 0.0098 -0.0421 0.0022 Constant 0.754 0.0 AP 0.097 0.0 0.159 0.0 0.024 0.0 Weighted ERM 0.8408 0.0031 AP 0.0575 0.0035 0.101 0.0106 -0.0443 0.0026 Adversarial 0.8358 0.0043 AP 0.0527 0.0028 0.0889 0.0066 -0.0401 0.0022 Reduction 0.7025 0.0144 AP 0.0388 0.0066 0.054 0.0151 -0.0525 0.0099 Fair Grad 0.7991 0.0036 AP 0.013 0.0051 0.0257 0.0138 -0.0125 0.0043 Unconstrained 0.8347 0.0129 Eodds 0.0523 0.0126 0.0495 0.0166 -0.1772 0.0512 Constant 0.75 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8199 0.002 Eodds 0.0287 0.0076 0.0274 0.0177 -0.1013 0.0543 Adversarial 0.8251 0.0064 Eodds 0.0223 0.0065 0.0451 0.0308 -0.0667 0.0559 Fair Batch 0.8212 0.0103 Eodds 0.0806 0.0137 0.0522 0.0076 -0.2545 0.0525 Reduction 0.7649 0.0241 Eodds 0.0386 0.011 0.044 0.02 -0.0954 0.0465 Fair Grad 0.8128 0.0102 Eodds 0.0196 0.0061 0.0392 0.0176 -0.0443 0.0342 Unconstrained 0.8373 0.0123 Eopp 0.0331 0.008 0.0183 0.0045 -0.1587 0.0643 Constant 0.762 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8216 0.0031 Eopp 0.0245 0.008 0.0243 0.0196 -0.1016 0.0543 Adversarial 0.8343 0.0036 Eopp 0.0209 0.0093 0.0327 0.013 -0.0927 0.0589 Fair Batch 0.821 0.0097 Eopp 0.067 0.0168 0.047 0.0113 -0.2484 0.0535 Reduction 0.8156 0.0204 Eopp 0.0259 0.0209 0.0472 0.0325 -0.0968 0.1117 Fair Grad 0.8341 0.0053 Eopp 0.0176 0.0059 0.0302 0.0272 -0.0731 0.0543 Published in Transactions on Machine Learning Research (08/2023) 0.50 0.55 0.60 0.64 0.69 0.73 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.50 0.55 0.60 0.64 0.69 0.73 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.50 0.55 0.60 0.64 0.69 0.73 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.50 0.55 0.60 0.64 0.69 0.73 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.50 0.55 0.60 0.64 0.69 0.73 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.50 0.55 0.60 0.64 0.69 0.73 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 10: Results for the Compas dataset with different fairness measures. Table 12: Results for the Compas dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.6644 0.0137 AP 0.0091 0.0025 0.0076 0.0031 -0.0107 0.004 Constant 0.545 0.0 AP 0.066 0.0 0.085 0.0 0.047 0.0 Weighted ERM 0.6671 0.0169 AP 0.0088 0.004 0.0061 0.0028 -0.0115 0.0051 Constrained 0.65 0.012 AP 0.014 0.005 0.018 0.006 0.009 0.003 Reduction 0.6141 0.011 AP 0.0107 0.0064 0.009 0.006 -0.0124 0.0086 Fair Grad 0.6708 0.0166 AP 0.0083 0.0068 0.0057 0.0048 -0.0108 0.0088 Unconstrained 0.6636 0.0104 Eodds 0.0827 0.0165 0.0758 0.0133 -0.1553 0.0259 Constant 0.527 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.6685 0.0073 Eodds 0.082 0.0137 0.0697 0.0115 -0.1618 0.0222 Constrained 0.564 0.014 Eodds 0.007 0.004 0.014 0.011 0.002 0.001 Bi Fair 0.672 0.021 Eodds 0.076 0.023 0.071 0.025 -0.15 0.039 Fair Batch 0.6847 0.0175 Eodds 0.09 0.0094 0.0854 0.0149 -0.1727 0.0304 Reduction 0.5493 0.027 Eodds 0.029 0.0058 0.0268 0.0062 -0.0622 0.0219 Fair Grad 0.6557 0.0075 Eodds 0.0593 0.0128 0.0524 0.0102 -0.1241 0.0202 Unconstrained 0.6609 0.0106 Eopp 0.052 0.0107 0.062 0.0145 -0.1461 0.0286 Constant 0.55 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.6695 0.0055 Eopp 0.0554 0.0074 0.0659 0.0107 -0.1557 0.0194 Constrained 0.565 0.015 Eopp 0.004 0.003 0.011 0.009 0.0 0.0 Bi Fair 0.68 0.013 Eopp 0.054 0.016 0.064 0.022 -0.15 0.044 Fair Batch 0.6865 0.0171 Eopp 0.0618 0.0134 0.0715 0.0173 -0.1755 0.0364 Reduction 0.5828 0.0457 Eopp 0.0252 0.0178 0.03 0.0216 -0.0707 0.0498 Fair Grad 0.6565 0.0152 Eopp 0.0467 0.0046 0.0554 0.0071 -0.1313 0.0119 Published in Transactions on Machine Learning Research (08/2023) Table 13: Results for the Compas dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.6593 0.0192 AP 0.0119 0.0072 0.0095 0.004 -0.0144 0.0107 Constant 0.545 0.0 AP 0.066 0.0 0.085 0.0 0.047 0.0 Weighted ERM 0.6687 0.0138 AP 0.0127 0.0061 0.011 0.0034 -0.0145 0.0099 Adversarial 0.6583 0.0157 AP 0.0078 0.0051 0.0066 0.0044 -0.009 0.0069 Reduction 0.6287 0.0117 AP 0.0118 0.0024 0.0103 0.0062 -0.0134 0.0024 Fair Grad 0.6672 0.0099 AP 0.0113 0.005 0.0095 0.0023 -0.0131 0.0082 Unconstrained 0.6562 0.0154 Eodds 0.0782 0.014 0.0715 0.0136 -0.1521 0.0277 Constant 0.527 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.6615 0.0175 Eodds 0.0789 0.0131 0.0726 0.0077 -0.1496 0.0313 Adversarial 0.6504 0.0157 Eodds 0.059 0.0138 0.0549 0.0107 -0.1294 0.0183 Bi Fair 0.661 0.009 Eodds 0.07 0.013 0.068 0.018 -0.133 0.016 Fair Batch 0.6792 0.0086 Eodds 0.071 0.0083 0.0663 0.0091 -0.1508 0.0304 Reduction 0.5631 0.0072 Eodds 0.0214 0.0112 0.024 0.0102 -0.0489 0.0363 Fair Grad 0.6457 0.0088 Eodds 0.061 0.0075 0.0564 0.0065 -0.127 0.0081 Unconstrained 0.6552 0.0137 Eopp 0.0553 0.0108 0.0659 0.015 -0.1552 0.0281 Constant 0.55 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.6604 0.0163 Eopp 0.0519 0.0111 0.0618 0.0148 -0.1458 0.0299 Adversarial 0.6494 0.0148 Eopp 0.0472 0.0072 0.0563 0.0108 -0.1327 0.0183 Bi Fair 0.669 0.01 Eopp 0.042 0.02 0.05 0.025 -0.117 0.055 Fair Batch 0.6802 0.0114 Eopp 0.0536 0.0133 0.062 0.0167 -0.1526 0.0367 Reduction 0.5801 0.0258 Eopp 0.025 0.0119 0.0296 0.0145 -0.0702 0.0333 Fair Grad 0.6586 0.0118 Eopp 0.0476 0.0056 0.0563 0.0067 -0.1339 0.0163 Published in Transactions on Machine Learning Research (08/2023) 0.64 0.68 0.72 0.77 0.81 0.85 Accuracy Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.64 0.68 0.72 0.77 0.81 0.85 Accuracy Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.64 0.68 0.72 0.77 0.81 0.85 Accuracy Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.64 0.68 0.72 0.77 0.81 0.85 Accuracy Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.64 0.68 0.72 0.77 0.81 0.85 Accuracy Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.64 0.68 0.72 0.77 0.81 0.85 Accuracy Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 11: Results for the Dutch dataset with different fairness measures. Table 14: Results for the Dutch dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8049 0.007 AP 0.0281 0.006 0.0281 0.006 -0.0282 0.0061 Constant 0.524 0.0 AP 0.151 0.0 0.152 0.0 0.15 0.0 Weighted ERM 0.8052 0.0073 AP 0.028 0.006 0.028 0.006 -0.0281 0.006 Constrained 0.799 0.009 AP 0.009 0.006 0.009 0.006 0.009 0.006 Reduction 0.723 0.0341 AP 0.0367 0.0172 0.0368 0.0172 -0.0367 0.0172 Fair Grad 0.8042 0.0046 AP 0.0048 0.0033 0.0048 0.0033 -0.0048 0.0032 Unconstrained 0.8071 0.0072 Eodds 0.0212 0.0018 0.0322 0.009 -0.0256 0.0052 Constant 0.522 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8074 0.0074 Eodds 0.0213 0.002 0.032 0.0086 -0.0254 0.0051 Constrained 0.79 0.005 Eodds 0.005 0.003 0.009 0.005 0.002 0.002 Bi Fair 0.804 0.008 Eodds 0.021 0.003 0.025 0.004 -0.033 0.01 Fair Batch 0.809 0.0096 Eodds 0.018 0.0016 0.0262 0.0039 -0.0211 0.004 Reduction 0.6716 0.0251 Eodds 0.0226 0.006 0.0333 0.0107 -0.0404 0.0213 Fair Grad 0.7978 0.0064 Eodds 0.0053 0.0019 0.007 0.0019 -0.009 0.0049 Unconstrained 0.8129 0.0021 Eopp 0.0075 0.0034 0.0107 0.0049 -0.0193 0.0086 Constant 0.524 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8077 0.0078 Eopp 0.0076 0.0034 0.011 0.0049 -0.0196 0.0087 Constrained 0.814 0.003 Eopp 0.003 0.002 0.007 0.006 0.0 0.0 Bi Fair 0.808 0.01 Eopp 0.005 0.005 0.008 0.007 -0.012 0.012 Fair Batch 0.8149 0.0117 Eopp 0.0031 0.0014 0.0044 0.002 -0.0079 0.0036 Reduction 0.7397 0.0176 Eopp 0.026 0.0058 0.0669 0.0149 -0.0372 0.0083 Fair Grad 0.8144 0.0021 Eopp 0.004 0.0037 0.006 0.0052 -0.0099 0.0097 Published in Transactions on Machine Learning Research (08/2023) Table 15: Results for the Dutch dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.7937 0.0052 AP 0.0252 0.0091 0.0252 0.009 -0.0252 0.0091 Constant 0.524 0.0 AP 0.151 0.0 0.152 0.0 0.15 0.0 Weighted ERM 0.7954 0.0023 AP 0.0257 0.0089 0.0257 0.0089 -0.0257 0.0089 Adversarial 0.7939 0.0043 AP 0.0232 0.0071 0.0232 0.0071 -0.0232 0.007 Reduction 0.7421 0.0168 AP 0.0227 0.0141 0.0227 0.0142 -0.0227 0.0141 Fair Grad 0.8043 0.0071 AP 0.0052 0.0026 0.0052 0.0026 -0.0052 0.0026 Unconstrained 0.7914 0.006 Eodds 0.0162 0.0062 0.0193 0.0071 -0.0263 0.0142 Constant 0.522 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7958 0.0027 Eodds 0.0168 0.0053 0.0202 0.0048 -0.0261 0.0131 Adversarial 0.7928 0.0077 Eodds 0.0148 0.0041 0.0202 0.0066 -0.0211 0.006 Bi Fair 0.819 0.003 Eodds 0.021 0.004 0.03 0.005 -0.028 0.007 Fair Batch 0.8091 0.012 Eodds 0.018 0.0021 0.0254 0.0058 -0.0248 0.0062 Reduction 0.7144 0.0176 Eodds 0.0253 0.0073 0.0347 0.0123 -0.0323 0.0064 Fair Grad 0.8013 0.0073 Eodds 0.0069 0.0031 0.0099 0.0038 -0.0095 0.0068 Unconstrained 0.8149 0.0034 Eopp 0.0055 0.0024 0.0079 0.0035 -0.014 0.0061 Constant 0.524 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8179 0.0044 Eopp 0.0066 0.0026 0.0095 0.0037 -0.017 0.0065 Adversarial 0.8156 0.0038 Eopp 0.004 0.0039 0.0058 0.0057 -0.0102 0.01 Bi Fair 0.819 0.003 Eopp 0.009 0.002 0.012 0.003 -0.022 0.006 Fair Batch 0.8174 0.0031 Eopp 0.002 0.0012 0.0029 0.0017 -0.0052 0.0031 Reduction 0.7571 0.0061 Eopp 0.0219 0.0021 0.0563 0.0054 -0.0313 0.0028 Fair Grad 0.8158 0.0051 Eopp 0.0036 0.0031 0.0051 0.0045 -0.0092 0.0079 Published in Transactions on Machine Learning Research (08/2023) 0.59 0.63 0.67 0.71 0.75 0.79 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.59 0.63 0.67 0.71 0.75 0.79 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.59 0.63 0.67 0.71 0.75 0.79 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.59 0.63 0.67 0.71 0.75 0.79 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.59 0.63 0.67 0.71 0.75 0.79 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.59 0.63 0.67 0.71 0.75 0.79 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 12: Results for the German dataset with different fairness measures. Table 16: Results for the German dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.692 0.0232 AP 0.0226 0.0181 0.0169 0.0111 -0.0284 0.0256 Constant 0.73 0.0 AP 0.05 0.0 0.069 0.0 0.031 0.0 Weighted ERM 0.707 0.0344 AP 0.0243 0.0191 0.0186 0.0113 -0.0299 0.027 Constrained 0.733 0.033 AP 0.024 0.025 0.032 0.033 0.015 0.017 Reduction 0.631 0.0396 AP 0.0323 0.0139 0.0286 0.0202 -0.036 0.0185 Fair Grad 0.744 0.0357 AP 0.0274 0.0212 0.0215 0.0123 -0.0334 0.0306 Unconstrained 0.69 0.0266 Eodds 0.0316 0.0207 0.0499 0.0341 -0.0618 0.0471 Constant 0.7 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.709 0.0296 Eodds 0.0324 0.0338 0.0461 0.046 -0.055 0.0626 Constrained 0.739 0.027 Eodds 0.037 0.012 0.072 0.025 0.01 0.004 Bi Fair 0.698 0.039 Eodds 0.033 0.01 0.052 0.023 -0.059 0.029 Fair Batch 0.7 0.0247 Eodds 0.0706 0.0184 0.1102 0.0489 -0.1134 0.0518 Reduction 0.707 0.0335 Eodds 0.0361 0.0175 0.0716 0.056 -0.0576 0.0266 Fair Grad 0.734 0.0358 Eodds 0.0464 0.0201 0.0784 0.0232 -0.0721 0.0496 Unconstrained 0.704 0.0193 Eopp 0.0053 0.0035 0.0096 0.004 -0.0116 0.0117 Constant 0.7 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.706 0.0328 Eopp 0.0048 0.0039 0.0097 0.0091 -0.0096 0.0092 Constrained 0.741 0.019 Eopp 0.005 0.002 0.015 0.006 0.0 0.0 Bi Fair 0.703 0.037 Eopp 0.007 0.006 0.014 0.015 -0.013 0.015 Fair Batch 0.718 0.0229 Eopp 0.0172 0.0124 0.0272 0.0187 -0.0416 0.0396 Reduction 0.717 0.0441 Eopp 0.0183 0.014 0.036 0.0254 -0.0372 0.0407 Fair Grad 0.723 0.0425 Eopp 0.0125 0.0043 0.0212 0.0087 -0.0288 0.0162 Published in Transactions on Machine Learning Research (08/2023) Table 17: Results for the German dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.695 0.0122 AP 0.0426 0.0241 0.0314 0.0144 -0.0537 0.0345 Constant 0.73 0.0 AP 0.05 0.0 0.069 0.0 0.031 0.0 Weighted ERM 0.703 0.0183 AP 0.035 0.0237 0.0265 0.0138 -0.0436 0.0338 Adversarial 0.681 0.0156 AP 0.041 0.0254 0.0327 0.0165 -0.0492 0.0368 Reduction 0.666 0.0198 AP 0.0173 0.0171 0.0131 0.0115 -0.0215 0.0231 Fair Grad 0.714 0.026 AP 0.037 0.0222 0.0291 0.0119 -0.0448 0.0331 Unconstrained 0.689 0.0213 Eodds 0.0089 0.0052 0.0117 0.0045 -0.0144 0.0116 Constant 0.7 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.703 0.034 Eodds 0.0211 0.0106 0.0305 0.0186 -0.0372 0.0158 Adversarial 0.684 0.0097 Eodds 0.0184 0.0122 0.0263 0.0201 -0.0339 0.0237 Bi Fair 0.725 0.031 Eodds 0.016 0.015 0.021 0.018 -0.027 0.018 Fair Batch 0.692 0.026 Eodds 0.0489 0.0382 0.0607 0.0446 -0.0882 0.0983 Reduction 0.706 0.0272 Eodds 0.0489 0.0217 0.0742 0.0266 -0.0717 0.051 Fair Grad 0.695 0.0237 Eodds 0.0095 0.004 0.0121 0.0046 -0.0175 0.0076 Unconstrained 0.686 0.0215 Eopp 0.0124 0.0075 0.0227 0.0128 -0.0269 0.0227 Constant 0.7 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7 0.0261 Eopp 0.0066 0.0057 0.0131 0.0071 -0.0133 0.0173 Adversarial 0.687 0.0129 Eopp 0.0085 0.0051 0.0203 0.0147 -0.0137 0.0099 Bi Fair 0.727 0.023 Eopp 0.015 0.013 0.023 0.019 -0.036 0.038 Fair Batch 0.697 0.025 Eopp 0.0084 0.0079 0.0235 0.0226 -0.0102 0.0094 Reduction 0.701 0.0397 Eopp 0.0102 0.008 0.0242 0.024 -0.0167 0.0134 Fair Grad 0.696 0.0166 Eopp 0.0052 0.0038 0.0093 0.0064 -0.0115 0.0108 Published in Transactions on Machine Learning Research (08/2023) 0.44 0.54 0.64 0.74 0.84 0.94 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.44 0.54 0.64 0.74 0.84 0.94 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.44 0.54 0.64 0.74 0.84 0.94 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.44 0.54 0.64 0.74 0.84 0.94 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.44 0.54 0.64 0.74 0.84 0.94 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.44 0.54 0.64 0.74 0.84 0.94 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 13: Results for the Gaussian dataset with different fairness measures. Table 18: Results for the Gaussian dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8689 0.0037 AP 0.0966 0.0029 0.0957 0.0028 -0.0974 0.0036 Constant 0.497 0.0 AP 0.001 0.0 0.001 0.0 0.001 0.0 Weighted ERM 0.869 0.0039 AP 0.0966 0.0026 0.0957 0.0023 -0.0974 0.0034 Constrained 0.799 0.004 AP 0.003 0.002 0.003 0.002 0.003 0.002 Reduction 0.7891 0.0266 AP 0.0575 0.0114 0.057 0.0118 -0.0579 0.0111 Fair Grad 0.8516 0.0064 AP 0.0558 0.0094 0.0553 0.0093 -0.0562 0.0096 Unconstrained 0.869 0.0037 Eodds 0.0971 0.0026 0.1872 0.0067 -0.1896 0.0056 Constant 0.499 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.869 0.0039 Eodds 0.0971 0.0023 0.1869 0.0063 -0.1894 0.0051 Constrained 0.497 0.003 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Bi Fair 0.873 0.004 Eodds 0.113 0.004 0.21 0.007 -0.213 0.004 Fair Batch 0.8649 0.0025 Eodds 0.0902 0.0035 0.1717 0.0046 -0.1719 0.0079 Reduction 0.6241 0.054 Eodds 0.0632 0.0164 0.0732 0.0198 -0.074 0.0226 Fair Grad 0.8459 0.01 Eodds 0.0786 0.0051 0.1504 0.0102 -0.1527 0.0142 Unconstrained 0.8598 0.0121 Eopp 0.0928 0.0012 0.1845 0.0041 -0.1869 0.0041 Constant 0.498 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8599 0.0121 Eopp 0.0931 0.0011 0.1849 0.004 -0.1874 0.004 Constrained 0.698 0.005 Eopp 0.004 0.002 0.008 0.005 0.0 0.0 Bi Fair 0.863 0.009 Eopp 0.1 0.003 0.2 0.007 -0.202 0.006 Fair Batch 0.8635 0.0024 Eopp 0.085 0.0023 0.17 0.0032 -0.1702 0.0065 Reduction 0.6251 0.0355 Eopp 0.0189 0.0138 0.0379 0.0271 -0.0378 0.0282 Fair Grad 0.8431 0.0065 Eopp 0.0752 0.0043 0.1494 0.0087 -0.1514 0.0094 Published in Transactions on Machine Learning Research (08/2023) Table 19: Results for the Gaussian dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.88 0.0038 AP 0.0897 0.0045 0.0888 0.0035 -0.0905 0.0055 Constant 0.497 0.0 AP 0.001 0.0 0.001 0.0 0.001 0.0 Weighted ERM 0.8809 0.0048 AP 0.0903 0.0045 0.0894 0.0033 -0.0911 0.0057 Adversarial 0.8725 0.0115 AP 0.0858 0.0077 0.0851 0.0076 -0.0866 0.0081 Reduction 0.718 0.0251 AP 0.0694 0.0237 0.0699 0.0236 -0.0689 0.0239 Fair Grad 0.8542 0.0047 AP 0.0352 0.0047 0.0349 0.0048 -0.0355 0.0046 Unconstrained 0.8814 0.0024 Eodds 0.093 0.0032 0.1807 0.0066 -0.183 0.005 Constant 0.499 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8821 0.0031 Eodds 0.0939 0.0013 0.1826 0.0042 -0.185 0.0033 Adversarial 0.8775 0.0091 Eodds 0.0852 0.007 0.1643 0.0125 -0.1666 0.0146 Bi Fair 0.868 0.013 Eodds 0.092 0.011 0.167 0.035 -0.168 0.031 Fair Batch 0.8735 0.0032 Eodds 0.0749 0.0041 0.1455 0.0059 -0.1456 0.0056 Reduction 0.7309 0.0189 Eodds 0.0262 0.0141 0.0438 0.0257 -0.0435 0.0265 Fair Grad 0.8539 0.0056 Eodds 0.0596 0.0068 0.1013 0.0147 -0.1025 0.0144 Unconstrained 0.8801 0.004 Eopp 0.0902 0.0017 0.1792 0.0041 -0.1816 0.0053 Constant 0.498 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.8805 0.0046 Eopp 0.0912 0.0008 0.1812 0.0024 -0.1837 0.0045 Adversarial 0.8754 0.0086 Eopp 0.0808 0.0066 0.1605 0.0128 -0.1628 0.0143 Bi Fair 0.88 0.003 Eopp 0.086 0.005 0.17 0.013 -0.172 0.009 Fair Batch 0.874 0.0035 Eopp 0.0733 0.0029 0.1465 0.0054 -0.1467 0.0066 Reduction 0.6868 0.0234 Eopp 0.0505 0.0179 0.1015 0.0359 -0.1005 0.036 Fair Grad 0.8543 0.0082 Eopp 0.0517 0.0095 0.1028 0.0191 -0.1041 0.0192 Published in Transactions on Machine Learning Research (08/2023) 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.57 0.61 0.65 0.69 0.73 0.77 Accuracy Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 14: Results for the Twitter Sentiment dataset with different fairness measures. Table 20: Results for the Twitter Sentiment dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.7211 0.004 AP 0.0426 0.0011 0.0426 0.0011 -0.0426 0.0011 Constant 0.5 0.0 AP 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7212 0.0044 AP 0.0426 0.0011 0.0426 0.0011 -0.0426 0.0011 Constrained 0.72 0.002 AP 0.04 0.003 0.04 0.003 0.04 0.003 Reduction 0.6008 0.022 AP 0.0159 0.0092 0.0159 0.0092 -0.0159 0.0092 Fair Grad 0.7219 0.0027 AP 0.0462 0.0021 0.0462 0.0021 -0.0462 0.0021 Unconstrained 0.7237 0.0054 Eodds 0.1867 0.0052 0.2287 0.0078 -0.2288 0.0078 Constant 0.5 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7234 0.0054 Eodds 0.188 0.0033 0.2314 0.0056 -0.2315 0.0056 Constrained 0.72 0.004 Eodds 0.012 0.002 0.019 0.005 0.006 0.005 Bi Fair 0.736 0.009 Eodds 0.041 0.012 0.056 0.022 -0.056 0.022 Fair Batch 0.7413 0.0014 Eodds 0.1391 0.0043 0.1755 0.0084 -0.1756 0.0084 Reduction 0.5962 0.0113 Eodds 0.0213 0.0108 0.0314 0.0211 -0.0314 0.021 Fair Grad 0.7193 0.0062 Eodds 0.0154 0.0051 0.0204 0.0098 -0.0204 0.0098 Unconstrained 0.7244 0.0051 Eopp 0.0719 0.0012 0.1439 0.0023 -0.1438 0.0023 Constant 0.5 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.72 0.0054 Eopp 0.0718 0.0013 0.1437 0.0026 -0.1436 0.0026 Constrained 0.752 0.004 Eopp 0.002 0.001 0.005 0.001 0.0 0.0 Bi Fair 0.746 0.009 Eopp 0.009 0.004 0.017 0.009 -0.017 0.009 Fair Batch 0.7426 0.001 Eopp 0.0429 0.0005 0.0858 0.0011 -0.0858 0.0011 Reduction 0.6381 0.0039 Eopp 0.0712 0.0117 0.1424 0.0234 -0.1425 0.0234 Fair Grad 0.7518 0.0069 Eopp 0.0024 0.002 0.0049 0.004 -0.0049 0.004 Published in Transactions on Machine Learning Research (08/2023) Table 21: Results for the Twitter Sentiment dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.715 0.0043 AP 0.0392 0.0055 0.0392 0.0055 -0.0392 0.0055 Constant 0.5 0.0 AP 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7183 0.0042 AP 0.0427 0.0019 0.0427 0.0019 -0.0427 0.0019 Adversarial 0.7385 0.0075 AP 0.0367 0.0027 0.0367 0.0027 -0.0368 0.0027 Reduction 0.6555 0.0162 AP 0.0101 0.0038 0.0101 0.0038 -0.0101 0.0038 Fair Grad 0.7154 0.0047 AP 0.0368 0.0079 0.0367 0.0078 -0.0368 0.0079 Unconstrained 0.7167 0.0126 Eodds 0.1854 0.0061 0.2349 0.0091 -0.235 0.0091 Constant 0.5 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.718 0.0137 Eodds 0.1882 0.0062 0.2379 0.0073 -0.2381 0.0073 Adversarial 0.7393 0.0024 Eodds 0.0382 0.0056 0.06 0.0151 -0.06 0.0151 Bi Fair 0.74 0.01 Eodds 0.039 0.016 0.058 0.017 -0.058 0.017 Fair Batch 0.7318 0.004 Eodds 0.1313 0.0057 0.1724 0.0055 -0.1725 0.0055 Reduction 0.6653 0.0134 Eodds 0.0133 0.0097 0.0199 0.0172 -0.0199 0.0173 Fair Grad 0.717 0.0082 Eodds 0.0109 0.0027 0.0165 0.0053 -0.0165 0.0053 Unconstrained 0.7147 0.0118 Eopp 0.0653 0.0062 0.1306 0.0124 -0.1306 0.0124 Constant 0.5 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7074 0.0158 Eopp 0.0672 0.0062 0.1346 0.0125 -0.1345 0.0125 Adversarial 0.7471 0.0042 Eopp 0.005 0.0035 0.0099 0.007 -0.0099 0.007 Bi Fair 0.747 0.009 Eopp 0.007 0.005 0.013 0.01 -0.013 0.01 Fair Batch 0.7359 0.0011 Eopp 0.0368 0.0012 0.0736 0.0025 -0.0736 0.0025 Reduction 0.681 0.0078 Eopp 0.0436 0.0071 0.0871 0.0143 -0.0871 0.0143 Fair Grad 0.7401 0.0059 Eopp 0.0049 0.0041 0.0099 0.0083 -0.0099 0.0083 Published in Transactions on Machine Learning Research (08/2023) 0.35 0.47 0.58 0.70 0.81 0.93 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Accuracy Parity Fair Grad Unconstrained Weighted Erm Reduction Constraints (a) Linear - AP 0.35 0.47 0.58 0.70 0.81 0.93 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (b) Linear - EOdds 0.35 0.47 0.58 0.70 0.81 0.93 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Equal Opportunity Fair Grad Fair Batch Unconstrained Weighted Erm Reduction Constraints Bi Fair (c) Linear - EOpp 0.35 0.47 0.58 0.70 0.81 0.93 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Accuracy Parity Fair Grad Unconstrained Adversarial Weighted Erm Reduction (d) Non Linear - AP 0.35 0.47 0.58 0.70 0.81 0.93 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (e) Non Linear - EOdds 0.35 0.47 0.58 0.70 0.81 0.93 Accuracy 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.10 0.11 0.12 0.13 Equal Opportunity Fair Grad Fair Batch Unconstrained Adversarial Weighted Erm Reduction Bi Fair (f) Non Linear - EOpp Figure 15: Results for the Folktables Adult dataset with different fairness measures. Table 22: Results for the Folktables Adult dataset with Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (L) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.7905 0.0033 AP 0.0131 0.0021 0.0123 0.0021 -0.0138 0.0022 Constant 0.666 0.0 AP 0.053 0.0 0.056 0.0 0.051 0.0 Weighted ERM 0.7906 0.0032 AP 0.0127 0.0023 0.0119 0.0022 -0.0134 0.0024 Constrained 0.467 0.115 AP 0.036 0.003 0.039 0.003 0.034 0.003 Reduction 0.733 0.0106 AP 0.0653 0.0114 0.0614 0.011 -0.0692 0.0118 Fair Grad 0.7837 0.0049 AP 0.0023 0.0009 0.0023 0.001 -0.0022 0.0008 Unconstrained 0.789 0.0026 Eodds 0.0301 0.011 0.0377 0.0153 -0.0458 0.0184 Constant 0.667 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7886 0.0032 Eodds 0.0294 0.012 0.0364 0.0169 -0.0443 0.0206 Constrained 0.663 0.032 Eodds 0.008 0.003 0.013 0.004 0.004 0.002 Bi Fair 0.768 0.007 Eodds 0.008 0.005 0.011 0.006 -0.011 0.008 Fair Batch 0.788 0.0027 Eodds 0.0045 0.0033 0.0069 0.0065 -0.0063 0.0049 Reduction 0.6922 0.0346 Eodds 0.077 0.0322 0.0761 0.0257 -0.0903 0.0378 Fair Grad 0.7885 0.0027 Eodds 0.0043 0.0019 0.0073 0.0037 -0.0068 0.0045 Unconstrained 0.7902 0.0038 Eopp 0.0094 0.0031 0.0162 0.0053 -0.0215 0.0071 Constant 0.667 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7893 0.0031 Eopp 0.009 0.003 0.0155 0.0051 -0.0206 0.0069 Constrained 0.706 0.002 Eopp 0.004 0.0 0.01 0.001 0.0 0.0 Bi Fair 0.77 0.002 Eopp 0.019 0.01 0.033 0.017 -0.044 0.023 Fair Batch 0.79 0.0031 Eopp 0.0012 0.0015 0.0022 0.0026 -0.0026 0.0034 Reduction 0.7388 0.0144 Eopp 0.0409 0.0111 0.0932 0.025 -0.0704 0.0194 Fair Grad 0.7893 0.0026 Eopp 0.0011 0.0009 0.0024 0.002 -0.0021 0.0016 Published in Transactions on Machine Learning Research (08/2023) Table 23: Results for the Folktables Adult dataset with Non Linear Models. All the results are averaged over 5 runs. Here MEAN ABS., MAXIMUM, and MINIMUM represent the mean absolute fairness value, the fairness level of the most well-offgroup, and the fairness level of the worst-offgroup, respectively. METHOD (NL) ACCURACY FAIRNESS MEASURE MEAN ABS. MAXIMUM MINIMUM Unconstrained 0.8037 0.0037 AP 0.0131 0.0017 0.0123 0.0016 -0.0139 0.0017 Constant 0.666 0.0 AP 0.053 0.0 0.056 0.0 0.051 0.0 Weighted ERM 0.8046 0.0049 AP 0.0131 0.0014 0.0123 0.0014 -0.0138 0.0015 Adversarial 0.8016 0.0053 AP 0.0122 0.0016 0.0115 0.0015 -0.0129 0.0016 Reduction 0.7293 0.0133 AP 0.0991 0.0149 0.0932 0.0139 -0.1051 0.016 Fair Grad 0.7917 0.0025 AP 0.0016 0.0011 0.0016 0.0011 -0.0016 0.001 Unconstrained 0.7947 0.0078 Eodds 0.0314 0.0059 0.0373 0.0058 -0.0454 0.0066 Constant 0.667 0.0 Eodds 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7902 0.0049 Eodds 0.0327 0.0061 0.04 0.0067 -0.0488 0.0077 Adversarial 0.806 0.0047 Eodds 0.0035 0.0018 0.0051 0.0021 -0.0053 0.0028 Bi Fair 0.793 0.006 Eodds 0.006 0.003 0.007 0.003 -0.007 0.004 Fair Batch 0.8061 0.0044 Eodds 0.0051 0.0015 0.0087 0.0048 -0.0084 0.0029 Reduction 0.7416 0.01 Eodds 0.0933 0.022 0.1517 0.0311 -0.1244 0.026 Fair Grad 0.7997 0.0087 Eodds 0.0045 0.0029 0.0067 0.0045 -0.0071 0.0058 Unconstrained 0.7902 0.0044 Eopp 0.0097 0.0026 0.0168 0.0045 -0.0222 0.006 Constant 0.667 0.0 Eopp 0.0 0.0 0.0 0.0 0.0 0.0 Weighted ERM 0.7947 0.0022 Eopp 0.0105 0.0027 0.0181 0.0047 -0.024 0.0062 Adversarial 0.8108 0.0161 Eopp 0.0034 0.0057 0.0041 0.0057 -0.0095 0.017 Bi Fair 0.793 0.008 Eopp 0.028 0.017 0.048 0.029 -0.064 0.039 Fair Batch 0.8038 0.0063 Eopp 0.0008 0.0005 0.0014 0.0009 -0.0018 0.0012 Reduction 0.7334 0.0155 Eopp 0.0573 0.0116 0.1307 0.0265 -0.0986 0.0199 Fair Grad 0.8058 0.0035 Eopp 0.0014 0.0014 0.003 0.0031 -0.0026 0.0024