# fair_ranking_with_noisy_protected_attributes__ffe9b077.pdf Fair Ranking with Noisy Protected Attributes Anay Mehrotra Yale University Nisheeth K. Vishnoi Yale University The fair-ranking problem, which asks to rank a given set of items to maximize utility subject to group fairness constraints, has received attention in the fairness, information retrieval, and machine learning literature. Recent works, however, observe that errors in socially-salient (including protected) attributes of items can significantly undermine fairness guarantees of existing fair-ranking algorithms and raise the problem of mitigating the effect of such errors. We study the fairranking problem under a model where socially-salient attributes of items are randomly and independently perturbed. We present a fair-ranking framework that incorporates group fairness requirements along with probabilistic information about perturbations in socially-salient attributes. We provide provable guarantees on the fairness and utility attainable by our framework and show that it is informationtheoretically impossible to significantly beat these guarantees. Our framework works for multiple non-disjoint attributes and a general class of fairness constraints that includes proportional and equal representation. Empirically, we observe that, compared to baselines, our algorithm outputs rankings with higher fairness, and has a similar or better fairness-utility trade-off compared to baselines. 1 Introduction Given a query and a set of m items, ranking problems require one to output an ordering of a small subset of items in decreasing order of relevance to the query. Such ranking problems have been extensively studied in the information retrieval [46] and the machine learning [45] literature, and algorithms for them are used in applications such as search engines, personalized feed generators, and online recruiting platforms [44, 12, 8]. Several studies have observed that when the outputs of ranking algorithms are consumed by end-users, e.g., image results for occupation-related queries, articles with different political leanings, and job applicants in online recruiting, the outputs can mislead or alter their perceptions about socially-salient groups [38], polarize their opinions [24, 50], and affect economic opportunities available to individuals [32]. A reason is that relevance (or utilities) input to ranking algorithms may be influenced by human or societal biases, leading to output rankings that skew representations of socially-salient, and often legally-protected, groups such as women and Black people [55]. A growing number of works aim to make the output of ranking algorithms fair with respect to sociallysalient attributes [74, 75, 58]. As for notions of fairness, in the case when each item belongs to one of two socially-salient groups (G1 or G2), equal representation requires that, for every k, (roughly) k 2 items from each of G1 and G2 appear in the first k positions of the output ranking. Proportional representation requires that at most k |Gℓ| m items from each Gℓappear in the first k positions. Fairness criteria that generalize proportional representation and involve p 2 groups G1, . . . , Gp, where each item may belong to multiple groups, have also been considered: Given values Ukℓ, they require that at most Ukℓitems from Gℓappear in the first k positions of the output ranking [61, 18]. One set of works in the fair-ranking literature tries to improve fairness in utility-estimation [72, 62, 73, 51]. Such approaches have the benefit that no changes to the existing ranking algorithm are necessary but they may be unable to guarantee that the output ranking satisfies the required fairness criteria [27]. Another set of works use the given utilities as-it-is and change the ranking algorithm to output 36th Conference on Neural Information Processing Systems (Neur IPS 2022). the ranking with the highest utility subject to satisfying the specified fairness criteria by including them as fairness constraints [61, 9, 18, 27, 30]. While these latter approaches can guarantee fairness, they require coming up with new algorithms to solve the arising constrained ranking problems. Both approaches, however, rely on knowledge of the socially-salient attributes of the items [56]. Assuming precise access to socially-salient attributes is reasonable in some contexts and has led to successful deployment of fair-ranking frameworks; see [27]. However, in several contexts, sociallysalient attributes can be erroneous, missing, or known only probabilistically. For instance, errors can arise due to misreporting, which is a common concern with self-reported attributes [4]. Attributes can also be missing, as is the case with images in web-search or in settings where it is illegal to collect certain socially-salient attributes [20]. Often attributes are predicted using ML-classifiers, but such prediction has inaccuracies [10]. In such cases, one can calibrate the confidence scores of classifiers to derive (aggregate) probabilistic information about the true attributes [35]. Moreover, probabilistic information about socially-salient (protected) attributes can be sometimes computed from other attributes. For instance, name and location of an individual, combined with aggregate census data may be used to get a conditional distribution of their race [23, 36, 20]. Even accurate attributes may be randomly and independently flipped to preserve user privacy, and the distribution of flipped attributes is determined by public parameters of, e.g., the randomized response mechanism [37, 69]. Several models of inaccuracies in data have been proposed [47, 26]. We consider one such model (due to [5]) to capture inaccuracies in socially-salient attributes. Each item i belongs to the ℓ-th group with a known probability Piℓ. For each item i, the distribution corresponding to Piℓs over groups is assumed to be independent of corresponding distributions of other items. This model can be used in cases where these probabilities are available or can be derived, as in some of the aforementioned examples (see Section 5 and Supplementary Material A). In other cases, e.g., when errors are strategic or adversarial, other models are needed. This model and its variants have also been used by works on designing fair algorithms in the presence of inaccuracies, for problems including classification [42, 67, 66, 14], subset selection [48], and clustering [25] (Supplementary Material B briefly discusses these works). In this noise model, while socially-salient attributes are not explicitly specified, one could still use existing fair-ranking algorithms by first sampling groups for items from the given probabilities. Indeed, [29] evaluate existing fair-ranking algorithms on attributes obtained from the probabilities derived from ML classifiers. They find that errors in [socially-salient attributes] can dramatically undermine fair-ranking algorithms and can cause [non-disadvantaged groups] to become disadvantaged after a fair re-ranking. We confirm this observation on a synthetic dataset when the goal is to finding a ranking that satisfies equal representation (Section 5.1). We assigned each item the sociallysalient group that is most likely and find that when existing fair-ranking algorithms (for equal representation) are run with this group information, they output rankings that significantly violate the equal representation criteria (Figure 1). Further, we mathematically analyze two natural methods to sample groups from probabilities and give examples where taking such information as input, existing fair-ranking algorithms output rankings which provably violate the equal representation criteria (Supplementary Material C). Thus, new ideas are needed to design fair-ranking frameworks that can guarantee given fairness criteria under this noise model. Our contributions. We present a fair-ranking framework that guarantees given fairness criteria when the socially-salient attributes are assumed to follow the probabilistic noise model mentioned above. In particular, it finds a utility maximizing ranking subject to a class of constraints that only rely on given probability distributions (Program (7)). These constraints relax the given fairness criteria by a carefully chosen factor: for equal representation, the relaxation is by roughly a 1 + 1 k multiplicative factor for position k for any k. Moreover, instead of sampling the attribute values and applying constraints on them, these constraints apply the relaxed-fairness criteria to the expected number of items from each group that appear in the first k positions. We show that these constraints ensure that any ranking approximately satisfying the given fairness criteria is feasible for them and any ranking feasible for them approximately satisfies the given fairness criteria (Theorem 4.1). Our fair-ranking framework works for the general class of fairness criteria introduced earlier, which involve multiple overlapping groups G1, . . . , Gp and upper bound Ukℓfor the ℓ-th group and k-th position (Theorem 4.1), and for their position-weighted versions (Theorem F.1). We show that our fair-ranking framework, besides nearly satisfying the given fairness criteria, has a provably high utility (Theorem 4.1). Complementing Theorem 4.1, we prove near-tightness of the fairness guarantee (Theorem 4.2): for equal representation fairness criteria, this results shows that it is information theoretically impossible to output a ranking that violates this criteria by less than a multiplicative factor of 1 + e O( 1 k) at the k-th position for any k. Finally, we give a polynomial-time algorithm to approximately solve Program (7) (Theorem 4.3). Empirically, we evaluate our framework on both synthetic and real-world data against standard metrics like weighted-risk difference (RD) that measure deviation from specific fairness criteria (Section 5). We compare its performance to key baselines [18, 61, 27, 48] on both single and multiple attributes. In all simulations, compared to baselines, our framework has a higher maximum fairness (2-10% for RD; Figures 1 to 3) and a similar/better fairness-utility trade-off (Figures 2, 8, 10 and 14 to 16). 2 Related work Work on automated information retrieval dates back to 1940s [43, 21]. Since then the IR literature has devoted a significant effort in measuring relevance of items to specific queries across different tasks: including, web search [7], personalization [34], and product rating [22]; we also refer the reader to [46] and the references therein. In the last three decades, works in the ML literature have also made significant contributions to relevance-estimation [45], by proposing methods that: (1) supplement traditional IR approaches, e.g., by automatically tuning their previously hard to tune parameters [65] and by improving their efficiency through clustering-based techniques [64, 3], and(2)substitute traditional IR approaches by neural-network based models to predict item relevance [12, 11, 68, 8]. Fair ranking. Existing works on the fair-ranking problem take diverse approaches: Among works that de-bias utilities, different approaches include, post-processing the utilities so that the post-processed utilities satisfy some fairness requirement [71], introducing a fairness penalty in the objective function used to train learning-to-rank models [62, 73, 49], and modifying feature representations generated by up-stream algorithms so that the utilities learned from the modified representations satisfy some fairness requirements [72]. Works that alter the ranking algorithms can also be further categorized into those which satisfy the constraints for each ranking [18, 70, 27, 30] and those that satisfy the constraints in aggregate over multiple rankings [61, 9]. Among aforementioned works, [49] uses a version of adversarial training to make (fair) learning-to-rank models robust to outliers but, unlike this work, they require socially-salient attributes of items to be accurately known to specify the fairness penalty. All of the other aforementioned works also need access to the socially-salient attributes of items. When protected attributes are inaccurate, these works can fail to satisfy their fairness and/or utility guarantees [29]. Effect of inaccuracies on fair-ranking algorithms. Some recent works have considered assessing fairness of rankings and ranking algorithms with missing or inaccurate protected attributes. [39] analyze the setting where all protected attributes are missing, but can be purchased at a fixed cost per item. They give statistical-techniques to estimate the fairness-value of a given ranking at a small cost. [29] use ML-classifiers to infer protected attributes from real-world data and study performance of the fair-ranking algorithm by [28] when given inferred attributes as input. While these works underscore the need for fair-ranking algorithms to be robust to inaccuracies in protected attributes, they only assess fairness in the presence of noisy protected attributes. 3 Model of fair ranking with noisy attributes Ranking problem. In ranking problems, given m items, one has to select a subset of n items and output a permutation of the selected items. This permutation is said to be a ranking. There is a large body of work on estimating the relevance of items and personalizing these estimates to specific users/queries [46, 45]. We consider a ranking problem where the relevance of items are known. Abstracting relevance estimation, in this problem, one is given an m n matrix W, such that placing the i-th item at the j-th position generates utility Wij. The utility of a ranking is the sum of utilities generated by each item in its assigned position. The algorithmic task in the ranking problem is to output a ranking with the highest utility. We denote rankings by assignment matrices R {0, 1}m n, where Rij = 1 indicates that item i appears in position j, and Rij = 0 indicates otherwise. In this notation, the utility of a ranking is R, W := Pm i=1 Pn j=1 Rij Wij. Then this ranking problem is to solve: max R R R, W . Where R is the set of all assignment matrices denoting a ranking: R := n X {0, 1}m n : i [m], Xn j=1 Xij 1, j [n], Xm i=1 Xij = 1 o . (1) Here, the constraint Pm i=1 Xij = 1 ensures position j has exactly one item and the constraint Pn j=1 Xij 1 ensures that item i occupies at most one position. While this model captures a variety of applications, in some cases, the entries of W may be skewed by an unknown amount [40, 16] or not known accurately [63] and the utility of the ranking may not be linear in the entries of W [2]. These are interesting directions but are not studied in this work. Fair-ranking problem. There are several versions of the fair-ranking problem. We consider a version with p 2 socially-salient groups G1, G2, . . . , Gp [m] (e.g., the group of all women or all Black people) which are often protected by law. Each of the m items belongs to one or more of these sociallysalient groups (henceforth referred to as just groups). This fair-ranking problem is to output the ranking with maximum utility subject to satisfying certain fairness criteria with respect to these groups. The appropriate notion of fairness is context dependent, and to capture different fairness criteria numerous fairness constraints have been proposed. We consider a class of general fairness constraints. Definition 3.1 (Fairness constraints). Given a matrix U Zn p + , a ranking R satisfies the upper bound constraint if P i Gℓ Pk j=1 Rij Ukℓ, for all ℓ [p] and k [n]. Existing works consider similar constraints and show that they can encapsulate a variety of fairness criteria [61]. For instance, when groups are disjoint, to capture equal and proportional representation, one can choose Ukℓ:= l k p m and Ukℓ:= l k |Gℓ| m m for all k and ℓrespectively. (That said, they do not capture qualitative differences among groups, such as, misrepresentation of demographics in image results [38, 55], which could arise even when rankings has sufficient individuals from each group.) As a running example, we consider the fair-ranking problem with equal representation with two disjoint groups, i.e., max R R R, W s.t. k [n] ℓ [2], X To ease readability, we omit ceilings-operators henceforth. Noise model. If the socially-salient attributes of items are known accurately, then one can solve the fair-ranking problem. However, as discussed, in many contexts, attributes are inaccurate, missing, or only probabilistically known. Several models have been proposed to capture different errors in attributes. Here, we consider a model (due to [5]) which has also appeared in [25, 42, 48]. Definition 3.2 (Noise model). Let P [0, 1]m p be a known matrix. The groups G1, . . . , Gp [m] are random variables, such that, for each i [m] and ℓ [p], Pr[Gℓ i] = Piℓ. Moreover, for different items i = j the events Gℓ i and Gk j are independent for all ℓ, k [p]. Definition 3.2 makes two key assumptions: the matrix P is known and for each item i, the events Gℓ i over groups ℓare independent of the corresponding events for other items. Both of these assumptions hold when attributes are flipped to preserve local differential privacy (Remark A.1). In other settings, P s estimate can be inaccurate and above events may be correlated. These can adversely affect the performance of our framework. We empirically study this in simulations where P is estimated using confidence scores of off-the-shelf classifiers and is miscalibrated (Figures 2 and 3). Supplementary Material A shows how Definition 3.2 captures both disjoint and overlapping groups. Fairness constraint with noisy attributes. Most existing fairness constraints assume that the groups are deterministic. Hence, it is not clear how to impose them when groups are random variables, as in Definition 3.2. One definition is to require the constraints to be approximately satisfied with high probability. Consider the instantiation of this definition for equal representation: A ranking R satisfies (ρ, δ)-equal representation, if with probability 1 δ, at most k 2(1 + ρ) items from Gℓappear in the first k positions in R places for all k [n] and ℓ [p]. Naturally, one would like to satisfy this definition for small δ, ρ. However, it turns out to be too stringent and is infeasible for any small δ, ρ. Proposition 3.3. No ranking satisfies (ρ, δ)-equal representation for ρ < 1, δ 1 2, and P = 1 The proof of Proposition 3.3 shows that any ranking R violates the equal-representation constraint at the 2nd position by a multiplicative factor of 2 with probability 1 2. The issue is that the same relaxation parameter ρ is used for each position k (whereas the information theoretically bestachievable relaxation parameter at k improves as k increases, this, e.g., follows by Theorem 4.1.) Motivated by this observation, we consider the following alternate version of upper bound constraints. Definition 3.4 ((ε, δ)-constraint). For any ε Rn 0 and δ (0, 1], a ranking R is said to satisfy (ε, δ)-constraint if with probability at least 1 δ over the draw of G1, . . . , Gp k [n] ℓ [p], X j=1 Rij Ukℓ(1 + εk). (3) We would like to output a ranking that satisfies Definition 3.4 for small δ and small ε1, ε2, . . . , εn. Problem 3.5 (Ranking problem with noisy attributes). Given matrices P, U, and W, find the ranking R maximizing utility R, W subject to satisfying (ε, δ)-constraint for some small ε and δ. 3.1 Challenges in solving Problem 3.5 In this section we discuss potential approaches for solving Problem 3.5. In other words, solving: max R R R, W , s.t., R satisfies (ε, δ)-constraint. (4) Even for two disjoint groups, given V , it is NP-hard to decide if the value of Program (4) is at least V (Theorem E.12). To bypass this hardness, one can consider approximation algorithms. Program (4) is an integer program (IP) because the entries of the matrix R are required to be integers (Equation (1)). A standard approach to (approximately) solve IPs is to: (1) consider their continuous relaxation that drops the integrality constraints, (2) compute the optimal solution Rc of the relaxed problem, and then (3) round Rc to satisfy integrality constraints while retaining its utility and fairness properties. To take this approach, we first need an efficient algorithm to find Rc. However, not just Program (4), but even its continuous relaxation is non-convex. Hence, it is unclear how to solve it to find Rc. Due to the independence assumption in Definition 3.2, the number of items from Gℓappearing in the first k positions of a ranking is concentrated around its expectation (for large k). This implies that if, in expectation, less that Ukℓitems from Gℓappear in the top k positions then, with high probability, the number of items from Gℓin the top k positions is not much larger than Ukℓ. Using this one can show that a ranking satisfying the following constraints k [n] ℓ [p], E X also satisfies (ε, δ)-constraint for small ε and δ. One idea is to find the ranking maximizing utility subject to satisfying Constraint (5). A feature of Constraint (5) is that it is linear in R as E[P i Gℓ Pk j=1 Rij] =Pm i=1 Pk j=1 PiℓRij and, hence, one may hope to find the ranking with the maximum utility subject to satisfying Constraint (5). However, the issue is that there are examples where any ranking satisfying Constraint (5) has 0 utility and there are rankings that satisfy (ε, δ)-constraint and have a large positive utility (Lemma E.10). Hence, this approach can output rankings whose utility is significantly smaller than the utility of the solution to Problem 3.5. To overcome this, we relax Constraint (5) by a carefully chosen position-dependent factor, such that, any ranking satisfying the (ε, δ)-constraint (for appropriate ε and δ) is also feasible for our framework. 4 Theoretical results In this section we present our optimization framework and its fairness and utility guarantees. Input:Matrices P [0, 1]m p , W Rm n 0, U Rn p Parameters: Constant c > 1, failure probability δ (0, 1], and k [n], relaxation parameter γk := 12 log 2np δ maxℓ [p] q Our Fair-Ranking Program max R R R, W , (Noise Resilient) (7) s.t. ℓ [p] k [n] P i [m], j [k] PiℓRij Ukℓ 1 + 1 1 2 c γk . (8) The above program modifies the program for fair ranking with accurate groups: It has the same objective but different constraints. Instead of sampling the attribute values and applying constraints on the sampled values, Constraint (8) applies upper bounds on the expected number of items in the first k positions from group ℓ(Section 3.1). Further, Constraint (8) relaxes upper bounds Ukℓby a small position-dependent factor. Like for Constraint (5), one can show that any ranking satisfying Constraint (8) also satisfies (ε, δ)-constraint (for small ε1, . . . , εn and δ). But unlike Constraint (5), and somewhat surprisingly, any rankingthatsatisfies(ε, δ)-constraint(for appropriateε1, . . . , εn and δ) must also satisfy Constraint (8). (In fact, γk is chosen to be the smallest, up to logarithmic factors, value such that this is true.) We use this to prove Theorem 4.1 s utility guarantee. Our first result bounds the fairness and utility of the optimal solution of Program (7). Theorem 4.1. Let γ Rn be as defined in Equation (6). There is an optimization program (Program (7)), parameterized by a constant c and failure probability δ, such that for any c > 1 and δ (0, 1 2] its optimal solution satisfies (cγ, δ)-constraint and has a utility at least as large as the utility of any ranking satisfying ((c c)γ, δ)-constraint. For equal representation, γk is e O 1 . Thus, Theorem 4.1 guarantees that, with high probability, the optimal solution of Program (7) multiplicatively violates equal representation at the k-th position by at most 1+ e O 1 . Further, this solution s utility is higher than the utility of any ranking satisfying a slight relaxation of this fairness guarantee. Theorem 4.1 can be extended to position-weighted versions of fairness constraints (Theorem F.1), where the fairness constraint is P j [k]vj Rij Ukℓ(for all k and ℓ) for specified discount factors v1 vn such as NDCG [33]. If we are also guaranteed Ukℓ ψk for some constant ψ > 0 and all k and ℓ, then we can improve γk s dependence on δ from δ (Supplementary Material E.3). The proof of Theorem 4.1 appears in Section 6. Lower bound on fairness guarantee. Our next result complements Theorem 4.1 s fairness guarantee. Theorem 4.2. There is a family of matrices U Zn p + such that for any U in the family and any parameters δ [0, 1) and ε1, . . . , εn 0, if for any position k [n], εk 1 and εk < 1 2Ukℓlog 1 4δ then there exists a matrix P [0, 1]m p, such that it is information theoreti- cally impossible to output a ranking that satisfies (ε, δ)-constraint. This family, in particular, contains the matrices U corresponding to equal representation and proportional representation constraints. Since γk is O log( np , Theorem 4.2 shows that Theorem 4.1 s fairness guarantee is optimal up to log-factors. Supplementary Material D.1 proves Theorem 4.2. An efficient algorithm. As for solving our optimization program, it is NP-hard to check its feasibility (Theorem E.7). However, because Constraint (8) is linear in R, the continuous relaxation of Program (7) is a standard linear program and can be solved efficiently. Our algorithm (Algorithm 1) solves the standard linear programming relaxation of Program (7) to find a solution Rc and then uses a dependent-rounding algorithm by [19] to convert Rc to a ranking. (See Supplementary Material D.2 for brief discussion of why straightforward rounding approaches are insufficient.) Theorem 4.3. There is a randomized algorithm (Algorithm 1) that given constants d > 2, a failure probability 0 < δ 1, and matrices P [0, 1]m p and W [0, 1]m n, outputs a ranking satisfying (O(dγ), δ)-constraint and with probability at least 1 δ, and has a utility at least 1 1 dn), where V is the utility of any ranking satisfying ((d d)γ, δ)-constraint. The algorithm runs in polynomial time in d and the bit complexity* of the input. The tension in setting d is that decreasing d improves the fairness guarantee and the utility guarantee s second term, but worsens the first term in the utility guarantee. Under the mild assumption that V =Ω(n), increasing d improves the utility guarantee because the first term in the utility guarantee dominates the second term. In this case, the utility guarantee improves to (1 1 d o(1)) V . Finally, while Theorem 4.3 requires utilities to be between 0 and 1, it can be extended to any non-negative and bounded utilities by scaling. The proof of Theorem 4.3 appears in Supplementary Material D.2. 5 Empirical results In this section we evaluate our framework s performance on synthetic and real-world data. Baselines and metrics. The correct choice of fairness metric is context-dependent and beyond the scope of this work [60]. To illustrate our results, we arbitrarily fix the fairness metric as weighted *The bit complexity of the inputs is the number if bits required to encode the input using the standard binary encoding (which, e.g., maps integers to their binary representation, rational numbers as pair of integers, and vectors/matrices as a tuple of their entries) [31, Section 1.3]. Code for our simulations is available at https://github.com/Anay Mehrotra/Fair Ranking With Noisy Attributes 1.0 1.2 1.4 1.6 1.8 2.0 (Looser constraint) Fairness const. ( ) (Stricter constrain Synthetic data (m,n,g)=(500,25,2),ITER=501,occ_lists=[NA,NA]. This work SJ CSV [Greedy] MC GAK [Det-Greedy] Uncons ......Weighted risk-difference...... ϕ (more fair) (less fair) Figure 1: Synthetic Data: Nonuniform Error Rate. We consider synthetic data where imputed socially-salient attributes have a higher false-discovery rate on the minority group. We vary the fairness constraint (ϕ) and observe the weighted risk-difference (RD) of algorithms. The y-axis plots RD and x-axis plots ϕ. (Note that the x-axis decreases toward the right). We observe that NResilient achieves the most fair RD, while obtaining a similar utility for all ϕ (Figure 8). Error-bars denote the error of the mean. risk-difference (RD). This is a position-weighted version of the standard risk-difference metric [13] and measures the extent to which a ranking violates equal representation. The RD of a ranking R is: k=5,10,... 1 log k maxℓ,q [p] i Gℓ,j [k], Rij X i Gq,j [k], Rij Where G denotes the ground-truth protected groups and Z is a constant so that RD has range [0, 1]. Here, RD = 1 is most fair and RD = 0 is least fair. We compare our framework, NResilient, against state-of-the-art fair-ranking algorithms: CSV ( greedy in [18]), SJ [61], and GAK ( Det Greedy in [27]). We also compare against MC, which ranks the items, in the subset output by [48] s algorithm, to maximize utility. Finally, we compare against the baseline, Uncons, which outputs the utility maximizing ranking without fairness considerations. We present additional discussion of baselines and results, additional plots for RD, and comparisons with weighted selection-lift (instead of RD) and different choices of U (than the ones below) in Supplementary Material G. Setup. We consider the DCG model of utilities [33] and a relaxation of equal representation constraints: (1) Given an intrinsic value wi 0, for each item i, we set Wij := wi (log (j + 1)) 1 j [n]. (2) Given a parameter ϕ [1, p], we set upper bounds Ukℓ:= ϕ p k for each k [n] and ℓ [p]. In simulations, we set m = 500, n = 25, and vary ϕ from p to 1. To gain some intuition about the relevant values of ϕ, note that satisfying the 80% rule requires ϕ 5p 5p 1, i.e., ϕ 1.11 for p = 2 and ϕ 1.05 for p = 4. For each ϕ, we draw m items uniformly without replacement and compute an estimate b P of the matrix P (from Definition 3.2) using, e.g., off-the-shelf ML classifiers or public APIs (see the paragraphs Estimating b P in simulation with image data and Setup in simulation on name data). We infer socially-salient groups b G1, . . . , b Gp via b P by assigning each item to its most-likely group. Finally, we run all algorithms using b P or b G1, . . . , b Gp as discussed next. Implementation details. NResilient and MC take probabilistic information about socially-salient attributes as input and are given b P. CSV, SJ, and GAK require access to socially-salient groups and are given b G1, . . . , b Gp. NResilient, SJ, and CSV use fairness constraints from Definition 3.1 and are given: for each k and ℓ, Ukℓ= ϕ p k. MC requires, for each ℓ [p], an upper bound on the number of items from Gℓthat can appear in top-n positions. It is given ϕ p n for each ℓ [p]. GAK requires the desired proportion αℓfor each group Gℓand, roughly, satisfies the constraint Ukℓ= αℓ k for each k [n] and ℓ [p]. It is given αℓ= 1 p for each ℓ [p], this corresponds to ϕ = 1 (hence, figures only plot the GAK at ϕ = 1). As a heuristic, we set γk = 1 20 maxℓ [p] q 1 Ukℓin all simulations. We find that this suffices and expect a more refined approach to improve the performance of NResilient. 5.1 Simulation on synthetic data We show that on synthetic data, where error-rates of given socially-salient attributes vary over groups, existing fair-ranking algorithms have worse RD than Uncons. Data. We generate w and P for two groups using code by [48] and fix b P = P. For all items i, wi is i.i.d. from the uniform distribution over [0, 1]. b P is constructed such that attributes inferred from b P have a higher false-discovery rate for the minority group compared to the majority (40% vs 10%). This 30% difference in false-discovery rates is comparable to the 34% difference in the false-discovery rates of dark-skinned females and light-skinned men observed by [10] for a commercial classifier. 0.55 0.60 0.65 0.70 0.75 0.80 This work SJ CSV [Greedy] MC GAK [Det-Greedy] Uncons ......Utility...... Weighted risk-difference (more fair) (less fair) Figure 2: Real-world image data. In this simulation, given non-gender labeled images and their utilities, our goal is to generate a high-utility gender-balanced ranking. We estimate P using an off-the-shelf ML-classifier and vary ϕ from p = 2 (less fair) to 1 (more fair). The y-axis plots the utility of algorithms and the x-axis plots RD. We observe that NResilient has the most fair RD and the best fairness-utility trade-off. Error bars show the error of the mean. Results. See Figure 1 for the observed RD averaged over 500 iterations. We observe that NResilient achieves best RD ( 0.81), while not losing significant utility ( 98% of max.; see Figure 8). MC achieves the next best RD ( 0.79). In contrast, CSV, SJ, and GAK, which do not account for noise in the socially-salient attributes, achieve a worse RD ( 0.68) for ϕ 1.2 than Uncons ( 0.75). This range of ϕ can be desired in practice: e.g., a platform must set ϕ 1.1 to guarantee the 80% rule is satisfied two groups. Thus, we observe that existing fair-ranking algorithms may achieve a worse RD than Uncons. 5.2 Simulation on real-world image data In this simulation, given non-gender labeled images-search results and their utilities, our goal is to generate a high-utility and gender-balanced ranking. Data. We use the Occupations dataset [15] which contains the top 100 Google Image results for 96 occupation-related queries. For each image, the data has its position in search results, gender (coded as male/female) of the individual depicted in the image, collected via MTurk. We use the (true) gender labels in the data to compute RD and to estimate b P, but do not provide them to algorithms. Setup. For each image i, with rank ri, we define wi := (log (1 + ri)) 1. We say an occupation is gender-stereotypical if more than 80% of images for this occupation have the same gender label (41/96 occupations). An image is said to be stereotypical if it is in a gender-stereotypical occupation and its gender label is the majority label for its occupation. We define the socially-salient groups as the sets of stereotypical and non-stereotypical images in gender-stereotypical occupations. Estimating b P. After pre-processing, we use a CNN-based gender-classifier f [59] to predict the (apparent) gender of the person depicted in each image. We calibrate the confidence scores output by f by binning and use these to estimate b P (see Supplementary Material G for more details). We perform this calibration once and on all occupations and, then, use it for gender-stereotypical occupations. Because of this b P is miscalibrated (and hence, inaccurate). For instance, among samples i for which 0.25 b P i,male 0.5, more than 75% are labeled as man (instead of some percentage between 25% and 50%). This violates the assumption that P is accurately known. Results. See Figure 2 for RD and utilities (NDCG) averaged over 1000 iterations. We observe that NResilient achieves the best RD ( 0.81) and has a better RD-utility trade-off than the other baselines. In contrast, CSV, SJ, and GAK, achieve a worse RD ( 0.77). MC achieves the worst RD ( 0.70) and a worst RD-utility trade-off. In particular, NResilient s RD-utility trade-off strictly dominates all baselines for RD 0.66. This value of RD can arise in practice: Figure 9 plots the RD vs ϕ for this simulation and shows that if ϕ 1.1 (as required to, e.g., guarantee satisfaction of the 80% rule), then all baselines have RD at least 0.66. We further evaluate the robustness of NResilient to varying levels of noise on the Occupations dataset in Supplementary Material G.3.2 and observe NResilient has a better or similar RD than each baseline at all noise levels. 5.3 Simulation on real-world name data We consider gender and race (encoded as binary) as socially-salient attributes. Our goal is to ensure equal representation across the four disjoint groups formed by combinations of these: non-White non-men, White non-men, non-White men, and White men. Data. We consider the chess ranking data [29] which has of 3,251 chess players. For each player, among other attributes, the data has their full-name, self-identified gender (coded as male/female), 1.0 1.5 2.0 2.5 3.0 3.5 4.0 (Looser constraint) Fairness const. ( ) (Stricter constrain ion compute_weighted_risk_diff at 0x7f13ab UPDATED Name data (DCG Utility) + Intersectional attributes (m,n,g)=(500,25,4),ITER=500. This work SJ CSV [Greedy] MC GAK [Det-Greedy] Uncons ......Weighted risk-difference...... ϕ (more fair) (less fair) Figure 3: Real-World Name Data: Multiple Attributes. In this simulation, the goal is to ensure equal representation across four disjoint groups formed by combinations of two attributes (non-White non-men, White non-men, non-White men, and White men). We estimate P by querying public APIs and libraries with names in the data. The y-axis plots RD and xaxis plots ϕ. (Note that the values decrease toward the right). We observe that all algorithms have a better RD than Uncons and NResilient has the best RD compared to all other baselines. Error bars represent the error of the mean. FIDE rating, and race (Asian, Black, Hispanic, White) collected via MTurk. We use the (true) gender and race labels in the data to evaluate RD, but do not provide them to algorithms. Setup. We partition the races into White (81.66%) and non-White (18.34%). For each player i, we query Genderize and Ethni Colr with i s full-name to obtain the probabilities pf(i) and pnw(i) that player i is labeled as a women and non-white respectively. We assume that these probabilities are correct and that the gender and race of players are drawn independently. Hence, e.g., we set the probability that i is a non-white women as b P i,nw+f=pnw(i)pf(i). Similarly, we set b P i,w+f=(1 pnw(i))pf(i), b P i,nw+m = pnw(i)(1 pf(i)), and b P i,w+m = (1 pnw(i))(1 pf(i)). Notably, we do not calibrate b P on this data. We verify that, like the previous simulation, b P is miscalibrated in this simulation. E.g., only 31% of the samples i for which b P i,nw+m > 0.75 are labeled as Non-white man (instead of 75%). Hence, the assumption that P is accurately known is violated in this simulation. We expect calibration to improve NResilient s performance. Results. See Figure 3 for RD averaged over 500 iterations. We observe that all algorithms (NResilient, CSV, GAK, SJ, and MC) have better RD than Uncons. Among these, NResilient achieves the best RD ( 0.67), next CSV, GAK, and SJ obtain RD ( 0.61), and MC achieves RD ( 0.53). More specifically, for all ϕ 1.75, NResilient has a strictly better RD than all baselines (this range of ϕ subsumes, e.g., the range ϕ 1.05 required guarantee satisfaction of the 80% rule with four groups.) Further, in Figure 10, we observe that all algorithms have a similar fairness-utility trade-off. 6 Proof of Theorem 4.1 In this section we prove Theorem 4.1. Some of the details are deferred to Supplementary Material E.3 due to space constraints. The proof is divided into two propositions: Proposition 6.1. For any δ (0, 1], any ranking feasible for Prog. (7) satisfies (cγ, δ)-constraint. Proposition 6.2. For any δ (0, 1 2) and c > 1, any ranking satisfying the ((c c)γ, δ)-constraint is feasible for Program (7). Proof of Theorem 4.1. Let R be the optimal solution of Program (7). Since R is feasible by definition, Proposition 6.1 implies that R satisfies the (cγ, δ)-constraint. Pick any R that satisfies the ((c c)γ, δ)-constraint. Proposition 6.2 implies that R is feasible for Program (7). Since R is an optimal solution of Program (7), R s utility is at least as large as the utility of R . Notation. For each item i and group ℓ, let Ziℓ {0, 1} be the indicator random variable Zi := I[Gℓ i]. By Definition 3.2, Pr[Ziℓ] = Piℓand Ziℓand Zjℓare independent for any i = j. Given ranking R R, group ℓ [p], and position k [n], let Z#(R, ℓ, k) be the number of items from Gℓ in the top k positions of R and let P#(R, ℓ, k) = E[Z#(R, ℓ, k)]. From the above, we get: P#(R, ℓ, k) = E [Z#(R, ℓ, k)] = P j [k] PiℓRij. We use the following concentration result (proved in Supplementary Material E.2) in the proof. Lemma 6.3. For any position k [n], group ℓ [p], parameters ε 0 and L, U R, and ranking R R, where R is possibly a random variable independent of {Ziℓ}i,ℓ, if P#(R, ℓ, k) U or gender-api.com and github.com/appeler/ethnicolr respectively P#(R, ℓ, k) L then the following equations hold respectively Pr [Z#(R, ℓ, k) < (1 + ε) U] 2+ε and Pr [Z#(R, ℓ, k) > (1 ε) L] 1 e Lε2 2(1 ε) . Proof of Proposition 6.1. Fix any k and ℓ. Let ϕ:=1 1 2 c, U :=Ukℓ(1 + ϕγk), and ζ := (1 ϕ)γk 1+ϕγk . (9) Here, U and ζ satisfy U (1 + ζ) = Ukℓ(1 + cγk). Fix any ranking R that is feasible for Program (7). Since R is feasible, it satisfies that ℓ [p], k [n], P#(R, ℓ, k) Uℓk (1 + ϕγk) . (10) Using U (1 + ζ) = Ukℓ(1 + cγk), Equation (10), and Lemma 6.3, we get that Pr [Z#(R, ℓ, k) U (1 + ζ)] e 2U ζ2 2+ζ (9)= e (1 ϕ)2c2γ2 k Ukℓ 2+(1+ϕ)cγk (ϕ 1) = e (1 ϕ)2c2γ2 k Ukℓ 2(1+cγk) . (11) Fact 6.4. For all x, y 0, if x y + y, then x2 1+x y. Using Fact 6.4 and Equation (6), we can show that for each k, c2γ2 k 1+cγk 2 (1 ϕ)2Ukℓ log 2np δ . (This uses δ < 1 2 and Ukℓ, n 1.) Substituting this in Equation (11) we get: Pr [Z#(R, ℓ, k) Uℓk(1 + cγk)] δ 2np. (12) Taking the union bound over all positions k and ℓ, we get (as desired) that with probability at least 1 δ, for all k [n] and ℓ , Z#(R, ℓ, k) Uℓk(1 + cγk). Proof of Proposition 6.2. Let ϕ := 1 1 2 c. Towards a contradiction, suppose that R satisfies ((c c)γ, δ)-constraint but is not feasible for Program (7). Then there exists ℓand k such that P#(R , k, ℓ) > Ukℓ (1 + ϕγk) . Fix any k and ℓsatisfying this. Let b:=1 1 c, L :=Ukℓ(1 + ϕγk) and ζ := (1+b)γk 1+ϕγk . (13) It holds that L (1 ζ) = Ukℓ(1 + bγk) and, hence, we get Pr [Z#(R , k, ℓ) L (1 ζ)] (13), Lem.6.3 e L ζ2 2(1 ζ) (13) = e (c b)2Ukℓγk Ukℓcγk 4(2 c 1) c(c>0) = e Since γk 8 log np 1 Ukℓ, δ < 1 2, and U 1, we have Pr [Z#(R , k, ℓ) Ukℓ] δ np < 1 δ. Since R satisfies ((c c)γ, δ)-constraint we have a contradiction, hence R must be feasible. 7 Limitations and conclusion Recent studies find that errors in socially-salient attributes can adversely affect fairness and utility of existing fair-ranking algorithms [29]. We consider a model of random and independent errors in socially-salient attributes and present a framework that can output rankings with high fairness and utility in this model. This framework works for a general class of fairness criteria, which involve multiple overlapping groups and upper bounds on the number of items that appear in the first k positions from each group. We also show near-tightness of the framework s fairness guarantee. Empirically, on both synthetic and real-world datasets, we observe that, compared to baselines, our framework can achieve higher fairness-values and a similar or better fairness-utility trade-off for standard metrics. Compared to existing fair-ranking frameworks, our framework does not need accurate socially-salient attributes, but assumes that errors in attributes are random and independent. When these assumptions do not hold, our framework may not satisfy its guarantees and a careful assessment of this on application-specific data would be important to avoid any (unintended) negative social impact. Our work only addresses one aspect of how bias may show up in rankings, and more generally, on the web. For instance, while we consider a large class of fairness constraints, it does not capture some important notions such as the qualitative representation of different groups [38, 55]. It is important to take an holistic approach to mitigate bias and incorporate our work as a part of such a broader effort. Finally, our work adds to the line of works that develop fair decision-making algorithms robust to inaccuracies in data [42, 6, 54, 25, 67, 66, 48, 14]. Acknowledgements. This research was supported in part by NSF Awards CCF-2112665 and IIS-2045951, and an AWS MLRA Award. [1] Open CV: Open Source Computer Vision Library. https://github.com/opencv/opencv_ 3rdparty/raw/dnn_samples_face_detector_20170830/res10_300x300_ssd_iter_ 140000.caffemodel. [2] Rakesh Agrawal, Sreenivas Gollapudi, Alan Halverson, and Samuel Ieong. Diversifying Search Results. In Proceedings of the Second ACM International Conference on Web Search and Data Mining, WSDM 09, page 5 14, New York, NY, USA, 2009. Association for Computing Machinery. [3] Ismail Sengor Altingovde, Engin Demir, Fazli Can, and Ozg ur Ulusoy. Incremental Cluster Based Retrieval Using Compressed Cluster-Skipping Inverted Files. ACM Transactions on Information Systems (TOIS), 26(3):1 36, 2008. [4] Mc Kane Andrus, Elena Spitzer, Jeffrey Brown, and Alice Xiang. What We Can t Measure, We Can t Understand: Challenges to Demographic Data Procurement in the Pursuit of Fairness. In FAcc T, pages 249 260. ACM, 2021. [5] Dana Angluin and Philip D. Laird. Learning From Noisy Examples. Mach. Learn., 2(4):343 370, 1987. [6] Pranjal Awasthi, Matth aus Kleindessner, and Jamie Morgenstern. Equalized Odds Postprocessing under Imperfect Group Information. In International Conference on Artificial Intelligence and Statistics, pages 1770 1780. PMLR, 2020. [7] Ziv Bar-Yossef and Maxim Gurevich. Random Sampling from a Search Engine s Index. Journal of the ACM (JACM), 55(5):1 74, 2008. [8] Michael Bendersky and Xuanhui Wang. Advances in TF-Ranking, July 2021. https://ai. googleblog.com/2021/07/advances-in-tf-ranking.html. [9] Asia J. Biega, Krishna P. Gummadi, and Gerhard Weikum. Equity of Attention: Amortizing Individual Fairness in Rankings. In SIGIR, pages 405 414. ACM, 2018. [10] Joy Buolamwini and Timnit Gebru. Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification. In FAT, volume 81 of Proceedings of Machine Learning Research, pages 77 91. PMLR, 2018. [11] Chris Burges, Tal Shaked, Erin Renshaw, Ari Lazier, Matt Deeds, Nicole Hamilton, and Greg Hullender. Learning to Rank Using Gradient Descent. In Proceedings of the 22nd international conference on Machine learning, pages 89 96, 2005. [12] Christopher J.C. Burges. From Rank Net to Lambda Rank to Lambda MART: An Overview. Learning, 2010. [13] Toon Calders and Sicco Verwer. Three Naive Bayes Approaches for Discrimination-Free Classification. Data Min. Knowl. Discov., 21(2):277 292, 2010. [14] L. Elisa Celis, Lingxiao Huang, Vijay Keswani, and Nisheeth K. Vishnoi. Fair Classification with Noisy Protected Attributes. In ICML, volume 120 of Proceedings of Machine Learning Research. PMLR, 2021. [15] L. Elisa Celis and Vijay Keswani. Implicit Diversity in Image Summarization. Proc. ACM Hum. Comput. Interact., 4(CSCW2):139:1 139:28, 2020. [16] L. Elisa Celis, Anay Mehrotra, and Nisheeth K. Vishnoi. Interventions for Ranking in the Presence of Implicit Bias. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, FAT* 20, page 369 380, New York, NY, USA, 2020. Association for Computing Machinery. [17] L. Elisa Celis, Anay Mehrotra, and Nisheeth K. Vishnoi. Fair Classification with Adversarial Perturbations. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, 2021. [18] L. Elisa Celis, Damian Straszak, and Nisheeth K. Vishnoi. Ranking with Fairness Constraints. In ICALP, volume 107 of LIPIcs, pages 28:1 28:15. Schloss Dagstuhl - Leibniz-Zentrum fuer Informatik, 2018. [19] Chandra Chekuri, Jan Vondr ak, and Rico Zenklusen. Multi-budgeted Matchings and Matroid Intersection via Dependent Rounding. In Dana Randall, editor, Proceedings of the Twenty Second Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2011, San Francisco, California, USA, January 23-25, 2011, pages 1080 1097. SIAM, 2011. [20] Jiahao Chen, Nathan Kallus, Xiaojie Mao, Geoffry Svacha, and Madeleine Udell. Fairness Under Unawareness: Assessing Disparity When Protected Class Is Unobserved. In FAT, pages 339 348. ACM, 2019. [21] Cyril W Cleverdon. The Significance of the Cranfield Tests on Index Languages. In Proceedings of the 14th annual international ACM SIGIR conference on Research and development in information retrieval, pages 3 12, 1991. [22] Kushal Dave, Steve Lawrence, and David M Pennock. Mining the Peanut Gallery: Opinion Extraction and Semantic Classification of Product Reviews. In Proceedings of the 12th international conference on World Wide Web, pages 519 528, 2003. [23] Marc Elliott, Peter Morrison, Allen Fremont, Daniel Mccaffrey, Philip Pantoja, and Nicole Lurie. Using the Census Bureau s Surname List to Improve Estimates of Race/Ethnicity and Associated Disparities. Health Services and Outcomes Research Methodology, 9:252 253, 06 2009. [24] Robert Epstein and Ronald E Robertson. The Search Engine Manipulation Effect (SEME) And Its Possible Impact on the Outcomes of Elections. Proceedings of the National Academy of Sciences, 112(33):E4512 E4521, 2015. [25] Seyed A. Esmaeili, Brian Brubach, Leonidas Tsepenekas, and John Dickerson. Probabilistic Fair Clustering. In Neur IPS, 2020. [26] Benoˆıt Fr enay and Michel Verleysen. Classification in the Presence of Label Noise: A Survey. IEEE Trans. Neural Networks Learn. Syst., 25(5):845 869, 2014. [27] Sahin Cem Geyik, Stuart Ambler, and Krishnaram Kenthapadi. Fairness-Aware Ranking in Search & Recommendation Systems with Application to Linked In Talent Search. In KDD, pages 2221 2231. ACM, 2019. [28] Sahin Cem Geyik and Krishnaram Kenthapadi. Building Representative Talent Search at Linked In. Linked In Engineering, October 2018. http://bit.ly/2x65HDJ. [29] Avijit Ghosh, Ritam Dutt, and Christo Wilson. When Fair Ranking Meets Uncertain Inference. In SIGIR, pages 1033 1043. ACM, 2021. [30] Sruthi Gorantla, Amit Deshpande, and Anand Louis. On the Problem of Underranking in Group-Fair Ranking. In ICML, volume 139 of Proceedings of Machine Learning Research, pages 3777 3787. PMLR, 2021. [31] M. Gr otschel, L. Lovasz, and A. Schrijver. Geometric Algorithms and Combinatorial Optimization. Algorithms and Combinatorics. Springer Berlin Heidelberg, 2012. [32] Anik o Hann ak, Claudia Wagner, David Garcia, Alan Mislove, Markus Strohmaier, and Christo Wilson. Bias in Online Freelance Marketplaces: Evidence from Task Rabbit and Fiverr. In CSCW, page 1914 1933, 2017. [33] Kalervo J arvelin and Jaana Kek al ainen. Cumulated Gain-Based Evaluation of IR Techniques. ACM Trans. Inf. Syst., 20(4):422 446, oct 2002. [34] Glen Jeh and Jennifer Widom. Scaling Personalized Web Search. In Proceedings of the 12th international conference on World Wide Web, pages 271 279. ACM, 2003. [35] Christopher Jung, Changhwa Lee, Mallesh Pai, Aaron Roth, and Rakesh Vohra. Moment Multicalibration for Uncertainty Estimation. In Mikhail Belkin and Samory Kpotufe, editors, Proceedings of Thirty Fourth Conference on Learning Theory, volume 134 of Proceedings of Machine Learning Research, pages 2634 2678. PMLR, 15 19 Aug 2021. [36] Nathan Kallus, Xiaojie Mao, and Angela Zhou. Assessing Algorithmic Fairness with Unobserved Protected Class Using Data Combination. In FAT*, page 110. ACM, 2020. [37] Shiva Prasad Kasiviswanathan, Homin K. Lee, Kobbi Nissim, Sofya Raskhodnikova, and Adam D. Smith. What Can We Learn Privately? SIAM J. Comput., 40(3):793 826, 2011. [38] Matthew Kay, Cynthia Matuszek, and Sean A. Munson. Unequal Representation and Gender Stereotypes in Image Search Results for Occupations. In CHI, pages 3819 3828. ACM, 2015. [39] Omer Kirnap, Fernando Diaz, Asia Biega, Michael D. Ekstrand, Ben Carterette, and Emine Yilmaz. Estimation of Fair Ranking Metrics with Incomplete Judgments. In WWW, pages 1065 1075. ACM / IW3C2, 2021. [40] Jon M. Kleinberg and Manish Raghavan. Selection Problems in the Presence of Implicit Bias. In ITCS, pages 33:1 33:17. Schloss Dagstuhl Leibniz-Zentrum fuer Informatik, 2018. [41] Nikola Konstantinov and Christoph H. Lampert. On the Impossibility of Fairness-Aware Learning from Corrupted Data. In Jessica Schrouff, Awa Dieng, Miriam Rateike, Kweku Kwegyir-Aggrey, and Golnoosh Farnadi, editors, Proceedings of The Algorithmic Fairness through the Lens of Causality and Robustness, volume 171 of Proceedings of Machine Learning Research, pages 59 83. PMLR, 13 Dec 2022. [42] Alexandre Louis Lamy and Ziyuan Zhong. Noise-Tolerant Fair Classification. In Neur IPS, pages 294 305, 2019. [43] Elizabeth D. Liddy. Automatic Document Retrieval. In Encyclopedia of Language and Linguistics. Elsevier, 2005. [44] Jiahui Liu, Peter Dolan, and Elin Rønby Pedersen. Personalized News Recommendation Based on Click Behavior. In Proceedings of the 15th international conference on Intelligent user interfaces, pages 31 40. ACM, 2010. [45] Tie-Yan Liu. Learning to Rank for Information Retrieval. 3(3):225 331, mar 2009. [46] Christopher Manning, Prabhakar Raghavan, and Hinrich Sch utze. Introduction to Information Retrieval. Natural Language Engineering, 16(1):100 103, 2010. [47] Naresh Manwani and P. S. Sastry. Noise Tolerance Under Risk Minimization. IEEE Trans. Cybern., 43(3):1146 1151, 2013. [48] Anay Mehrotra and L. Elisa Celis. Mitigating Bias in Set Selection with Noisy Protected Attributes. In FAcc T, pages 237 248. ACM, 2021. [49] Omid Memarrast, Ashkan Rezaei, Rizal Fathony, and Brian D. Ziebart. Fairness for robust learning to rank. Co RR, abs/2112.06288, 2021. [50] Christopher Mims. Why Social Media Is So Good at Polarizing Us, October 2020. https://www.wsj.com/articles/ why-social-media-is-so-good-at-polarizing-us-11603105204. [51] Marco Morik, Ashudeep Singh, Jessica Hong, and Thorsten Joachims. Controlling Fairness and Bias in Dynamic Learning-to-Rank. In SIGIR, pages 429 438. ACM, 2020. [52] Rajeev Motwani and Prabhakar Raghavan. Randomized Algorithms. Cambridge university press, 1995. [53] Nima Mousavi. How Tight Is Chernoff Bound, 2010. [54] Hussein Mozannar, Mesrob I. Ohannessian, and Nathan Srebro. Fair Learning with Private Demographic Data. In ICML, volume 119 of Proceedings of Machine Learning Research, pages 7066 7075. PMLR, 2020. [55] Safiya Umoja Noble. Algorithms of Oppression: How Search Engines Reinforce Racism. NYU Press, 2018. [56] Gourab K. Patro, Lorenzo Porcaro, Laura Mitchell, Qiuyue Zhang, Meike Zehlike, and Nikhil Garg. Fair Ranking: A Critical Review, Challenges, and Future Directions. In 2022 ACM Conference on Fairness, Accountability, and Transparency, FAcc T 22, page 1929 1942, New York, NY, USA, 2022. Association for Computing Machinery. [57] Ulrich Pferschy, Joachim Schauer, and Clemens Thielen. Approximating the Product Knapsack Problem. Optimization Letters, 15(8):2529 2540, 2021. [58] Evaggelia Pitoura, Kostas Stefanidis, and Georgia Koutrika. Fairness in Rankings and Recommendations: An Overview. The VLDB Journal, 2021. [59] Rasmus Rothe, Radu Timofte, and Luc Van Gool. Deep Expectation of Real and Apparent Age from a Single Image without Facial Landmarks. International Journal of Computer Vision, 126(2-4):144 157, 2018. [60] Andrew D. Selbst, Danah Boyd, Sorelle A. Friedler, Suresh Venkatasubramanian, and Janet Vertesi. Fairness and Abstraction in Sociotechnical Systems. In Proceedings of the Conference on Fairness, Accountability, and Transparency, FAT* 19, pages 59 68, New York, NY, USA, 2019. ACM. [61] Ashudeep Singh and Thorsten Joachims. Fairness of Exposure in Rankings. In KDD, pages 2219 2228. ACM, 2018. [62] Ashudeep Singh and Thorsten Joachims. Policy Learning for Fairness in Ranking. In Neur IPS, pages 5427 5437, 2019. [63] Ashudeep Singh, David Kempe, and Thorsten Joachims. Fairness in Ranking under Uncertainty. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, 2021. [64] Pavan Kumar C Singitham, Mahathi S Mahabhashyam, and Prabhakar Raghavan. Efficiencyquality tradeoffs for vector score aggregation. In Proceedings of the Thirtieth international conference on Very large data bases-Volume 30, pages 624 635, 2004. [65] Michael Taylor, Hugo Zaragoza, Nick Craswell, Stephen Robertson, and Chris Burges. Optimisation Methods for Ranking Functions with Multiple Parameters. In Proceedings of the 15th ACM international conference on Information and knowledge management, pages 585 593, 2006. [66] Jialu Wang, Yang Liu, and Caleb Levy. Fair Classification with Group-Dependent Label Noise. In FAcc T, pages 526 536. ACM, 2021. [67] Serena Wang, Wenshuo Guo, Harikrishna Narasimhan, Andrew Cotter, Maya R. Gupta, and Michael I. Jordan. Robust Optimization for Fairness with Noisy Protected Groups. In Neur IPS, 2020. [68] Jason Weston, Samy Bengio, and Nicolas Usunier. Large Scale Image Annotation: Learning to Rank with Joint Word-Image Embeddings. Machine learning, 81(1):21 35, 2010. [69] Grace Hui Yang and Sicong Zhang. Differential Privacy for Information Retrieval. In WSDM, pages 777 778. ACM, 2018. [70] Ke Yang, Vasilis Gkatzelis, and Julia Stoyanovich. Balanced Ranking with Diversity Constraints. In IJCAI, pages 6035 6042. ijcai.org, 2019. [71] Ke Yang, Joshua R. Loftus, and Julia Stoyanovich. Causal Intersectionality and Fair Ranking. In FORC, volume 192 of LIPIcs, pages 7:1 7:20. Schloss Dagstuhl - Leibniz-Zentrum f ur Informatik, 2021. [72] Ke Yang and Julia Stoyanovich. Measuring Fairness in Ranked Outputs. In SSDBM, pages 22:1 22:6. ACM, 2017. [73] Meike Zehlike and Carlos Castillo. Reducing disparate exposure in ranking: A learning to rank approach. In WWW, pages 2849 2855. ACM / IW3C2, 2020. [74] Meike Zehlike, Ke Yang, and Julia Stoyanovich. Fairness in Ranking, Part I: Score-Based Ranking. ACM Comput. Surv., apr 2022. Just Accepted. [75] Meike Zehlike, Ke Yang, and Julia Stoyanovich. Fairness in Ranking, Part II: Learning-to-Rank and Recommender Systems. ACM Comput. Surv., apr 2022. Just Accepted.