# multigroup_proportional_representation_in_retrieval__52be68c7.pdf Multi-Group Proportional Representation in Retrieval Alex Oesterling Harvard University Claudio Mayrink Verdun Harvard University Carol Xuan Long Harvard University Alexander Glynn Harvard University Lucas Monteiro Paes Harvard University Sajani Vithana Harvard University Martina Cardone University of Minnesota Flavio P. Calmon Harvard University Image search and retrieval tasks can perpetuate harmful stereotypes, erase cultural identities, and amplify social disparities. Current approaches to mitigate these representational harms balance the number of retrieved items across population groups defined by a small number of (often binary) attributes. However, most existing methods overlook intersectional groups determined by combinations of group attributes, such as gender, race, and ethnicity. We introduce Multi-Group Proportional Representation (MPR), a novel metric that measures representation across intersectional groups. We develop practical methods for estimating MPR, provide theoretical guarantees, and propose optimization algorithms to ensure MPR in retrieval. We demonstrate that existing methods optimizing for equal and proportional representation metrics may fail to promote MPR. Crucially, our work shows that optimizing MPR yields more proportional representation across multiple intersectional groups specified by a rich function class, often with minimal compromise in retrieval accuracy. Code is provided at https://github.com/ alex-oesterling/multigroup-proportional-representation. 1 Introduction A recognized objective in fair machine learning (ML) is discovering, reporting, and mitigating representational harms [1 3]. Representational harms arise when systems reinforce and perpetuate the marginalization of population groups based on characteristics such as race, socioeconomic status, cultural background, and gender [1, 4 6]. These harms often manifest through the biased portrayal or misrepresentation of these groups, stereotype reinforcement, and erasure of cultural identities [7, 8]. Representational harms have been widely documented in ML systems. For instance, many freely available datasets used in ML are not geo-diverse [9], and generative models can output images that under-represent demographic minorities across gender and racial identities [10 12]. We focus on representation in retrieval tasks. Representational harms in retrieval and ranking tasks arise when retrieved items do not accurately reflect the true diversity and proportions present in reality, perpetuating harmful stereotypes and biases [13]. For instance, Kay et al. [2] demonstrated that, in 2015, only 10% of the top 100 results for CEO in Google Image Search were women, despite 28% of CEOs in the US being women, and recent studies [14, 15] have demonstrated that this bias is still present in modern image search engines. Later, Otterbacher et al. [16] showed that the search engine Bing usually produced twice as many men as compared to women when queried with the word person . Finally, [17] systematically studied gender biases in image representations by auditing the four most important image search engines: Google, Bing, Baidu, and Yandex. In particular, among other findings, this study showed that when a qualifier such as intelligent is added to the query person , the engines still exhibit a significant gender discrepancy; see also [1, Chapter 7]. When Equal contribution, correspondence to aoesterling@g.harvard.edu 38th Conference on Neural Information Processing Systems (Neur IPS 2024). retrieval is based on a search over vector embeddings of images and text, biases in embedding models can propagate to retrieved results, leading to disparate representation of people by demographic groups such as gender or race [18 20], as well as propagating spurious correlations from the dataset more broadly [21] or distorting results due to stereotypes present in the representation [22]. Though correlations between unrelated semantic concepts do not necessarily constitute representational harms, a wealth of research charts undesirable bias in embedding models, including publicly available vision-language embedding models such as CLIP [23] used for retrieval. Bias in embeddings can lead to bias in retrieval [24 26]. In fact, recent work by Srinivasan et al. [27] argues that many image retrieval systems lack social diversity as perceived by humans. Several interventions aim to control gross statistical deviations in group representations in retrieval. A common approach is ensuring equal representation of pre-defined population groups in retrieved items [28, 29]. An alternative goal is proportional representation (PR) [20, 30], where the target representation of different groups is proportional to some reference population or statistic. Various interventions have been proposed to achieve equal or proportional representation, including optimizing for diversity in addition to the similarity in retrieval [27, 31], directly selecting the objects retrieved based on balancing known or predicted sensitive attributes [29] and, for vector databases, modifying embeddings directly to remove information about group-defining attributes [18, 20]. Achieving proportional representation across multiple intersectional groups is challenging, as existing fairness interventions typically consider only a small number of pre-defined groups. Ensuring representation across individual groups (e.g., given by gender or race) does not guarantee representation across intersectional groups (e.g., given by gender and race), as demonstrated in Table 1 (and Table 5 in the Appendix). The study of intersectionality and its consequences is rooted in work in law and the social sciences [32 34] and has been a well-discussed problem for several decades. In machine learning systems, lack of intersectional representation can contribute to the invisibility of historically marginalized groups determined by multiple axes of identity [35], or their mistreatment through fairness gerrymandering, where interventions on fairness for specific groups may harm fairness on intersectional subgroups [36]. However, as the number of group-denoting attributes increases, the number of intersectional groups grows exponentially, quickly surpassing the number of retrieved items and limiting the achievable proportional representation. In fact, achieving optimal proportional representation across multiple attributes has been shown to be NP-hard [37]. We propose a metric called Multi-group Proportional Representation (MPR) to quantify the representation of intersectional groups in retrieval tasks. MPR measures the worst-case deviation between the average values of a collection of representation statistics computed over retrieved items relative to a reference population whose representation we aim to match. The set of representation statistics is given by a function class C, where each function c C maps a retrieved item x to a real number. For instance, c(x) { 1, 1} may denote binary group membership for an input x. In theory, C can be given by bounded complexity function classes, such as linear functions or shallow decision trees. In practice, C are functions defined over item attributes, such as vector embeddings or group-denoting labels, enabling MPR to measure proportional representation across complex, intersectional subgroups. Compared to naively counting the number of retrieved items across a (potentially exponential) number of pre-defined groups, MPR offers a more flexible, scalable, and theoretically grounded metric for multi-group representation in retrieval. In addition to introducing and carefully motivating the MPR metric, cf. Section 2, we make three contributions. First, we show how to compute MPR for several function classes in Section 3. Computing MPR relies on a curated dataset that represents the desired population whose proportional representation we aim to reflect in retrieved items. We derive sample complexity bounds for the size of this curated dataset based on the complexity of the function class C. We also prove that MPR can be computed in closed form for certain function classes or with a regression oracle. Second, we propose MOPR, an algorithm that retrieves items from a vector database to maximize their average similarity to a query embedding while satisfying an MPR constraint relative to the curated dataset (Section 4). MOPR achieves this by iteratively calling an oracle that computes MPR violations, yielding both relevant and representative retrieved items. Third, we evaluate MOPR on retrieval tasks using datasets such as Celeb A [38], the Occupations dataset [31], Fairface [39], and UTKFace [40]. In section 5, we show that MOPR Pareto-dominates competing approaches in balancing retrieval similarity and MPR. Related Work. Broadly speaking, existing work on fair and diverse retrieval with pretrained embeddings either: 1) modifies the embedding space to prioritize downstream fairness, or 2) provides retrieval algorithms optimizing for diversity or fairness. Mitigating Bias in Embedding Space. Various approaches have been proposed to produce vector embeddings of text and images that target diversity or mitigate bias. These include modifying loss functions to encourage group fairness [41], adversarial training [42, 43], disentangling representations [44], and in-processing fair sampling methods for gender imbalance [18]. Many popular embedding models are not trained with such approaches, so post-processing methods for fairness have also been developed for multimodal models such as CLIP [23], including CLIP-clip [18], CLIP-debias [20], and Fair CLIP [19]. Recent work by Srinivasan et al. [27] builds upon the pretrained image-text model Co Ca and leverages text-guided projections to extract an embedding, called PATHS, that captures complex notions of diversity in people, and then conduct diverse retrieval with a greedy algorithm. However, code for implementing and reproducing PATHS embeddings was not available at the time of publication. In our experiments, we report the MPR achieved by CLIP-clip and CLIP-debias in retrieval tasks and approximate the greedy retrieval algorithm proposed in [27] with vanilla CLIP embeddings. Unlike the aforementioned methods, MOPR is not based on modifying embeddings. Retrieval Algorithms with Diversity or Fairness. Instead of modifying embeddings, several methods aim to promote diversity, given possibly biased embeddings. Given an additional diversity constraint, a popular approach is Maximal Marginal Relevance (MMR) [45], where items are greedily retrieved in order to maximize a linear combination of similarity and diversity metrics. Celis et al. [30] and Celis and Keswani [31] focus on diversity and fairness in image retrieval, using a reweighing method that combines similarity and MMR to select diverse yet relevant images. Alternatively, the Post Hoc Bias Mitigation [29] method aims to achieve equal representation over a pre-defined attribute (such as gender) through calls to an oracle classifier and selecting a balanced group of images. We extend this prior work in two key ways: (i) we aim to achieve proportional representation across intersectional groups, going beyond the representation of two groups defined by binary attributes; (ii) MOPR explicitly balances MPR and retrieval similarity via an optimization that controls for both, offering an efficient alternative to greedy MMR-based methods. Diversity Metrics. A traditional measure of diversity for a set is the pairwise similarity of the retrieved items [46]. However, the general visual (dis)similarity metric is insufficient in optimizing for peoplerelated diversity such as skin tone or gender [30]. When the attribute (e.g., race) along which diversity is optimized is known, or, in other words, we have access to a ground truth attribute label, one can use fairness definition-derived metrics such as proportional representation (see next section and [47]). For more complex notions of diversity that aim to capture intersectional or subtle sociocultural identities (e.g., cultural backgrounds, lifestyles, nationalities, religions), human annotators can be employed to determine if the selected set of images is more diverse than the baseline (retrieval based on similarity to query only) [27]. MPR complements existing metrics by providing a rigorous approach to measuring representation across complex, intersectional groups defined by a rich class of functions. However, as with any fairness metric, MPR has limitations see Section 6 and may not always align with human perceptions of representation. In particular, we underscore the importance of involving stakeholders in defining representation goals and auditing retrieval systems. Multi-Group Fairness. Our work is directly informed by the burgeoning literature on multi-group fairness in classification, particularly the work of Hébert-Johnson et al. [48], Kim et al. [49, 50] who proposed rigorous frameworks of multi-group fairness auditing and post-processing to ensure fair predictions across identifiable subgroups. Recent efforts in this field include [51 53], which analyze sample complexity aspects of measuring and reporting multi-group representation, and [54, 55] who develop stronger variants of multi-group fairness notions. Unlike prior work, we focus on multi-group representation in retrieval rather than multi-group fairness in classification and regression. Multi-Attribute Proportional Representation. A recognized challenge in computational social choice research is building committees, groups, or sets of items that ensure proportional representation across several attributes [56 59]. The work most closely related to ours is by Lang and Skowron [37], which introduced the problem of multi-attribute proportional representation, where a set of items must be selected to reflect desired distributions over multiple attributes simultaneously. Their work established fundamental computational complexity results, showing that finding optimal proportional representation is NP-hard, and developed approximation algorithms and integer linear programs for promoting proportional representation across attributes. While our work shares similar goals, our MPR metric takes a different approach by measuring the worst-case deviation over computationally-identifiable groups defined by a function class C. This formulation connects naturally with statistical learning theory through MMD [60] and sample complexity bounds (Propositions 1 and 2). Though we use the term multi-group rather than multi-attribute to emphasize connections with the burgeoning literature on multi-group fairness [48, 49], both terms describe related problems of ensuring representation across multiple dimensions. We also highlight that work on multi-attribute representation in computational social choice (including [37]) strives for proportional representation marginally on each attribute, while our primary goal with this work is to achieve intersectional representation. Finally, while [37] similarly seeks to match retrieved candidates to a target distribution, we focus on database retrieval rather than social choice applications. The rich history of proportional representation in social choice theory and political sciences provides, however, an important context for our work. We briefly discuss these connections in Appendix E. There, we elaborate on how our approach relates to and differs in terms of representation formulation from [37]. 2 A Multi-Group Proportional Representation Metric Preliminaries. Consider a retrieval dataset of items from which we aim to retrieve relevant entries, denoted by DR = {xr 1, ..., xr n}. We assume that xr i Rd G, where G is a set of additional labels for each item. For example, items xr i can be d-dimensional embeddings of images, videos, or textual content, in which case G = . Alternatively, xr i = (ei, gi), where ei Rd is an embedding and gi is a vector of labels indicating group membership (e.g., gender, race, ethnicity). Given a query embedding q Rd, the goal of retrieval is to search and return the top-k most similar items to q in DR. The similarity between an embedding of q and items in DR is measured according to a metric κ : Rd Rd R. Throughout the paper, we assume that DR is a vector database and κ is cosine similarity between embeddings, though this formulation can be generalized. To retrieve items from DR, users prompt a query q (e.g., "Fortune 500 CEOs") which is then embedded Rd. Ideally, the user receives k retrieved items R(q) = {xr 1, ..., xr k} DR such that κ(q, xr i ) κ(q, x) for all x / R(q) and i [k]. The simplest setting for measuring representation is to consider only two population groups denoted by a binary variable a setting commonly found in the fair retrieval and generation literature [18, 19]. In this case, group membership is determined by a group-denoting function c : Rd G { 1, 1}. For an item xr i , c(xr i ) indicates membership to group 1 or 1 (e.g., male/female) based on its embedding ei and associated labels gi. If the retrieval dataset DR contains annotations for group membership, c(xr i ) can simply return the relevant feature from gi. However, when group labels are not present (i.e., G = ), c(xr i ) can be implemented as a classifier that predicts group membership based on the item s embedding ei. With a group-denoting function c(xr) in hand, we can measure the representation of each group in a set of retrieved items R(q). One popular constraint for representation is equal representation [2, 28], which aims to ensure that the number of retrieved items in each group is approximately the same, i.e., 1 xr R(q) c(xr) 0. An alternative metric is proportional representation [20, 61], which quantifies the deviation of group membership from a reference distribution Q. Methods that promote proportional representation aim to ensure 1 xr R(q) c(xr) EQ [c(X)]. Here, the measure Q captures the distribution of a reference population whose representation statistics we aim to match. For example, if Q were the distribution of individuals in the US, EQ [c(X)] could measure the proportion of men vs. women in the US and be approximated using Census data. Proportional representation generalizes equal representation since different groups are rarely uniformly distributed over a given population. Naturally, the choice of distribution Q is application and context-dependent we revisit the choice of Q below. Importantly, if Q is biased, then biases will be propagated to the retrieved items. Multi-Group Proportional Representation. We aim to ensure that retrieved items represent individuals from diverse and intersectional population groups. Instead of measuring proportional representation in terms of the average of a single group-denoting function c(xr), we consider a class of Q-measurable functions, the representation statistics class, C c : Rd G R . This set C may represent multiple, potentially uncountable overlapping groups. We formally define multi-group proportional representation next. Definition 1. For a reference representation distribution Q, a set of Q-measurable set of representation statistics C, and a set of k retrieved items R, we define the Multi-Group Proportional Representation (MPR) metric as MPR(C, R, Q) sup c C xr R c(xr) EQ[c(X)] A set of items R is (C, ρ)-multi-group proportional representative of Q if MPR(C, R, Q) ρ. The MPR metric quantifies the representativeness of a rich class of statistics, denoted by functions in C, within a set of retrieved items R. This generalization is crucial for capturing intersectional groups. For example, C could contain functions that map items to demographic groups based on race, gender, age, or combinations thereof. Alternatively, when labels in G contain such group-denoting attributes, C can represent decision trees of a given depth over features (e.g., all combinations of pairs of race, gender, and age). The MPR metric compares the empirical average of each c C over the retrieved set R to its expectation under the reference distribution Q. By defining Q appropriately, we can flexibly specify different statistical representation goals, such as equal representation (EQ[c(X)] = 0 for binary c) or proportional representation w.r.t. a target population. Measuring representation requires defining what constitutes fair and proportional representation. While equal representation may suffice for binary groups, proportional representation of more complex intersectional groups requires care. The choice of representation reference statistics (given by the distribution Q in the definition of MPR) should be application-dependent, context-aware, and culturally sensitive. Remark 1. MPR is equivalent to the maximum mean discrepancy (MMD) of representation statistics in C between the empirical distribution over retrieved items R and the reference distribution Q. MMD-based metrics have a long history [60, 62] and are used in hypothesis testing for comparing distributions. MMD is a particular case of the Integral Probability Metric (IPM) [63, 64], allowing us to borrow from the rich literature on IPMs to measure and ensure MPR in practice. Remark 2. MPR can be viewed as the representation in retrieval counterpart of multi-group fairness metrics found in classification, such as multiaccuracy [49] and multicalibration [48]. Whereas multiaccuracy and multicalibration measure if classification error residuals are correlated with any group represented within a class C, MPR captures proportional representation of groups in C within a set of retrieved items a fundamentally different problem. The idea of representing groups in terms of a computationally identifiable class of functions C is directly inspired by the multi-group fairness in classification literature, and we adopt similar notation (i.e., c and C). Curated Datasets for Proportional Representation. A key challenge in computing MPR is selecting the reference representation distribution Q and measuring the expectation of functions in C against this distribution. In the simple case where C consists of a small number of functions indicating membership in individual groups, we could potentially set the proportional representation targets EQ[c(X)] for each group c C manually, e.g., by defining a target fraction of men/women or individuals from different ethnic backgrounds. This quickly becomes infeasible when there are many intersectional groups and impossible if C is uncountable. Moreover, in most practical settings, Q will very likely not have a simple closed-form analytical expression. In practice, i.i.d. samples drawn for Q may be available. For instance, the retrieval dataset DR itself may be drawn from the target population for which we aim to preserve proportional representation in retrieved items. Alternatively, we may have access to a dataset that was carefully curated to be representative of a diverse population. Examples include the Fair Face dataset [39], which was designed to be balanced across race, gender, and age groups, and the AVFS dataset [65]. More generally, we refer to datasets with samples drawn from the target population Q as a curated dataset. Definition 2 (Curated Dataset). Let Q be a probability distribution over Rd G tailored to account for diversity and representation of stakeholders who query the retrieval dataset. The curated dataset with m samples drawn the distribution Q is denoted as DC {xc 1, ..., xc m}. We denote the MPR of a set of retrieved items R relative to the empirical distribution over DC as MPR(C, R, DC). Constructing a proper DC is critical to properly measuring bias and preventing downstream harms, and the nature of DC is context-dependent and may vary based on the specific application and desired representation goals. We reiterate that if the curated dataset is biased, then these biases will be propagated to downstream usages of MPR. For these reasons, we recommend that curated datasets used for MPR measurements be developed and verified through participatory design approaches [66, 67] in collaboration with diverse stakeholders. By involving stakeholders in defining representation goals, we can help ensure that the proportional representation in information retrieval systems measured by MPR aligns with the values and needs of the user base these systems serve. Finally, we note that we can also condition the curated dataset on a given query. Specifically, for a query q, we can retrieve relevant samples from both the curated dataset DC and the retrieval dataset DR. The samples retrieved from DC can then serve as a conditional curated dataset, denoted as DC|q, which captures the desired representation target specific to the query q. This approach allows for a more granular and context-aware proportional representation. In the next two sections, we introduce theoretical guarantees and algorithms that are agnostic to the specific choice of DC. In other words, our proposed methods for measuring and promoting MPR in retrieval are general and can be applied regardless of how the curated dataset is constructed. In our numerical experiments, presented in Section 5, we use Fair Face [39] as the curated dataset. 3 Computing Multi-Group Proportional Representation Computing MPR in Defnition 1 requires approximating two quantities: the representation distribution Q and the supremum supc C. We first establish generalization bounds for approximating Q using i.i.d. samples. We then show how to compute supc C for several classes of representation statistics C. Error in approximation Q via a curated dataset. Proposition 1, proved in Appendix C, bounds the deviation between the empirical MPR computed over the curated dataset DC drawn i.i.d. from reference distribution Q and the true MPR measured over Q. Proposition 1 (Generalization Gap of MPR). Let R(q) = {xr i }k i=1 be a set of k retrieved samples, DC = {xc i}m i=1 be a curated dataset comprised of m i.i.d. samples from a target representation distribution Q, and δ > 0. If C = {c : Rd G { 1, 1}} with Rademacher complexity Rm(C) then, with probability at least 1 δ, |MPR(C, R, DC) MPR(C, R, Q)| Rm(C) + We can extend Proposition 1 to any set of bounded functions C (see, e.g., bounds on empirical MMD estimates in [60]). Note that the guarantee in (2) only holds for a single set of retrieved items R in response to a query. Proposition 2 provides a bound on the size m of an i.i.d. curated dataset DC that ensures an ϵ-accurate estimate of MPR for a set of M queries. Proposition 2 (Query Budget Guarantee). Consider any set of M queries Q = {q1, ..., q M} where M N. Let VC(C) denote the VC-dimension of the class C with range in { 1, 1}. For ϵ > 0, if DC consists of m i.i.d. samples drawn from Q where m satisfies ϵ2 + 2 log 2M then, with probability at least 1 δ, |MPR(C, R, DC) MPR(C, R, Q)| ϵ. The above results provide guidelines on the size of the curated dataset required to accurately estimate MPR relative to a true target representation distribution Q. If the class of representation statistics C is very complex (in the VC-dimension sense), then the size m of the curated dataset DC must be proportionally large to represent Q accurately. Hence, the curation dataset should be designed having in mind (i) the number of diverse queries being asked and (ii) the complexity of the function class C. While Proposition 2 provides a conservative bound, as generalization bounds are often not tight [68], these results nevertheless offer insight into the relationship between the complexity of C and the accuracy of MPR estimation. In the remainder of the paper, we assume that MPR is computed against a fixed curated dataset DC of size m. We consider four specific instantiations of the set of representation statistics C: (i) C is closed under scalar multiplication, i.e., if c C, then λc C for any λ R, (ii) C is a set of functions taking values in { 1, 1}; (iii) C consists of linear functions; and (iv) C consists of functions in a Reproducing Kernel Hilbert Space (RKHS). Next, we show that for cases (i) and (ii) MPR can be computed using calls to an oracle that performs regression over C and, for linear functions (iii), MPR has a simple closed-form expression. Due to its technical nature, we defer the calculation of MPR for functions in an RKHS to Appendix D.2. Computing MPR via Mean Square Error (MSE) Minimization. When C is closed under multiplication or consists of binary functions with values 1, 1, MPR can be expressed as an MSE minimization. This enables us to compute MPR by leveraging existing black-box oracles that perform regression under quadratic loss, such as regressors implemented in scikit-learn [69]. The key observation for expressing MPR as an MSE minimization is that MPR can be formulated as a maximum correlation problem. Consider the (row-wise) concatenation of the retrieval dataset DR and the curated dataset DC, given by X [xr 1, ..., xr n, xc 1, ..., xc m], where xi is the i-th entry of X. For the remainder of this section, we consider a fixed set of retrieved items R. Let a {0, 1}n be a vector indicating items that are retrieved from DR for a given query, i.e., ai = 1 xr i R. Under this notation, where retrieved items are indicated by the vector a, MPR can be reformulated as MPR(C, R, DC) = sup c C i=1 aic(xr i ) 1 i=1 c(xc i) i=1 c(xi)eai where ea Rn+m has i-th entry given by eai 1i n ai m. This reformulation will be useful for explicitly casting MPR-constrained retrieval as an optimization problem. When C consists of binary functions in range { 1, 1}, MPR is equivalent to regression over C of ea under quadratic loss, since i=1 (c(xi) eai)2 = n + m + ea ea + 2 sup c C i=1 c(xi)eai. (5) The absolute value in (4) can be recovered by regressing ea instead of ea. Most classes of regression functions in R are closed under scalar multiplication (e.g., multiplying the output of a decision tree regressor by a constant still yields a decision tree regressor). However, in this case, it follows from (4) that MPR is unbounded. Even for bounded C, without proper normalization MPR can scale with dataset size, which is undesirable for a representational measure. We constrain C to a set of normalized functions for MPR to be bounded and independent of dataset size. This constraint also allows us to cast MPR as an MSE minimization over C. We formally state this result next, proven in Appendix C, which will be applied in Section 4 to develop a cutting-plane-based algorithm for ensuring MPR in retrieval called MOPR. Proposition 3. Let C {c C | Pm+n i=1 c(xi)2 = mk m+k} where C is closed under scalar multiplication. Let c(X) = [c(x1), ..., c(xn+m)]. Then 0 MPR(C , R, DC) 1, and for any c arg inf c C i=1 (c(xi) eai)2, (6) let ˆc : X R be defined as ˆc(x) = mk/(m+k) c (X) 2 c (x). Then, we have that ˆc arg sup c C i=1 c(xi)eai Computing MPR for Bounded-Norm Linear Regression. Recall that each item in the retrieval or curated datasets is of the form xi = (ei, gi), where ei is an embedding and gi are labels. When xi Rl (i.e., the labels have numerical values), we can consider C as a linear set of functions over retrieved items. In this case, MPR enjoys a closed-form expression, as stated in the next proposition. Proposition 4. Let items in the retrieval and curated datasets be vectors in Rl for l < m + n and C = x 7 w x | w Rk . Moreover, let X R(m+n) l be the matrix formed by concatenating items in the retrieval and curated dataset DR and DC, respectively. For C in Proposition 3, we have MPR(C , R, DC) = mk m + k U l ea 2 , (8) Algorithm 1 MOPR (Multi-group Optimized Proportional Retrieval) Input: DR, query q, number of iterations T, number of items to retrieve k, MPR constraint ρ, Oracle(a) that computes solves (4) and returns c C that approximates the sup. 1: e C , si = κ(xr i , q) 2: for iter {1, 2, ..., T} do 3: a arg maxa [0,1]n a s subject to 1 k Pn i=1 aic(xr i ) 1 m Pm i=1 c(xc i) ρ for c e C 4: (mpr_violation, c) Oracle(a) Solve (4) 5: if mpr_violation ρ or iter = T then 6: return R {xr i DR corresponding to k largest entries of ai} 7: else 8: e C e C {c} 9: end if 10: end for where X = UΣV is the SVD of X and Ul R(m+n) l are the left singular vectors in U corresponding to the top-l largest singular values.1 Proposition 4 can be directly adapted to linear functions defined only over embeddings or a subset of features of retrieved items. The closed-form expression for MPR in (8) also allows MPR-constrained retrieval to be computed by a quadratic program, as discussed in Appendix D.1. 4 Promoting Multi-Group Proportional Representation in Retrieval Tasks We develop an optimization framework for retrieving the k most similar items in a vector database to a given query q while satisfying a target MPR threshold of ρ. We formulate the retrieval goal as maximizing the average similarity between a query and retrieved items, given by si = κ(xr i , q) (recall Section 2 for notation), where κ(xr i , q) is given by cosine similarity in our experiments. The problem of maximizing utility in retrieval while satisfying an MPR constraint (expressed as Eq. (4)) can be formulated as the integer program: max a {0,1}n a s subject to sup c C i=1 aic(xr i ) 1 i=1 c(xc i) i=1 ai = k. (9) When C is given by normalized linear functions as in Proposition 4, the integer constraints can be relaxed to a [0, 1]n and the optimization can be approximated via standard quadratic solvers; see Appendix D.1. Multi-Group Optimized Proportional Retrieval. We approximate (9) via the Multi-Group Optimized Proportional Retrieval Algorithm (MOPR), described in Algorithm 1. MOPR is essentially a cutting-plane method that aims to find a set of k items that maximize utility while approximately satisfying an MPR constraint of ρ. The algorithm iterates between (i) a call to an oracle that computes MPR and returns a function c C that achieves the supremum in (4) and (ii) a call to a linear program (LP) solver that approximates the top-k most similar items to a query subject to linear constraints on the violation measured by c. Each oracle call adds an additional constraint to the LP solver which, in turn, approximates the top-k items under an increasing set of constraints. The solution of the LP is rounded to satisfy the integer constraint a {0, 1}n. In our implementation, we assume that MPR can be computed via MSE minimization (cf. Section 3) in which case we consider the normalized class C . The oracle call consists of running a black-box quadratic loss minimization over C and normalizing. Interestingly, we observe that, despite relaxing the integer constraints in each LP call, the solution to the relaxed problem is very sparse and (after rounding) approximates well the solution of the ideal integer program (9) for moderate values of ρ. However, the method can fail to accurately approximate the IP solution when ρ is small. We present a more detailed analysis of MOPR in Appendix D, which discusses potential convergence issues with the cutting plane method as well as stopping conditions. 1For analysis of functions in an RKHS, see Appendix D.2 0.65 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP Figure 1: Cosine similarity vs MPR for query A photo of a programmer" with k = 50 images retrieved. From left to right: Celeb A, UTKFaces, Occupations. Values are normalized so MPR and similarity when retrieving the top-k items is the point (1,1). MOPR Pareto-dominates baselines and significantly closes the MPR gap. 5 Numerical Experiments In this section, we show that MOPR is effective in promoting more proportional representation across intersectional groups while preserving similarity between retrieved items and a given query. Notably, MOPR Pareto-dominates competing benchmarks in terms of achieved MPR gap and utility. Datasets. We conduct retrieval over three image datasets of faces: Celeb A [38], which includes labels for gender, age, and various other attributes, UTKFace [40], which contains gender, age, and race attributes, and Occupations [2], which contains gender attributes. We compute MPR for equal representation by constructing a synthetic dataset DC balanced over all attributes and for proportional representation using Fair Face [39] as the curated dataset DC, since it is a carefully designed dataset of faces with subgroup attributes for race, gender, and age. In all cases, each dataset entry consists of an image (CLIP) embedding ei and a set of labels gi. Benchmarks. We compare our method with four baselines. Debias CLIP [20] modifies text queries with a prepended learned embedding for debiased retrieval but does not allow for tunable control of the representation-similarity trade-off, so it only results in a single retrieval set over which we compute similarity and MPR. CLIP-Clip [18] trims features from CLIP embeddings that have high mutual information with gender, allowing for control of how many features are clipped from the embedding. PBM [29] post-processes CLIP embeddings to mitigate bias by predicting gender attributes and subsampling from each gender equally. PBM also allows tuning the representation-similarity tradeoff by controlling the likelihood of sampling at random or in a balanced manner from the retrieval dataset. Finally, the state-of-the-art for fair retrieval first constructs PATHS embeddings [27], which use a set of adjective-noun keywords to capture a broad set of identity vectors in vision-language models, and then greedily samples using the MMR [45] algorithm to navigate the representation-similarity tradeoff. However, the authors method of computing PATHS embeddings is not public, so we used CLIP embeddings to replicate their algorithm. We refer to this method as MMR. For a detailed discussion of these algorithms and their approaches to representation, see Appendix E. Experimental Setup. We consider three classes of representation statistics C: linear regression, decision trees, and MLPs (the last two are presented in Appendix G). In both cases, we compute MPR over C , i.e., C normalized over the retrieval and curated datasets (see Prop. 3). We report results over 10 queries for occupations suggested by Chat GPT 3.5 (Appendix H) and an additional 45 occupations for the the Occupations dataset. To accelerate retrieval, we query each of our retrieval datasets for the top 10k items according to the prompt A photo of a {query}" in terms of raw similarity using FAISS [70]. We similarly retrieve the top 10k items from Fair Face as the query-conditioned curated dataset DC. One reason we first filter for the top 10k items is the runtime of MMR, which is a greedy algorithm that traverses the full dataset for each retrieval k (even for 10k entries and k = 50 retrieved items, MMR takes hours to run for 10 queries). In order to evaluate representation over interpretable population groups, we consider that representation statistics C are computed only over labels gi given in Fair Face. These labels are not present in all retrieval datasets. Thus, for Celeb A and Occupations, we train linear probes on Fair Face s CLIP embeddings to predict their race and age attributes, and when using UTKFace, we map from Fair Face s race labels to UTKFace s race labels by mapping Southeast Asian and East Asian Table 1: Retrieval averaged over 10 queries with 50 retrievals on UTKFace dataset [40] for MOPR, Top-k, and MMR [27]. Entries indicate average percentage representation in retrieved items. MOPR is able to balance representation across classes, whereas Top-k and MMR miss intersectional groups (highlighted in red). White Black Asian Indian Others Sum Top-k Male 21.2 11.84 13.2 7.70 10.8 7.54 21.2 20.70 1.4 1.56 67.8 23.72 Female 17.2 21.04 6.6 6.14 4.0 2.82 2.8 2.56 1.6 1.50 32.2 23.72 Sum 38.4 17.18 19.8 12.56 14.8 7.38 24 19.38 3.0 2.04 MMR Male 28.8 8.96 11.0 4.58 10.2 4.04 10.8 2.40 3.8 1.06 64.6 7.74 Female 16.2 9.06 7.6 3.66 5.0 3.00 4.0 1.26 2.6 1.56 35.4 7.74 Sum 45.0 7.76 18.6 5.38 15.2 5.82 14.8 3.12 6.4 1.96 MOPR (Ours) Male 7.8 3.74 10.2 2.90 13.0 2.56 11.8 2.44 7.2 3.70 50 0 Female 12.2 3.74 9.8 2.90 7.0 2.56 8.2 2.44 12.8 3.70 50 0 Sum 20 0 20 0 20 0 20 0 20 0 to Asian and Middle Eastern and Latino_Hispanic to Other. While the fair ML community should engage in broader discussions addressing the ethics of predicting sensitive attributes from CLIP embeddings [71] (especially given that CLIP itself has been found to be biased) and the issues surrounding the grouping and re-grouping of diverse racial identities, we acknowledge that these are larger-scale issues that our work does not presume to address. We retrieve k = 50 items for each of the above queries. For a given function class and query, we compute the baseline MPR and average cosine similarity given by the top 50 most similar items. Then, we conduct a parameter sweep over ρ starting from this max-MPR value, inputting each value to MOPR in Algorithm 1. We normalize our results such that the Top-k MPR and similarity are the point (1,1) on each graph, and each point measures the fraction of Top-k MPR and similarity. Results. In Fig. 1 we report results for proportional representation for the query A photo of a programmer with respect to Fair Face for linear regression models, with additional results for other queries, as well as decision trees and MLPs in Appendix G (Figs. 8-19). We observe that MOPR Pareto-dominates benchmark methods, and is able to preserve similarity while reaching lower levels of MPR. Methods based on directly modifying embeddings or queries for a single group attribute such as CLIP-Clip and Debias CLIP only partially reduce MPR at a high utility cost. This is because these methods do not modify the retrieval algorithm but use an unbiased embedding and hope by chance that the retrieved items will be representational. The most competitive method to ours is the retrieval algorithm MMR, which is our attempt to replicate PATHS. Though MMR is competitive to ours for large MPR, it fails to achieve small values of MPR relative to MOPR. It is also notable that MOPR can drive MPR to near zero for UTKFace, yet gaps remain in Celeb A and Occupations, indicating this gap may be caused by the use of probes to estimate group attribute labels. In Table 1, we construct a synthetic curation set with an equal number of each group attribute. This allows us to evaluate the performance of MOPR in achieving equal representation. We conduct a similar hyperparameter sweep as above, and take the retrieval set with the minimum MPR averaged over all 10 queries to fairly compare the best possible representation of each method.2 Our results demonstrate that MOPR can exactly achieve our equal representation goals when provided with an equally-balanced curation dataset, whereas prior baselines (see Debias CLIP, CLIP-Clip, and PBM in Appendix Table 5) underrepresent several intersectional groups. 6 Concluding Remarks In this work, we introduced a novel retrieval metric called Multi-group Proportional Representation (MPR) and developed algorithms to ensure MPR in retrieval tasks. By measuring deviations in a set of representation statistics, MPR provides a scalable approach to quantifying and enforcing proportional representation for complex, intersectional groups. We analyzed the generalization properties and realizable regimes of MPR and, through experiments in image retrieval, demonstrated its favorable utility-fairness trade-off compared to existing fair retrieval algorithms. 2Aggregate results should be handled with care. As each query-conditioned retrieval setting results in a different distribution, average-case performance does not preclude the existence of a query where baseline methods achieve superior results. See Limitations (Appendix B) for further discussion. 7 Acknowledgements The authors would like to thank the discussions with Filipe Goulart Cabral. This material is based upon work supported by the National Science Foundation under awards CAREER-1845852, CIF-1900750, CIF-2231707, and CIF-2312667, FAI-2040880, and also by the National Science Foundation Graduate Research Fellowship under Grant No. DGE-2140743. The views expressed here are those of the authors and do not reflect the official policy or position of the funding agencies. [1] Solon Barocas, Moritz Hardt, and Arvind Narayanan. Fairness and machine learning: Limitations and opportunities. MIT Press, 2023. [2] Matthew Kay, Cynthia Matuszek, and Sean A Munson. Unequal representation and gender stereotypes in image search results for occupations. In Proceedings of the 33rd annual acm conference on human factors in computing systems, pages 3819 3828, 2015. [3] Kate Crawford. The trouble with bias, 2017. Accessed May 1, 2024. https://www.youtube. com/watch?v=f Mym_BKWQzk. [4] Sarah Myers West, Meredith Whittaker, and Kate Crawford. Discriminating systems. AI Now, pages 1 33, 2019. [5] Solon Barocas, Kate Crawford, Aaron Shapiro, and Hanna Wallach. The problem with bias: Allocative versus representational harms in machine learning. In 9th Annual conference of the special interest group for computing, information and society, page 1. Philadelphia, PA, USA, 2017. [6] Sanjana Gautam, Pranav Narayanan Venkit, and Sourojit Ghosh. From melting pots to misrepresentations: Exploring harms in generative ai. ar Xiv preprint ar Xiv:2403.10776, 2024. [7] Harini Suresh and John Guttag. A framework for understanding sources of harm throughout the machine learning life cycle. In Proceedings of the 1st ACM Conference on Equity and Access in Algorithms, Mechanisms, and Optimization, pages 1 9, 2021. [8] Jared Katzman, Angelina Wang, Morgan Scheuerman, Su Lin Blodgett, Kristen Laird, Hanna Wallach, and Solon Barocas. Taxonomizing and measuring representational harms: A look at image tagging. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 14277 14285, 2023. [9] Shreya Shankar, Yoni Halpern, Eric Breck, James Atwood, Jimbo Wilson, and D Sculley. No classification without representation: Assessing geodiversity issues in open data sets for the developing world. ar Xiv preprint ar Xiv:1711.08536, 2017. [10] Sasha Luccioni, Christopher Akiki, Margaret Mitchell, and Yacine Jernite. Stable bias: Evaluating societal representations in diffusion models. Advances in Neural Information Processing Systems, 36, 2024. [11] Federico Bianchi, Pratyusha Kalluri, Esin Durmus, Faisal Ladhak, Myra Cheng, Debora Nozza, Tatsunori Hashimoto, Dan Jurafsky, James Zou, and Aylin Caliskan. Easily accessible text-to-image generation amplifies demographic stereotypes at large scale. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, pages 1493 1504, 2023. [12] Alexander Lin, Lucas Monteiro Paes, Sree Harsha Tanneru, Suraj Srinivas, and Himabindu Lakkaraju. Word-level explanations for analyzing bias in text-to-image models. ar Xiv preprint ar Xiv:2306.05500, 2023. [13] Latanya Sweeney. Discrimination in online ad delivery. Communications of the ACM, 56(5): 44 54, 2013. [14] Danaë Metaxa, Michelle A Gan, Su Goh, Jeff Hancock, and James A Landay. An image of society: Gender and racial representation and impact in image search results for occupations. Proceedings of the ACM on Human-Computer Interaction, 5(CSCW1):1 23, 2021. [15] Yunhe Feng and Chirag Shah. Has CEO Gender Bias Really Been Fixed? Adversarial Attacking and Improving Gender Fairness in Image Search . In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 11882 11890, 2022. [16] Jahna Otterbacher, Jo Bates, and Paul Clough. Competent men and warm women: Gender stereotypes and backlash in image search results. In Proceedings of the 2017 chi conference on human factors in computing systems, pages 6620 6631, 2017. [17] Roberto Ulloa, Ana Carolina Richter, Mykola Makhortykh, Aleksandra Urman, and Celina Sylwia Kacperski. Representativeness and face-ism: Gender bias in image search. new media & society, 26(6):3541 3567, 2024. [18] Jialu Wang, Yang Liu, and Xin Wang. Are gender-neutral queries really gender-neutral? mitigating gender bias in image search. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 1995 2008, 2021. [19] Junyang Wang, Yi Zhang, and Jitao Sang. Fairclip: Social bias elimination based on attribute prototype learning and representation neutralization. ar Xiv preprint ar Xiv:2210.14562, 2022. [20] Hugo Berg, Siobhan Hall, Yash Bhalgat, Hannah Kirk, Aleksandar Shtedritski, and Max Bain. A prompt array keeps the bias away: Debiasing vision-language models with adversarial learning. In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 806 822, 2022. [21] Jae Myung Kim, A. Sophia Koepke, Cordelia Schmid, and Zeynep Akata. Exposing and mitigating spurious correlations for cross-modal retrieval. 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pages 2585 2595, 2023. [22] Mohsen Abbasi, Sorelle A Friedler, Carlos Scheidegger, and Suresh Venkatasubramanian. Fairness in representation: quantifying stereotyping as a representational harm. In Proceedings of the 2019 SIAM International Conference on Data Mining, pages 801 809. SIAM, 2019. [23] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748 8763. PMLR, 2021. [24] Siobhan Mackenzie Hall, Fernanda Gonçalves Abrantes, Hanwen Zhu, Grace Sodunke, Aleksandar Shtedritski, and Hannah Rose Kirk. Visogender: A dataset for benchmarking gender bias in image-text pronoun resolution. Advances in Neural Information Processing Systems, 36, 2024. [25] Robert Wolfe and Aylin Caliskan. Markedness in visual semantic ai. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, pages 1269 1279, 2022. [26] Tolga Bolukbasi, Kai-Wei Chang, James Y Zou, Venkatesh Saligrama, and Adam T Kalai. Man is to computer programmer as woman is to homemaker? debiasing word embeddings. Advances in neural information processing systems, 29, 2016. [27] Hansa Srinivasan, Candice Schumann, Aradhana Sinha, David Madras, Gbolahan Oluwafemi Olanubi, Alex Beutel, Susanna Ricco, and Jilin Chen. Generalized people diversity: Learning a human perception-aligned diversity representation for people images. ar Xiv preprint ar Xiv:2401.14322, 2024. [28] Anay Mehrotra and Elisa Celis. Mitigating bias in set selection with noisy protected attributes. Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, 2020. [29] Fanjie Kong, Shuai Yuan, Weituo Hao, and Ricardo Henao. Mitigating test-time bias for fair image retrieval. Advances in Neural Information Processing Systems, 36, 2024. [30] Elisa Celis, Vijay Keswani, Damian Straszak, Amit Deshpande, Tarun Kathuria, and Nisheeth Vishnoi. Fair and diverse dpp-based data summarization. In International conference on machine learning, pages 716 725. PMLR, 2018. [31] L Elisa Celis and Vijay Keswani. Implicit diversity in image summarization. Proceedings of the ACM on Human-Computer Interaction, 4(CSCW2):1 28, 2020. [32] Kimberle Crenshaw. Demarginalizing the intersection of race and sex: A black feminist critique of antidiscrimination doctrine, feminist theory and antiracist politics. In University of Chicago Legal Forum, volume 1989, page 8, 1989. [33] Kimberlé Williams Crenshaw. Mapping the margins: Intersectionality, identity politics, and violence against women of color. In The public nature of private violence, pages 93 118. Routledge, 2013. [34] Sumi Cho, Kimberlé Williams Crenshaw, and Leslie Mc Call. Toward a field of intersectionality studies: Theory, applications, and praxis. Signs: Journal of women in culture and society, 38 (4):785 810, 2013. [35] Joy Buolamwini and Timnit Gebru. Gender shades: Intersectional accuracy disparities in commercial gender classification. In Conference on fairness, accountability and transparency, pages 77 91. PMLR, 2018. [36] Michael Kearns, Seth Neel, Aaron Roth, and Zhiwei Steven Wu. Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. In International conference on machine learning, pages 2564 2572. PMLR, 2018. [37] Jérôme Lang and Piotr Skowron. Multi-attribute proportional representation. Artificial Intelligence, 263:74 106, 2018. [38] Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the wild. In Proceedings of International Conference on Computer Vision (ICCV), December 2015. [39] Kimmo Karkkainen and Jungseock Joo. Fairface: Face attribute dataset for balanced race, gender, and age for bias measurement and mitigation. In Proceedings of the IEEE/CVF winter conference on applications of computer vision, pages 1548 1558, 2021. [40] Zhifei Zhang, Yang Song, and Hairong Qi. Age progression/regression by conditional adversarial autoencoder. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, 2017. [41] Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. Learning fair representations. In Sanjoy Dasgupta and David Mc Allester, editors, Proceedings of the 30th International Conference on Machine Learning, volume 28 of Proceedings of Machine Learning Research, pages 325 333, Atlanta, Georgia, USA, 17 19 Jun 2013. PMLR. [42] Harrison Edwards and Amos Storkey. Censoring representations with an adversary. In 4th International Conference on Learning Representations, pages 1 14, 2016. [43] David Madras, Elliot Creager, Toniann Pitassi, and Richard Zemel. Learning adversarially fair and transferable representations. In Jennifer Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 3384 3393. PMLR, 10 15 Jul 2018. [44] Elliot Creager, David Madras, Jörn-Henrik Jacobsen, Marissa Weis, Kevin Swersky, Toniann Pitassi, and Richard Zemel. Flexibly fair representation learning by disentanglement. In International conference on machine learning, pages 1436 1445. PMLR, 2019. [45] Jaime Carbonell and Jade Goldstein. The use of mmr, diversity-based reranking for reordering documents and producing summaries. In Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval, pages 335 336, 1998. [46] Barry Smyth and Paul Mc Clave. Similarity vs. diversity. In International conference on case-based reasoning, pages 347 361. Springer, 2001. [47] Evaggelia Pitoura, Kostas Stefanidis, and Georgia Koutrika. Fairness in rankings and recommendations: an overview. The VLDB Journal, pages 1 28, 2022. [48] Ursula Hébert-Johnson, Michael Kim, Omer Reingold, and Guy Rothblum. Multicalibration: Calibration for the (computationally-identifiable) masses. In International Conference on Machine Learning, pages 1939 1948. PMLR, 2018. [49] Michael P Kim, Amirata Ghorbani, and James Zou. Multiaccuracy: Black-box post-processing for fairness in classification. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, pages 247 254, 2019. [50] Michael P Kim, Christoph Kern, Shafi Goldwasser, Frauke Kreuter, and Omer Reingold. Universal adaptability: Target-independent inference that competes with propensity scoring. Proceedings of the National Academy of Sciences, 119(4):e2108097119, 2022. [51] Eliran Shabat, Lee Cohen, and Yishay Mansour. Sample complexity of uniform convergence for multicalibration. Advances in Neural Information Processing Systems, 33:13331 13340, 2020. [52] Lucas Monteiro Paes, Ananda Theertha Suresh, Alex Beutel, Flavio Calmon, and Ahmad Beirami. Multi-group fairness evaluation via conditional value-at-risk testing. ar Xiv preprint ar Xiv:2312.03867, 2023. [53] Parikshit Gopalan, Lunjia Hu, and Guy N Rothblum. On computationally efficient multi-class calibration. ar Xiv preprint ar Xiv:2402.07821, 2024. [54] Parikshit Gopalan, Michael P Kim, Mihir A Singhal, and Shengjia Zhao. Low-degree multicalibration. In Conference on Learning Theory, pages 3193 3234. PMLR, 2022. [55] Zhun Deng, Cynthia Dwork, and Linjun Zhang. Happymap: A generalized multicalibration method. In 14th Innovations in Theoretical Computer Science Conference, ITCS 2023, page 41. Schloss Dagstuhl-Leibniz-Zentrum fur Informatik Gmb H, Dagstuhl Publishing, 2023. [56] Bailey Flanigan, Paul Gölz, Anupam Gupta, Brett Hennig, and Ariel D Procaccia. Fair algorithms for selecting citizens assemblies. Nature, 596(7873):548 552, 2021. [57] Edith Elkind, Piotr Faliszewski, Piotr Skowron, and Arkadii Slinko. Properties of multiwinner voting rules. Social Choice and Welfare, 48:599 632, 2017. [58] Virginie Do, Jamal Atif, Jérôme Lang, and Nicolas Usunier. Online selection of diverse committees. ar Xiv preprint ar Xiv:2105.09295, 2021. [59] Piotr Skowron, Piotr Faliszewski, and Jérôme Lang. Finding a collective set of items: From proportional multirepresentation to group recommendation. Artificial Intelligence, 241:191 216, 2016. [60] Arthur Gretton, Karsten M Borgwardt, Malte J Rasch, Bernhard Schölkopf, and Alexander Smola. A kernel two-sample test. The Journal of Machine Learning Research, 13(1):723 773, 2012. [61] Ajil Jalal, Sushrut Karmalkar, Jessica Hoffmann, Alex Dimakis, and Eric Price. Fairness for image generation with uncertain sensitive attributes. In International Conference on Machine Learning, pages 4721 4732. PMLR, 2021. [62] Karsten M Borgwardt, Arthur Gretton, Malte J Rasch, Hans-Peter Kriegel, Bernhard Schölkopf, and Alex J Smola. Integrating structured biological data by kernel maximum mean discrepancy. Bioinformatics, 22(14):e49 e57, 2006. [63] Alfred Müller. Integral probability metrics and their generating classes of functions. Advances in applied probability, 29(2):429 443, 1997. [64] Bharath K Sriperumbudur, Kenji Fukumizu, Arthur Gretton, Bernhard Schölkopf, and Gert RG Lanckriet. On the empirical estimation of integral probability metrics. Electronic Journal of Statistics, 6:1550 1599, 2012. [65] Jerone Theodore Alexander Andrews, Przemyslaw Joniak, and Alice Xiang. A view from somewhere: Human-centric face representations. In The Eleventh International Conference on Learning Representations, 2022. [66] Fernando Delgado, Stephen Yang, Michael Madaio, and Qian Yang. Stakeholder participation in ai: Beyond" add diverse stakeholders and stir". ar Xiv preprint ar Xiv:2111.01122, 2021. [67] Douglas Zytko, Pamela J. Wisniewski, Shion Guha, Eric PS Baumer, and Min Kyung Lee. Participatory design of ai systems: opportunities and challenges across diverse users, relationships, and application domains. In CHI Conference on Human Factors in Computing Systems Extended Abstracts, pages 1 4, 2022. [68] Yiding Jiang, Behnam Neyshabur, Hossein Mobahi, Dilip Krishnan, and Samy Bengio. Fantastic generalization measures and where to find them. In International Conference on Learning Representations, 2019. [69] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825 2830, 2011. [70] Jeff Johnson, Matthijs Douze, and Hervé Jégou. Billion-scale similarity search with GPUs. IEEE Transactions on Big Data, 7(3):535 547, 2019. [71] Foad Hamidi, Morgan Klaus Scheuerman, and Stacy M Branham. Gender recognition or gender reductionism? the social implications of embedded gender recognition systems. In Proceedings of the 2018 chi conference on human factors in computing systems, pages 1 13, 2018. [72] Peter L. Bartlett and Shahar Mendelson. Rademacher and gaussian complexities: Risk bounds and structural results. Journal of Machine Learning Research, 2002. [73] Martin J Wainwright. High-dimensional statistics: A non-asymptotic viewpoint, volume 48. Cambridge university press, 2019. [74] Jun Yang, Shengyang Sun, and Daniel M Roy. Fast-rate pac-bayes generalization bounds via shifted rademacher processes. Advances in Neural Information Processing Systems, 32, 2019. [75] Yuheng Bu, Shaofeng Zou, and Venugopal V Veeravalli. Tightening mutual information-based bounds on generalization error. IEEE Journal on Selected Areas in Information Theory, 1(1): 121 130, 2020. [76] Amir Asadi, Emmanuel Abbe, and Sergio Verdú. Chaining mutual information and tightening generalization bounds. Advances in Neural Information Processing Systems, 31, 2018. [77] Ron Amit, Baruch Epstein, Shay Moran, and Ron Meir. Integral probability metrics pac-bayes bounds. Advances in Neural Information Processing Systems, 35:3123 3136, 2022. [78] James E Kelley, Jr. The cutting-plane method for solving convex programs. Journal of the society for Industrial and Applied Mathematics, 8(4):703 712, 1960. [79] Elliot Ward Cheney and Allen A Goldstein. Newton s method for convex programming and tchebycheff approximation. Numerische Mathematik, 1(1):253 268, 1959. [80] B Curtis Eaves and WI Zangwill. Generalized cutting plane algorithms. SIAM Journal on Control, 9(4):529 542, 1971. [81] Dimitri Bertsekas. Convex optimization algorithms. Athena Scientific, 2015. [82] Donald M Topkis. Cutting-plane methods without nested constraint sets. Operations Research, 18(3):404 413, 1970. [83] Jean-Louis Goffin and Jean-Philippe Vial. Convex nondifferentiable optimization: A survey focused on the analytic center cutting plane method. Optimization methods and software, 17 (5):805 867, 2002. [84] Laurence A Wolsey and George L Nemhauser. Integer and combinatorial optimization. John Wiley & Sons, 2014. [85] Bernhard Schölkopf and Alexander J Smola. Learning with kernels: support vector machines, regularization, optimization, and beyond. MIT press, 2002. [86] T Hofmann, B Schölkopf, and AJ Smola. Kernel methods in machine learning. The Annals of Statistics, 36(3):1171 1220, 2008. [87] Weihao Gao, Sreeram Kannan, Sewoong Oh, and Pramod Viswanath. Estimating mutual information for discrete-continuous mixtures, 2018. [88] Duncan Black. Theory Committees and Elections. Cambridge University Press, 1958. [89] John R Chamberlin and Paul N Courant. Representative deliberations and representative decisions: Proportional representation and the borda rule. American Political Science Review, 77(3):718 733, 1983. [90] Burt L Monroe. Fully proportional representation. American Political Science Review, 89(4): 925 940, 1995. [91] Friedrich Pukelsheim. Proportional representation. Springer, 2017. [92] Jonathan K Hodge and Richard E Klima. The mathematics of voting and elections: a hands-on approach, volume 30. American Mathematical Soc., 2018. [93] Thomas Hare. A Treatise on the Election of Representatives, Parliamentary and Municipal. Longman, Green, Longman, & Roberts, 1861. [94] Michel Balinski. Fair majority voting (or how to eliminate gerrymandering). The American Mathematical Monthly, 115(2):97 113, 2008. [95] Ariel D Procaccia, Jeffrey S Rosenschein, and Aviv Zohar. On the complexity of achieving proportional representation. Social Choice and Welfare, 30(3):353 362, 2008. [96] Nadja Betzler, Arkadii Slinko, and Johannes Uhlmann. On the computation of fully proportional representation. Journal of Artificial Intelligence Research, 47:475 519, 2013. [97] Felix Brandt, Vincent Conitzer, Ulle Endriss, Jérôme Lang, and Ariel D Procaccia. Handbook of computational social choice. Cambridge University Press, 2016. [98] Gurobi Optimization, LLC. Gurobi Optimizer Reference Manual, 2024. URL https://www. gurobi.com. [99] Steven Diamond and Stephen Boyd. CVXPY: A Python-embedded modeling language for convex optimization. Journal of Machine Learning Research, 17(83):1 5, 2016. [100] Akshay Agrawal, Robin Verschueren, Steven Diamond, and Stephen Boyd. A rewriting system for convex optimization problems. Journal of Control and Decision, 5(1):42 60, 2018. Supplementary material to the paper Multi-Group Proportional Representation in Retrieval In this supplement to the paper, we begin in Section A and Section B by noting some ethical considerations and limitations for our work, respectively. Then, we present our proofs of Propositions 1, 3, 2, and 4 in Appendix C. In Appendix D, we discuss the MOPR algorithm, its convergence aspects, and MOPR for linear functions and functions belonging to an RKHS. We also compare the linear relaxation to the integer program optimization problem in Equation 9 empirically and similarly compare the result of MOPR to directly solving the closed form linear program in the case of linear models. In Appendix E, we describe the baselines reported in this paper and provide further retrieval statistics. In particular, we describe the algorithms CLIP-clip Wang et al. [18], CLIP-debias Berg et al. [20], Post-Hoc Bias Mitigation ([29] and PATHS [27]. In F, we provide a brief discussion about proportional representation in political sciences, give a few historical comments about social choice theory, and discuss the contribution given by [37]. Then, in Appendix G, we include an additional table of results for equal representation with the other baseline methods (Table 5). We also include plots for additional queries for the linear regression model (Figs. 8, 9, and 10) for as well as two additional models, 1) decision tree models in Figs. 11, 12, and 13 and for 2) multilayer perceptrons in Figs. 14, 15, and 16. We provide additional details on the experiment setting in Appendix H. Finally, we include some example sets of retrieved images for qualitative comparison. A Ethics considerations It is important to consider the long-term societal impact of enforcing MPR in real-world retrieval systems to understand its potential benefits and risks. First, applying MPR, such as with MOPR in deployment settings, could result in ethics-washing" where a company claims their system is fair because they measured MPR when, in fact, there are still representational harms or could provide users and developers with a false sense of fairness. Second, MPR is just one of many ways of analyzing representation, and as a statistical metric, it fails to consider aspects such as human perceptions of representation. There are legal and regulatory risks with overreliance on a single metric for fairness, especially if this metric is used to inform policy and decision-making. As mentioned previously, any biases present in the curation dataset will propagate to the results and downstream uses of MPR; furthermore, the choice of what dataset and what attributes to consider is an inherently social and political choice made by humans, subject to its own limitations and biases. Finally, the introduction of MPR in real-world retrieval systems could introduce new stereotypes or erase identities if not conducted properly. B Limitations Our aggregate results (e.g., Table 1) should be approached with caution, as each MPR-similarity curve is sampled from a different distribution of data conditioned on a query, limiting our ability to draw statistical conclusions. This does not preclude the existence of a query where a competing method may achieve an MPR-similarity trade-off point not achieved by MOPR. However, we note that MOPR is specifically optimized to promote MPR while preserving similarity, explaining its favorable performance. In all instances observed by the authors, MOPR Pareto dominates competing benchmarks and was the most successful method in achieving low values of MPR while maintaining high levels of similarity with a given query. Second, while MPR is designed to handle a large number of groups, the computational complexity of the proposed cutting-plane algorithm in the case of a general function class C can be a concern for high-dimensional feature spaces. Developing more efficient algorithms or approximation techniques could help address this issue. Finally, MPR is inherently limited by the curated dataset. If this dataset is biased against a given group, this bias will be propagated by MOPR. The creation and selection of curated datasets must carefully account for stakeholder interest and target representation statistics. Additionally, we note that not every combination of intersectional identities (such as white male" as opposed to black female") may be considered intersectional from a sociological perspective, and furthermore, these datasets do not contain multiracial labels. We do not presume to address representational issues in ML datasets, and instead, our experiments aim to demonstrate the flexibility and scalability of the MPR metric in handling complex intersectional groups. We hope that MPR will only prove more useful as datasets become more representative. C Proofs from Section 3 We start by proving Proposition 1 from Section 3 which bounds the error in MPR estimates when Q is approximated via a curated dataset DC. We do so by using tools from standard Rademacher complexity results in statistical learning theory, e.g., [72]. See also [73, Chapter 4]. We restate the theorems here to facilitate readability. We start with a lemma about the difference of suprema. Lemma 1. Let P and P be two distributions over X and C = {c : X [0, 1]}. Then sup c C EP [c(X)] sup c C EP [c(X)] sup c C {EP [c(X)] EP [c(X)]} . (10) sup c C EP [c(X)] sup c C EP [c(X)], = sup c C {EP [c(X)] EP [c(X)] + EP [c(X)]} sup c C EP [c(X)], sup c C {EP [c(X)] EP [c(X)]} + sup c C EP [c(X)] sup c C EP [c(X)], sup c C {EP [c(X)] EP [c(X)]} . Proposition 1 (Generalization Gap of MPR). Let R(q) = {xr i }k i=1 be a set of k retrieved samples, DC = {xc i}m i=1 be a curated dataset comprised of m i.i.d. samples from a target representation distribution Q, and δ > 0. If C = {c : Rd G { 1, 1}} with Rademacher complexity Rm(C) then, with probability at least 1 δ, |MPR(C, R, DC) MPR(C, R, Q)| Rm(C) + Proof. First, note that MPR(C, R, DC) MPR(C, R, Q) (11) i=1 c(xr i ) 1 i=1 c(xc i) i=1 c(xr i ) EQ[c(X)] i=1 c(xr i ) 1 i=1 c(xc i) i=1 c(xr i ) EQ[c(X)] i=1 c(xr i ) 1 i=1 c(xc i) i=1 c(xr i ) EQ[c(X)] i=1 c(xc i) EQ[c(X)] Where (13) comes from Lemma 1 and (15) comes from the triangular inequality. Additionally, by switching MPR(C, R, Q) by supc C 1 k Pk i=1 c(xi) EQ[c(X)] in (11) and repeating the previous argument, we conclude that: MPR(C, R, DC) sup c C i=1 c(xi) EQ[c(X)] i=1 c(xc i) EQ[c(X)] Therefore, from (16) we show MPR(C, R, DC) sup c C i=1 c(xi) EQ[c(X)] i=1 c(xc i) EQ[c(X)] c(xc i) 2 EQ here (19) comes from the generalization bound using Rademacher complexity from [72] and the fact that c(x) 2] hence the image of c 2 is in an interval of size 1. We now prove Proposition 2, which provides a bound on the size m of an i.i.d. curated dataset DC that ensures an ϵ-accurate estimate of MPR for a set of M queries. Proposition 2 (Query Budget Guarantee). Consider any set of M queries Q = {q1, ..., q M} where M N. Let VC(C) denote the VC-dimension of the class C with range in { 1, 1}. For ϵ > 0, if DC consists of m i.i.d. samples drawn from Q where m satisfies ϵ2 + 2 log 2M then, with probability at least 1 δ, |MPR(C, R, DC) MPR(C, R, Q)| ϵ. Proof. First, recall that the Rademacher complexity Rm(C) is upper bounded by the VC-dimension of C by the relation v u u t2VC(C) log em VC(C) where e is Euler s number. Denote the samples retrieved for a given query q by R(q) = {xr i (q)}k i=1. From Proposition 1 we have that for all δ |MPR(C, R, DC) MPR(C, R, Q)| Rm(C) + Therefore, using the previous equation if we take δ = δ M and denote R(q) = R we have that sup q Q |MPR(C, R, DC) MPR(C, R, Q)| Rm(C) + |MPR(C, R, DC) MPR(C, R, Q)| Rm(C) + M = δ. (21) Hence, we have that, with probability at least 1 δ for all q Q i=1 c(xr i (q)) EQ[c(X)] MPR(C, R(q), DC) + Rm(C) + MPR(C, R(q), DC) + v u u t2VC(C) log em VC(C) Where the inequality in (23) comes from the upper bound in (20). Moreover, if m 32VC(C) 2VC(C) log em VC(C) 2 and if m log( 2M δ ) 2ϵ2 then q 2. Therefore, we conclude that if m is such that ϵ2 + log 2M i=1 c(xr i (q)) EQ[c(X)] MPR(C, R(q), DC) + ϵ. (25) As described in Section 3, such bounds tend to be conservative and these bounds can be improved by doing a sharper analysis with PAC-Bayesian [74] or mutual information bounds [75]. See also [76] and [77]. We now prove Proposition 3. This proposition states that for specific function classes, computing MPR is equivalent to solving an MSE regression problem. This proposition allows us to efficiently estimate MPR with standard regression models from packages such as scikit-learn [69]. Proposition 3. Let C {c C | Pm+n i=1 c(xi)2 = mk m+k} where C is closed under scalar multiplication. Let c(X) = [c(x1), ..., c(xn+m)]. Then 0 MPR(C , R, DC) 1, and for any c arg inf c C i=1 (c(xi) eai)2, (6) let ˆc : X R be defined as ˆc(x) = mk/(m+k) c (X) 2 c (x). Then, we have that ˆc arg sup c C i=1 c(xi)eai Proof. We begin proving 0 MPR(C , R, DC) 1,. As a shorthand, let c(X) = [c(x1), ..., c(xn+m)]. The Cauchy-Schwarz inequality gives, i=1 c(xi)eai c(X) 2 ea 2, (26) where X R(m+n) l is the matrix formed by the row-wise concatenation of DR and DC and ea is the vector [ea0, ..., eam+n]. C gives us a constraint on the norm of c(X), and we can compute ea given there are k terms with value 1/k, m terms with value 1/m, and n k terms with value 0. Also note that mk m+k = ( 1 i=1 c(xi)eai m) 1 2 = 1. (27) Next, we prove the second statement. Consider the minimizer of the MSE problem, arg inf c C i=1 (c(xi) eai)2, (28) As C is closed under scalar multiplication, we can equivalently write this a two-part optimization problem arg inf c C i=1 (λc(xi) eai)2 ! where the inner optimization is quadratic and yields a solution λ = Pn+m i=1 c(xi)eai c(X) 2 2 , (30) as c(X) 2< . Plugging this back into Equation (29), the MSE problem becomes arg inf c C i=1 ( Pn+m i=1 c(xi)eai c(X) 2 2 c(xi) eai)2, = arg inf c C c(X) 4 2 c(xi)2 2 c(X), ea c(X) 2 2 c(xi)eai + ea2 i = arg inf c C c(X) 4 2 c(X) 2 2 2 c(X), ea c(X) 2 2 c(X), ea + = arg inf c C c(X), ea 2 i=1 ea2 i . As eai does not depend on c, this optimization is equivalent to arg sup c C c(X) 2 2 = arg sup c C | c(X), ea |2 Note that the supremum can be multiplied by a constant. Then we can multiply by some > 0 and get arg sup c C | c(X), ea |2 c(X) 2 2 = arg sup c C c(X) 2= | c(X), ea |2. (31) In other words, for > 0 and c arg sup c C | c(X), ea |2 and letting ˆc : X R be defined as ˆc(x) = c (X) 2 c (x), we have that ˆc arg sup c C, c(X) 2= | c(X), ea |2= arg sup c C, c(X) 2= i=1 c(xi)eai = arg sup c C, c(X) 2= i=1 c(xi)eai Letting = mk m+k, we can conclude that solving an MSE regression problem is equivalent to finding the representation statistic for MPR. Now, we establish Proposition 4, restated below for convenience. It says that MPR enjoys a closedform expression whenever the class C is given by linear functions. This proposition allows us to simplify MOPR since, in this case, the optimization problem can be cast as a quadratic program. Proposition 4. Let items in the retrieval and curated datasets be vectors in Rl for l < m + n and C = x 7 w x | w Rk . Moreover, let X R(m+n) l be the matrix formed by concatenating items in the retrieval and curated dataset DR and DC, respectively. For C in Proposition 3, we have MPR(C , R, DC) = mk m + k U l ea 2 , (8) where X = UΣV is the SVD of X and Ul R(m+n) l are the left singular vectors in U corresponding to the top-l largest singular values.3 3For analysis of functions in an RKHS, see Appendix D.2 The closed-form expression for MPR in (8) allows MPR-constrained retrieval to be approximated by a quadratic program, whose proof can be found in Appendix D. Moreover, the above proposition can be directly adapted to linear functions defined only for embeddings or a subset of features of retrieved items. Proof. Since we are considering the class C {c C | Pm+n i=1 c(xi)2 = mk m+k}, where C = x 7 w x | w Rd , the computation of the Multi-Group Proportional Representation metric MPR(C , R, DC) = arg sup c C i=1 c(xi)eai boils down to the following optimization problem i=1 (w x) eai (34) i=1 [w x]2 = mk m + k , which can be written as max w Rn+m ea T Xw (35) subject to Xw 2 2= mk m + k , where X R(n+m) l is the matrix formed by concatenating items in the retrieval and curated dataset DR and DC, respectively. Let X = UΣV be the SVD of X and Ul R(m+n) l are the left singular vectors in U corresponding to the top-l largest singular values. Then, the problem becomes max w Rm+n ea T UΣV w (36) subject to ΣV w 2 2= mk m + k . Now, by a change of variables ΣV w = ew and U l ea = z max ew,w Rm+n z T ew (37) subject to ew 2 2= mk m + k , We can use the method of Lagrange multipliers. Let us define the Lagrangian function: L(ew, w, λ, µ) = z T ew λ µT (ΣV w ew). (38) To find the optimal solution, we set the partial derivatives of the Lagrangian with respect to ew, w, λ, and µ to zero: L ew = z λew + µ = 0, (39) L w = VΣµ = 0, (40) L µ = ΣV w ew = 0. (42) From (42), we have VΣµ = 0. Assuming V and Σ are full rank, this implies µ = 0. Substituting this into (39), we get: z λew = 0 = ew = 1 Using (43) and the constraint ew 2 2= mk m+k, we can solve for λ: λ2 z 2 2= mk m + k = λ = z 2 Since we are maximizing z T ew, we choose the positive value of λ. Thus, the optimal ew is: mk m+k z 2 z. (45) Finally, using (45), we can find the optimal w: ΣV w = ew = w = (ΣV ) 1ew = mk m+k z 2 (ΣV ) 1z = mk m+k z 2 (ΣV ) 1U l ea. (46) The optimal w is then obtained by solving the linear system ΣV w = ew . Finally, plugging the solution back in to our objective and canceling terms, we arrive at our solution. Note that we can replace U with Ul in our original SVD as U R(m+n) l, so we can truncate the vectors in the null space of U. max w Rm+n ea T UΣV w = ea T UlΣV mk m+k z 2 (ΣV ) 1U l ea mk m+k U l ea 2 ea T Ul U l ea mk m + k U l ea 2 D The Multi-group Optimized Retrieval Algorithm In this section, we discuss the convergence of MOPR, the Multi-group Optimized Proportional Retrieval algorithm. As described in Section 4, the MOPR is essentially a cutting plane method (also known as the Kelley-Cheney-Goldstein method [78, 79]) applied to the linear function a T s; see also [80]. The algorithm begins by identifying the k-most similar items to a given query in DR. Then, an MSE-minimizing oracle is called and returns the statistic (or group-denoting function) c C with the most disproportionate representation for the set of retrieved items. The function c is then added as a constraint to a linear program that outputs a vector a [0, 1]n maximizing average similarity as in (9) subject to a single constraint 1 k P i = 1naic(xr i ) 1 m Pm i=1 c(xc i) ρ, where integer constraints are relaxed. The optimal a is rounded so only the largest k entries are marked as 1, corresponding to a new set of retrieved items. The process repeats with an increasing number of constraints to the linear program and halts when the oracle does not return a function that violates the target MPR constraint ρ. In mathematical terms, we start by solving the problem max a {0,1}na s (47) i=1 ai = k. Note that the solution to this problem is given by the vector a that finds the kth largest entries of s, i.e., it identifies the k-most similar items to a given query in DR. Then, Proposition 3 provides a way to generate cuts given the vector ak Rn found as a solution of the previous iteration. These cuts are given by the function c C with the most disproportionate representation for the set of retrieved items. At each iteration, we test if the MPR constraint, less or equal to ρ, is satisfied. If not, this gives us a new inequality to be incorporated into the linear problem (D) that outputs a vector a [0, 1]n maximizing average similarity as in (9) subject to a single constraint 1 k Pn i=1 aic(xr i ) 1 m Pm i=1 c(xc i) ρ, where integer constraints are relaxed. Therefore, the algorithm consists of iteratively adding a new constraint (cut) of the form G(a) =: sup c C i=1 aic(xr i ) 1 i=1 c(xc i) and solving the problem again until all the inequalities are fulfilled, i.e., until the oracle does not return a function that violates the target MPR constraint ρ. At this point, the algorithm halts. It remains to be shown that the description above indeed corresponds to a traditional cutting plane method for convex functions. For the subsequent discussion, we will denote the term inside the supremum in (48) by g(a, c), i.e., g(a, c) = 1 i=1 aic(xr i ) 1 i=1 c(xc i), for c e C. (49) The next proposition links the oracle via MSE, described in Proposition 3, with the subdifferential of the function G(a). Proposition 5. Let C be in infinite set and let G(a) = supc C|g(a, c)| be the convex restriction of the optimization problem (9) given by the supremum, i.e., G(a) ρ, a Rn. If c arg supc C|g(a , c)| and c denote the function c evaluated at the points x1, . . . , xn, i.e., c = [c (x1), . . . , c (xm)], then g := sign(g(a , c )) c G(a ), i.e., g is a subgradient of G at a point a . Proof. We start by noting that G(a) is a convex function since it is a supremum of a composition of an affine function with the absolute value. Also, note that G(a) can be written as G(a) = supc C|g(a, c)|= supc C|c a + bc|, where b : C R. We need to prove that G(a ) + g (a a ) G(a), a Rn (50) In fact, starting from the left-hand side leads to |g(a , c )|+ sign(g(a , c ))c (a a ) = |c a + bc|+ c a + bc |c a + bc|c (a a ) = |c a + bc|+ c a + bc |c a + bc| c a bc + c a + bc = |c a + bc| |c a + bc|2 |c a + bc| + sign(g(a , c ))(c a + bc ) |c a + bc | G(a), a Rn This shows that c = [c (x1), . . . , c (xm)] is a subgradient of G(a). The cutting plane strategy described in Algorithm 1 addresses the constraints of the form G(a) ρ that are not fulfilled. Since (50) holds, the algorithm adds linear approximations (or cutting planes) of the form G(a ) + g (a a ) ρ (51) to the problem at each iteration until the solution satisfies all the inequalities. This reformulation demonstrates that the MOPR algorithm can be viewed as a traditional cutting plane method for minimizing a linear function over the polyhedron P = {ai [0, 1] Pn i=1 ai = k}. If, at a certain iteration, the algorithm finds an iterate that satisfies all the inequalities, it terminates since the lower bound provided by the cutting plane method matches the upper bound. In the case where the algorithm continues adding constraints, the accumulation point (guaranteed to exist due to the compactness of the set P) will converge to an optimal solution, as proven in [81, Proposition 4.1.2] or the main theorem in [78, Section 2]. However, it is important to note that even when the algorithm terminates in a finite number of steps, the cutting plane method often requires a substantial number of iterations to converge and may exhibit numerical instabilities. Addressing these instabilities or exploring techniques to discard certain cuts for improved convergence, such as partial cutting plane methods, is beyond the scope of this paper. For further information on these topics, refer to [82, 83] and the references therein. The integer program may not be as scalable as its convex relaxation. If we relax it to ai [0, 1], this becomes a convex optimization problem solvable by standard methods (even for very large C) and provides an upper bound to the original program for reasonable values of the MPR violation (and for which Proposition 5 applies). However, the optimal solution of the relaxed problem might not be feasible for the original problem, because it might have fractional values for some ai, while the original problem requires these to be binary. There are certain special cases where the optimal solution of the relaxed problem is guaranteed to be integral (i.e., all ai are 0 or 1), in which case it is also an optimal solution for the original problem. This is the case, for example, when the constraint matrix is totally unimodular and the right-hand sides of the constraints are integral [84]. Still, we observe from empirical experiments that selecting the top k ai of the relaxed problem is a computationally efficient alternative to solving the integer program in Equation (9), since results from rounding produce a negligible difference to the upper bound (see Figures 2 and 3). 0.825 0.850 0.875 0.900 0.925 0.950 0.975 1.000 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor LP LP Top-k IP Figure 2: Comparison of the linear program with and without taking the top-k to the integer program. Solving the relaxed program is much more computationally efficient and achieves similar performance after rounding to solving the integer program. D.1 Multi-group Optimized Retrieval via Quadratic Programming Proposition 4 allowed us to have a simple closed-form representation of MPR in the case of bounded- norm linear regression functions, i.e., when C {c C | q mk m+k Pm+n i=1 c(xi)2 = 1}, and C = x 7 w x | w Rd . In this case, MOPR becomes 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect LP LP Top-k IP 0.88 0.90 0.92 0.94 0.96 0.98 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist LP LP Top-k IP 0.84 0.86 0.88 0.90 0.92 0.94 0.96 0.98 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef LP LP Top-k IP 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer LP LP Top-k IP 0.84 0.86 0.88 0.90 0.92 0.94 0.96 0.98 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer LP LP Top-k IP 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse LP LP Top-k IP 0.65 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer LP LP Top-k IP 0.88 0.90 0.92 0.94 0.96 0.98 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist LP LP Top-k IP 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher LP LP Top-k IP Figure 3: Additional comparisons of relaxed problem and top-k selection to integer program. chef", nurse", artist", lawyer", teacher", engineer", architect", scientist", and programmer". max a,y a s (52) mk m + k U l ea 2 ρ, a 1 = k, ai {0, 1}, where ea Rn+m has i-th entry given by eai 1i n ai This result allows us to explore the efficiency of the cutting plane algorithm in MOPR by comparing it to the solution to (52), which is a closed-form, convex quadratic program that can be easily optimized with existing solvers. Note that in these experiments, top-k refers to the process of solving the relaxed linear program and then rounding to get a solution in {0, 1}n by selecting the k items with the highest score given by a. This is different than the Top-k described elsewhere in the paper, which denotes conducting vanilla retrieval without fairness considerations. In Figures 4 and 6, we compare the quadratic optimization problem (Eqn. 52) to MOPR (a linear program, Algorithm 1) for linear regression and measure the MPR of our solutions with the closed form value from Proposition 4. As we can see, while the quadratic program traces a smoother curve, MOPR s cutting plane approach well-approximates the closed form solution. In these experiments, we retrieve 50 items from Celeb A, without a curation set. Finally, in Figures 5 and 7, we run MOPR with the quadratic program oracle and compare our measurement of MPR with the closed form solution and linear regressor-approximated measurement of MPR. In other words, where Figures 4 and 6 measure the quality of our retrieval algorithm given a perfect (closed form) MPR oracle, Figures 5 and 7 measure the quality of our MSE estimator of MPR in comparison to the perfect oracle. We observe that the MSE oracle is a perfect approximator of the closed form for MPR, as expected. 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer MOPR (Ours) MOPR Top-k QP Top-k QP Figure 4: Similarity vs MPR for Quadratic Program (Eqn. 52) and MOPR. MOPR well approximates the quadratic program along the Pareto frontier. Measured over a single query A photo of a lawyer" for 50 retrieved samples on Celeb A. 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: lawyer Closed Form MPR MSE MPR Estimator Figure 5: Similarity vs MPR for MSE estimated (Prop 3 and closed form (Prop 4) measures of MPR. For the class of linear models, a linear regression oracle perfectly achieves the analytical solution for MPR. Measured over a single query A photo of a lawyer" for 50 retrieved samples on Celeb A. D.2 Computing MPR for Functions in an RKHS Kernel methods are a generalized but practical, non-linear, and easy technique to implement algorithms for a plethora of important problems in machine learning [85, 86]. As MPR is closely related to the maximum mean discrepancy problem, a closed-form expression of MPR can be easily deduced by using results from [60, Lemma 6] when C is a reproducing kernel Hilbert space, which includes, for example, linear regression, logistic regression, Gaussian kernel ridge regression, and hard-margin SVMs. Proposition 6 (MPR for RKHS [60]). Let {xr i }k i=1 be retrieved samples R and {xc i}m i=1 the samples from the curation dataset DC. Let C be a reproducing kernel hilbert space with kernel K(., .), then MPR(C, R, Q) = i,j=1 K(xr i , xr j) 2 mk i,j=1 K(xr i , xc j) + 1 i,j=1 K(xc i, xc j) Therefore, given a kernel K(., .), we can easily compute MPR(C, R, Q) by evaluating the kernel at the points xr i , xc j and use this in Algorithm 1. 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef MOPR (Ours) MOPR Top-k QP Top-k QP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse MOPR (Ours) MOPR Top-k QP Top-k QP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist MOPR (Ours) MOPR Top-k QP Top-k QP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor MOPR (Ours) MOPR Top-k QP Top-k QP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher MOPR (Ours) MOPR Top-k QP Top-k QP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer MOPR (Ours) MOPR Top-k QP Top-k QP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect MOPR (Ours) MOPR Top-k QP Top-k QP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist MOPR (Ours) MOPR Top-k QP Top-k QP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer MOPR (Ours) MOPR Top-k QP Top-k QP Figure 6: Additional comparisons of Quadratic Program (Eqn. 52) and MOPR over queries chef", nurse", artist", doctor", teacher", engineer", architect", scientist", and programmer". 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: chef Closed Form MPR MSE MPR Estimator 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: nurse Closed Form MPR MSE MPR Estimator 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: artist Closed Form MPR MSE MPR Estimator 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: doctor Closed Form MPR MSE MPR Estimator 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: teacher Closed Form MPR MSE MPR Estimator 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: engineer Closed Form MPR MSE MPR Estimator 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: architect Closed Form MPR MSE MPR Estimator 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: scientist Closed Form MPR MSE MPR Estimator 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity MSE Estimator vs QP, Query: programmer Closed Form MPR MSE MPR Estimator Figure 7: Additional comparisons of MSE to closed form MPR estimator for queries chef", nurse", artist", doctor", teacher", engineer", architect", scientist", and programmer". Embedding Change Embedding Change Retrieval Pre-defined features Approach to Diversity Debias CLIP Custom Ignore/Decorrelate Attributes PBM Any VLM Equal Representation over Attributes PATHS Custom High Variance Representation over Attributes CLIP-Clip Any VLM Ignore/Decorrelate Attributes MOPR (ours) Any Multigroup Proportional Representation over Attributes Table 2: Axis of comparison for ours and existing fair retrieval methods. The blue color indicates a positive feature of the method, while the red color indicates a negative one. E Description of Fair Retrieval Algorithms and Retrieval Statistics In this section, we describe the methods we benchmark against that aim to promote fairness multimodal models like CLIP [23]. We also briefly overview proportional representation in political sciences. CLIP-clip (Wang et al. [18]): CLIP-clip is a post-processing algorithm with two stages. First, the mutual information between each positional feature of an embedding and some sensitive attributes is estimated (treating each sensitive attribute as a discrete random variable and each position as a continuous random variable [87]). Then, in order of greatest to lowest mutual information, at test-time some number of CLIP features are dropped from both the text and image embedding, resulting in less bias (at cost to the fidelity of the embedding). By varying the number of features dropped, a recall-bias trade-off is induced. CLIP-debias (Berg et al. [20]): The algorithm learns a series of tokens that are pre-pended to text queries into CLIP via adversarial training to minimize the ability to predict a sensitive attribute of the resulting embedding (in the default case, gender). Then, retrieval is conducted with the debiased embeddings. Post-Hoc Bias Mitigation ([29]): Post-Hoc Bias Mitigation is a post-processing method that relies on an oracle classifier to label elements that might be retrieved with either one of several sensitive attributes or a neutral label. Then, iteratively, Post-Hoc Bias Mitigation adds either a set of elements (one of each predicted sensitive attribute) or a predicted neutral element to target equal representation, based on which option has higher average cosine similarity. The oracle may be a pretrained classifier, a zero-shot method using CLIP embeddings, or the true labels. PATHS (Srinivasan et al. [27]): PATHS is a new embedding space built upon the pretrained image-text model Co Ca and leveraged text-guided projections to extract an embedding that captures complex notions of diversity in people, upon which a diverse retrieval method (such as the MMR method detailed above, or even MOPR) can be run to effectively increase the diversity of the returned samples. F Connection to Social Choice Theory In this section, we comment on the role of proportional representation in social sciences. As described in the related works in Section 1, proportional representation (broadly construed) has a long history in social choice theory and political sciences [88 92] and has been a guiding principle in the design of political systems, aiming to ensure representation reflects underlying population preferences and demographics. The foundations of social choice theory can be traced back to the late 18th century, with the seminal works of Jean-Charles de Borda and the Marquis de Condorcet, who developed fundamental voting methods and paradoxes that continue to influence modern computational approaches. During this same period, early proponents of proportional representation emerged, such as Honoré Gabriel Riqueti, Comte de Mirabeau, who discussed the idea in 1780. The 19th century saw significant contributions to the development of proportional representation, including the works of Thomas Wright Hill, Thomas Hare, and Charles Dodgson (better known as Lewis Carroll), who made substantial contributions to voting theory and proportional representation systems. Hare s 1859 publication, A Treatise on the Election of Representatives, Parliamentary and Municipal [93], outlined a system of single transferable vote for proportional representation. In the late 19th and early 20th centuries, the works of Carl Andrae, Victor D Hondt, and John Stuart Mill further explored and contributed to the discussion of proportional representation in the political science literature. The field was revolutionized in the 20th century by Kenneth Arrow s seminal work, whose impossibility theorem fundamentally shaped our understanding of social choice mechanisms. See [88 92]. Similarly, in ML, our aim is proportional representation in data and outputs, ensuring equitable performance across groups and mirroring representation goals in political systems. PR electoral systems [94], used by countries like Uruguay, Sweden, South Africa, and New Zealand, allocate legislative seats proportionally to votes received by parties. Challenges include, e.g., gerrymandering [94] and the complexity of apportionment and seat allocation [95, 96]. The emergence of computational social choice theory [97] in recent decades has built upon these classical foundations while introducing new algorithmic perspectives to address modern challenges. Researchers began studying algorithmic aspects of proportional representation, particularly focusing on the complexity of achieving different notions of proportionality in committee selection [95, 96]. The Hamilton and d Hondt methods, initially developed for parliamentary seat allocation, have been extended and generalized to handle more complex scenarios. As mentioned at the end of Section 1, a notable contribution in this direction is the work of Lang and Skowron [37], who introduced multiattribute proportional representation. Their framework generalizes classical apportionment methods to scenarios where candidates possess multiple attributes (such as gender, profession, and age), and representation goals must be satisfied across all these dimensions simultaneously. Their work provides both theoretical guarantees through approximation algorithms and practical methods for achieving proportional representation in modern committee selection problems, bridging the gap between classical political science approaches and contemporary computational challenges. In particular, they introduced the problem of multi-attribute proportional representation in committee selection. Their framework can be viewed as solving an (m, t)-filling problem: given m buckets (representing intersectional groups) with target capacities {t1, ..., tm} (derived from marginal probabilities), the goal is to optimally fill these buckets with appropriately labeled balls (committee members). While groundbreaking in extending classical apportionment methods to multiple attributes, their approach differs from ours in several key aspects. First, their work considers multiple attributes, but they primarily focus on matching target proportions derived from marginal distributions for each attribute independently rather than truly addressing intersectional representation. That is, their approach aims to satisfy target proportions for each individual attribute (like gender or profession separately) but does not explicitly handle the complex interactions between attributes (like representation of specific gender-profession combinations). This is in contrast to our framework, where the function class C can directly capture and measure such intersectional relationships. Moreover, their formulation focuses solely on matching target distributions without additional objectives, whereas our work explicitly balances representation with retrieval utility through cosine similarity maximization. Second, while they work with pre-specified target probabilities, we estimate the desired representation from a reference dataset, making our approach more flexible and data-driven. Third, and perhaps most importantly, our MPR metric measures representation through the lens of computationally identifiable groups defined by the aforementioned function class C rather than explicitly enumerating all possible intersectional groups. Through C, our framework provides a unified way to handle both simple and complex representations: it can capture discrete indicators for each attribute-value pair (like the Hamilton method; see Definition 5 in [37]) but also complex nonlinear relationships (like d Hondt method; see Definition 6 in [37]), and goes beyond both to measure intersectional representation via decision trees, linear functions, neural networks, or more generally, any function in a reproducing kernel Hilbert space (RKHS). This allows us to handle more complex notions of representation while maintaining computational tractability. In essence, while [37] provides fundamental insights for multi-attribute committee selection, our framework offers a more general approach suitable for modern retrieval systems where both representation and relevance matter. G Additional Results In this section, we include additional experimental results. First, we include a comparison of runtimes for each method in Table 3. We find that MOPR provides up to a 100x speedup when compared with the most competitive retrieval algorithm, MMR, while remaining competitive to methods such as CLIP-Clip and Debias CLIP that do not modify the retrieval algorithm at all and just use embedding similarity search. We additionally provide the similarity score for the minimum MPR achieved by each algorithm, which was used to report the results in Tables 1 and 5. These similarity scores can be found in Table 4. Even when significantly closing the MPR gap and achieving more representation than other methods, MOPR also preserves high cosine similarity with respect to the query. Second, we include full tables of intersectional retrieval performance on UTKFace for our method, Top-k, MMR, CLIP-Clip, PBM, and Debias CLIP. We report averaged results for each of the 10 queries while retrieving 50 images. In Table 5, we compute MPR with a linear regression oracle and optimize our method with the same linear regression oracle. We are able to perfectly balance across intersectional groups while other methods often fail to account for the Others race category and the Female gender category. Some methods, such as PBM, which optimize for equal representation, are able to achieve balanced representation across the gender axis but fail to consider diversity in race. In Table 6, we see a similar story for MPR measured and optimized with a decision tree. Next, we include additional similarity-MPR tradeoff curves for the other 9 queries not reported in the main paper (Figs. 8,9,10), as well as for all 10 queries on two additional models: a decision tree (Figs. 11, 12, 13) and with an MLP with 1 hidden layer of dimension 64 (Figs. 14, 15, 16). We observe a similar Pareto-domination to that observed in the main paper, as our method is able to lower MPR the most while preserving similarity even with more nonlinear oracle estimators of MPR. In the MLP setting, while PBM preserves high similarity on the whole, it is still unable to decrease MPR as low as MOPRis able to at 95% of top-k similarity. We also include results for an additional 45 queries for the Occupations dataset (with Fair Face as the curation dataset), as this dataset comes with 45 ground-truth occupations labels. These results again highlight the Pareto-dominance of MOPR, as shown in Figures 17, 18, and 19. In Figure 20, we explore the performance of MOPR in terms of precision, a common retrieval metric. We can do this using the Occupations dataset, as it has ground truth occupations labels with which we can benchmark retrieval precision. We see that MOPR performs competitively in terms of precision while also traversing more of the precision-MPR Pareto frontier. In Figure 21, we include nine additional precision-representation curves. Similar to the result for the query chief executive officer", MOPR maintains similar precision-representation performance to baseline methods. Finally, we include example sets of retrievals for architect and teacher queries over UTKFace with each of our baselines. Similar to the tables, we select each algorithm s solution with the lowest MPR after doing a hyperparameter sweep for each respective algorithm and plot the corresponding images below. The rest of the Appendix is solely tables and figures. k=10 k=50 k=150 MOPR(LP, 10 iterations) 1.434 0.304 1.6 0.54 1.371 0.361 MOPR(LP, 50 iterations) 5.14 0.352 4.199 1.74 4.342 1.803 MOPR(QP, linear regression) 0.528 0.061 0.533 0.066 0.584 0.089 MMR 151.961 1.126 267.412 3.863 556.184 6.324 PBM 0.275 0.162 0.241 0.013 0.276 0.151 CLIP-Clip 21.303 0.361 21.907 0.525 21.34 0.353 Debias CLIP 2e-4 1e-6 2e-4 1e-6 2e-4 1e-6 Table 3: Average running time (in seconds) over ten queries for k=10, 50, and 150 items for the different methods. MOPR is able to achieve 100x speedup over the SOTA retrieval algorithm, MMR, and maintains competitive performance with vanilla KNN-based methods such as CLIP-Clip and Debias CLIP. For more discussion, see the general comment in the rebuttal. Normalized Mean Cosine Similarity Top-k 1.0000 0.0000 MOPR 0.9049 0.0692 MMR 0.8047 0.1477 PBM 0.9416 0.0802 CLIP-Clip 0.7758 0.1433 Debias CLIP 0.8497 0.0985 Table 4: Normalized mean cosine similarity between the query embedding and retrieved image embeddings for retrieved images from Tables 1 and 5. The normalization is performed by dividing the mean similarity by the mean similarity given by the Top-k nearest neighbors of a query. Table 5: Retrieval averaged over 10 queries with 50 retrievals on UTKFace dataset [40] for MOPR, k-NN, MMR [27], CLIP-Clip, PBM, and Debias CLIP with MPR optimized and measured with a linear regression oracle. Entries indicate the average percentage representation in retrieved items. MOPR is able to balance representation across classes, whereas other baselines miss intersectional groups (highlighted in red). White Black Asian Indian Others Sum MOPR (Ours) Male 7.8 3.74 10.2 2.90 13.0 2.56 11.8 2.44 7.2 3.70 50 0 Female 12.2 3.74 9.8 2.90 7.0 2.56 8.2 2.44 12.8 3.70 50 0 Sum 20 0 20 0 20 0 20 0 20 0 Top-K Male 21.2 11.84 13.2 7.70 10.8 7.54 21.2 20.70 1.4 1.56 67.8 23.72 Female 17.2 21.04 6.6 6.14 4.0 2.82 2.8 2.56 1.6 1.50 32.2 23.72 Sum 38.4 17.18 19.8 12.56 14.8 7.38 24 19.38 3.0 2.04 MMR Male 28.8 8.96 11.0 4.58 10.2 4.04 10.8 2.40 3.8 1.06 64.6 7.74 Female 16.2 9.06 7.6 3.66 5.0 3.00 4.0 1.26 2.6 1.56 35.4 7.74 Sum 45.0 7.76 18.6 5.38 15.2 5.82 14.8 3.12 6.4 1.96 Male 23.2 9.96 11.0 5.16 10.8 5.88 14.2 8.70 1.8 2.08 61.0 17.08 Female 15.2 9.44 7.0 4.32 7.4 5.30 6.6 5.52 2.8 1.84 39.0 17.08 CLIP-Clip Sum 38.4 11.42 18.0 8.20 18.2 5.68 20.8 9.52 4.6 2.2 Male 19.8 7.82 13.0 7.12 9.4 5.66 15.2 8.72 1.2 1.6 58.6 11.62 Female 18.6 12.96 8.4 5.92 5.6 3.88 7.2 5.16 1.6 1.50 41.4 11.62 PBM Sum 38.4 13.50 21.4 11.38 15.0 6.70 22.4 13.44 2.8 2.04 Male 20.4 7.48 8.2 3.94 5.4 2.54 11.0 3.50 3.8 2.6 48.8 5.52 Female 19.4 6.98 9.4 5.14 8.4 4.28 10.0 4.64 4.0 1.78 51.2 5.52 Debias CLIP Sum 39.8 9.78 17.6 3.66 13.8 5.62 21.0 6.52 7.8 3.16 Table 6: Retrieval averaged over 10 queries with 50 retrievals on UTKFace dataset [40] for MOPR, k-NN, MMR [27], PBM, and Debias CLIP with MPR optimized and measured with a decision tree oracle. Entries indicate the average percentage representation in retrieved items. MOPR is able to balance representation across classes, whereas other baselines miss intersectional groups (highlighted in red). White Black Asian Indian Others Sum Ours Male 11.0 1.34 10.4 0.8 10.0 0 10.2 0.6 8.8 0.98 50.4 1.64 Female 10.6 1.28 9.6 0.80 9.6 1.20 10.2 1.08 9.6 1.2 49.6 1.64 Sum 21.6 1.64 20 1.26 18.6 1.2 20.4 1.2 18.4 1.2 Top-K Male 21.2 11.84 13.2 7.70 10.8 7.54 21.2 20.70 1.4 1.56 67.8 23.72 Female 17.2 21.04 6.6 6.14 4.0 2.82 2.8 2.56 1.6 1.50 32.2 23.72 Sum 38.4 17.18 19.8 12.56 14.8 7.38 24 19.38 3.0 2.04 MMR Male 27.4 9.13 11.8 5.02 10.2 4.33 10.2 3.03 3.2 2.4 62.8 11.77 Female 17 9.13 9 5.67 5.4 3.47 3.6 1.5 2.2 1.08 37.2 11.77 Sum 44.4 6.62 20.8 9.39 15.6 5.64 13.8 3.84 5.4 2.37 Male 19.6 13.11 12.4 6.56 9.4 6 15 8.91 1.2 1.6 57.6 12.74 Female 19.6 13.11 8.4 5.92 5.4 3.69 7.2 5.15 1.8 1.66 42.4 12.74 PBM Sum 39.2 13.45 20.8 11.07 14.8 6.46 22.2 13.61 3 2.05 Male 20.4 6.99 8.2 3.94 5.4 2.54 11 3.49 3.8 2.6 48.8 5.53 Female 19.4 6.99 9.4 5.14 8.4 4.27 10 4.65 4 1.79 51.2 5.53 Debias CLIP Sum 39.8 9.78 17.6 3.67 13.8 5.62 21 6.53 7.8 3.16 MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0.95 1.00 1.05 1.10 1.15 1.20 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.90 0.95 1.00 1.05 1.10 1.15 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0.65 0.70 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.850 0.875 0.900 0.925 0.950 0.975 1.000 1.025 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0.875 0.900 0.925 0.950 0.975 1.000 1.025 1.050 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 8: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over linear regression models for 9 additional queries for the Celeb A dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.0 0.2 0.4 0.6 0.8 1.0 1.2 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0.0 0.2 0.4 0.6 0.8 1.0 1.2 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 9: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over linear regression models for 9 additional queries for the UTKFace dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. 0.6 0.7 0.8 0.9 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0.8 0.9 1.0 1.1 1.2 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0.70 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.925 0.950 0.975 1.000 1.025 1.050 1.075 1.100 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0.6 0.7 0.8 0.9 1.0 1.1 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.70 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 10: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over linear regression models for 9 additional queries for the Occupations dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.65 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0.95 1.00 1.05 1.10 1.15 1.20 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.90 0.95 1.00 1.05 1.10 1.15 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0.65 0.70 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.850 0.875 0.900 0.925 0.950 0.975 1.000 1.025 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0.875 0.900 0.925 0.950 0.975 1.000 1.025 1.050 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 11: Fraction of Top-k cosine similarity vs. Fraction of Top-k MPR over decision trees for 10 queries for the Celeb A dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer 0.0 0.2 0.4 0.6 0.8 1.0 1.2 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0.0 0.2 0.4 0.6 0.8 1.0 1.2 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 12: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over decision trees for 10 queries for the UTKFace dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer 0.6 0.7 0.8 0.9 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0.8 0.9 1.0 1.1 1.2 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0.70 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.925 0.950 0.975 1.000 1.025 1.050 1.075 1.100 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0.6 0.7 0.8 0.9 1.0 1.1 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.70 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 13: Fraction of Top-k cosine similarity vs. Fraction of Top-k MPR over decision trees for 10 queries for the Occupations dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Paretodominates baselines and significantly closes the MPR gap. 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0.2 0.4 0.6 0.8 1.0 1.2 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0 2 4 6 8 10 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0 5 10 15 20 25 30 35 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 14: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over 2-layer MLPs (hidden dimension 64) for 10 queries for the Celeb A dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0.0 0.2 0.4 0.6 0.8 1.0 1.2 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0 2 4 6 8 10 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0 5 10 15 20 25 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0 20 40 60 80 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0.0 0.5 1.0 1.5 2.0 2.5 3.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 15: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over 2-layer MLPs (hidden dimension 64) for 10 queries for the UTKFace dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: programmer 0.0 0.5 1.0 1.5 2.0 2.5 Fraction of Top-k MPR Fraction of Top-k Similarity Query: architect 0 2 4 6 8 10 Fraction of Top-k MPR Fraction of Top-k Similarity Query: artist 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0.00 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0 1 2 3 4 5 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lawyer 0.0 0.5 1.0 1.5 2.0 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0 1 2 3 4 5 6 Fraction of Top-k MPR Fraction of Top-k Similarity Query: scientist 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 Fraction of Top-k MPR Fraction of Top-k Similarity Query: teacher Figure 16: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over 2-layer MLPs (hidden dimension 64) for 10 queries for the Occupations dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.8000.8250.8500.8750.9000.9250.9500.9751.000 Fraction of Top-k MPR Fraction of Top-k Similarity Query: administrative_assistant 0.80 0.85 0.90 0.95 1.00 1.05 1.10 Fraction of Top-k MPR Fraction of Top-k Similarity Query: announcer 0.825 0.850 0.875 0.900 0.925 0.950 0.975 1.000 Fraction of Top-k MPR Fraction of Top-k Similarity Query: bartender 0.92 0.94 0.96 0.98 1.00 1.02 1.04 1.06 Fraction of Top-k MPR Fraction of Top-k Similarity Query: biologist 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: building_inspector 0.850 0.875 0.900 0.925 0.950 0.975 1.000 1.025 Fraction of Top-k MPR Fraction of Top-k Similarity Query: bus_driver 0.8250.8500.8750.9000.9250.9500.9751.0001.025 Fraction of Top-k MPR Fraction of Top-k Similarity Query: butcher 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chef 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chemist 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: chief_executive_officer 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: childcare_worker 0.88 0.90 0.92 0.94 0.96 0.98 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: computer_programmer 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: construction_worker 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: cook 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: crane_operator Figure 17: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over linear regression models for 45 (First 15, see Figures 18 and 19) occupations present in the Occupations dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.85 0.90 0.95 1.00 1.05 1.10 1.15 1.20 Fraction of Top-k MPR Fraction of Top-k Similarity Query: custodian 0.86 0.88 0.90 0.92 0.94 0.96 0.98 1.00 1.02 Fraction of Top-k MPR Fraction of Top-k Similarity Query: customer_service_representative 0.94 0.96 0.98 1.00 1.02 1.04 Fraction of Top-k MPR Fraction of Top-k Similarity Query: doctor 0.850 0.875 0.900 0.925 0.950 0.975 1.000 1.025 Fraction of Top-k MPR Fraction of Top-k Similarity Query: drafter 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: electrician 0.70 0.75 0.80 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: engineer 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: garbage_collector 0.850 0.875 0.900 0.925 0.950 0.975 1.000 1.025 Fraction of Top-k MPR Fraction of Top-k Similarity Query: housekeeper 0.8750.9000.9250.9500.9751.0001.0251.050 Fraction of Top-k MPR Fraction of Top-k Similarity Query: insurance_sales_agent 0.88 0.90 0.92 0.94 0.96 0.98 1.00 1.02 Fraction of Top-k MPR Fraction of Top-k Similarity Query: lab_tech 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: librarian 0.8750.9000.9250.9500.9751.0001.0251.050 Fraction of Top-k MPR Fraction of Top-k Similarity Query: mail_carrier 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: nurse_practitioner 0.88 0.90 0.92 0.94 0.96 0.98 1.00 1.02 1.04 Fraction of Top-k MPR Fraction of Top-k Similarity Query: optician Figure 18: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over linear regression models for 45 (Second 15, see Figures 17 and 19) occupations present in the Occupations dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.8000.8250.8500.8750.9000.9250.9500.9751.000 Fraction of Top-k MPR Fraction of Top-k Similarity Query: administrative_assistant 0.90 0.95 1.00 1.05 1.10 1.15 Fraction of Top-k MPR Fraction of Top-k Similarity Query: pharmacist 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: pilot 0.70 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: plumber 0.85 0.90 0.95 1.00 1.05 Fraction of Top-k MPR Fraction of Top-k Similarity Query: police_officer 0.900 0.925 0.950 0.975 1.000 1.025 1.050 1.075 Fraction of Top-k MPR Fraction of Top-k Similarity Query: primary_school_teacher 0.86 0.88 0.90 0.92 0.94 0.96 0.98 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: receptionist 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: roofer 0.8000.8250.8500.8750.9000.9250.9500.9751.000 Fraction of Top-k MPR Fraction of Top-k Similarity Query: security_guard 0.84 0.86 0.88 0.90 0.92 0.94 0.96 0.98 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: software_developer 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: special_ed_teacher 0.85 0.90 0.95 1.00 1.05 1.10 1.15 Fraction of Top-k MPR Fraction of Top-k Similarity Query: technical_writer 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: telemarketer 0.75 0.80 0.85 0.90 0.95 1.00 Fraction of Top-k MPR Fraction of Top-k Similarity Query: truck_driver 0.8500.8750.9000.9250.9500.9751.0001.0251.050 Fraction of Top-k MPR Fraction of Top-k Similarity Query: welder Figure 19: Fraction of Top-k cosine similarity vs Fraction of Top-k MPR over linear regression models for 45 (Third 15, see Figures 17 and 18) occupations present in the Occupations dataset. Values are normalized so Top-k achieves point (1,1) in each case. MOPR Pareto-dominates baselines and significantly closes the MPR gap. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Query: chief_executive_officer Figure 20: Precision-MPR curve on Occupations dataset over linear regression. Retrieving 50 items for the query A photo of a chief executive officer." MPR remains competitive in terms of retrieval performance. MOPR (Ours) MMR PBM CLIP-Clip Debias CLIP 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Query: bartender 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Query: bus_driver 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Query: chemist 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Query: construction_worker 0.2 0.4 0.6 0.8 1.0 1.2 Fraction of Top-k MPR Query: customer_service_representative 0.0 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Query: housekeeper 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Query: nurse_practitioner 0.2 0.4 0.6 0.8 1.0 Fraction of Top-k MPR Query: receptionist 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 Fraction of Top-k MPR Query: special_ed_teacher Figure 21: Additional precision-representation plots over queries bartender", bus driver", chemist", construction worker", customer service representative", housekeeper", nurse practitioner", receptionist", and sepecial ed teacher". Figure 22: Retrievals for MOPR with query A photo of a teacher" Figure 23: Retrievals for k-NN with query A photo of a teacher" Figure 24: Retrievals for MMR with query A photo of a teacher" Figure 25: Retrievals for CLIP-Clip with query A photo of a teacher" Figure 26: Retrievals for PBM with query A photo of a teacher" Figure 27: Retrievals for Debias CLIP with query A photo of a teacher" Figure 28: Retrievals for MOPR with query A photo of an architect" Figure 29: Retrievals for k-NN with query A photo of an architect" Figure 30: Retrievals for MMR with query A photo of an architect" Figure 31: Retrievals for CLIP-Clip with query A photo of an architect" Figure 32: Retrievals for PBM with query A photo of an architect" Figure 33: Retrievals for Debias CLIP with query A photo of an architect" H Experiment Notes H.1 GPT Generated Queries. The 10 queries used in experiments and generated by GPT are as follows: [programmer, nurse, architect, scientist, artist, chef, lawyer, teacher, engineer, doctor]. H.2 Experiment Setting and Details. These experiments were conducted mainly on CPU after initially using a single A100 GPU to compute CLIP and Debias CLIP embeddings and run FAISS to compute the top 10k candidates for each. Then, all embeddings are saved so that retrieval across all methods can be run on the CPU. When training linear probes, we conduct a cross-validated hyperparameter search from 0.01 to 100 over ℓ2 regularized logistic regression models from scikit-learn [69] to find the optimal probe. To run MOPR, we use Gurobi [98] with an Academic License to solve Eqn. (9) each iteration of our cutting plane method and use CVXPY [99, 100] to solve our closed-form, quadratic program. To train our oracle estimators, we call scikit-learn s implementations of Linear Regression, Decision Trees with depth 3, and MLPs with one hidden layer of size 64. On average, the runtime for MOPR is approximately 1-5 minutes with 50 queries and 10000 elements in each of our retrieval and curation sets. However, with more features (such as using CLIP embeddings instead of group attributes), this runtime can increase. We note, however, that our most competitive baseline, MMR, is a greedy algorithm that requires 1-2 hours per query to compute. In live image retrieval settings, this runtime is too slow to be of any practical use. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: In this paper, we introduce the novel metric Multi-Group Proportional Representation (MPR) in Section 4, we provide theoretical guarantees in Section 3, we develop a practical algorithm in Section 4 and we numerically show that existing methods optimizing for equal and proportional representation metrics may fail to promote MPR in Section 5 Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: We describe them at the conclusion. First, while MPR is designed to handle a large number of groups, the computational complexity of the proposed cutting-plane algorithm in the case of a general function class C can be a concern for high-dimensional feature spaces. Developing more efficient algorithms or approximation techniques could help address this issue. Second, while we focused on retrieval tasks, adapting MPR to other domains, such as ranking, recommendation systems, and image generation, could broaden its impact and applicability. Future research could also investigate the long-term societal impact of enforcing MPR in real-world retrieval systems to understand its potential benefits and risks. Finally, MPR is inherently limited by the curated dataset. If this dataset is biased against a given group, this bias will be propagated by MOPR. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: We prove all the theorems in the paper in Section C and Section D. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: All the necessary information to reproduce the main results will be available in a Github repository with precise instructions for each figure generated in this paper. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: The datasets Celeb A [38], occupations Celis and Keswani [31], Fairface [39], and UTKFace [40] are all publicly available and the code to generated all the figures presented in the paper and to reproduce all the experiments will be available in a public Github repository. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: The experimental setting is described in H. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: Tables 1, 5 and 6 have deviation bars, and the experiments are averaged over multiple queries. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: The computer resources are described in Appendix H. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: The research conducted in this paper conforms in every respect to the Neur IPS Code of Ethics. The paper aims to develop techniques for improving multi-group representation in image retrieval systems, which aligns with the Code s principles of avoiding discrimination and promoting fairness. The paper does not raise any ethical concerns related to safety, security, privacy, or human rights. The techniques are designed to be applied to standard, publicly available image datasets and do not involve any surveillance, tracking or identifying of individuals. No deception or harassment is involved. In the paper, we include a discussion of broader impacts, both positive and negative. On the positive side, the techniques can help mitigate biases and underrepresentation of certain groups in image search results. This can have significant societal benefits by reducing stereotyping and increasing visibility of marginalized groups. We do not anticipate any direct negative societal impacts, as the methods aim to improve fairness without privileging any groups. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [Yes] Justification: As discussed in the previous answer, we included a discussion of broader impacts, both positive and negative. Regarding potential negative impacts, we acknowledge in the paper that if the reference distribution is biased, then biases will be propagated to the retrieved items. Still, we believe that our method represents an improvement over current approaches that do not consider representational fairness at all. On the positive side, our method has the potential to significantly reduce representational harms in image retrieval systems. By ensuring that results reflect the diversity of the population, we can help combat stereotyping, erasure, and discrimination. This can have wide-ranging societal benefits in shaping perceptions and providing visibility for marginalized groups. We discuss these limitations openly. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: We are not generating any new datasets or models and, therefore, we dont need safeguard. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: We are citing all the publicly available datasets and codes used in the paper. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [Yes] Justification: We have added a repository https://github.com/alex-oesterling/ multigroup-proportional-representation with all the code and instructions to reproduce our experiments. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: [NA] Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: [NA] Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.