# adaptive_data_debiasing_through_bounded_exploration__033bfdcb.pdf Adaptive Data Debiasing through Bounded Exploration Yifan Yang Ohio State University yang.5483@osu.edu Yang Liu University of California, Santa Cruz yangliu@ucsc.edu Parinaz Naghizadeh Ohio State University naghizadeh.1@osu.edu Biases in existing datasets used to train algorithmic decision rules can raise ethical and economic concerns due to the resulting disparate treatment of different groups. We propose an algorithm for sequentially debiasing such datasets through adaptive and bounded exploration in a classification problem with costly and censored feedback. Exploration in this context means that at times, and to a judiciouslychosen extent, the decision maker deviates from its (current) loss-minimizing rule, and instead accepts some individuals that would otherwise be rejected, so as to reduce statistical data biases. Our proposed algorithm includes parameters that can be used to balance between the ultimate goal of removing data biases which will in turn lead to more accurate and fair decisions, and the exploration risks incurred to achieve this goal. We analytically show that such exploration can help debias data in certain distributions. We further investigate how fairness criteria can work in conjunction with our data debiasing algorithm. We illustrate the performance of our algorithm using experiments on synthetic and real-world datasets. 1 Introduction Data-driven algorithmic decision making is being adopted widely to aid humans decisions, in applications ranging from loan approvals to determining recidivism in courts. Despite their ability to process vast amounts of data and make accurate predictions, these algorithms can also exhibit and amplify existing social biases (e.g., [11, 23, 33]). There are at least two possible sources of unfairness in algorithmic decision rules: (data) biases in the training datasets, and (prediction) biases arising from the algorithm s decisions [29]. The latter problem has been receiving increasing attention, and is often addressed by imposing fairness constraints on the algorithm. In contrast, in this paper, we are primarily focused on the former problem of statistical biases in the training data itself. The datasets used for training machine learning algorithms might not accurately represent the agents they make decisions on, due to, e.g., historical biases in decision making and feature selection, or changes in the populations characteristics or participation rates since the data was initially collected. Such data biases in turn can result in disparate treatment of underrepresented or disadvantaged groups; i.e., data bias can cause prediction/model bias, as also verified by recent work [18, 40, 43, 25]. Motivated by this, we focus on data biases, and propose an algorithm which, while attempting to make accurate (and fair) decisions, also aims to collect data in a way that helps it recover unbiased estimates of the characteristics of agents interacting with it. In particular, we study a classification problems with censored and costly feedback. Censored feedback means that the decision maker only observes the true qualification state of those individuals it admits (e.g., a bank will only observe whether an individual defaults on or repays a loan if the loan is extended in the first place; an employer only assesses the performance of applicants it hires). In such settings, any mismatch between the available training data and the true population may grow over time due to adaptive sampling bias: once a decision rule is adopted based on the current training 36th Conference on Neural Information Processing Systems (Neur IPS 2022). data, the algorithm s decisions will impact new data collected in the future, in that only agents passing the requirements set by the current decision rule will be admitted going forward. In response, the decision maker may attempt to collect more data from the population; however, such data collection is costly (e.g., in the previous examples, may require extending loans to/hiring unqualified individuals). Given these challenges, we present an active debiasing algorithm with bounded exploration: our algorithm admits some agents that would otherwise be rejected (i.e., it explores), yet adaptively and judiciously limits the extent and frequency of this exploration. Formally, consider a population of agents with features x, true qualification/labels y, and group memberships g based on their demographic features. To design a (fair) algorithm that can minimize classification loss, the decision maker (implicitly) relies on estimates ˆf y g,t(x) of the feature-label distribution of agents from group g, obtained from the current training dataset Ht = {(xn, yn, gn)}Nt n=1. However, the resulting assumed distribution ˆf y g,t(x) may be different from the true underlying distribution f y g (x); this is the statistical data bias issue we focus on herein. Specifically, we consider distribution shifts between the estimates and the true distributions (Assumption 1). Our algorithm. We propose an active debiasing algorithm (Algorithm 1), which actively adjusts its decisions with the goal of ensuring unbiased estimates of the underlying distributions f y g (x) over time. In particular, at each time t, the algorithm selects a (fairness-constrained) decision rule that would minimize classification error based on its current, possibly biased estimates ˆf y g,t(x); adopting this decision rule corresponds to exploitation of the current information by the algorithm. At the same time, to circumvent the censored feedback nature of the problem, our algorithm also deviates from the prescriptions of this loss-minimizing classifier to a judiciously chosen extent (the extent is chosen adaptively, based on the current estimates); this will constitute exploration. Our algorithm includes two parameters to limit the costs of this exploration: one modulates the frequency of exploration (an exploration probability t which can be adjusted using current bias estimates), and another limits the depth of exploration (by setting a threshold LBt on how far from the classifier one is willing to go when exploring). We show that these choices can strike a balance between the ultimate goal of removing statistical biases in the training data which will in turn lead to more accurate and fair decisions, and the cost of exploration incurred to achieve this goal. Summary of findings and contributions. Our main findings and contributions are as follows: 1. Comparison with baselines. We contrast our proposed algorithm against two baselines: an exploitation-only baseline (one that does not include any form of exploration), and a pure exploration baseline (which may randomly accept some of the agents rejected by the classifier, but does not bound exploration). We show (Theorem 1) that exploitation-only always leads to overestimates of the underlying distributions. Further, while pure exploration can debias the distribution estimates in the long-run (Theorem 2), it does so at the expense of accepting any agent, no matter how far from the classifier s threshold, leading to more costly exploration (Section 5). 2. Analytical support for our proposed algorithm. We show (Theorem 3) that our proposed active debiasing algorithm with bounded exploration can correct biases in unimodal distribution estimates. We also provide an error bound for our algorithm (Theorem 4). 3. Interplay with fairness criteria. We analyze the impact of fairness constraints on our algorithm s performance, and show (Proposition 1) that existing fairness criteria may speed up debiasing of the data in one group, while slowing it down for another. 4. Numerical experiments. We provide numerical support for the performance of our algorithm using experiments on synthetic and real-world (Adult and FICO) datasets. Related work. Our paper is closely related to the works of [4, 20, 14, 6, 17], which study the impact of data biases on (fair) algorithmic decision making. Among these works, Bechavod et al. [4] and Kilbertus et al. [20] study fairness-constrained learning in the presence of censored feedback. While these works also use exploration, the form and purpose of exploration is different: the algorithm in [4] starts with a pure exploration phase, and subsequently explores with the goal of ensuring the fairness constraint is not violated; the stochastic (or exploring) policies in [20] conduct (pure) exploration to address the censored feedback issue. In contrast, we start with a biased dataset, and conduct bounded exploration to debias data; fairness constraints may or may not be enforced separately and are orthogonal to our debiasing process. Also, as shown in Section 5, such pure exploration processes can incur higher exploration costs than our proposed bounded exploration algorithm. Our work is also closely related to [10, 31, 30, 41], which study adaptive sampling biases induced by a decision rule, particularly when feedback is censored. Among these, Neel and Roth [30] also consider an adaptive data gathering procedure, and show that no debiasing will be necessary if the data is collected through a differentially private method. We similarly propose an adaptive debiasing algorithm, but unlike [30], account for the costs of exploration in our data collection procedure. The recent work of Wei [41] studies data collection in the presence of censored feedback, and similar to our work, accounts for the cost of exploration in data collection, by formulating the problem as a partially observable Markov decision processes. Using dynamic programming methods, the data collection policy is shown to be a threshold policy that becomes more stringent (in our terminology, reduces exploration) as learning progresses. Our works are similar in that we both propose using adaptive and cost-sensitive exploration, but we differ in the problem setup and our analysis of the impact of fairness constraints. More importantly, in contrast to both [30, 41], our starting point is a biased dataset (which may be biased for reasons other than adaptive sampling in its collection, including historical biases); we then show how, while attempting to debias this dataset by collecting new data, any additional adaptive sampling bias during data collection can be prevented. Our work also falls within the fields of selective labeling bias, fair learning, and active learning. From the selective labeling bias perspective, Lakkaraju et al. [22] propose a contraction technique to compare the performance of the predictive model and a human judge while they are forced to have the same acceptance rate. De-Arteaga et al. [9] propose a data augmentation scheme by adding more samples that are more likely to be rejected (we refer to this as exploration) to correct the sample selection bias. From the fair learning perspective, Kallus and Zhou [18] propose a re-weighting technique (re-weighting ideas are also explored in [1, 6, 17]) to solve the residual unfairness issue while accounting for adaptive sampling bias. From the active learning perspective, Noriega-Campero et al. [32] adaptively acquire additional information according to the needs of different groups or individuals given information budgets, to achieve fair classification. Similar to the approaches of these papers, we also compensate for adaptive sampling bias through exploration; the main difference, aside from the application, is in our analytical guarantees as well as our study of the interplay of data debiasing with fairness constraints. More broadly, our work has similarities to Bandit learning and its focus on exploration-exploitation trade-offs. A key difference of our work with existing bandit algorithms ( -greedy, UCB, EXP3, etc.) is our focus on bounded exploration. We provide additional discussion on this, and review other related works [35, 3, 19, 26, 27, 42, 22, 9, 18, 32, 1] in more detail, in Appendix B. 2 Model and Preliminaries The environment. We consider a firm or decision maker, who selects an algorithm to make decisions on a population of agents. The firm observes agents arriving over times t = 1, 2, . . . , makes a decision for agents arriving at time t based on the current algorithm, and can subsequently adjust its algorithm for times t + 1 onward based on the observed outcomes. Each agent has an observable feature or score x 2 X R.1 These represent the agent characteristics that are leveraged by the firm in its decision; examples include credit scores or exam scores. Each agent is either qualified or unqualified to receive a favorable decision; this is captured by the agent s true label or qualification state y 2 {0, 1}, with y = 1 and y = 0 denoting qualified and unqualified agents, respectively. In addition, each agent in the population belongs to a different group based on its demographic or protected attributes (e.g., race, gender); the agent s group membership is denoted g 2 {a, b}. We consider threshold-based, group-specific, binary classifiers h g,t(x) : X ! {0, 1} as (part of) the algorithm adopted by the firm, where g,t denotes the classifier s decision threshold. An agent from group g with feature x arriving at time t is admitted if x g,t. Quantifying bias. Let f y g (x) denote the true underlying probability density function for the feature distribution of agents from group g with qualification state y. The algorithm has an estimate of these unknown distributions, at each time t, based on the data collected so far (or an initial training set). Denote the algorithm s estimate at t by ˆf y g,t(x). In general, there can be a mismatch between the estimates ˆf y g,t(x) and the true f y g (x); this is what we refer to as bias. We assume the following. 1We use a one-dimensional feature setting in our analysis, and generalize to X Rn in Section 5. Discussions and numerical experiments on potential loss of information due to our feature dimension reduction technique is given in Appendix A. Assumption 1. The firm updates its estimates ˆf y g,t(x) by updating a single parameter ˆ!y This type of assumption is common in the multi-armed bandit learning literature [38, 39, 34, 24, 36] (there, the algorithm aims to learn the mean arm rewards). In our setting, it holds when the assumed underlying distribution is single-parameter, or when only one of the parameters of a multi-parameter distribution is unknown. Alternatively, it can be interpreted as identifying and correcting distribution shifts by updating a reference point in the distribution (e.g., adjusting the mean).2 More specifically, we will let ˆ!y g,t be the -th percentile of ˆf y g,t(x). We discuss potential limitations of Assumption 1 in Appendix A, and present an extension to a case with two unknown parameters in Appendix I. Under Assumption 1, the bias can be captured by the mismatch between the estimated and true parameters ˆ!y g. In particular, we set the mean absolute error E[|ˆ!y g|] as the measure for quantifying bias, where the randomness is due to that in ˆ!y g,t, the estimate of the unknown parameter based on data collected up to time t. Algorithm choice without debiasing. Let y g be the fraction of group g agents with label y. A loss-minimizing fair algorithm selects its thresholds g,t at time t as follows: min a,t, b,t g,t(x)dx + 0 g,t(x)dx, s.t. C( a,t, b,t) = 0 . (1) Here, the objective is the misclassification error, and C( a, b) = 0 is the fairness constraint imposed by the firm, if any. For instance, C( a,t, b,t) = a,t b,t for same decision rule, or C( a,t, b,t) = R 1 b,t(x)dx for equality of opportunity. Note that both the objective function and the fairness constraint are affected by any inaccuracies in the current estimates ˆf y g,t. As such, a biased training dataset can lead to both loss of accuracy and loss in desired fairness. 3 An Active Debiasing Algorithm with Bounded Exploration In this section, we present the active debiasing algorithm which uses both exploitation (the decision rules of (1)) and exploration (some deviations) to remove any biases from the estimates ˆf y g,t. Although the deviations may lead to admission of some unqualified agents, they can be beneficial to the firm in the long-run: by reducing biases in ˆf y g,t, both classification loss estimates and fairness constraint evaluations can be improved. In this section, we drop the subscripts g from the notation; when there are multiple groups, our algorithm can be applied to each group s estimates separately. As noted in Section 1, our algorithm is one of bounded exploration: it includes a lower bound LBt, which captures the extent to which the decision maker is willing to deviate from the current classifier t, based on its current estimate ˆf 0 t of the unqualified agents underlying distribution. Formally, Definition 1. At time t, the firm selects a lower bound LBt such that LBt = ( ˆF 0 t ) 1(2 ˆF 0 where t is the (current) loss-minimizing threshold determined from (1), ˆF 0 t ) 1 are the cdf and inverse cdf of the estimated distribution ˆf 0 t , respectively, and ˆ!0 t is (wlog) the -th percentile of ˆf 0 In more detail, we choose LBt such that ˆF 0 t (LBt) = ˆF 0 t ( t) ˆF 0 t ); that is, such that ˆ!0 t is the median in the interval (LBt, t) based on the current estimate of the distribution ˆF 0 t at the beginning of time t. Then, once a new batch of data is collected, we update ˆ!0 t+1, the realized median of the distribution between (LBt, t) based on the data observed during [t, t + 1). Once the underlying distribution is correctly estimated, (in expectation) we will observe the same number of samples between (LBt, !0 t ) and between (!0 t , t), and hence !0 t will no longer change. We also note that by selecting a high -th percentile in the above definition, LBt can be increased so as to limit the depth of exploration. As shown in Theorem 3, and in our numerical experiments, these thresholding choice will enable debiasing of the distribution estimates while controlling its costs. Our active debiasing algorithm is summarized below. A pseudo-code is given in Appendix C. 2For instance, a bank may want to adjust for increases in average credit scores [15, 7] over time. Algorithm 1 (The active debiasing algorithm). Denote the loss-minimizing decision threshold determined from (1) by t, and let LBt be given by Definition 1. Let { t} be a sequence of exploration probabilities. At each time t, and for agents (x , y ) arriving at t: Step I: Admit agents and collect data. Admit all agents with x t. Additionally, if LBt x < t, admit the agent with probability t. Step II: Update the distribution estimates based on new data collected in Step I. Qualified agents distribution update: Identify new data with LBt x and y = 1. Use all such x with LBt x < t, and such x with t x with probability t, to update ˆ!1 t . Unqualified agents distribution update: Identify new data with LBt x and y = 0. Use all such x with LBt x < t, and such x with t x with probability t, to update ˆ!0 In more detail, our algorithm repeatedly performs the following two steps: Step I: Data collection. At the beginning of a time period t, a loss-minimizing classifier with threshold t (according to (1)) and the exploration lower bound LBt (Definition 1) are selected based on the data collected so far. Then, given t, the new data collected during period t will consist of arriving agents with features x t. Additionally, to address the censored feedback issues, with probability t, the algorithm will also accept agents with LBt x < t. Note that this step balances between exploration and exploitation through its choice of both LBt (which limit the depth of exploration) and exploration probabilities t (which limits the frequency of exploration). Step II: Updating estimates. At the end of period t, the data collected in Step I will be used to update ˆf 0 t . Under Assumption 1, the estimates ˆf y t are updated by updating the parameter ˆ!y t . We assume, without loss of generality, that the firm sets ˆ!y t to the -th percentile of ˆf y t . This -th percentile is the reference point that will be adjusted over time as new data is collected. As an example, when the reference point ˆ!1 t is set to the median (the 50-th percentile), the parameter can be adjusted so that half the label 1 data collected in Step I will lie on each side of the reference point. 4 Theoretical Analysis We begin by analyzing two baselines: exploitation-only (which only accepts agents with x t, and uses no exploration or thresholding) and pure exploration (which accepts arriving agents at time t who have x < t with probability t, without setting any lower bound). The motivation for the choice of these two baselines is as follows: the exploitation-only baseline tracks the performance of a decision maker who is unaware of underlying data biases, and makes no attempt at fixing them. The pure exploration baseline, on the other hand, is motivated by the Bandit learning literature, and is also akin to debiasing algorithms proposed in recent work (see Section 1, Related Work). We, in contrast, propose and show the benefits of bounded exploration through our active debiasing algorithm. 4.1 The exploitation-only baseline Our first baseline algorithm only updates its estimates of the underlying distributions based on agents with x t who pass the (current) loss-minimizing classifier (1). The following result shows that this approach consistently suffers from adaptive sampling bias, ultimately resulting in overestimation of the underlying distributions. Theorem 1. An exploitation-only algorithm overestimates !y, i.e., limt!1 E[ˆ!y t ] > !y, 8y. A detailed proof is given in Appendix D. 4.2 The pure exploration baseline In this second baseline, at each time t, the algorithm may accept any agent with x < t with probability t. The following result establishes that using the data collected this way, the distributions can be debiased in the long-run, if the data collected above the classifier is also sampled with probability t when updating the distributions. Theorem 2. Using the pure exploration algorithm, ˆ!y t ! !y as t ! 1, 8y. The proof follows from assuming (wlog) that the unknown parameter !y being estimated is the distribution s mean (can be generalized to arbitrary statistics under Assumption 1). Then, as we are collecting i.i.d. samples from across the distribution, ˆ!y t can be set to the sample mean of the collected data, and the conclusion follows from the strong law of large numbers. Note also that if all the data above the classifier was considered when making the updates, following similar arguments to those in the proof of Theorem 1, the algorithm would obtain overestimates of the distributions. Lastly, we could equivalently balance data by resampling the exploration data (rather than downsampling the exploitation data), to debias data through this procedure. 4.3 The active debiasing algorithm While pure exploration can successfully debias data in the long-run, it does so at the expense of accepting agents with any x < t. Below, we provide analytical support that our proposed exploration and thresholding procedure in the active debiasing algorithm can still debias data in certain distributions, while limiting the depth of exploration to LBt < x < t. Theorem 3. Let f y and ˆf y t denote the true feature distribution and their estimates at the beginning of time t, with respective -th percentiles !y and ˆ!y t . Assume these are unimodal distributions, t > 0, 8t, and ˆ!0 t , 8t. Then, using the active debiasing algorithm, (a) If ˆ!y t is underestimated (resp. overestimated), then E[ˆ!y t , (resp. E[ˆ!y t ) 8t, 8y. (b) The sequence {ˆ!y t } converges, with ˆ!y t ! !y as t ! 1, 8y. We provide a proof sketch for debiasing ˆf 0 t which highlights the main technical challenges addressed in our analysis. The detailed proof is given in Appendix E. Proof sketch: Our proof involves the analysis of statistical estimates ˆ!0 t based on data collected from truncated distributions. In particular, by bounding exploration, our algorithm will only collect data with features x LBt, and can use only this truncated data to build estimates of the unknown parameter of the distributions. Part (a) establishes that the sequence of {ˆ!y t } produced by our active debiasing algorithm moves in the right direction over time, and ultimately converges. The main challenge in this analysis is that as the exploration and update intervals [LBt, 1) are themselves adaptive, there is no guarantee on the number of samples in each interval, and therefore we need to analyze the estimates in finite sample regimes. To proceed with the analysis, we assume the feature distribution estimates follow unimodel distributions (such as Gaussian, Beta, and the family of alpha-stable distributions) with !0 as reference points. We then consider the expected parameter update following the arrival of a batch of agents; Denote the current left portion in (LBt, ˆ!0 t ) as p1 := t ) ˆ F 0(LBt) ˆ F 0( t) ˆ F 0(LBt) . Based on Definition 1, we can also obtain the current portion in (ˆ!0 t , t) denoted as p2 := ˆ F 0( t) ˆ F 0(ˆ!0 t ) ˆ F 0( t) ˆ F 0(LBt) = p1. The new expected estimates E[ˆ!0 t+1] is the sample median in (LBt, t), where samples come from the true distribution. We establish that this expected update will be higher/lower than !0 t if the current estimate is an under/over estimate of the true parameter. Then, in Part (b) we first show that the sequence of overand under-estimation errors in {ˆ!y t } relative to the true parameter !y are supermartingales. By the Doobs Convergence theorem and using results from part (a), these will converge to zero mean random variables with variance going to zero as the number of samples increases. This establishes that {ˆ!y t } converges. It remains to show that this convergence point is the true parameter of the distribution. To do so, as detailed in the proof, we note that the density function of the sample median estimated on label 0 data collected in [LBt, t] is t = )d = (2m + 1)! m!m! ( F 0( ) F 0(LBt) F 0( t) F 0(LBt))m( F 0( t) F 0( ) F 0( t) F 0(LBt))m f 0( ) F 0( t) F 0(LBt)d (2) which is a beta distribution pushed forward by H( ) := F 0( ) F 0(LBt) F 0( t) F 0(LBt); this is the CDF of the truncated F 0 distribution in [LBt, t]. We then establish that the convergence point will be the true median of the underlying distribution. 4.4 Error bound analysis Our error bound analysis compares the errors (measured as the number of wrong decisions made) of our adaptive debiasing algorithm against the errors that would be made by an oracle which knows the true underlying distributions. We measure the performance using 0-1 loss, ( ˆyi, yi) = [ ˆyi 6= yi], where ˆyi and yi denote the predicted and true label of agent i, respectively. We consider the error accumulated when updating the estimates using a total of m batches of data. We split the total T samples that have arrived during [t, t + 1) into four groups, corresponding to four different distributions f y g . Specifically, we use by g,t to denote the number of samples from each label-group pair at round t 2 {0, . . . , m}. We update the unknown distribution estimates once all batches meet a size requirement s, i.e, once min(by g,t) s, 8y, 8g. The error of our algorithm is given by: Error = E[Error Adaptive Error Oracle] E (xi,yi,gi) D (h t,g(xi, gi), yi) E (xi,yi,gi) D g(xi, gi), yi) The following theorem provides an upper bound on the error incurred by active debiasing. Theorem 4. Let ˆf y g,t(x) be the estimated feature-label distributions at round t 2 {0, . . . , m}. We consider the threshold-based, group-specific, binary classifier h g,t, and denote the Rademacher complexity of the classifier family H with n training samples by Rn(H). Let g,t be a v-approximately optimal classifier based on data collected up to time t. At round t, let Ng,t be the number of exploration errors incurred by our algorithm, ng,t be the sample size at time t from group g, d H H( Dg,t, Dg) be the distance between the true unbiased data distribution Dg and the current biased estimate Dg,t, and c( Dg,t, Dg) be the minimum error on an algorithm trained on unbiased and biased data. Then, with probability at least 1 4δ with δ > 0, the active debiasing algorithm s error is bounded by: 2v |{z} v-approx. + 4Rng,t(H) + 4 png,t + ng,t | {z } empirical estimation errors + Ng,t |{z} explor. + d H H( Dg,t, Dg) + 2c( Dg,t, Dg) | {z } source-target distribution mismatch More details on the definitions of the distance measure d H H, and the error term c( ), and the exploration error term Ng,t, along with a a detailed proof, are given in Appendix F. From the expression above, we can see that the error incurred by our algorithm consists of four types of error: errors due to approximation of the optimal (fair) classifier at each round, empirical estimation errors, exploration errors, and errors due to our biased training data (viewed as source-target distribution mismatches); the latter two are specific to our active debiasing algorithm. In particular, as we collect more samples, ng,t will increase. Hence, the empirical estimation errors decrease over time. Moreover, as the mismatch between Dg,t and Dg decreases using our algorithm (by Theorem 3), the error due to target domain and source domain mismatches also decrease. In the meantime, our exploration probability t also becomes smaller over time, decreasing Ng,t. 4.5 Active debiasing and fairness criteria We next consider our proposed active debiasing algorithm when used in conjunction with demographic fairness constraints (e.g., equality of opportunity, same decision rule, and statistical parity [29]). Imposing such fairness rules will lead to changes to the selected classifiers compared to the fairness-unconstrained case. Let F g,t denote the fairness constrained and unconstrained decision rules obtained from (1) at time t for group g, respectively. We say group g is being overselected (resp. under-selected) following the introduction of fairness constraints if F g,t (resp. F g,t). Below, we show how such over/under-selections can differently affect the debiasing of estimates on different agents. In particular, let the speed of debiasing be the rate at which E[|ˆ!y t !y|] decreases with respect to t; then, for a given t, an algorithm for which this error is larger has a slower speed of debiasing. The following proposition identifies the impacts of different fairness constraints on the speed of debiasing attained by our active debiasing algorithm. The proof is given in Appendix G. Proposition 1. Let f y g,t be the true and estimated feature distributions, with respective -th percentiles !y and ˆ!y t . Assume these are unimodal distributions, and active debiasing is applied. If group g is over-selected (resp. under-selected) under a fairness constraint, i.e., F g,t (resp. F g,t), the speed of debiasing on the estimates ˆf y g,t will decrease (resp. increase). Proposition 1 highlights the following implications of using both fairness rules and our active debiasing efforts. Some fairness constraints (such as equality of opportunity) can lead to an increase in opportunities for (here, over-selection of) agents from disadvantaged groups, while others (such as same decision rule) can lead to under-selection from that group. Proposition 1 shows that active debiasing may in turn become faster or slower at debiasing estimates on this group. Intuitively, over-selection provides increased opportunities to agents from a group (compared to an unconstrained classifier). In fact, the reduction of the decision threshold to F g,t can itself be interpreted as introducing exploration (which is separate from that introduced by our debiasing algorithm). When a group is over-selected under a fairness constraint, the fairness-constrained threshold F g,t will be lower than the unconstrained threshold U g,t. Therefore, the exploration range will be narrower, which means by adding a fairness constraint, the algorithm needs to wait and collect more samples (takes a longer time) before it manages to collect sufficient data to accurately update the unknown distribution parameter, and hence, it has a slower debiasing speed. More broadly, these findings contribute to our understanding of how fairness constraints can have long-term implications beyond the commonly studied fairness-accuracy tradeoff when we consider their impacts on data collection and debiasing efforts. 5 Numerical Experiments In this section, we illustrate the performance of our algorithm through numerical experiments on both Gaussian and Beta distributed synthetic datasets, and on two real-world datasets: the Adult dataset [12] and the FICO credit score dataset [37] pre-processed by [16]. Additional details (groundtruth information) on the experiments, and larger versions of all figures, are available in Appendix H. Our code is available at: https://github.com/Yifankevin/adaptive_data_debiasing. Throughout, we either choose a fixed schedule for reducing the exploration frequencies { t}, or reduce these adaptively as a function of the estimated error. For the latter, the algorithm can select a range (e.g., above the classifier for label 0/1) and adjust the exploration frequency proportional to the discrepancy between the number of observed classification errors in this interval relative to the number expected given the distribution estimates. Comparison with the exploitation-only and pure exploration baselines: Our first experiments in Fig. 1, compare our algorithm against two baselines. The underlying distributions are Gaussian and no fairness constraint is imposed. Our algorithm sets 1 = 50 and 0 = 60 percentiles, and exploration frequencies t are selected adaptively by both our algorithm and pure exploration. (a) Rate of debiasing, f 1 and f 0 underestimated. (b) Rate of debiasing, f 1 and f 0 overestimated. (c) Regret. (d) Weighted regret. Figure 1: Speed of debiasing, regret, and weighted regret, of active debiasing vs. exploitation-only and pure exploration (larger figures in Appendix H). Speed of debiasing: Figs. 1(a) and 1(b) show that consistent with Theorem 1, exploitation-only overestimates the distributions due to adaptive sampling biases. Further, consistent with Theorem 2, pure exploration successfully debiases data. We also observe that as expected, pure exploration debiases faster than active debiasing. The difference is more pronounced in the label 0 distributions compared to label 1, where pure exploration collects more diverse observations than our algorithm. For this same reason, the gap between pure exploration and our algorithm is larger when f 0 is overestimated. This is because pure exploration observes samples with lower features x than active debiasing, and so can use these to reduce its estimate faster. Regret: Figs. 1(c) and 1(d) compare the regret and weighted regret of the algorithms. Regret is measured as the difference between the number of FN+FP decisions of an algorithm vs the oracle loss-minimizing algorithm derived on unbiased data. Formally, regret is defined as in Section 4.4; weighted regret is defined similarly, but also adds a weight to each FN or FP decision, with the weight exponential in the distance of the feature of the admitted agent from the classifier. We observe that exploitation-only s regret is super-linear, as not only it fails to debias, but has increasing error due to biases from overestimating. On the other hand, while algorithms that explore deeper have lower regret (e.g. pure exploration < active debiasing with 0 = 50 < active debiasing with 0 = 60 in Figs. 1(c)), they have higher weighted regret (the order is reversed in Fig. 1(d)). In other words, exploring to admit agents with low features x leads to some errors, but ultimately helps reduce future mistakes, leading to sub-linear regret. However, if the risk/cost of these wrong decisions is taken into account, the firm may be better off adopting slower, but less risky exploration thresholds (e.g. 0 = 70). Figure 2: Debiasing under Beta distributions. Performance of active debiasing on Beta distributions: Fig. 2 shows that our algorithm can debias data for which the underlying feature-label distributions follow Beta distributions. We have assumed a mistmach between the parameter of the true and estimated distributions, and selected these so that the estimated and true distributions have different relative skewness. This verifies that Theorem 3 holds for asymmetric distributions. (a) Advantaged label 0. (b) Disadvantaged label 0. Figure 3: Debiasing used with fairness constraints. Interplay of debiasing and fairness constraints: Fig. 3 compares the performance of active debiasing when there are two groups of agents with underlying Gaussian distributions, and the algorithm is chosen subject to three different fairness settings: no fairness, equality of opportunity (EO), and the same decision rule (SD). The findings are consistent with Proposition 1. For instance, SD will over-select the majority group (i.e., SD a,t) so that, as shown in the left panel in Fig. 3, the speed of debiasing on the estimates ˆf y a,t will decrease. In contrast, an opposite effect will happen in the minority group b which is under-selected (i.e., SD b,t). The effects of EO can be similarly explained by noting that it under-selects the majority group and over-selects the minority group. Active debiasing on the Adult dataset: Fig. 4 illustrates the performance of our algorithm on the Adult dataset. Data is grouped based on race (White Ga and non-White Gb), with labels y = 1 for income > $50k/year. A one-dimensional feature x 2 R is constructed by conducting logistic regression on four quantitative and qualitative features (education number, sex, age, workclass), based on the initial training data.3 Using an input analyzer, we found Beta distributions as the best fit to the underlying distributions. We use 2.5% of the data to obtain a biased estimate of the parameter . The remaining data arrives sequentially. We use 1 = 50 and 0 = 60 and a fixed decreasing { t}, with the equality of opportunity fairness constraint imposed throughout. (a) Debiasing Ga, Adult. (b) Debiasing Gb, Adult. (c) Gb, augmented with synthetic data (Adult). (d) Debiasing on FICO. Figure 4: Active debiasing on the Adult and FICO datasets. We observe that our proposed algorithm can debias estimates across groups and for both labels, but that this happens in the long-run and given access to sufficient samples. In particular, we note that 3While this experiment maintains the same mapping throughout, the mapping could be periodically revised. for label 1 agents from Gb, as there are only 1080 samples in the dataset, although the bias initially decreases, the final estimate still differs from the true value. Fig. 4(c) verifies that this estimate would have been debiased in the long-run, had additional samples from the underlying population become available (i.e., as more such agents arrive). Active debiasing on the FICO dataset: Fig. 4 also illustrates the performance of our algorithm on the FICO dataset [37, 16], and shows that it is successful in debiasing distribution estimates on both groups and on both labels. 6 Conclusion, Limitations, and Future Work We proposed an active debiasing algorithm which recovers unbiased estimates of the underlying data distribution of agents interacting with it over time. We also analyzed the interplay of our proposed statistical/data debiasing effort with existing social/model debiasing efforts, shedding light on the potential alignments and conflicts between these two goals in fair algorithmic decision making. We further illustrated the performance of our proposed algorithm, and its interplay with fairness constraints, through numerical experiments on both synthetic and real-world datasets. The single-unknown parameter assumption. Our work focuses on learning of a single unknown parameter (Assumption 1). Despite the commonality of this assumption in the multi-armed bandit learning literature, it also entails parametric knowledge of the underlying distribution with the other parameters such as variance or spread being known. We extend our algorithm to a Gaussian distribution with two unknown parameters in Appendix I. Extensions beyond this, especially those not requiring parametric assumptions on the underlying distributions, remain a main direction of future work. On one-dimensional features and threshold classifiers. Our analytical results have been focused on one-dimensional feature data and threshold classifiers. These assumptions may not be too restrictive in some cases: the optimality of threshold classifiers has been established in the literature by, e.g., [8, Thm 3.2] and [36], as long as a multi-dimensional feature can be mapped to a properly defined scalar. Moreover, the recent advances in deep learning have helped enable this possibility: one can take the last layer outputs from a deep neural network and use it as the single dimensional representation. That said, any reduction of multi-dimensional features to a single-dimensional score may lead to some loss of information. In particular, our experiments have considered the use of our active debiasing algorithm on the Adult dataset with multi-dimensional features by first performing a dimension reduction to a single-dimensional score; we find that this reduction can lead to a 5% loss in performance (see Appendix A for details). One potential solution to this is to adopt a mapping from high-dimensional features to scores that is revised repeatedly as the algorithm collects more data. Alternatively, one may envision a debiasing algorithm which targets its exploration towards collecting data on features that are believed to be highly biased; these remain as potential extensions of our algorithm. Potential social impacts. More broadly, while our debiasing algorithm imposes fairness constraints on its exploitation decisions (see problem (1)), it does not consider fairness constraints in its exploration decisions. That means that our proposed algorithm could be disproportionate in the way it increases opportunities for qualified or unqualified agents in different groups during exploration. Imposing fairness rules on exploration decisions, as well as identifying algorithms that can improve the speed of debiasing of estimates on underrepresented populations, can be explored to address these potential social impacts, and remain as interesting directions of future work. Additional discussions on limitations, extensions, and social impacts, are given in Appendix A. Acknowledgments and Disclosure of Funding We sincerely thank the three reviewers and the area chair for their comments and feedback which helped improve our paper. We are also grateful for support from the National Science Foundation (NSF) program on Fairness in AI in collaboration with Amazon under Award No. IIS-2040800, the NSF under grant IIS-2143895, and Cisco Research. Any opinion, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the NSF, Amazon, or Cisco. [1] J. Abernethy, P. Awasthi, M. Kleindessner, J. Morgenstern, C. Russell, and J. Zhang. Active sampling for min-max fairness. ar Xiv preprint ar Xiv:2006.06879, 2020. [2] A. Agarwal, A. Beygelzimer, M. Dudík, J. Langford, and H. Wallach. A reductions approach to fair classification. In International Conference on Machine Learning, pages 60 69. PMLR, 2018. [3] M.-F. Balcan, A. Broder, and T. Zhang. Margin based active learning. In International Conference on Computational Learning Theory, 2007. [4] Y. Bechavod, K. Ligett, A. Roth, B. Waggoner, and S. Z. Wu. Equal opportunity in online classification with partial feedback. In Advances in Neural Information Processing Systems, pages 8974 8984, 2019. [5] S. Ben-David, J. Blitzer, K. Crammer, A. Kulesza, F. Pereira, and J. W. Vaughan. A theory of learning from different domains. Machine learning, 79(1):151 175, 2010. [6] A. Blum and K. Stangl. Recovering from biased data: Can fairness constraints improve accuracy? In 1st Symposium on Foundations of Responsible Computing (FORC 2020). Schloss Dagstuhl-Leibniz-Zentrum für Informatik, 2020. [7] CNBC. The average U.S. FICO score is up 8 points from last year. https://www.cnbc.com/select/ heres-how-the-average-american-increased-their-fico-score-last-year/, 2021. [8] S. Corbett-Davies, E. Pierson, A. Feller, S. Goel, and A. Huq. Algorithmic decision making and the cost of fairness. In Proceedings of the 23rd acm sigkdd international conference on knowledge discovery and data mining, pages 797 806, 2017. [9] M. De-Arteaga, A. Dubrawski, and A. Chouldechova. Learning under selective labels in the presence of expert consistency. ar Xiv preprint ar Xiv:1807.00905, 2018. [10] Y. Deshpande, L. Mackey, V. Syrgkanis, and M. Taddy. Accurate inference for adaptive linear models. In International Conference on Machine Learning, pages 1194 1203, 2018. [11] J. Dressel and H. Farid. The accuracy, fairness, and limits of predicting recidivism. Science advances, 4(1): eaao5580, 2018. [12] D. Dua and C. Graff. UCI machine learning repository, 2017. URL http://archive.ics.uci.edu/ml. [13] C. Dwork, M. Hardt, T. Pitassi, O. Reingold, and R. Zemel. Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference, 2012. [14] D. Ensign, S. A. Friedler, S. Neville, C. Scheidegger, and S. Venkatasubramanian. Runaway feedback loops in predictive policing. In Conference on Fairness, Accountability and Transparency, pages 160 171. PMLR, 2018. [15] Experian. What Is the Average Credit Score in the U.S.? https://www.experian.com/blogs/ ask-experian/what-is-the-average-credit-score-in-the-u-s/, 2020. [16] M. Hardt, E. Price, and N. Srebro. Equality of opportunity in supervised learning. In Advances in neural information processing systems, pages 3315 3323, 2016. [17] H. Jiang and O. Nachum. Identifying and correcting label bias in machine learning. In International Conference on Artificial Intelligence and Statistics, pages 702 712, 2020. [18] N. Kallus and A. Zhou. Residual unfairness in fair machine learning from prejudiced data. In International Conference on Machine Learning, pages 2439 2448. PMLR, 2018. [19] A. Kazerouni, Q. Zhao, J. Xie, S. Tata, and M. Najork. Active learning for skewed data sets. ar Xiv preprint ar Xiv:2005.11442, 2020. [20] N. Kilbertus, M. G. Rodriguez, B. Schölkopf, K. Muandet, and I. Valera. Fair decisions despite imperfect predictions. In International Conference on Artificial Intelligence and Statistics, pages 277 287. PMLR, 2020. [21] T. L. Lai and Z. Ying. Estimating a distribution function with truncated and censored data. The Annals of Statistics, pages 417 442, 1991. [22] H. Lakkaraju, J. Kleinberg, J. Leskovec, J. Ludwig, and S. Mullainathan. The selective labels problem: Evaluating algorithmic predictions in the presence of unobservables. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 275 284, 2017. [23] A. Lambrecht and C. Tucker. Algorithmic bias? an empirical study of apparent gender-based discrimination in the display of stem career ads. Management Science, 65(7):2966 2981, 2019. [24] T. Lattimore and C. Szepesvári. Bandit algorithms. Cambridge University Press, 2020. [25] Y. Liao and P. Naghizadeh. Social bias meets data bias: The impacts of labeling and measurement errors on fairness criteria. ar Xiv preprint ar Xiv:2206.00137, 2022. [26] L. T. Liu, S. Dean, E. Rolf, M. Simchowitz, and M. Hardt. Delayed impact of fair machine learning. In International Conference on Machine Learning, pages 3150 3158. PMLR, 2018. [27] L. T. Liu, A. Wilson, N. Haghtalab, A. T. Kalai, C. Borgs, and J. Chayes. The disparate equilibria of algorithmic decision making when individuals invest rationally. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, 2020. [28] J. Maritz and R. Jarrett. A note on estimating the variance of the sample median. Journal of the American Statistical Association, 73(361):194 196, 1978. [29] N. Mehrabi, F. Morstatter, N. Saxena, K. Lerman, and A. Galstyan. A survey on bias and fairness in machine learning. ar Xiv preprint ar Xiv:1908.09635, 2019. [30] S. Neel and A. Roth. Mitigating bias in adaptive data gathering via differential privacy. In International Conference on Machine Learning, pages 3720 3729. PMLR, 2018. [31] X. Nie, X. Tian, J. Taylor, and J. Zou. Why adaptively collected data have negative bias and how to correct for it. In International Conference on Artificial Intelligence and Statistics, pages 1261 1269, 2018. [32] A. Noriega-Campero, M. A. Bakker, B. Garcia-Bulle, and A. Pentland. Active fairness in algorithmic decision making. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, pages 77 83, 2019. [33] Z. Obermeyer, B. Powers, C. Vogeli, and S. Mullainathan. Dissecting racial bias in an algorithm used to manage the health of populations. Science, 366(6464):447 453, 2019. [34] V. Patil, G. Ghalme, V. Nair, and Y. Narahari. Achieving fairness in the stochastic multi-armed bandit problem. J. Mach. Learn. Res., 22:174 1, 2021. [35] J. Perdomo, T. Zrnic, C. Mendler-Dünner, and M. Hardt. Performative prediction. In International Conference on Machine Learning, pages 7599 7609. PMLR, 2020. [36] R. Raab and Y. Liu. Unintended selection: Persistent qualification rate disparities and interventions. Advances in Neural Information Processing Systems, 34:26053 26065, 2021. [37] U. F. Reserve. Report to the congress on credit scoring and its effects on the availability and afford- ability of credit. https://www.federalreserve.gov/boarddocs/rptcongress/creditscore/ creditscore.pdf, 2007. [38] C. Schumann, Z. Lang, N. Mattei, and J. P. Dickerson. Group fairness in bandit arm selection. ar Xiv preprint ar Xiv:1912.03802, 2019. [39] A. Slivkins. Introduction to multi-armed bandits. ar Xiv preprint ar Xiv:1904.07272, 2019. [40] J. Wang, Y. Liu, and C. Levy. Fair classification with group-dependent label noise. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, pages 526 536, 2021. [41] D. Wei. Decision-making under selective labels: Optimal finite-domain policies and beyond. In Interna- tional Conference on Machine Learning, pages 11035 11046. PMLR, 2021. [42] X. Zhang, M. Khaliligarekani, C. Tekin, and M. Liu. Group retention when using machine learning in sequential decision making: the interplay between user dynamics and fairness. In Advances in Neural Information Processing Systems, pages 15269 15278, 2019. [43] Z. Zhu, T. Luo, and Y. Liu. The rich get richer: Disparate impact of semi-supervised learning. In International Conference on Learning Representations, 2021. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] (c) Did you discuss any potential negative societal impacts of your work? [Yes] Fairness implications of our algorithm are discussed in Section 4.5. A limitation of our algorithm for groups with smaller representation are discussed at the end of Section 5. See also Appendix A for more discussion. (d) Did you describe the limitations of your work? [Yes] See the remark in Section 3 and the paragraph before the proof of Theorem 3 in Section 4, and Appendix A. 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] See Section 4. (b) Did you include complete proofs of all theoretical results? [Yes] See Section 4 for the main proofs, and additional proofs in the supplementary material, Appendix E, F, and G. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main ex- perimental results (either in the supplemental material or as a URL)? [Yes] Code is included in the supplementary material. Dataset used are the Adult dataset, publicly available from the UCI machine learning repository [12], and the FICO credit score dataset [37] pre-processed by [16] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Section 5 for details. (c) Did you report error bars (e.g., with respect to the random seed after running ex- periments multiple times)? [Yes] See Figure 2, 4 Adult Dataset, 12(a), 12(b) for details (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [N/A] All experiments were run on a local computer. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] See Section 5. (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] Our algorithm s code is available in the supplementary material. No new data was curated or generated. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] We use the publicly available Adult dataset from the UCI machine learning repository [12], for which data was extracted from a 1994 Census database; and the FICO credit score dataset [37] pre-processed by [16] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] We use the publicly available Adult dataset from the UCI machine learning repository [12], for which data was extracted from a 1994 Census database; and the FICO credit score dataset [37] pre-processed by [16]. They do not contain PII or offensive content. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]