# online_learning_with_an_unknown_fairness_metric__f5abe876.pdf Online Learning with an Unknown Fairness Metric Stephen Gillen University of Pennsylvania stepe@math.upenn.edu Christopher Jung Michael Kearns Aaron Roth University of Pennsylvania {chrjung, mkearns, aaroth}@cis.upenn.edu We consider the problem of online learning in the linear contextual bandits setting, but in which there are also strong individual fairness constraints governed by an unknown similarity metric. These constraints demand that we select similar actions or individuals with approximately equal probability [?], which may be at odds with optimizing reward, thus modeling settings where profit and social policy are in tension. We assume we learn about an unknown Mahalanobis similarity metric from only weak feedback that identifies fairness violations, but does not quantify their extent. This is intended to represent the interventions of a regulator who knows unfairness when he sees it but nevertheless cannot enunciate a quantitative fairness metric over individuals. Our main result is an algorithm in the adversarial context setting that has a number of fairness violations that depends only logarithmically on T, while obtaining an optimal O( T) regret bound to the best fair policy. 1 Introduction The last several years have seen an explosion of work studying the problem of fairness in machine learning. Yet there remains little agreement about what fairness should mean in different contexts. In broad strokes, the literature can be divided into two families of fairness definitions: those aiming at group fairness, and those aiming at individual fairness. Group fairness definitions are aggegrate in nature: they partition individuals into some collection of protected groups (say by race or gender), specify some statistic of interest (say, positive classification rate or false positive rate), and then require that a learning algorithm equalize this quantity across the protected groups. On the other hand, individual fairness definitions ask for some constraint that binds on the individual level, rather than only over averages of people. Often, these constraints have the semantics that similar people should be treated similarly ?. Individual fairness definitions have substantially stronger semantics and demands than group definitions of fairness. For example, ? lay out a compendium of ways in which group fairness definitions are unsatisfying. Yet despite these weaknesses, group fairness definitions are by far the most prevalent in the literature (see e.g. ??????? and ? for a survey). This is in large part because notions of individual fairness require making stronger assumptions on the setting under consideration. In particular, the definition from ? requires that the algorithm designer know a task-specific fairness metric. Learning problems over individuals are also often implicitly accompanied by some notion of merit, embedded in the objective function of the learning problem. For example, in a lending setting we might posit that each loan applicant is either creditworthy and will repay a loan, or is not creditworthy and will default which is what we are trying to predict. ? take the approach that this measure of merit already present in the model, although initially unknown to the learner can be taken to be the similarity metric in the definition of ?, requiring informally that creditworthy individuals have at least the same probability of being accepted for loans as defaulting individuals. (The implicit and coarse fairness metric here assigns distance zero between pairs of creditworthy individuals and pairs of defaulting individuals, and some non-zero distance between a creditworthy 32nd Conference on Neural Information Processing Systems (Neur IPS 2018), Montréal, Canada. and a defaulting individual.) This resolves the problem of how one should discover the fairness metric , but results in a notion of fairness that is necessarily aligned with the notion of merit (creditworthiness) that we are trying to predict. However, there are many settings in which the notion of merit we wish to predict may be different or even at odds with the notion of fairness we would like to enforce. For example, notions of fairness aimed at rectifying societal inequities that result from historical discrimination can aim to favor the disadvantaged population (say, in college admissions), even if the performance of the admitted members of that population can be expected to be lower than that of the advantaged population. Similarly, we might desire a fairness metric incorporating only those attributes that individuals can change in principle (and thus excluding ones like race, age and gender), and that further expresses what are and are not meaningful differences between individuals, outside the context of any particular prediction problem. These kinds of fairness desiderata can still be expressed as an instantiation of the definition from ?, but with a task-specific fairness metric separate from the notion of merit we are trying to predict. In this paper, we revisit the individual fairness definition from ?. This definition requires that pairs of individuals who are close in the fairness metric must be treated similarly (e.g. in an allocation problem such as lending, served with similar probability). We investigate the extent to which it is possible to satisfy this fairness constraint while simultaneously solving an online learning problem, when the underlying fairness metric is Mahalanobis but not known to the learning algorithm, and may also be in tension with the learning problem. One conceptual problem with metric-based definitions, that we seek to address, is that it may be difficult for anyone to actually precisely express a quantitative metric over individuals but they nevertheless might know unfairness when they see it. We therefore assume that the algorithm has access to an oracle that knows intuitively what it means to be fair, but cannot explicitly enunciate the fairness metric. Instead, given observed actions, the oracle can specify whether they were fair or not, and the goal is to obtain low regret in the online learning problem measured with respect to the best fair policy while also limiting violations of individual fairness during the learning process. 1.1 Our Results and Techniques We study the standard linear contextual bandit setting. In rounds t = 1, . . . , T, a learner observes arbitrary and possibly adversarially selected d-dimensional contexts, each corresponding to one of k actions. The reward for each action is (in expectation) an unknown linear function of the contexts. The learner seeks to minimize its regret. The learner also wishes to satisfy fairness constraints, defined with respect to an unknown distance function defined over contexts. The constraint requires that the difference between the probabilities that any two actions are taken is bounded by the distance between their contexts. The learner has no initial knowledge of the distance function. Instead, after the learner makes its decisions according to some probability distribution πt at round t, it receives feedback specifying for which pairs of contexts the fairness constraint was violated. Our goal in designing a learner is to simultaneously guarantee near-optimal regret in the contextual bandit problem (with respect to the best fair policy), while violating the fairness constraints as infrequently as possible. Our main result is a computationally efficient algorithm that guarantees this for a large class of distance functions known as Mahalanobis distances (these can be expressed as d(x1, x2) = ||Ax1 Ax2||2 for some matrix A). Theorem (Informal): There is a computationally efficient learning algorithm L in our setting that guarantees that for any Mahalanobis distance, any time horizon T, and any error tolerance ϵ: 1. (Learning) With high probability, L obtains regret O k2d2 log (T) + d T to the best fair policy (See Theorem 3 for a precise statement.) 2. (Fairness) With probability 1, L violates the unknown fairness constraints by more than ϵ on at most O k2d2 log(d/ϵ) many rounds. (Theorem 4.) We note that the quoted regret bound requires setting ϵ = O(1/T), and so this implies a number of fairness violations of magnitude more than 1/T that is bounded by a function growing logarithmically in T. Other tradeoffs between regret and fairness violations are possible. These two goals: of obtaining low regret, and violating the unknown constraint a small number of times are seemingly in tension. A standard technique for obtaining a mistake bound with respect to fairness violations would be to play a halving algorithm , which would always act as if the unknown metric is at the center of the current version space (the set of metrics consistent with the feedback observed thus far) so that mistakes necessarily remove a non-trivial fraction of the version space, making progress. On the other hand, a standard technique for obtaining a diminishing regret bound is to play optimistically i.e. to act as if the unknown metric is the point in the version space that would allow for the largest possible reward. But optimistic points are necessarily at the boundary of the version space, and when they are falsified, the corresponding mistakes do not necessarily reduce the version space by a constant fraction. We prove our theorem in two steps. First, in Section 3, we consider the simpler problem in which the linear objective of the contextual bandit problem is known, and the distance function is all that is unknown. In this simpler case, we show how to obtain a bound on the number of fairness violations using a linear-programming based reduction to a recent algorithm which has a mistake bound for learning a linear function with a particularly weak form of feedback ?. A complication is that our algorithm does not receive all of the feedback that the algorithm of ? expects. We need to use the structure of our linear program to argue that this is ok. Then, in Section 4, we give our algorithm for the complete problem, using large portions of the machinery we develop in Section 3. We note that in a non-adversarial setting, in which contexts are drawn from a distribution, the algorithm of ? could be more simply applied along with standard techniques for contextual bandit learning to give an explore-then-exploit style algorithm. This algorithm would obtain bounded (but suboptimal) regret, and a number of fairness violations that grows as a root of T. The principal advantages of our approach are that we are able to give a number of fairness violations that has only logarithmic dependence on T, while tolerating contexts that are chosen adversarially, all while obtaining an optimal O( T) regret bound to the best fair policy. 1.2 Additional Related Work There are two papers, written concurrently to ours, that tackle orthogonal issues in metric-fair learning. ? consider the problem of generalization when performing learning subject to a known metric constraint. They show that it is possible to prove relaxed PAC-style generalization bounds without any assumptions on the metric, and that for worst-case metrics, learning subject to a metric constraint can be computationally hard, even when the unconstrained learning problem is easy. In contrast, our work focuses on online learning with an unknown metric constraint. Our results imply similar generalization properties via standard online-to-offline reductions, but only for the class of metrics we study. ? considers a group-fairness like relaxation of metric-fairness, asking that on average, individuals in pre-specified groups are classified with probabilities proportional to the average distance between individuals in those groups. They show how to learn such classifiers in the offline setting, given access to an oracle which can evaluate the distance between two individuals according to the metric (allowing for unbiased noise). The similarity to our work is that we also consider access to the fairness metric via an oracle, but our oracle is substantially weaker, and does not provide numeric valued output. There are also several papers in the algorithmic fairness literature that are thematically related to ours, in that they both aim to bridge the gap between group notions of fairness (which can be semantically unsatisfying) and individual notions of fairness (which require very strong assumptions). ? attempt to automatically learn a representation for the data in a batch learning problem (and hence, implicitly, a similarity metric) that causes a classifier to label an equal proportion of two protected groups as positive. They provide a heuristic approach and an experimental evaluation. Two recent papers (? and ?) take the approach of asking for a group notion of fairness, but over exponentially many implicitly defined protected groups, thus mitigating what ? call the fairness gerrymandering problem, which is one of the principal weaknesses of group fairness definitions. Both papers give polynomial time reductions which yield efficient algorithms whenever a corresponding agnostic learning problem is solvable. In contrast, in this paper, we take a different approach: we attempt to directly satisfy the original definition of individual fairness from ?, but with substantially less information about the underlying similarity metric. Starting with ?, several papers have studied notions of fairness in classic and contextual bandit problems. ? study a notion of meritocratic fairness in the contextual bandit setting, and prove upper and lower bounds on the regret achievable by algorithms that must be fair at every round. This can be viewed as a variant of the ? notion of fairness, in which the expected reward of each action is used to define the fairness metric . The algorithm does not originally know this metric, but must discover it through experimentation. ? extend the work of ? to the setting in which the algorithm is faced with a continuum of options at each time step, and give improved bounds for the linear contextual bandit case. ? extend this line of work to the reinforcement learning setting in which the actions of the algorithm can impact its environment. ? consider a notion of fairness based on calibration in the simple stochastic bandit setting. Finally, ? consider a notion of online group fairness in the stochastic contextual bandit setting by constraining how much probability mass can be placed on each pre-specified group of arms. There is a large literature that focuses on learning Mahalanobis distances see ? for a survey. In this literature, the closest paper to our work focuses on online learning of Mahalanobis distances (?). However, this result is in a very different setting from the one we consider here. In ?, the algorithm is repeatedly given pairs of points, and needs to predict their distance. It then learns their true distance, and aims to minimize its squared loss. In contrast, in our paper, the main objective of the learning algorithm is orthogonal to the metric learning problem i.e. to minimize regret in the linear contextual bandit problem, but while simultaneously learning and obeying a fairness constraint, and only from weak feedback noting violations of fairness. 2 Model and Preliminaries 2.1 Linear Contextual Bandits We study algorithms that operate in the linear contextual bandits setting. A linear contextual bandit problem is parameterized by an unknown vector of linear coefficients θ Rd, with ||θ||2 1. Algorithms in this setting operate in rounds t = 1, . . . , T. In each round t, an algorithm L observes k contexts xt 1, . . . , xt k Rd, scaled such that ||xt i||2 1. We write xt = (xt 1, . . . , xt k) to denote the entire set of contexts observed at round t. After observing the contexts, the algorithm chooses an action it. After choosing an action, the algorithm obtains some stochastic reward rt it such that rt it is subgaussian1 and E[rt it] = xt it, θ . The algorithm does not observe the reward for the actions not chosen. When the action it is clear from context, and write rt instead of rt it. Remark 1. For simplicity, we consider algorithms that select only a single action at every round. However, this assumption is not necessary. In the appendix of the full version (?), we show how our results extend to the case in which the algorithm can choose any number of actions at each round. This assumption is sometimes more natural: for example, in a lending scenario, a bank may wish to make loans to as many individuals as will be profitable, without a budget constraint. In this section, we will be discussing algorithms L that are necessarily randomized. To formalize this, we denote a history including everything observed by the algorithm up through but not including round t as ht = ((x1, i1, r1), . . . , (xt 1, it 1, rt 1)) The space of such histories is denoted by Ht = (Rd k [k] R)t 1. An algorithm L is defined by a sequence of functions f 1, . . . , f T each mapping histories and observed contexts to probability distributions over actions: f t : Ht Rd k [k]. We write πt to denote the probability distribution over actions that L plays at round t: πt = f t(ht, xt). We view πt as a vector over [0, 1]k, and so πt i denotes the probability that L plays action i at round t. We denote the expected reward of the algorithm at day t as E[rt] = Ei πt[rt i]. It will sometimes also be useful to refer to the vector of expected rewards across all actions on day t. We denote it as rt = ( xt 1, θ , . . . , xt k, θ ). 2.2 Fairness Constraints and Feedback We study algorithms that are constrained to behave fairly in some manner. We adapt the definition of fairness from ? that asserts, informally, that similar individuals should be treated similarly . We imagine that the decisions that our contextual bandit algorithm L makes correspond to individuals, and that the contexts xt i correspond to features pertaining to individuals. We adopt the following (specialization of) the fairness definition from Dwork et al, which is parameterized by a distance function d : Rd Rd R. 1A random variable X with µ = E[X] is sub-gaussian, if for all t R, E[et(X µ)] e t2 Definition 1 (?). Algorithm L is Lipschitz-fair on round t with respect to distance function d if for all pairs of individuals i, j: |πt i πt j| d(xt i, xt j). For brevity, we will often just say that the algorithm is fair at round t, with the understanding that we are always talking about this one particular kind of fairness. Remark 2. Note that this definition requires a fairness constraint that binds between individuals at a single round t, but not between rounds t. This is for several reasons. First, at a philosophical level, we want our algorithms to be able to improve with time, without being bound by choices they made long ago before they had any information about the fairness metric. At a (related) technical level, it is easy to construct lower bound instances that certify that it is impossible to simultaneously guarantee that an algorithm has diminishing regret to the best fair policy, while violating fairness constraints (now defined as binding across rounds) a sublinear number of times. One of the main difficulties in working with Lipschitz fairness (as discussed in ?) is that the distance function d plays a central role, but it is not clear how it should be specified. In this paper, we concern ourselves with learning d from feedback. In particular, algorithms L will have access to a fairness oracle, which models a regulator who knows unfairness when he sees it . Informally, the fairness oracle will take as input: 1) the set of choices available to L at each round t, and 2) the probability distribution πt that L uses to make its choices at round t, and returns the set of all pairs of individuals for which L violates the fairness constraint. Definition 2 (Fairness Oracle). Given a distance function d, a fairness oracle Od is a function Od : Rd k [k] 2[k] [k] defined such that: Od(xt, πt) = {(i, j) : |πt i πt j| > d(xt i, xt j)} Formally, algorithms L in our setting will operate in the following environment: 1. An adversary fixes a linear reward function θ Rd with ||θ|| 1 and a distance function d. L is given access to the fairness oracle Od. 2. In rounds t = 1 to T: (a) The adversary chooses contexts xt Rd k with ||xt i|| 1 and gives them to L. (b) L chooses a probability distribution πt over actions, and chooses action it πt. (c) L receives reward rt it and observes feedback Od(πt) from the fairness oracle. Because of the power of the adversary in this setting, it is not possible to avoid arbitrarily small violations of the fairness constraint. Instead, we will aim to limit significant violations. Definition 3. Algorithm L is ϵ-unfair on pair (i, j) at round t with respect to distance function d if |πt i πt j| > d(xt i, xt j) + ϵ. Given a sequence of contexts and a history ht (which fixes the distribution on actions at day t) We write Unfair(L, ϵ, ht) = Pk 1 i=1 Pk j=i+1 1(|πt i πt j| > d(xt i, xt j) + ϵ) to denote the number of pairs on which L is ϵ-unfair at round t. Given a distance function d and a history h T +1, the ϵ-fairness loss of an algorithm L is the total number of pairs on which it is ϵ-unfair: Fairness Loss(L, h T +1, ϵ) = PT t=1 Unfair(L, ϵ, ht) For a shorthand, we write Fairness Loss(L, T, ϵ). We will aim to design algorithms L that guarantee that their fairness loss is bounded with probability 1 in the worst case over the instance: i.e. in the worst case over both θ and x1, . . . , x T , and in the worst case over the distance function d (within some allowable class of distance functions see Section 2.4). 2.3 Regret to the Best Fair Policy In addition to minimizing fairness loss, we wish to design algorithms that exhibit diminishing regret to the best fair policy. We first define a linear program that we will make use of throughout the paper. Given a vector a Rd and a vector c Rk2, we denote by LP(a, c) the following linear program: maximize π={p1,...,pk} subject to |pi pj| ci,j, (i, j) We write π(a, c) [k] to denote an optimal solution to LP(a, c). Given a set of contexts xt, recall that rt is the vector representing the expected reward corresponding to each context (according to the true, unknown linear reward function θ). Similarly, we write dt to denote the vector representing the set of distances between each pair of contexts i, j (according to the true, unknown distance function d): dt i,j = d(xt i, xt j). Observe that π( rt, dt) corresponds to the distribution over actions that maximizes expected reward at round t, subject to satisfying the fairness constraints i.e. the distribution that an optimal player, with advance knowledge of θ would play, if he were not allowed to violate the fairness constraints at all. This is the benchmark with respect to which we define regret: Definition 4. Given an algorithm L (f1, . . . , f T ), a distance function d, a linear parameter vector θ, and a history h T +1 (which includes a set of contexts x1, . . . , x T ), its regret is defined to be: Regret(L, θ, d, h T +1) = t=1 E i π( rt, dt) [ rt i] t=1 E i f t(ht,xt)[ rt i] Our goal will be to design algorithms for which we can bound regret with high probability over the randomness of h T +1 in the worst case over θ, d, and (x1, . . . , x T ). 2.4 Mahalanobis Distance In this paper, we will restrict our attention to a special family of distance functions which are parameterized by a matrix A: Definition 5 (Mahalanobis distances). A function d : Rd Rd R is a Mahalanobis distance function if there exists a matrix A such that for all x1, x2 Rd: d(x1, x2) = ||Ax1 Ax2||2 where || ||2 denotes Euclidean distance. Note that if A is not full rank, then this does not define a metric but we will allow this case (and be able to handle it in our algorithmic results). Mahalanobis distances will be convenient for us to work with, because squared Mahalanobis distances can be expressed as follows: d(x1, x2)2 = ||Ax1 Ax2||2 2 = A(x1 x2), A(x1 x2) = (x1 x2) A A(x1 x2) = i,j=1 Gi,j(x1 x2)i(x1 x2)j where G = A A. Observe that when x1 and x2 are fixed, this is a linear function in the entries of the matrix G. We will use this property to reason about learning G, and thereby learning d. 3 Warmup: The Known Objective Case In this section, we consider an easier case of the problem in which the linear objective function θ is known to the algorithm, and the distance function d is all that is unknown. In this case, we show via a reduction to an online learning algorithm of ?, how to simultaneously obtain a logarithmic regret bound and a logarithmic (in T) number of fairness violations. The analysis we do here will be useful when we solve the full version of our problem (in which θ is unknown) in Section 4. Here, we sketch our solution. Details are in the full version of the paper (?). 3.1 Outline of the Solution Recall that since we know θ, at every round t after seeing the contexts, we know the vector of expected rewards rt that we would obtain for selecting each action. Our algorithm will play at each round t the distribution π( rt, ˆdt) that results from solving the linear program LP( rt, ˆdt), where ˆdt is a guess for the pairwise distances between each context dt. (Recall that the optimal distribution to play at each round is π( rt, dt).) The main engine of our reduction is an efficient online learning algorithm for linear functions recently given by ?. Their algorithm, which we refer to as Distance Estimator, works in the following setting. There is an unknown vector of linear parameters α Rm. In rounds t, the algorithm observes a vector of features ut Rm, and produces a prediction gt R for the value α, ut . After it makes its prediction, the algorithm learns whether its guess was too large or not, but does not learn anything else about the value of α, ut . The guarantee of the algorithm is that the number of rounds in which its prediction is off by more than ϵ is bounded by O(m log(m/ϵ))2. Our strategy will be to instantiate k 2 copies of this distance estimator one for each pair of actions to produce guesses ( ˆdt i,j)2 intended to approximate the squared pairwise distances d(xt i, xt j)2. From this we derive estimates ˆdt i,j of the pairwise distances d(xt i, xt j). Note that this is a linear estimation problem for any Mahalanobis distance, because by our observation in Section 2.4, a squared Mahalanobis distance can be written as a linear function of the m = d2 unknown entries of the matrix G = A A which defines the Mahalanobis distance. The complication is that the Distance Estimator algorithms expect feedback at every round, which we cannot always provide. This is because the fairness oracle Od provides feedback about the distribution π( rt, ˆdt) used by the algorithm, not directly about the guesses ˆdt. These are not the same, because not all of the constraints in the linear program LP( rt, ˆdt) are necessarily tight it may be that |π( rt, ˆdt)i π( rt, ˆdt)j| < ˆdt i,j. For any copy of Distance Estimator that does not receive feedback, we can simply roll back its state and continue to the next round. But we need to argue that we make progress that whenever we are ϵ-unfair, or whenever we experience large per-round regret, then there is at least one copy of Distance Estimator that we can give feedback to such that the corresponding copy of Distance Estimator has made a large prediction error, and we can thus charge either our fairness loss or our regret to the mistake bound of that copy of Distance Estimator. As we show, there are three relevant cases. 1. In any round in which we are ϵ-unfair for some pair of contexts xt i and xt j, then it must be that ˆdt i,j d(xt i, xt j)+ϵ, and so we can always update the (i, j)th copy of Distance Estimator and charge our fairness loss to its mistake bound. 2. For any pair of arms (i, j) such that we have not violated the fairness constraint, and the (i, j)th constraint in the linear program is tight, we can provide feedback to the (i, j)th copy of Distance Estimator (its guess was not too large). There are two cases. Although the algorithm never knows which case it is in, we handle each case separately in the analysis. (a) For every constraint (i, j) in LP( rt, ˆdt) that is tight in the optimal solution, | ˆdt i,j d(xt i, xt j)| ϵ. In this case, we show that our algorithm does not incur very much per round regret. (b) Otherwise, there is a tight constraint (i, j) such that | ˆdt i,j d(xt i, xt j)| > ϵ. In this case, we may incur high per-round regret but we can charge such rounds to the mistake bound of the (i, j)th copy of Distance Estimator. Theorem 1. Fairness Loss(Lknown θ, T, ϵ) O k2d2 log d ||AT A||F Theorem 2. For any time horizon T: Regret(Lknown θ, T) O k2d2 log d ||A A||F Setting ϵ = O(1/(k3T)) yields a regret bound of O(d2 log(||A A||F dk T)). 2If the algorithm also learned whether or not its guess was in error by more than ϵ at each round, variants of the classical halving algorithm could obtain this guarantee. But the algorithm does not receive this feedback, which is why the more sophisticated algorithm of ? is needed. 4 The Full Algorithm 4.1 Outline of the Solution At a high level, our plan will be to combine the techniques used for the case where the linear objective θ is known with a standard optimism in the face of uncertainty strategy for learning the parameter vector θ. Our algorithm will maintain a ridge-regression estimate θ together with confidence regions derived in ?. After it observes the contexts xt i at round t, it uses these to derive upper confidence bounds on the expected rewards, corresponding to each context represented as a vector ˆrt. The algorithm continues to maintain distance estimates, ˆdt, the same way as the case where the linear objective θ is known, using ? as a subroutine. At ever round, the algorithm then chooses its action according to the distribution πt = π(ˆrt, ˆdt). The regret analysis of the algorithm follows by decomposing the per-round regret into two pieces. The first can be bounded by the sum of the expected widths of the confidence intervals corresponding to each context xt i that might be chosen at each round t, where the expectation is over the randomness of the algorithm s distribution πt. A theorem of ? bounds the sum of the widths of the confidence intervals corresponding to arms actually chosen by the algorithm. Using a martingale concentration inequality, we are able to relate these two quantities. We show that the second piece of the regret bound can be manipulated into a form that can be bounded using machinery built in section 3, which is described in further details in the full version (?). Theorem 3. For any time horizon T, with probability 1 δ: Regret(Lfull, T) O k2d2 log d ||A A||F If ϵ = 1/k3T, this is a regret bound of O k2d2 log kd T ||A A||F + d Finally, the bound on the fairness loss is identical to the bound we proved in Theorem 1 (because our algorithm for constructing distance estimates ˆd is unchanged). We have: Theorem 4. For any sequence of contexts and any Mahalanobis distance d(x1, x2) = ||Ax1 Ax2||2: Fairness Loss(Lfull, T, ϵ) O k2d2 log d ||A A||F 5 Conclusion and Future Directions We have initiated the study of fair sequential decision making in settings where the notions of payoff and fairness are separate and may be in tension with each other, and have shown that in a stylized setting, optimal fair decisions can be efficiently learned even without direct knowledge of the fairness metric. A number of extensions of our framework and results would be interesting to examine. At a high level, the interesting question is: how much can we further relax the information about the fairness metric available to the algorithm? For instance, what if the fairness feedback is only partial, identifying some but not all fairness violations? What if it only indicates whether or not there were any violations, but does not identify them? What if the feedback is not guaranteed to be exactly consistent with any metric? Or what if the feedback is consistent with some distance function, but not one in a known class: for example, what if the distance is not exactly Mahalanobis, but is approximately so? In general, it is very interesting to continue to push to close the wide gap between the study of individual fairness notions and the study of group fairness notions. When can we obtain the strong semantics of individual fairness without making correspondingly strong assumptions? Yasin Abbasi-Yadkori, Dávid Pál, and Csaba Szepesvári. Improved algorithms for linear stochastic bandits. In Advances in Neural Information Processing Systems 24: 25th Annual Conference on Neural Information Processing Systems 2011. Proceedings of a meeting held 12-14 December 2011, Granada, Spain., pages 2312 2320, 2011. URL http://papers.nips.cc/paper/ 4417-improved-algorithms-for-linear-stochastic-bandits. Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, and Aaron Roth. Fairness in criminal justice risk assessments: the state of the art. ar Xiv preprint ar Xiv:1703.09207, 2017. L Elisa Celis, Sayash Kapoor, Farnood Salehi, and Nisheeth K Vishnoi. An algorithmic framework to control bias in bandit-based personalization. ar Xiv preprint ar Xiv:1802.08674, 2018. Alexandra Chouldechova. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. ar Xiv preprint ar Xiv:1703.00056, 2017. Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference, pages 214 226. ACM, 2012. Sorelle A Friedler, Carlos Scheidegger, and Suresh Venkatasubramanian. On the (im) possibility of fairness. ar Xiv preprint ar Xiv:1609.07236, 2016. Stephen Gillen, Christopher Jung, Michael Kearns, and Aaron Roth. Online learning with an unknown fairness metric. ar Xiv preprint ar Xiv:1802.06936, 2018. Sara Hajian and Josep Domingo-Ferrer. A methodology for direct and indirect discrimination prevention in data mining. IEEE transactions on knowledge and data engineering, 25(7):1445 1459, 2013. Moritz Hardt, Eric Price, and Nathan Srebro. Equality of opportunity in supervised learning. Advances in Neural Information Processing Systems, 2016. Ursula Hébert-Johnson, Michael P Kim, Omer Reingold, and Guy N Rothblum. Calibration for the (computationally-identifiable) masses. ar Xiv preprint ar Xiv:1711.08513, 2017. Shahin Jabbari, Matthew Joseph, Michael Kearns, Jamie Morgenstern, and Aaron Roth. Fairness in reinforcement learning. In International Conference on Machine Learning, pages 1617 1626, 2017. Prateek Jain, Brian Kulis, Inderjit S Dhillon, and Kristen Grauman. Online metric learning and fast similarity search. In Advances in neural information processing systems, pages 761 768, 2009. Matthew Joseph, Michael Kearns, Jamie H Morgenstern, and Aaron Roth. Fairness in learning: Classic and contextual bandits. pages 325 333, 2016a. Matthew Joseph, Michael J. Kearns, Jamie Morgenstern, Seth Neel, and Aaron Roth. Fair algorithms for infinite and contextual bandits. Co RR, abs/1610.09559, 2016b. URL http://arxiv.org/ abs/1610.09559. Faisal Kamiran and Toon Calders. Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems, 33(1):1 33, 2012. Michael Kearns, Seth Neel, Aaron Roth, and Zhiwei Steven Wu. Preventing fairness gerrymandering: Auditing and learning for subgroup fairness. ar Xiv preprint ar Xiv:1711.05144, 2017. Michael P Kim, Omer Reingold, and Guy N Rothblum. Fairness through computationally-bounded awareness. ar Xiv preprint ar Xiv:1803.03239, 2018. Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan. Inherent trade-offs in the fair determination of risk scores. In Proceedings of the 2017 ACM Conference on Innovations in Theoretical Computer Science, Berkeley, CA, USA, 2017, 2017. Brian Kulis et al. Metric learning: A survey. Foundations and Trends R in Machine Learning, 5(4): 287 364, 2013. Yang Liu, Goran Radanovic, Christos Dimitrakakis, Debmalya Mandal, and David C Parkes. Calibrated fairness in bandits. ar Xiv preprint ar Xiv:1707.01875, 2017. Ilan Lobel, Renato Paes Leme, and Adrian Vladu. Multidimensional binary search for contextual decision-making. In Proceedings of the 2017 ACM Conference on Economics and Computation, EC 17, Cambridge, MA, USA, June 26-30, 2017, page 585, 2017. doi: 10.1145/3033274.3085100. URL http://doi.acm.org/10.1145/3033274.3085100. Guy N Rothblum and Gal Yona. Probably approximately metric-fair learning. ar Xiv preprint ar Xiv:1803.03242, 2018. Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, and Krishna P Gummadi. Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In Proceedings of the 26th International Conference on World Wide Web, pages 1171 1180. International World Wide Web Conferences Steering Committee, 2017. Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. Learning fair representations. In International Conference on Machine Learning, pages 325 333, 2013.