# learning_losses_for_strategic_classification__d13dfa69.pdf Learning Losses for Strategic Classification Tosca Lechner1, Ruth Urner2 1 University of Waterloo, Cheriton School of Computer Science, Waterloo, Canada 2 York University, Lassonde School of Engineering, EECS Department, Toronto, Canada tlechner@uwaterloo.ca, ruth@eecs.yorku.ca Strategic classification, i.e. classification under possible strategic manipulations of features, has received a lot of attention from both the machine learning and the game theory community. Most works focus on analysing properties of the optimal decision rule under such manipulations. In our work we take a learning theoretic perspective, focusing on the sample complexity needed to learn a good decision rule which is robust to strategic manipulation. We perform this analysis by introducing a novel loss function, the strategic manipulation loss, which takes into account both the accuracy of the final decision rule and its vulnerability to manipulation. We analyse the sample complexity for a known graph of possible manipulations in terms of the complexity of the function class and the manipulation graph. Additionally, we initialize the study of learning under unknown manipulation capabilities of the involved agents. Using techniques from transfer learning theory, we define a similarity measure for manipulation graphs and show that learning outcomes are robust with respect to small changes in the manipulation graph. Lastly, we analyse the (sample complexity of) learning of the manipulation capability of agents with respect to this similarity measure, providing novel guarantees for strategic classification with respect to an unknown manipulation graph. 1 Introduction In many scenarios where a decision rule is learned from data, the publication of this decision rule has an effect on the distribution of the underlying population that may harm the quality of the rule. For example, applicants for a loan may change details in their bank account to receive a better score, people may join a gym or sports club without ever intending to participate, in order to get a better health insurance policy, or students may employ different strategies such as registering to volunteer, or joining rare clubs (without attending either) to appear better on college applications. Effects and incentives resulting from strategic behavior in classification scenarios have received substantial attention from both machine learning and game-theoretic perspectives in recent years (Hardt et al. 2016; Milli et al. 2019; Haghtalab et al. 2020; Tsirtsis and Rodriguez 2020; Zhang and Conitzer 2021). Most works study this as a two-player game Copyright 2022, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. between an institution that publishes a decision rule and a population of best responding agents to be classified. Given the classifier, these agents may change their feature representations in order to obtain a more favorable classification outcome. To prevent the induced additional classification error the institution will publish a modified predictor, not transparently reflecting the underlying intent and potentially causing additional harm to sub-populations that may be less equipped to perform the required changes to their representations (Hu, Immorlica, and Vaughan 2019). In this work, we propose a learning theoretic take on this scenario. In machine learning, it is common to model desiderata for a learning outcome in form of a loss function. The goal of the learning process is then to identify a predictor that minimizes this loss in expectation over a datagenerating distribution. Thus, we here define a novel loss function for learning under strategic manipulations. The aim of this loss is to induce a combination of two (potentially competing) requirements: achieving low classification error taking into account that individuals being classified may manipulate their features, and discouraging such feature manipulations overall. Prior work has shown that these may be conflicting requirements (Zhang and Conitzer 2021). Our proposed loss function thus aims to induce a balanced combination of these requirements rather than strictly enforcing one, and then only observing the effect on the other (as is implied by frameworks that aim to minimize classification error under best-responding agents (Hardt et al. 2016; Milli et al. 2019) or enforcing incentive-compatibility (Zhang and Conitzer 2021)). To define our strategic manipulation loss we employ an abstraction of the plausible feature manipulations in form of a manipulation graph (Zhang and Conitzer 2021). An edge x x in this graph indicates that an individual with feature vector x may change their features to present as x if this leads to a positive classification, for example since the utility of this change in classification exceeds the cost of the change between these vectors. We define our strategic loss in dependence of this graph and carefully motivate the proposed loss in terms of requirements and effects from previous literature. We then analyze the sample complexity of learning with this loss function. We identify sufficient conditions for proper learnability that take into account the interplay between a hypothesis class and an underlying ma- The Thirty-Sixth AAAI Conference on Artificial Intelligence (AAAI-22) nipulation graph. Moreover, we show that every class that has finite VC-dimension is learnable with respect to this loss by drawing a connection to results in the context of learning under adversarial perturbations (Montasser, Hanneke, and Srebro 2019). This effect may be surprising, since it presents a contrast to learning VC-classes with the sole requirement of minimizing classification error under strategic feature manipulations, which has been shown can lead to some VC-classes not being learnable (Zhang and Conitzer 2021). Thus, our analysis shows that balancing classification error with disincentivizing feature manipulations can reduce the complexity of the learning problem. Moreover, we show that the quality of learning outcomes under our loss function is robust to inaccuracies in the manipulation graph. Such a robustness property is important, since an assumed graph might not exactly reflect agents responses. In fact, it has recently been argued that the model of best-responding agents is not backed up by empirical observations on agent distributions after strategic responses (Jagadeesan, Mendler-D unner, and Hardt 2021). Moreover, different sub-populations may have differences in their manipulation graphs (different capabilities to manipulate their features) or a manipulation graph may be inferred from data and therefore exhibit statistical errors. We introduce a novel distance measure between manipulation graphs by drawing connections to learning bounds in transfer learning (Ben David et al. 2010; Mansour, Mohri, and Rostamizadeh 2009) and show that the strategic loss of a learned predictor when employing a different manipulation graph can be bounded in terms of this distance measure. Finally, we present some initial results on how manipulation graphs may be learned from data. 1.1 Related Work That learning outcomes might be compromised by agents responding to published classification rules with strategic manipulations of their feature vectors was first pointed out over a decade ago (Dalvi et al. 2004; Br uckner and Scheffer 2011) and has received substantial interest from the research community in recent years initiated by a study by Hardt et al. that differentiated the field from the more general context of learning under adversarial perturbations (Hardt et al. 2016). That study considered strategic responses being induced by separable cost functions for utility maximizing agents and studied the resulting decision boundaries for certain classes of classifiers. Recent years have seen a lot of interest in better understanding the interplay of various incentives in settings where a decision rule is published and thereby has an effect on how the entities that are to be classified might present themselves to the decision-maker. In particular, various externalities to this scenario have been analyzed. A general cost to society formalized in form of a social burden incurred by the costs of enforced feature manipulation, has been shown to occur when institutions anticipate strategic responses (Milli et al. 2019; Jagadeesan, Mendler-D unner, and Hardt 2021). Further, it has been demonstrated how such a burden may be suffered to differing degrees by various subgroups of a population that may differ in their capabilities to adapt their features in ways that are favorable to them (Milli et al. 2019; Hu, Immorlica, and Vaughan 2019), raising concerns over fairness in such scenarios. Recent studies have extended the original game-theoretic model of a classifier publishing intuition and best responding subjects. For example, a recent work studied how strategic modification can also be a positive effect and how that should be taken into consideration by the institution (Haghtalab et al. 2020). Such a perspective has been connected to underlying causal relations between features and classification outcome and resulting strategic recommendations (Miller, Milli, and Hardt 2020; Tsirtsis and Rodriguez 2020). Further, a very recent study has explored how the model of a best responding agent may be relaxed to better reflect empirically observed phenomena (Jagadeesan, Mendler-D unner, and Hardt 2021). Much of previous work considers the scenario of classification with strategic agents on a population level. A few recent studies have also analyzed how phenomena observed on samples reflect the underlying population events (Haghtalab et al. 2020). Notably, very recent studies provided first analyses of learning with strategically responding agents in a PAC framework (Zhang and Conitzer 2021; Sundaram et al. 2021). The former work studied the sample complexity of learning VC-classes in this setup and analyzed effects on sample complexity of enforcing incentive compatibility for the learned classification rules. Our work can be viewed as an extension of this analysis. We propose to combine aspects of incentive compatibility and minimizing negative externalities such as social burden in form of a novel loss function that may serve as a learning objective when strategic responses are to be expected. Our sample complexity analysis is then hinging on techniques developed in the context of learning under adversarial perturbations, a learning scenario which has received considerable research attention in recent years (Feige, Mansour, and Schapire 2015; Cullina, Bhagoji, and Mittal 2018; Montasser, Hanneke, and Srebro 2019, 2021). While the learning problems are not identical, we present how strategic behaviour can be modeled as a form of one-sided adversarial perturbation and inheritance of resulting learning guarantees. 1.2 Overview on Contributions In Section 2 we review our notation and then introduce our new notion of strategic loss and motivate it. Our main contributions can be summarized as follows: Strategic manipulation loss We propose a novel loss function for learning in the presence of strategic feature manipulations. We carefully motivate this loss by relating it to concepts of social burden and incentive compatibility (and their potential trade-offs with accuracy) in prior literature. Sample complexity analysis We analyze (PAC type) learnability of VC-classes with the strategic loss. We provide sufficient conditions (and examples of when they are satisfied) for learnability with a proper learner. By drawing connections and adapting results from learning under adversarial perturbations to our setup, we also show that, while proper learnability can not always be guaranteed, every VC-class is learnable under the strategic loss with an improper learner. Robustness to inaccurate manipulation information We investigate the impact of using an approximate manipulation graph to yield a surrogate strategic loss function for cases where the true manipulation graph is not known or not accessible. For this, we introduce a novel similarity measure on graphs and show that if graphs are similar with respect to our notion then they yield reasonable surrogate strategic losses for each other (Theorem 6). Learning the manipulation graph We explore the question of whether it is possible to learn a manipulation graph that yields a good surrogate strategic loss. We identify a sufficient condition for a class of graphs G being learnable with respect to our previously defined similarity measure for graphs (Theorem 7), which in turn guaranteed the learning of a reasonable surrogate loss. All proofs can be found in the appendix of the full version (Lechner and Urner 2022). 2.1 Basic Learning Theoretic Notions for Classification We employ a standard setup of statistical learning theory for classification. We let X Rd denote the domain and Y (mostly Y = {0, 1}) a (binary) label space. We model the data generating process as a distribution P over X Y and let PX denote the marginal of P over X. We use the notation (x, y) P to indicate that (x, y) is a sample from distribution P and S P n to indicate that set S is a sequence (for example a training or test data set) of n i.i.d. samples from P. Further, we use notation ηP (x) = P(x,y) P [y = 1 | x] to denote the regression or conditional labeling function of P. We say that the distribution has deterministic labels if ηP (x) {0, 1} for all x X. A classifier or hypothesis is a function h : X Y. A classifier h can naturally be viewed a subset of X Y, namely h = {(x, y) X Y | x X, y = h(x)}. We let F denote the set of all Borel measurable functions from X to Y (or all functions in case of a countable domain). A hypothesis class is a subset of F, often denoted by H F. For a loss function ℓ: F X Y R we denote the expected loss for a distribution P as LP and the empirical loss for a sample S as LS. We use standard definitions like PAC learnability, sample complexity, and approximation error. For further elaborations on these definitions, we refer the reader to the appendix for an extended definitions section or a textbook (Shalev-Shwartz and Ben-David 2014). 2.2 Strategic Classification Learning objectives in prior work The possibilities for strategic manipulations of a feature vector are often modeled in terms of a cost function c : X X R+ 0 , so that c(x, x ) indicates how expensive it is for an individual with feature vector x to present as x . A natural minimal assumption a cost function should satisfy is c(x, x) = 0 for all feature vectors x. It is then typically assumed that instances best-respond to a published classifier, in that the individual x would choose to pay the cost of presenting as x as long as the cost doesn t exceed the utility that would be gained from the difference in classification outcome. Assuming the benefit of individual x receiving classification 1 over classification 0 is γ, the manipulation would happen if c(x, x ) γ and h(x) = 0 while h(x ) = 1 for a given classifier. That is, we can define the best response of an individual with feature vector x facing classifier h as br(x, h) = argmaxx X [γ h(x ) c(x, x )], with ties broken arbitrarily, and assuming that, if the original feature vector x is among those maximizing the above, then the individual would choose to maintain the original features. An often assumed learning goal is then performative optimality (Perdomo et al. 2020; Jagadeesan, Mendler D unner, and Hardt 2021), which stipulates that a learner should aim to maximize accuracy on the distribution it induces via the agent responses. That is, this objective can be phrased as minimizing E (x,y) P 1 [h(br(x, h)) = y] An alternative view on this setup, if the agent responses are deterministic, is to view the above as minimizing the binary loss of the effective hypothesis ˆh : X {0, 1} that is induced by h and the agents best responses br( , ) (Zhang and Conitzer 2021), defined as ˆh(x) = h(br(x, h)). (1) The goal of performative optimality has been combined with the notion of social burden that is induced by a classifier (Milli et al. 2019; Jagadeesan, Mendler-D unner, and Hardt 2021). This notion reflects that it is undesirable for a (truly) positive instance to be forced to manipulate its features to obtain a (rightfully) positive classification. This is modeled by considering the burden on a positive individual to be the cost that is incurred by reaching for a positive classification and the social burden incurred by a classifier to be the expectation with respect to the data-generating process over these costs: brd P (h) = E (x,y) P min x X{c(x, x ) | h(x ) = 1} | y = 1 It has been shown that optimizing for performative optimality (under the assumption of deterministic best-responses) also incurs maximal social burden (Jagadeesan, Mendler D unner, and Hardt 2021). A new loss function for learning under strategic manipulations Arguably, to seek performative optimality (or minimize the binary loss over the effective hypothesis class) the cost function as well as the value γ (or function γ : X R) of positive classification needs to be known (or at least approximately known). To take best responses into account, a learner needs to know what these best responses may look like. In that case, we may ignore the details of the cost function and value γ, and simply represent the collection of plausible manipulations as a directed graph structure M = (X, E) over the feature space X (Zhang and Conitzer 2021). The edge-set E consists of all pairs (x, x ) with c(x, x ) γ, and we will also use the notation x x for (x, x ) E, and write M = (X, E) = (X, ). We note that this formalism is valid for both countable (discrete) and uncountable domains. Given the information in the so obtained manipulation graph M = (X, ), we now design a loss function for classification in the presence of strategic manipulation that reflects both classification errors and the goal of disincentivizing manipulated features as much as possible. Our proposed loss function below models that it is undesirable for a classifier to assign h(x) = 0 and h(x ) = 1 if feature vector x can present as x . This is independent of a true label y (e.g. if (x, y) is sampled from the data generating process). If the label y = 0 is not positive, the point gets misclassified when x presents as x . On the other hand, if the true label is 1, then either a true positive instance is forced to manipulate their features to obtain a rightly positive outcome (and this contributes to the social burden), or, if the choice is to not manipulate the features, the instance will be misclassified (prior work has also considered models where true positive instance are honest and will not manipulate their features (Dong et al. 2018)). Here, we propose to incorporate both misclassification and contributions to social burden into a single loss function that a learner may aim to minimize. Definition 1. We define the strategic loss ℓ : F X Y {0, 1} with respect to manipulation graph (X, ) as follows: ℓ (h, x, y) = 1 if h(x) = y 1 if h(x) = 0 and x with x x and h(x ) = 1 0 else Note that the first two cases are not mutually exclusive. The above loss function discretizes the social burden by assigning a loss of 1 whenever a positive individual is required to manipulate features. As for the standard classification loss, the above point-wise definition of a loss function allows to define the true strategic loss L P (h) and empirical strategic loss L S (h) of a classifier with respect to a distribution P or a data sample S. 2.3 Comparison With Alternative Formalisms for Strategic Classification To motivate our proposed loss, we here discuss several scenarios where, we d argue, minimizing the strategic loss leads to a more desirable learning outcome than learning with a binary loss, while taking strategic manipulations into account. As discussed above, a common approach to modeling classification in a setting where strategic manipulations may occur is to assume that all agents will best-respond to a published classifier. That is, if h(x) = 0, h(x ) = 1 and x x , then the agent with initial feature vector x will effectively receive classification 1. A natural modeling is then to consider the effective hypothesis ˆh induced h (see Equation 1) and aim to minimize the classification error with the effective class ˆH = {f | f = ˆh for some h H} (Zhang and Conitzer 2021). However it has been shown, that the VC-dimension of ˆH may be arbitrarily larger than the VC-dimension of H, and may even become infinite (Zhang and Conitzer 2021). When learning this effective class ˆH with respect to the binary loss (which corresponds to aiming for performative optimality), this will imply that the class is not learnable. By contrast, we will show below that any class of finite VCdimension remains learnable with respect to the strategic loss (Theorem 5). It has also been shown that the negative effects in terms of sample complexity of considering the effective hypothesis class can be avoided by considering only incentivecompatible hypotheses in H, that is outputting only such hypotheses that will not induce any feature manipulations in response to the published classifier (Zhang and Conitzer 2021). While this avoids the growths in terms of VCdimension, it may prohibitively increase the approximation error of the resulting (pruned) class as we show in the example below. We would argue that this illustrates how low sample complexity, in itself, is not a sufficient criterion for learning success. Example 1. Consider X = N and a manipulation graph that includes edges n n + 1 and n n 1 for all n N. This is a reasonable structure, considering that the cost of moving the (one-dimensional) feature by 1 is worth a positive classification outcome. However, the only two hypotheses that are incentive compatible in this case are the two constant functions h0 : X {0} and h1 : X {1}. Thus, requiring incentive compatiblity forces the learner to assign all points in the space with the same label. This class, in fact, has low sample complexity. However, arguably, restricting the learning to such a degree (and suffering the resulting classification error, which will be close to 0.5 for distributions with balanced classes), is, in most cases not a reasonable price to pay for dis-incentivising feature manipulations. The following example illustrates how our loss function can be viewed as incorporating the notion of social burden directly into the loss. Example 2. Let s again consider a domain X = N and a manipulation graph M with edges n n + 1 for all n N. We consider distributions that have support {(1, 0), (2, 0), (3, 1), (4, 1)}, thus only these four points have positive probability mass and a hypothesis class of thresholds H = {ha | a R}, with ha(x) = 1 [x a]. The true labeling on these distributions is η(x) = h2.5(x). On all distributions, where all four points have positive mass the performatively optimal hypothesis (or effective hypothesis of minimal binary loss) however is h3.5. The social burden incurred then is brd P (h3.5) = P((3, 1)) c(3, 4). It is important to note that the performative optimality of h3.5 is independent of the distribution P over the points. A learner that minimizes the strategic loss, on the other hand, will take the distribution P into account and output h2.5 if P((2, 0)) < P((3, 1)), while outputting h3.5 if P((2, 0)) > P((3, 1)). If the difference in mass of these points (or the margin areas in a more general setting) is significant, then minimizing the strategic loss will opt for allowing a small amount of manipulation in turn for outputting a correct classification rule in case P((2, 0)) P((3, 1)); and it will opt for changing the classification rule, accept a small amount of social burden in exchange for preventing a large amount of manipulations and resulting classification errors, in case P((2, 0)) P((3, 1)). We would argue that this reflects a desirable learning outcome. 3 Learnability With the Strategic Loss 3.1 Warm Up: Loss Classes and Learnability It is well known that a class H is learnable (with respect to the set of all distributions) if the loss class induced by a 0/1-valued loss function ℓhas finite VC-dimension. In the case of the binary classification loss, this is in fact a characterization for learnability (and the VC-dimension of the loss class is identical to the VC-dimension of the hypothesis class H). In general, bounded VC-dimension of the loss class is a sufficient condition for learnability (the VC-dimension provides an upper bound on the sample complexity), but it is not a necessary condition (it doesn t, in general, yield a lower bound on the sample complexity of learning a class H with respect to some loss ℓ). We start by reviewing these notions for the classification loss and then take a closer look at the loss class induced by the strategic loss. Let ℓbe a loss function and h be a classifier. We define the loss set hℓ X Y as the set of all labeled instances (x, y) on which h suffers loss 1. The loss class Hℓis the collection of all loss sets (in the literature, the loss class is often described as the function class of indicator functions over these sets). In the case of binary classification loss ℓ0/1, the loss set of a classifier h is exactly the complement of h in X Y. That is, in this case the loss set of h is also a binary function over the domain X (namely the function x 7 |h(x) 1|). For the strategic loss on the other hand, the loss set of a classifier h is not a function, since it can contain both (x, 0) and (x, 1) for some points x X, namely if h(x) = 0 and there exists an x with x x and h(x ) = 1. For a class H we let Hℓ0/1 denote the loss class with respect to the binary loss and Hℓ the loss class with respect to the strategic loss. Definition 2. Let Z be some set and U 2Z be a collection of subsets of Z. We say that a set S Z is shattered by U if {U S | U U} = 2S, that is, every subset of S can be obtained by intersecting S with some set U from the collection U. The VC-dimension of U is the largest size of a set that is shattered by U (or if U can shatter arbitrarily large sets). It is easy to verify that for the binary loss, the VCdimension of H as a collection of subsets of X Y is identical with the VC-dimension of Hℓ0/1 (and this also coincides with the VC-dimension of H as a binary function class (Shalev-Shwartz and Ben-David 2014); VC-dimension is often defined for binary functions rather than for collection of subsets, however, this is limiting for cases where the loss class is not a class of functions). We now show that the VC-dimension of a class H and its loss class with respect to the strategic loss can have an arbitrarily large difference. Similar results have been shown for the binary loss class of the effective class ˆH induced by a manipulation graph (Zhang and Conitzer 2021). However the binary loss class of ˆH is different from the strategic loss class of H and, as we will see, the implications for learnability are also different. Observation 1. For any d N { } there exists a class H and a manipulation graph M = (X, ) with VC(H) = 1 and VC(Hℓ ) d. On the other hand, we prove that the VC-dimension of the strategic loss class Hℓ is always at least as large as the VC-dimension of the original class. Observation 2. For any hypothesis class H and any manipulation graph M = (X, ), we have VC(H) VC(Hℓ ). Standard VC-theory tells us that, for the binary classification loss, any learner that acts according to the ERM (Empirical Risk Minimization) principle is a successful learner for classes of bounded VC-dimension d. For a brief recap of the underpinnings of this result, we refer the reader to the supplementary material in the full version (Lechner and Urner 2022) or to (Shalev-Shwartz and Ben-David 2014). In the case of general loss classes with values in {0, 1}, the VC-dimension does not characterize learnability. In particular, we next show that the VC-dimension of the strategic loss class does not imply a sample complexity lower bound. Theorem 3. For every d N { }, there exists a hypothesis class H with VC(Hℓ ) = d that is learnable with sample complexity O(log(1/δ)/ϵ) in the realizable case. 3.2 Sufficient Conditions for Strategic Loss Learnability In the previous section, we have seen that the loss class having a finite VC-dimension is a sufficient (but not necessary) condition for learnability with respect to the strategic loss. We have also seen that the VC-dimension of Hℓ can be arbitrarily larger than the VC-dimension of H. To start exploring what determines learnability under the strategic loss, we provide a sufficient condition for a class to be properly learnable with respect to the strategic loss. Note that for a hypothesis h, the strategic loss set hℓ can be decomposed into the loss set of h with respect to the binary loss and the component that comes from the strategic manipulations. We next define the strategic component loss: Definition 3. We let the strategic component loss with respect to manipulation graph be defined as ℓ , (h, x) = 1 [h(x) = 0 x : x x , h(x ) = 1] We note that ℓ (h, x, y) ℓ0/1(h, x, y) + ℓ , (h, x). We will denote the true strategic component loss with respect to marginal distribution PX as L , PX . For the loss sets, we then get hℓ0/1 = {(x, y) X Y | h(x) = y}, and hℓ , = {(x, y) X Y | h(x) = 0 x : x x , h(x ) = 1}. This implies hℓ = hℓ0/1 hℓ , for all classifiers h F, and thereby Hℓ = {hℓ0/1 hℓ , | h H}. By standard counting arguments on the VC-dimension of such unions (see, for example Chapter 6 of (Shalev-Shwartz and Ben-David 2014) and exercises in that chapter), it can be shown this decomposition implies that VC(Hℓ ) d log d for d = VC(Hℓ0/1)+VC(Hℓ , ) = VC(H)+VC(Hℓ , ). Thus, if both the class H itself and the class of strategic components have finite VC-dimension, then H is properly learnable by any learner that is an ERM for the strategic loss: Theorem 4. Let H be a hypothesis class with finite VC(H) + VC(Hℓ , ) = d < . Then H is properly PAC learnable with respect to the strategic loss (both in the realizable and the agnostic case). Whether the class of strategic components has finite VCdimension intrinsically depends on the interplay between the hypothesis class H and the graph structure of the manipulation graph. In Observation 1, we saw that the graph structure can yield the strategic component sets to have much larger complexity than the original class. In the appendix, Section B, we provide a few natural examples, where the VC-dimension of the strategic components is finite. Theorem 4 provides a strong sufficient condition under which any empirical risk minimizer for the strategic loss is a successful agnostic learner for a class of finite VCdimension. We believe, in many natural situations the conditions in that theorem will hold, and analyzing in more detail which graph structure, combinations of graphs structures and hypothesis classes or classes of cost function lead to the strategic component sets having finite VC-dimension is an intriguing direction for further research. We close this section with two results, both stated in Theorem 5, on the learnability under the strategic loss in the general case where the VC-dimension of the strategic component sets may be infinite. First, there are classes and manipulation graphs for which no proper learner is (PAC-) successful, even in the realizable case. Second, for any class of finite VC-dimension and any manipulation graph, there exists an improper PAC learner. These results follow by drawing a connection from learning under the strategic loss to learning under an adversarial loss (Montasser, Hanneke, and Srebro 2019). In the general adversarial loss setup, every domain instance x is assigned a set of potential perturbation U(x), and the adversarial loss of a hypothesis h is then defined as ℓU(h, x, y) = 1 [ x U(x) : h(x ) = y] . The strategic loss can be viewed as a one-sided version of the adversarial loss, where the perturbation sets differ conditional on the label of a point, and where U(x, 1) = {x}, while U(x, 0) = {x X | x x }. The following results on learnability with the strategic loss then follow by slight modifications of the corresponding proofs for learning under adversarial loss. Theorem 5. (Adaptation of Theorem 1 and Theorem 4 in (Montasser, Hanneke, and Srebro 2019)) There exists a hypothesis class H with VC(H) = 1 that is not learnable with respect to the strategic loss by any proper learner A for H even in the realizable case. On the other hand, every class H of finite VC-dimension is learnable (by some improper learner). 4 Strategic Loss With Respect to an Approximate Manipulation Graph In many situations one might not have direct access to the true manipulation graph M = (V, E), but only to some approximate graph M = (V, E ). In this section, we will investigate how this change of manipulation graph impacts the corresponding loss function. We define a criterion for measuring the similarity of graphs with respect to hypothesis class H and show that similar graphs will yield similar strategic losses. That is, we show an upper bound on the true strategic loss of a hypothesis h (i.e., strategic loss with respect to the true manipulation graph) in terms of the graph similarity and the surrogate strategic loss of h (i.e., the strategic loss with respect to the approximate graph). We will use x x to denote (x, x ) E . As the set of vertices V is always equal to X in our setting, the graphs M and M are uniquely defined by and respectively. We will therefore use and M, as well as and M interchangeably. We now define the distance between graphs with respect to a hypothesis class H by the impact a change of manipulation graph has on the strategic component loss of elements of H. This definition and its later use are inspired by works on domain adaptation (Ben-David et al. 2010; Mansour, Mohri, and Rostamizadeh 2009). Definition 4. For two manipulation graphs, given by and , we let their H-PX -distance be defined as d H,PX ( , ) = sup h H E x PX [|ℓ , (h, x) ℓ , (h, x)|] We will now bound the strategic manipulation loss L P (h) with respect to the true graph in terms of the strategic manipulation loss L P (h) with respect to the approximate graph and the H-PX -distance between and . Theorem 6. Let H be any hypothesis class and , two manipulation graphs. Then for any distribution P over X Y and any h H we have L P (h) L0/1 P (h) + L , P (h) + d H,PX ( , ) 2L P (h) + d H,PX ( , ). Furthermore, by rearranging the result, we get 1 2L P (h) d H,PX ( , ) L P (h). We note that the expression d H,PX ( , ) is independent of the labeling and can therefore be estimated from unlabeled data. Furthermore, we have seen that small d H,PX ( , ) tightens the upper as well as the lower bound on L P (h). Therefore, d H,PX ( , ) is a suitable distance measure for approximating the structure of the manipulation graph. In the following section, we will explore learning with low d H,PX ( , ) from finite samples. 5 Learning a Manipulation Graph In the last section, we have assumed to be given an approximate manipulation graph which we can use to learn a classifier with low strategic loss. We now want to go one step further and pose the goal of learning a manipulation graph from a predefined class of graphs G such that ℓ , serves as a good strategic surrogate loss for ℓ , . From Theorem 6 we already know that ℓ , is a good surrogate loss for ℓ , if d H,PX ( , ) is small. This section will thus focus on learning an approximate manipulation graph G with small d H,PX ( , ). In order to further specify our learning problem, we will now describe what the input of such a learning procedure will look like. For a manipulation graph , let B : X 2X be the function that maps an instance x to its set of children, i.e., B (x) = {x X : x x }. We note that a manipulation graph is uniquely defined by B . Thus we will sometimes use B and interchangeably. The input to our learning procedure will be of the form of samples S = {(x1, B (x1)), . . . , (xn, B (xn))} from the true manipulation graph . As a next step in formulating our learning problem, we will need to define a loss function. As stated above, our goal is to learn G with small d H,PX ( , ). As the definition of d H,PX ( , ) contains a supremum over all h H, we cannot use it as a loss directly (as a loss needs to be defined point-wise). However, we can formulate a loss that is closely related and will serve to guarantee low d H,PX ( , ). Let the graph loss for a manipulation graph , a domain point x, a manipulation set B X and a hypothesis h as ℓgr(h, , x, B) = 1 if h(x) = 0 x B : h(x ) = 1 x : x x implies h(x ) = 0 1 if h(x) = 0 x B : h(x ) = 0 x : x x and h(x) = 1 0 otherwise. This loss is indeed closely related to the H-PX -distance as ℓgr(h, , x, B (x)) = |ℓ , (h, x) ℓ , (h, x)|. The true graph loss with respect to some marginal PX and true manipulation graph is then defined by Lgr (PX , )(h, ) = Ex PX [ℓgr(h, , x, B (x))]. Furthermore for a sample S = {(x1, B1), . . . (xn, Bn)} we define the empirical graph loss as Lgr S (h, ) = X (xi,Bi) S ℓgr(h, , xi, Bi). Similar to previous sections, we now want to define a loss class for H G. We define g(h, ) to be the set of all pairs (x, B) X 2X on which ℓgr(h, , x, B) = 1. Then the graph loss class of H G is defined as (H G)ℓgr = {g(h, ) : h H and G}. We will now show that if the VC-dimension of the loss class (H G)ℓgr is finite, we can indeed learn G with respect to ℓgr. For some examples and more discussion on the VCdimension with respect to the loss class (H G)ℓgr, we refer the reader to the appendix. Lemma 1. Let VC((H G)ℓgr) = d. Then there is ngraph : (0, 1)2 7 N, such that for any marginal distribution PX and any manipulation graph for a sample S = {(x1, B (x1)), . . . , (xn, B (xn))} of size n n(ϵ, δ), we have with probability at least 1 δ over the sample generation SX = (x1, . . . , xn) P n X for any h H and any G |Lgr (PX , )(h, ) Lgr S (h, )| < ϵ. Furthermore, ngraph(ϵ, δ) O( d+log 1 δ ϵ2 ). We note that the above lemma is agnostic in the sense that it did not require G. We will now introduce an empirical version of the H-PX -distance. This will allow us to state the main theorem of this section and show that it is indeed possible to learn G with low d H,PX ( , ) if VC((H G)ℓgr) is finite. Definition 5. Given a sample SX = {(x1, . . . , xn)} of domain elements xi and two manipulation graphs and we can define the empirical H-SX -distance as d H,SX ( , ) = sup h H xi SX ℓgr(h, , xi, B (xi)). Theorem 7. Let VC((H G)ℓgr) = d. Then there is ndist : (0, 1)2 7 N, such that for every marginal distribution PX and every manipulation graph for a sample S = {(x1, B (x1)), . . . , (xn, B (xn))} of size n n(ϵ, δ), we have with probability at least 1 δ over the sample generation SX = (x1, . . . , xn) P n X for any G d H,PX ( , ) < d H,SX ( , ) + ϵ. Furthermore, ndist(ϵ, δ) O( d+log 1 δ ϵ2 ). Combining Theorem 7 and Theorem 6 we can thus conclude that it is indeed possible to learn G such that using ℓ as a surrogate loss function guarantees a good approximation on the true strategic loss ℓ . 6 Conclusion In this paper, we introduced a new strategic loss, which incentivizes correct classification under strategic manipulations. We also incorporate the idea of social burden into our notion of loss. We differentiated this loss from previous formal frameworks designed to mitigate strategic manipulation. In particular, we showed that optimizing for our strategic loss can yield satisfactory classification rules, even if there is no incentive-compatible hypothesis in the class that performs well on the classification task at hand. In addition, the loss formulation yields desirable effects in terms of sample complexity. Our work opens various avenues for further investigations and we hope it will inspire follow-up studies on the connections between a hypothesis class and the underlying manipulation graphs, effects of these connections, as well as learnability of the manipulation graph. Acknowledgements Ruth Urner is also a faculty affiliate member of Toronto s Vector Institute. Her research is supported by an NSERC Discovery grant. Tosca Lechner is also a graduate student of the Vector Institute and was supported by a Vector Research Grant. References Ben-David, S.; Blitzer, J.; Crammer, K.; Kulesza, A.; Pereira, F.; and Vaughan, J. W. 2010. A theory of learning from different domains. Mach. Learn., 79(1-2): 151 175. Br uckner, M.; and Scheffer, T. 2011. Stackelberg games for adversarial prediction problems. In Proceedings of the 17th ACM International Conference on Knowledge Discovery and Data Mining SIGKDD, 547 555. Cullina, D.; Bhagoji, A. N.; and Mittal, P. 2018. PAClearning in the presence of adversaries. In Advances in Neural Information Processing Systems, 230 241. Dalvi, N. N.; Domingos, P. M.; Mausam; Sanghai, S. K.; and Verma, D. 2004. Adversarial classification. In Proceedings of the Tenth ACM International Conference on Knowledge Discovery and Data Mining SIGKDD, 99 108. Dong, J.; Roth, A.; Schutzman, Z.; Waggoner, B.; and Wu, Z. S. 2018. Strategic Classification from Revealed Preferences. In Proceedings of the 2018 ACM Conference on Economics and Computation, EC, 55 70. Feige, U.; Mansour, Y.; and Schapire, R. 2015. Learning and inference in the presence of corrupted inputs. In Conference on Learning Theory, 637 657. Haghtalab, N.; Immorlica, N.; Lucier, B.; and Wang, J. Z. 2020. Maximizing Welfare with Incentive-Aware Evaluation Mechanisms. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI, 160 166. Hardt, M.; Megiddo, N.; Papadimitriou, C. H.; and Wootters, M. 2016. Strategic Classification. In Proceedings of the 2016 ACM Conference on Innovations in Theoretical Computer Science, ITCS, 111 122. Hu, L.; Immorlica, N.; and Vaughan, J. W. 2019. The Disparate Effects of Strategic Manipulation. In Proceedings of the Conference on Fairness, Accountability, and Transparency, FAT*, 259 268. Jagadeesan, M.; Mendler-D unner, C.; and Hardt, M. 2021. Alternative Microfoundations for Strategic Classification. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 4687 4697. Lechner, T.; and Urner, R. 2022. Learning Losses for Strategic Classification. Co RR, abs/2203.13421. Mansour, Y.; Mohri, M.; and Rostamizadeh, A. 2009. Domain Adaptation: Learning Bounds and Algorithms. In The 22nd Conference on Learning Theory, COLT. Miller, J.; Milli, S.; and Hardt, M. 2020. Strategic Classification is Causal Modeling in Disguise. In Proceedings of the 37th International Conference on Machine Learning, ICML, 6917 6926. Milli, S.; Miller, J.; Dragan, A. D.; and Hardt, M. 2019. The Social Cost of Strategic Classification. In Proceedings of the Conference on Fairness, Accountability, and Transparency, FAT*, 230 239. Montasser, O.; Hanneke, S.; and Srebro, N. 2019. VC Classes are Adversarially Robustly Learnable, but Only Improperly. In Conference on Learning Theory, COLT, 2512 2530. Montasser, O.; Hanneke, S.; and Srebro, N. 2021. Adversarially Robust Learning with Unknown Perturbation Sets. In Conference on Learning Theory, COLT 2021, 3452 3482. Perdomo, J. C.; Zrnic, T.; Mendler-D unner, C.; and Hardt, M. 2020. Performative Prediction. In Proceedings of the 37th International Conference on Machine Learning, ICML, 7599 7609. Shalev-Shwartz, S.; and Ben-David, S. 2014. Understanding Machine Learning: From Theory to Algorithms. Cambridge University Press. Sundaram, R.; Vullikanti, A.; Xu, H.; and Yao, F. 2021. PAC-Learning for Strategic Classification. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 9978 9988. Tsirtsis, S.; and Rodriguez, M. G. 2020. Decisions, Counterfactual Explanations and Strategic Behavior. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems Neur IPS. Zhang, H.; and Conitzer, V. 2021. Incentive-Aware PAC Learning. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI, 5797 5804.