# equilibrium_analysis_of_multidefender_security_games__48ff9c48.pdf Equilibrium Analysis of Multi-Defender Security Games Jian Lou and Yevgeniy Vorobeychik Electrical Engineering and Computer Science Vanderbilt University {jian.lou,yevgeniy.vorobeychik}@vanderbilt.edu Stackelberg game models of security have received much attention, with a number of approaches for computing Stackelberg equilibria in games with a single defender protecting a collection of targets. In contrast, multi-defender security games have received significantly less attention, particularly when each defender protects more than a single target. We fill this gap by considering a multidefender security game, with a focus on theoretical characterizations of equilibria and the price of anarchy. We present the analysis of three models of increasing generality, two in which each defender protects multiple targets. In all models, we find that the defenders often have the incentive to overprotect the targets, at times significantly. Additionally, in the simpler models, we find that the price of anarchy is unbounded, linearly increasing both in the number of defenders and the number of targets per defender. Surprisingly, when we consider a more general model, this results obtains only in a corner case in the space of parameters; in most cases, however, the price of anarchy converges to a constant when the number of defenders increases. 1 Introduction With terrorism and cyber threats ever on people s minds, developing and refining our understanding of both security threats and responses to these is both an important research area, and of great practical value. Game theory has come to play an important role in the security domain, with considerable modeling and algorithmic advances, as well as actual deployment of security systems in practice that are based on such models and algorithms, including LAX Airport [Jain et al., 2008; Pita et al., 2009], US Coast Guard [Shieh et al., 2012], and the Federal Air Marshals Service [Jain et al., 2010a; 2010b; Kiekintveld et al., 2009], among others. A popular game theoretic model of security that has received much attention both in the research and in practice is as a Stackelberg game between a single defender and a single attacker, in which the defender commits to a randomized strategy, while the attacker, upon learning this strategy, chooses an optimal target or a subset of targets to attack [Conitzer and Sandholm, 2006]. In most of the associated literature, it is assumed that a single defender is responsible for all the targets that need protection, and that she has control over all of the security resources. However, there are many domains in which there are multiple defender agencies who are in charge of different subsets of all targets. While sometimes such agencies can be aligned to follow the same set of goals, in general different defender entities exhibit at least some disparities in goals. In particular, a defender is typically responsible (financially, politically, or legally) for targets in their direct charge, rather than other targets that may have social importance. This is certainly the case for the private sector, where different corporations secure their own resources without necessarily much concern for those of others, but is also common for the public sector, with different government agencies held accountable for their own assets, and not for those of others. In such non-cooperative security scenarios, the typical single-defender Stackelberg game model is clearly inadequate. Instead, we must consider the consequences of strategic interactions among multiple defenders, each charged with protecting their assets from common adversaries. An important consideration in such games is the negative externalities that security decisions impose on others: specifically, when a defender chooses a high level of security investment, budget-constrained attackers are more likely to choose others to attack. The resulting dynamics is likely to lead to over-investment in security, a phenomenon observed in several related efforts [Bachrach et al., 2013]. We consider a problem with multiple defenders protecting a collection of homogeneous targets. Each defender chooses a probability distribution over protection levels for all targets in their charge. A single attacker then best responds to the defenders action by attacking the target with the lowest probability to be protected, breaking ties uniformly at random. Our analysis is focused on three models of such multidefender games, with defenders acting non-cooperatively in all of these. We show that a Nash equilibrium among defenders in this two-stage game model need not always exist, even when the defenders utilize randomized strategies (i.e., probability distributions over target protection levels); this is distinct from a model in which the attacker moves simultaneously with the defenders, where a mixed strategy equilibrium is guaranteed to exist. When an equilibrium does exist, we show that the defenders protect all of their targets Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI 2015) with probability 1 in all three models, whereas the socially optimal protection levels are generally significantly lower. When no equilibrium exists, we characterize the best approximate Nash equilibrium (that is, one in which defenders have the least gain from deviation), showing that over-investment is substantial in this case as well. Our price of anarchy (Po A) analysis, which relies on the unique equilibrium when it exists, and the approximate equilibrium otherwise, demonstrates a surprising finding: whereas Po A is unbounded in the simpler models, increasing linearly with the number of defenders, the more general model shows this to be an atypical special case achieved when several parameters are exactly zero. More generally, Po A tends to a constant as the number of defenders increases. Related Work Although most work in Stackelberg models of security concerns computing a Stackelberg equilibrium in a single-defender single-attacker scenario, there are several exceptions. [Jiang et al., 2013] consider (mis)-coordination in cases where there are multiple defenders who are responsible for different sets of targets and share the common utility function over all targets. In this work, the defenders are fundamentally cooperative (sharing identical goals), however, making it distinct from our contribution. [Bachrach et al., 2013] examined non-cooperative security games among many defenders, in a two-stage model, but imposed strong assumptions on the model structure, and only considered onedimensional continuous security investment strategies for the defender (departing significantly from the typical structure of Stackelberg security games, in which defensive strategies are discrete protection choices). [Smith et al., 2014] extend the standard computational Stackelberg game framework to analyze games with multiple defenders, but offer no theoretical analysis. [Chan et al., 2012] propose interdependent defense (IDD) games, to study aspects of the interdependence of risk and security in a natural extension of interdependent security (IDS) games previous proposed by Heal and Kunreuther [Heal and Kunreuther, 2002] to consider attackers as explicit players in the game. In IDD games, unlike our setting, defenders and the attacker move simultaneously. To our knowledge, no previous work considers a theoretical analysis of multi-defender games with defenders protecting multiple targets, making previous literature on the subject qualitatively distinct from the typical practical considerations of single-defender settings, where resource allocation among multiple targets is a fundamental concern. 2 Modeling Multi-Defender Security Games Our modeling effort proceeds in three steps, each generalizing the previous. As we see below, each generalization step reveals new and surprising insights about the multi-defender security setting, allowing us to appreciate the fundamental incentive forces. 2.1 The Baseline Model We start with a model which most reflects the related literature: in particular, this model involves n defenders and a single attacker, with each defender engaged in protecting a single target. Each target has the same value to the defender v > 0. We suppose that the defender has two discrete choices: to protect the target, or not. In addition, the defender can randomly choose among these; our focus is on these coverage probabilities (i.e., the probability of protecting, or covering, the target), which we denote by si for a given defender i. The attacker is strategic and could observe the defenders strategies to choose a target so as to maximize the damage. We assume that attacker is indifferent among the targets, and attacks the target with the lowest coverage probability, breaking ties uniformly at random. In a given scenario, for all defenders, the attacker s strategy is a vector of probabilities P =< p1, p2, ..., pn >, where pi is the probability of attacking a target i, with Pn i=1 pi = 1. We assume that if the attacker chooses to attack a target corresponding to defender i and defender i chooses to protect the target, then the utility of the defender i is 0, and if the attacker attacks the target but it is not protected, then the utility of the defender is v. If a defender chooses to cover a target, it will incur a cost c > 0. Additionally, we assume that the defender gets a utility of zero whenever another defender s target is attacked. We can thus define the expected utility of a defender i as ui = piua i + (1 pi)uu i , where ua i is the utility of i if it is attacked, and uu i is the utility of i if it is not attacked. By the assumptions above, ua i = (1 si)v sic = v + si(v c) uu i = sic. 2.2 The Multi-Target Model Our key conceptual departure from related work is in allowing each defender to protect multiple targets, aligning it better with practical security domains. Specifically, suppose that there are n defenders, each protecting k 1 targets. Then the strategy of defender i will be a vector < si1, si2, ...sik >. The strategy profile of the attacker can then be described as a matrix of probabilities, p11 p12 . . . p1k p21 p22 . . . p2k ... ... ... ... pn1 pn2 . . . pnk in which Pn i=1 Pk j=1 pij = 1 and pij 0 for each i and j. The expected utility of a defender i in this model is j=1 pijua ij + (1 pij)uu ij, where ua ij is the utility of target j to defender i if it is attacked, and uu ij is the utility of target j to i if it is not attacked. Using the notation introduced earlier, we have ua ij = (1 sij)v sijc = v + sij(v c) uu ij = sijc. 2.3 The General Model Finally, we analyze the most general version of the model considered. We assume that if the attacker chooses to attack a target i and the defender i chooses to protect the target, then the utility of the defender i is Uc, and if the attacker attacks the target but it is not protected, then the utility of the defender is Uu. It is reasonable to assume that Uc Uu. If the target of defender i is not attacked, then we assume that the utility of defender i is T Uc. Other assumptions are same as those in the multi-target model. In the general model, therefore, ua ij = sij Uc + (1 sij)Uu sijc = (Uc Uu c)sij + Uu uu ij = T sijc. 2.4 Solution Concepts We consider a class of solutions to multi-defender security games where the defenders simultaneously commit to a probability distribution s over their pure strategies (corresponding to protection decisions for each target), and the attacker subsequently chooses an optimal target to attack, breaking ties uniformly at random. Since the attacker s behavior is straightforward (given s), we will focus on the simultaneous-move game among the defenders, and Nash equilibria thereof (giving rise to subgame perfect equilibria of the two-stage game of interest). As we demonstrate below, Nash equilibria are not guaranteed to exist, in which case we focus on ϵ-equilibria, in which no player gains more than ϵ by deviating; in particular, we will consider ϵ-equilibria with the smallest attainable ϵ. In fact, we could see Nash equilibrium as a special case of ϵ-equilibrium in which ϵ = 0. To measure how the efficiency of the game degrades due to selfish behavior of its defenders, we consider Utilitarian Social Welfare and (ϵ)-Price of Anarchy in our paper. Utilitarian Social Welfare means the sum of all defenders payoffs. For the smallest attainable ϵ, we define ϵ-Price of Anarchy (ϵ-Po A) as follows: ϵ-Po A = SWO SWE where SWO means the optimal (utilitarian) social welfare we can get from the game, SWE means the worst-case (utilitarian) social welfare in ϵ-equilibrium. An underlying assumption of this definition is that the value of SWO and SWE are both positive. If they are both negative, then ϵ-Po A will be the reciprocal of above equation. We should also note that the ordinary Price of Anarchy can be seen as a special case of ϵ-Price of Anarchy in which ϵ = 0. 3 The Baseline Model Our first result presents necessary and sufficient conditions for the existence of a Nash equilibrium in the baseline model, and characterizes it when it does exist. We omit the proofs of Baseline Model (this section) and Multi-target Model (next section) due to space constraints, and they can be found in extended version of the paper on the website. Theorem 1. In the Baseline model, Nash equilibrium exists if and only if v c. In this equilibrium all targets are protected with probability 1. Thus, if a Nash equilibrium does exist, it is unique, with all defenders always protecting their target. But what if the equilibrium does not exist? Next, we characterize the (unique) ϵ-equilibrium with the minimal ϵ that arises in such a case. We will use this approximate equilibrium strategy profile as a prediction of the defenders strategies. Theorem 2. In the Baseline model, if v < c, the optimal ϵ-equilibrium is for all defenders to cover their target with probability v c . The corresponding ϵ is v(c v) Armed with a complete characterization of predictions of strategic behavior among the defenders, we can now consider how this behavior related to socially optimal protection decisions. Since the solutions are unique, there is no distinction between the notions of price of anarchy and price of stability; we term the ratio of socially optimal welfare to welfare in equilibrium as the price of anarchy for convenience. First, we characterize the socially optimal outcome. Theorem 3. In the Baseline model, the optimal social welfare SWO is SWO = cn, if v cn; v, if v < cn. From this result, it is already clear that defenders systematically over-invest in security, except when values of the targets are quite high. This stems from the fact that the attacker creates a negative externality of protection: if a defender protects his target with higher probability than others, the attacker will have an incentive to attack another defender. In such a case, we can expect a dynamic adjustment process with defenders increasing their security investment well beyond what is socially optimal. To see just how much the defenders lose in the process, we now characterize the price of anarchy of our game. If v c, it is one and only one Nash equilibrium when all defenders have the coverage probability 1 for their targets. And the corresponding social welfare is Because it is the only Nash equilibrium, we could get the Price of Anarchy as follows: Po A = 1, if v cn; nc v , if c < v < cn. Figure 1 shows the relationship among Price of Anarchy, number of defenders, and ratio of cost c and value v. From the figure we could find that when number of defenders and ratio of c and v are small enough (e.g. n 5 and c v = 0.2), the price of anarchy is close to 1. Otherwise, the price of anarchy is unbounded, growing linearly with n. If v < c, there is no Nash equilibrium. However, we could get the optimal ϵ-equilibrium when all defenders have the same coverage probability v c for their targets. The corresponding Social Welfare is SWE = (v cn)v 2 3 4 5 6 7 8 9 10 0 Number of Defender Price of Anarchy c /v = 0. 2 c /v = 0. 4 c /v = 0. 6 c /v = 0. 8 c /v = 1. 0 Figure 1: Price of Anarchy when v c Similarly, we could get the v(c v) cn -Price of Anarchy as follows, v(c v) cn -Po A = cn + c v which is, again, linear in n. 4 The Multi-Target Model Armed with observations from the model with a single target for each defender, we now extend the model to a case not as yet considered in the literature in a theoretical light: each defender protects a set of k targets. This gives rise to a combinatorial set of possible decisions for each defender, so that even computing a best response is not necessarily easy. Remarkably, we are able to characterize equilibria and approximate equilibria in this setting as well. Our first result is almost a mirror-image of the corresponding result in the baseline model: when a Nash equilibrium exists, all defenders protect all of their targets with probability 1. Theorem 4. In the Multi-Target model, Nash equilibrium exists if and only if v kc. In this equilibrium all targets are protected with probability 1. Next, we consider scenarios when v < kc, in which there is no Nash equilibrium. Our next result characterizes optimal (lowest-ϵ) approximate equilibria. Theorem 5. In the Multi-Target model, if v < kc, then in the optimal ϵ-equilibrium all targets are protected with probability v kc. The corresponding ϵ is v(kc v) Thus, as n increases, the optimal approximate equilibrium approaches a Nash equilibrium. Figure 2 illustrates the relationship between ϵ and the number of targets each defender protects when v = 10 and c = 1. In this figure, ϵ = 0 when k 10, which means that an exact Nash equilibrium exists; ϵ increases with k when k > 10, but at a decreasing rate, converging to v n when k . Finally, we characterize socially optimal welfare, and, subsequently, put everything together in describing the price of anarchy. 2 4 6 8 10 12 14 16 18 20 0 Number of Targets Each Defender Has n=2 n=3 n=4 n=5 n=6 Figure 2: ϵ value when v = 10, c = 1 Theorem 6. In the Multi-Target model, the optimal social welfare SWO is SWO = cnk, if v cnk; v, if v < cnk. Thus, just as in the baseline case, the defenders will generally over-invest in security. If v kc, there is a unique Nash equilibrium with all targets protected with probability 1. The corresponding social welfare is SWE = cnk Because it is the only Nash equilibrium, the Price of Anarchy is Po A = 1, if v cnk; nkc v , if ck v < cnk. If v < kc, there is no Nash equilibrium. However, we could get the optimal approximate equilibrium when all defenders have the same coverage probability v kc for their targets. The corresponding Social Welfare is SWE = (v cnk) v And the v(kc v) cnk -Price of Anarchy is cnk -Po A = n + 1 v Clearly, in either case, and just as in the baseline model, the price of anarchy is unbounded, growing linearly with n. We now consider how Po A changes as a function of k, i.e. the number of targets each defender has. When k v cn, a Nash equilibrium exists and the Po A is 1; when v cn < k v c , Po A increases linearly in k with the slope nc v . However, when k > v c , a Nash equilibrium does not exist and the approximate Po A is n + 1 v kc, which increases very slowly with k, and is bounded by n + 1 when k . Figure 3 illustrates the relationship between (approximate) Price of Anarchy and k for n = 2. When k is very small, Po A = 1. For intermediate k, Po A increases linearly, and when k is sufficiently large, Nash equilibrium no longer exists, and ϵ-Po A increases quite slowly, converging to 3 when k . 2 4 6 8 10 12 14 16 18 20 0 Number of Targets Each Defender Has (Approximate) Price of Anarchy n=2 n=3 n=4 n=5 n=6 Figure 3: (Approximate) Price of Anarchy when c 5 The General Model Both the baseline and the multi-target models made rather strong assumptions about the structure of the utility functions of the players. In the general model, we relax these assumptions, allowing for arbitrary utilities for the players when the target is attacked or not, and when it is protected or not (when attacked). Quite surprisingly, our findings here are qualitatively different: the special case of the baseline and multitarget models turns out to be an exception, rather than the rule when more general models are considered. Just as before, we start by characterizing Nash and approximate Nash equilibria. Theorem 7. In the General model, Nash equilibrium exists if and only if Uc Uu kc (n 1)(T Uc) n . In this equilibrium all targets are protected with probability 1. Proof. We firstly claim that Nash equilibrium can appear only if coverage probability of all of targets tij are identical. Otherwise, there will be a target tik which has the possibility of 0 to be attacked, and defender i has incentive to decrease sik. To find Nash equilibrium, we need only consider scenarios in which all targets have the same coverage probability. When all targets have the same possibility s to be protected. Then for each defender, her expected utility is u = (Uc Uu nkc)s + Uu + (nk 1)T n If s < 1, then some defender i could slightly increase s to s + δ (δ is a very small positive real number) for all of her targets to make sure them not be attacked and get the value u = k T k(s + δ)c, u u = (Uc Uu)(1 s) + (T Uc) nkcδ As Uc Uu, T Uc, and δ can be very small, u u > 0 when s < 1. We could know that the defender has incentive to improve s when s < 1. So the Nash equilibrium can appear only if sij = 1 for all defenders targets sij. When all targets have the same possibility s = 1 to be covered, for each defender, her expected utility is u = Uc nkc + (nk 1)T We claim that, if a defender i has incentive to deviate, the optimal deviation could appear only if defender i has the same protection probability s for all her targets. Otherwise, for some target tik which has the probability 0 to be attacked, she could always decrease s to get higher utility. If probabilities of targets protected by defender i are all s (0 s < 1), then her expected utility is u = (Uc Uu c)s + Uu + (k 1)(T s c), and u u = (Uc Uu kc)(s 1) + (n 1)(Uc T) 1) If Uc Uu kc, then u u 0, we could know that it is a Nash equilibrium when all targets have the probability 1 to be protected. 2) If Uc Uu < kc, the maximal value of u u corresponds to s = 0, max 0 s <1 u u = (Uc Uu kc) (n 1)(T Uc) If kc (n 1)(T Uc) n Uc Uu < kc, u u 0, it is a Nash equilibrium; otherwise, it is not. To sum up, Nash equilibrium exists if and only if Uc Uu kc (n 1)(T Uc) n , and the equilibrium corresponds to all targets having probability 1 of being protected. Next, we characterize the optimal approximate equilibrium when no Nash equilibrium exists. Theorem 8. In the General model, in the optimal ϵequilibrium all targets are protected with probability T Uu kc . The corresponding ϵ is (T Uu)(kc Uc+Uu) Proof sketch. When all targets have the same probability s of being protected, the expected utility of each defender is u = (Uc Uu nkc)s + Uu + (nk 1)T n Assume 0 s < 1. If some defender i slightly increases s to s + δij for target tij, then she would obtain utility u = Pk j=1 T (s + δij)c, u u = T (Uc Uu)s Uu T (Uc Uu)s Uu Then we consider scenarios in which a defender i could get higher utility by decreasing protection probability. We claim that the optimal deviation could appear only if defender i has the same protection probability s for all of her targets. Otherwise, for some target tik which has the probability 0 to be attacked, she could always decrease coverage probability of tik to get higher utility. Then we need only consider cases in which a defender deviates by decreasing probabilities of all her targets to s δ. And her utility will be u = (Uc Uu kc)(s δ) + Uu + (k 1)T, As Uc Uu < kc, when δ = s (the maximal value of δ), we could get maximal value of u u, max 0<δ s u u = T (Uc Uu)s Uu nk +kcs+Uu T (2) By comparing the value of equation (1) and equation (2), we could get different value of ϵ for for ϵ-equilibrium as shown below: ( T (Uc Uu)s Uu n , if 0 s T Uu kc ; T (Uc Uu)s Uu n + kcs + Uu T, if T Uu When s = T Uu kc , we could get the minimal ϵ = (T Uu)(kc Uc+Uu) cnk . We claim that the (T Uu)(kc Uc+Uu) cnk -equilibrium can appear only if all targets have the same coverage probability s. We prove this by contradiction. Suppose that targets have different coverage probabilities. This gives rise to two cases: 1) Each defender uses an identical coverage probability for each target she owns (these may differ between defenders); 2) There exists a defender, who has a different probability to protect her own targets. In case 1), there exist β defenders (1 β < n) who have the same minimal probability s to protect all of their targets. The expected utility for each defender among these β defenders is: u = (Uc Uu kβc)s + Uu + (kβ 1)T kc < s 1, some defender i among these β defenders could decrease probability of all her targets to 0 to get value u1 = Uu + (k 1)T, u1 u = T (Uc Uu)s Uu β + kcs + Uu T > T (Uc Uu)s Uu n + kcs + Uu T When 0 s T Uu kc , some defender i among these β defenders could slightly increase probabilities of all her targets to s + δ3 to get the utility u2 = k T k(s + δ3)c u2 u = T (Uc Uu)s Uu kβcδ3 > T (Uc Uu)s Uu n The above inequality holds because δ3 can be arbitrarily small. Thus, no profile in case 1) can be a (T Uu)(kc Uc+Uu) cnk -equilibrium. In case 2), for defenders who have different coverage probability for their own targets, they could always increase payoff by decreasing some of their targets probability to get a corresponding profile in case 1). Then we could know that any profile in case 2) cannot be a (T Uu)(kc Uc+Uu) cnk - equilibrium. As the final step towards characterizing the Price of Anarchy, we derive optimal social welfare in this model. Theorem 9. In the General model, the optimal social welfare SWO is SWO = Uc nkc + (nk 1)T, if Uc Uu nkc; Uu + (n 1)T, if Uc Uu < nkc. Proof sketch. We firstly claim that we could get optimal social welfare only if all targets have the same probability s to be protected their. Otherwise, some target tij has the probability of 0 to be attacked. Then we could decrease sij to get a better social welfare. Consequently, we need only to consider an optimal identical coverage probability s to obtain optimal social welfare, which can be done in a relatively straightforward way. If Uc Uu kc (n 1)(T Uc) n , the Nash equilibrium is unique, with all targets protected with probability 1. The corresponding social welfare in equilibrium is SWE = Uc nkc + (nk 1)T. So far we have not yet added any constrains to value of T, Uc, and Uu (except that T Uc Uu). In order to make Price of Anarchy well-defined, we need to add constraints that values of T, Uc, and Uu are all non-positive (just as in the previous two models) or all non-negative. To be consistent with previous models, we add constraints that Uc, Uu and T are all non-positive (little changes if all are non-negative). In the case of a unique Nash equilibrium, the price of anarchy is 1, if Uc Uu nkc; Uc Uu nkc Uu+(nk 1)T + 1, if kc (n 1)(T Uc) n Uc Uu < nkc. If Uc Uu < kc (n 1)(T Uc) n , there is no Nash equilibrium. However, we could get the optimal approximate Nash equilibrium when all defenders use the same coverage probability T Uu kc for all targets. The corresponding Social Welfare is SWE = (Uc Uu nkc)T Uu kc + Uu + (nk 1)T, and the (T Uu)(kc Uc+Uu) cnk -Price of Anarchy is (T Uu)(kc Uc + Uu) =(Uc Uu nkc)(T Uu) kc Uu + (nk 1)kc T + 1 We now analyze the relationship between (ϵ-)Po A and the values of n and k. Here are the key differences from Multi Target Model. First we consider (ϵ-)Po A as the function of n. If T = 0, then we could find that the result is same as that in Multi-Target Model and (ϵ-)Po A linearly increases in n, which is unbounded. However, if T = 0, Po A and ϵ-Po A are increasing in n, and as n , approach 1 c T and 1+ Uu T k T , respectively. In other words, Po A (exact and approximate) is bounded by a constant, for a constant k! Consider now approximate price of anarchy as a function of k. If T = 0, it is bounded by n + 1. However, if T = 0, 2 4 6 8 10 12 14 16 18 20 Number of Targets Each Defender Has (Approximate) Price of Anarchy n=2 n=3 n=4 n=5 n=6 Figure 4: (Approximate) Price of Anarchy when c = 1, T = 1, Uc = 2 and Uu = 10 when kc (n 1)(T Uc) n Uc Uu, it is an increasing function of k. When kc (n 1)(T Uc) n > Uc Uu, it may at first increase or decrease in k, depending on the the values of the model parameters. However, when k is large enough, price of anarchy will invariably be decreasing in k, and as k , ϵ-Po A 1. Figure 4 provides an example of the relationship between ϵ-Po A and k. Observe that all the curves begin to decrease when k > 10; they all approach 1 as k . Thus, price of anarchy in the general model is only unbounded in the special case when T = 0, whereas when T = 0, price of anarchy is always bounded by a constant. This observation is particularly surprising and significant considering the fact that the baseline and simplified multi-target models are quite natural, and seemingly innocuous, restrictions of the general case. 6 Conclusion We examined a non-cooperative multi-defender security game in which defenders may protect multiple targets, offering complete characterization of Nash and approximate equilibria, socially optimal solutions, and price of anarchy. Our results show that defenders generally over-protect the targets in this model, but different modeling assumptions give rise to qualitatively different outcomes: a simpler model gives rise to an unbounded price of anarchy, whereas a more general model sees Po A converge to a constant when the number of defenders increases. There are a number of directions for further work. Firstly, in our work we assume that strategy spaces of defenders are symmetric, and the impact of asymmetry is far from clear. Secondly, defenders in our model always have incentive to over-invest for their targets. Some public policies such as taxation and penalties may be helpful to improve the overall efficiency. Finally, it would be important to consider how interdependence among targets affects the levels of security investment and the price of anarchy. Acknowledgments This work was partially supported by the Air Force Research Laboratory under award FA8785-14-2-0180, National Sci- ence Foundation under award CNS-1238959, and Sandia National Laboratories. References [Bachrach et al., 2013] Yoram Bachrach, Moez Draief, and Sanjeev Goyal. Contagion and observability in security domains. Allerton Conference, 2013. [Chan et al., 2012] Hau Chan, Michael Ceyko, and Luis E. Ortiz. Interdependent defense games: Modeling interdependent security under deliberate attacks. In Nando de Freitas and Kevin P. Murphy, editors, UAI, pages 152 162. AUAI Press, 2012. [Conitzer and Sandholm, 2006] Vincent Conitzer and Tuomas Sandholm. Computing the optimal strategy to commit to. In Proceedings of the 7th ACM Conference on Electronic Commerce, EC 06, pages 82 90, New York, NY, USA, 2006. ACM. [Heal and Kunreuther, 2002] Geoffrey Heal and Howard Kunreuther. Interdependent security. Journal of Risk and Uncertainty, 26:231 249, 2002. [Jain et al., 2008] Manish Jain, James Pita, Milind Tambe, Fernando Ord o nez, Praveen Paruchuri, and Sarit Kraus. Bayesian stackelberg games and their application for security at los angeles international airport. SIGecom Exch., 7(2):10:1 10:3, June 2008. [Jain et al., 2010a] Manish Jain, Erim Kardes, Christopher Kiekintveld, Fernando Ordez, and Milind Tambe. Security games with arbitrary schedules: A branch and price approach. In Maria Fox and David Poole, editors, AAAI. AAAI Press, 2010. [Jain et al., 2010b] Manish Jain, Jason Tsai, James Pita, Christopher Kiekintveld, Shyamsunder Rathi, Milind Tambe, and Fernando Ord o oez. Software assistants for randomized patrol planning for the lax airport police and the federal air marshal service. Interfaces, 40(4):267 290, July 2010. [Jiang et al., 2013] Albert Xin Jiang, Ariel D. Procaccia, Yundi Qian, Nisarg Shah, and Milind Tambe. Defender (mis)coordination in security games. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, IJCAI 13, pages 220 226. AAAI Press, 2013. [Kiekintveld et al., 2009] Christopher Kiekintveld, Manish Jain, Jason Tsai, James Pita, Fernando Ord o nez, and Milind Tambe. Computing optimal randomized resource allocations for massive security games. In Proceedings of The 8th International Conference on Autonomous Agents and Multiagent Systems - Volume 1, AAMAS 09, pages 689 696, Richland, SC, 2009. International Foundation for Autonomous Agents and Multiagent Systems. [Pita et al., 2009] James Pita, Manish Jain, Fernando Ord o nez, Christopher Portway, Milind Tambe, Craig Western, Praveen Paruchuri, and Sarit Kraus. Using game theory for los angeles airport security. AI Magazine, 30(1):43 57, 2009. [Shieh et al., 2012] Eric Shieh, Bo An, Rong Yang, Milind Tambe, Craig Baldwin, Joseph Di Renzo, Ben Maule, and Garrett Meyer. Protect: A deployed game theoretic system to protect the ports of the united states. In Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems - Volume 1, AAMAS 12, pages 13 20, Richland, SC, 2012. International Foundation for Autonomous Agents and Multiagent Systems. [Smith et al., 2014] Andrew Smith, Yevgeniy Vorobeychik, and Joshua Letchford. Multi-defender security games on networks. ACM SIGMETRICS Performance Evaluation Review, 41(4):4 7, 2014.