# individual_regret_in_cooperative_nonstochastic_multiarmed_bandits__54b5b1fb.pdf Individual Regret in Cooperative Nonstochastic Multi-Armed Bandits Yogev Bar-On Tel Aviv University, Israel baronyogev@gmail.com Yishay Mansour Tel Aviv University, Israel and Google Research, Israel mansour.yishay@gmail.com We study agents communicating over an underlying network by exchanging messages, in order to optimize their individual regret in a common nonstochastic multi-armed bandit problem. We derive regret minimization algorithms that guar- antee for each agent v an individual expected regret of e O r 1 + K |N (v)| T , where T is the number of time steps, K is the number of actions and N (v) is the set of neighbors of agent v in the communication graph. We present algorithms both for the case that the communication graph is known to all the agents, and for the case that the graph is unknown. When the graph is unknown, each agent knows only the set of its neighbors and an upper bound on the total number of agents. The individual regret between the models differs only by a logarithmic factor. Our work resolves an open problem from [Cesa-Bianchi et al., 2019b]. 1 Introduction The multi-armed bandit (MAB) problem is one of the most basic models for decision making under uncertainty. It highlights the agent s uncertainty regarding the losses it suffers from selecting various actions. The agent selects actions in an online fashion - each time step the agent selects a single action and suffers a loss corresponding to that action. The agent s goal is to minimize its cumulative loss over a fixed horizon of time steps. The agent observes only the loss of the action it selected each step. Therefore, the MAB problem captures well the crucial trade-off between exploration and exploitation, where the agent needs to explore various actions in order to gather information about them. MAB research discusses two main settings: the stochastic setting, where the losses of each action are sampled i.i.d. from an unknown distribution, and the nonstochastic (adversarial) setting, where we make no assumptions about the loss sequences. In this work we consider the nonstochastic setting and the objective of minimizing the regret - the difference between the agent s cumulative loss and the cumulative loss of the best action in hindsight. It is known that a regret of the order of Θ is the best that can be guaranteed, where K is the number of actions and T is the time horizon. In contrast, when the losses of all actions are observed (full-information feedback) the regret can be of the order of Θ T ln K (see, e.g., [Cesa-Bianchi and Lugosi, 2006, Bubeck et al., 2012]). The main focus of our work is to consider agents that are connected in a communication graph, and can exchange messages in each step, in order to reduce their individual regret. This is possible since the losses depend only on the action and the time step, but not on the agent. One extreme case is when the communication graph is a clique, i.e., any pair of agents can communicate directly. In this case, the agents can run the well known Exp3 algorithm [Auer et al., 2002], and guarantee each a regret of O T ln K , assuming there are at least K agents (see [Seldin et al., 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. 2014, Cesa-Bianchi et al., 2019b]). However, in many motivating applications, such as distributed learning, or communication tasks such as routing, the communication graph is not a clique. The work of Cesa-Bianchi et al. [2019b] studies a general communication graph, where the agents can communicate in order to reduce their regret. The paper presents the Exp3-Coop algorithm, which achieves an expected regret when averaged over all agents of e O q 1 + K N α (G) T , where α (G) is the independence number of the communication graph G, and N is the number of agents. The question of whether it is possible to obtain a low individual regret, that holds simultaneously for all agents, was left as an open question. We answer this question affirmatively in this work. Our main contribution is an individual expected regret bound, which holds for each agent v, of order s 1 + K |N (v)| where N (v) is the set of neighbors of agent v in the communication graph. We remark that our result also implies the previous average regret bound. The main idea of our algorithm is to artificially partition the graph into disjoint connected components. Each component has a center agent, which is in some sense the leader of the component. The center agent has (almost) the largest degree in the component, and it selects actions using the Exp3-Coop algorithm. By observing the outcomes of its immediate neighboring agents, the center agent can guarantee its own desired individual regret. The main challenge is to create such components with a relatively small diameter, so that the center will be able to broadcast its information in a short time to all the agents in the component. Special care is given to relate the agents local parameters (degree) to the global component parameters (degree of the center agent and the broadcast time). We consider both the case that the communication graph is known to all the agents in advance (the informed setting), and the case that the graph is unknown (the uninformed setting). In the uninformed setting, we assume each agent knows its local neighborhood (i.e., the set of its neighbors), and an upper bound on the total number of agents. The regret bound in the uninformed setting is higher by a logarithmic factor and the algorithm is more complex. In the next section, we formally define our model, and review preliminary material. Section 3 shows the center-based policy, given a graph partition. We then present our graph partitioning algorithms in Section 4. Overview of the analysis is given in Section 5, while all proofs are differed to the supplementary material. Our work is concluded in Section 6. 1.1 Additional related works The cooperative nonstochastic MAB setting was introduced by Awerbuch and Kleinberg [2008], where they bound the average regret, when some agents might be dishonest and the communication is done through a public channel (clique network). The previously mentioned [Cesa-Bianchi et al., 2019b], also considers the issue of delays, and presents a bound on the average regret for a general graph of order e O q d + K N α (G) T + d , when messages need d steps to arrive. Dist-Hedge, introduced by Sahu and Kar [2017], considers a network of forecasting agents, with delayed and inexact losses, and derives a sub-linear individual regret bound, that also depends on spectral properties of the graph. More recently, Cesa-Bianchi et al. [2019a] studied an online learning model where only a subset of the agents play at each time step, and showed matching upper and lower bounds on the average regret of order p α (G) T when the set of agents that play each step is chosen stochastically. When the set of agents is chosen arbitrarily, the lower bound becomes T. In the stochastic setting, Landgren et al. [2016a,b] presented a cooperative variant of the well-known UCB algorithm, that uses a consensus algorithm for estimating the mean losses, to obtain a low average regret. More cooperative variants of the UCB algorithm that yield a low average regret were presented by Kolla et al. [2018]. They also showed a policy, where like in the methods in this work, agents with a low degree follow the actions of agents with a high degree. Stochastic MAB over P2P communication networks were studied by Szörényi et al. [2013], which showed that the probability to select a sub-optimal arm reduces linearly with the number of peers. The case where only one agent can observe losses was investigated by Kar et al. [2011]. This agent needs to broadcast information through the network, and it was shown this is enough to obtain a low average regret. Another multi-agent research area involve agents that compete on shared resources. The motivation comes from radio channel selection, where multiple devices need to choose a radio channel, and two or more devices that use the same channel simultaneously interfere with each other. In this setting, many papers assume agents cannot communicate with each other, and do not receive a reward upon collision - where more than one agent tries to choose the same action at the same step. The first to give regret bounds on this variant are Avner and Mannor [2014], that presented an average regret bound of order O T 2 3 in the stochastic setting. Also in the stochastic setting, Rosenski et al. [2016] showed an expected average regret bound of order O K δ + N that holds with probability 1 δ, where is the minimal gap between the mean rewards (notice that this bound is independent of T). In the same paper, they also studied the case that the number of agents may change each step, and presented a regret bound of e O x T , where x is the total number of agents throughout the game. Bistritz and Leshem [2018] consider the case that different agents have different mean rewards, and each agent has a different unique action it should choose to maximize the total regret. They showed an average regret of order O log2+ϵ T for every ϵ > 0, where the O-notation hides the dependency on the mean rewards. 2 Preliminaries We consider a nonstochastic multi-armed bandit problem over a finite action set A = {1, . . . , K} played by N agents. Let G = V, E be an undirected connected communication graph for the set of agents V = {1, . . . , N}, and denote by N (v) the neighborhood of v V , including itself. Namely, N (v) = {u V | u, v E} {v} . At each time step t = 1, 2, . . . , T, each agent v V draws an action It (v) A from a distribution pv t = pv t (1) , . . . , pv t (K) on A. It then suffers a loss ℓt (It (v)) [0, 1] which it observes. Notice the loss does not depend on the agent, but only on the time step and the chosen action. Thus, agents that pick the same action at the same step will suffer the same loss. We also assume the adversary is oblivious, i.e., the losses do not depend on the agents realized actions. In the end of step t, each agent sends a message mt (v) = v, t, It (v) , ℓt (It (v)) , pv t to all the agents in its neighborhood, and also receives messages from its neighbors: mt (v ) for all v N (v). Our goal is to minimize, for each v V , its expected regret over T steps: t=1 ℓt (It (v)) min i A A well-known policy to update pv t is the exponential-weights algorithm (Exp3) with weights wv t (i) for all i A, such that pv t (i) = wv t (i) W v t where W v t = P i A wv t (i) (see, e.g., [Cesa-Bianchi and Lugosi, 2006]). The weights are updated as follows: let Bv t (i) be the event that v observed the loss of action i at step t; in our case Bv t (i) = I { v N (v) : It (v ) = i}, where I is the indicator function. Also, let ˆℓv t (i) = ℓt(i) Et[Bv t (i)]Bv t (i) be an unbiased estimated loss of action i at step t, where Et [ ] is the expectation conditioned on all the agents choices up to step t (hence, Et h ˆℓv t (i) i = ℓt (i)). Then wv t+1 (i) = wv t (i) exp η (v) ˆℓv t (i) , where η (v) is a positive parameter chosen by v, called the learning rate of agent v. Exp3 is given explicitly in the supplementary material. Notice that in our setting all agents v V have the information needed to compute ˆℓv t (i), since Et [Bv t (i)] = Pr [ v N (v) : It (v ) = i] = 1 Y 1 pv t (i) , and if agent v does not observe ℓt (i), then ˆℓv t (i) = 0. We proceed with two useful lemmas that will help us later. For completeness, we provide their proofs in the supplementary material as well. The first lemma is the usual analysis of the exponential-weights algorithm: Lemma 1. Assuming agent v uses the exponential-weights algorithm, its expected regret satisfies RT (v) ln K η (v) + η (v) i=1 pv t (i) ˆℓv t (i)2 # The next lemma is from [Cesa-Bianchi et al., 2019b], and it bounds the change of the action distribution in the exponential-weights algorithm. Lemma 2. Assuming agent v uses the exponential-weights algorithm with a learning rate η (v) 1 2K , then for all i A: 1 η (v) ˆℓv t (i) pv t (i) pv t+1 (i) 2pv t (i) . Also, the following definition will be needed for our algorithm. We denote by Gr the r-th power of G, in which v1, v2 V are adjacent if and only if dist G (v, v ) r; and by G|U the sub-graph of G induced by U V . Definition 3. Let G = V, E be an undirected connected graph and let W U V . W is called an r-independent set of G, if it is an independent set of Gr. Namely, w, w W : dist G (w, w ) r + 1. If W is also a maximal independent set of (Gr)|U, it is called a maximal r-independent subset (r-MIS) of U. Namely, there is no r-independent set W U such that W W . 3 Center-based cooperative multi-armed bandits We now present the center-based policy for the cooperative multi-armed bandit setting, which will give us the desired low individual regret. In the center-based cooperative MAB, not all the agents behave similarly. We partition the agents to three different types. Center agents are the agents that determine the action distribution for all other agents. They work together with their neighbors to minimize their regret. The neighbors of the center agents in the communication graph, center-adjacent agents, always copy the action distribution from their neighboring center, and thus the centers gain more information about their own distribution each step. Other (not center or center-adjacent) agents are simple agents, which simply copy the action distribution from one of the centers. Since they are not center-adjacent, they receive the action distribution with delay, through other agents that copy from the same center. We artificially partition the graph to connected components, such that each center c has its own component, and all the simple agents in the component of c copy their action distribution from it. To obtain a low individual regret, we require the components to have a relatively small diameter, and the center agents to have a high degree in the communication graph. Namely, center agents have the highest or nearly highest degree in their component. In more detail, we select a set C V of center agents. All center agents c C use the exponential- weights algorithm with a learning rate η (c) = 1 (ln K) min{|N (c)|,K} KT . The agent set V is partitioned into disjoint subsets {Vc V | c C}, such that N (c) Vc for all c C, and the sub-graph Gc G|Vc induced by Vc is connected. Notice that since the components are disjoint, the condition N (c) Vc implies C is a 2-independent set. For all non-centers v V \C, we denote by C (v) C the center agent such that v VC(v), and call it the center of v. All non-center agents v V \ C copy their distribution from their origin neighbor U (v), which is their neighbor in GC(v) closest to C (v), breaking ties arbitrarily. Namely, U (v) = arg min v N(v) VC(v) dist GC(v) (v , C (v)) . Thus, agent v receives its center s distribution with a delay of d (v) = dist GC(v) (v, C (v)) steps, so for all t d (v) + 1: pv t = p C(v) t d(v). Notice that if v N (c), then v is center-adjacent and it holds U (v) = C (v) and d (v) = 1. For completeness, we define U (c) = C (c) = c and d (c) = 0 for all c C. To express the regret of the center-based policy, we introduce a new concept: Algorithm 1 Center-based cooperative MAB - v is a center agent Parameters: Number of arms K; Time horizon T. Initialize: η (v) 1 KT ; wv 1 (i) 1 K for all i A. 1: for t T do 2: Set pv t (i) wv t (i) W v t for all i A, where W v t = P i A wv t (i). 3: Play an action It (v) drawn from pv t = pv t (1) , . . . , pv t (K) . 4: Observe loss ℓt (It (v)). 5: Send the following message to the set N (v): mt (v) = v, t, It (v) , ℓt (It (v)) , pv t . 6: Receive all messages mt (v ) from v N (v). 7: Update for all i A: wv t+1 (i) wv t (i) exp η (v) ˆℓv t (i) , where ˆℓv t (i) = ℓt (i) Et [Bv t (i)]Bv t (i) , Bv t (i) = I { v N (v) : It (v ) = i} , Et [Bv t (i)] = 1 Y 1 pv t (i) . Algorithm 2 Center-based cooperative MAB - v is a non-center agent Parameters: Number of arms K; Time horizon T; Origin neighbor U (v). Initialize: pv 1 (i) 1 K for all i A. 1: for t T do 2: Play an action It (v) drawn from pv t = pv t (1) , . . . , pv t (K) . 3: Observe loss ℓt (It (v)). 4: Send the following message to the set N (v): mt (v) = v, t, It (v) , ℓt (It (v)) , pv t . 5: Receive the message mt (U (v)) from U (v). 6: Update pv t+1 (i) = p U(v) t (i) for all i A. 7: end for Definition 4. The mass of a center agent c C is defined to be M (c) min {|N (c)| , K} , and the mass of non-center agent v V \ C is 6 d(v)M (C (v)) . Notice the mass depends only on how the graph is partitioned, and it satisfies M (v) = e 1 6 M (U (v)) for all non-centers v V \ C. Intuitively, the mass of agent v captures the idea that as the degree of the center is larger and as the agent is closer to its center, the lower the regret of v. We prove that the regret is e O q K M(v)T . Our partitioning algorithms, presented in the next section, show that the mass of agent v satisfies M (v) = Ω(min {|N (v)| , K}), so we obtain an individual regret of the order of e O r 1 + K |N (v)| T . We specify the center-based policy in Algorithms 1 and 2. We emphasize that before the agents use the center-based policy they must partition the graph with one of the algorithms we present in the next section. While the agents partition the graph, they play arbitrary actions. 4 Partitioning the graph The goal now is to show that we can partition the graph such that the mass is large for every v V . In particular, we want to show that any graph can be partitioned such that M (v) = Ω(min {|N (v)| , K}). We consider two cases: the informed and uninformed settings. In the informed setting, all of the agents have access to the graph structure. Each agent can partition the graph by itself in advance, Algorithm 3 Centers-to-Components Parameters: Number of arms K; Center set C. Initialize: Number of iterations ΘK 12 ln K . 1: if v C then 2: Initialize: C0 (v) v; U0 (v) v; M0 (v) min {|N (v)| , K}. 3: else 4: Initialize: C0 (v) nil; U0 (v) nil; M0 (v) 0. 5: end if 6: for 0 t ΘK do 7: Send the following message to the set N (v): µt (v) = v, t, Ct (v) , Mt (v) . 8: Receive all messages µt (v ) from v N (v). 9: if Ut (v) / C then The center-based policy requires N (c) Vc for all c C. 10: Find the best origin neighbor for v: Ut+1 (v) arg max v N(v)\{v} Mt (v ) . 11: Update: Ct+1 (v) Ct (Ut+1 (v)) ; Mt+1 (v) e 1 6 Mt (Ut+1 (v)). 12: else 13: Keep old values: Ct+1 (v) Ct (v) ; Ut+1 (v) Ut (v) ; Mt+1 (v) Mt (v). 14: end if 15: end for 16: return C (v) = CΘK+1 (v) ; U (v) = UΘK+1 (v) ; M (v) = MΘK+1 (v) . to know the role it plays: whether it is a center or not, and which agent is its origin neighbor. In the uninformed setting, the graph structure is not known to the agents, only their neighbors and an upper bound on the total number of agents N N. The agents partition the graph using a distributed algorithm while playing actions and suffering loss. The basic structure of the partitioning algorithm in both settings is the same. First, we show an algorithm that computes the connected components given a center set C. Then, we show an algorithm that computes a center set C. The second algorithm is specifically designed to be used with the first, and together they partition the graph to connected components such that every agent has a large mass. 4.1 Computing graph components given a center set Given a center set C, we show a distributed algorithm called Centers-to-Components, which computes the connected components, and present it in Algorithm 3. Although it is distributed, in the informed setting agents can simply simulate it locally in advance. Centers-to-Components runs simultaneous distributed BFS graph traversals, originating from every center c C. When the traversal of center c arrives to a simple agent v V \ C, v decides if c is the best center for it so far, and if it is, v switches its component to Vc. Notice each agent needs to know only if itself is a center or not. 4.2 Computing centers To compute the center set C, we show two algorithms; one for the informed setting and one for the uninformed setting. The regret bound for the informed setting is slightly better, and the algorithm is simpler. The informed setting The algorithm that computes the center set in the informed setting is called Compute-Centers-Informed and is presented in Algorithm 4. The center set is built in a greedy way: each iteration, all of the agents test if they are satisfied with the current center set (i.e., M (v) min {|N (v)| , K}). If there are unsatisfied agents left, the agent with the highest degree is added to the center set. Algorithm 4 Compute-Centers-Informed Parameters: Undirected connected graph G = V, E ; Number of arms K. Initialize: Center set C0 ; Unsatisfied agents S0 V . 1: t 0. 2: while St = do 3: Choose the next center: ct arg maxv St |N (v)|. 4: Update Ct+1 Ct {ct}. 5: Run Centers-to-Components with center set Ct+1, and obtain mass Mt+1 (v) for each v V . 6: Update St+1 v V | Mt+1 (v) < min {|N (v)| , K} min c Ct+1 dist G (v, c) 3 . 7: t t + 1. 8: end while 9: return C = Ct. The uninformed setting At first, it may seem that the uninformed setting can be solved the same way as the informed setting, with some distributed version of Compute-Centers-Informed. However, such algorithm will require Ω(N) steps in the worst case, since at each iteration only one agent becomes a center. In the informed setting we do not care about this, since the components are computed in advance. In the uninformed setting however, at each step of the algorithm the agents suffer a loss, and thus the regret bound will be at least linear in the number of agents, which can be very large. To avoid this problem, we need to add many centers each iteration, and not just one as in Compute Centers-Informed. To do this, we exploit the fact that there are only K possible values for a center s mass. In our algorithm, there are K iterations, and in each iteration t, as many agents as possible with degree K t become centers. To ensure the final center set is 2-independent, only a 2-MIS of the potential center agents are added to the center set each iteration. To compute a 2-MIS in a distributed manner, we use Luby s algorithm [Luby, 1986, Alon et al., 1986] on the sub-graph of G2 induced by the potential center agents. Briefly, at each iteration of Luby s algorithm, every potential center agent picks a number uniformly from [0, 1]. Agents that picked the maximal number among their neighbors of distance 2 join the 2-MIS, and their neighbors of distance 2 stop participating. A 2-MIS is computed after l 3 ln N m iterations with probability 1 δ. Each iteration requires exchanging 4 messages - 2 for communicating the random numbers and 2 for communicating the new agents in the 2-MIS. Hence, 4 l 3 ln N m steps suffice to compute a 2-MIS with probability 1 δ. A more detailed explanation of Luby s algorithm can be found in the supplementary material. We present Compute-Centers-Uninformed in Algorithm 5. Since this is a distributed algorithm, we have the variables C (v) and S (v) as indicators for whether v is a center or unsatisfied, respectively. 5 Regret analysis We will now provide an overview for the analysis of our algorithms. We remind that all proofs are differed to the supplementary material. 5.1 Individual regret of the center-based policy We start by bounding the expected regret of the agents when they are using the center-based policy. Theorem 5. Let T K2 ln K. Using the center-based policy, the regret of each agent v V satisfies (ln K) K M (v)T. Algorithm 5 Compute-Centers-Uninformed - agent v Parameters: Number of arms K; Upper bound on the total number of agents N; Time horizon T. Initialize: Center indicator C (v) FALSE; Unsatisfied indicator S (v) TRUE. 1: for 0 t K 1 do 2: Participate for 4 l 3 ln N KT m steps in Luby s algorithm on G2 St = {v V | S (v) = TRUE min {|N (v)| , K} = K t} , to compute Wt, a 2-MIS of St, with probability 1 1 T K . 3: If v Wt, set C (v) TRUE. 4: Participate in Centers-to-Components with center set Ct = {v V | C (v ) = TRUE}; obtain mass Mt (v) and whether minc Ct dist G (v, c) 3. 5: minc Ct dist G (v, c) 3 if and only if C2 (v) = nil in Centers-to-Components. 6: Update S (v) I Mt (v) < min {|N (v)| , K} min c Ct dist G (v, c) 3 . 7: end for 8: return C = CK 1. This individual regret bound holds simultaneously for all agents in the graph, and it depends only on the graph structure and components. 5.2 Analyzing Centers-to-Components We need to show the results of Centers-to-Components follow their definitions, and the derived components satisfy all the properties required by the center-based policy. The following lemma show it under some requirements from the center set C. Lemma 6. Let C V be a center set that is 2-independent, such that every v V holds minc C dist G (v, c) 6 ln K 1. Let C (v) , U (v) , M (v) be the results of Centers-to-Components. For each c C, let Vc be its corresponding component, namely, Vc = {v V | C (v) = c}. Then the following properties are satisfied: 1. {Vc | c C} are pairwise disjoint and V = S 2. N (c) Vc and Gc is connected for all c C. 3. M (v) = e 1 6 d(v)M (C (v)) and U (v) = arg minv N(v) VC(v) d (v ) for all v V \ C. 5.3 Analyzing Compute-Centers-Informed The first thing we need to show is that the center set returned by Compute-Centers-Informed satisfies the conditions of Lemma 6: Lemma 7. Let C V be the center set returned by Compute-Centers-Informed. Then: 1. C is 2-independent. 2. For all v V , minc C dist G (v, c) 6 ln K 1. Now, we can show that by using our informed graph partitioning algorithms, the mass of all agents is large: Theorem 8. Let C V be the center set returned by Compute-Centers-Informed, and let {Vc V | c C} be the components resulted from Centers-to-Components. For every v V : M (v) e 1 min {|N (v)| , K} . Together with Theorem 5, we obtain the desired regret bound. Corollary 9. Let T K2 ln K. Let C V be the center set returned by Compute-Centers Informed, and let {Vc V | c C} be the components resulted from Centers-to-Components. Using the center-based policy, we obtain for every v V : (ln K) 1 + K |N (v)| s 1 + K |N (v)| 5.4 Analyzing Compute-Centers-Uninformed First, we show that Compute-Centers-Uninformed terminates after a relatively small number of steps, and thus the loss suffered while running it is insignificant. Lemma 10. Compute-Centers-Uninformed runs for less than 12K ln K2 NT steps. As in the informed setting, we now need to show the center set resulted from Compute-Centers Uninformed satisfies the conditions of Lemma 6. Lemma 11. Let C V be the center set resulted from Compute-Centers-Uninformed, such that Luby s algorithm succeeded at all iterations of the algorithm. Then: 1. C is 2-independent. 2. For all v V , minc C dist G (v, c) 6 ln K 1. We can now obtain the same result as in the informed setting: Theorem 12. Let C V be the center set resulted from Compute-Centers-Uninformed, such that Luby s algorithm succeeded at all iterations of the algorithm, and also let {Vc V | c C} be the components resulted from Centers-to-Components. For every v V : M (v) e 1 min {|N (v)| , K} . Again we can use Theorem 5 to obtain the desired regret bound. Corollary 13. Let T K2 ln K and N N. Let C V be the center set resulted from Compute-Centers-Uninformed, and let {Vc V | c C} be the components resulted from Centersto-Components. Using the center-based policy, we obtain for every v V : K ln K2 NT + (ln K) 1 + K |N (v)| s 1 + K |N (v)| 5.5 Average regret of the center-based policy As mentioned before, we strictly improve the result of Cesa-Bianchi et al. [2019b], and our algorithms imply the same average expected regret bound. Corollary 14. Let T K2 ln K. Let C V be the center set resulted from Compute-Centers Informed or Compute-Centers-Uninformed, and let {Vc V | c C} be the components resulted from Centers-to-Components. Using the center-based policy, we get: v V RT (v) = e O 6 Conclusions We investigated the cooperative nonstochastic multi-armed bandit problem, and presented the centerbased cooperation policy (Algorithms 1 and 2). We provided partitioning algorithms that provably yield a low individual regret bound that holds simultaneously for all agents (Algorithms 3, 4 and 5). We express this bound in terms of the agents degree in the communication graph. This bound strictly improves a previous regret bound from [Cesa-Bianchi et al., 2019b] (Corollary 14), and also resolves an open question from that paper. Note that our regret bound in the informed setting does not depend on the total number of agents, N, and in the uninformed setting it depends on N only logarithmically. It is unclear whether in the uninformed setting, any dependence on N in the individual regret is required. Acknowledgments This work was supported in part by the Yandex Initiative in Machine Learning and by a grant from the Israel Science Foundation (ISF). Noga Alon, László Babai, and Alon Itai. A fast and simple randomized parallel algorithm for the maximal independent set problem. Journal of algorithms, 7(4):567 583, 1986. Peter Auer, Nicolo Cesa-Bianchi, Yoav Freund, and Robert E Schapire. The nonstochastic multiarmed bandit problem. SIAM journal on computing, 32(1):48 77, 2002. Orly Avner and Shie Mannor. Concurrent bandits and cognitive radio networks. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pages 66 81. Springer, 2014. Baruch Awerbuch and Robert Kleinberg. Competitive collaborative learning. Journal of Computer and System Sciences, 74(8):1271 1288, 2008. Ilai Bistritz and Amir Leshem. Distributed multi-player bandits-a game of thrones approach. In Advances in Neural Information Processing Systems, pages 7222 7232, 2018. Sébastien Bubeck, Nicolo Cesa-Bianchi, et al. Regret analysis of stochastic and nonstochastic multi-armed bandit problems. Foundations and Trends R in Machine Learning, 5(1):1 122, 2012. Nicolo Cesa-Bianchi and Gabor Lugosi. Prediction, learning, and games. Cambridge university press, 2006. Nicolò Cesa-Bianchi, Tommaso R Cesari, and Claire Monteleoni. Cooperative online learning: Keeping your neighbors updated. ar Xiv preprint ar Xiv:1901.08082, 2019a. Nicolo Cesa-Bianchi, Claudio Gentile, and Yishay Mansour. Delay and cooperation in nonstochastic bandits. The Journal of Machine Learning Research, 20(1):613 650, 2019b. Soummya Kar, H Vincent Poor, and Shuguang Cui. Bandit problems in networks: Asymptotically efficient distributed allocation rules. In 2011 50th IEEE Conference on Decision and Control and European Control Conference, pages 1771 1778. IEEE, 2011. Ravi Kumar Kolla, Krishna Jagannathan, and Aditya Gopalan. Collaborative learning of stochastic bandits over a social network. IEEE/ACM Transactions on Networking (TON), 26(4):1782 1795, 2018. Peter Landgren, Vaibhav Srivastava, and Naomi Ehrich Leonard. On distributed cooperative decisionmaking in multiarmed bandits. In 2016 European Control Conference (ECC), pages 243 248. IEEE, 2016a. Peter Landgren, Vaibhav Srivastava, and Naomi Ehrich Leonard. Distributed cooperative decisionmaking in multiarmed bandits: Frequentist and bayesian algorithms. In 2016 IEEE 55th Conference on Decision and Control (CDC), pages 167 172. IEEE, 2016b. Michael Luby. A simple parallel algorithm for the maximal independent set problem. SIAM journal on computing, 15(4):1036 1053, 1986. Jonathan Rosenski, Ohad Shamir, and Liran Szlak. Multi-player bandits a musical chairs approach. In International Conference on Machine Learning, pages 155 163, 2016. Anit Kumar Sahu and Soummya Kar. Dist-hedge: A partial information setting based distributed non-stochastic sequence prediction algorithm. In 2017 IEEE Global Conference on Signal and Information Processing (Global SIP), pages 528 532. IEEE, 2017. Yevgeny Seldin, Peter L Bartlett, Koby Crammer, and Yasin Abbasi-Yadkori. Prediction with limited advice and multiarmed bandits with paid observations. In ICML, pages 280 287, 2014. Balázs Szörényi, Róbert Busa-Fekete, István Heged us, Róbert Ormándi, Márk Jelasity, and Balázs Kégl. Gossip-based distributed stochastic bandit algorithms. In Journal of Machine Learning Research Workshop and Conference Proceedings, volume 2, pages 1056 1064. International Machine Learning Societ, 2013. VK Wei. A lower bound on the stability number of a simple graph. Technical report, Bell Laboratories Technical Memorandum 81-11217-9, Murray Hill, NJ, 1981.