# global_rewards_in_restless_multiarmed_bandits__8b3ca93b.pdf Global Rewards in Restless Multi-Armed Bandits Naveen Raman Carnegie Mellon University naveenr@cmu.edu Zheyuan Ryan Shi University of Pittsburgh ryanshi@pitt.edu Fei Fang Carnegie Mellon University feifang@cmu.edu Restless multi-armed bandits (RMAB) extend multi-armed bandits so pulling an arm impacts future states. Despite the success of RMABs, a key limiting assumption is the separability of rewards into a sum across arms. We address this deficiency by proposing restless-multi-armed bandit with global rewards (RMAB-G), a generalization of RMABs to global non-separable rewards. To solve RMAB-G, we develop the Linearand Shapley-Whittle indices, which extend Whittle indices from RMABs to RMAB-Gs. We prove approximation bounds but also point out how these indices could fail when reward functions are highly non-linear. To overcome this, we propose two sets of adaptive policies: the first computes indices iteratively, and the second combines indices with Monte-Carlo Tree Search (MCTS). Empirically, we demonstrate that our proposed policies outperform baselines and index-based policies with synthetic data and real-world data from food rescue. 1 Introduction Restless multi-armed bandits (RMAB) are models of resource allocation that combine multi-armed bandits with states for each bandit s arm. Such a model is restless because arms can change state even when not pulled. RMABs are an enticing framework because optimal decisions can be efficiently made using pre-computed Whittle indices [1, 2]. As a result, RMAB have been used in scheduling [3], autonomous driving [4], multichannel access [5], and maternal health [6]. The existing RMAB model assumes that rewards are separable into a sum of per-arm rewards. However, many real-world objectives are non-separable functions of the arms pulled. We find this situation in food rescue platforms. These platforms notify subsets of volunteers about upcoming food rescue trips and aim to maximize the trip completion rate [7]. When modeling this with RMABs, arms correspond to volunteers, arm pulls correspond to notifications, and the reward corresponds to the trip completion rate. The trip completion rate is non-separable as we only need one volunteer to complete each trip (more details in Section 3). Beyond food rescue, RMABs can potentially model problems in peer review [8], blood donation [9], and emergency dispatch [10], but the rewards are again non-separable. Motivated by these situations, we propose the restless multi-armed bandit with global rewards (RMABG), a generalization of RMABs to non-separable rewards. Whittle indices from RMABs [1] fail for RMAB-Gs due to the non-separability of the reward. Because of this, we propose a generalization of Whittle indices to global rewards called the Linear-Whittle and Shapley-Whittle indices. Our approximation bounds demonstrate that these policies perform well for near-linear rewards but struggle for highly non-linear rewards. We address this by developing adaptive policies that combine 38th Conference on Neural Information Processing Systems (Neur IPS 2024). Linearand Shapley-Whittle indices with search techniques. We empirically verify that adaptive policies outperform index-based and baseline approaches on synthetic and food rescue datasets. 1 Our contributions are: (1) We propose the RMAB-G problem, which extends RMABs to situations with global rewards. We additionally characterize the difficulty of solving and approximating RMAB-Gs; (2) We develop the Linear-Whittle and Shapley-Whittle indices, which extend Whittle indices for global rewards, and detail approximation bounds; (3) We design a set of adaptive policies which combine Linearand Shapley-Whittle indices with greedy and Monte Carlo Tree Search (MCTS) algorithms; and (4) We empirically demonstrate that adaptive policies improve upon baselines and pre-computed index-based policies for the RMAB-G problem with synthetic and food rescue datasets. 2 Background and Notation An instance of restless multi-armed bandits (RMABs) consists of N arms, each of which is defined through the following Markov Decision Process (MDP): (S, A, Ri, Pi, γ). Here, S is the state space, A is the action space, Ri : S A R is a reward function, Pi : S A S [0, 1] is a transition function detailing the probability of an arm transitioning into a new state, and γ [0, 1) is a discount factor. For a time period t, we define the state of all arm as s(t) = {s(t) 1 , , s(t) N }, s(t) i S, and for each round, a planner takes action a(t) = {a(t) 1 , , a(t) N }, a(t) i A; we drop the superscript t when clear from context. Following prior work [11, 12], we assume A = {0, 1} throughout this paper to represent not pulling and pulling an arm respectively. A planner chooses a(t) subject to a budget K, so that at most K arms can be pulled in each time step: PN i=1 a(t) i K, t. The objective is to find a policy, π : SN AN that maximizes the discounted total reward, summed over all arms: max π E(s,a) (P,π)[ i=1 γt Ri(s(t) i , a(t) i )] (1) To solve an RMAB, one can pre-compute the Whittle index for each arm i and possible state si: wi(si) = min w {w|Qi,w(si, 0) = Qi,w(si, 1)}, where Qi,w(si, ai) = wai + Ri(si, ai) + s Pi(si, ai, s )Vi,w(s ), and Vi,w(s ) = max a Qi,w(s , a). Here, Qi,w(si, ai) represents the expected future reward for playing action ai, given a penalty w for pulling an arm. The Whittle index computes the minimum penalty needed to prevent arm pulls; larger penalties (wi(si)) indicate more value for pulling an arm. In each round, the planner pulls the arms with the K largest Whittle indices given the states of the arms. Such a Whittle index-based policy is asymptotically optimal [1, 2]. 3 Defining Restless Multi-Armed Bandits with Global Rewards Objectives in RMABs are written as the sum of per-arm rewards (Equation 1), while real-world objectives involve non-separable rewards. For example, food rescue organizations, which notify volunteers about food rescue trips, aim to maximize the trip completion rate, i.e., the fraction of completed trips. Suppose we model volunteers as arms and volunteer notifications as arm pulls. Then arm states correspond to volunteer engagement (e.g., active or inactive) and rewards are the probability that any notified volunteers complete the rescue. Here, we cannot split the reward into per-volunteer functions. Inspired by situations such as food rescue, we propose the restless multi-armed bandit with global rewards (RMAB-G) problem, which generalizes RMABs to problems with a global reward function: Definition 1. We define an instance of RMAB-G through the following MDP for each arm: (S, A, Ri, Rglob, Pi, γ). The aim is to find a policy, π : SN AN, that maximizes: max π E(s,a) (P,π)[ t=0 γt(Rglob(s(t), a(t)) + i=1 Ri(s(t) i , a(t) i ))] (2) Throughout this paper, we focus on scenarios where Rglob(s, a) is monotonic and submodular in a; Rglob(s, a) = FS({i|ai = 1}), where F is submodular and monotonic. Such rewards have 1All code is available here https://github.com/naveenr414/food-rescue-rmab diminishing returns as extra arms are pulled. We select this because a) monotonic and submodular functions can be efficiently optimized [13] and b) submodular functions are ubiquitous [14, 15, 8]. Our problem formulation can model various real-world scenarios including notification decisions in food rescue and reviewer assignments in peer review [8]. In food rescue, let pi(si) be the probability that a volunteer picks up a given trip, given their state, si. The global reward is the probability that any notified volunteer accepts a trip: Rglob(s, a) = 1 QN i=1(1 aipi(si)). In peer review, journals select reviewers for submissions where selection impacts future reviewer availability [8]. The goal is to select a subset of available reviewers with comprehensive subject-matter expertise. If reviewer expertise is a set Yi, then we maximize the overlap between the required expertise, Z, and the expertise across selected (ai = 1) and available (si = 1) reviewers: Rglob(s, a) = | S i,si=1,ai=1 Yi Z|. Our formulation inherits the per-arm reward Ri from RMABs as there may be per-arm rewards. For example, food rescue platforms might care about the number of active volunteers, so Ri(si, ai) = si. Let R(s, a) = Rglob(s, a) + PN i=1 Ri(si, ai). We characterize the difficulty of RMAB-G: Theorem 1. The restless multi-armed bandit with global rewards is PSPACE-Hard. Theorem 2. No polynomial time algorithm can achieve better than a 1 1/e approximation to the restless multi-armed bandit with global rewards problem. Theorem 3. Consider an RMAB-G instance (S, A, Ri, Rglob, Pi, γ). Let Qπ(s, a) = E(s ,a ) (P,π)[P t=0 γt R(s , a ))|s(0) = s, a(0) = a] and let π be the optimal policy. Let a = arg maxa Qπ (s, a). Suppose Pi(s, 1, 1) Pi(s, 0, 1) s, i, Pi(1, a, 1) Pi(0, a, 1) i, a, R(s, a) monotonic and submodular in s and a, and g(s) = max a R(s, a) submodular in s. Then, with oracle access to Qπ we can compute ˆa in O(N 2) time so Qπ (s, ˆa) (1 1 e)Qπ (s, a ). Theorem 1 and 2 demonstrate the difficulty of finding solutions to RMAB-Gs while Theorem 3 describes a potential panacea via Qπ . Leveraging Qπ is a polynomial-time algorithm, but the exponential state space makes it difficult to learn such a Q, motivating the need for better algorithms (we empirically demonstrate this in Section 6.1). In essence, Lemma 3 demonstrates that it is possible to achieve a 1 1 e approximation for the Q-value; however, even computing such an approximation is difficult, and motivates the need for computationally feasible solutions. Our assumptions for Lemma 3 essentially capture the idea that rewards are submodular in both the state s and action a; these assumptions are satisfied for all of our synthetic experiments in Section 6. We prove Theorem 1 through reduction from RMAB, Theorem 2 through reduction from submodular optimization, and Theorem 3 through induction on value iteration. Full proofs are in Appendix K. 4 Linear-Whittle and Shapley-Whittle Indices 4.1 Defining Linear-Whittle and Shapley-Whittle Because the Whittle index policy is asymptotically optimal for RMABs [2], we investigate its extension to RMAB-Gs. We develop two extensions: the Linear-Whittle policy, which computes marginal rewards for each arm, and the Shapley-Whittle policy, which computes Shapley values for each arm. To define Linear-Whittle, we let the marginal reward be pi(si) = max s |s i=si Rglob(s , ei), where ei is a vector with 1 at index i and 0 elsewhere. The Linear-Whittle policy linearizes the global reward by optimistically estimating Rglob(s, a) PN i=1 pi(si)ai and computing the Whittle index from this: Definition 2. Linear-Whittle - Consider an instance of RMAB-G (S, A, Ri, Rglob, Pi, γ). Define QL i,w(si, ai, pi) = aiw + Ri(si, ai) + aipi(si) + γ P s i Pi(si, ai, s i)Vi,w(s i), where Vi,w(s i) = max a i QL i,w(s i, a i, pi). Then the Linear-Whittle index is w L i (si, pi) = min w {w|QL i,w(si, 0, pi) > QL i,w(si, 1, pi)}. The Linear-Whittle policy pulls arms with the K highest Linear-Whittle indices. To improve on the Linear-Whittle policy, we develop the Shapley-Whittle policy, which uses Shapley values to linearize the global reward. The Linear-Whittle policy linearizes global rewards via pi(si), but this approximation could be loose as PN i=1 pi(si)ai Rglob(s, a). The Shapley value, ui(si), improves this by averaging marginal rewards across subsets of arms (see Shapley et al. [16]). Let be the element-wise maximum, so a ei means arms in a and arm i are pulled. Define: ui(si) = min s |s i=si a {0,1}N,ai=0, a 1 K 1 a 1!(N a 1 1)! N! (N K)! (Rglob(s , a ei) Rglob(s , a)) Here, Rglob(s , a ei) Rglob(s , a) captures the added value of pulling arm i, when a are already pulled. While the Linear-Whittle policy estimates the global reward as Rglob(s, a) PN i=1 pi(si)ai, the Shapley-Whittle policy estimates the global reward as Rglob(s, a) PN i=1 ui(si)ai. The Linear Whittle policy overestimates marginal contributions, as PN i=1 pi(si)ai Rglob(s, a) (proof in Appendix L), so by using Shapley values, we take a pessimistic approach (by taking the minimum over all s ). This approach could lead to more accurate estimates of Rglob(s, a) because marginal contributions are averaged across many combinations of arms: Definition 3. Shapley-Whittle - Consider an instance of RMAB-G (S, A, Ri, Rglob, Pi, γ). Let QS i,w(si, ai, ui) = aiw + Ri(si, ai) + aiui(si) + γ P s i Pi(si, ai, s i)Vi,w(s i), where Vi,w(s i) = max a i QS i,w(s i, a i, ui). Then the Shapley-Whittle index is w S i (si, ui) = min w {w|QS i,w(si, 0, ui) > QS i,w(si, 1, ui)}. The Shapley-Whittle policy pulls arms with the K highest Shapley-Whittle indices. 4.2 Approximation Bounds To characterize the performance of our policies, we prove approximation bounds (proofs in Appendix L). We first define three properties that characterize situations when policies based on the Whittle index are asymptotically optimal; we use these characteristics to define when our Linearand Shapley-Whittle policies perform well: 1. Indexability refers to a property of an RMAB where increasing the penalty for pulling arms, w, from 0 to monotonically changes the arms that are pulled from all the arms to none of the arms (see Akbarzadeh and Mahajan [17] for details) 2. Irreducibility is a property of a Markov Chain if every pair of states is visitable; that is, there exists no subset of sink states such that one cannot escape from these sink states 3. Uniform Global Attractor is a property of an RMAB, where the state distribution converges to the optimal state distribution (in terms of expected reward) from any initial state; for more discussion, see Gast et al. [18]. We define two Theorems which capture lower bounds on the performance of our policies: Theorem 4. For any fixed set of transitions, P, let OPT = max π E(s,a) (P,π)[ 1 T PT 1 t=0 R(s, a)] for some T. For an RMAB-G (S, A, Ri, Rglob, Pi, γ), let R i(si, ai) = Ri(si, ai) + pi(si)ai, and let the induced linear RMAB be (S, A, R i, Pi, γ). Let πlinear be the Linear-Whittle policy, and let ALG = E(s,a) (P,πlinear)[ 1 T PT 1 t=0 R(s, a)]. Define βlinear as βlinear = min s SN,a [0,1]N, a 1 K R(s, a) PN i=1(Ri(si, ai) + pi(si)ai) (4) If the induced linear RMAB is irreducible and indexable with the uniform global attractor property, then ALG βlinear OPT asymptotically in N for any set of transitions, P. Theorem 5. For any fixed set of transitions, P, let OPT = max π E(s,a) (P,π)[ 1 T PT 1 t=0 R(s, a)] for some T. For an RMAB-G (S, A, Ri, Rglob, Pi, γ), let R i(si, ai) = Ri(si, ai) + ui(si)ai, and let the induced Shapley RMAB be (S, A, R i, Pi, γ). Let πshapley be the Shapley-Whittle policy, and let ALG = E(s,a) (P,πshapley)[ 1 T PT 1 t=0 R(s, a)]. Define βshapley as βshapley = min s SN,a [0,1]N, a 1 K R(s,a) PN i=1(Ri(si,ai)+ui(si)ai) max s SN,a [0,1]N, a 1 K R(s,a) PN i=1(Ri(si,ai)+ui(si)ai) (5) If the induced Shapley RMAB is irreducible and indexable with the uniform global attractor property, then ALG βshapley OPT asymptotically in N for any choice of transitions, P. Corollary 5.1. Let πlinear be the Linear-Whittle policy, ALG = E(s,a) (P,πlinear)[ 1 T PT 1 t=0 R(s, a)] and OPT = max π E(s,a) (P,π)[ 1 T PT 1 t=0 R(s, a)] for some T. Then ALG OPT K for any P. β lower bounds policy performance; small β does not guarantee poor performance but large β guarantees good performance. For example, βlinear and βshapley are near 1 for near-linear global rewards, so Linearand Shapley-Whittle policies perform well in those situations. We prove all three Theorems by bounding the gap between the linear (or Shapley) approximation and R(s, a) (proofs in Appendix L). Our key insight is that we can bound the performance of Linearand Shapely-Whittle policies using similarities between the global rewards and the rewards of an induced RMAB. We note that the assumptions made by all Theorems are the same assumptions needed for the optimality of Whittle indices [2]. We present an example of these lower bounds: Example 4.1. Consider an N = 4, K = 2 scenario for RMAB-G with S = {0, 1}. Let Rglob(s, a) = | S i|ai=1,si=1 Yi| and Ri(si, ai) = 0. Here, Yi are sets corresponding to each arm, with Y1 = Y2 = {1, 2, 3}, Y3 = {1, 2} and Y4 = {3, 4}. Let si = 1 for all i, and consider the Linear-Whittle policy. Next, note that pi(si) = R(siei, ei), so p1(1) = p2(1) = 3, p3(1) = p4(1) = 2. Now let a = [1, 1, 0, 0] and let s = [1, 1, 1, 1]. Here, R(s, a) = |Y1 Y2| = 3, while PN i=1 pi(si)ai = 6. We note that we select such an a and s because this minimizes the lower bound from βlinear; other selections will not lead to the correct lower bound due to the minimum in the equation for βlinear. Therefore, βlinear = 3 2, so the Linear-Whittle policy is a 1 2 approximation to this problem. We describe upper bounds and apply these bounds to Example 4.1. Theorem 6. Let ˆa(s) = arg max a [0,1]N, a 1 K PN i=1(Ri(si, ai) + pi(si)ai). For a set of transi- tions P = {P1, P2, , PN}, let OPT = max π E(s,a) (P,π)[P t=0 γt R(s, a)|s(0)] and ALG = E(s,a) (P,πlinear)[P t=0 γt R(s, a)|s(0)]. Define θlinear as follows: θlinear = min s SN R(s, ˆa(s)) max a [0,1]N, a 1 K R(s, a) (6) Then there exists some transitions, P, and initial state s(0), so that ALG θlinear OPT Theorem 7. Let ˆa(s) = arg max a [0,1]N, a 1 K PN i=1(Ri(si, ai) + ui(si)ai). For a set of transitions P = {P1, P2, , PN} and initial state s(0), let OPT = max π E(s,a) (P,π)[P t=0 γt R(s, a)|s(0)] and ALG = E(s,a) (P,πshapley)[P t=0 γt R(s, a)|s(0)]. Let θshapley be: θshapley = min s R(s, ˆa(s)) max a [0,1]N, a 1 K R(s, a) (7) Then there exists some transitions, P and initial state s(0) so that ALG θshapley OPT Example 4.2. Recall the situation from Example 4.1. Let si = 1 for all i, and consider the Linear Whittle policy. Construct construct Pi, so that Pi(s, a, 1) = 1 i, s, a or in other words, s(t) i = 1 i, t. The Linear-Whittle policy will pull arms 1 and 2, as p1(1) + p2(1) pi(1) + pj(1) i, j. This results in an action a = [1, 1, 0, 0], so that R(s, a) = 3, while the optimal action is a = [0, 0, 1, 1] with R(s, a ) = 4. Computing θlinear matches this upper bound of 3 4 in this scenario. Our upper bounds demonstrate that, for any reward function, there exists a set of transitions where our policies perform a θ-fraction worse than optimal. θlinear is small when the optimal action diverges from the action taken by the Linear-Whittle policy (analogously for Shapley-Whittle). Such situations occur because policies fail to account for inter-arm interactions (e.g. Example 4.2). We present lower bounds using the average reward because Whittle optimality holds in that situation. Otherwise, we follow prior work, and focus on discounted reward [11] (for more discussion see Ghosh et al. [19]) 5 Adaptive Approaches for RMAB-G 5.1 Limitations of Pre-Computed Index-Based Policies Despite the approximation bounds given in Appendix L, Linearand Shapley-Whittle policies can exhibit poor performance when upper bounds θlinear and θshapley are small. We detail such a situation: Lemma 8. (Informal) Let π {πlinear, πshapley} be either the Linear or Shapley-Whittle policies. There exist problem settings where π will achieve an approximation ratio no better than 1 γ 1 γK . In general, pre-computed index-based policies perform poorly because of their inability to consider inter-arm interactions (proof in Appendix L). This approximation ratio matches lower bounds, lim γ 1 1 γ 1 γK = 1 K = βlinear = βshapley, which motivates new policies which select arms adaptively. 5.2 Iterative Whittle Indices Lemma 8 motivates the need for policies that select arms adaptively and go beyond pre-computed indices. Based on this, we propose two new policies: Iterative Linear-Whittle and Iterative Shapley Whittle. Both policies compute current time step rewards, ri, based on the arms pulled, X. Definition 4. Iterative Linear-Whittle - Consider an instance of RMAB-G (S, A, Ri, Rglob, Pi, γ). Consider a state s, and construct a new state, s, with P si = Pi. Let QIL i,w( si, ai, pi, r L i , X) = QL i,w(si, ai, pi) + ai(r L i (s, X) pi(si)) and QIL i,w(si, ai, pi, r L i , X) = QL i,w(si, ai, pi) for si = si. Then the Iterative Linear-Whittle index is w IL i (si, pi, X) = min w {w|QIL i,w(si, 0, pi, r L i , X) > QIL i,w(si, 1, pi, r L i , X)}, where r L i (s, X) = Rglob(s, ei Wj l=1 exl) Rglob(s, Wj l=1 exl) Definition 5. Iterative Shapley-Whittle - Consider an instance of RMAB-G (S, A, Ri, Rglob, Pi, γ), with ui(si) defined by Equation 3. Construct a new state, si, with P si = Pi. Let QIS i,w( si, ai, ui, r S i , X) = QS i,w(si, ai, ui) + ai(r S i (si, X) ui(si)) and QIS i,w(si, ai, ui, r S i , X) = QS i,w(si, ai, ui) for si = si. Then the Iterative Shapley-Whittle Index is w IS i (si, ui, X) = min w {w|QIS i,w(si, 0, ui, r S i , X) > QIS i,w(si, 1, ui, r S i , X)}. Here ri(si, X) is computed as: min s |s i=si a {0,1}N, a 1 K 2 ai=a X1=a X2= =a Xj =0 a 1!(n a 1 1)! N! (N K 1)! (Rglob(s , a ei X) Rglob(s , a X)) (8) By incorporating the arms pulled, X, both policies better estimate the present reward. Here, QIL i,w( si, ai, pi, r L i , X) represents the expected discounted reward when receiving r L i (s, X) currently and pi(si) in the future (similarly for Shapley-Whittle). To model this, we construct a new state, si, which mirrors si except for the reward in the current timestep. The Iterative Linear-Whittle and Shapley-Whittle policies then iteratively select arms that maximize w IL i and w IS i respectively and stop after selecting K arms. 5.3 MCTS-based Approaches To improve upon the greedy selection done by iterative policies, we combine Monte Carlo Tree Search (MCTS) with Whittle indices. MCTS allows us to look beyond greedy selection by searching through various arm combinations. We propose two policies: MCTS Linear-Whittle and MCTS Shapley-Whittle. For both, we search through arm combinations, a, compute the present value (R(s, a)), and estimate the future value via Linearand Shapley-Whittle indices. Nodes in the MCTS tree correspond to arm combinations, edges to individual arms, and leaf nodes to actions a. Our implementation of MCTS matches the upper confidence tree algorithm for selection, expansion, and backpropogation [20], while we differ for rollout (details in Appendix C). During rollout, we select nodes according to an ϵ-greedy algorithm [21] and leverage the Linearand Shapley-Whittle policies for greedy selection. We do so because ϵ-greedy is a simple strategy that could potentially improve upon random rollouts. Once at a leaf node, the reward is r = R(s, a)+PN i=1[QL i,0(si, ai) pi(si)ai Ri(si, ai)] for MCTS Linear-Whittle and r = R(s, a) + PN i=1[QS i,0(si, ai) ui(si)ai Ri(si, ai)] Probability 1.0 Linear 1.00 Probability 1.00 Normalized Reward Vanilla Whittle Greedy MCTS DQN DQN Greedy Linear-Whittle Shapley-Whittle Iterative Linear Iterative Shapley MCTS Linear MCTS Shapley Optimal Figure 1: We compare baselines to our index-based and adaptive policies across four reward functions. All of our policies outperform baselines. Across all rewards, our best policy is within 3% of optimal for N = 4. Among our policies, Iterative and MCTS Shapley-Whittle consistently perform best. for MCTS Shapley-Whittle. Here, R(s, a) is the known current value, and PN i=1(QL i,0(si, ai) pi(si)ai Ri(si, ai)) is the estimated future value when accounting for the present reward R(s, a). 6 Experiments We evaluate our policies across both synthetic and real-world datasets. We compare our six policies (Section 4.1, Section 5.2, and Section 5.3) to the following baselines (more details in Appendix B): 1. Random - We uniform randomly select K arms at each timestep. 2. Vanilla Whittle - We run the vanilla Whittle index, which optimizes only for Ri(si, ai). 3. Greedy - We select actions which have the highest value of pi(si). 4. MCTS - We run MCTS up to some depth, c K, where c is a constant. 5. DQN - We train a Deep Q Network (DQN) [22], and provide details in Appendix B. 6. DQN Greedy - Inspired by Theorem 3, we first learn a DQN then greedily optimize Q(s, a). 7. Optimal - We compute the optimal policy through value iteration when N 4. We run all experiments for 15 seeds and 5 trials per seed. We report the normalized reward, which is the accumulated discounted reward divided by that of the random policy. This normalization is due to differences in magnitudes within and across problem instances and we use discounted rewards due to prior work [11]. We additionally compute the standard error across all runs. 6.1 Results with Synthetic Data To understand performance across synthetic scenarios, we develop synthetic problem instances that leverage each of the following global reward functions (details on mi and Yi choices in Appendix A): 1. Linear - Given m1, , m N R, then Rglob(s, a) = PN i=1 misiai. 2. Probability - Given m1, , m N [0, 1], then Rglob(s, a) = 1 QN i=1(1 miaisi). 3. Max - Given m1, , m N R, then Rglob(s, a) = maxi siaimi. 4. Subset - Given sets, Y1, , YN {1, , m}, m Z, then Rglob(s, a) = | i|siai=1 Yi|. We select these rewards, as they range from linear or near-linear (Linear and Probability) to highly non-linear (Max and Subset). The reward functions can also be employed to model real-world situations; the Probability reward models food rescue and the Subset reward models peer review (see Section 3). For all rewards, we let Ri(s, a) = si/N and S = {0, 1}, which matches prior work [12]. We construct synthetic transitions parameterized by q, so Pi(0, 0, 1) U(0, q), where U is the uniform distribution (see Appendix A). We construct transition probabilities so that having ai = 1 or 0.65 0.70 0.75 0.80 Linearity ( linear) (4 arms) Normalized Reward Linear-Whittle Shapley-Whittle Iterative Linear Iterative Shapley MCTS Linear MCTS Shapley 0.4 0.6 0.8 Linearity ( linear) (10 arms) Figure 2: We plot policy performance for instances of the Subset reward which vary in linearity. We see that the Iterative and MCTS Shapley-Whittle policies outperform alternatives for non-linear rewards (small θlinear) while policy performances converge for linear rewards (large θlinear). 4 10 15 20 Number of Arms Time Taken (s) Iterative Shapley Iterative Shapley 100 Iterative Shapley 10 Iterative Shapley 1 MCTS Shapley MCTS Shapley 40 MCTS Shapley 4 4 10 15 20 Number of Arms Normalized Reward Figure 3: We compare the efficiency and performance of Iterative Shapley-Whittle methods which vary in Monte Carlo samples used for Shapley estimation and MCTS Shapley-Whittle methods which vary in MCTS iterations. While Iterative Shapley-Whittle is the slowest, decreasing the number of Monte Carlo samples can improve efficiency without impacting performance. si = 1 only improves transition probabilities (see Wang et al. [11]). We vary q in Appendix G and find that our choice of q does not impact findings. Unless otherwise specified, we fix K = N/2. Figure 1 shows that our policies consistently outperform baselines. Among our policies, Iterative and MCTS Shapley-Whittle policies perform best. For N = 10, Iterative and MCTS Shapley-Whittle significantly outperform baselines (paired t-test p < 0.04) and index-based policies except for the Linear reward (p < 0.001). DQN-based policies regress from on average 4% worse for N = 4 to on average 9% worse for N = 10 compared to our best policy, which occurs due to the exponential state space. In Appendix I, we consider more combinations of N and K and find similar results. To understand policy performance across rewards, we analyze how reward linearity impacts performance. We quantify linearity via θlinear (see Theorem 6) and compute θlinear for Subset instances varying in Yi (see Appendix A). For non-linear reward functions (small θlinear), Iterative and MCTS-Shapley-Whittle are best, outperforming Linear-Whittle by 56%, while for near-linear reward functions (large θlinear), all policies are within 18% of each other (Figure 2, N = 10). Appendix H shows that, for some rewards, adaptive policies outperform non-adaptive policies by 50%. To understand the tradeoff between efficiency and performance we plot the time taken and normalized reward for various N (Figure 3). We evaluate variants of Iterative Shapley-Whittle and MCTS Shapley-Whittle because both perform well yet are computationally expensive. For Iterative Shapley Whittle, 1000 samples are used for Shapley estimation by default, so we consider variants that use 100, 10, and 1 samples. For MCTS Shapley-Whittle, we run 400 iterations of MCTS by default, so we consider variants that run 40 and 4 iterations. Iterative Shapley-Whittle policies are the slowest Vanilla Whittle Greedy MCTS DQN Linear Whittle Notifications 0.975 0.012 1.829 0.016 1.047 0.012 1.547 0.025 1.932 0.017 Phone Calls 0.989 0.009 1.228 0.008 1.070 0.016 1.209 0.007 1.288 0.008 Shapley Whittle Iter. Linear Iter. Shapley MCTS Linear MCTS Shapley Notifications 1.931 0.016 1.921 0.017 Phone Calls 1.290 0.008 1.287 0.009 1.294 0.009 1.291 0.008 1.293 0.008 Table 1: We evaluate policies in real-world food rescue contexts by developing situations that mirror notifications (N = 100, K = 50) and phone calls (N = 20, K = 10). Our policies outperform baselines for both situations and achieve similar results due to the nearly linear reward function. but can be made faster without impacting performance by reducing the number of samples. Our results demonstrate how we can develop efficient adaptive policies which perform well. 6.2 Results with Real-World Data in Food Rescue To evaluate our policies in real-world contexts, we apply our policies to a food rescue dataset. We leverage data from a partnering multi-city food rescue organization and construct an RMAB-G instance using their data (details in Appendix D). Here, arms correspond to volunteers, and states correspond to engagement, with si = 1 indicating an engaged volunteer. Notifications are budget-limited because sending out too many notifications can lead to burned-out volunteers. We compute transition probabilities from volunteer participation data and compute match probabilities from volunteer response rates. Using this, we study whether our policies maintain a high trip completion rate and volunteer engagement. We assume that food rescue organizations are indifferent to the tradeoff between trip completion and engagement. Because such an assumption might not hold in reality, we explore alternative tradeoffs in Appendix I. If volunteer i is both notified (ai = 1) and engaged (si = 1), then they match to a trip with probability mi. We do so because unengaged volunteers are less likely to accept food rescue trips. Under this formulation, our global reward is the probability that any volunteer accepts a trip, which is exactly the Probability reward. For engagement, we let Ri(si, ai) = si/N, which captures the fraction of engaged volunteers. We model two food rescue scenarios, and detail scenarios with more volunteers in Appendix J: Notifications - Food rescue organizations deliver notifications to volunteers as new trips arise [7]. To model this scenario, we consider N = 100 volunteers and K = 50 notifications. Phone Calls - When a trip remains unclaimed after notifications, food rescue organizations call a small set of experienced volunteers. We model this with N = 20, K = 10, and using only arms corresponding to experienced volunteers with more than 100 trips completed (details in Appendix D). In Table 1, we compare our policies against baselines and find that our policies achieve higher normalized rewards. All of our policies outperform all baselines by at least 5%. Comparing between policies, we find that Iterative and MCTS Shapley-Whittle perform best, performing slightly better than the Shapley-Whittle policy. Currently, food rescue organizations perform notifications greedily, so an RMAB-G framework could improve the engagement and trip completion rate. 7 Related Works and Problem Setup Multi-armed Bandits In our work, we explore multi-armed bandits with two properties: combinatorial actions [23] and non-separable rewards. Combinatorial bandits can be solved through follow-the-perturbed-leader algorithms [24], which are necessary because the large action space causes traditional linear bandit algorithms to be inefficient [21]. Prior work has tackled non-separable rewards in slate bandits [25], through an explore-then-commit framework, and adversarial combinatorial bandits [26], from a minimax optimality perspective. In this work, we leverage the structure of RMAB-Gs to tackle both combinatorial actions and non-separable rewards. Submodular Bandits Additionally related to our work are prior investigations into submodular bandits. Submodular bandits combine submodular functions with bandits and have been investigated from an explore-then-commit perspective [27] and through upper confidence bound (UCB) algorithms [28]. Similar to prior work, we focus on optimizing submodular functions under uncertainty. However, prior work focuses on learning the reward structure when the submodular function is unknown, while we focus on learning optimal decisions when state transitions are stochastic. Restless Multi-Armed Bandits RMABs extend multi-armed bandits to situations where unpulled arms can transition between states [1]. While finding optimal solutions to this problem is PSPACEHard [29], index-based solutions are efficient and asymptotically optimal [2]. As a result, RMABs have been used across a variety of domains [4, 3, 30] and have been deployed in public health scenarios [12]. While prior work has relaxed assumptions by considering RMABS that have contexts [31], combinatorial arms [32], graph-based interactions [33], and global state [34], we focus on combinatorial situations with a global reward. 8 Limitations and Conclusion Limitations - While we present approximation bounds for index-based policies, our adaptive policies lack theoretical guarantees. Developing bounds for adaptive policies is difficult due to the difficulty in leveraging Whittle index optimality. However, these bounds may explain the empirical performance of adaptive policies. Our empirical analysis is limited to one real-world dataset (food rescue). We aimed to mitigate this by exploring different scenarios within food rescue, but empirical evaluation in domains such as peer review would be valuable. Finally, evaluating policies when global rewards are neither submodular nor monotonic would improve our understanding of RMAB-G. Conclusion - RMABs fail to account for situations with non-separable global rewards, so we propose the RMAB-G problem. We tackle RMAB-G by developing index-based, iterative, and MCTS policies. We prove performance bounds for index-based policies and empirically find that iterative and MCTS policies perform well on synthetic and real-world food rescue datasets. Our results show how RMABs can extend to scenarios where rewards are global and non-separable. Acknowledgements We thank Gati Aher and Gaurav Ghosal for the comments they provided on an earlier draft of this paper. We additionally thank Milind Tambe for his comments. This work was supported in part by NSF grant IIS-2200410. Co-author Raman is also supported by an NSF GRFP Fellowship. Co-author Shi is supported in part by a Pitt Mascaro Center for Sustainable Innovation fellowship. [1] Peter Whittle. Restless bandits: Activity allocation in a changing world. Journal of applied probability, 25(A):287 298, 1988. [2] Richard R Weber and Gideon Weiss. On an index policy for restless bandits. Journal of applied probability, 27(3):637 648, 1990. [3] Vivek S Borkar, Gaurav S Kasbekar, Sarath Pattathil, and Priyesh Y Shetty. Opportunistic scheduling as restless bandits. IEEE Transactions on Control of Network Systems, 5(4):1952 1961, 2017. [4] Mushu Li, Jie Gao, Lian Zhao, and Xuemin Shen. Adaptive computing scheduling for edgeassisted autonomous driving. IEEE Transactions on Vehicular Technology, 70(6):5318 5331, 2021. [5] Keqin Liu and Qing Zhao. Indexability of restless bandit problems and optimality of whittle index for dynamic multichannel access. IEEE Transactions on Information Theory, 56(11): 5547 5567, 2010. [6] Aditya Mate, Lovish Madaan, Aparna Taneja, Neha Madhiwalla, Shresth Verma, Gargi Singh, Aparna Hegde, Pradeep Varakantham, and Milind Tambe. Field study in deploying restless multiarmed bandits: Assisting non-profits in improving maternal and child health. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 12017 12025, 2022. [7] Zheyuan Ryan Shi, Leah Lizarondo, and Fei Fang. A recommender system for crowdsourcing food rescue platforms. In Proceedings of the Web Conference 2021, pages 857 865, 2021. [8] Justin Payan and Yair Zick. I will have order! optimizing orders for fair reviewer assignment. ar Xiv preprint ar Xiv:2108.02126, 2021. [9] Duncan C Mc Elfresh, Christian Kroer, Sergey Pupyrev, Eric Sodomka, Karthik Abinav Sankararaman, Zack Chauvin, Neil Dexter, and John P Dickerson. Matching algorithms for blood donation. In Proceedings of the 21st ACM Conference on Economics and Computation, pages 463 464, 2020. [10] Pieter L van den Berg, Shane G Henderson, Caroline Jagtenberg, and Hemeng Li. Modeling emergency medical service volunteer response. Available at SSRN 3825060, 2021. [11] Kai Wang, Lily Xu, Aparna Taneja, and Milind Tambe. Optimistic whittle index policy: Online learning for restless bandits. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 10131 10139, 2023. [12] Aditya Mate, Jackson Killian, Haifeng Xu, Andrew Perrault, and Milind Tambe. Collapsing bandits and their application to public health intervention. Advances in Neural Information Processing Systems, 33:15639 15650, 2020. [13] George L Nemhauser, Laurence A Wolsey, and Marshall L Fisher. An analysis of approximations for maximizing submodular set functions i. Mathematical programming, 14:265 294, 1978. [14] Alain Chateauneuf and Bernard Cornet. Submodular financial markets with frictions. Economic Theory, 73(2):721 744, 2022. [15] Manish Prajapat, Mojmír Mutn y, Melanie N Zeilinger, and Andreas Krause. Submodular reinforcement learning. ar Xiv preprint ar Xiv:2307.13372, 2023. [16] Lloyd S Shapley et al. A value for n-person games. 1953. [17] Nima Akbarzadeh and Aditya Mahajan. Restless bandits with controlled restarts: Indexability and computation of whittle index. In 2019 IEEE 58th conference on decision and control (CDC), pages 7294 7300. IEEE, 2019. [18] Nicolas Gast, Bruno Gaujal, and Chen Yan. Exponential convergence rate for the asymptotic optimality of whittle index policy. ar Xiv preprint ar Xiv:2012.09064, 2020. [19] Abheek Ghosh, Dheeraj Nagaraj, Manish Jain, and Milind Tambe. Indexability is not enough for whittle: Improved, near-optimal algorithms for restless bandits. ar Xiv preprint ar Xiv:2211.00112, 2022. [20] Levente Kocsis and Csaba Szepesvári. Bandit based monte-carlo planning. In European conference on machine learning, pages 282 293. Springer, 2006. [21] Tor Lattimore and Csaba Szepesvári. Bandit algorithms. Cambridge University Press, 2020. [22] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. Playing atari with deep reinforcement learning. ar Xiv preprint ar Xiv:1312.5602, 2013. [23] Nicolo Cesa-Bianchi and Gábor Lugosi. Combinatorial bandits. Journal of Computer and System Sciences, 78(5):1404 1422, 2012. [24] Adam Kalai and Santosh Vempala. Efficient algorithms for online decision problems. Journal of Computer and System Sciences, 71(3):291 307, 2005. [25] Jason Rhuggenaath, Alp Akcay, Yingqian Zhang, and Uzay Kaymak. Algorithms for slate bandits with non-separable reward functions. ar Xiv preprint ar Xiv:2004.09957, 2020. [26] Yanjun Han, Yining Wang, and Xi Chen. Adversarial combinatorial bandits with general nonlinear reward functions. In International Conference on Machine Learning, pages 4030 4039. PMLR, 2021. [27] Fares Fourati, Christopher John Quinn, Mohamed-Slim Alouini, and Vaneet Aggarwal. Combinatorial stochastic-greedy bandit. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 38, pages 12052 12060, 2024. [28] Sho Takemori, Masahiro Sato, Takashi Sonoda, Janmajay Singh, and Tomoko Ohkuma. Submodular bandit problem under multiple constraints. In Conference on Uncertainty in Artificial Intelligence, pages 191 200. PMLR, 2020. [29] Christos H Papadimitriou and John N Tsitsiklis. The complexity of optimal queueing network control. In Proceedings of IEEE 9th annual conference on structure in complexity Theory, pages 318 322. IEEE, 1994. [30] Biswarup Bhattacharya. Restless bandits visiting villages: A preliminary study on distributing public health services. In Proceedings of the 1st ACM SIGCAS Conference on Computing and Sustainable Societies, pages 1 8, 2018. [31] Biyonka Liang, Lily Xu, Aparna Taneja, Milind Tambe, and Lucas Janson. A bayesian approach to online learning for contextual restless bandits with applications to public health. ar Xiv preprint ar Xiv:2402.04933, 2024. [32] Lily Xu, Bryan Wilder, Elias B. Khalil, and Milind Tambe. Reinforcement learning with combinatorial actions for coupled restless bandits. In ar Xiv, 2023. [33] Christine Herlihy and John P Dickerson. Networked restless bandits with positive externalities. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 11997 12004, 2023. [34] Tomer Gafni, Michal Yemini, and Kobi Cohen. Learning in restless bandits under exogenous global markov process. IEEE Transactions on Signal Processing, 70:5679 5693, 2022. [35] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019. [36] Uriel Feige. A threshold of ln n for approximating set cover. Journal of the ACM (JACM), 45 (4):634 652, 1998. [37] Alexander Schrijver. A course in combinatorial optimization. CWI, Kruislaan, 413:1098, 2003. Broader Impact In this paper, we analyze the use of restless multi-armed bandits to tackle general reward functions and apply this to real-world problems in food rescue. Our goal is to develop practically motivated and theoretically insightful algorithms. With this goal, we evaluate our algorithms on real-world data to understand whether such decision-making algorithms have real-world validity. Our proposed algorithms and analysis are a step towards improving the deployment of restless multi-armed bandit algorithms in real-world situations such as food rescue. A Experimental Details We provide additional details on our experiments and experimental conditions. For all experiments in the main paper, we run experiments for 15 seeds, and for each seed, we run experiments across 5 choices of initial state (5 trials). For all ablations in the Appendix, we run for 9 seeds and run experiments across 5 choices of initial state. Random seeds change the reward parameters (mi or Yi) and the transitions. For all plots, we plot the normalized reward averaged across seeds along with the standard error. We compute the discounted reward with an infinite time horizon and normalize this as the improvement over the random policy. We do this so that we can standardize the magnitude of rewards, as such a magnitude can vary across rewards and selections of reward parameters. For all experiments γ = 0.9. We estimate the infinite-horizon discounted reward by computing the discounted reward for T = 50 timesteps; we do this because for t > 50, γt 0.005, and so we can compute the discounted reward with only a small difference from the infinite time horizon. To construct synthetic transitions, we consider a parameter q, and let Pi(0, 0, 1) U(0, q) where U is the uniform distribution. Then Pi(1, 0, 1) U(Pi(0, 0, 1), 1) and Pi(0, 1, 1) U(Pi(0, 0, 1), 1). Finally, Pi(1, 1, 1) U(max(Pi(1, 0, 1), Pi(0, 1, 1)), 1). We construct transitions in such a way so that being in si = 1 or playing action ai = 1 can only help with these synthetic transitions, which matches from prior work [11]. We vary q in Appendix G, and find similar results across selections of q. We additionally detail some of the choices for parameters for each of the rewards used in Section 6.1. For the Linear, Probability, and Max rewards, we select m1, , m N U(0, 1). For the Probability distribution, such a selection is natural (as probabilities are between 0 and 1), while for the Linear and Max reward, we find similar results no matter the distribution of mi chosen. We consider different selections for the sizes of the subset and use these different sizes to investigate the impact of linearity on performance. By default, we let m = 20 and let Yi be sets of size 6. We select such a size so that Rglob(s, a) is non-linear in |Yi|, as when |Yi| is too small the problem becomes more linear (because Yi Yj is empty). To understand the impact of linearity, we construct the sets Yi in such a way to control θlinear. In particular, we consider two parameters a and b. We construct K (where K is the budget) sets which where Yi = {1, 2, , b}, and we construct N K disjoint sets which are size a < b. For the N K disjoint sets, they would be {1, 2, , a}, {a + 1, a + 2 , 2a}, . Under this construction, θlinear = b Ka. We then vary b {3, 4, 5}, and vary a {1, , b 1}, then compute the performance of each policy for each of these rewards. Constructing sets in this way allows us to directly measure the impact of linearity on performance, as we can control the linearity of the reward function by varying a and b. We run all experiments on a TITAN Xp with 12 GB of GPU RAM running on Ubuntu 20.04 with 64 GB of RAM. Each policy runs in under 30 minutes, and we use 300-400 hours of computation time for all experiments. We develop DQN policies using Py Torch [35]. B Baseline Details We provide additional details on select baselines here: 1. DQN - We develop a Deep Q Network (DQN), where a state consists of the set of arms pulled and the current state of all arms. The arms correspond to a new arm to pull; we structure the DQN to avoid the number of actions being exponential in the budget (which is often linear in the number of arms). We train the DQN for 100 epochs and evaluate with Not Engaged No Notify Notify Not Engaged No Notify Notify (a) Transitions 0% 5% 10% Match Probability (b) Match Probabilities Figure 4: We plot the (a) transition probabilities obtained by clustering volunteers with various levels of experience and (b) the distribution of match probabilities across all volunteers. We see that as volunteers gain more experience, they are more likely to become engaged when notified. However, we see that most volunteers have a low probability of responding, which motivates the need for large budgets in notification. other choices in Appendix F. We let the learning rate be 5 10 4, a batch size of 16, and an Adam optimizer. We use an MLP with a 128-dimension as our hidden layer, and use 2 such hidden layers. We use Re Lu activation for all layers. We provide more details on training DQNs in our supplemental codebase. 2. DQN Greedy - Inspired by Theorem 3, we combine a DQN with greedy search. We first learn a DQN network where states directly correspond to the states of the arms, and actions correspond to subsets of arms pulled. Due to the large number of actions, such a method does not scale when the number of arms is increased (as the number of actions is exponential in the budget). Using this learned Q function, we greedily select arms one-by-one, until K total arms are selected. We let the learning rate be 5 10 4, a batch size of 16, and an Adam optimizer. 3. MCTS - We run our MCTS baseline algorithm and use the upper confidence tree algorithm [20]. We run this for 400 iterations, and recurse until a depth of 2 K is reached; we detail other choices for this in Appendix E. We select a depth of 2K so that arm selections are non-myopic. By letting the depth be K, the MCTS algorithm would simplify into a better version of greedy search, whereas we want to see if we can non-myopically select arms using MCTS. C Policy Details We provide details for each of our policies below. For Shapley-based methods, we approximate the Shapley value by computing 1000 random samples of combinations. We do this because computing the exact Shapley value takes exponential time in the number of arms. In Section 6, we consider the impact of other choices on the efficiency and performance of models. We additionally provide the pseudo-code for our iterative algorithms (Algorithm 2) and our MCTS Shapley-Whittle and MCTS Linear-Whittle algorithms (Algorithm 1). For our MCTS-based policies, we note that this follows from the typical upper confidence trees framework, varying only in the rollout step. We let ϵ = 0.1 for the rollout function, and let c = 5. D Food Rescue Details To apply the food rescue scenario to RMABs, we compute aggregate transition probabilities for volunteers through clustering. Each volunteer in the food rescue performs a certain number of trips, Algorithm 1 Monte Carlo Tree Search for RMAB-Gs Input: State s, reward function R(s, a), Q functions QL i,w(si, ai, pi) or QS i,w(si, ai, ui), and transitions P1, , PN, number of MCTS trials M, and hyperparameter c Output: Action a Initialize a tree, T, with leafs representing K arms pulled Let vm represent the total value of a node m Let nm represent the number of visits to a node m for t = 1 to M do Let m be the root of T Let l be empty set while m is non-leaf do Add m to l if m has an unexplored child then while m is non-leaf do With probability ϵ, let m be a random unexplored child of m Otherwise, let m be the unexplored child of m, with the largest value of wi(si), where wi(si) is either the Linear or Shapley-Whittle index Add m to l Let m = m end while else Let m = argmaxm vm nm nm Let m = m end if end while Compute action a from nodes played, l For MCTS Linear-Whittle, let r = R(s, a) + PN i=1[QL i,w(si, ai, pi) pi(si)ai Ri(si, ai)] For MCTS Shapley-Whittle, r = R(s, a) + PN i=1[QS i (si,w, ai, ui) ui(si)ai Ri(si, ai)] for m l do Update wm wm + r Update nm nm + 1 end for end for return The action corresponding to the leaf, m, with the highest wm and we use this information to construct clusters. We ignore volunteers who have performed 1 or 2 trips, as those volunteers have too little information to construct transition functions for. We construct 100 clusters of roughly equal size, where each cluster consists of volunteers who perform a certain number of trips; for example, the 1st cluster consists of all volunteers who perform 3 trips, and the 90th cluster is those who perform 107-115 trips. From this, we compute transition probabilities by letting the state (which represents engagement) denote whether a volunteer has completed a trip in the past two weeks. We then see if completing a trip in the current two weeks impacts the chance that they complete a trip in the next two weeks. We use this as a proxy for engagement due to the lack of ground truth information on whether a volunteer is truly engaged. We average this over all volunteers in a cluster and across time periods to compute transition probabilities. We present examples of such transition probabilities in Figure 4. To compute the corresponding match probabilities for each volunteer, we compute the fraction of notifications they accept, across all notifications seen. This way, each arm corresponds to a cluster of volunteers, with a transition probability and match probability. We compute match probabilities on a per-volunteer basis so that multiple arms can have the same transition probability (corresponding to volunteers being in the same cluster) while differing in match probability. Such a situation corresponds to the real world as well; not all volunteers who complete between 107-115 trips have the same match probability. We analyze the distribution of match probabilities for each volunteer, across all trips that they are notified for. We find that most volunteers accept less than 2% of the trips they ve seen, while a few volunteers answer more than 10% (Figure 4). We note the difference between our two scenarios: Algorithm 2 Iterative Policies for RMAB-Gs Input: State s, reward function R(s, a), Whittle functions w L i (si, pi, X) or w S i (si, ui, X), and transitions P1, , PN Output: Action a Let X = for t = 1 to K do Let vi = 0 for 1 i N for i = 1 to N do if i / X then Update vi = w L i (si, pi, X) or vi = w S i (si, ui, X) end if end for if max v > 0 then Update X X {arg max v} end if end for Let a {0, 1}N, with ai = 1 i S 1 2 4 MCTS Depth Normalized Reward Linear-Whittle MCTS (a) MCTS Depth 100 200 400 800 MCTS Test Iterations Normalized Reward Linear-Whittle MCTS (b) MCTS Test Iterations Figure 5: We compare the performance of vanilla MCTS algorithms when varying (a) depth of exploration and (b) the number of test iterations. We find that, no matter the choice of depth or test iterations, MCTS performs worse than Linear-Whittle. Moreover, as expected, increasing test iterations improves performance, while lower depth improves MCTS performance due to the difficulty in estimating rewards from transitions. notifications and phone calls. For notifications, we consider all volunteers and notify a large fraction of these volunteers, while for phone calls, we consider only volunteers who have completed more than 100 trips, as they generally have a higher match probability. We then cluster those volunteers into 20 groups and similarly compute transition probabilities. We do this because food rescue organizations target more active volunteers when making phone calls due to the limited number of phone calls that can be made. E Vanilla MCTS Experiments To understand the impact of our choice of hyperparameters for the MCTS baseline algorithm, we vary the depth of the MCTS search, along with the number of iterations. We vary the depth of MCTS in {1K, 2K, 4K} (while setting the iterations to 400), and we vary the number of iterations run for MCTS in {50, 100, 200, 400} (while searching up to depth 2K). For all runs, we let K = 5 and N = 10 for the Max reward. In Figure 5, we see that increasing MCTS depth only lowers performance, while increasing the number of iterations MCTS is run for increases performance. This is due to the stochasticity in rewards as the depth of exploration increases; there are |S|N possible combinations of states, making exploration difficult. We find that for any combination of iterations and depth, the performance of MCTS lags significantly behind Linear-Whittle. While running MCTS for additional iterations 50 100 200 400 Training Episodes Normalized Reward Linear-Whittle DQN Figure 6: We compare the impact of additional training epochs on the performance of a DQN baseline. We find that, regardless of the number of training episodes, Linear-Whittle outperforms the DQN baseline. Additional training episodes do not result in a significantly better performance (compared with Linear-Whittle), showing that additional training episodes do not lead to a much smaller gap between Linear-Whittle and the DQN baseline. 0.25 0.5 0.75 1.0 Transition Probability (q) Normalized Reward Linear-Whittle Shapley-Whittle Iterative Linear Iterative Shapley MCTS Linear MCTS Shapley Figure 7: We assess the impact of changing q, which parametrizes transition probabilities. Smaller q makes it less likely that arms transition to the 1 state. We find that, regardless of the transition probability chosen, MCTS Shapley-Whittle policies improve upon all other policies. would be helpful, this comes at a cost of time, so we run MCTS for 400 iterations throughout our experiments to balance these. F Deep Q Networks We compare the performance of our reinforcement learning baseline (DQN) to the Linear-Whittle policy as we vary the number of epochs that the DQN is trained for. We vary the number of train epochs in {50, 100, 200, 400} and plot our results in Figure 6. We see that increasing the number of iterations does not provide a big advantage in performance; this is because DQN fails to learn past 50 epochs, and so additional training steps do not lead to much better performance. Moreover, we see that, no matter the selection of train iterations, DQN performs worse than Linear-Whittle. We believe the reason for this is the stochasticity in the dataset; because the number of actions is large, it becomes difficult to learn state values in this scenario, which leads to the poor performance of DQN. Normalized Reward Vanilla Whittle Greedy MCTS DQN DQN Greedy Linear-Whittle Shapley-Whittle Iterative Linear Iterative Shapley MCTS Linear MCTS Shapley Optimal Figure 8: We construct situations where adaptive policies are needed to perform better than baselines. When N = 4, Iterative Shapley-Whittle significantly outperforms alternatives, and when N = 10, MCTS Shapley-Whittle outperforms alternatives. Probability 0.95 Subset 0.95 Probability 0.95 Normalized Reward Linear-Whittle Shapley-Whittle Iterative Linear Iterative Shapley MCTS Linear MCTS Shapley Figure 9: We repeat Figure 1 for N = 25 and N = 100. We find that many of the same trends appear here; namely, that MCTS Shapley-Whittle outperforms all other policies, and Shapley-based policies do better in general. G Synthetic Transitions Throughout our experiments, we sample synthetic transitions, with a maximum probability for Pi(0, 1, 0) of q; we vary the value of q to understand if this impacts the performance of policies. We vary q {0, 0.25, 0.5, 0.75, 1.0}, and evaluate the performance of our policies on the Max reward. We find that, across all values of q, the MCTS Shapley-Whittle policy performs best. We find large gaps between MCTS Shapley-Whittle and Linear-Whittle for small q; in these scenarios, the transitions are more stochastic, leading to greater impact from MCTS Shapley-Whittle. However, regardless of the choice of q, our conclusion remains that Iterative and MCTS Shapley-Whittle policies outperform other alternatives. H Other Constructions of Transitions We investigate whether situations detailed in Lemma 8 can be tackled by adaptive algorithms. We construct a situation for the Max reward where transitions are Pi(s, 1, 0) = 1 and Pi(s, 0, s) = 1; pulling an arm always leads to si = 0, and otherwise, arms stay in their current state. We design such a situation to see whether adaptive policies can perform well in situations where index-based policies are proven to perform poorly. We let the values m1 = m2 = 1, then let the other choices of mi = 0. We do this to encourage arms to be pulled one at a time, rather than pulling m1 and m2 together. In Figure 8, we compare the performance of policies in this situation and see that Iterative Shapley-Whittle significantly outperforms all other policies when N = 4, and that MCTS Linearand Shapley-Whittle outperform all other policies when N = 10. When N = 10, we see that all non-adaptive Whittle policies achieve similar performance, demonstrating the need for adaptive policies. 0 0.25 0.5 0.75 1.0 1.0 Normalized Reward Vanilla Whittle Greedy Linear-Whittle Shapley-Whittle Iterative Linear Iterative Shapley MCTS Linear MCTS Shapley Figure 10: We vary the parameter α, which modulates between the individual rewards Ri and the global reward Rglob. When α = 1, all policies perform similarly, while for α {0.25, 0.5, 0.75}, we see that Linear-Whittle performs worse than other policies. Across values of α, baselines perform worse than all policies, which is most notable when α = 0. I Other Combinations of Arms and Budget More arms We evaluate the performance of policies when N = 25 or N = 100 and plot this in Figure 9. We run Iterative Shapley-Whittle, MCTS Linear-Whittle, and MCTS Shapley-Whittle policies for N = 25 and refrain from running them when N = 100 due to the time needed. We find that MCTS Shapley-Whittle policies perform better than alternatives, and that they perform significantly better than Linear-Whittle policies. The results with N = 25 and N = 100 confirm the trends from Section 6.1 even when the number of arms is large. Weighting Individual and Global Rewards Throughout our experiments, we weight the impact of the global reward, Rglob(s,a), and the individual reward, Ri(si, ai), equally. We evaluate the impact of various weighting factors so that R(s, a) = (1 α)Rglob(s, a) + α PN i=1 Ri(s, a). By default, α = 0.5 throughout our experiments, so we vary α {0, 0.25, 0.5, 0.75, 1.0} and evaluate the reward for baselines and our policies on the maximum reward. In Figure 10, we see that, regardless of the choice of α, our policies outperform baselines. We additionally see that the Shapley-Whittle and Iterative Shapley-Whittle perform well across the choices of α, implying that the results we found with α = 0.5 are not limited to one choice of α. Varying Budget Throughout our experiments, we let the budget, K = 0.5N, so we analyze the impact of varying K when fixing N = 10. We vary K {2, 5, 8}, and note that with K = 10 or K = 0, all policies perform the same (as all or no arms are selected). We see that, across all choices of budget, Linear-Whittle performs worse than all other policies (Figure 11). Moreover, we see again that regardless of the choice of K, MCTS Shapley-Whittle is the best policy. J Food Rescue More Arms In addition to the experiments in Section 6.2, we evaluate the impact of increasing the number of arms in food rescue on the performance of various policies. We consider a situation with N = 1000 and K = 250, for the notification scenario, and another situation with N = 250 and K = 5, which corresponds to the phone calls scenario. We plot both situations in Figure 12, and we see that all policies perform similarly in both of these situations. When the number of arms is this large, all policies will select the top arms as arms are likely to have a large match probability due to the sheer number of arms. We can differentiate between policies in the notification scenario, where Shapley-Whittle policies perform better than Linaer-Whittle policies, but in the phone calls settings, all policies perform similarly. Additionally, due to the large budget, the match probability will be 2 5 8 Budget Normalized Reward Linear-Whittle Shapley-Whittle Iterative Linear Iterative Shapley MCTS Linear MCTS Shapley Figure 11: We vary the budget, K, while fixing N = 10 for the maximum reward. We see that, for any choice of K, all policies have a larger reward than greedy, and that Linear-Whittle policies perform worse than other policies. Moreover, for K = 5 and K = 8, Iterative Shapley-Whittle policies outperform Iterative Linear-Whittle policies, while for K = 2 policies, Iterative Linear-Whittle policies perform well, potentially due to the linearity of the problem with small K. Notificaitons (1000) 1.0 Phone Calls (250) 1.0 Normalized Reward Linear-Whittle Shapley-Whittle Iterative Linear Figure 12: We compare the performance of policies on the food rescue dataset for two situations: (a) notifications, when N = 1000 and K = 500, and (b) phone calls, when N = 250 and K = 50. We find that when N is large, all policies perform similarly, due to the linearity of the problem for large N. large, so policies should focus on engagement in this scenario. In reality, the choice of whether to use pre-computed index policies or adaptive policies depends on the time available, the size of the problem, and the structure of the reward. K Proofs of Problem Difficulty Theorem 1. The restless multi-armed bandit with global rewards is PSPACE-Hard. Proof. We reduce RMAB to RMAB-G. Note that RMAB-G aims to maximize t=0 γt(Rglob(s, a) + i=1 Ri(si, ai)) (9) Consider an RMAB instance with reward R i(si, ai). We can construct an instance of RMAB-G with Ri = R i and with Rglob(s, a) = 0. Therefore, we have reduced RMAB to RMAB-G. Because RMAB is PSPACE-Hard [29], RMAB-G is as well. Theorem 2. No polynomial time algorithm can achieve better than a 1 1 e approximation to the restless multi-armed bandit with global rewards problem. Proof. Consider any submodular function F(S) : 2Ω R and a cardinality constraint |S| K. Unless P = NP, maximizing F(S) with a cardinality constraint can only be approximated up to a factor of 1 1 e [36]. Next, we will show that we can reduce submodular maximization to RMAB-G. For a given submodular function F, we construct an instance of RMAB-G with S = {0, 1}. Let s(0) = 1, where 1 is the vector of N 1s. Finally let Pi(s, a, 0) = 1, so that all arms are in state 0 for timesteps t > 0. Let R(s, a) = 0 if s = 1, and let R(s, a) = F({i|ai = 1}) otherwise. Note that only at time t = 0 can you achieve any reward, as afterwards, F(S) = 0. The optimal solution to the RMAB-G problem is the action, a that maximizes F({i|ai = 1}) subject to the constraint, PN i=1 ai K. Note that this is the same as the cardinality constraint for submodular maximization, |S| = PN i=1 ai, along with the function of interest, F(S). Therefore, if we can efficiently solve RMAB-G, then we can efficiently solve the original submodular maximization problem in F. Because we cannot do better than a 1 1 e approximation to the submodular maximization problem in polynomial time, we cannot do better than a 1 1 e approximation to the RMAB-G problem. Lemma 9. Let P be a set of transition functions, P1, , PN S A S [0, 1], where N is the number of arms. Let S = A = {0, 1}. Let Qπ(s, a) = E(s ,a ) (P,π)[P t=0 γt R(s , a ))|s0 = s, a0 = a] for some policy π : S A, and let π be the optimal policy. Let Vπ (s) = max a Qπ(s, a). If R(s, a) is monotonic in s and Pi(1, a, 1) Pi(0, a, 1) for any a, then Vπ (s) is monotonic in s Proof. We construct Vπ (s) by stepping through inductive steps in value iteration, and showing that the monotonicity property is preserved throughout these steps. We note that performing Value iteration from any initialization results in the optimal policy, Vπ (s) Let Q(t)(s, a) represent the Q value for iteration t of Value iteration, and let V (t)(s) be the value function for iteration t. We initialize V (0)(s) = 0 for any s, and note that the initialization should not impact the process. Running an iteration of value iteration leads Q(1)(s, a) = R(s, a), so that V (1)(s) = max a Q(1) π (s, a). Base Case: Initially, V (1)(s) = max a R(s, a). Next consider any two states, x and y, so that x y. Here x y implies that xi yi i. Let a = argmaxa R(y, a). Then V (1)(y) = R(y, a ) R(x, a ) V (1)(x). Therefore, V (1)(s) is monotonic. Inductive Hypothesis: Assume that V (t 1)(s) is monotonic in s. Next, we will prove that V (t)(s ei) V (t)(s) 0. Here ei is the vector with 1 at location i and 0 everywhere else. Before proving this, we will first show that this implies monotonicity for any two states s, s . Suppose that V (t)(s ei) V (t)(s) 0. Then consider some pair of states, s s . Because S = A = {0, 1}, if s s , then we can write s = s W i X ei for some set X = {x1, x2, , xn}. If V (t)(s ei) V (t)(s ), then V (t)((s ex1) ei) V (t)(s ex1) Therefore, this implies that V (t)(s) V (t)(s ) = i=1 (V (t)(s j=1 exj) V (t)(s j=1 exj)) 0 (10) We get this by expanding V (t)(s) V (t)(s ) into a telescoping series. Because each term in the sum is non-negative, the whole sum is non-negative. Next we prove our original statement: V (t)(s + ei) V (t)(s). We first expand V (t)(s) = R(s, a) + γ P s P(s, a, s )V (t 1)(s ) for some a, where P(s, a, s ) = QN j=1 Pj(sj, aj, s j). We note that V (t)(s ei) R(s ei, a) + γ P s P(s ei, a, s )V (t 1)(s ) V (t)(s ei) V (t)(s) R(s ei, a) R(s, a)+γ X s (P(s ei, a, s ) P(s, a, s ))V (t 1)(s ) (11) We note that R(s ei, a) R(s, a) 0 due to the monotonicity of R with fixed a. Let Ci {0, 1}N be the set of 0-1 vectors of length N so that if x Ci then xi = 0. Additionally, let δi = P(s ei, a, s ei) P(s, a, s ei). We note that δi = P(s ei, a, s ei) P(s, a, s ei) k=1 Pk((s ei)k, ak, (s ei)k) k=1 Pk(sk, ak, (s ei)k) = (Pi(1, ai, 1) Pi(si, ai, 1)) k=1,k =i Pk(sk, ak, s k) Let 1i be the vector of all 1s except at i. Let be the element-wise minimum operator. Then s 1i sets si = 0 while leaving other indices unchanged. Because transition probabilities add to one, P(s ei, a, s 1i) P(s, a, s 1i) = δi V (t)(s ei) V (t)(s) R(s ei, a) R(s, a) + γ X s (P(s ei, a, s ) P(s, a, s ))V (t 1)(s ) s (P(s ei, a, s ) P(s, a, s ))V (t 1)(s ) s Ci (P(s ei, a, s 1i) P(s, a, s 1i))V (t 1)(s 1i) s Ci (P(s ei, a, s ei) P(s, a, s ei))V (t 1)(s ei) s Ci δi V (t 1)(s 1i) + δi V (t 1)(s ei) s Ci δi(V (t 1)(s ei) V (t 1)(s 1i)) The second step splits the sum into transitions for states with s i = 0 and those with s i = 1. The final step applies the inductive hypothesis and noting that s ei s 1i. Lemma 10. Let P be a set of transition functions, P1, , PN S A S [0, 1], where N is the number of arms. Let S = A = {0, 1}. Let Qπ(s, a) = E(s ,a ) (P,π)[P t=0 γt R(s , a ))|s0 = s, a0 = a] for some policy π : S A, and let π be the optimal policy. Let Vπ (s) = max a Qπ(s, a). If R(s, a) is submodular in both s and a, g(s) = max a R(s, a) is submodular in s, and Pi(1, a, 1) Pi(0, a, 1) for any a, then Vπ (s) is submodular in s. Proof. We similarly construct Vπ (s) through induction in the value iteration steps. Base Case: We start with the base case: V (1)(s) = max a R(s, a) = g(s), which is submodular by assumption. Inductive Hypothesis: Again let V (t)(s) be the value function at step t of value iteration. We assume that V (t 1)(s) is submodular, and aim to prove that V (t)(s ei) V (t)(s) V (t)(s ei ej) + V (t)(s ej) 0. We note that this is an equivalent definition of submodularity [37]. Now, consider V (t)(s ei) V (t)(s) V (t)(s ei ej) + V (t)(s ej). The sum of submodular functions is submodular and note that V (t)(s) = max a R(s, a) + γ P s V (t 1)(s )P(s, a, s ). We note that the first term is submodular by assumption, so we focus on proving the submodularity of γ P s V (t 1)(s )P(s, a, s ). Let Ci,j {0, 1}N be the set of 0-1 vectors of length N so that if x Ci,j then xi = 0, xj = 0, and let δi = Pi(1, ai, 1) Pi(si, ai, 1); note the different definition of δi from the previous proof. Finally, let C(s, a, s ) = QN l=1,l =i,j Pl(sl, al, s l), then we can write the following: V (t)(s ei) V (t)(s) V (t)(s ei ej) + V (t)(s ej) s Ci,j C(s, a, s )(V (t 1)(s ei) V (t 1)(s ))δi Pj(sj, aj, 0) s Ci,j C(s, a, s )(V (t 1)(s ei ej) V (t 1)(s ej))δi(1 Pj(sj, aj, 0)) s Ci,j C(s, a, s )(V (t 1)(s ei) V (t 1)(s ))δi(Pj(1, aj, 0)) s Ci,j C(s, a, s )(V (t 1)(s ei ej) V (t 1)(s ej))δi(1 Pj(1, aj, 0)) s Ci,j C(s, a, s )δi(V (t 1)(s ei) V (t 1)(s ))(Pj(sj, aj, 0) Pj(1, aj, 0)) s Ci,j C(s, a, s )δi(V (t 1)(s ei ej) V (t 1)(s ej))(Pj(1, aj, 0) Pj(sj, aj, 0)) s Ci,j C(s, a, s )δiδj(V (t 1)(s ei) + V (t 1)(s ei) V (t 1)(s ) V (t 1)(s ei ej)) The last step arises because δi, C(s, a, s ) are all non-negative, and the inductive hypothesis. Theorem 3. Consider an RMAB-G instance (S, A, Ri, Rglob, Pi, γ). Let Qπ(s, a) = E(s ,a ) (P,π)[P t=0 γt R(s , a ))|s(0) = s, a(0) = a] and let π be the optimal policy. Let a = arg maxa Qπ (s, a). Suppose Pi(s, 1, 1) Pi(s, 0, 1) s, i, Pi(1, a, 1) Pi(0, a, 1) i, a, R(s, a) monotonic and submodular in s and a, and g(s) = max a R(s, a) submodular in s. Then with oracle access to Qπ we can compute ˆa in O(N 2) time so Qπ (s, ˆa) (1 1 e)Qπ (s, a ). Proof. We first demonstrate that Qπ (s, a) is submodular and monotonic in a. Prior work demonstrates that monotonic submodular function can be approximated within an approximation factor of 1 1 e with O(N 2) evaluations of the function, so all that remains is to demonstrate submodularity [13]. We note that Qπ (s, a) = R(s, a) + γ P s Vπ (s )P(s, a, s ), and that R(s, a) is both monotonic and submodular in a. Therefore, we demonstrate that γ P s Vπ (s )P(s, a, s ) is monotonic and submodular in a. Let H(a) = γ P s Vπ (s )P(s, a, s ) for fixed s. We start by demonstrating monotonicity. First, let ηi = Pi(si, 1, 1) Pi(si, 0, 1), and by assumption, ηi 0. Then: H(a ei) H(a) s Ci Vπ (s ei)(P(s, a ei, s ei) P(s, a, s ei)) s Ci Vπ (s ei)(P(s, a ei, s 1i) P(s, a, s 1i)) s Ci ηi(Vπ (s ei) Vπ (s 1i)) j=1,j =i Pj(sj, aj, s j) The last steps follows from the monotonicity of Vπ (s), which was proven in Lemma 9. Next, we prove submodularity of H(a), by showing that H(a ei) H(a) H(a ei ej) + H(a ej) 0. That is H(a ei) H(a) H(a ei ej) + H(a + ej) s Ci,j ηi Pj(sj, aj, 0)(Vπ (s ei) Vπ (s)) + ηi(1 Pj(sj, aj, 0))(Vπ (s ei ej) Vπ (s)) s Ci,j ηi Pj(sj, 1, 0)(Vπ (s ei) Vπ (s)) + ηi(1 Pj(sj, 1, 0))(Vπ (s ei ej) Vπ (s ei)) s Ci,j ηiηj(Vπ (s ei) + Vπ (s ej) Vπ (s) V (s ei ej)) The last line follows from the submodularity of V , as proven in Lemma 10. Therefore, we have shown that Qπ (s, a) is submodular, which completes the proof. L Proofs of Index-Based Bounds L.1 Lower Bounds Theorem 4. For any fixed set of transitions, P, let OPT = max π E(s,a) (P,π)[ 1 T PT 1 t=0 R(s, a)] for some T. For an RMAB-G (S, A, Ri, Rglob, Pi, γ), let R i(si, ai) = Ri(si, ai) + pi(si)ai, and let the induced linear RMAB be (S, A, R i, Pi, γ). Let πlinear be the Linear-Whittle policy, and let ALG = E(s,a) (P,πlinear)[ 1 T PT 1 t=0 R(s, a)]. Define βlinear as βlinear = min s SN,a [0,1]N, a 1 K R(s, a) PN i=1(Ri(si, ai) + pi(si)ai) (12) If the induced linear RMAB is irreducible and indexable with the uniform global attractor property, then ALG βlinear OPT asymptotically in N for any set of transitions, P. Proof. By the definition of the minimum, βlinear PN i=1(pi(si)ai + Ri(si, ai)) R(s, a) a, s. Additionally, note that PN i=1 pi(si)ai Rglob(s, a). This holds because pi(si) = max s |s i=si Rglob(s, a). So, PN i=1 pi(si)ai R(s, PN i=1 eiai) = R(s, a) by submodularity of Rglob(s, a). We compare the linearized reward to the submodular reward for the set of actions played by the Linear-Whittle policy πlinear. Let the actions be a(1), a(2), . If we play for T iterations, then: i=1 (pi(s(t) i )a(t) i + Ri(s(t) i , a(t) i )) 1 t=0 R(s(t), a(t)) (13) Next, consider the following induced RMAB with R i(si, ai) = Ri(si, ai) + pi(si)ai. By assumption, our relaxed RMAB is indexable, irreducible, and has a global attractor. In this scenario, prior work [2] states that the Whittle index is asymptotically optimal. For this relaxed RMAB problem, we note that the Whittle Index solution corresponds exactly to the Linear-Whittle policy, πlinear. That is max π E(s,a) (P,π)[ 1 i=1 R i(si, ai)] = E(s,a) (P,πlinear)[ 1 i=1 R i(si, ai)] (14) Then, asymptotically, = max π E(s,a) (P,π)[ 1 t=0 R(s, a)] max π E(s,a) (P,π)[ 1 i=1 R i(si, ai)] = E(s,a) (P,πlinear)[ 1 i=1 R i(si, ai)] 1 βlinear E(s,a) (P,πlinear)[ 1 t=0 R(s, a)] = 1 βlinear ALG Theorem 5. For any fixed set of transitions, P, let OPT = max π E(s,a) (P,π)[ 1 T PT 1 t=0 R(s, a)] for some T. For an RMAB-G (S, A, Ri, Rglob, Pi, γ), let R i(si, ai) = Ri(si, ai) + ui(si)ai, and let the induced Shapley RMAB be (S, A, R i, Pi, γ). Let πshapley be the Shapley-Whittle policy, and let ALG = E(s,a) (P,πshapley)[ 1 T PT 1 t=0 R(s, a)]. Define βshapley as βshapley = min s SN,a [0,1]N, a 1 K R(s,a) PN i=1(Ri(si,ai)+ui(si)ai) max s SN,a [0,1]N, a 1 K R(s,a) PN i=1(Ri(si,ai)+ui(si)ai) (15) If the induced Shapley RMAB is irreducible and indexable with the uniform global attractor property, then ALG βshapley OPT asymptotically in N for any choice of transitions, P. Proof. First, consider OPT = max π E(s,a) (P,π)[P t=0 γt R(s, a)]. First, note that R(s , a ) (PN i=1 ui(s i)a i + Ri(s i, a i)) max s SN,a [0,1]N, a 1 K R(s,a) PN i=1(ui(si)ai+Ri(si,ai)). Next, note that sim- ilar to Theorem 4, πshapley is the optimal policy for the RMAB instance where R i(si, ai) = ui(si)ai + Ri(si, ai) for the induced RMAB (due to the indexability, irreducibility, and global attractor properties [2]), or in other words, max π E(s,a) (P,π)[ 1 i=1 R i(si, ai)] = E(s,a) (P,πshapley)[ 1 i=1 R i(si, ai)] (16) = max π E(s,a) (P,π)[ 1 t=0 R(s, a)] E(s,a) (P,π)[ 1 i=1 R i(si, ai)] max s SN,a [0,1]N, a 1 K R(s, a) PN i=1(ui(si)ai + Ri(si, ai)) E(s,a) (P,πshapley)[ 1 i=1 R i(si, ai)] max s SN,a [0,1]N, a 1 K R(s, a) PN i=1(ui(si)ai + Ri(si, ai)) E(s,a) (P,πshapley)[ 1 Corollary 5.1. Consider an RMAB-G instance with a monotonic, submodular reward function R(s, a) = Rglob(s, a), and let πlinear be the Linear-Whittle policy. Let ALG = E(s,a) (P,πlinear)[ 1 T PT 1 t=0 R(s, a)] and OPT = max π E(s,a) (P,π)[ 1 T PT 1 t=0 R(s, a)] for some T. Then ALG OPT K for any transitions P. Proof. By Theorem 4, we know that ALG βlinear OPT. We next show that βlinear 1 For a monotonic, submodular function, F : 2Ω R, F({x1, x2, , xn}) max i F({xi}). Next, applying this to our global reward function, for a fixed state s, we note that Rglob(s, a) max i Rglob(s, eiai). Next, by construction, R(s, a) = PN i=1 Ri(si, ai) + Rglob(s, a) PN i=1(Ri(si, ai) + pi(si)ai). Additionally, R(s, a) PN i=1 Ri(si, ai) + max i pi(si)ai Next, for any state s, PN i=1 pi(si)ai K max i pi(si)ai. Combining the bounds on R(s, a) and PN i=1 pi(s, a), gives that for any state s, R(s, a) PN i=1(Ri(si, 0) + pi(si)ai) PN i=1 Ri(si, ai) + max i pi(si)ai PN i=1 Ri(si, ai) + K max i pi(si)ai 1 Therefore, βlinear 1 K , completing our proof. L.2 Upper Bounds Theorem 6. Let ˆa(s) = arg max a [0,1]N, a 1 K PN i=1(Ri(si, ai) + pi(si)ai). For a set of transi- tions P = {P1, P2, , PN}, let OPT = max π E(s,a) (P,π)[P t=0 γt R(s, a)|s(0)] and ALG = E(s,a) (P,πlinear)[P t=0 γt R(s, a)|s(0)]. Define θlinear as follows: θlinear = min s SN R(s, ˆa(s)) max a [0,1]N, a 1 K R(s, a) (18) Then there exists some transitions, P, and initial state s(0), so that ALG θlinear OPT Proof. We prove this by constructing such a set of transitions such that ALG θlinear OPT. Let s(0) = argmins R(s,ˆa) max a [0,1]N , a 1 K R(s,a). Next, let Pi be the identity transition; Pi(si, ai, s(0) i ) = 1 for any s, a, i. Under these transitions, all arms stay in the state s(0) for all time periods t. Because the state is constant, the set of actions played, a(t) is constant across time periods (as the Linear-Whittle algorithm is deterministic). We next show that the Linear-Whittle policy plays the action arg max a [0,1]N, a 1 K PN i=1(pi(si)ai + Ri(si, ai)). We note that all arms have the same transitions; therefore, arms are played in order of pi(s(0) i )+Ri(si, ai), as each arm is predicted to receive a reward aipi(s(0) i )+Ri(si, ai). This is because Linear-Whittle aims to maximize the sum of marginal rewards, and when the transitions are homogeneous, chooses the largest values for pi(si)ai + Ri(si, ai). In other words, the Linear-Whittle policy chooses an action, alinear so that alinear = arg max a [0,1]N, a 1 K i=1 (pi(s(0) i )ai + Ri(si, ai)) = ˆa(s) (19) Finally, we observe that the optimal action to play is argmaxa [0,1]N, a 1 KR(s(0), a). The ratio of rewards between these two actions is exactly θlinear, and therefore, ALG θlinear OPT Theorem 7. Let ˆa(s) = arg max a [0,1]N, a 1 K PN i=1(Ri(si, ai) + ui(si)ai). For a set of transitions P = {P1, P2, , PN} and initial state s(0), let OPT = max π E(s,a) (P,π)[P t=0 γt R(s, a)|s(0)] and ALG = E(s,a) (P,πshapley)[P t=0 γt R(s, a)|s(0)]. Let θshapley be: θshapley = min s R(s, ˆa(s)) max a [0,1]N, a 1 K R(s, a) (20) Then there exists some transitions, P and initial state s(0) so that ALG θshapley OPT Proof. We follow the same strategy as the proof for the Linear-Whittle case and construct a set of transition probabilities so that ALG θshapley OPT. Again, let s(0) = argmaxs R(s,ˆa(s)) max a [0,1]N , a 1 K R(s,a) and Pi be the identity transition. By the same reasoning as Theorem 6, the Shapley-Whittle policy will play the action ˆa; the Linearand Shapley-Whittle situations are analogous, with the only difference being that Shapley-Whittle assumes each arm receives a reward of aiui(s(0) i ) instead of aipi(s(0) i ). The optimal action played receives a reward of max a [0,1]N, a 1 K R(s, a), so that the ratio between the reward of the Shapley-Whittle policy and optimal is min s SN R(s, ˆa(s)) max a [0,1]N, a 1 K R(s, a) (21) This is exactly θshapley when s = s(0). Lemma 8. Define an index-based policy as any policy that can be described through a function g(si, Pi, R), such that in state s, the policy, π selects the K largest values of g(si, Pi, R), where such a function is evaluated for each arm. Then index-based algorithms will achieve a discounted reward approximation ratio no better than 1 γ 1 γK Proof. Consider an N = K arm system, with S = {0, 1}. Suppose our reward is R(s, a) = max i siai. Next, define the transition probabilities for all arms as follows: Pi(s, 0, s) = 1 and Pi(s, 1, 0) = 1; if an arm is pulled, then the arm will transition to state 0, and otherwise, the arm remains in its current state. Then note that because of symmetry between the arms, g(si, Pi, R) is identical, as R is symmetric across arms, and si and Pi is the same across arms. Moreover, because there are N = K arms, all arms will be selected in the first time step. Therefore, any index-based strategy will lead to a reward of 1. However, selecting arm i in timestep i instead leads to rewards in each of the K timesteps, leading to a total reward of PN 1 i=0 γi > 1 when N > 1. Applying the geometric sum formula reveals that this is a 1 γ 1 γK approximation. Neur IPS Paper Checklist The checklist is designed to encourage best practices for responsible machine learning research, addressing issues of reproducibility, transparency, research ethics, and societal impact. Do not remove the checklist: The papers not including the checklist will be desk rejected. The checklist should follow the references and follow the (optional) supplemental material. The checklist does NOT count towards the page limit. Please read the checklist guidelines carefully for information on how to answer these questions. For each question in the checklist: You should answer [Yes] , [No] , or [NA] . [NA] means either that the question is Not Applicable for that particular paper or the relevant information is Not Available. Please provide a short (1 2 sentence) justification right after your answer (even for NA). The checklist answers are an integral part of your paper submission. They are visible to the reviewers, area chairs, senior area chairs, and ethics reviewers. You will be asked to also include it (after eventual revisions) with the final version of your paper, and its final version will be published with the paper. The reviewers of your paper will be asked to use the checklist as one of the factors in their evaluation. While "[Yes] " is generally preferable to "[No] ", it is perfectly acceptable to answer "[No] " provided a proper justification is given (e.g., "error bars are not reported because it would be too computationally expensive" or "we were unable to find the license for the dataset we used"). In general, answering "[No] " or "[NA] " is not grounds for rejection. While the questions are phrased in a binary way, we acknowledge that the true answer is often more nuanced, so please just use your best judgment and write a justification to elaborate. All supporting evidence can appear either in the main paper or the supplemental material, provided in appendix. If you answer [Yes] to a question, in the justification please point to the section(s) where related material for the question can be found. IMPORTANT, please: Delete this instruction block, but keep the section heading Neur IPS paper checklist", Keep the checklist subsection headings, questions/answers and guidelines below. Do not modify the questions and only use the provided macros for your answers. Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: We present algorithms for RMAB-Gs, prove approximation bounds, and verify empirical performance. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: In Section 8, listed out limitations. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: Proofs for all Theorems can be found in Appendix K and Appendix L. All Theorems have assumptions, etc. listed out. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: List out experimental details in Appendix A, Appendix C, and Appendix B. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: We plan to provide the code in the supplemental materials. We plan to release our synthetic dataset; our food rescue dataset is proprietary, so we are unable to release this. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: Detail optimizers, etc. for the DQN baseline in Appendix B. General experimental details are in Appendix A. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: Plot the standard error for plots to show statistical significance Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: In Appendix A, detail the resources used Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: We conform to the code of ethics Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [Yes] Justification: We have a broader impacts sections which discusses the societal impacts Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: No pretrained models, etc. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: Code packages that are used are referenced, and any datasets are referenced Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: No new assets from this paper Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: No human subjects Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: No human studies Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.