# choice_bandits__d3e2ca60.pdf Choice Bandits Arpit Agarwal University of Pennsylvania Philadelphia, PA 19104, USA aarpit@seas.upenn.edu Nicholas Johnson University of Minnesota Minneapolis, MN 55455, USA njohnson@cs.umn.edu Shivani Agarwal University of Pennsylvania Philadelphia, PA 19104, USA ashivani@seas.upenn.edu There has been much interest in recent years in the problem of dueling bandits, where on each round the learner plays a pair of arms and receives as feedback the outcome of a relative pairwise comparison between them. Here we study a natural generalization, that we term choice bandits, where the learner plays a set of up to k 2 arms, and receives limited relative feedback in the form of a single multiway choice among the pulled arms, drawn from an underlying multiway choice model. We study choice bandits under a very general class of choice models that is characterized by the existence of a unique best arm (which we term generalized Condorcet winner), and includes as special cases the well-studied multinomial logit (MNL) and multinomial probit (MNP) choice models, and more generally, the class of random utility models with i.i.d. noise (IID-RUMs). We propose an algorithm for choice bandits, termed Winner Beats All (WBA), with a distribution dependent O(log T) regret bound under all these choice models. The challenge in our setting is that the decision space is Θ(nk), which is large for even moderate k. Our algorithm addresses this challenge by extracting just O(n2) statistics from multiway choices and exploiting the existence of a unique best arm to find arms that are competitive to this arm in order to construct sets with low regret. Since these statistics are extracted from the same choice observations, one needs a careful martingale analysis in order to show that these statistics are concentrated. We complement our upper bound result with a lower bound result, which shows that our upper bound is order-wise optimal. Our experiments demonstrate that for the special case of k = 2, our algorithm is competitive with previous dueling bandit algorithms, and for the more general case k > 2, outperforms the recently proposed Max Min UCB algorithm designed for the MNL model. 1 Introduction The dueling bandit problem has received a lot of interest in recent years [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14]. Here there are n arms {1, . . . , n}; on each trial t, the learner pulls a pair of arms (it, jt), and receives relative feedback indicating which of the two arms has a better quality/reward. In the regret minimization setting, the goal is to identify the best arm(s) while also minimizing the regret due to playing sub-optimal arms in the learning (exploration) phase. Work done while at the University of Pennsylvania. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. In many applications, however, it can be natural for the learner to pull more than two arms at a time, and seek relative feedback among them. For example, in recommender systems, it is natural to display several items or products to a user, and seek feedback on the most preferred item among those shown. In online advertising, it is natural to display several ads at a time, and observe which of them is clicked (preferred). In online ranker evaluation for information retrieval, one can easily imagine a generalization of the setting studied by Yue & Joachims [15], where one may want to "multi-leave" several rankers at a time to help identify the best ranking system while also presenting good/acceptable results to users using the system during the exploration phase. In general, there is also support in the marketing literature for showing customers more than two items at a time [16]. Motivated by such applications, we consider a framework that generalizes the dueling bandit problem to allow the learner to pull more than two arms at a time. Here, on each trial t, the learner pulls a set St of up to k arms (for fixed k {2, . . . , n}), and receives relative feedback in the form of a multiway choice yt St indicating which arm in the set has the highest quality/reward. The goal of the learner is again to identify a best arm (to be formalized below) while minimizing a suitable notion of regret that penalizes the learner for playing sub-optimal arms during the exploration phase. We term the resulting framework choice bandits. In the (stochastic) dueling bandits framework, the underlying probabilistic model from which feedback is observed is a pairwise comparison model, which for each pair of arms (i, j), defines a probability Pij that arm i has higher reward/quality than arm j. In our choice bandits framework, the underlying probabilistic model is a multiway choice model, which for each set of arms S [n] with |S| k and each arm i S, defines a probability Pi|S that arm i has the highest reward/quality in the set S. We study choice bandits under a new class of choice models, that are characterized by the existence of a unique generalized Condorcet winner (GCW), which we define to be an arm that has larger probability of being chosen than any other arm in any choice set. This class includes as special cases the well-studied multinomial logit (MNL) [17, 18, 19] and multinomial probit (MNP) [20] choice models, and more generally, the class of random utility models with i.i.d. noise (IID-RUMs) [21, 22]. Our main contribution is a computationally efficient algorithm, termed Winner Beats All (WBA), that achieves a distribution dependent O(n2 log n + n log T) regret bound under any choice model that exhibits a unique GCW, where T is the time-horizon. We complement our upper bound result with an order-wise lower bound of Ω(n log T) for any no-regret algorithm, showing that our algorithm has asymptotically order optimal regret under our general class of choice models. If the underlying model is MNL, then WBA achieves an instance-wise asymptotically optimal regret bound, which is better than the regret bound for the recent Max Min UCB algorithm under MNL [23]. The main challenge in designing an algorithm under our framework is that the space of exploration (number of possible sets the learner can play) is Θ(nk) which is large even for moderate k. Therefore, it can be challenging to simultaneously explore/learn the choice sets with low regret out of the possible Θ(nk) sets and exploit these low regret sets. We overcome these challenges by extracting just O(n2) pairwise statistics from the observed multiway choices under different sets, and using these statistics to find choice sets with low regret. Since these pairwise statistics are extracted from multiway choices under different sets, a technical challenge is to show that these statistics are concentrated. We resolve this challenge by using a novel coupling argument that couples the stochastic process generating choices with another stochastic process, and showing that pairwise estimates according to this other process are concentrated. We believe that our results for efficient learning under this large class of choice models that is considerably more general than the MNL class are of independent interest. We also run experiments on several synthetic and real-world datasets. Our experiments on these datasets show that our algorithm for the special case of k = 2 is competitive as compared to previous dueling bandit algorithms, even though it is designed for a more general setting. For the case of k > 2, we compare our algorithm with the Max Min UCB algorithm of [23] which was designed for the MNL model. We observe that our algorithm performs better in terms of regret than Max Min UCB under all datasets (even under synthetic MNL datasets). We further observe that under several datasets the regret achieved by our algorithm for k > 2 is better than the regret for k = 2. Related Work. There has been some recent interest in bandit settings where more than two arms are pulled at a time, although no work that we are aware of considers the types of general choice models that we do. (1) A related setting to ours is that of multi-dueling bandits [24, 25, 26], where the learner also pulls a set St of k items; however, the feedback received by the learner is assumed to be drawn from a pairwise comparison model (in particular, the learner observes some subset of the k 2 possible pairwise comparisons among items in St). In contrast, in our choice bandits setting, the learner receives the outcome of a direct multiway choice among the items in St, generated from a multiway choice model. (2) In combinatorial bandit with relative feedback [23], the learner pulls a set St of up to k arms, and observes top-m ordered feedback drawn according to the MNL model, for some m k. In contrast, we only observe the (top-1) choice feedback from the set St that is played, but, we study a much more general class of choice models than the MNL model. (3) Another related setting is that of battling bandits [27], where the learner pulls a multiset of k arms and receives feedback indicating which arm was chosen. However, their setting considers a specific pairwise-subset (PS) choice model that is defined in terms of a pairwise comparison model, whereas we consider more general choice models. (4) In stochastic click bandits [28], the learner pulls an ordered set of k arms/documents, and observes clicks on a subset of these documents, drawn according to an underlying click model which is a probabilistic model for click generation over an ordered set. However, click models in their setting are different than choice models in our setting, and neither can be cast as a special case of the other. (5) Another related setting is that of best-of-k bandits [29], where again the learner pulls a set St of k arms. Of the various types of feedback considered in [29], the marked bandit feedback corresponds to the type of feedback that we study, however, the choice models studied in [29] correspond only to a subclass of random utility choice models, and moreover, the analysis in [29] is in the PAC/pure exploration setting, while ours is in the regret minimization setting. (6) Other recent work has specifically considered active learning problems, either in the context of dynamic assortment optimization under MNL where the goal is to maximize expected revenue [30, 31, 32, 33, 34]; or in the context of best arm(s) identification under MNL or IID-RUMs [35, 36] in a PAC/pure exploration setting. (7) Finally, we also mention combinatorial bandits, which have a different goal but also involve pulling subsets of arms [37, 38, 39, 40]. See the supplementary material for more detailed discussion. Organization. We set up the choice bandits problem in Section 2. We give our lower bound result in Section 3. We present our algorithm in Section 4, and its regret analysis in Section 5. We give experimental results in Section 6. We finally conclude with a brief discussion in Section 7. All the proofs can be found in the supplementary material. 2 Problem Setup and Preliminaries In the choice bandits problem, there are n arms [n] := {1, . . . , n}, and a set size parameter 2 k n. On each trial t, the learner pulls (selects/plays) a choice set St [n] of up to k arms, i.e. with |St| k, and receives as feedback yt St, indicating the arm that is most preferred in St. We assume the feedback yt is generated probabilistically from an underlying multiway choice model, which defines for each S [n] such that |S| k, and arm i S, a choice probability Pi|S which corresponds to the probability that arm i is the most preferred arm in S.2 Before defining appropriate notions of best arm and regret for the learner we will give some examples of multiway choice models. Random utility models with i.i.d. noise (IID-RUMs). IID-RUMs are a well-known class of choice models that have origins in the econometrics and marketing literature [21, 41]. Under an IID-RUM, the (random) utility associated with arm i [n] is given by Ui = vi + ϵi where vi R is a deterministic utility and ϵi R is the noise drawn i.i.d. from a distribution D over reals. For a set S, the probability of choosing i S is given by Pi|S = Pr Ui > Uj, j S \{i}). We will sometimes also refer to vi as the weight of item i. Under any IID-RUM if vi > vj for some i, j [n] then arm i will be more likely to be chosen than arm j in any set. The IID-RUM class contains some popular models, such as the multinomial logit (MNL) [17, 18, 19], and multinomial probit (MNP) [20]. Example 1 (MNL). Under MNL, the noise distribution D is a Gumbel(0, 1) and the probability Pi|S of choosing an item i from a set S has the following closed form expression: Pi|S := evi/(P j S evj). It is clear from this expression that arms with higher weights are more likely to be chosen. Example 2 (MNP). Under the MNP model, the noise distribution D is the standard Normal distribution N(0, 1), however, unlike the MNL there is no closed form expression for the choice probabilities. Under IID-RUMs there is a clear notion of best arm: an arm that has the highest weight maxi [n] vi. We now define a strictly more general class of models where there is a clear notion of best arm. 2Note that for the special case of k = 2, our framework reduces to dueling bandits; the pairwise comparison probabilities Pij := Pr (i j) in dueling bandits can be viewed as pairwise choice probabilities Pi|{i,j}. A New Class of Choice Models. We introduce a new class of multiway choice models that are characterized by the following condition that requires the existence of a unique best arm. Definition 1 (Generalized Condorcet Condition (GCC)). A choice model is said to satisfy the GCC condition if there exists a unique arm i [n] such that for every choice set S [n] that contains i , we have Pi |S > Pj|S for all j S \ {i }. Intuitively, the above condition requires the existence of a unique arm that is always (stochastically) preferred to all other arms, no matter what other arms are shown with it. This condition is a generalization of the Condorcet condition studied for pairwise comparison models [6, 11]. Just as the Condorcet condition need not be satisfied for all pairwise comparison models, similarly, GCC need not be satisfied by all multiway choice models. Below we show that the GCC condition is satisfied for all IID-RUMs subject to a minor technical condition. Lemma 1 (IID-RUMs satisfy GCC). For any IID-RUM choice model with utility for arm i [n] given by Ui = vi + ϵi, the GCC condition is satisfied if | argmaxi [n] vi| = 1. In this paper, we study the class of all choice models where the GCC is satisfied. Under GCC, we will refer to this unique best arm as the generalized Condorcet winner (GCW) and denote it by i . Note that for any set S containing the GCW i , we must have Pi |S 1 |S|. Regret Notion. Similar to dueling bandits, the goal of the learner in our setting is to identify the best arm while also playing good/competitive sets with respect to this arm during the exploration phase.3 Hence, our notion of regret measures the sub-optimality of a choice set S relative to i , and is a generalization of the regret defined by [23] for the special case of MNL choice model. Moreover, under our notion of regret it is optimal to play S = {i }, i.e. regret of playing S is 0. The regret of a set is defined to be the sum of regret due to individual arms in the set, and the regret for an arm corresponds to the margin by which the best arm i beats this arm. In other words, the regret of an arm corresponds to the shortfall in preference probability due to pulling this arm over the best arm. Definition 2. The regret r(S) for S [n] is defined as: r(S) := P i S Pi |S {i } Pi|S {i } . This notion of regret can be interpreted as: r(S) is the sum over all arms i S, the fraction of consumers that will choose i minus the fraction of consumers that will choose i when i is played together with S. It is easy to see that r({i }) = 0, and 0 r(S) |S| for any set S [n]. Example 3. Consider a choice model where arm 1 is the GCW, and for each set S containing arm 1, we have P1|S = 0.51 and Pi|S = 0.49 |S| 1 i S \{1}. Then r({1, . . . , m}) = 0.51 (m 1) 0.49. In the above example, the regret increases linearly as we increase m. The following gives an example where the arms are much more competitive and regret is smaller. Example 4. Consider the MNL choice model with weights v1 = log(1 + ϵ), for ϵ > 0, and v2 = = vn = 0. Then r({1, , m}) = P i S ev1 evi P j S evj = ϵ(m 1) The regret here increases much more slowly in terms of m. Note that our regret is not necessarily well-defined in the dueling bandits setting, due to the need to consider choice probabilities for sets of size 3 even when one plays only sets of size 2. In the supplementary material, we give results for an additional notion of regret that is a direct generalization of the dueling bandit regret, and allows for a more direct comparison between our framework and the dueling bandits framework. Under the above notion of regret, the goal of an algorithm A is to minimize its cumulative regret over T trials defined as: R(T) = PT t=1 r(St). 3 A Fundamental Lower Bound In this section we present a regret lower bound for our choice bandits problem. We say that an algorithm is strongly consistent under GCC if its expected regret over T trials is o(T a) for any a > 0 under any model in this class. Before presenting the lower bound let us define the following distribution dependent quantities. i i|S = Pi |S Pi|S Pi |S + Pi|S , GCC max := max S:|S| k max i S i i|S , GCC min := min S:|S| k min i S i i|S . (3.1) 3Note that we are not working in the pure exploration setting, where all sets incur equal cost during exploration. The following theorem presents a lower bound for any strongly consistent algorithm. Theorem 1. Given a set of arms [n], choice set size bound k n, parameter (0, 1), and any strongly consistent algorithm A under GCC, there exists a GCC choice model with GCC min = such that when choice outcomes are drawn from this model we have lim inf T E [R(T)] log T = Ω n 1 where T is the time-horizon. If the underlying model is MNL with parameters v1, v2, vn R, then: lim inf T E[R(T )] log T = Ω P i [n]\{i } 1 MNL i i where MNL i i = evi evi evi +evi , for i [n] \ {i }. Discussion. The above bound shows that any algorithm for the choice bandits problem needs to incur Ω(n log T) regret in the worst case. Note that the above lower bound does not depend on the choice set size parameter k. If the choices are generated from an underlying MNL model, then the above theorem gives an instance-dependent lower bound for the regret of any algorithm. Note that [23] also provided a lower bound under MNL for our notion of regret, however, their bound depends on the worst-case gap between i and any other arm i = i , while we provide a more fine-grained bound under MNL which depends on gaps between i and each individual arm i [n]. In order to prove the above bound we construct a pair of instances that have different GCW arms, and use the information divergence lemma of [42] in order to characterize the minimum number of samples needed in order to collect the information needed to separate these two instances. Remark 1. In order to prove a lower bound for our choice bandits problem one may also be able to use the lower bound given in [43], by casting our problem as a structured bandit problem. However, the lower bound of [43] is in terms of the solution of a linear program, and one will then need to design a distribution over hard instances in order to quantify the solution of this linear program in terms of the gap parameter GCC min . One of the main novelty of our bound is this construction of hard instances that allows us to quantify the lower bound in terms of GCC min . 4 Algorithm In this section we will present our algorithm for the choice bandits problem, termed Winner Beats All (WBA). WBA divides its execution into rounds and each round can contain multiple trials. We will use r as an global index for a round, and t as an global index for a trial. For each round r, WBA maintains a set Ar of active arms, which are a set of arms for which the algorithm is still not confident enough that these are bad arms. Note that an arm that is inactive in a particular round, can become active in a later round. We also maintain a set Q that is initialized to being empty at the beginning of each round and keeps track of the arms in Ar that have been played so far in the round. Given a trial t belonging to round r, WBA selects a special arm termed the anchor arm, and a set S Ar \ Q (arbitrarily) of up to k 1 arms in Ar that have not been played so far in round r. The set S and at are selected such that at empirically performs better than each arm in S. The set S is then played together with arm at (if |S| < k 1, then other arbitrary arms from Ar are added to the played set). The anchor arm is updated in every trial and is chosen so that one can quickly find evidence that arms in S are not good. Let yt be the feedback received in trial t when St was played including anchor at. For all i, j [n], let Nij(t) denote the number of times (up to round t) that either arm i or j was chosen when arm j is the anchor, i.e. Nij(t) := Pt t =1 1(at = j, {i, j} St , yt {i, j}). For each i, j [n] and trial t, such that Nij(t) > 0, the algorithm maintains an estimate of the marginal probability of arm i beating the arm j as ˆPij(t) := 1 Nij(t) t =1 1(at = j, {i, j} St , yt = i) , (4.1) which is the fraction of times i was selected (compared to j) when both i and j were played together and j was the anchor. (When Nij(t) = 0, we can simply take ˆPia(t) to be 1/2.) Similar to [44], let us define an empirical divergence Ii(t, S) which provides a certificate that an arm i is worse than (some) arms in S, as Ii(t, S) = P j S 1[ ˆPij(t) 1 2] Nij(t) d( ˆPij(t), 1 2), where d( ˆPij, 1 2) is the KL-divergence defined as d(P, Q) = P log( P Q) + (1 P) log( 1 P 1 Q), for P, Q [0, 1]. If Ii(t, S) is 0, it means that arm i is empirically at least as good as all other arms in S, and a higher Ii(t, S) would suggest that arm i is most likely bad . For a constant C, we define the condition Ji(t, C) for arm i [n] and round t as Ji(t, C) = 1 n S [n] : Ii(t, S) |S| log(n C) + log(t) o . If Ji(t, C) = 1 for some i, it means that there exists a certificate S to show that i is not likely the best arm as it loses to some arms in S by a large margin .4 The larger the set S the larger the margin needs to be. This condition can be evaluated in polynomial time by computing argmax S [n] Ii(t, S) |S| log(n C) and checking if it is greater than log(t) (details in supplementary material). Finally, let t be the final round in a round r. In order to decide which arms should be included in the next set of active arms Ar+1 we simply check the condition Ji(t, C) for each i [n] and include all arms for which Ji(t, C) = 0 holds. Note that the set of active arms Ar+1 can be empty, in which case we will simply play the anchor arm until it becomes non-empty in the future. The anchor arm in each trial is the arm which empirically beats the maximum number of unplayed arms in the current round. Detailed pseudo-code for WBA is given in Algorithm 1. 5 Regret Analysis In this section we will prove a regret upper bound for our WBA algorithm. The following theorem gives the upper bound. Theorem 2. Let n be the number of arms, k n be the choice set size parameter, and i be the GCW arm . If the multiway choices are drawn according to a GCC choice model with GCC min and GCC max defined in Equation 3.1, then for any C 1/( GCC min )4, the expected regret incurred by WBA is upper bounded by E [R(T)] O n2 log n ( GCC min )2 + O n log(TC) GCC max ( GCC min )2 where T is the (unknown) time-horizon. If the underlying model is MNL with weights v1, , vn R, then for any C 1/( MNL min )4, we have E [R(T)] O n2 log n ( MNL min )2 where MNL i i = evi evi evi +evi and MNL min := mini =i MNL i i . Remark 2 (Selecting C). A value of T 4 for the parameter C suffices for Theorem 2 to hold, giving a regret upper bound of O(log(TC)) = O(log(T 5)) = O(log(T)). (If T is not known, one can use the doubling trick.) To see this note that in order to obtain any non-trivial upper bound for our algorithm, min has to be larger than 1/T. Hence, either min is upper bounded by 1/T, or the instance is too hard to allow any non-trivial upper bound. Therefore, C T 4 would suffice whenever the instance is not already too hard. We actually believe setting C = T 4 may be somewhat pessimistic (it arises from taking a union bound over all possible states of the algorithm in our regret analysis indeed, in our experiments, we set C = 1 for all datasets, and our algorithm still demonstrates sublinear regret with this choice but it certainly suffices, and the regret bound with C = T 4 is at most a constant factor 5 times what one might get with C = 1 if the regret bound holds in that case. Discussion. The above theorem yields a O(n2 log n + n log T) upper bound on regret. Comparing this bound with the lower bound given in Section 3, one can observe this upper bound is asymptotically order-optimal. This upper bound is similar (in order-wise sense) to the upper bounds obtained for some popular dueling bandit algorithms such as RUCB [6], RMED [44], DTS [45] etc. It is also important to note that our regret bound does not depend directly on the choice set size k. However, the behavior of this bound is more subtle and depends on the specific multiway choice model through the gap parameters GCC max and GCC min . We also note that while in general the regret can behave differently for different models, in our experiments, we find that there are choice models (including some in real data) where our algorithm empirically achieves smaller regret when allowed to play sets of size k > 2 as compared to k = 2. If the underlying model is MNL, then our algorithm achieves asymptotically 4Note that the above condition is similar to condition used in [44], except that they only use the set [n] as a certificate instead of all possible subsets S [n]. In our analysis and experiments will show that this condition is an improvement over the condition used in [44] for the case of dueling bandits. Algorithm 1 Winner Beats All (WBA) 1: Input: set of arms [n], size of choice set k, parameter C 2: t 1, r 1, Ar [n], at Unif([n]), Q 3: ˆPij 1 2, i, j [n] 4: while t T do 5: Select largest S Ar \ {Q at} with |S| k 1 and ˆPiat 1 2, i S 6: Let St S {at}; while |St| < k and Ar \ St = : add an (arbitrary) arm from Ar \ St to St 7: Play set St and receive yt St as feedback; Q Q S 8: For all i St, calculate ˆPiat(t) and Ji(t, C) 9: if i Ar \ {Q at} such that ˆPiat(t) 1 2 then 10: at+1 argmaxi [n] P j [n]\Q 1[ ˆPji(t) 1 2] 11: else 12: at+1 at 13: if Q = Ar or S = then 14: Ar+1 , r r + 1 15: for i [n] do 16: if Ji(t, C) = 0, then Ar Ar {i} 17: at+1 argmaxi [n] P j [n] 1[ ˆPji(t) 1 2], Q 18: t t + 1 Figure 1: Datasets used in our experiments 1. MNL-Exp: MNL weights drawn i.i.d. from Exp(λ = 3.5); 2. MNL-Geom: Geometrically decreasing MNL weights: 1, 1 2, . . . , 1 2n 1 ; 3. GCC-One: GCC model defined in Example 3; 4. GCC-Two: GCC model similar to Example 3, but with different choice probabilities; 5. GCC-Three: GCC model similar to Example 3, but with different choice probabilities; 6. Sushi: Choice model extracted from the Sushi dataset [47]; 7. Irish-Dublin: Choice model extracted from Irish-Dublin election dataset; 8. Irish-Meath: Choice model extracted from Irish-Meath election dataset. optimal instance-wise regret that does not depend on k. This instance-wise bound under MNL is an improvement over the upper bound for the Max Min UCB algorithm under MNL for (top-1) choice feedback which depends on worst-case gap parameters [23]. An important point to note is that we do not need a specialized algorithm for MNL in order to achieve an instance-wise bound under MNL. Proof Ideas. Our algorithm is based on the idea of isolating a good anchor arm and playing arms that are competitive against this anchor. Hence, in order to prove a regret upper bound we need to show that the GCW i would eventually beat every other arm i, i.e. ˆPi i(t) (Equation 4.1) would eventually become larger than 1/2. In this case i would become the anchor arm. However, an important technical challenge here is to bound the deviation in these pairwise estimates ˆPi i(t) obtained from multiway choices. In the past, [46] have shown that if one uses rank breaking to extract pairwise estimates under the MNL model, then these pairwise estimates will be concentrated. However, this concentration result relies crucially on the independence from irrelevant attributes (IIA) property of MNL which states that for any two arms, the odds of choosing one over the other in any set remains the same regardless of which set is shown. This concentration result does not apply to our setting as the IIA property does not hold for general GCC models beyond the MNL. Below we outline a novel coupling argument that allows us to prove concentration for the extracted pairwise estimates between the GCW arm i and any other arm i [n] Lemma 2 (Concentration). Consider a GCC choice model with GCW i . Fix i [n]. Let S1, , ST be a sequence of subsets of [n] and y1, , y T be a sequence of choices according to this model, let Ft = {S1, y1, , St, yt} be a filtration such that St+1 is a measurable function of Ft. We have Pr( ˆPi i(t) P GCC i i ϵ and Ni i(t) m) e d(P GCC i i ϵ,P GCC i i ) m (5.1) where P GCC i i = min S:|S| k,{i ,i} S Pi |S Pi |S+Pi|S , and d( , ) is the KL-divergence. Proof Sketch. Let us consider an alternate process for generating multiway choices y t from sets St. In this process, given any t and a set St such that i , i St with at = i, we first generate a Bernoulli random variable Xt with probability Pi |S +Pi|S. If Xt = 0 we set y t = j with probability Pj|S 1 Pi |S Pi|S , for j S \ {i, i }. If Xt = 1 then we sample another Bernoulli random variable Zt with probability P GCC i i . If Zt = 1 then we let y t = i , otherwise if Zt = 0 we set y t = i. Let Pi i|St = Pi |St/(Pi |St + Pi|St). Now, we couple y t and yt as follows: if y t St \ {i} then we let yt = y t, otherwise if y t = i then we let yt = i with probability (Pi i|St P GCC i i )/(1 P GCC i i ) and let yt = i with probability (1 Pi i|St)/(1 P GCC i i ). One can verify that yt is distributed according to the correct underlying choice distribution. It is now easy to observe that the estimates ˆPi i(t) under yt will always be larger than the estimates ˆP i i(t) under y t, hence, we will have that 100 103 106 t: round 101 R(t): regret WBA, k=2 DTS RMED1 BTM RUCB 100 103 106 t: round 101 R(t): regret WBA, k=2 DTS RMED1 BTM RUCB 100 103 106 t: round 101 R(t): regret WBA, k=2 DTS RMED1 BTM RUCB 100 103 106 t: round 101 R(t): regret WBA, k=2 DTS RMED1 BTM RUCB 100 103 106 t: round 101 R(t): regret WBA, k=2 DTS RMED1 BTM RUCB 100 103 106 t: round R(t): regret WBA, k=2 DTS RMED1 BTM RUCB 100 103 106 t: round R(t): regret irish-dublin WBA, k=2 DTS RMED1 BTM RUCB 100 103 106 t: round R(t): regret irish-meath WBA, k=2 DTS RMED1 BTM RUCB Figure 2: Regret v/s trials for our algorithm WBA (for k = 2) against dueling bandit algorithms (DTS, BTM, RUCB and RMED1) (the shaded region corresponds to std. deviation). As can be observed, our algorithm is competitive against these algorithms. 100 103 106 t: round 101 R(t): regret WBA, k=5 WBA, k=2 MMU, k=5 MMU, k=2 100 103 106 t: round 101 R(t): regret WBA, k=5 WBA, k=2 MMU, k=5 MMU, k=2 100 103 106 t: round 101 R(t): regret WBA, k=5 WBA, k=2 MMU, k=5 MMU, k=2 100 103 106 t: round 101 R(t): regret WBA, k=5 WBA, k=2 MMU, k=5 MMU, k=2 100 103 106 t: round 101 R(t): regret WBA, k=5 WBA, k=2 MMU, k=5 MMU, k=2 100 103 106 t: round R(t): regret WBA, k=5 WBA, k=2 MMU, k=5 MMU, k=2 100 103 106 t: round R(t): regret irish-dublin WBA, k=5 WBA, k=2 MMU, k=5 MMU, k=2 100 103 106 t: round R(t): regret irish-meath WBA, k=5 WBA, k=2 MMU, k=5 MMU, k=2 Figure 3: Regret v/s trials for our algorithm WBA against the Max Min UCB (MMU) algorithm for k = 2 and k = 5 (the shaded region corresponds to std. deviation). We observe that our algorithm is better than Max Min UCB on all datasets for both values of k. We further observe that under several datasets the regret achieved by our algorithm for k > 2 is better than the regret of our algorithm for k = 2. Pr( ˆPi i(t) x) Pr( ˆP i i(t) x) for any x > 0. One can then show concentration for the coupled estimates ˆP i i(t), and use it to bound the deviation in ˆPi i(t). Note that the above lemma only shows concentration for the pairwise estimates ˆPi i(t) between i and any other arm i [n], but not for estimates ˆPij(t) between two arbitrary arms i [n] and j [n]. However, in order to prove our result we only need concentration of estimates between i and any other arm i [n]. We believe that the above concentration lemma is of independent interest, and might be useful in other learning from multiway choice settings beyond MNL. Once we have bounded the deviation for the pairwise estimates, we bound the number of rounds r in which i is not a part of the active set Ar. We then bound the expected number of times that there exists an arm i such that ˆPi i(t) < 1 2, thus bounding the number of trials until i becomes the anchor. Finally, once i is the anchor arm, we bound the regret incurred due to sub-optimal arms. 6 Experiments We compared the performance of our WBA algorithm and other existing algorithms on our choice bandit problem under different choice models. The first two choice models were MNL models, the next three were from the GCC class, and the last three we extracted from real-world datasets. Details of these models are in Figure 1 (additional details can be found in the supplementary material). Below we describe the different sets of experiments that were performed. Each experiment was repeated 10 times. The value of n was 100 for all synthetic datasets, 16 for Sushi, 8 for Irish-Dublin, and 12 for Irish-Meath. The parameter C in our algorithm was set to 1. Comparison with Dueling Bandit Algorithms (k = 2). For the special case of k = 2, we compared our algorithms with a representative set of dueling bandit algorithms (RMED1 [11], DTS [45], RUCB [6], BTM [2]). Note that the purpose of these experiments is merely to perform a sanity check and ensure that our algorithm performs reasonably well compared with dueling bandit baselines when k = 2; the goal is not to argue that our choice bandit algorithm beats the state-of-the-art for the specialized dueling bandit (k = 2) setting. We set α = 0.51 for RUCB and DTS, and f(K) = 0.3K1.01 for RMED, and γ = 1.3 for BTM. Figure 2 contain plots for these comparisons. Our algorithm either performs better or similar to RMED1, RUCB, and BTM on all datasets; and is competitive with DTS on most of the datasets. Comparison with Max Min UCB Algorithm [23] (k > 2). We compared the performance of our algorithm with the recent Max Min UCB algorithm [23] that was designed and analyzed primarily for MNL choice models under the same notion of regret as ours. 5 We set the parameter α to be 0.51 for Max Min UCB. Figure 3 contain plots for these experiments for k = 2 and k = 5. We observe that our algorithm is much better in terms of regret than Max Min UCB under all datasets for both values of k. One should note that WBA performs better than Max Min UCB even under the MNL datasets, even though Max Min UCB is specialized to MNL while our algorithm works under more general models. We further observe that under several datasets (GCC-One, GCC-Two, Sushi, Irish-Dublin) the regret achieved by our algorithm for k > 2 is better than for k = 2. Note that even though our study of more general choice feedback is motivated by applications where it might be desirable to pull sets of size larger than 2 due to reasons other than improving regret, these experimental results show that there exist settings of choice models (including some in real data) where our algorithm empirically achieves a smaller regret when allowed to play sets of size k > 2 as compared to k = 2. 7 Conclusion We have introduced a new framework for bandit learning from choice feedback that generalizes the dueling bandit framework. Our main result is to show that computationally efficient learning is possible in this more general framework under a wide class of choice models that is considerably more general than the previously studied class of MNL models. Our algorithm for this general setting, termed Winner Beats All (WBA), achieves order-wise optimal regret for the general class of GCC models. For the special case k = 2, WBA is competitive with previous dueling bandit algorithms; for k > 2, WBA outperforms the recently proposed Max Min UCB (MMU) algorithm even on MNL models for which MMU was designed. Acknowledgement Thanks to Aadirupa Saha for early discussions related to this work, and to the anonymous referees for helpful comments. This material is based upon work supported in part by the US National Science Foundation (NSF) under Grant Nos. 1717290 and 1934876. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation. Broader Impact The purpose of this paper is to understand whether efficient learning is possible in a bandit setting where one does not receive quantitative feedback for an individual arm but rather relative feedback in the form of a multiway choice. It is well-known that quantitative judgments of humans can have biases; our algorithm, which learns from relative multiway choices, can help alleviate these biases. Moreover, by receiving larger choice sets from our algorithm, humans can have a better sense of the quality distribution of arms, and can make more informed choices. 5 We also considered the Self Sparring algorithm of [26] and the battling bandit algorithms of [27], which are applicable to choice models defined in terms of an underlying pairwise comparison model P. However, these algorithms all return multisets St, and any simple reduction of such multisets to strict sets as considered in our setting (as well as the setting of [23]) can end up throwing away important information learned by the algorithms, resulting in a comparison that could be unfair to those algorithms. We did explore such reductions and our algorithm easily outperformed them, but we chose not to include the results here due to this issue of fairness. (Moreover, under the MNL model, [23] already established that Max Min UCB outperforms those algorithms presumably under similar reductions so in the end, we decided such a comparison would provide little additional value here.) Another advantage of our setting is that we do not rely on historic data as our data collection is online. Hence, one does not need to worry about past biases being reflected in the choice datasets. However, one has to be cautious about the use of our algorithm in applications where arms represent individuals/entities such as job applicants, property renters etc. In these applications, the choices of people can be biased against certain individuals/groups, thereby hurting the chances of these individuals/groups to be selected by our algorithm. Here, depending on the application, one might need to consider imposing some form of fairness constraints on the choice sets output by our algorithm in order to prevent any discrimination against such individuals/groups. [1] Yisong Yue, Josef Broder, Robert Kleinberg, and Thorsten Joachims. The K-armed Dueling Bandits Problem. In Proceedings of the 22nd Conference on Learning Theory, 2009. [2] Yisong Yue and Thorsten Joachims. Beat the mean bandit. In Proceedings of the 28th International Conference on Machine Learning, 2011. [3] Yisong Yue, Josef Broder, Robert Kleinberg, and Thorsten Joachims. The k-armed dueling bandits problem. J. Comput. Syst. Sci., 78(5):1538 1556, 2012. [4] Tanguy Urvoy, Fabrice Clerot, Raphael Feraud, and Sami Naamane. Generic Exploration and K-armed Voting Bandits. In Proceedings of the 30th International Conference on Machine Learning, 2013. [5] Nir Ailon, Zohar Karnin, and Thorsten Joachims. Reducing Dueling Bandits to Cardinal Bandits. In Proceedings of the 31st International Conference on Machine Learning, 2014. [6] Masrour Zoghi, Shimon Whiteson, Remi Munos, and Maarten de Rijke. Relative Upper Confidence Bound for the K-Armed Dueling Bandit Problem. In Proceedings of the 31st International Conference on Machine Learning, 2014. [7] Masrour Zoghi, Zohar Karnin, Shimon Whiteson, and Maarten de Rijke. Copeland Dueling Bandits. In Advances in Neural Information Processing Systems 28, 2015. [8] Masrour Zoghi, Shimon Whiteson, and Maarten de Rijke. Merge RUCB: A method for large-scale online ranker evaluation. In Proceedings of the 8th ACM International Conference on Web Search and Data Mining, 2015. [9] Miroslav Dudik, Katja Hofmann, Robert E. Schapire, Aleksandrs Slivkins, and Masrour Zoghi. Contextual Dueling Bandits. In Proceedings of the 28th Conference on Learning Theory, 2015. [10] Kevin Jamieson, Sumeet Katariya, Atul Deshpande, and Robert Nowak. Sparse Dueling Bandits. In Proceedings of the 18th International Conference on Artificial Intelligence and Statistics, 2015. [11] Junpei Komiyama, Junya Honda, Hisashi Kashima, and Hiroshi Nakagawa. Regret Lower Bound and Optimal Algorithm in Dueling Bandit Problem. In Proceedings of the 28th Conference on Learning Theory, 2015. [12] Junpei Komiyama, Junya Honda, and Hiroshi Nakagawa. Copeland Dueling Bandit Problem: Regret Lower Bound, Optimal Algorithm, and Computationally Efficient Algorithm. In Proceedings of the 33rd International Conference on Machine Learning, 2016. [13] Siddartha Ramamohan, Arun Rajkumar, and Shivani Agarwal. Dueling Bandits : Beyond Condorcet Winners to General Tournament Solutions. In Advances in Neural Information Processing Systems 29, 2016. [14] Bangrui Chen and Peter I. Frazier. Dueling Bandits with Weak Regret. In Proceedings of the 34th International Conference on Machine Learning, 2017. [15] Yisong Yue and Thorsten Joachims. Interactively Optimizing Information Retrieval Systems as a Dueling Bandits Problem. In Proceedings of the 26th International Conference on Machine Learning, 2009. [16] Eric J Johnson, Suzanne B Shu, Benedict GC Dellaert, Craig Fox, Daniel G Goldstein, Gerald Häubl, Richard P Larrick, John W Payne, Ellen Peters, David Schkade, et al. Beyond nudges: Tools of a choice architecture. Marketing Letters, 23(2):487 504, 2012. [17] Robert Duncan Luce. Individual choice behavior: A theoretical analysis. Wiley, 1959. [18] Robin L. Plackett. The Analysis of Permutations. Journal of the Royal Statistical Society. Series C (Applied Statistics), 24(2):193 202, 1975. [19] Daniel Mc Fadden. Conditional Logit Analysis of Qualitative Choice Analysis. New York: Academic Press, 1974. [20] Louis L Thurstone. A law of comparative judgment. Psychological review, 34(4):273, 1927. [21] Jacob Marschak. Binary choice constraints and random utility indicators. In Stanford Symposium on Mathematical Methods in the Social Sciences, page 312 329, 1960. [22] TA Domencich and D Mc Fadden. Urban travel demand; a behavioural analysis. North-Holland, 1975. [23] Aadirupa Saha and Aditya Gopalan. Combinatorial bandits with relative feedback. In Advances in Neural Information Processing Systems, pages 983 993, 2019. [24] Brian Brost, Yevgeny Seldin, Ingemar J. Cox, and Christina Lioma. Multi-Dueling Bandits and Their Application to Online Ranker Evaluation. In Proceedings of the 25th ACM International Conference on Information and Knowledge Management, 2016. [25] Anne Schuth, Harrie Oosterhuis, Shimon Whiteson, and Maarten de Rijke. Multileave Gradient Descent for Fast Online Learning to Rank. In Proceedings of the 9th ACM International Conference on Web Search and Data Mining, 2016. [26] Yanan Sui, Vincent Zhuang, Joel W. Burdick, and Yisong Yue. Multi-dueling Bandits with Dependent Arms. In Proceedings of the 33rd Conference on Uncertainty in Artificial Intelligence, 2017. [27] Aadirupa Saha and Aditya Gopalan. Battle of bandits. In UAI, pages 805 814, 2018. [28] Masrour Zoghi, Tomas Tunys, Mohammad Ghavamzadeh, Branislav Kveton, Csaba Szepesvari, and Zheng Wen. Online learning to rank in stochastic click models. In ICML, pages 4199 4208, 2017. [29] Max Simchowitz, Kevin Jamieson, and Benjamin Recht. Best-of-K Bandits. In Proceedings of the 29th Annual Conference on Learning Theory, 2016. [30] Paat Rusmevichientong, Zuo-Jun Max Shen, and David B. Shmoys. Dynamic Assortment Optimization with a Multinomial Logit Choice Model and Capacity Constraint. Operations Research, 58(6):1666 1680, 2010. [31] Denis Sauré and Assaf Zeevi. Optimal Dynamic Assortment Planning with Demand Learning. Manufacturing & Service Operations Management, 15(3):387 404, 2013. [32] Shipra Agrawal, Vashist Avadhanula, Vineet Goyal, and Assaf Zeevi. A Near-Optimal Exploration Exploitation Approach for Assortment Selection. In Proceedings of the 17th ACM Conference on Economics and Computation, 2016. [33] Shipra Agrawal, Vashist Avadhanula, Vineet Goyal, and Assaf Zeevi. Thompson Sampling for the MNL-Bandit. In Proceedings of the 30th Conference on Computational Learning Theory, 2017. [34] Xi Chen and Yining Wang. A Note on Tight Lower Bound for MNL-Bandit Assortment Selection Models. Technical report, ar Xiv:1709.06109v2, 2017. [35] Xi Chen, Yuanzhi Li, and Jieming Mao. A Nearly Instance Optimal Algorithm for Top-k Ranking under the Multinomial Logit Model. In Proceedings of the 29th Annual ACM-SIAM Symposium on Discrete Algorithms, 2018. [36] Aadirupa Saha and Aditya Gopalan. Best-item learning in random utility models with subset choices. In AISTATS, 2020. [37] Yi Gai, Bhaskar Krishnamachari, and Rahul Jain. Combinatorial Network Optimization With Unknown Variables: Multi-Armed Bandits With Linear Rewards and Individual Observations. IEEE/ACM Transactions on Networking, 20(5):1466 1478, 2012. [38] Wei Chen, Yajun Wang, and Yang Yuan. Combinatorial Multi-Armed Bandit: General Framework, Results and Applications. In Proceedings of the 30th International Conference on Machine Learning, 2013. [39] Branislav Kveton, Zheng Wen, Azin Ashkan, and Csaba Szepesvari. Tight Regret Bounds for Stochastic Combinatorial Semi-Bandits. In Proceedings of the 18th International Conference on Artificial Intelligence and Statistics, 2015. [40] Richard Combes, M. Sadegh Talebi, Alexandre Proutiere, and Marc Lelarge. Combinatorial Bandits Revisited. In Advances in Neural Information Processing Systems 28, 2015. [41] Kenneth E. Train. Discrete Choice Methods with Simulation. Cambridge University Press, 2003. [42] Emilie Kaufmann, Olivier Cappe, and Aurelien Garivier. On the Complexity of Best-Arm Identification in Multi-Armed Bandit Models. Journal of Machine Learning Research, 17(1):1 42, 2016. [43] Richard Combes, Stefan Magureanu, and Alexandre Proutière. Minimal exploration in structured stochastic bandits. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 4-9 December 2017, Long Beach, CA, USA, pages 1763 1771, 2017. [44] Junpei Komiyama, Junya Honda, and Hiroshi Nakagawa. Optimal Regret Analysis of Thompson Sampling in Stochastic Multi-armed Bandit Problem with Multiple Plays. In Proceedings of the 32nd International Conference on Machine Learning, 2015. [45] Huasen Wu and Xin Liu. Double thompson sampling for dueling bandits. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, December 5-10, 2016, Barcelona, Spain, pages 649 657, 2016. [46] Aadirupa Saha and Aditya Gopalan. PAC battling bandits in the plackett-luce model. In Algorithmic Learning Theory, ALT 2019, 22-24 March 2019, Chicago, Illinois, USA, pages 700 737, 2019. [47] Toshihiro Kamishima. Nantonac collaborative filtering: recommendation based on order responses. In Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, August 24 - 27, 2003, pages 583 588, 2003.