# multiarmed_bandit_requiring_monotone_arm_sequences__11b25cb0.pdf Multi-armed Bandit Requiring Monotone Arm Sequences Ningyuan Chen Rotman School of Management, University of Toronto 105 St George St, Toronto, ON, Canada ningyuan.chen@utoronto.ca In many online learning or multi-armed bandit problems, the taken actions or pulled arms are ordinal and required to be monotone over time. Examples include dynamic pricing, in which the firms use markup pricing policies to please early adopters and deter strategic waiting, and clinical trials, in which the dose allocation usually follows the dose escalation principle to prevent dose limiting toxicities. We consider the continuum-armed bandit problem when the arm sequence is required to be monotone. We show that when the unknown objective function is Lipschitz continuous, the regret is O(T). When in addition the objective function is unimodal or quasiconcave, the regret is O(T 3/4) under the proposed algorithm, which is also shown to be the optimal rate. This deviates from the optimal rate O(T 2/3) in the continuous-armed bandit literature and demonstrates the cost to the learning efficiency brought by the monotonicity requirement. 1 Introduction In online learning problems such as the multi-armed bandits (MAB), the decision maker chooses from a set of actions/arms with unknown reward. The goal is to learn the expected rewards of the arms and choose the optimal one as much as possible within a finite horizon. The framework and its many variants have been successfully applied to a wide range of practical problems, including news recommendation [31], dynamic pricing [9] and medical decision-making [8]. Popular algorithms such as UCB [4, 5] and Thompson sampling [3, 34] typically explore the arms sufficiently and as more evidence is gathered, converge to the optimal arm. These algorithms are designed to handle the classic MAB problems, in which the arms are discrete and do not have any order. However, in some practical problems, the arms are ordinal and even continuous. This is the focus of the study on the so-called continuum-armed bandit [2, 26, 6, 11, 38]. By discretizing the arm space properly, algorithms such as UCB and Thompson sampling can still be applied. Therefore, the continuous decision space has readily been incorporated into the framework. Another practical issue, which is not studied thoroughly but emerges naturally from some applications, is the requirement that the action/arm sequence is has to be monotone. Such a requirement is usually imposed due to external reasons that cannot be internalized to be part of the regret. We give two concrete examples below. Dynamic pricing. When launching a new product, how does the market demand reacts to the price is typically unknown. Online learning is a popular tool for the firm to learn the demand and make profits simultaneously. See [21] for a comprehensive review. In this case, the arms are the prices charged by the firm. However, changing prices arbitrarily over time may have severe adverse effects. First, early adopters of the product are usually loyal customers. Price drops at a later stage may cause a backlash in this group of customers, whom the firm is the least willing to offend. Second, customers 35th Conference on Neural Information Processing Systems (Neur IPS 2021). are usually strategic, as studied extensively in the literature [39, 32, 44, 31, 16]. The possibility of a future discount may make consumers intentionally wait and hurt the revenue. A remedy for both issues is a markup pricing policy, in which the firm increases the price over time. This is particularly common for digital services growing rapidly: both Dropbox and Google Drive went through the phases from free, to low fees, to high fees. In a slightly different context, markup pricing is often used in successful crowdfunding campaigns: to reward early backers, the initial products may be given for free; later after some goals are achieved, backers may get the product at a discount; after the release, the product is sold at the full price. It is not a good idea to antagonize early backers by lowering the price later on. In the language of multi-armed bandit, it is translated to the requirement that the pulled arms in each period are increasing over time. Clinical trials. Early stage clinical trials for the optimal dose are usually conducted in Phase I/II trials sequentially. For Phase I, the toxicity of the doses is examined to determine a safe dose range. In this phase, the dose allocation commonly follows the dose escalation principle [30] and the 3+3 design is the most common dose allocation approach. For Phase II, the efficacy of the doses are tested within the safe dose range (by random assignment, for example). Recently, especially in oncology, there are proposals to integrate both phases [42] which is called seamless Phase I/II trials. In this case, dose escalation has to be observed when looking for the most efficacious dose. Our study is thus motivated by seamless trials. In the online learning framework, the dosage is treated as arms and the efficacy is the unknown reward. Multi-armed bandit has been applied to the dose allocation to identify the optimal dose [41, 7]. In particular, after a dose has been assigned to a cohort of patients, their responses are recorded before the next cohort is assigned a higher dose. If any of the patients experiences dose limiting toxicities (DLTs), then the trial stops immediately. The dose escalation principle requires the allocated doses to be increasing over time. A recent paper incorporates the principle in the design of adaptive trials [13]. To accommodate the above applications, in this paper, we incorporate the requirement of monotone arm sequences in online learning, or more specifically, the continuum-armed bandit framework. We consider an unknown continuous function f(x) and the decision maker may choose an action/arm Xt in period t. The observed reward Zt is a random variable with mean f(Xt). The requirement of monotone arm sequences is translated to X1 X2 XT , where T is the length of the horizon. The goal of the decision maker is to minimize the regret T max f(x) PT t=1 E[f(Xt)]. Main results. Next we summarize the main results of the paper informally. We first show that for a general continuous function f(x), the extra requirement makes learning impossible. Result. For the class of continuous functions f(x), the regret is at least T/4. This diverges from the results in the continuum-armed bandit literature. In particular, [26] shows that the optimal regret for continuous functions scales with T 2/3 without the monotonicity requirement. Therefore, to prove meaningful results, we need to restrict the class of functions f(x). From both the practical and theoretical points of view, we find that assuming f(x) to be unimodal or quasiconcave is reasonable. In practice, for the two applications mentioned above, the revenue function in dynamic pricing is typically assumed to be quasiconcave in the price [47]; for clinical trials, the majority of the literature assumes the dose-response curve to be increasing [10], increasing with a plateau effect [33], or unimodal [45], all of which satisfy the assumption. In theory, with the additional assumption, we can show that Result. There is an algorithm that has monotone arm sequences and achieves regret O(log(T)T 3/4) for continuous and quasiconcave functions f(x). Note that the rate T 3/4 is still different from that in the continuum-armed bandit literature. We then proceed to close the case by arguing that this is the fundamental limit of learning, not due to the design of the algorithm. Result. No algorithm can achieve less regret than T 3/4/32 for the instances we construct. The constructed instances that are hard to learn (see Figure 2) are significantly different from the literature. In fact, the classic construction of needle in a haystack [27] does not leverage the requirement of monotone arm sequences and does not give the optimal lower bound. We prove the lower bound by using a novel random variable that is only a stopping time under monotone arm sequences. By studying the Kullback-Leibler divergence of the probabilities measures involving Table 1: The optimal regret under different settings in the continuum-armed bandit. The regret in [18] is achieved under additional assumptions. Paper Concavity Quasiconcavity Monotone Arm Sequences Regret [26] N N N O(T 2/3) This paper N N Y O(T) [20] Y Y N O(T 1/2) [35] Y Y Y O(T 1/2) [18] N Y N O(T 1/2) This paper N Y Y O(T 3/4) the stopping time, we are able to obtain the lower bound. We also show numerical studies that complement the theoretical findings. Related literature. The framework of this paper is similar to the continuum-armed bandit problems, which are widely studied in the literature. The earlier studies focus on univariate Lipschitz continuous functions [2, 26] and identify the optimal rate of regret O(T 2/3). Later papers study various implications including smoothness [6, 15, 24], the presence of constraints and convexity [9, 43], contextual information and high dimensions [38, 12], and so on. The typical reduction used in this stream of literature is to adopt a proper discretization scheme and translate the problem into discrete arms. In the regret analysis, the Lipschitz continuity or smoothness can be used to control the discretization error. Our setting is similar to the study in [26] with quasiconcave objective functions and the additional requirement of monotone arm sequences. The optimal rate of regret, as a result of the difference, deteriorates from T 2/3 to T 3/4. This deterioration can be viewed as the cost of the monotonicity requirement. In a recent paper, [35] study a similar problem while assuming the function f(x) is strictly concave. The motivation is from the notion of fairness in sequential decision making [46]. The concavity allows them to develop an algorithm based on the gradient information. The resulting regret is O(T 1/2), the same as continuum-armed bandit with convexity/concavity [20, 1]. Comparing to this set of results, the strong concavity prevents the deterioration of regret rates under the monotonicity requirement. In contrast, [18, 19] show that the same rate of regret O(T 1/2) can be achieved if the function is not strictly concave, but just unimodal or quasiconcave, under additional technical assumptions. It is worth noting that many results in this paper are discovered independently in a recent paper [25]. The comparison of the regret rates are shown in Table 1. Other requirements on the arm sequence motivated by practical problems are also considered in the literature. [17, 36, 37] are motivated by a similar concern in dynamic pricing: customers are hostile to frequent price changes. They propose algorithms that have a limited number of switches and study the impact on the regret. Motivated by Phase I clinical trials, [22] study the best arm identification problem when the reward of the arms are monotone and the goal is to identify the arm closest to a threshold. [14] consider the setting when the reward is nonstationary and periodic, which is typical in the retail industry as demand exhibits seasonality. Notations. We use Z+ and R to denote all positive integers and real numbers, respectively. Let T Z+ be the length of the horizon. Let [m] := {1, . . . , m} for any m Z+. Let x = arg maxx [0,1] f(x) be the set of maximizers of a function f(x) in [0, 1]. The indicator function is represented by 1. 2 Problem Setup The objective function f(x) : [0, 1] [0, 1] is unknown to the decision maker. In each period t [T], the decision maker chooses Xt [0, 1] and collects a random reward Zt which is independent of everything else conditional on Xt and satisfies E[Zt|Xt] = f(Xt). After T periods, the total (random) reward collected by the decision maker is thus PT t=1 Zt. Since the decision maker does not know f(x) initially, the arm chosen in period t only depends on the previously chosen arms and observed rewards. That is, for a policy π of the decision maker, we have Xt = πt(X1, Z1, . . . , Xt 1, Zt 1, Ut), where Ut is an independent random variable associated with the internal randomizer of πt (e.g., Thompson sampling). Before proceeding, we first introduce a few standard assumptions. Assumption 1. The function f(x) is c-Lipschitz continuous, i.e., |f(x1) f(x2)| c|x1 x2|, for x1, x2 [0, 1]. This is a standard assumption in the continuum-armed bandit literature [26, 6, 11, 38]. Indeed, without continuity, the decision maker has no hope to learn the objective function well. In this study, we do not consider other smoothness conditions. Assumption 2. The reward Zt is σ-subgaussian for all t [T]. That is, conditional on Xt, E[exp(λ(Zt f(Xt)))|Xt] exp(λ2σ2/2) for all λ R. Subgaussian random variables are the focus of most papers in the literature, due to their benign theoretical properties. See [29] for more details. We can now define an oracle that has the knowledge of f(x). The oracle would choose arm x arg maxx [0,1] f(x) and collect the total reward with mean T maxx [0,1] f(x). Because of Assumption 1, the maximum is well defined. We define the regret of a policy π as Rf,π(T) = T max x [0,1] f(x) t=1 E[f(Xt)]. Note that the regret depends on the unknown function f(x), which is not very meaningful. As a standard setup in the continuum-armed bandit problem, we consider the worst-case regret for all f satisfying Assumption 2. Rπ(T) := max f Rf,π(T). Next we present the key requirement in this study: the sequence of arms must be increasing. Requirement 1 (Increasing Arm Sequence). The sequence {Xt}T t=1 must satisfy almost surely under π. Note that increasing can be relaxed to monotone without much effort. For the simplicity of exposition, we only consider the increasing arm sequence. As explained in the introduction, this is a practical requirement in several applications. In phase I/II clinical trials, the quantity Xt is the dose applied to the t-th patient and f(x) is the average efficacy of dose x. The dose escalation principle protects the patients from DLTs. More precisely, the doses applied to the patients must increase gradually over time such that the trial can be stopped before serious side effects occur to a batch of patients under high doses. In dynamic pricing, Xt is the price charged for the t-th customer and f(x) is the expected revenue under price x. The increasing arm sequence corresponds to a markup pricing strategy that is typical for new products. Because of the increasing price, early adopters, who are usually loyal to the product, wouldn t feel betrayed by the firm. Moreover, it deters strategic consumer behavior that may cause a large number of customers to wait for promotions, hurting the firm s profits. We point out that under Assumption 1 and 2, which typically suffice to guarantee sublinear regret, the regret Rπ(T) is at least T/4 under Requirement 1 for all policies π. Proposition 1. For all π satisfying Requirement 1, we have Rπ(T) T/4 under Assumptions 1 with c = 2 and Assumption 2. The result can be easily illustrated by the two functions (f1(x) in the left and f2(x) in the right) in Figure 1. Because of Requirement 1 and the fact that f1(x) = f2(x) for x [0, 0.5], we must have 0 0.2 0.4 0.6 0.8 1 0 0 0.2 0.4 0.6 0.8 1 0 Figure 1: Two functions (f1(x) in the left and f2(x) in the right) that any policy cannot differentiate before any t such that Xt 0.5 under Requirement 1. PT t=1 1Xt 0.5 being equal under f1(x) and f2(x) for any given policy π. Moreover, it is easy to see that t=1 1Xt 0.5, Rf2,π(T) T t=1 1Xt>0.5 1 t=1 1Xt 0.5. Recall the fact PT t=1 1Xt 0.5 + PT t=1 1Xt>0.5 = T. The above inequalities lead to Rf1,π(T) + Rf2,π(T) T/2, which implies Proposition 1. Therefore, to obtain sublinear regret, we need additional assumptions for f(x). In particular, we assume Assumption 3. The function f(x) is quasiconcave, i.e., for any x1, x2, λ [0, 1], we have f(λx1 + (1 λ)x2) min {f(x1), f(x2)} . In other words, a quasiconcave function is weakly unimodal and does not have multiple separated local maximums. Therefore, f2(x) in Figure 1 is not quasiconcave while f1(x) is. Note that a quasiconcave function may well have a plateau around the local maximum. We impose Assumption 3 on the function class due to two reasons. First, quasiconcavity is probably the weakest technical assumption one can come up with that is compatible with Requirement 1. It is much weaker than concavity [35], and it is easy to see from Figure 1 that multiple local maximums tend to cause linear regret under Requirement 1. Moreover, unimodal functions are the focus of other studies in the online learning literature [18, 19]. Second, objective functions in practice are commonly assumed to be quasiconcave. For the dynamic pricing example, the revenue function is typically assumed to be quasiconcave in the price [47], supported by theoretical and empirical evidence. In clinical trials, most literature assumes the dose-response curve to be quasiconcave [10, 45, 33]. 3 The Algorithm and the Regret Analysis In this section, we present the algorithm and analyze its regret. We first introduce the high-level idea of the algorithm. The algorithm discretizes [0, 1] into K evenly-spaced intervals with end points {0, 1/K, . . . , (K 1)/K, 1}, where K is a tunable hyper-parameter. The algorithm pulls arm k/K for m times sequentially for k = 0, 1, . . . , for a hyper-parameter m. It stops escalating k after a batch of m pulls if the reward of arm k/K is not as good as that of a previous arm. More precisely, the stopping criterion is that the upper confidence bound for the average reward of arm k/K is less than the lower confidence bound for that of some arm i/K for i < k. When the stopping criterion is triggered, the algorithm always pulls arm k/K afterwards. The detailed steps of the algorithm are shown in Algorithm 1. The intuition behind Algorithm 1 is easy to see. Because of Assumption 3, the function f(x) is unimodal. Therefore, when escalating the arms, the algorithm keeps comparing the upper confidence Algorithm 1 Increasing arm sequence Input: K, m, σ Initialize: k 0, t 0, S 0 while S = 0 do Stopping criterion Choose Xt+1 = = Xt+m = k/K and observe rewards Zt+1, . . . , Zt+m Calculate the confidence bounds for arm k/K i=t+1 Zi + σ if UBk < LBi for some i < k then t t + m, k k + 1 end if end while Set Xt k/K for the remaining t bound of the current arm k/K with the lower confidence bound of the historically best arm. If the upper confidence bound of k/K is lower, then it means k/K is likely to have passed the mode of f(x). At this point, the algorithm stops escalating and keeps pulling k/K for the rest of the horizon. Clearly, Algorithm 1 satisfies Requirement 1. The regret of Algorithm 1 is provided in Theorem 2. Theorem 2. By choosing K = T 1/4 and m = T 1/2 , the regret of Algorithm 1 satisfies Rπ(T) 3c + 11 log T T 3/4 when T 16, under Assumptions 1 to 3. We note that to achieve the regret in Theorem 2, the algorithm requires the information of T and σ but not the Lipschitz constant c. The requirement for σ may be relaxed using data-dependent confidence intervals such as bootstrap, proposed in [28, 23]. The requirement for T is typically relaxed using the well-known doubling trick. However, it may not work in this case because of Requirement 1: once T is doubled, the arm cannot start from Xt = 0 again. It remains an open question whether Algorithm 1 can be extended to the case of unknown T. Next we introduce the major steps of the proof of Theorem 2. The discretization error of pulling arms k/K, k {0, . . . , K}, is of the order O(T/K). Therefore, choosing K = O(T 1/4) controls the regret at the desired rate. The total number of pulls wasted in this stage is O(m K) = O(T 3/4). We consider the good event, on which f(k/K) is inside the confidence interval [LBk, UBk] for all k = 0, 1, . . . until the stopping criterion is triggered. We choose m such that for a given k this happens with a probability no less than 1 O(1/m). Therefore, applying the union bound to at most K discretized arms, the probability of the good event is no less than 1 O(K/m) = 1 O(T 1/4). As a result, the regret incurred due to the bad event is bounded by TO(T 1/4) = O(T 3/4), matching the desired rate. On the good event, the stopping criterion is triggered because arm k/K has passed arg maxx [0,1] f(x). But it is not much worse than the optimal arm because otherwise the stopping criterion would have been triggered earlier. The gap between f(k/K) and f(x ) is controlled by the width of the confidence interval, O( p log m/m). Multiplying it by T (because k/K is pulled for the rest of the horizon once the stopping criterion is triggered) gives O(T p log m/m) = O(T 3/4 log T). Putting the pieces together gives the regret bound in Theorem 2. 0 0.2 0.4 0.6 0.8 1 0 0 0.2 0.4 0.6 0.8 1 0 Figure 2: The constructed instances for the lower bound in continuum-armed bandit (left panel) and Theorem 3 (right panel). 4 Lower Bound for the Regret In the classic continuum-armed bandit problem, without Requirement 1, it is well known that there exist policies and algorithms that can achieve regret O(T 2/3), which has also been shown to be tight [26, 6, 38]. In this section, we show that in the presence of Requirement 1, no policies can achieve regret less than Ω(T 3/4) under Assumptions 1, 2 and 3. Theorem 3 (Lower Bound). When T 16, under Assumption 1 with c = 1, Assumptions 2 and 3, we have Rπ(T) 1 for any policy π satisfying Requirement 1. Compared to the literature, Theorem 3 implies that Requirement 1 substantially impedes the learning efficiency by increasing the rate of regret from T 2/3 to T 3/4. The proof for Theorem 3 is significantly different from that in the literature. In particular, a typical construction for the lower bound in the continuum-armed bandit literature involves a class of functions that have humps in a small interval and are zero elsewhere, as illustrated in the left panel of Figure 2. The idea is to make the functions indistinguishable except for two small intervals. In our case, it is not sufficient to achieve O(T 3/4). To utilize Requirement 1, the constructed functions in the right panel of Figure 2 share the same pattern before reaching the mode. Afterwards, the functions diverge. The objective is to make it hard for the decision maker to decide whether to keep escalating the arm or stop, and punish her severely if a wrong decision is made. (Consider stopping at the mode of the red function while the actual function is yellow.) Such regret is much larger than the hump functions and thus gives rise to the higher rate. Besides the construction of the instances, the proof relies on the following random time τ = max {t|Xt a} for some constant a. Note that τ is typically not a stopping time for a stochastic process Xt. Under Requirement 1, however, it is a stopping time. This is the key observation that helps us to fully utilize the monotonicity. Technically, this allows us to study the Kullback-Leibler divergence for the probability measures on (X1, Z1, . . . , Xτ, Zτ, τ). By decomposing the KL-divergence on the events {τ = 1}, {τ = 2}, and so on, we can focus on the periods when Xt a, which cannot possibly distinguish functions whose mode are larger than a (right panel of Figure 2). This observation and innovation is essential in the proof of Theorem 3. 5 Numerical Experiment In this section, we conduct numerical studies to compare the regret of Algorithm 1 with a few benchmarks. The experiments are conducted on a Windows 10 desktop with Intel i9-10900K CPU. To make a fair comparison, we generate the instances randomly so that it doesn t bias toward our algorithm. In particular, we consider T {1000, 11000, 21000, . . . , 101000}. For each T, we simulate 100 independent instances. For each instance, we generate x1 U(0, 1) and x2 U(0.5, 1), and let f(x) be the linear interpolation of the three points (0, 0), (x1, x2), and (1, 0). The reward Zt N(f(Xt), 0.1) for all instances and T.1 We consider the following algorithms for each instance: 1. Algorithm 1. As stated in Theorem 2, we use K = T 1/4 , m = T 1/2 and σ = 0.1. 2. The classic UCB algorithm for the continuum-armed bandit. We first discretize the interval [0, 1] into K + 1 equally spaced grid points, {k/K}K k=0, where K = T 1/3 according to [26]. Then we implement the version of UCB in Chapter 8 of [29] on the multi-armed bandit problem with K + 1 arms. More precisely, we have Xt = arg max k/K (UCBk(t)) ( 1 Tk(t 1) = 0 ˆµk(t 1) + σ q 2 log(1+t log(t)2) Tk(t 1) Tk(t 1) 1 where µk(t 1) is the empirical average of arm k, x = k/K, in the first t 1 periods; Tk(t 1) is the number that arm k is chosen in the first T 1 periods. Note that in this benchmark we do not impose Requirement 1. 3. UCB with Requirement 1. We follow the same discretization scheme and UCB definition in the last benchmark, except that we impose Xt+1 Xt. That is Xt = max{Xt 1, arg max k/K (UCBk(t))}. 4. Deflating UCB with Requirement 1. To improve the performance of UCB with Requirement 1, we need to make the algorithm more likely to stick to the current arm. Moreover, when exploring new arms, we need to encourage the algorithm to explore the arm slightly larger than the current arm. We design a heuristic for this purpose by deflating the UCB initialization of the arms. More precisely, we let UCBk(t) = 1 k/K, if Tk(t 1) = 0. In this way, the decision maker is encouraged to explore arms with small k first and gradually escalate the arms. The average regret over the 100 instances is shown in Figure 3. We may draw the following observations from the experiment. First, the regret of UCB is significantly less than the other three algorithms which are constrained by Requirement 1. Therefore, the cost brought to the learning efficiency by Requirement 1 is huge. Second, simply imposing Requirement 1 to the UCB algorithm doesn t perform well (the green curve). It more than triples the regret of Algorithm 1. Third, the deflating UCB heuristic seems to perform well. Its regret is on par with that of Algorithm 1, although the trend implies a fast scaling with T. The regret analysis and the optimal choice of the deflating factor remain an open question. Overall, Algorithm 1 performs the best among the algorithms that are constrained by Requirement 1. The result is consistent with the theoretical studies. 6 Conclusion In this study, we investigate the online learning or multi-armed bandit problem, when the chosen actions are ordinal and required to be monotone over time. The requirement appears naturally in several applications: firms setting markup pricing policies in order not to antagonize early adopters and clinical trials following dose escalation principles. It turns out that the requirement may severely limit the capability of the decision maker to learn the objective function: the worst-case regret is linear in the length of the learning horizon, when the objective function is just continuous. However, we show that if the objective function is also quasiconcave, then the proposed algorithm can achieve 1Here U(a, b) denotes a uniform random variable between in [a, b]; N(a, b) denotes a normal random variable with mean a and standard deviation b. 0.2 0.4 0.6 0.8 1 Algorithm 1 UCB UCB Requirement 1 Deflating UCB Figure 3: The average regret of Algorithm 1 and three benchmarks. sublinear regret O(T 3/4). It is still worse than the optimal rate in the continuum-armed bandit literature, but is proved to be optimal in this setting. The study has a few limitations and opens up potential research questions. First, since the doubling trick no longer works, the assumption of knowing T in advance cannot be simply relaxed. It is not clear if one can design a new algorithm that can achieve the optimal regret even when T is unknown. Second, it is unclear if the monotone requirement can be incorporated in contextual bandits. This is a highly relevant question in clinical trials, which increasingly rely on patient characteristics for the dose allocation. A novel framework is needed to study how the dose escalation principle plays out in this setting. [1] Alekh Agarwal, Dean P Foster, Daniel J Hsu, Sham M Kakade, and Alexander Rakhlin. Stochastic convex optimization with bandit feedback. In J. Shawe-Taylor, R. Zemel, P. Bartlett, F. Pereira, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 24. Curran Associates, Inc., 2011. [2] Rajeev Agrawal. The continuum-armed bandit problem. SIAM journal on control and optimization, 33(6):1926 1951, 1995. [3] Shipra Agrawal and Navin Goyal. Further optimal regret bounds for thompson sampling. In Artificial intelligence and statistics, pages 99 107. PMLR, 2013. [4] Peter Auer. Using confidence bounds for exploitation-exploration trade-offs. Journal of Machine Learning Research, 3(Nov):397 422, 2002. [5] Peter Auer, Nicolo Cesa-Bianchi, and Paul Fischer. Finite-time analysis of the multiarmed bandit problem. Machine learning, 47(2):235 256, 2002. [6] Peter Auer, Ronald Ortner, and Csaba Szepesvári. Improved rates for the stochastic continuumarmed bandit problem. In International Conference on Computational Learning Theory, pages 454 468. Springer, 2007. [7] Maryam Aziz, Emilie Kaufmann, and Marie-Karelle Riviere. On multi-armed bandit designs for dose-finding clinical trials. Journal of Machine Learning Research, 22:1 38, 2021. [8] Hamsa Bastani and Mohsen Bayati. Online decision making with high-dimensional covariates. Operations Research, 68(1):276 294, 2020. [9] Omar Besbes and Assaf Zeevi. Dynamic pricing without knowing the demand function: Risk bounds and near-optimal algorithms. Operations Research, 57(6):1407 1420, 2009. [10] Thomas M Braun. The bivariate continual reassessment method: extending the CRM to phase I trials of two competing outcomes. Controlled clinical trials, 23(3):240 256, 2002. [11] Sébastien Bubeck, Rémi Munos, Gilles Stoltz, and Csaba Szepesvári. X-armed bandits. Journal of Machine Learning Research, 12(5), 2011. [12] Ningyuan Chen and Guillermo Gallego. Nonparametric pricing analytics with customer covariates. Operations Research, Forthcoming. [13] Ningyuan Chen and Amin Khademi. Adaptive seamless dose-finding trials. Working Paper, 2020. [14] Ningyuan Chen, Chun Wang, and Longlin Wang. Learning and optimization with seasonal patterns. Working paper, 2020. [15] Qi Chen, Stefanus Jasin, and Izak Duenyas. Nonparametric self-adjusting control for joint learning and optimization of multiproduct pricing with finite resource capacity. Mathematics of Operations Research, 44(2):601 631, 2019. [16] Yiwei Chen and Vivek F Farias. Robust dynamic pricing with strategic customers. Mathematics of Operations Research, 43(4):1119 1142, 2018. [17] Wang Chi Cheung, David Simchi-Levi, and He Wang. Dynamic pricing and demand learning with limited price experimentation. Operations Research, 65(6):1722 1731, 2017. [18] Richard Combes and Alexandre Proutiere. Unimodal bandits: Regret lower bounds and optimal algorithms. In International Conference on Machine Learning, pages 521 529. PMLR, 2014. [19] Richard Combes, Alexandre Proutière, and Alexandre Fauquette. Unimodal bandits with continuous arms: Order-optimal regret without smoothness. Proceedings of the ACM on Measurement and Analysis of Computing Systems, 4(1):1 28, 2020. [20] Eric W Cope. Regret and convergence bounds for a class of continuum-armed bandit problems. IEEE Transactions on Automatic Control, 54(6):1243 1253, 2009. [21] Arnoud V Den Boer. Dynamic pricing and learning: historical origins, current research, and new directions. Surveys in operations research and management science, 20(1):1 18, 2015. [22] Aurélien Garivier, Pierre Ménard, Laurent Rossi, and Pierre Menard. Thresholding bandit for dose-ranging: The impact of monotonicity. Working paper, 2017. [23] Botao Hao, Yasin Abbasi-Yadkori, Zheng Wen, and Guang Cheng. Bootstrapping upper confidence bound. Working Paper, 2019. [24] Yichun Hu, Nathan Kallus, and Xiaojie Mao. Smooth contextual bandits: Bridging the parametric and non-differentiable regret regimes. In Conference on Learning Theory, pages 2007 2010. PMLR, 2020. [25] Su Jia, Andrew Li, and R Ravi. Markdown pricing under unknown demand. Working Paper, 2021. [26] Robert Kleinberg. Nearly tight bounds for the continuum-armed bandit problem. Advances in Neural Information Processing Systems, 17:697 704, 2004. [27] Robert Kleinberg, Aleksandrs Slivkins, and Eli Upfal. Multi-armed bandits in metric spaces. In Proceedings of the fortieth annual ACM symposium on Theory of computing, pages 681 690, 2008. [28] Branislav Kveton, Csaba Szepesvari, Sharan Vaswani, Zheng Wen, Tor Lattimore, and Mohammad Ghavamzadeh. Garbage in, reward out: Bootstrapping exploration in multi-armed bandits. In International Conference on Machine Learning, pages 3601 3610. PMLR, 2019. [29] Tor Lattimore and Csaba Szepesvári. Bandit algorithms. Cambridge University Press, 2020. [30] Christophe Le Tourneau, J Jack Lee, and Lillian L Siu. Dose escalation methods in phase I cancer clinical trials. JNCI: Journal of the National Cancer Institute, 101(10):708 720, 2009. [31] Lihong Li, Wei Chu, John Langford, and Robert E Schapire. A contextual-bandit approach to personalized news article recommendation. In Proceedings of the 19th international conference on World wide web, pages 661 670, 2010. [32] Qian Liu and Garrett J Van Ryzin. Strategic capacity rationing to induce early purchases. Management Science, 54(6):1115 1131, 2008. [33] Marie-Karelle Riviere, Ying Yuan, Jacques-Henri Jourdan, Frédéric Dubois, and Sarah Zohar. Phase I/II dose-finding design for molecularly targeted agent: plateau determination using adaptive randomization. Statistical methods in medical research, 27(2):466 479, 2018. [34] Daniel J. Russo, Benjamin Van Roy, Abbas Kazerouni, Ian Osband, and Zheng Wen. A tutorial on thompson sampling. Found. Trends Mach. Learn., 11(1):1 96, July 2018. [35] Jad Salem, Swati Gupta, and Vijay Kamble. Taming wild price fluctuations: Monotone stochastic convex optimization with bandit feedback. Working paper, 2021. [36] David Simchi-Levi and Yunzong Xu. Phase transitions and cyclic phenomena in bandits with switching constraints. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. [37] David Simchi-Levi, Yunzong Xu, and Jinglong Zhao. Blind network revenue management and bandits with knapsacks under limited switches. Working paper, 2019. [38] Aleksandrs Slivkins. Contextual bandits with similarity information. Journal of Machine Learning Research, 15(73):2533 2568, 2014. [39] Xuanming Su. Intertemporal pricing with strategic customer behavior. Management Science, 53(5):726 741, 2007. [40] Alexandre B Tsybakov. Introduction to nonparametric estimation. Springer Science & Business Media, 2008. [41] Sofía S Villar, Jack Bowden, and James Wason. Multi-armed bandit models for the optimal design of clinical trials: benefits and challenges. Statistical science: a review journal of the Institute of Mathematical Statistics, 30(2):199, 2015. [42] Nolan A Wages and Christopher Tait. Seamless phase i/ii adaptive design for oncology trials of molecularly targeted agents. Journal of biopharmaceutical statistics, 25(5):903 920, 2015. [43] Zizhuo Wang, Shiming Deng, and Yinyu Ye. Close the gaps: A learning-while-doing algorithm for single-product revenue management problems. Operations Research, 62(2):318 331, 2014. [44] Rui Yin, Yossi Aviv, Amit Pazgal, and Christopher S Tang. Optimal markdown pricing: Implications of inventory display formats in the presence of strategic customers. Management Science, 55(8):1391 1408, 2009. [45] Wei Zhang, Daniel J Sargent, and Sumithra Mandrekar. An adaptive dose-finding design incorporating both toxicity and efficacy. Statistics in Medicine, 25(14):2365 2383, 2006. [46] Xueru Zhang and Mingyan Liu. Fairness in learning-based sequential decision algorithms: A survey. Working paper, 2020. [47] Serhan Ziya, Hayriye Ayhan, and Robert D Foley. Relationships among three assumptions in revenue management. Operations Research, 52(5):804 809, 2004.