# improving_the_expected_improvement_algorithm__d8181261.pdf Improving the Expected Improvement Algorithm Chao Qin Columbia Business School New York, NY 10027 cqin22@gsb.columbia.edu Diego Klabjan Northwestern University Evanston, IL 60208 d-klabjan@northwestern.edu Daniel Russo Columbia Business School New York, NY 10027 djr2174@gsb.columbia.edu The expected improvement (EI) algorithm is a popular strategy for information collection in optimization under uncertainty. The algorithm is widely known to be too greedy, but nevertheless enjoys wide use due to its simplicity and ability to handle uncertainty and noise in a coherent decision theoretic framework. To provide rigorous insight into EI, we study its properties in a simple setting of Bayesian optimization where the domain consists of a finite grid of points. This is the so-called best-arm identification problem, where the goal is to allocate measurement effort wisely to confidently identify the best arm using a small number of measurements. In this framework, one can show formally that EI is far from optimal. To overcome this shortcoming, we introduce a simple modification of the expected improvement algorithm. Surprisingly, this simple change results in an algorithm that is asymptotically optimal for Gaussian best-arm identification problems, and provably outperforms standard EI by an order of magnitude. 1 Introduction Recently Bayesian optimization has received much attention in the machine learning community [21]. This literature studies the problem of maximizing an unknown black-box objective function by collecting noisy measurements of the function at carefully chosen sample points. At first a prior belief over the objective function is prescribed, and then the statistical model is refined sequentially as data are observed. Expected improvement (EI) [13] is one of the most widely-used Bayesian optimization algorithms. It is a greedy improvement-based heuristic that samples the point offering greatest expected improvement over the current best sampled point. EI is simple and readily implementable, and it offers reasonable performance in practice. Although EI is reasonably effective, it is too greedy, focusing nearly all sampling effort near the estimated optimum and gathering too little information about other regions in the domain. This phenomenon is most transparent in the simplest setting of Bayesian optimization where the function s domain is a finite grid of points. This is the problem of best-arm identification (BAI) [1] in a multiarmed bandit. The player sequentially selects arms to measure and observes noisy reward samples with the hope that a small number of measurements enable a confident identification of the best arm. Recently Ryzhov [20] studied the performance of EI in this setting. His work focuses on a link between EI and another algorithm known as the optimal computing budget allocation [3], but his analysis reveals EI allocates a vanishing proportion of samples to suboptimal arms as the total number of samples grows. Any method with this property will be far from optimal in BAI problems [1]. 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA. In this paper, we improve the EI algorithm dramatically through a simple modification. The resulting algorithm, which we call top-two expected improvement (TTEI), combines the top-two sampling idea of Russo [19] with a careful change to the improvement-measure used by EI. We show that this simple variant of EI achieves strong asymptotic optimality properties in the BAI problem, and benchmark the algorithm in simulation experiments. Our main theoretical contribution is a complete characterization of the asymptotic proportion of samples TTEI allocates to each arm as a function of the true (unknown) arm means. These particular sampling proportions have been shown to be optimal from several perspectives [4, 12, 9, 19, 8], and this enables us to establish two different optimality results for TTEI. The first concerns the rate at which the algorithm gains confidence about the identity of the optimal arm as the total number of samples collected grows. Next we study the so-called fixed confidence setting, where the algorithm is able to stop at any point and return an estimate of the optimal arm. We show that when applied with the stopping rule of Garivier and Kaufmann [8], TTEI essentially minimizes the expected number of samples required among all rules obeying a constraint on the probability of incorrect selection. One undesirable feature of our algorithm is its dependence on a tuning parameter. Our theoretical results precisely show the impact of this parameter, and reveal a surprising degree of robustness to its value. It is also easy to design methods that adapt this parameter over time to the optimal value, and we explore one such method in simulation. Still, removing this tuning parameter is an interesting direction for future research. Further related literature. Despite the popularity of EI, its theoretical properties are not well studied. A notable exception is the work of Bull [2], who studies a global optimization problem and provides a convergence rate for EI s expected loss. However, it is assumed that the observations are noiseless. Our work also relates to a large number of recent machine learning papers that try to characterize the sample complexity of the best-arm identification problem [5, 18, 1, 7, 14, 10, 11, 15 17]. Despite substantial progress, matching asymptotic upper and lower bounds remained elusive in this line of work. Building on older work in statistics [4, 12] and simulation optimization [9], recent work of Garivier and Kaufmann [8] and Russo [19] characterized the optimal sampling proportions. Two notions of asymptotic optimality are established: sample complexity in the fixed confidence setting and rate of posterior convergence. Garivier and Kaufmann [8] developed two sampling rules designed to closely track the asymptotic optimal proportions and showed that, when combined with a stopping rule motivated by Chernoff [4], this sampling rule minimizes the expected number of samples required to guarantee a vanishing threshold on the probability of incorrect selection is satisfied. Russo [19] independently proposed three simple Bayesian algorithms, and proved that each algorithm attains the optimal rate of posterior convergence. TTEI proposed in this paper is conceptually most similar to the top-two value sampling of Russo [19], but it is more computationally efficient. 1.1 Main Contributions As discussed below, our work makes both theoretical and algorithmic contributions. Theoretical: Our main theoretical contribution is Theorem 1, which establishes that TTEI a simple modification to a popular Bayesian heuristic converges to the known optimal asymptotic sampling proportions. It is worth emphasizing that, unlike recent results for other top-two sampling algorithms [19], this theorem establishes that the expected time to converge to the optimal proportions is finite, which we need to establish optimality in the fixed confidence setting. Proving this result required substantial technical innovations. Theorems 2 and 3 are additional theoretical contributions. These mirror results in [19] and [8], but we extract minimal conditions on sampling rules that are sufficient to guarantee the two notions of optimality studied in these papers. Algorithmic: On the algorithmic side, we substantially improve a widely used algorithm. TTEI can be easily implemented by modifying existing EI code, but, as shown in our experiments, can offer an order of magnitude improvement. A more subtle point involves the advantages of TTEI over algorithms that are designed to directly target convergence on the asymptotically optimal proportions. In the experiments, we show that TTEI substantially outperforms an oracle sampling rule whose sampling proportions directly track the asymptotically optimal proportions. This phenomenon should be explored further in future work, but suggests that by carefully reasoning about the value of information TTEI accounts for important factors that are washed out in asymptotic analysis. Finally as discussed in the conclusion although we focus on uncorrelated priors we believe our method can be easily extended to more complicated problems like that of best-arm identification in linear bandits [22]. 2 Problem Formulation Let A = {1, . . . , k} be the set of arms. At each time n N = {0, 1, 2, . . .}, an arm In A is measured, and an independent noisy reward Yn,In is observed. The reward Yn,i R of arm i at time n follows a normal distribution N(µi, σ2) with common known variance σ2, but unknown mean µi. The objective is to allocate measurement effort wisely in order to confidently identify the arm with highest mean using a small number of measurements. We assume that µ1 > µ2 > . . . > µk. Our analysis takes place in a frequentist setting, in which the true means (µ1, . . . , µk) are fixed but unknown. The algorithms we study, however, are Bayesian in the sense that they begin with prior over the arm means and update the belief to form a posterior distribution as evidence is gathered. Prior and Posterior Distributions. The sampling rules studied in this paper begin with a normally distributed prior over the true mean of each arm i A denoted by N(µ0,i, σ2 0,i), and update this to form a posterior distribution as observations are gathered. By conjugacy, the posterior distribution after observing the sequence (I0, Y0,I0, . . . , In 1, Yn 1,In 1) is also a normal distribution denoted by N(µn,i, σ2 n,i). The posterior mean and variance can be calculated using the following recursive equations: µn+1,i = (σ 2 n,iµn,i + σ 2Yn,i)/(σ 2 n,i + σ 2) if In = i, µn,i, if In = i, σ2 n+1,i = 1/(σ 2 n,i + σ 2) if In = i, σ2 n,i, if In = i. . We denote the posterior distribution over the vector of arm means by Πn = N(µn,1, σ2 n,1) N(µn,2, σ2 n,2) N(µn,k, σ2 n,k) and let θ = (θ1, . . . , θk). For example, with this notation The posterior probability assigned to the event that arm i is optimal is θi > max j =i θj To avoid confusion, we always use θ = (θ1, . . . , θk) to denote a random vector of arm means drawn from the algorithm s posterior Πn, and µ = (µ1, . . . , µk) to denote the vector of true arm means. Two notions of asymptotic optimality. Our first notion of optimality relates to the rate of posterior convergence. As the number of observations grows, one hopes that the posterior distribution definitively identifies the true best arm, in the sense that the posterior probability 1 αn,1 assigned by the event that a different arm is optimal tends to zero. By sampling the arms intelligently, we hope this probability can be driven to zero as rapidly as possible. Following Russo [19], we aim to maximize the exponent governing the rate of decay, lim inf n 1 n log (1 αn,1) , among all sampling rules. The second setting we consider is often called the fixed confidence setting. Here, the agent is allowed at any point to stop gathering samples and return an estimate of the identity of the optimal. In addition to a sampling rule, we require a stopping rule that selects a time τ at which to stop, and a decision rule that returns an estimate ˆIτ of the optimal arm based on the first τ observations. We consider minimizing the average number of observations E[τδ] required by an algorithm (that consists of a sampling rule, a stopping rule and a decision rule) guaranteeing a vanishing probability δ of incorrect identification, i.e., P(ˆIτδ = 1) δ. Following Garivier and Kaufmann [8], the number of samples required scales with log(1/δ), and so we aim to minimize lim sup δ 0 E[τδ] log(1/δ) among all algorithms with probability of error no more than δ. In this setting, we study the performance of sampling rules when combined with the stopping rule studied by Chernoff [4] and Garivier and Kaufmann [8]. 3 Sampling Rules In this section, we first introduce the expected improvement algorithm, and point out its weakness. Then a simple variant of the expected improvement algorithm is proposed. Both algorithms make calculations using function f(x) = xΦ(x) + φ(x) where Φ( ) and φ( ) are the CDF and PDF of the standard normal distribution. One can show that as x , log f( x) x2/2, and so f( x) e x2/2 for very large x. One can also show that f is an increasing function. Expected Improvement. Expected improvement [13] is a simple improvement-based sampling rule. The EI algorithm favors the arm that offers the largest amount of improvement upon a target. The EI algorithm measures the arm In = arg maxi A vn,i where vn,i is the EI value of arm i at time n. Let I n = arg maxi A µn,i denote the arm with largest posterior mean at time n. The EI value of arm i at time n is defined as vn,i Eθ Πn h θi µn,I n +i . where x+ = max{x, 0}. The above expectation can be computed analytically as follows, vn,i = µn,i µn,I n Φ µn,i µn,I n σn,i + σn,iφ µn,i µn,I n σn,i = σn,if µn,i µn,I n σn,i The EI value vn,i measures the potential of arm i to improve upon the largest posterior mean µn,I n at time n. Because f is an increasing function, vn,i is increasing in both the posterior mean µn,i and posterior standard deviation σn,i. Top-Two Expected Improvement. The EI algorithm can have very poor performance for selecting the best arm. Once the posterior indicates a particular arm is the best with reasonably high probability, EI allocates nearly all future samples to this arm at the expense of measuring other arms. Recently Ryzhov [20] showed that EI only allocates O(log n) samples to suboptimal arms asymptotically. This is a severe shortcoming, as it means n must be extremely large before the algorithm has enough samples from suboptimal arms to reach a confident conclusion. To improve the EI algorithm, we build on the top-two sampling idea in Russo [19]. The idea is to identify in each period the two most promising arms based on current observations, and randomize to choose which to sample. A tuning parameter β (0, 1) controls the probability assigned to the top arm. A naive top-two variant of EI would identify the two arms with largest EI value, and flip a β weighted coin to decide which to measure. However, one can prove that this algorithm is not optimal for any choice of β. Instead, what we call the top-two expected improvement algorithm uses a novel modified EI criterion which more carefully accounts for the decision-maker s uncertainty when deciding which arm to sample. For i, j A, define vn,i,j Eθ Πn [(θi θj)+]. This measures the expected magnitude of improvement arm i offers over arm j, but unlike the typical EI criterion, this expectation integrates over the uncertain quality of both arms. This measure can be computed analytically as σ2 n,i + σ2 n,jf µn,i µn,j q σ2 n,i + σ2 n,j TTEI depends on a tuning parameter β > 0, set to 1/2 by default. With probability β, TTEI measures the arm I(1) n by optimizing the EI criterion, and otherwise it measures an alternative I(2) n that offers the largest expected improvement on the arm I(1) n . Formally, TTEI measures the arm ( I(1) n = arg maxi A vn,i, with probability β, I(2) n = arg maxi A vn,i,I(1) n , with probability 1 β. Note that vn,i,i = 0, which implies I(2) n = I(1) n . We notice that TTEI with β = 1 is the standard EI algorithm. Comparing to the EI algorithm, TTEI with β (0, 1) allocates much more measurement effort to suboptimal arms. We will see that TTEI allocates β proportion of samples to the best arm asymptotically, and it uses the remaining 1 β fraction of samples for gathering evidence against each suboptimal arm. 4 Convergence to Asymptotically Optimal Proportions For all i A and n N, we define Tn,i Pn 1 ℓ=0 1{Iℓ= i} to be the number of samples of arm i before time n. We will show that under TTEI with parameter β, limn Tn,1/n = β. That is, the algorithm asymptotically allocates β proportion of the samples to true best arm. Dropping for the moment questions regarding the impact of this tuning parameter, let us consider the optimal asymptotic proportion of effort to allocate to each of the k 1 remaining arms. It is known that the optimal proportions are given by the unique vector (wβ 2 , , wβ k) satisfying Pk i=2 wβ i = 1 β and 1/β + 1/wβ 2 = . . . = (µ1 µk)2 1/β + 1/wβ k . (2) We set wβ 1 = β, so wβ = wβ 1 , . . . , wβ k encodes the sampling proportions of each arm. To understand the source of equation (2), imagine that over the first n periods each arm i is sampled exactly wβ i n times, and let ˆµn,i N µi, σ2 denote the empirical mean of arm i. Then ˆµn,1 ˆµn,i N µ1 µi, σ2 i where σ2 i = σ2 The probability ˆµn,1 ˆµn,i 0 leading to an incorrect estimate of which arm has highest mean is Φ ((µi µ1)/ σi) where Φ is the CDF of the standard normal distribution. Equation (2) is equivalent to requiring (µ1 µi)/ σi is equal for all arms i, so the probability of falsely declaring µi µ1 is equal for all i = 1. In a sense, these sampling frequencies equalize the evidence against each suboptimal arm. These proportions appeared first in the machine learning literature in [19, 8], but appeared much earlier in the statistics literature in [12], and separately in the simulation optimization literature in [9]. As we will see in the next section, convergence to this allocation is a necessary condition for both notions of optimality considered in this paper. Our main theoretical contribution is the following theorem, which establishes that under TTEI sampling proportions converge to the proportions wβ derived above. Therefore, while the sampling proportion of the optimal arm is controlled by the tuning parameter β, the remaining 1 β fraction of measurement is optimally distributed among the remaining k 1 arms. Such a result was established for other top-two sampling algorithms in [19]. The second notion of optimality requires not just convergence to wβ with probability 1, but also a sense in which the expected time until convergence is finite. The following theorem presents such a stronger result for TTEI. To make this precise, we introduce a time after which for each arm, the empirical proportion allocated to it is accurate. Specifically, given β (0, 1) and ϵ > 0, we define M ϵ β inf N N : max i A |Tn,i/n wβ i | ϵ n N . (3) It is clear that P(M ϵ β < ) = 1 for all ϵ > 0 if and only if Tn,i/n wβ i with probability 1 for each arm i A. To establish optimality in the fixed confidence setting , we need to prove in addition that E[M ϵ β] < for all ϵ > 0, which requires substantial new technical innovations. Theorem 1. Under TTEI with parameter β (0, 1), E[M ϵ β] < for any ϵ > 0. This result implies that under TTEI, P(M ϵ β < ) = 1 for all ϵ > 0, or equivalently n = wβ i i A. 4.1 Problem Complexity Measure Given β (0, 1), define the problem complexity measure Γ β (µ1 µ2)2 2σ2 1/β + 1/wβ 2 = . . . = (µ1 µk)2 2σ2 1/β + 1/wβ k , which is a function of the true arm means and variances. This will be the exponent governing the rate of posterior convergence, and also characterizing the average number of samples in the fixed confidence stetting. The optimal exponent comes from maximizing over β. Let us define Γ = maxβ (0,1) Γ β and β = arg maxβ (0,1) Γ β and set w = wβ = β , wβ 2 , . . . , wβ Russo [19] has proved that for β (0, 1), Γ β Γ / max n β 1 β o , and therefore Γ 1/2 Γ /2. This demonstrates a surprising degree of robustness to β. In particular, Γβ is close to Γ if β is adjusted to be close to β , and the choice of β = 1/2 always yields a 2-approximation to Γ . 5 Implied Optimality Results This section establishes formal optimality guarantees for TTEI. Both results, in fact, hold for any algorithm satisfying the conclusions of Theorem 1, and are therefore of broader interest. 5.1 Optimal Rate of Posterior Convergence We first provide upper and lower bounds on the exponent governing the rate of posterior convergence. The same result has been has been proved in Russo [19] for bounded correlated priors. We use different proof techniques to prove the following result for uncorrelated Gaussian priors. This theorem shows that no algorithm can attain a rate of posterior convergence faster than e Γ n and that this is attained by any algorithm that, like TTEI with optimal tuning parameter β , has asymptotic sampling ratios (w 1, . . . , w k). The second part implies TTEI with parameter β attains convergence rate e nΓ β and that it is optimal among sampling rules that allocation β fraction of samples to the optimal arm. Recall that, without loss of generality, we have assumed arm 1 is the arm with true highest mean µ1 = maxi A µi. We will study the posterior mass 1 αn,1 assigned to the event that some other has the highest mean. Theorem 2 (Posterior Convergence - Sufficient Condition for Optimality). The following properties hold with probability 1: 1. Under any sampling rule that satisfies Tn,i/n w i for each i A, n log (1 αn,1) = Γ . Under any sampling rule, lim sup n 1 n log(1 αn,1) Γ . 2. Let β (0, 1). Under any sampling rule that satisfies Tn,i/n wβ i for each i A, n log(1 αn,1) = Γ β. Under any sampling rule that satisfies Tn,1/n β, lim sup n 1 n log(1 αn,1) Γ β. This result reveals that when the tuning parameter β is set optimally to β , TTEI attains the optimal rate of posterior convergence. Since Γ 1/2 Γ /2, when β is set to the default value 1/2, the exponent governing the convergence rate of TTEI is at least half of the optimal one. 5.2 Optimal Average Sample Size Chernoff s Stopping Rule. In the fixed confidence setting, besides an efficient sampling rule, a player also needs to design an intelligent stopping rule. This section introduces a stopping rule proposed by Chernoff [4] and studied recently by Garivier and Kaufmann [8]. This stopping rule makes use of the Generalized Likelihood Ratio statistic, which depends on the current maximum likelihood estimates of all unknown means. For each arm i A, the maximum likelihood estimate of its unknown mean µi at time n is its empirical mean ˆµn,i = T 1 n,i Pn 1 ℓ=0 1{Iℓ= i}Yℓ,Iℓwhere Tn,i = Pn 1 ℓ=0 1{Iℓ= i}. Next we define a weighted average of empirical means of arms i, j A: ˆµn,i,j Tn,i Tn,i + Tn,j ˆµn,i + Tn,j Tn,i + Tn,j ˆµn,j. Then if ˆµn,i ˆµn,j, the Generalized Likelihood Ratio statistic Zn,i,j has the following explicit expression: Zn,i,j Tn,id(ˆµn,i, ˆµn,i,j) + Tn,jd(ˆµn,j, ˆµn,i,j) where d(x, y) = (x y)2/(2σ2) is the Kullback-Leibler (KL) divergence between Gaussian distributions N(x, σ2) and N(y, σ2). Similarly, if ˆµn,i < ˆµn,j, Zn,i,j = Zn,j,i 0 where Zn,j,i is well defined as above. If either arm has never been sampled before, these quantities are not well defined and we take the convention that Zn,i,j = Zn,j,i = 0. Given a target confidence δ (0, 1), to ensure that one arm is better than the others with probability at least 1 δ, we use the stopping time τδ inf n N : Zn max i A min j A\{i} Zn,i,j > γn,δ where γn,δ > 0 is an appropriate threshold. By definition, minj A\{i} Zn,i,j is nonnegative if and only if ˆµn,i ˆµn,j for all j A \ {i}. Hence, whenever ˆI n arg maxi A ˆµn,i is unique, Zn = minj A\{ˆI n} Zn,ˆI n,j. Next we introduce the exploration rate for normal bandit models that can ensure to identify the best arm with probability at least 1 δ. We use the following result given in Garivier and Kaufmann [8]. Proposition 1 (Garivier and Kaufmann [8] Proposition 12). Let δ (0, 1) and α > 1. There exists a constant C = C(α, k) such that under any sampling rule, using the Chernoff s stopping rule with the threshold γα n,δ = log(Cnα/δ) guarantees P τδ < , arg max i A ˆµτδ,i = 1 δ. Sample Complexity. Garivier and Kaufmann [8] recently provided a general lower bound on the number of samples required in the fixed confidence setting. In particular, they show that for any normal bandit model, under any sampling rule and stopping time τδ that guarantees a probability of error no more than δ, lim inf δ 0 E[τδ] log(1/δ) 1 Recall that M ϵ β, defined in (3), is the first time after which the empirical proportions are within ϵ of their asymptotic limits. The next result provides a condition in terms of M ϵ β that is sufficient to guarantee optimality in the fixed confidence setting. Theorem 3 (Fixed Confidence - Sufficient Condition for Optimality). Let δ, β (0, 1) and α > 1. Under any sampling rule which, if applied with no stopping rule, satisfies E[M ϵ β] < for all ϵ > 0, using the Chernoff s stopping rule with the threshold γα n,δ = log(Cnα/δ) (where C = C(α, k)) guarantees lim sup δ 0 E[τδ] log(1/δ) 1 When β = β the general lower bound on sample complexity of 1/Γ is essentially matched. In addition, when β is set to the default value 1/2, the sample complexity of TTEI combined with the Chernoff s stopping rule is at most twice the optimal sample complexity since 1/Γ 1/2 2/Γ . 6 Numerical Experiments To test the empirical performance of TTEI, we conduct several numerical experiments. The first experiment compares the performance of TTEI with β = 1/2 and EI. The second experiment compares the performance of different versions of TTEI, top-two Thompson sampling (TTTS) [19], knowledge gradient (KG) [6] and oracle algorithms that know the optimal proportions a priori. Each algorithm plays arm i = 1, . . . , k exactly once at the beginning, and then prescribe a prior N(Yi,i, σ2) for unknown arm-mean µi where Yi,i is the observation from N(µi, σ2). In both experiments, we fix the common known variance σ2 = 1 and the number of arms k = 5. We consider three instances [µ1, . . . , µ5] = [5, 4, 1, 1, 1], [5, 4, 3, 2, 1] and [2, 0.8, 0.6, 0.4, 0.2]. The optimal parameter β equals 0.48, 0.45 and 0.35, respectively. Recall that αn,i, defined in (1), denotes the posterior probability that arm i is optimal. Tables 1 and 2 show the average number of measurements required for the largest posterior probability assigned to some arm being the best to reach a given confidence level c, i.e., maxi αn,i c. In a Bayesian setting, the probability of correct selection under this rule is exactly c. The results in Table 1 are averaged over 100 trials. We see that TTEI with β = 1/2 outperforms standard EI by an order of magnitude. Table 1: Average number of measurements required to reach the confidence level c = 0.95 TTEI-1/2 EI [5, 4, 1, 1, 1] 14.60 238.50 [5, 4, 3, 2, 1] 16.72 384.73 [2, .8, .6, .4, .2] 24.39 1525.42 The second experiment compares the performance of different versions of TTEI, TTTS, KG, a random sampling oracle (RSO) and a tracking oracle (TO). The random sampling oracle draws a random arm in each round from the distribution w encoding the asymptotically optimal proportions. The tracking oracle tracks the optimal proportions at each round. Specifically, the tracking oracle samples the arm with the largest ratio its optimal and empirical proportions. Two tracking algorithms proposed by Garivier and Kaufmann [8] are similar to this tracking oracle. TTEI with adaptive β (a TTEI) works as follows: it starts with β = 1/2 and updates β = ˆβ every 10 rounds where ˆβ is the maximizer of equation (2) based on plug-in estimators for the unknown arm-means. Table 2 shows the average number of measurements required for the largest posterior probability being the best to reach the confidence level c = 0.9999. The results in Table 2 are averaged over 200 trials. We see that the performances of TTEI with adaptive β and TTEI with β are better than the performances of all other algorithms. We note that TTEI with adaptive β substantially outperforms the tracking oracle. Table 2: Average number of measurements required to reach the confidence level c = 0.9999 TTEI-1/2 a TTEI TTEI-β TTTS-β RSO TO KG [5, 4, 1, 1, 1] 61.97 61.98 61.59 62.86 97.04 77.76 75.55 [5, 4, 3, 2, 1] 66.56 65.54 65.55 66.53 103.43 88.02 81.49 [2, .8, .6, .4, .2] 76.21 72.94 71.62 73.02 101.97 96.90 86.98 In addition to the Bayesian stopping rule tested above, we have run some experiments with the Chernoff stopping rule discussed in Section 5.2. Asymptotic analysis shows these two rules are similar when the confidence level c is very high. However, the Chernoff stopping rule appears to be too conservative in practice; it typically yields a probability of correct selection much larger than the specified confidence level c at the expense of using more samples. Since our current focus is on allocation rules, we focus on this Bayesian stopping rule, which appears to offer a more fundamental comparison than one based on ad hoc choice of tuning parameters. Developing improved stopping rules is an important area for future research. 7 Conclusion and Extensions to Correlated Arms We conclude by noting that while this paper thoroughly studies TTEI in the case of uncorrelated priors, we believe the algorithm is also ideally suited to problems with complex correlated priors and large sets of arms. In fact, the modified information measure vn,i,j was designed with an eye toward dealing with correlation in a sophisticated way. In the case of a correlated normal distribution N(µ, Σ), one has vn,i,j = Eθ N(µ,Σ)[(θi θj)+] = p Σii + Σjj 2Σijf µn,i µn,j p Σii + Σjj 2Σij This closed form accommodates efficient computation. Here the term Σi,j accounts for the correlation or similarity between arms i and j. Therefore vn,i,I(1) n is large for arms i that offer large potential improvement over I(1) n , i.e. those that (1) have large posterior mean, (2) have large posterior variance, and (3) are not highly correlated with arm I(1) n . As I(1) n concentrates near the estimated optimum, we expect the third factor will force the algorithm to experiment in promising regions of the domain that are far away from the current-estimated optimum, and are under-explored under standard EI. [1] Jean-Yves Audibert, Sébastien Bubeck, and Rémi Munos. Best arm identification in multiarmed bandits. In COLT 2010 - The 23rd Conference on Learning Theory, Haifa, Israel, June 27-29, 2010, pages 41 53, 2010. [2] Adam D. Bull. Convergence rates of efficient global optimization algorithms. Journal of Machine Learning Research, 12:2879 2904, 2011. URL http://dblp.uni-trier.de/db/ journals/jmlr/jmlr12.html#Bull11. [3] Chun-Hung Chen, Jianwu Lin, Enver Yücesan, and Stephen E Chick. Simulation budget allocation for further enhancing the efficiency of ordinal optimization. Discrete Event Dynamic Systems, 10(3):251 270, 2000. [4] Herman Chernoff. Sequential design of experiments. Ann. Math. Statist., 30(3):755 770, 09 1959. doi: 10.1214/aoms/1177706205. URL http://dx.doi.org/10.1214/aoms/ 1177706205. [5] Eyal Even-dar, Shie Mannor, and Yishay Mansour. Pac bounds for multi-armed bandit and markov decision processes. In In Fifteenth Annual Conference on Computational Learning Theory (COLT), pages 255 270, 2002. [6] Peter I Frazier, Warren B Powell, and Savas Dayanik. A knowledge-gradient policy for sequential information collection. SIAM Journal on Control and Optimization, 47(5):2410 2439, 2008. [7] Victor Gabillon, Mohammad Ghavamzadeh, and Alessandro Lazaric. Best arm identification: A unified approach to fixed budget and fixed confidence. In F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 25, pages 3212 3220. Curran Associates, Inc., 2012. [8] Aurélien Garivier and Emilie Kaufmann. Optimal best arm identification with fixed confidence. In Proceedings of the 29th Conference on Learning Theory, COLT 2016, New York, USA, June 23-26, 2016, pages 998 1027, 2016. [9] P. Glynn and S. Juneja. A large deviations perspective on ordinal optimization. In Simulation Conference, 2004. Proceedings of the 2004 Winter, volume 1. IEEE, 2004. [10] Kevin Jamieson, Matthew Malloy, Robert Nowak, and Sébastien Bubeck. lil ucb : An optimal exploration algorithm for multi-armed bandits. In Maria Florina Balcan, Vitaly Feldman, and Csaba Szepesvári, editors, Proceedings of The 27th Conference on Learning Theory, volume 35 of Proceedings of Machine Learning Research, pages 423 439, Barcelona, Spain, 13 15 Jun 2014. PMLR. URL http://proceedings.mlr.press/v35/jamieson14.html. [11] Kevin G. Jamieson and Robert D. Nowak. Best-arm identification algorithms for multi-armed bandits in the fixed confidence setting. In 48th Annual Conference on Information Sciences and Systems, CISS 2014, Princeton, NJ, USA, March 19-21, 2014, pages 1 6, 2014. [12] C. Jennison, I. M. Johnstone, and B. W. Turnbull. Asymptotically optimal procedures for sequential adaptive selection of the best of several normal means. Statistical decision theory and related topics III, 2:55 86, 1982. [13] Donald R. Jones, Matthias Schonlau, and William J. Welch. Efficient global optimization of expensive black-box functions. Journal of Global Optimization, 13(4):455 492, 1998. ISSN 1573-2916. doi: 10.1023/A:1008306431147. URL http://dx.doi.org/10.1023/A: 1008306431147. [14] Zohar Karnin, Tomer Koren, and Oren Somekh. Almost optimal exploration in multi-armed bandits. In Sanjoy Dasgupta and David Mc Allester, editors, Proceedings of the 30th International Conference on Machine Learning, volume 28 of Proceedings of Machine Learning Research, pages 1238 1246, Atlanta, Georgia, USA, 17 19 Jun 2013. PMLR. URL http://proceedings.mlr.press/v28/karnin13.html. [15] Emilie Kaufmann and Shivaram Kalyanakrishnan. Information complexity in bandit subset selection. In Shai Shalev-Shwartz and Ingo Steinwart, editors, Proceedings of the 26th Annual Conference on Learning Theory, volume 30 of Proceedings of Machine Learning Research, pages 228 251, Princeton, NJ, USA, 12 14 Jun 2013. PMLR. URL http://proceedings. mlr.press/v30/Kaufmann13.html. [16] Emilie Kaufmann, Olivier Cappé, and Aurélien Garivier. On the complexity of a/b testing. In Maria Florina Balcan, Vitaly Feldman, and Csaba Szepesvári, editors, Proceedings of The 27th Conference on Learning Theory, volume 35 of Proceedings of Machine Learning Research, pages 461 481, Barcelona, Spain, 13 15 Jun 2014. PMLR. URL http://proceedings.mlr. press/v35/kaufmann14.html. [17] Emilie Kaufmann, Olivier Cappé, and Aurélien Garivier. On the complexity of best-arm identification in multi-armed bandit models. Journal of Machine Learning Research, 17(1): 1 42, 2016. URL http://jmlr.org/papers/v17/kaufman16a.html. [18] Shie Mannor, John N. Tsitsiklis, Kristin Bennett, and Nicolò Cesa-bianchi. The sample complexity of exploration in the multi-armed bandit problem. Journal of Machine Learning Research, 5:2004, 2004. [19] Daniel Russo. Simple bayesian algorithms for best arm identification. In 29th Annual Conference on Learning Theory, pages 1417 1418, 2016. [20] Ilya O. Ryzhov. On the convergence rates of expected improvement methods. Operations Research, 64(6):1515 1528, 2016. doi: 10.1287/opre.2016.1494. URL http://dx.doi.org/ 10.1287/opre.2016.1494. [21] Bobak Shahriari, Kevin Swersky, Ziyu Wang, Ryan P. Adams, and Nando de Freitas. Taking the human out of the loop: A review of Bayesian optimization. Proceedings of the IEEE, 104(1): 148 175, 2016. doi: 10.1109/JPROC.2015.2494218. URL http://dx.doi.org/10.1109/ JPROC.2015.2494218. [22] Marta Soare, Alessandro Lazaric, and Rémi Munos. Best-arm identification in linear bandits. In Advances in Neural Information Processing Systems, pages 828 836, 2014.