# bootstrapping_upper_confidence_bound__b45073ac.pdf Bootstrapping Upper Confidence Bound Botao Hao Purdue University haobotao000@gmail.com Yasin Abbasi-Yadkori Vin AI yasin.abbasi@gmail.com Zheng Wen Deepmind zhengwen@google.com Guang Cheng Purdue University chengg@purdue.edu Upper Confidence Bound (UCB) method is arguably the most celebrated one used in online decision making with partial information feedback. Existing techniques for constructing confidence bounds are typically built upon various concentration inequalities, which thus lead to over-exploration. In this paper, we propose a nonparametric and data-dependent UCB algorithm based on the multiplier bootstrap. To improve its finite sample performance, we further incorporate second-order correction into the above construction. In theory, we derive both problem-dependent and problem-independent regret bounds for multi-armed bandits with symmetric rewards under a much weaker tail assumption than the standard sub-Gaussianity. Numerical results demonstrate significant regret reductions by our method, in comparison with several baselines in a range of multi-armed and linear bandit problems. 1 Introduction In artificial intelligence, learning to make decisions online plays a critical role in many fields, such as personalized news recommendation [1], robotics [2] and the game of Go [3]. To learn to make optimal decisions as soon as possible, the decision-makers must carefully design an algorithm to balance the trade-off between the exploration and exploitation [4, 5]. Over-exploration could be expensive and unethical in practice, e.g., medical decision making [6, 7, 8]. On the other hand, insufficient exploration tends to make an algorithm stuck at a sub-optimal solution. The delicate design of exploration methods stands in the heart of online learning and decision making. Upper Confidence Bound (UCB) [9, 10, 11, 12, 13] is a class of highly effective algorithms in dealing with the exploration-exploitation trade-off in bandits and reinforcement learning. The tightness of confidence bound, as is known, is the key ingredient to achieve the optimal degree of explorations. To the best of our knowledge, nearly all the existing works construct confidence bounds based on various concentration inequalities, e.g. Hoeffding-type [10], empirical Bernstein type [14] or self-normalized type [13]. Those concentration-based confidence bounds, however, are typically conservative since they are data-independent. Concentration inequalities only exploit tail information, e.g., bounded or sub-Gaussian, rather than the whole distribution knowledge. In general, the loose constant factor may result in confidence bounds that are too wide to be informative [15]. In this paper, we propose a non-parametric and data-dependent UCB algorithm based on the multiplier bootstrap [16, 17, 18, 19, 20], called bootstrapped UCB. The principle is to use the multiplier 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. bootstrapped quantile as the confidence bound to enforce the exploration. Inspired by recent advances on non-asymptotic guarantee and non-asymptotic inference such as [18, 19, 20, 21], we develop an explicit second-order correction for the multiplier bootstrapped quantile that ensures the nonasymptotic validity. Our algorithm is easy to implement and has the potential to be generalized to more complicated models such as structured contextual bandits. In theory, we develop both problem-dependent and problem-independent regret bounds for multiarmed bandits with symmetric rewards under a much weaker tail assumption, i.e., sub-Weibull distribution, than the classical sub-Gaussianity. In this case, it is proven that the mean estimator can still achieve the same problem-independent regret bound as the one under the sub-Gaussian assumption. Note that our result does not rely on other sophisticated approaches such as medianof-means or Catoni s M-estimator in [22]. A key technical tool we propose is a new concentration inequality for the sum of sub-Weibull random variables. Empirically, we evaluate our method in several multi-armed and linear bandit models. When the exact posterior is unavailable or the noise variance is mis-specified, the bootstrapped UCB demonstrates superior performance over variants of Thompson sampling and concentration-based UCB due to its non-parametric and data-dependent nature. Recently, an increasing number of works [23, 24, 25, 26] study bootstrap methods for multi-armed and contextual bandits as an alternative to Thompson sampling. Most treat the bootstrap just as a way to randomize historical data (without any theoretical guarantee). One exception is [27] who derive a regret bound for Bernoulli bandit by adding pseudo observations. However, their method cannot be easily extended to unbounded cases, and their analyses heavily limit to the Bernoulli assumption. In contrast, our method applies to a broader class of bandit models with rigorous regret analysis. The rest of the paper is organized as follows. Section 2 introduces the basic setup and our bootstrapped UCB algorithm. Section 3 provides the regret analysis and Section 4 conducts several experiments. Notations. Throughout the paper, we denote Pw( ), Ew( ) as the probability and expectation operator with respect to the distribution of the vector w only, conditioning on other random variables. We use similar notations for Py( ), Ey( ) with respect to y only. [n] means the set {1, 2, . . . , n}. We denote boldface lower letters (e.g. x, y) as a vector. For a set E, we define its complement as Ec. 2 Bootstrapped UCB Problem setup. As a fruit fly, we illustrate our idea on the stochastic multi-armed bandit problem [28, 5]. In detail, the decision-makers interact with an environment for T rounds. In round t [T], the decision-makers pull an arm It [K] and observes its reward y It which is drawn from a distribution associated with the arm It, denoted by PIt with an unknown mean µIt. Without loss of generality, we assume arm 1 is the optimal arm, that is, µ1 = maxk [K] µk. In multi-armed bandit problems, the objective is to minimize the expected cumulative regret, defined as, R(T) = Tµ1 E h T X k=2 k E h T X t=1 I{It = k} i , (2.1) where k = µ1 µk is the sub-optimality gap for arm k, and I{ } is an indicator function. Here, the second equality is from the regret decomposition Lemma (Lemma 4.5 in [5]). We call an upper bound of R(T) problem-independent if the bound only depends on the distributional assumption and not on the specific bandit problem, say the gap k. Upper Confidence Bound. The upper confidence bound (UCB) algorithm [10] is based on the principle of optimism in the face of uncertainty. The key idea is to act as if the environment (parameterized by µk in multi-armed bandits) is as nice as plausibly possible. Concretely, a plausible environment refers to an upper confidence bound G(yn, 1 α) for the true mean µ, of the form G(yn, 1 α) = x R, x yn hα(yn) , (2.2) where yn = (y1, . . . , yn) is the sample vector, yn is the empirical mean, α (0, 1) is the confidence level, and hα : Rn R+ is a threshold that could be either data-dependent or data-independent. Definition 2.1. We define G(yn, 1 α) as a non-asymptotic upper confidence bound if for any sample size n 1, the following inequality holds P µ G(yn, 1 α) 1 α. (2.3) In bandit problems, a non-asymptotic control on the confidence level is more commonly used. This is rather different from the asymptotic validity of confidence bound in statistics literature [29]. A generic UCB algorithm will select the action based on its UCB index yn + hα(yn) for different arms. As is well known, the sharper the threshold is, the better exploration and exploitation trade-off one can achieve [5]. By the definition of quantile, the sharpest threshold in (2.2) is the (1 α)-quantile of the distribution of yn µ. However, this quantile relies on the knowledge of the exact reward distribution and is therefore itself unknown. To evaluate this value, we construct a data-dependent confidence bound based on the multiplier bootstrap. 2.1 Confidence Bound Based on Multiplier Bootstrap Multiplier Bootstrap. Multiplier bootstrap is a fast and easy-to-implement alternative to the standard bootstrap, and has been successfully applied in various statistical contexts [18, 19, 20]. Its goal is to approximate the distribution of the target statistic by reweighing its summands with random multipliers independent of the data. For instance, in a mean estimation problem, we define a multiplier bootstrapped estimator as n 1 Pn i=1 wi(yi yn) = n 1 Pn i=1(wi wn)yi, where {wi}n i=1 are some random variables independent of yn, called bootstrap weights. Some classical weights are as follows: Efron s bootstrap weights. (w1, . . . , wn) is a multinomial random vector with parameters (n; n 1, . . . , n 1). This is the standard nonparameteric bootstrap [30]. Gaussian weights. wi s are i.i.d standard Gaussian random variables. This is closely related to Gaussian approximation in statistics [19]. Rademacher weights. wi s are i.i.d Rademacher variables. This is closely related to symmetrization in learning theory. The bootstrap principle suggests that the (1 α)-quantile of the distribution of n 1 Pn i=1 wi(yi yn) conditionally on yn could be used to approximate the (1 α)-quantile of the distribution of yn µ. As the first building block, the multiplier bootstrapped quantile is defined as, qα(yn yn) := inf n x R|Pw 1 i=1 wi(yi yn) > x α o . (2.4) The question is whether qα(yn yn) is a valid threshold for any sample size n 1. 2.2 Second-order Correction Most statistical theories guarantee the asymptotic validity of qα(yn yn) by the multiplier central limit theorem [31]. However, we show that such a claim is valid non-asymptotically at the cost of adding a second-order correction. Next theorem rigorously characterizes this phenomenon under a symmetric assumption on the reward. Moreover, in Section A in the supplement, we show that without the second-order correction, a naive bootstrapped UCB will result in linear regret. Theorem 2.2 (Non-asymptotic Second-order Correction). Suppose {yi}n i=1 are i.i.d symmetric random variables with respect to its mean µ, and the bootstrap weights {wi}n i=1 are i.i.d Rademacher random variables. For two arbitrary parameters α, δ (0, 1), the following inequality holds for any sample size n 1, Py yn µ > qα(1 δ)(yn yn) + n ϕ(yn) | {z } bootstrapped threshold where ϕ(yn) is a non-negative function satisfying Py(| yn µ| ϕ(yn)) α. The detailed proof is deferred to Section B.1 in the supplement. In (2.5), the bootstrapped threshold may be interpreted as a main term, i.e., qα(1 δ)(yn yn) (at a shrunk confidence level), plus a second-order correction term, i.e., (log(2/αδ)/n)1/2ϕ(yn). The latter is added to guarantee the non-asymptotic validity of the bootstrapped threshold. In the above, ϕ(yn) could be any preliminary upper bound on yn µ. Hence, Theorem 2.2 transforms a possibly coarse prior bound ϕ(yn) on quantiles into a more accurate version that is based on a main term estimated by multiplier bootstrap plus a second-order correction term based on ϕ(yn) multiplied by a O(n 1/2) factor. Remark 2.3 (Choice of ϕ(yn)). If {yi}n i=1 are independent 1-sub-Gaussian random variables, a natural choice of ϕ(yn) is (2 log(1/α)/n)1/2 by Hoeffding s inequality (Lemma 2). Plugging it into (2.5) and letting δ = 1/2, the bootstrapped threshold in (2.5) becomes qα/4(yn yn) | {z } main term + 2 log(8/α) n | {z } second order correction Lemma B.3 in the supplement shows that the main term is of order at least O(n 1/2) as n grows, which implies the second order correction is just a remainder term. We emphasize that the reminder term is obviously not sharp and will be sharpened as a future work. Remark 2.4. Existing works on UCB-type algorithms typically utilized various concentration inequalities, e.g. Hoeffding s inequality [10] or empirical Bernstein s inequality [14], to find a valid threshold hα(yn). However, they are not data-dependent and only use the tail information, rather than fully exploit the whole distribution knowledge. This is typically conservative, and leads to over-exploration. Remark 2.5. Empirical KL-UCB [32] used empirical likelihood to build confidence intervals for general distributions that have support in [0, 1]. Although empirical KL-UCB is also data-dependent, our proposed method is from a very different non-parametric perspective and uses different tools by bootstrap. In practice, resampling tends to be more efficient computationally, without solving a convex optimization each round like empirical KL-UCB. Moreover, our method can work with unbounded rewards and we believe it is easier to generalize to structured bandits, e.g. linear bandit. In Figure 1, we compare different approaches to calculate 95% confidence bound for the population mean based on samples from a truncated-normal distribution. When the sample size is extremely small ( 10), the naive bootstrap (without any correction) cannot output a valid threshold since the bootstrapped quantile is smaller than the true 95% quantile. This confirms the necessity of the second-order correction. When the sample size increases, our bootstrapped threshold converges to the truth rapidly. This confirms the correction term is just a small remainder term. Additionally, the bootstrapped threshold is shown to be sharper than Hoeffding s bound and empirical Bernstein bound when sample size is large (see the right panel of Figure 1). 2.3 Main Algorithm: Bootstrapped UCB Based on the above theoretical findings, we conclude that bootstrapped UCB will select the arm according to its UCB index defined as below: UCBk(t) = ynk,t + qα(1 δ)(ynk,t ynk,t) + nk,t ϕ(ynk,t) , (2.7) Figure 1: 95% confidence bound of the sample mean. where nk,t is the number of pulls for arm k until time t. Practically, we may use Monte Carlo quantile approximation to get an approximated bootstrapped quantile eqα(1 δ)(ynk,t ynk,t, w B) and corresponding theorem for the control of the approximation of the bootstrapped quantile is also derived (see Section D in the supplement for details). The algorithm is summarized in Algorithm 1. The computational complexity at step t is e O(Bt) e O(BT). Comparing with vanilla UCB, the extra Bt is due to resampling. In practice, the choice of B is seldom treated as a tuning parameter, but usually determined by the available computational resource. Algorithm 1 Bootstrapped UCB Input: the number of bootstrap repetitions B, hyper-parameter δ. for t = 1 to K do Pull each arm once to initialize the algorithm. end for t = K + 1 to T do Set confidence level α = 1/(t + 1). Calculate the boostrapped quantile eqα(1 δ)(ynk,t ynk,t, w B). Pull the arm It = argmax k [K] ( ynk,t + eqα(1 δ)(ynk,t ynk,t, w B) + (log(2/αδ)/nk,t)1/2ϕ(ynk,t)). Receive reward y It. end 3 Regret Analysis In Section 3.1, we derive regret bounds for bootstrapped UCB. Moreover, we show that naive bootstrapped UCB will result in linear regret in some cases in Section A in the supplement. 3.1 Regret Bound for Bootstrapped UCB For multi-armed bandit problems, most literature [5] consider sub-Gaussian rewards. In this work, we move beyond sub-Gaussianity and consider the reward under a much weaker tail assumption, so-called sub-Weibull distribution. As shown in [33, 34], it is characterized by the right tail of the Weibull distribution and generalizes sub-Gaussian and sub-exponential distributions. Definition 1 (Sub-Weibull Distribution). We define y as a sub-Weibull random variable if it has a bounded ψβ-norm. The ψβ-norm of y for any β > 0 is defined as y ψβ := inf n C (0, ) : E[exp(|y|β/Cβ)] 2 o . Particularly, when β = 1 or 2, sub-Weibull random variables reduce to sub-exponential or sub Gaussian random variables, respectively. It is obvious that the smaller β is, the heavier tail the random variable has. Next theorem provides a corresponding concentration inequality for the sum of independent sub-Weibull random variables. Theorem 3.1 (Concentration Inequality for Sub-Weibull Distribution). Suppose {yi}n i=1 are independent sub-Weibull random variables with yi ψβ σ. Then there exists an absolute constant Cβ only depending on β such that for any a = (a1, . . . , an) Rn and 0 < α < 1/e2, i=1 aiyi E( i=1 aiyi) Cβσ a 2(log α 1)1/2 + a (log α 1)1/β with probability at least 1 α. The proof relies on a precise characterization of p-th moment of a Weibull random variable and standard symmetrization arguments. Details are deferred to Section B.2 in the supplement. This theorem generalizes the Hoeffding-type concentration inequalities for sub-Gaussian random variables (see, e.g. Proposition 5.10 in [35]), and Bernstein-type concentration inequalities for sub-exponential random variables (see, e.g. Proposition 5.16 in [35]) up to some constants. In Theorem 3.2, we provide both problem-dependent and problem-independent regret bounds. Theorem 3.2. Consider a stochastic K-armed sub-Weibull bandit, where the noise follows a symmetric sub-Weibull distribution with its ψβ-norm upper bounded by σ. Denote nk,t as the number of pulls for arm k until time t. We choose ϕ according to Theorem 3.1 as follows ϕ(ynk,t) = Cβσ s nk,t + (log 2/α)1/β and let the confidence level α = 1/T 2. For any round T, the problem-dependent regret of bootstrapped UCB is upper bounded by k: k>0 128C2 βσ2 log T k + 23+1/βCβσK(log T)1/β + 4 k=2 k, (3.2) where Cβ is some absolute constant from Theorem 3.1, and k is the sub-optimality gap. Moreover, if the round T 22/β 3K(log T)2/β 1, the problem-independent regret of bootstrapped UCB is upper bounded by TK log T + 4Kµ 1. (3.3) The main proof structure follows the standard analysis of UCB [5] and relies on a sharp upper bound for the (data-dependent) bootstrapped quantile term by Theorem 3.1. Details are deferred to Section B.3 in the supplement. When β 1, (3.2) provides a logarithm regret that matches the state-of-art result [5]. When β < 1, we have a non-negligible term (log T)1/β that is the price paid for heavy-tailedness. However, this term does not depend on the gap k. Therefore, we have an optimal problem-independent regret bound. Remark 3.3. The choice of α = 1/T 2 led to an easy analysis. Using similar techniques in Chapter 8.2 of [5], we can achieve a similar regret bound by setting αt = 1/(t logτ(t)) for any τ > 0. Remark 3.4. [22] consider bandit with heavy-tail (moment of order (1 + ε)) based on a median-ofmeans estimator. As mentioned in [36], there are two disadvantages for median-of-means approach: (a) it involves an additional tuning parameter; (b) it is numerically unstable for small sample size. In contrast, we identify a class of heavy-tailed bandits (sub-Weibull bandit) where mean estimators can still achieve regret bounds of the same order as those under sub-Gaussian reward distributions. The reason is that although sub-Weibull r.v. has heavier tail than sub-Gaussian r.v., its tail still has an exponential-like decay. 4 Experiments In Section 4.1, we consider multi-armed bandits with both symmetric and asymmetric rewards. In Section 4.2, we extend our method to linear bandits. Implementation details and some additional experimental results are deferred to Section E in the supplement. 4.1 Multi-armed Bandit In this section, we compare bootstrapped UCB (Algorithm 1) with three baselines: Upper Confidence Bound based on concentration inequalities (Vanilla UCB), Thompson sampling with normal Jeffery prior [37] (Jeffery-TS) and Thompson sampling with Beta prior [38] (Bernoulli-TS). For bounded rewards, we also compare with Giro [27]1, that is a sampling-based exploration method by adding artificial pseudo observations {0, 1} to escape from local optima, and empirical KL-UCB [39] using package: Pyma Bandits. For the preliminary bound ϕ(yn), we simply choose the one derived by the concentration inequality. Note that the second-order correction term in (2.5) is conservative. For practitioners, we suggest to set the correction term to be ϕ(yn)/ n. To be fair, we choose the confidence level α = 1/(1 + t) for both UCB1 and bootstrapped UCB, and δ = 0.1 in (2.5). All algorithms above require knowledge of an upper bound on the noise standard deviation. The number of bootstrap repetitions is B = 200, and the number of arms is K = 5. First, we consider symmetric rewards with a mean parameter µk generated from Uniform( 1, 1). The noise follows either truncated-normal distribution within [ 1, 1], or standard Gaussian distribution. From Figure 2, bootstrapped UCB outperforms Jeffery-TS and Vanilla-UCB for truncated-normal bandit and has comparable or sometimes better performance over empirical KL-UCB. It s obvious that if the reward distribution is exactly Gaussian and the plug-in estimate for the noise standard deviation is the truth, Jeffery-TS should be the best. However, when the posterior (plots (a),(b)) or noise standard derivation (plot (c)) are mis-specified, the performance of TS deteriorates fast. Since (concentration-based) Vanilla UCB only uses the tail information (bounded or sub-Gaussian), it is very conservative and results in bad regret as expected. Figure 2: Cumulative regrets for truncated-normal bandit and Gaussian bandit. Sigma is the upper bound on the standard deviation of the noise. The results are averaged over 200 realizations. 1We have implemented Giro in the unbounded reward case, which could result in linear regret in most cases. See Figure 7 in the supplement. So, it s unclear what is the best way to add pseudo observations in this case. Second, we consider asymmetric rewards with a mean parameter µk generated from Uniform(0.25, 0.75). For Bernoulli bandit, the reward follows Ber(µk); for Beta bandit, the reward follows 2 Beta(vµk, v(1 µk)) for v = 8. From Figure 3, bootstrapped UCB outperforms Vanilla UCB and Giro in both cases, and outperforms Bernoulli-TS for Beta bandit. In fact, we are supposed not to beat Bernoulli-TS for Bernoulli bandit since TS fully makes use of the distribution knowledge in this case. One possible explanation is that our method is non-parametric. Figure 3: Cumulative regrets for Bernoulli bandit and Beta bandit. The results are averaged over 200 realizations. Third, we demonstrate that the robustness of bootstrapped UCB over mis-specifications of the noise standard deviation. In the left panel of Figure 4, we consider the cumulative regret at round T = 2000 of standard Gaussian bandit. As one can see, when we increase the plug-in upper bound of the standard deviation of the noise, bootstrapped UCB is more robust than Bernoulli-TS and Vanilla UCB. Figure 4: The left panel is the cumulative regret over noise levels while the right panel is the instancedependent regret of various algorithms as a function of gaps. The results are averaged over 200 realizations. Last, we present a frequentist instance-dependent regret curve for truncated-normal bandit and the experiment set up follows [40]. We plot cumulative regrets at T = 1000 of various algorithms with respect to the instance gap and the mean vector µ = ( , 0, 0, 0, 0). The results are summarized in the right panel of Figure 4. 4.2 Linear Bandit We extend our method to linear bandit case. The basic set up follows the one in [15]. In detail, θ Rd is drawn from a multivariate Gaussian distribution with mean vector µ = 0 and covariance matrix Σ = 10Id. The noise follows a standard Gaussian distribution. There are 100 actions with feature vector components drawn uniformly at random from [ 1/ 10]. We consider two state-of-art methods: Thompson sampling for linear bandit [41] (TSL) and optimism in the face of uncertainty for linear bandits [13] (OFUL). Following the principle of constructing second-order 2We adopt the technique in [38] to run Thompson Sampling with [0, 1] rewards. In particular, for any reward yt [0, 1], we draw pseudo reward byt Ber(yt), and then use byt instead of yt in the algorithm. correction in mean problems (Theorem 2.2), we construct the bootstrapped UCB for linear bandit (BUCBL) as follows: At each round t, the action is selected as argmaxx(x bθt + βBUCBL t,1 δ x V 1 t ), where βBUCBL t,1 δ = qα(bθ(b) t bθt) + βOFUL t,1 δ,σ/ n. The formal definition of bθt, bθ(b) t , βOFUL t,1 δ,σ and some basic setups are given in Section E.2 in the supplement. To be fair, the confidence level for all methods is set to be δ = 1/(1 + t) and we plug in the true standard deviation of the noise for each method. From Figure 5, we can see that bootstrapped UCB greatly improves the cumulative regret over TSL and OFUL. Figure 5: Cumulative regret for linear bandit. 5 Conclusion In this paper, we propose a novel class of non-parametric and data-driven UCB algorithms based on multiplier bootstrap. It is easy to implement and has the potential to be generalized to other complex structured problems. As future works, we will evaluate our idea on other structured contextual bandits and reinforcement learning problems. Acknowledgments We thank Tor Lattimore for helpful discussions. Guang Cheng would like to acknowledge support by NSF DMS-1712907, DMS-1811812, DMS-1821183, and Office of Naval Research (ONR N0001418-2759). In addition, Guang Cheng is a visiting member of Institute for Advanced Study, Princeton (funding provided by Eric and Wendy Schmidt) and visiting Fellow of SAMSI for the Deep Learning Program in the Fall of 2019; he would like to thank both Institutes for their hospitality. [1] Lihong Li, Wei Chu, John Langford, and Robert E. Schapire. A contextual-bandit approach to personalized news article recommendation. In Proceedings of the 19th International Conference on World Wide Web, WWW 10, pages 661 670, New York, NY, USA, 2010. ACM. [2] Jens Kober, J Andrew Bagnell, and Jan Peters. Reinforcement learning in robotics: A survey. The International Journal of Robotics Research, 32(11):1238 1274, 2013. [3] David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, et al. Mastering the game of go with deep neural networks and tree search. Nature, 529(7587):484, 2016. [4] Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT press, 2018. [5] Tor Lattimore and Csaba Szepesvári. Bandit algorithms. preprint, 2018. [6] Hamsa Bastani and Mohsen Bayati. Online decision-making with high-dimensional covariates. Available at SSRN 2661896, 2015. [7] Hamsa Bastani, Mohsen Bayati, and Khashayar Khosravi. Mostly exploration-free algorithms for contextual bandits. ar Xiv preprint ar Xiv:1704.09011, 2017. [8] Sarah Bird, Solon Barocas, Kate Crawford, Fernando Diaz, and Hanna Wallach. Exploring or exploiting? social and ethical implications of autonomous experimentation in ai. In Workshop on Fairness, Accountability, and Transparency in Machine Learning, 2016. [9] Peter Auer. Using confidence bounds for exploitation-exploration trade-offs. Journal of Machine Learning Research, 3(Nov):397 422, 2002. [10] Peter Auer, Nicolo Cesa-Bianchi, and Paul Fischer. Finite-time analysis of the multiarmed bandit problem. Machine learning, 47(2-3):235 256, 2002. [11] Varsha Dani, Thomas P Hayes, and Sham M Kakade. Stochastic linear optimization under bandit feedback. 2008. [12] Lihong Li, Wei Chu, John Langford, and Robert E Schapire. A contextual-bandit approach to personalized news article recommendation. In Proceedings of the 19th international conference on World wide web, pages 661 670. ACM, 2010. [13] Yasin Abbasi-Yadkori, Dávid Pál, and Csaba Szepesvári. Improved algorithms for linear stochastic bandits. In Advances in Neural Information Processing Systems, pages 2312 2320, 2011. [14] Volodymyr Mnih, Csaba Szepesvári, and Jean-Yves Audibert. Empirical bernstein stopping. In Proceedings of the 25th international conference on Machine learning, pages 672 679. ACM, 2008. [15] Daniel Russo and Benjamin Van Roy. Learning to optimize via posterior sampling. Mathematics of Operations Research, 39(4):1221 1243, 2014. [16] Donald B Rubin. The bayesian bootstrap. The annals of statistics, pages 130 134, 1981. [17] Chien-Fu Jeff Wu et al. Jackknife, bootstrap and other resampling methods in regression analysis. the Annals of Statistics, 14(4):1261 1295, 1986. [18] Sylvain Arlot, Gilles Blanchard, Etienne Roquain, et al. Some nonasymptotic results on resampling in high dimension, i: confidence regions. The Annals of Statistics, 38(1):51 82, 2010. [19] Victor Chernozhukov, Denis Chetverikov, Kengo Kato, et al. Gaussian approximation of suprema of empirical processes. The Annals of Statistics, 42(4):1564 1597, 2014. [20] Vladimir Spokoiny, Mayya Zhilova, et al. Bootstrap confidence sets under model misspecification. The Annals of Statistics, 43(6):2653 2675, 2015. [21] Yun Yang, Zuofeng Shang, and Guang Cheng. Non-asymptotic theory for nonparametric testing. ar Xiv preprint ar Xiv:1702.01330, 2017. [22] Sébastien Bubeck, Nicolo Cesa-Bianchi, and Gábor Lugosi. Bandits with heavy tail. IEEE Transactions on Information Theory, 59(11):7711 7717, 2013. [23] Adam N. Elmachtoub, Ryan Mc Nellis, Sechan Oh, and Marek Petrik. A practical method for solving contextual bandit problems using decision trees. In Proceedings of the Thirty-Third Conference on Uncertainty in Artificial Intelligence, UAI 2017, Sydney, Australia, August 11-15, 2017, 2017. [24] Ian Osband, Charles Blundell, Alexander Pritzel, and Benjamin Van Roy. Deep exploration via bootstrapped dqn. In Advances in neural information processing systems, pages 4026 4034, 2016. [25] Liang Tang, Yexi Jiang, Lei Li, Chunqiu Zeng, and Tao Li. Personalized recommendation via parameterfree contextual bandits. In Proceedings of the 38th international ACM SIGIR conference on research and development in information retrieval, pages 323 332. ACM, 2015. [26] Dean Eckles and Maurits Kaptein. Thompson sampling with the online bootstrap. ar Xiv preprint ar Xiv:1410.4009, 2014. [27] Branislav Kveton, Csaba Szepesvari, Zheng Wen, Mohammad Ghavamzadeh, and Tor Lattimore. Garbage in, reward out: Bootstrapping exploration in multi-armed bandits. ar Xiv preprint ar Xiv:1811.05154, 2018. [28] Tze Leung Lai and Herbert Robbins. Asymptotically efficient adaptive allocation rules. Advances in applied mathematics, 6(1):4 22, 1985. [29] George Casella and Roger L Berger. Statistical inference, volume 2. Duxbury Pacific Grove, CA, 2002. [30] Bradley Efron. The jackknife, the bootstrap, and other resampling plans, volume 38. Siam, 1982. [31] Aad W Van der Vaart. Asymptotic statistics, volume 3. Cambridge university press, 2000. [32] Olivier Cappé, Aurélien Garivier, Odalric-Ambrym Maillard, Rémi Munos, Gilles Stoltz, et al. Kullback leibler upper confidence bounds for optimal sequential allocation. The Annals of Statistics, 41(3):1516 1541, 2013. [33] Arun Kumar Kuchibhotla and Abhishek Chakrabortty. Moving beyond sub-gaussianity in high-dimensional statistics: Applications in covariance estimation and linear regression. ar Xiv preprint ar Xiv:1804.02605, 2018. [34] Mariia Vladimirova and Julyan Arbel. Sub-weibull distributions: generalizing sub-gaussian and subexponential properties to heavier-tailed distributions. ar Xiv preprint ar Xiv:1905.04955, 2019. [35] Roman Vershynin. Compressed sensing, chapter Introduction to the non-asymptotic analysis of random matrices, pages 210 268. Cambridge Univ. Press, 2012. [36] Xi Chen and Wen-Xin Zhou. Robust inference via multiplier bootstrap. The Annals of Statistics, to appear, 2019. [37] Nathaniel Korda, Emilie Kaufmann, and Remi Munos. Thompson sampling for 1-dimensional exponential family bandits. In Advances in Neural Information Processing Systems, pages 1448 1456, 2013. [38] Shipra Agrawal and Navin Goyal. Further optimal regret bounds for thompson sampling. In Artificial intelligence and statistics, pages 99 107, 2013. [39] Aurélien Garivier and Olivier Cappé. The kl-ucb algorithm for bounded stochastic bandits and beyond. In Proceedings of the 24th annual Conference On Learning Theory, pages 359 376, 2011. [40] Tor Lattimore. Refining the confidence level for optimistic bandit strategies. The Journal of Machine Learning Research, 19(1):765 796, 2018. [41] Shipra Agrawal and Navin Goyal. Thompson sampling for contextual bandits with linear payoffs. In International Conference on Machine Learning, pages 127 135, 2013. [42] Sharan Vaswani, Branislav Kveton, Zheng Wen, Anup Rao, Mark Schmidt, and Yasin Abbasi-Yadkori. New insights into bootstrapping for bandits. ar Xiv preprint ar Xiv:1805.09793, 2018. [43] P Hitczenko, SJ Montgomery-Smith, and K Oleszkiewicz. Moment inequalities for sums of certain independent symmetric random variables. Studia Math, 123(1):15 42, 1997. [44] Michel Talagrand. The supremum of some canonical processes. American Journal of Mathematics, 116(2):283 325, 1994. [45] Michel Ledoux and Michel Talagrand. Probability in Banach Spaces: isoperimetry and processes. Springer Science & Business Media, 2013. [46] Joseph P Romano and Michael Wolf. Exact and approximate stepdown methods for multiple hypothesis testing. Journal of the American Statistical Association, 100(469):94 108, 2005. [47] Noga Alon and Joel H Spencer. The probabilistic method. John Wiley & Sons, 2004. [48] Radoslaw Adamczak, Alexander E Litvak, Alain Pajor, and Nicole Tomczak-Jaegermann. Restricted isometry property of matrices with independent columns and neighborly polytopes by random sampling. Constructive Approximation, 34(1):61 88, 2011. [49] Robert Bogucki. Suprema of canonical weibull processes. Statistics & Probability Letters, 107:253 263, 2015. [50] Victor De la Pena and Evarist Giné. Decoupling: from dependence to independence. Springer Science & Business Media, 2012.