# bandit_learning_with_implicit_feedback__485a585a.pdf Bandit Learning with Implicit Feedback Yi Qi1, Qingyun Wu2, Hongning Wang2, Jie Tang1, Maosong Sun1 1 State Key Lab of Intell. Tech. & Sys., Institution for Artificial Intelligence, Dept. of Comp. Sci. & Tech., Tsinghua University, Beijing, China 2 Department of Computer Science, University of Virginia qi-y16@mails.tsinghua.edu.cn, {jietang, sms}@tsinghua.edu.cn {qw2ky,hw5x}@virginia.edu Implicit feedback, such as user clicks, although abundant in online information service systems, does not provide substantial evidence on users evaluation of system s output. Without proper modeling, such incomplete supervision inevitably misleads model estimation, especially in a bandit learning setting where the feedback is acquired on the fly. In this work, we perform contextual bandit learning with implicit feedback by modeling the feedback as a composition of user result examination and relevance judgment. Since users examination behavior is unobserved, we introduce latent variables to model it. We perform Thompson sampling on top of variational Bayesian inference for arm selection and model update. Our upper regret bound analysis of the proposed algorithm proves its feasibility of learning from implicit feedback in a bandit setting; and extensive empirical evaluations on click logs collected from a major MOOC platform further demonstrate its learning effectiveness in practice. 1 Introduction Contextual bandit algorithms [4, 20, 19] provide modern information service systems an effective solution to adaptively find good mappings between available items and users. This family of algorithms sequentially select items to serve users using side information about user and item, while adapting their selection strategies based on the immediate user feedback to maximize users long-term satisfaction. They have been popularly deployed in practical systems for content recommendation [20, 5, 26] and display advertising [6, 22]. However, the most dominant form of user feedback in such systems is implicit feedback, such as clicks, which is known to be biased and incomplete about users evaluation of system s output [16, 11]. For example, a user skips a recommended item might not be because he/she does not like the item, but he/she just does not examine that display position, i.e., position bias [13]. Unfortunately, a common practice in contextual bandit applications simply treats no click as a form of negative feedback [20, 25, 6]. This introduces inconsistency to model update, since the skipped items might not be truly irrelevant, and it inevitably leads to suboptimal outputs of bandit algorithms over time. In this work, we focus on learning contextual bandits with user click feedback, and model such implicit feedback as a composition of user result examination and relevance judgment. Examination hypothesis [8], which is a fundamental assumption in click modeling, postulates that a user clicks on a system s returned result if and only if that result has been examined by the user and it is relevant to the user s information need at the moment. Because a user s examination behavior is unobserved, we model it as a latent variable, and realize the examination hypothesis in a probabilistic model. We define the conditional probabilities of result examination and relevance judgment via logistic functions over the corresponding contextual features. To perform model update, we take a variational 32nd Conference on Neural Information Processing Systems (Neur IPS 2018), Montréal, Canada. Bayesian approach to develop a closed form approximation to the posterior distribution of model parameters on the fly. This approximation also paves the way for an efficient Thompson sampling strategy for arm selection in bandit learning. Our finite time analysis proves that, despite the increased complexity in parameter estimation introduced by the latent variables, our Thompson sampling policy based on the true posterior is guaranteed to achieve a sub-linear Bayesian regret with a high probability. We also demonstrate that the regret of Thompson sampling based on the approximated posterior is well-bounded. In addition, we prove that when one fails to model result examination in click feedback, a linearly increasing regret is possible, as the model cannot differentiate examination driven skips from relevance driven skips in the negative feedback. We tested the algorithm in Xuetang X1, a major Massive Open Online Course (MOOC) platform in China, for personalized education. To personalize students learning experience on this platform, we recommend quiz-like questions in a form of banners on top of the lecture videos when students are watching the videos. The algorithm needs to decide where in a video to display which question to a target student. If the student feels the displayed question is helpful for him/her to understand the lecture content, he/she could click on the banner to read the answer and more related online content about the question. Therefore, our goal is to maximize the click through rate (CTR) on the selected questions. There are several properties of this application that amplifies the bias and incompleteness of click feedback. First, based on the consideration of user experience, to minimize the risk of annoying any student, the displayed time of a banner is limited to a few seconds. Second, as this feature is newly introduced to the platform, many users might not realize that they can click on the question to read more related content about it. As a result, no click on a question does not necessarily indicate its irrelevance. We tested the algorithm in this application in a four-month period, where a total of 69 questions are manually compiled for the algorithm to select over 20 major videos with more than 100 thousands student video watching sessions. Based on the unbiased offline evaluation policy [21], our algorithm achieved a 8.9% CTR lift compared to standard contextual bandits [20, 9] which do not model users examination behavior. 2 Related Works As having been extensively studied in click modeling of user search results [7], various factors affect users click decisions; and among them result examination plays a central role [13, 8]. Unfortunately, most applications of bandit algorithms simply treat user clicks as explicit feedback for model update [20, 25, 6, 26], where no click on a selected result is considered as negative feedback. This inevitably leads to inaccurate model update and sub-optimal arm selection. There is a line of related research that develops click model based bandit algorithms for learning to rank problems. For example, by assuming that skipped documents are less attractive than later clicked ones in a ranked list, Kveton et al. [17] developed a cascading bandit model to learn from both clicks and skips in search results. To enable learning from multiple clicks in the same result ranking list, they adopted the dependent click model [10] to infer user satisfaction after a sequence of clicks [14], and later further extended it to broader types of click models [27]. However, such algorithms aim at estimating the best ranking of results in a per-query basis, without specifying any specific ranking function. Hence, it is hard for them to generalize to unseen queries. This directly limits their application scenario in practice. The solution developed in Lagrée et al. [18] is the closest to ours, which exploits bias in reward distribution induced by different examination probabilities at different display positions. Yet they assumed the examination probability only depends on position, while we allow any reasonable feature to be a determinant. Besides, they postulated that the probability of examination at each position is either heuristically set or empirically estimated, and henceforth fixed; while we estimate it on the fly from the observations obtained by interacting with users. Another line of related research is bandit learning with latent variables. Maillard and Mannor studied the problem of latent bandit [23], which assumes reward distributions are clustered and the clusters are determined by some latent variables. They only studied the problem in a context-free setting, and a very weak performance guarantee is provided when the reward distribution is unknown in those clusters. Kawale et al. developed a Thompson sampling scheme for online matrix-factorization [15]. Latent features are extracted via an online low-rank matrix completion based on samples selected from Thompson sampling on the fly. Due to the ad-hoc combination of factorization method and bandit method, little theoretical analysis was provided. Wang et al. studied the problem of latent 1http://www.xuetangx.com/ feature learning for contextual bandits [25]. They extended arms context vectors with latent features under a linear reward structure, and applied the upper confidence bound principle over coordinate descent to iteratively estimate the hidden features and model parameters. The linear reward structure prohibits it from recognizing the nonlinear dependency between result examination and relevance judgment in click feedback. And their regret analysis depends heavily on the initialization of the algorithm, which could be hard to achieve in practice. 3 Problem Setup We consider a contextual bandit problem with finite, but possibly large, number of arms. Denote the arm set as A. At each trial t = 1, ..., T, the learner observes a subset of candidate arms At with At A, where each arm a is associated with a context vector xa summarizing the side information about the arm. Once an arm at 2 At is chosen according to some policy , corresponding implicit binary feedback Cat, e.g., user click, will be given to the learner as the reward. The learner s goal is to adjust its arm selection strategy to maximize its cumulative reward over time. What makes this problem unique and challenging is that Cat does not truly reflect users evaluation of the selected arm at. Based on the examination hypothesis [13, 8], when Cat = 1, the chosen at must be relevant to the user s information need at time t; but when Cat = 0, at might be relevant but the user just does not examine it. Unfortunately, the result examination condition is unobserved to the learner. We model a user s result examination via a binary latent variable Eat and assume that the context vector xa t of arm a can be decomposed into (xa E,t), where the dimension of xa E,t are d C and d E respectively. Accordingly, users result examination and relevance judgment decisions are assumed to be governed by a conjecture of (xa E,t) and the corresponding bandit parameter = ( E). In the rest of this paper, when no ambiguity is introduced, we drop the index a to simplify the notations. As a result, we make the following generative assumption about an observed click Ct on arm at, P(Ct = 1|Et = 0, x C,t) = 0 P(Ct = 1|Et = 1, x C,t) = (x T P(Et = 1|x E,t) = (x T where (x) = 1 1+e x . Based on this assumption, we have E[Ct|xt] = (x T E). As a result, the observed click feedback Ct is a sample from this generative process. Define f (x) := E[C|x, ] = (x T E E). The accumulated regret of a policy up to time T is formally defined as, Regret(T, , ) = max a2At f (xa) f (xat) where xat := (xat E ) is the context vector of the arm at 2 At selected by the policy at time t based on the history Ht := {(Ai, xi, Ci)}t 1 i=1. The Bayesian regret is defined by E Regret(T, , ) , where the expectation is taken with respect to the prior distribution over , and it can be written as, Bayes Regret(T, ) = max a2At f (xa) f (xat) In our online learning setting, the objective is to find the policy that minimizes the accumulated regret over T. 4 Algorithm The learner needs to estimate the bandit parameters E based on its interactively obtained click feedback {xi, Ci}t i=1 over time. Ideally, this estimation can be obtained by maximizing the data likelihood with respect to the bandit model parameters. However, the inclusion of examination as a latent variable in our bandit learning setting poses serious challenges to both parameter estimation and arm selection. Neither conventional least square estimator nor maximum likelihood estimator can be easily obtained, let alone computational efficiency, due to the non-convexity of the corresponding optimization problem. To make things even worse, the two popular bandit learning paradigms, upper confidence bound principle [1] and Thompson sampling [3], both demand an accurate estimation of bandit parameters and their uncertainty. In this section, we present an efficient new solution to tackle these two challenges, which makes use of variational Bayesian inference technique to learn parameters approximately on the fly, as well as to bridge parameter estimation and arm selection policy design. 4.1 Variational Bayesian for parameter estimation To complete the generative process defined in Section 3, we further assume C and E follow Gaussian distribution N(ˆ C, C) and N(ˆ E, E) respectively. We are interested in developing a closed form approximation to their posteriors, when a newly obtained observation (x C, x E, C) becomes available. By applying Bayes rule in the log space, we have, log P( C, E|x C, x E, C) = log P(C| C, E, x C, x E) + log P( C, E) + log const =C log (x T E E) + (1 C) log 2( C ˆ C)T 1 C ( C ˆ C) 1 2( E ˆ E)T 1 E ( E ˆ E) + log const The key idea is to develop a variational lower bound in the quadratic form of C and E for the loglikelihood function. Because of the convexity of log (x) x 2 with respect to x2 (See Appendix B.1) and the Jensen s inequality for log x (See Appendix B.2), a lower bound of the required form is achievable. When C = 1, by Eq (16) in Appendix B.3, we have, l C=1(x C, x E, ) := log C , C) + g(x T where g(x, ) := x 2 + log ( ) λ( )(x2 2), λ( ) = tanh 2 4 , x, 2 R. More specifically, g(x, ) is a polynomial of degree 2 with respect to x. When C = 0, by Eq (17) in Appendix B.3, we have, l C=0(x C, x E, ) := log H(q) + qg( x T C , C) + qg(x T E , E,1) + (1 q)g( x T where H(q) := q log q (1 q) log(1 q). Once the lower bound in the quadratic form is established, we can use a Gaussian distribution to approximate our target posterior, whose mean and covariance matrix are determined by the following equations, C + 2q1 Cλ( C)x Cx T ˆ C,post = C,post( 1 2( q)1 Cx C) (4) E + 2λ( E)x Ex T ˆ E,post = E,post( 1 2(2q 1)1 Cx E) (6) where the subscript post denotes the parameters in the Gaussian distributions that approximate the desired posteriors. Consecutive observations can be incorporated into the approximated posteriors sequentially. There is one thing left to decide, i.e., the choice of variational parameters ( C, E, q). A typical criterion is to choose the values such that the likelihood on the observations is maximized. Similar to the choice made by [12], we choose the closed form update formulas of those variational parameters as, q = exp (g(x T C C, C) + g(x T E E, E) g( x T E E, E)) 1 + exp (g(x T C C, C) + g(x T E E, E) g( x T Algorithm 1 Thompson sampling for E-C Bandit 1: Initiate C = λI, E = λI, ˆ C = C,0, ˆ E = E,0. 2: for t = 0, 1, 2... do 3: Observe the available arm set At A and its corresponding context set Xt := {(xa E) : a 2 At}. 4: Randomly sample C N(ˆ C, C), E N(ˆ E, E). 5: Select: at = arg max C)T C) ((xa 6: Play the selected arm at and Observe the reward Ct. 7: Update C, ˆ C, E, ˆ E according to Eq (3) to Eq (6) respectively. 8: end for where all the expectations are taken under the approximated posteriors. Empirically, we found the iterative update of the approximated posterior and the variational parameters converge quite rapidly, such that it usually only needs a few rounds of iterations to get a satisfactory local maximum in our experiments. 4.2 Thompson sampling with approximated lower bound Thompson sampling, also known as probability matching, is widely used in bandit learning to balance exploration and exploitation, and it shows great empirical performance [6]. Thompson sampling requires a distribution of the model parameters to sample from. In a standard Thompson sampling [3], one is required to sample from the true posterior of model parameters. But as logistic regression does not have a conjugate prior, the model defined in our problem does not have an exact posterior. We decide to sample from the approximated posterior as derived in Eq (3) to Eq (6). Later we will demonstrate this is a very tight posterior approximation. Once the sampling of ( C, E) is complete, we can select the corresponding arm at 2 At which maximizes (x T E E). We name the resulting bandit algorithm as examination-click bandit, or E-C Bandit in short, and summarize it in Algorithm 1. 5 Regret Analysis Recall our object is to find the policy that minimizes the Beyesian regret, Bayes Regret(T, ) = max a2At f (xa) f (xat) where f (x) := E[C|x, ] = (x T E E). Our algorithm, which is based on a maximum likelihood estimator, is equivalent to an estimator that minimizes a log-loss with binary random variables. In this section, we will first bound the aggregate empirical discrepancy of the log-loss estimator used in our model in Proposition 1. This prepares for the upper bound of the generic Bayeisan regret under a log-loss estimator with Thompson sampling policy in Theorem 1. Based on this generic Bayesian regret bound, we study the upper bound of Bayesian regret for our proposed E-C Bandit. Due to space limit, we provide all the detailed proofs in the Appendix. To further simplify our notations, we use f for f , which is the reward function based the estimated bandit parameter , and fk for f (xak), i.e., the reward for arm ak. We use f for f , which is the reward function based on the ground-truth bandit parameter, and correspondingly f k for f (xak). We assume that f lies in a known function space F, where any f 2 F is a function mapping from the arm set A to the range (0, 1). Define the log-loss estimator by ˆf LOGLOSS t 2 arg minf2F L2,t(f) where L2,t(f) is the aggregate log-loss written as Pt 1 k=1 lk(f) where lk(f) = Ck log fk + (1 Ck) log(1 fk) . We have the following proposition, Proposition 1. Denote the aggregate empirical discrepancy Pt k)2 by kf f k2 all δ > 0 and > 0, if Ft = )))f ˆf LOGLOSS for all t 2 N, t (F, δ, ) is an appropriately constructed confidence parameter. In particular, it is defined as β t (F, δ, ) := 2 λ0 log(N(F, , k k1)/δ)+2 t, where N(F, , k k1) denotes the alpha-covering number of F, λ0 = 3 ( 1 mf + 1 1 Mf )2 and t = 4Mf + 1 min{mf ,1 Mf } t, in which mf, Mf 2 R are upper and lower bounds of f such at 0 < mf f Mf < 1 for any f 2 F. Remark 1. The proof is provided in Appendix C. Here we discuss two important details related to our later proof about E-C Bandit s regret. First, the precise optimization of ˆf LOGLOSS t 2 arg minf2F L2,t(f) could be hard in some instances of F. For example, when F is a set of non-convex functions. Nevertheless, we can always resort to approximation methods to solve the optimization problem as long as the approximation error can be bounded. Indeed, in our E-C Bandit, we resort to variational inference to estimate ˆf LOGLOSS t on the fly and find it works quite well in practice. Second, when f 62 F, this corresponds to the problem of model mis-specification. In this situation, the regret bound could be very poor, as the real regret could be linear with respect to time. To show this clearly in our case, in Appendix F we construct a situation in which the regret is inevitably linear if one fails to model the examination condition in click feedback and simply treats no click as negative feedback. With Proposition 1, we have the following theorem which bounds the Bayesian regret of the Thompson Sampling strategy under a log-loss estimator. Theorem 1. For all T 2 N, > 0 and δ < 1 2T , if T S denotes the policy derived from the log-loss estimator and a Thompson sampling strategy along the time steps, then Bayes Regret(T, T S) 1 + E(F, T 1) + 1 T (F, , δ)T (8) where C = supf2F{sup f}, dim A E(F, T 1) is the eluder dimension (see Definition 3 in Russo and Van Roy [24]) of F with respect to A. Remark 2. We can choose C = 1 in our click feedback case since f 2 (0, 1). C is kept in the theorem to show the same form compared to the Proposition 8 in Russo and Van Roy [24]. In fact, the proof is almost the same once we have Proposition 1. Hence, we omit the proof in our paper. Now we turn to provide an upper regret bound of our E-C Bandit, based on the above generic Bayeisan regret analysis under a log-loss estimator. We add the following two assumptions which are standard in the literature of contextual bandits. Assumption 1. The optimal lies in Bs := { 2 Rd : k k2 s}, and s is known as a prior. Assumption 2. The norm of context vectors are bounded by x, i.e., (x C, x E) 2 Bx, where Bx := {x 2 Rd : kxk2 x} and x is known as a prior. Based on these two assumptions, it is straightforward to verify that (x T E E) and f (x) are bounded. Let M = max 2Bs,x2Bx (x T ) and m = min 2Bs,x2Bx (x T ). Hence, 0 < m M < 1. Similarly, denote the maximum of f (x) by Mf and the minimum by mf, we have 0 < mf Mf < 1. Once the arm set is restricted to a finite cardinality, we have dim A E(F, T 1) |A| by Appendix C.1. in Russo and Van Roy [24]. Choosing the function class as that in our E-C Bandit, i.e., F = {f : Bx ! R|f = (x T E E), 2 Bs}, by Lemma 8 (See Appendix 8 for its proof), we have N(F, , k k1) = (γ/ )d where γ = 2M k x (k is the Lipschitz constant of , see Lemma 4). Hence, choosing = 1/t2 and δ = 1/t leads to t (F, 1/t, 1/t2) = 2 d log(γt3) + 1 Therefore, the upper bound of Bayesian regret of our proposed E-C Bandit takes the following form, Bayes Regret(T, T S) = O(|A| + d|A|T log T). (10) When T |A| and T d, which is a typical case in practice, it becomes O(p T log T). 6 Experiment We perform empirical evaluations in simulation and on the online student click logs collected from our MOOC platform to verify the effectiveness of our proposed algorithm. In particular, we compare with those that fail to model the examination condition and directly use click as feedback. 6.1 Algorithms for comparison We list the models used for empirical comparisons below, and explain how we adjust them in our evaluations. Logistic Bandit. This model has been extensively used for online advertisement CTR optimization. In [6, 21], the authors model user clicks by a regularized logistic regression model over observed context features and make decisions by applying Thompson sampling over the learnt model. In particular, no click is treated as negative feedback. Following their setting in [6], we used the Laplace approximation and Gaussian prior presented in to update the model parameters on the fly. We also want to highlight that despite mountains of works focusing on generalized linear bandits, most of them are not truly online algorithms, because the estimation of their parameters at each iteration has to involve all historical observations iteratively. This incurs a space complexity at least O(T) and time complexity at least O(T 2) (e.g., Filippi et al. [9] requires exact optimum of logistic regression on all historical observations at each round). h Lin UCB. This is an algorithm proposed by Wang et al. [25] for bandit learning with latent variables. It is related to our model in a sense that both models estimate hidden features. In particular, h Lin UCB extends linear contextual bandit by inclusion of hidden features and operates under a UCB-like strategy. However, it still treats click as direct feedback, but aims at learning more expressive features to describe the observed clicks. E-C Bandit. This is the algorithm we present in Algorithm 1. We should note that in the experiments on real-world data, the manual separation of examination feature x E and click feature x C in the context vector x offers a principled approach to incorporate one s domain knowledge about what affects user examination and what affects user result relevance. We explain in detail what features are chosen for which component in Appendix G. Thanks to the tight approximation achieved by Bayesian variational inference presented in Section 4, truly online model update is feasible in this algorithm. This provides both computational and storage efficiency. 6.2 Experiments on simulations First we demonstrate the effectiveness of our algorithm by experiment with simulated data. The experiment is performed as follows. The context vector s dimension d C and d E are set to 5, and thus d = d C + d E = 10. We set |A| = 100, each of which is associated with a unique context vector (x C, x E). The ground-truth parameter ( E) and the specific value of (x C, x E) are all randomly sampled from the unit ball B = {x 2 Rd : kxk2 1}. Since ( E) and (x C, x E) are both sampled from B, mf and Mf can be obtained by taking the minimum and maximum of (x T E E) on B, respectively, i.e., mf = 1 (1+e)2 and Mf = 1 (1+e 1)2 . At each time t, an arm set At is randomly sampled from A such that | At| = 10, i.e., each time we offer 10 randomly selected arms for the algorithm to choose from. An algorithm selects an arm from At and observes the corresponding reward Calg t generated by the Bernoulli distribution B( (x T E)). The regret of this algorithm at time t is defined by its received click reward, i.e., regret(t) = Ca t Cat, where a t is the optimum arm to be chosen based on the ground-truth bandit parameters ( We repeat the experiment 100 times using the same simulation setting, each containing 10,000 iterations. The average cumulative regret over 100 runs and the corresponding standard deviation (plotted per thousand iterations) are illustrated in Figure 1. One can clearly notice that the Logistic Bandit suffers from a linear regret with respect to time t, as it mistakenly treats no click as negative feedback. Our E-C Bandit achieves a fast converging sub-linear regret. The result that h Lin UCB performs the worst is expected, since it assumes a linear relation between click and context feature vectors. We further investigate how the aggregate empirical discrepancy between E-C bandit and Figure 1: Comparison of cumulative regret over 100 runs of simulation. Figure 2: Comparison of discrepancy bound provided by Proposition 1. Logistic Bandit increases with respect to time. Figure 2 illustrates that the aggregate empirical discrepancy of E-C Bandit is well bounded by the upper bound provided by Proposition 1, while the Logistic Bandit s aggregate empirical discrepancy increases linearly. This directly explains their accumulative regret in this experiment comparison. 6.3 Experiments on MOOC video watching data The MOOC data we used for evaluation is collected from a single course in a 4-month period. The course has 503 lecture videos in total. About 500 high-quality quiz-like questions have been manually crafted and each video is assigned with a subset of them based on human-judged relatedness. We selected 21 videos, whose accumulated watching time are ranked in the first 21 places, for this evaluation. Over the selected videos, on average a video is assigned with 5.5 questions, each of which is associated with 6 possible displaying positions within the video, leading to a total of 33 arms in average (as each question can be placed in all positions). The data set with our manually crafted features and our model implementation have been made publicially available here: https://github.com/qy7171/ec_bandit. We picked one video as an example to analyze students click behavior. 9 arms are picked and projected by a random Gaussian matrix to a two-dimension plane in Figure 3. Thus, their relative distance are kept. The number in the parenthesis indicates the empirical CTR of the corresponding arm. It can be clearly seen that while arm c and arm f have the same empirical CTR, the arms between them, such as arm a and d, have lower CTRs. Logistic Bandit is never able to capture this non-monotonicity relation, since its reward prediction increases monotonically with respect to a linear predictor. We construct a more general case in Appendix F to illustrate the scenario that failing to model examination condition would lead to a linear regret. Mapping the illustration back to the MOOC data set, arm a and arm f are two different questions displayed at the same position in the video, while arm a and arm c are the same question displayed at different positions. This phenomenon strongly suggests bias in users implicit feedback, which again justifies our decomposition of examination and relevance in click feedback. We followed [21] to develop our online data collection policy in our MOOC platform so as to prepare our offline evaluation data set. In particular, any related questions with respect to a video will have an equal probability to be selected and displayed at all positions in this video. We name this policy as Similarity. We create an instance of a bandit model for each video to learn its own optimal question placing policy. See Appendix G for detailed explanations of our examination feature choice. We also added a new baseline here, i.e., PBMUCB[18], which assumes a position-based examination probability in any ranking result. To adjust it to our setting, we assumed that the examination probability of any question chosen in a video is determined by and only by its position. Therefore, the key difference between our model and PBMUCB is that ours utilizes the available contextual information to estimate the examination probability, while PBMUCB is context-free. Yet, another important difference is that PBMUCB assumes the probability of examination at different position is known, and is estimated from offline data. Figure 3: 9 arms feature vectors projected onto a two-dimension plane, such that the relative distances between points are kept. The number in the parenthesis is the arm s empirical CTR. Figure 4: Performance comparison on MOOC videos data of different bandit algorithms. Li et al. [21] proposed a method to calculate a near-unbiased estimate of CTR of any bandit algorithm based on the collected history data, so that offline evaluation and performance comparison are possible. We take their offline evaluation protocol here and report the estimated CTR in Figure 4, which is averaged over 100 runs. To avoid disclosing any proprietary information about the platform, all algorithms CTRs are normalized by that from the Similarity policy. As shown in the figure, independently estimating E-C Bandits across videos achieves an average 40.6% increase in CTR over the Similarity baseline. Meanwhile, E-C Bandit consistently outperforms the other three baseline bandits, i.e., h Lin UCB, Logistic Bandit and PBMUCB. The improvement of our model compared to Logistic Bandit clearly suggests the necessity of modeling examination condition in user clicks for improving the online recommendation performance, and the improvement against PBMUCB provides strong evidence of the importance of modelling examination with available contextual information. In addition, the standard of error of E-C Bandit, h Lin UCB, Logistic Bandit and PBMUCB s relative CTR performance among 100 trials are 0.032, 0.031, 0.030, 0.041, respectively. Therefore, the variance of our offline evaluation is small and the improvement from our solution to the baselines are statistically significant. 7 Conclusion Motivated by the examination hypothesis in user click modeling, in this paper we developed E-C Bandit, which differentiates result examination and content relevance in user clicks and actively learns from such implicit feedback. We developed an efficient and effective learning algorithm based on variational inference and demonstrated its effectiveness on both simulated and real-world datasets. We proved that despite the complexity of underlying reward generation assumption and the resulting parameter estimation procedure, the proposed learning algorithm enjoys a sub-linear regret bound. Currently we only studied click feedback on single items; it is important for us to study it in a more general setting, e.g., a list of ranked items, where sequential result examination and relevance judgment introduce richer inter-dependency. In addition, our current regret analysis does not account for the additional discrepancy introduced by the variational inference. Abeille et al. [2] suggests that an exact posterior is not a necessary condition for a Thompson sampling policy to be optimal. It is important to study a tighter upper regret bound under our approximated posterior in general. Acknowledgements. We thank the anonymous reviewers for their insightful comments. This paper is based upon work supported by a research fund from Xuetang X.com and the National Science Foundation under grant IIS-1553568 and IIS-1618948. [1] Yasin Abbasi-yadkori, Dávid Pál, and Csaba Szepesvári. Improved algorithms for linear stochastic bandits. In NIPS, pages 2312 2320. 2011. [2] Marc Abeille, Alessandro Lazaric, et al. Linear thompson sampling revisited. Electronic Journal of Statistics, 11(2):5165 5197, 2017. [3] Shipra Agrawal and Navin Goyal. Thompson sampling for contextual bandits with linear payoffs. In International Conference on Machine Learning, pages 127 135, 2013. [4] Peter Auer. Using confidence bounds for exploitation-exploration trade-offs. Journal of Machine Learning Research, 3:397 422, 2002. URL http://www.jmlr.org/papers/v3/auer02a. html. [5] Djallel Bouneffouf, Amel Bouzeghoub, and Alda Lopes Gançarski. A contextual-bandit algorithm for mobile context-aware recommender system. In Neural Information Processing, pages 324 331. 2012. [6] Olivier Chapelle and Lihong Li. An empirical evaluation of thompson sampling. In Advances in neural information processing systems, pages 2249 2257, 2011. [7] Aleksandr Chuklin, Ilya Markov, and Maarten de Rijke. Click models for web search. Synthesis Lectures on Information Concepts, Retrieval, and Services, 7(3):1 115, 2015. [8] Nick Craswell, Onno Zoeter, Michael Taylor, and Bill Ramsey. An experimental comparison of click position-bias models. In Proceedings of the 2008 international conference on web search and data mining, pages 87 94. ACM, 2008. [9] Sarah Filippi, Olivier Cappe, Aurélien Garivier, and Csaba Szepesvári. Parametric bandits: The generalized linear case. In Advances in Neural Information Processing Systems, pages 586 594, 2010. [10] Fan Guo, Chao Liu, and Yi Min Wang. Efficient multiple-click models in web search. In Proceedings of the Second ACM International Conference on WSDM, pages 124 131. ACM, 2009. [11] Yifan Hu, Yehuda Koren, and Chris Volinsky. Collaborative filtering for implicit feedback datasets. In Data Mining, 2008. ICDM 08. Eighth IEEE International Conference on, pages 263 272. Ieee, 2008. [12] Tommi S Jaakkola and Michael I Jordan. Bayesian parameter estimation via variational methods. Statistics and Computing, 10(1):25 37, 2000. [13] Thorsten Joachims, Laura Granka, Bing Pan, Helene Hembrooke, and Geri Gay. Accurately interpreting clickthrough data as implicit feedback. In ACM SIGIR Forum, volume 51, pages 4 11. Acm, 2017. [14] Sumeet Katariya, Branislav Kveton, Csaba Szepesvari, and Zheng Wen. Dcm bandits: Learning to rank with multiple clicks. In International Conference on Machine Learning, pages 1215 1224, 2016. [15] Jaya Kawale, Hung H Bui, Branislav Kveton, Long Tran-Thanh, and Sanjay Chawla. Efficient thompson sampling for online matrix-factorization recommendation. In NIPS, pages 1297 1305, 2015. [16] Diane Kelly and Jaime Teevan. Implicit feedback for inferring user preference: a bibliography. In Acm Sigir Forum, volume 37, pages 18 28. ACM, 2003. [17] Branislav Kveton, Csaba Szepesvari, Zheng Wen, and Azin Ashkan. Cascading bandits: Learning to rank in the cascade model. In International Conference on Machine Learning, pages 767 776, 2015. [18] Paul Lagrée, Claire Vernade, and Olivier Cappe. Multiple-play bandits in the position-based model. In Advances in Neural Information Processing Systems, pages 1597 1605, 2016. [19] John Langford and Tong Zhang. The epoch-greedy algorithm for multi-armed bandits with side information. In NIPS, pages 817 824, 2008. [20] Lihong Li, Wei Chu, John Langford, and Robert E Schapire. A contextual-bandit approach to personalized news article recommendation. In Proceedings of the 19th international conference on World wide web, pages 661 670. ACM, 2010. [21] Lihong Li, Wei Chu, John Langford, and Xuanhui Wang. Unbiased offline evaluation of contextual-bandit-based news article recommendation algorithms. In Proceedings of the fourth ACM international conference on Web search and data mining, pages 297 306. ACM, 2011. [22] Wei Li, Xuerui Wang, Ruofei Zhang, Ying Cui, Jianchang Mao, and Rong Jin. Exploitation and exploration in a performance based contextual advertising system. In Proceedings of 16th SIGKDD, pages 27 36. ACM, 2010. [23] Odalric-Ambrym Maillard and Shie Mannor. Latent bandits. In International Conference on Machine Learning, pages 136 144, 2014. [24] Daniel Russo and Benjamin Van Roy. Learning to optimize via posterior sampling. Mathematics of Operations Research, 39(4):1221 1243, 2014. [25] Huazheng Wang, Qingyun Wu, and Hongning Wang. Learning hidden features for contextual bandits. In Proceedings of the 25th ACM International on Conference on Information and Knowledge Management, pages 1633 1642. ACM, 2016. [26] Yisong Yue and Carlos Guestrin. Linear submodular bandits and their application to diversified retrieval. In NIPS, pages 2483 2491, 2011. [27] Masrour Zoghi, Tomas Tunys, Mohammad Ghavamzadeh, Branislav Kveton, Csaba Szepesvari, and Zheng Wen. Online learning to rank in stochastic click models. In International Conference on Machine Learning, pages 4199 4208, 2017.