# locally_differentially_private_contextual_bandits_learning__59dffd50.pdf Locally Differentially Private (Contextual) Bandits Learning Kai Zheng1 zhengk92@gmail.com Tianle Cai2,3 caitianle1998@pku.edu.cn Weiran Huang4 weiran.huang@outlook.com Zhenguo Li4 li.zhenguo@huawei.com Liwei Wang5,6, wanglw@cis.pku.edu.cn We study locally differentially private (LDP) bandits learning in this paper. First, we propose simple black-box reduction frameworks that can solve a large family of context-free bandits learning problems with LDP guarantee. Based on our frameworks, we can improve previous best results for private bandits learning with one-point feedback, such as private Bandits Convex Optimization, and obtain the first result for Bandits Convex Optimization (BCO) with multi-point feedback under LDP. LDP guarantee and black-box nature make our frameworks more attractive in real applications compared with previous specifically designed and relatively weaker differentially private (DP) context-free bandits algorithms. Further, we extend our (ε, δ)-LDP algorithm to Generalized Linear Bandits, which enjoys a sub-linear regret O(T 3/4/ε) and is conjectured to be nearly optimal. Note that given the existing Ω(T) lower bound for DP contextual linear bandits [35], our result shows a fundamental difference between LDP and DP contextual bandits learning. 1 Introduction As a general and powerful model, (contextual) bandits learning has attracted lots of attentions both in theoretical study and real applications [8, 28], from personalized recommendation to clinical trails. However, existing algorithms designed for these applications heavily rely on user s sensitive data, and an off-the-shelf use of such algorithms may leak user s privacy and bring concerns to future users for sharing their data with related institutions or corporations. For example, in classification or regression tasks, we update our model according to the feature and label of each user. In Multi-Armed Bandits (MAB), we estimate underlying rewards of all arms based on user s feedback. A solid notion of data privacy is Differential Privacy (DP) proposed by Dwork et al. [13] in 2006. Since then, differentially private bandits learning has been studied extensively. Among context-free bandits learning, Bandits Convex Optimization (BCO) is one of the fundamental problems. Thakurta and Smith [37] designed the first (ε, δ)-differentially private adversarial BCO 1 Kwai Inc. 2 School of Mathematical Sciences, Peking University 3 Haihua Institute for Frontier Information Technology 4 Huawei Noah s Ark Lab 5 Key Laboratory of Machine Perception, MOE, School of EECS, Peking University 6 Center for Data Science, Peking University Corresponding author. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. Type Problem Our Regret Bound Best Non-Private Regret Context-Free Convex O T 3/4/ε O T 3/4 [16] Convex + Smooth O T 2/3/ε O T 2/3 [32] S.C O T 2/3/ε O T 2/3 [3] S.C + Smooth O T 1/2/ε O T 1/2 [18] MP-BCO Convex O T 1/2/ε2 O T 1/2 [3] Strongly Convex O log T/ε2 O (log T) [3] Context-Based Contextual Linear Bandits O(T 3/4/ε) O(T 1/2) [1] Generalize Linear Bandits O(T 3/4/ε) O(T 1/2) [30] Table 1: Summary of our main results under (ε, δ)-LDP, where O notation hides dependence over dimension d and other poly-logarithmic factors. (S.C means Strongly Convex, MP means Multi-Point) algorithm with O T 3/4/ε regret for convex loss and O T 2/3/ε regret for strongly convex loss, which nearly match current best non-private results under the same conditions [3, 16]1. However, when loss functions are further smooth, current best non-private bounds for convex/strongly convex bandits are O T 2/3 [32] and O T 1/2 [18] respectively, and previous approaches [37, 5] seem hard to achieve such regret bounds in the same setting under privacy constraint (see Section 3.1 for more discussions). Besides BCO and its extension to multi-point feedback [3], context-free bandits also include other important cases, such as Multi-Armed Bandits (MAB), and there have been lots of algorithms designed for differentially private MAB [37, 31, 38, 39, 5, 33], either in stochastic or adversarial environment. As one can see, there are many different settings in context-free bandits learning, and existing differentially private algorithms are carefully designed for each one of them, which makes them relatively inconvenient to be used. Besides, their theoretical performance is analyzed separately and rather complicated. Some of them do not match corresponding non-private results. Different with context-free bandits, usually there are certain contexts in real applications, such as user profile that contains user s features. Advanced bandit model uses these contexts explicitly to find the corresponding best action at each round, which is called contextual bandits. Two representatives are contextual linear bandits [29] and Generalized Linear Bandits [15]. Given benefits of contextual bandits, one may also wish to design corresponding private mechanisms. However, Shariff and Sheffet [35] proved that any differentially private contextual bandit algorithm would cause an Ω(T) regret bound. Hence, they considered a relaxed definition of DP called joint differential privacy, and proposed an algorithm based on Lin UCB [1] with regret bound O T 1/2/ε [35] under ε-joint differential privacy. Note all of previous study focus on differential privacy or its relaxed version. Compared with Differential Privacy, most of time Local Differential Privacy (LDP) [26, 11] is a much stronger and user-friendly standard of privacy and is more appealing in real applications [10], as LDP requires protecting each user s data before collection. For context-free bandits, it is not hard to see algorithms with LDP guarantee protects DP automatically. However in contextual bandits, things become more delicate. These two definitions are not comparable as they have different interpretations about the output sequence, and traditional post-processing property cannot be used here to imply LDP is more rigorous than DP. In detail, DP regards predicted actions for contexts as the output sequence. Since optimal action varies from round to round in contextual bandits, it is not surprising there is a lower bound of linear regret in this case [35], as DP requires outputs to be nearly the same for any two neighboring datasets/contexts, which essentially contradicts with the goal of personalized prediction in contextual bandits. In contrast, LDP regards the collected information from users as output sequence and has no restriction on predicted actions, which is more reasonable as these actions are predicted on the local side based on local personal information and will not be released to public. Therefore, LDP seems like a more appropriate standard 1Though Bubeck et al. [9] designed a polynomial time algorithm for general BCO with O(T 1/2) regret, it is far from practical, so we don t consider its result in this paper, but of course we can plug that algorithm into our framework to obtain optimal O(T 1/2/ε) bound for general private BCO. for contextual bandits compared with DP, and maybe there is hope to bypass the lower bound proved for DP contextual bandits. Given above discussions, a natural question arises: can we design simple and effective algorithms for bandits learning with LDP guarantee? Our Contributions: In this work, we study both context-free bandits2 and contextual bandits with LDP guarantee. Our contributions are summarized as follows: (see Table 1 for more details) (1) We propose a simple reduction framework motivated by Agarwal and Singh [5] for a large class of context-free bandits learning problems with LDP guarantee, including BCO, MAB and Best Arm Identification (see Section 3.1 and Appendix3 B). Equipped with different non-private algorithms, the utility of our framework can match corresponding best non-private performances, and these results are obtained through a unified and simple analysis; (2) By modifying above framework slightly, we extend our algorithm to BCO with multi-point feedback [3], and design the first LDP multi-point BCO algorithm with nearly optimal guarantees; (3) For contextual bandits including contextual linear bandits and more difficult generalized linear bandits, we propose algorithms with regret bounds O(T 3/4/ε) under (ε, δ)-LDP , which are conjectured to be optimal. Note that these results show a fundamental difference between LDP and DP contextual bandits as discussed above. All our results can be extended in parallel to ε-LDP if using Laplacian noise instead of Gaussian noise. Here, we only focus on (ε, δ)-LDP. Comparison with Prior Work: As mentioned earlier, for context-free bandits, nearly all of previous work focused on differentially private bandits learning, rather than stronger LDP guarantee. Only algorithms proposed in Tossou and Dimitrakakis [39] and Agarwal and Singh [5] for adversarial MAB can be converted to LDP version easily and obtain almost the same results. Though both their algorithms and ours are nearly the same in MAB, which is a very special case of bandits learning, our analysis is different, and we prove a new result for MAB with LDP guarantee as a side-product, which achieves nearly optimal regret bound under both adversarial and stochastic environment simultaneously (Appendix B.1). What s more, our results apply to more general bandits learning. For more comparison with Agarwal and Singh [5], see Section 3.1. Note, even in stronger LDP context-free bandits, our framework can achieve improved regret bounds for smooth BCO compared with previous results under weaker DP guarantee [37]. Besides, to the best of our knowledge, we give the first results for contextual bandits under LDP. 2 Preliminaries Notations: [p] = {1, 2, , p}. d is the dimension of decision space, and ei represents i-th basis vector. For a vector x and a matrix M, define x M := x Mx. Given a set W, we define the projection into this set as ΠW( ). Suppose the server collects certain information from each user with data domain C. C can be the range of loss values in context-free bandits, or both contexts and losses/rewards in contextual bandits. Now we define LDP rigorously: Definition 1 (LDP). A mechanism Q : C Z is said to protect (ε, δ)-LDP, if for any two data x, x C, and any (measurable) subset U Z, there is Pr[Q(x) U] eε Pr[Q(x ) U] + δ In particular, if Q preserves (ε, 0)-LDP, we call it ε-LDP. Now, we introduce a basic mechanism in LDP literature Gaussian Mechanism. Given any function h : C Rd. Define := maxx,x C h(x) h(x ) 2, then Gaussian Mechanism is defined as h(x) + Y , where random vector Y is sampled from Gaussian distribution N(0, σ2Id) with 2 ln(1.25/δ) ε . One can prove Gaussian Mechanism preserves (ε, δ)-LDP [12]. 2Note that adaptive adversary is ambiguous in bandits setting [6], so we only consider oblivious adversary throughout the paper. 3Appendix could be found in the full version [43]. Algorithm 1: One-Point Bandits Learning-LDP 1 Input: non-private algorithm A, privacy parameters ε, δ 2 Initialize: set σ = 2B 2 ln(1.25/δ) ε 3 for t = 1, 2, . . . do 4 Server plays xt X returned by A; 5 User t suffers loss ft(xt) and sends ft(xt) + Zt to A in the server, where Zt N(0, σ2); 6 A receives ft(xt) + Zt and calculates xt+1 Next, we define the common strong convexity and smoothness for a function f. Definition 2. We say that a function f : X R is µ-strongly convex if there is: f(x) f(y) f(x) (x y) µ 2 x y 2 2. We say that a function f : X R is β-smooth if it satisfies the following inequality: f(x) f(y) f(y) (x y) β 3 Nearly Optimal Context-Free Bandits Learning with LDP Guarantee In this section, we consider private context-free bandits learning with LDP guarantee, including bandits with one-point and multi-point feedback. As the following theorem shows, LDP is much stronger than DP in this setting (see Appendix A for the definition of DP in streaming setting and the proof), therefore it is more difficult to design algorithms under LDP with nearly optimal guarantee. Theorem 1. If an algorithm A protects ε-LDP, then any algorithm based on the output of A on a sequence of users guarantees ε-DP in streaming setting. 3.1 Private Bandits Learning with One-Point Feedback Bandits learning with one-point feedback includes several important cases, such as BCO, MAB, and Best Arm Identification (BAI). Generally speaking, we need to choose an action in the decision set at each round based on all previous information, then receive corresponding loss value of the action we choose. Most of time, our goal is to design an algorithm to minimize regret (it will be defined clearly later) compared with any fixed competitor. Different with previous work [37, 31, 38, 39, 33], which designed delicate algorithms for different bandit learning problems under DP, here we propose a general framework to solve all of them within a unified analysis under stronger LDP. Our general private framework is shown in Algorithm 1, based on a pre-chosen non-private black-box bandits learning algorithm A. Definitions of X, ft and the choice of A in Algorithm 1 will be made clear in concrete settings below. Here we only assume all ft(x) are bounded by a constant B, i.e., x X, t [T], |ft(x)| B. For private linear bandits learning, Agarwal and Singh [5] also propose a general reduction framework that can achieve nearly optimal regret. The key idea is to inject a linear perturbation nt, xt to the observed value ft(xt) at each round, where xt is the current decision strategy and nt is fresh noise vector sampled from a predefined distribution. Because of the special form of linear loss, their approach actually protects data sequence in the functional sense, i.e., it is equivalent to disturbing original linear loss function ft(x) with noisy function n t x. However, this approach cannot protect privacy when loss functions are nonlinear, as injected noise depends on strategy xt. Just consider xt = 0, then it may leak the information of ft as values of different nonlinear functions can be different at point xt = 0 and there is no noise at all if we use perturbation nt, xt . Instead, our main idea is to inject fresh noise variable directly to the observed loss value at each round, which doesn t rely on xt any more. Intuitively, this approaches looks more natural as bandits learning algorithms only use the information of these observed loss values instead of loss functions. Obviously, the LDP guarantee of Algorithm 1 is followed directly from basic Gaussian mechanism. Theorem 2. Algorithm 1 guarantees (ε, δ)-LDP. To show the power of Algorithm 1, here we consider its main application, Bandits Convex Optimization. For another two concrete applications, MAB and BAI, see Appendix B for more details. Besides, it also looks promising to extend the technique to pure exploration in combinatorial bandits (e.g., [21]). In bandit convex optimization [20], X is a bounded convex constraint set. At each round, the server chooses a prediction xt based on previous collected information, then suffers and observers a loss value ft(xt). The goal is to design an algorithm with low regret defined as maxx X E[PT t=1 ft(xt) ft(x)]. There are two different environments which generate underlying loss function sequence {ft(x)|t [T]}. For adversarial BCO, there is no further assumption about {ft(x)|t [T]} and they are fixed functions given before games starts. For stochastic BCO [4], feedback ft(xt) is generated as f(xt) + qt, where f(x) is an unknown convex function and {qt} are independently and identically distributed noise sampled from a sub-Gaussian distribution Q with mean 0. A critical ingredient in BCO is the gradient estimator constructed through the observed feedback. Besides convexity, when ft have additional properties like smoothness or strong convexity, usually we need to construct different gradient estimators and use different efficient non-private algorithms A to achieve better performance [16, 3, 32, 18]. Denote ut as a uniform random vector sampled from the unit sphere, then two representatives of gradient estimators are sphere sampling estimator d ρft(xt)ut used in [16, 3] (ρ is a parameter), and advanced ellipsoidal sampling estimator dft(xt)A 1 t ut which is the key part in [32, 18] to further improve the performance, where At is the Hessian matrix induced by certain loss function with self-concordant barrier. When it comes to private setting, Thakurta and Smith [37] designed a delicate differentially private algorithm with O T 3/4/ε and O T 2/3/ε guarantees for convex and strongly convex loss functions respectively, based on classical sphere sampling estimator and tree-based aggregation technique [14]. To achieve better bounds under additional smoothness assumption, it seems natural to combine their method with advanced ellipsoidal sampling estimator. However, this approach doesn t work even under DP guarantee, let alone LDP guarantee. In detail, to protect privacy, usually we need to add noise proportional to the range of information we use. For classical sphere sampling estimator, it is bounded by d B/ρ. However, for the advanced ellipsoidal sampling estimator, the spectral norm of inverse Hessian of self-concordant barrier (i.e., A 1 t ) can be unbounded, which makes it hard to protect privacy. Besides, tree-based aggregation techniques fail in LDP setting. Instead of adding noise to the accumulated estimated gradient like Thakurta and Smith [37], our general reduction Algorithm 1 injects noise directly to the loss value that is already bounded. Based on the critical observation that the regret defined for original loss functions {ft(x)|t [T]} equals to the regret defined for virtual loss functions {ft(x) + Zt|t [T]} in expectation, we avoid complex analysis which is based on a connection with non-private solutions [37], and obtain the utility of our private algorithm through the guarantee of non-private algorithm A directly as the following shows: Theorem 3. Suppose non-private algorithm A achieves regret B Reg T A for BCO, where B is the range of loss function. We have the following guarantee for Algorithm 1: for any x X, there is t=1 ft(xt) ft(x) O B ln(T/δ) where expectation is taken over the randomness of non-private algorithm A and all injected noise.4 With above theorem, by plugging different non-private optimal algorithms under variant cases, we obtain corresponding regret bounds with LDP guarantee: Corollary 4. When loss functions are convex and β-smooth, Algorithm 1 achieves O(T 2/3/ε) regret by setting A as Algorithm 1 in [32]. When loss functions are µ-strongly convex and β-smooth, Algorithm 1 achieves O( T/ε) regret by setting A as Algorithm 1 in [18]. For private Stochastic BCO, using Algorithm 2 in [4] as the black-box algorithm will achieve O( T/ε) regret. Note this result improves previous result [37] in three aspects. First, our Algorithm 1 guarantees stronger LDP rather than DP. Second, it achieves better regret bounds when loss functions are further smooth, and matches corresponding non-private results. Third, our algorithm is easy to be implemented, admits a unified analysis, and also obtains new results in stochastic BCO. 4Actually, if using the high probability guarantee of black-box algorithm A, we can also obtain corresponding high probability guarantee of our Algorithm 1. See Appendix E for more details, and the same argument there can be extended to results in section 3.2 as well. Algorithm 2: Two-Point Feedback Private Bandit Convex Optimization via Black-box Reduction 1 Input: set A as Algorithm 4 (in Appendix C) with parameters η, ρ, ξ, privacy parameters ε, δ 2 Initialize: set σ = 2G 2 ln(1.25/δ) T , ρ = log T 3 for t = 1, 2, . . . do 4 Server plays xt,1, xt,2 X received from A 5 User suffers ft(xt,1), ft(xt,2) and passes ft(xt,1) ft(xt,2) + n t (xt,1 xt,2) to A in the server, where nt N(0, σ2Id) 3.2 Private Bandits Convex Optimization with Multi-Point Feedback Now we consider BCO with Multi-Point Feedback. Different with one-point bandit feedback setting, where we can only query one point at each round, now we can query multiple points. This is natural in many applications, such as in personalized recommendation, we can recommend multiple items to each user and receive their feedback. Suppose we are permitted to query K points per round (denote them as xt,1, . . . , xt,K at round t), then we observe ft(xt,1), . . . , ft(xt,K). Suppose decision set X satisfies r B X RB like in Agarwal et al. [3], where B is the unit ball in Rd. The expected regret is defined as k=1 ft(xt,k) where {ft(x)} are G-Lipschitz convex functions, and expectation is taken over the randomness of algorithm. With the relaxation of amount about queries, there is a significant difference about regret bound of BCO between one-point feedback and K-point feedback for K 2 [3]. In detail, the minimax regret for general BCO with one-point feedback is in order O( T) (even for strongly convex and smooth losses [34]), whereas one can design algorithms for BCO under multi-point feedback with O( T) regret for convex loss and O(log T) regret for strongly convex loss, just like full information online convex optimization. As there is not much difference between K = 2 and K > 2, so we focus on K = 2 in this paper. An optimal non-private algorithm can be found in [3] and is given as Algorithm 4 in Appendix C for completion, which will be used as our black-box algorithm later. For private version of this problem, note our previous reduction framework no longer fits in this new setting, mainly because of multiple feedback. If we add the same noise Zt to observed values ft(xt,1), ft(xt,2), then it cannot guarantee privacy. If we use different noise Zt,1, Zt,2 to perturb observed values respectively, though it protects privacy, previous utility analysis fails. Based on the non-private algorithm, we design a slightly modified reduction framework that resembles the approach in Agarwal and Singh [5] but for Multi-Point BCO, as shown in Algorithm 2. The key observation is that now we play two pretty close points xt,1, xt,2 at each round, and critical information we use about user t is only the difference ft(xt,1) ft(xt,2) of two observed values. Note xt,1 xt,2 = 2ρut (see Algorithm 4 in Appendix C), which implies we can add noise n t (xt,1 xt,2) to ft(xt,1) ft(xt,2) to protect its privacy. As ft(x) is G-Lipschitz, hence |ft(xt,1) ft(xt,2)| 2ρG ut 2 and adding Gaussian noise with standard deviation σ = 2G 2 ln(1.25/δ) ε is enough to protect privacy as ut 2 = 1. Theorem 5. Algorithm 2 guarantees (ε, δ)-LDP. For utility analysis of Algorithm 2, as now the noise depends on strategies xt,1, xt,2 at round t, hence both output and regret in terms of original loss functions {ft(x)|t [T]} are the same as output and regret in terms of virtual loss functions {ft(x) + n t x|t [T]} in expectation. Therefore we can obtain the utility of our private Algorithm 2 through the guarantee of non-private algorithm A: Theorem 6. For any x X, Algorithm 2 guarantees t=1 (ft(xt,1) + ft(xt,2)) ft(x) If {ft} are further µ strongly convex, set η = 1 µt, ρ = log T r , then for any x X, we have t=1 (ft(xt,1) + ft(xt,2)) ft(x) From above results, one can see there is also a significant difference about regret bounds between BCO and Multi-Point BCO under LDP setting, which is exactly the same as non-private settings. 4 Contextual Bandits Learning with LDP Guarantee In this section, we turn our attention to more practical contextual bandits learning. At each round t, the learner needs to choose an action xt Xt in the local side, where Xt contains the personal information and features about underlying arms. Then the user generates a reward which is assumed to be yt = g(x t θ ) + ηt, where θ is an unknown true parameter in the domain W , g : R R is a known function, and ηt is a random noise in [ 1, 1] with mean 0 5. If we know θ , xt, := argmaxx Xt g(x θ ) is apparently the optimal choice at round t. For an algorithm A, we define its regret over T rounds as Reg A T := PT t=1 g(x t, θ ) g(x t θ ), where {xt, t [T]} is the output of A. We omit the superscript A when it is clear. There are two critical parts in contextual bandits. One is to estimate θ , and corresponding estimated parameter is used to find best action for exploitation. Another one is to construct certain term for the purpose of exploration, since we are in the environment of partial feedback. Throughout this section, we assume both {Xt} and W are bounded by a d-dimensional L2 ball with radius 1 for simplicity. Compared with private context-free bandits, private contextual bandits learning is more difficult, not only because of relatively complicated setting, but we need to protect more information including both contexts and rewards, which causes additional difficulty in the analysis of regret. As a warm-up, we show how to design algorithm with LDP guarantee for contextual linear bandits, which resembles a recent work [35] but under a relaxed version of DP. Next, we propose a more complicated algorithm for generalized linear bandits with LDP guarantee. 4.1 Warm-Up: LDP Contextual Linear Bandits In contextual linear bandits, mapping g is an identity, or equivalently, the reward generated by user t for action xt is yt = x t θ +ηt. To estimate θ , the straightforward method is to use linear regression based on collected data. Combined with classic principal for exploration, optimism in the face of uncertainty, it leads to Lin UCB [1], which is nearly optimal for contextual linear bandits. To protect privacy, it s not surprising that we adopt the same technique as LDP linear regression [36], i.e. injecting noise to xtx t and ytxt collected from user t. However, the injected noise have influence not only over the parameter estimation, but also for further exploration part, due to more complex bandit model, thus we need to set parameters more carefully. See Algorithm 5 in Appendix D. Now, we state the theoretical guarantee of Algorithm 5. Theorem 7. Algorithm 5 guarantees (ε, δ)-LDP. Theorem 8. With probability at least 1 α, the regret of Algorithm 5 satisfies the following bound: Given the Ω(T) lower bound for DP contextual linear bandits [35], Theorem 8 implies a fundamental difference between LDP and DP in contextual bandit learning, which also verifies that LDP is a more appropriate standard about privacy for contextual bandits as discussed in the introduction. One may think we can still prove DP based on LDP guarantee and post-processing property. Recall post-processing property holds only for the output of a DP algorithm which doesn t use private data any more. However, in our algorithms for LDP contextual bandits, though we can use post-processing property to prove estimation sequence { θt} satisfies DP, it doesn t imply the output action sequence {xt} satisfies DP, as these actions are made in the local side which use private local data. 5It s not hard to relax this constraint to a sub-Gaussian noise. Algorithm 3: Generalized Linear Bandits with LDP 1 Input: privacy parameters ε, δ, failure probability α 2 Initialize: V0 = 0d d, u0 = 0d, θ0 = ˆθ1 = 0d, ζ = Θ(1/ T), σ = 6 p 2 ln(3.75/δ)/ε 3 Notations: Υt = σ d + 2 ln(2T/α)), ct = 2Υt, β2 t = O( Cσ 4 for t = 1, 2, . . . do 5 For the local user t: 6 Receive information Vt 1, θt 1, ˆθt from the server 7 Play action xt = argmaxx Dt D θt 1, x E + βt 1 x V 1 t 1 8 Observe reward yt = g(x t θ ) + ηt, set zt = x t ˆθt. 9 Send xtx t + Bt, ztxt + ξt, ℓt(ˆθt) + rt to the server, where ℓt(θ) = ℓ(x t θ, yt), Bt(i, j) i.i.d N(0, σ2), i j, and B(j, i) = B(i, j), ξt N(0d, σ2Id d), rt N(0d, C2σ2Id d) 10 For the server: 11 Update Vt = Vt 1 + xtx t + Bt, ut = ut 1 + ztxt + ξt θt = V 1 t ut, where Vt = Vt + ct Id d ˆθt+1 = ΠW ˆθt ζ( ℓt(ˆθt) + rt) 4.2 LDP Generalized Linear Bandits In generalized linear bandits, mapping g can be regarded as the inverse link function of exponential family model. Here we suppose function g is G-Lipschitz, continuously differentiable on [ 1, 1], |g(a)| C, and infa ( 1,1) g (a) = µ > 0, which implies g is strictly increasing. These assumptions are common either in real applications or previous work [30, 25]. We also define corresponding negative log-likelihood function ℓ(a, b) := ab + m(a), where m( ) is the integral of function g. As a concrete example, if reward y is a Bernoulli random variable, then the form of g is g(a) = (1 + exp( a)) 1, m(a) = log(1 + exp(a)), ℓ(a, b) = log(1 + exp( a(2b 1))), b {0, 1}, and noise η is 1 g(a) with probability g(a) and g(a) otherwise. Note the non-linearity of g makes things much more complicated either from the view of bandits learning or privacy preservation. The counterpart of Contextual Linear Bandits is linear regression, the locally private version of which is relatively easy and well-studied. However, the counterpart of Generalized Linear Bandit is Empirical Risk Minimization (ERM) with respect to generalized linear loss, and the optimal approach of parameter estimation for GLM bandit is to solve ERM at each round [30]. Different with linear regression, for learning ERM with LDP guarantee, in general there is no efficient private algorithm that can achieve optimal performance in the non-interactive environment [36, 42, 41], let alone calculating an accurate parameter estimation needed in our problem. Therefore, it seems hard to learn generalized linear bandit under LDP guarantee. Luckily, we can make full use of the interactive environment in bandit problems. In detail, we build our private mechanism based on GLOC framework proposed in [25]. Compared with previous nearly optimal approach [30], GLOC framework enjoys much better time efficiency, which calculates estimator θ in an online fashion instead of solving ERM at each round. Its main idea is to maintain a rough estimation for unknown parameter θ through an adversarial online learning algorithm and use it to relabel current reward, and then solve the corresponding linear regression for a refined estimator. To achieve optimal O( T) regret, the online learning algorithm is set as Online Newton Step [19]. Though the original goal of GLOC framework proposed in Jun et al. [25] is to improve time efficiency, the update form of estimated parameter for unknown θ shares the same form of linear regression, therefore we can use nearly the same technique as in previous subsection to protect LDP, which avoids solving complex ERM with LDP guarantee. Besides, since internal online learning algorithm also utilizes users data, we also need to guarantee its privacy. Different with Jun et al. [25] which adopts Online Newton Step, we choose basic noisy Online Gradient Descent as our online black-box algorithm. See Algorithm 3 for the full implementation. For clarity, we just write the LDP Online Gradient Descent explicitly in Line 11. Though Algorithm 3 is based on the framework proposed by Jun et al. [25], we want to emphasize that both finding the right approach and proving the rigorous guarantee are non-trivial because of stringent LDP constraint. Following theorems give both the privacy guarantee and utility bound of our Algorithm 3 for generalized linear bandits. Theorem 9. Algorithm 3 guarantees (ε, δ)-LDP. Theorem 10. With probability at least 1 α, the regret of Algorithm 3 satisfies the following bound: Note that both our upper bounds (5) and (6) are in order O T 3/4 , which differ from common O( T) regret bound in corresponding non-private settings. We conjecture this order is nearly the best one can achieve in LDP setting, mainly because we need to protect more information, i.e., both contexts and corresponding rewards. See Appendix G for more discussions and intuitions. 5 Conclusions In this paper, we propose a simple black-box reduction framework that can solve a large class of context-free bandits learning problems with LDP guarantee in a unified way, including BCO, MAB, Best Arm Identification. We also extend the reduction framework to BCO with Multi-Point Feedback. This black-box reduction mainly has three advantages compared with previous work. First it guarantees a more rigorous LDP guarantee instead of DP. Second, this framework gives us a unified analysis for all above private bandit learning problems instead of analyzing each of them separately, and it easily improves previous best results or obtains new results for some problems, as well as matching corresponding non-private optimal bounds. Third, such a black-box reduction is more attractive in real applications, as we only need to modify the input to black-box algorithms. Besides, we also propose new algorithms for more practical contextual bandits with LDP guarantee, including contextual linear bandits and generalized linear bandits. Our algorithms can achieve O(T 3/4) regret bound, which is conjectured to be nearly optimal. We leave the rigorous proof of this lower bound as an interesting open problem. Broader Impact This work is mostly theoretical, with no negative outcomes. (Contextual) bandits learning has been widely used in real applications, which heavily relies on user s data that may contain personal private information. To protect user s privacy, we adopt the appealing solid notion of privacy Local Differential Privacy (LDP) that can protect each user s data before collection, and design (contextual) bandit algorithms under the guarantee of LDP. Our algorithms can be easily used in real applications, such as recommendation, advertising, to protect data privacy and ensure the utility of private algorithms simultaneously, which will befit everyone in the world. Acknowledgments and Disclosure of Funding This work was supported by National Key R&D Program of China (2018YFB1402600), Key-Area Research and Development Program of Guangdong Province (No. 2019B121204008), Beijing Academy of Artificial Intelligence, and in part by the Zhongguancun Haihua Institute for Frontier Information Technology. [1] Y. Abbasi-Yadkori, D. Pál, and C. Szepesvári. Improved algorithms for linear stochastic bandits. In Advances in Neural Information Processing Systems, pages 2312 2320, 2011. [2] Y. Abbasi-Yadkori, D. Pal, and C. Szepesvari. Online-to-confidence-set conversions and application to sparse stochastic bandits. In Artificial Intelligence and Statistics, pages 1 9, 2012. [3] A. Agarwal, O. Dekel, and L. Xiao. Optimal algorithms for online convex optimization with multi-point bandit feedback. In COLT, pages 28 40. Citeseer, 2010. [4] A. Agarwal, D. P. Foster, D. J. Hsu, S. M. Kakade, and A. Rakhlin. Stochastic convex optimization with bandit feedback. In Advances in Neural Information Processing Systems, pages 1035 1043, 2011. [5] N. Agarwal and K. Singh. The price of differential privacy for online learning. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 32 40. JMLR. org, 2017. [6] R. Arora, O. Dekel, and A. Tewari. Online bandit learning against an adaptive adversary: from regret to policy regret. In Proceedings of the 29th International Coference on International Conference on Machine Learning, pages 1747 1754. Omnipress, 2012. [7] D. Basu, C. Dimitrakakis, and A. Tossou. Differential privacy for multi-armed bandits: What is it and what is its cost? ar Xiv preprint ar Xiv:1905.12298, 2019. [8] S. Bubeck, N. Cesa-Bianchi, et al. Regret analysis of stochastic and nonstochastic multi-armed bandit problems. Foundations and Trends R in Machine Learning, 5(1):1 122, 2012. [9] S. Bubeck, Y. T. Lee, and R. Eldan. Kernel-based methods for bandit convex optimization. In Proceedings of the 49th Annual ACM SIGACT Symposium on Theory of Computing, pages 72 85. ACM, 2017. [10] G. Cormode, S. Jha, T. Kulkarni, N. Li, D. Srivastava, and T. Wang. Privacy at scale: Local differential privacy in practice. In Proceedings of the 2018 International Conference on Management of Data, pages 1655 1658, 2018. [11] J. Duchi, M. J. Wainwright, and M. I. Jordan. Local privacy and minimax bounds: Sharp rates for probability estimation. In Advances in Neural Information Processing Systems, pages 1529 1537, 2013. [12] C. Dwork and A. Roth. The algorithmic foundations of differential privacy. Foundations and Trends R in Theoretical Computer Science, 9(3 4):211 407, 2014. [13] C. Dwork, F. Mc Sherry, K. Nissim, and A. Smith. Calibrating noise to sensitivity in private data analysis. In Theory of cryptography, pages 265 284, Berlin, Germany, March 2006. Springer. [14] C. Dwork, M. Naor, T. Pitassi, and G. N. Rothblum. Differential privacy under continual observation. In Proceedings of the forty-second ACM symposium on Theory of computing, pages 715 724. ACM, 2010. [15] S. Filippi, O. Cappe, A. Garivier, and C. Szepesvári. Parametric bandits: The generalized linear case. In Advances in Neural Information Processing Systems, pages 586 594, 2010. [16] A. D. Flaxman, A. T. Kalai, and H. B. Mc Mahan. Online convex optimization in the bandit setting: gradient descent without a gradient. In Proceedings of the sixteenth annual ACM-SIAM symposium on Discrete algorithms, pages 385 394. Society for Industrial and Applied Mathematics, 2005. [17] A. Garivier and O. Cappé. The kl-ucb algorithm for bounded stochastic bandits and beyond. In Proceedings of the 24th annual Conference On Learning Theory, pages 359 376, 2011. [18] E. Hazan and K. Levy. Bandit convex optimization: Towards tight bounds. In Advances in Neural Information Processing Systems, pages 784 792, 2014. [19] E. Hazan, A. Agarwal, and S. Kale. Logarithmic regret algorithms for online convex optimization. Machine Learning, 69(2-3):169 192, 2007. [20] E. Hazan et al. Introduction to online convex optimization. Foundations and Trends R in Optimization, 2 (3-4):157 325, 2016. [21] W. Huang, J. Ok, L. Li, and W. Chen. Combinatorial pure exploration with continuous and separable reward functions and its applications. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, pages 2291 2297, 2018. [22] P. Jain, P. Kothari, and A. Thakurta. Differentially private online learning. In Conference on Learning Theory, pages 24 1, 2012. [23] K. Jamieson and R. Nowak. Best-arm identification algorithms for multi-armed bandits in the fixed confidence setting. In 2014 48th Annual Conference on Information Sciences and Systems (CISS), pages 1 6. IEEE, 2014. [24] K. Jamieson, M. Malloy, R. Nowak, and S. Bubeck. lil ucb: An optimal exploration algorithm for multi-armed bandits. In Conference on Learning Theory, pages 423 439, 2014. [25] K.-S. Jun, A. Bhargava, R. Nowak, and R. Willett. Scalable generalized linear bandits: Online computation and hashing. In Advances in Neural Information Processing Systems, pages 99 109, 2017. [26] S. P. Kasiviswanathan, H. K. Lee, K. Nissim, S. Raskhodnikova, and A. Smith. What can we learn privately? SIAM Journal on Computing, 40(3):793 826, 2011. [27] E. Kaufmann, O. Cappé, and A. Garivier. On the complexity of best-arm identification in multi-armed bandit models. The Journal of Machine Learning Research, 17(1):1 42, 2016. [28] T. Lattimore and C. Szepesvári. Bandit algorithms. preprint, 2018. [29] L. Li, W. Chu, J. Langford, and R. E. Schapire. A contextual-bandit approach to personalized news article recommendation. In Proceedings of the 19th international conference on World wide web, pages 661 670, 2010. [30] L. Li, Y. Lu, and D. Zhou. Provably optimal algorithms for generalized linear contextual bandits. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 2071 2080. JMLR. org, 2017. [31] N. Mishra and A. Thakurta. (nearly) optimal differentially private stochastic multi-arm bandits. In Proceedings of the Thirty-First Conference on Uncertainty in Artificial Intelligence, pages 592 601. AUAI Press, 2015. [32] A. Saha and A. Tewari. Improved regret guarantees for online smooth convex optimization with bandit feedback. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, pages 636 642, 2011. [33] T. Sajed and O. Sheffet. An optimal private stochastic-mab algorithm based on optimal private stopping rule. In International Conference on Machine Learning, pages 5579 5588, 2019. [34] O. Shamir. On the complexity of bandit and derivative-free stochastic convex optimization. In Conference on Learning Theory, pages 3 24, 2013. [35] R. Shariff and O. Sheffet. Differentially private contextual linear bandits. In Advances in Neural Information Processing Systems, pages 4296 4306, 2018. [36] A. Smith, A. Thakurta, and J. Upadhyay. Is interaction necessary for distributed private learning? In 2017 IEEE Symposium on Security and Privacy (SP), pages 58 77. IEEE, 2017. [37] A. G. Thakurta and A. Smith. (nearly) optimal algorithms for private online learning in full-information and bandit settings. In Advances in Neural Information Processing Systems, pages 2733 2741, 2013. [38] A. C. Tossou and C. Dimitrakakis. Algorithms for differentially private multi-armed bandits. In Thirtieth AAAI Conference on Artificial Intelligence, 2016. [39] A. C. Y. Tossou and C. Dimitrakakis. Achieving privacy in the adversarial multi-armed bandit. In Thirty-First AAAI Conference on Artificial Intelligence, 2017. [40] R. Vershynin. High-dimensional probability: An introduction with applications in data science, volume 47. Cambridge University Press, 2018. [41] D. Wang, M. Gaboardi, and J. Xu. Empirical risk minimization in non-interactive local differential privacy revisited. In Advances in Neural Information Processing Systems, pages 973 982, 2018. [42] K. Zheng, W. Mou, and L. Wang. Collect at once, use effectively: Making non-interactive locally private learning possible. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 4130 4139. JMLR. org, 2017. [43] K. Zheng, T. Cai, W. Huang, Z. Li, and L. Wang. Locally differentially private (contextual) bandits learning. ar Xiv preprint ar Xiv:2006.00701, 2020. [44] J. Zimmert and Y. Seldin. An optimal algorithm for stochastic and adversarial bandits. In Proceedings of Machine Learning Research, volume 89 of Proceedings of Machine Learning Research, pages 467 475. PMLR, 16 18 Apr 2019.