# optimistic_policy_optimization_with_bandit_feedback__d5d4c3da.pdf Optimistic Policy Optimization with Bandit Feedback Yonathan Efroni * 1 Lior Shani * 1 Aviv Rosenberg 2 Shie Mannor 1 Policy optimization methods are one of the most widely used classes of Reinforcement Learning (RL) algorithms. Yet, so far, such methods have been mostly analyzed from an optimization perspective, without addressing the problem of exploration, or by making strong assumptions on the interaction with the environment. In this paper we consider model-based RL in the tabular finite-horizon MDP setting with unknown transitions and bandit feedback. For this setting, we propose an optimistic policy optimization algorithm for which we establish O( S2AH4K) regret for stochastic rewards. Furthermore, we prove O( S2AH4K2/3) regret for adversarial rewards. Interestingly, this result matches previous bounds derived for the bandit feedback case, yet with known transitions. To the best of our knowledge, the two results are the first sub-linear regret bounds obtained for policy optimization algorithms with unknown transitions and bandit feedback. 1. Introduction Policy Optimization (PO) is among the most widely used methods in Reinforcement Learning (RL) (Peters & Schaal, 2006; 2008; Deisenroth & Rasmussen, 2011; Lillicrap et al., 2015; Levine et al., 2016; Gu et al., 2017). Unlike valuebased approaches, e.g., Q-learning, these types of methods directly optimize the policy by incrementally changing it. Furthermore, PO methods span wide variety of popular algorithms such as policy-gradient algorithms (Sutton et al., 2000), natural policy gradient (Kakade, 2002), trust region policy optimization (TRPO) (Schulman et al., 2015) and soft actor-critic (Haarnoja et al., 2018). *Equal contribution 1Technion - Israel Institute of Technology, Haifa, Israel 2Tel Aviv University, Tel Aviv, Israel. Correspondence to: Lior Shani , Yonathan Efroni . Proceedings of the 37 th International Conference on Machine Learning, Vienna, Austria, PMLR 119, 2020. Copyright 2020 by the author(s). Due to their popularity, there is a rich literature that provides different types of theoretical guarantees for different PO methods (Scherrer & Geist, 2014; Abbasi-Yadkori et al., 2019; Agarwal et al., 2019; Liu et al., 2019; Bhandari & Russo, 2019; Shani et al., 2019; Wei et al., 2019) for both the approximate and tabular settings. However, previous results, concerned with regret or PAC bounds for the RL setting when the model is unknown and only bandit feedback is given, provide guarantees which critically depend on concentrability coefficients (Kakade & Langford, 2002; Munos, 2003; Scherrer, 2014) or on a unichain MDP assumption (Abbasi-Yadkori et al., 2019). However, these coefficients might be infinite and are usually small only for highly stochastic domains, while the unichain assumption is often very restrictive. Recently, Cai et al. (2019) established an O( K) regret bound for an optimistic PO method in the case of an unknown model and assuming full-information feedback on adversarially chosen instantaneous costs, where K is the number of episodes seen by the agent. In this work, we eliminate the full-information assumption on the cost, as in most practical settings only bandit feedback on the cost is given, i.e., the cost is observed through interacting with the environment. Specifically, we establish regret bounds for an optimistic PO method in the case of an unknown model and bandit feedback on the instantaneous cost in two regimes: 1. For stochastic cost, we establish an O( S2AH4K) regret bound for a PO method (Section 6). 2. For adversarially chosen cost, we establish an O( S2AH4K2/3) regret bound for a PO method. The regret bound matches the O K2/3 upper bound obtained by Neu et al. (2010a) for PO methods which have an access to the true model and observe bandit adversarial cost feedback (Section 7). 2. Preliminaries Stochastic MDPs. A finite horizon stochastic Markov Decision Process (MDP) M is defined by a tuple (S, A, H, {ph}H h=1, {ch}H h=1), where S and A are finite state and action spaces with cardinality S and A, respectively, and H N is the horizon of the MDP. On time Optimistic Policy Optimization with Bandit Feedback 2 Table 1. Comparison of our bounds with several state-of-the-art bounds for policy-based RL and occupancy measure RL in tabular finite-horizon MDPs. The time complexity of the algorithms is per episode; S and A are the sizes of the state and action sets, respectively; H is the horizon of the MDP; K is the total number of episodes; Env. describes the environment of the algorithm: stochastic (Sto) or adversarial (Adv); Policy based describes if an algorithm is based on policy updates or on occupancy measure updates. Costs and model terms describes how optimism is used in the estimators: For costs, a bonus term (Bonus) or an importance sampling estimator (IS). For transition model: a bonus term (Bonus) or a confidence interval of models (CI); The update procedure describes how the optimization problem is solved, using a state-wise closed-form solution (Closed form), or by solving an optimization problem over the entire state-action space (Optimization). The algorithms proposed in this paper are highlighted in gray. The other algorithms are OMD-BP (Neu et al., 2010b), UC-O-REPS (Rosenberg & Mansour, 2019a), OPPO (Cai et al., 2019) and UOB-REPS (Jin et al., 2019). (**) represents the different setting of the average cost criterion. Algorithm Regret Env. Bandit Feedback Unknown Model Policy Based Costs Model Update Procedure S2AH4K) Sto. ! ! ! Bonus Bonus Closed form OMDP-BP(**) O(K2/3) Adv. ! % ! IS - Closed form UC-O-REPS O( S2AH4K) Adv. % ! % - CI Optimization OPPO O( S3A3H4K) Adv. % ! ! - Bonus Closed form UOB-REPS O( S2AH4K) Adv. ! ! % IS CI Optimization POMD O( S2AH4K2/3) Adv. ! ! ! IS CI Closed form step h, and state s, the agent performs an action a, transitions to the next state s according to a time-dependent transition function ph(s | s, a), and suffers a random cost Ch(s, a) [0, 1] drawn i.i.d from a distribution with expectation ch(s, a). A stochastic policy π : S [H] A is a mapping from states and time-step indices to a distribution over actions, i.e., A = π RA : P a π(a) = 1, π(a) 0 . The performance of a policy π when starting from state s at time h is measured by its value function, which is defined as V π h (s) = E h =h ch (sh , ah ) | sh = s, π, p where the expectation is with respect to the randomness of the transition function, the cost function and the policy. The Q-function of a policy given the state action pair (s, a) at time-step h is defined by Qπ h(s, a) = E h =h ch (sh , ah ) | sh = s, ah = a, π, p The two satisfy the following relation: Qπ h(s, a) = ch(s, a) + ph( | s, a)V π h+1, V π h (s) = Qπ h(s, ), πh( | s) , (2.3) with ph( |s, a)V =P s ph(s |s, a)V (s ) for V RS, and , is the dot product. An optimal policy π minimizes the value for all states s and time-steps h simultaneously (Puterman, 2014), and its corresponding optimal value is denoted by V h (s) = minπ V π h (s), for all h [H]. We consider an agent that repeatedly interacts with an MDP in a sequence of K episodes such that the starting state at the k-th episode, sk 1, is initialized by a fixed state s1 . The agent does not have access to the model, and the costs are received by bandit feedback, i.e., the agent only observes the costs of encountered state-action pairs. At the beginning of the k-th episode, the agent chooses a policy πk and samples a trajectory sk h, ak h, Ck h(sk h, ak h) H h=1 by interacting with the stochastic MDP using this policy, where (sk h, ak h) are the state and action at the h-th time-step of the k-th episode. The performance of the agent for stochastic MDPs is measured by its regret relatively to the value of the optimal policy, defined as Regret(K ) = PK k=1 V πk 1 (sk 1) V 1 (sk 1) for all K [K], and πk is the policy of the agent at the k-th episode. Adversarial MDPs. Unlike stochastic MDP, in adversarial MDP, we let the cost to be determined by an adversary at the beginning of every episode, whereas the transition function is fixed. Thus, we denote the MDP at the k-th episode by Mk = (S, A, H, {ph}H h=1, {ck h}H h=1). As in (2.1), (2.2), we define the value function and Q-function of a policy π at the k-th episode by V k,π h (s) = E h =h ck h (sh , ah ) | sh = s, π, p Qk,π h (s, a) = E h =h ck h (sh , ah ) | sh = s, ah = a, π, p for simplicity we fix the initial state, but the results hold when it is drawn from a fixed distribution. Optimistic Policy Optimization with Bandit Feedback 3 Notably, V k,π h and Qk,π h satisfy the relations in relation (2.3). We consider an agent which repeatedly interacts with an adversarial MDP in a sequence of K episodes. Each episode starts from a fixed initial state, sk 1 = s1. As in the stochastic case, at the beginning of the k-th episode, the agent chooses a policy πk and samples a trajectory sk h, ak h, ck h(sk h, ak h) H h=1 by interacting with the adversarial MDP. In this case, the performance of the agent is measured by its regret relatively to the value of the best policy in hindsight. The objective is to minimize Regret(K ) = maxπ PK k=1 V k,πk 1 (s1) V k,π 1 (s1) for all K [K]. Notations and Definitions. The filtration Fk includes all events (states, actions, and costs) until the end of the k-th episode, including the initial state of the k + 1 episode. We denote by nk h(s, a), the number of times that the agent has visited state-action pair (s, a) at the h-th step, and by Xk, the empirical average of a random variable X. Both quantities are based on experience gathered until the end of the kth episode and are Fk measurable. We also define the probability to visit the state-action pair (s, a) at the k-th episode at time-step h by wk h(s, a) = Pr sk h = s, ak h = a | sk 1, πk, p . Since πk is Fk 1 measurable, so is wk h(s, a). In what follows, we refer to wk h(s, a) as the state-action occupancy measure. Furthermore, we define the state visitation frequency of a policy π in state s given a transition model p as dπ h(s; p) = E[1{sh = s} | s1, π, p]. By the two definitions, it holds that wk h(s, a) = dπk h (s; p)πk h(a | s). We use O(X) to refer to a quantity that depends on X up to a poly-log expression of a quantity at most polynomial in S, A, K, H and δ 1. Similarly, represents up to numerical constans or poly-log factors. We define X Y := max{X, Y }. Mirror Descent. The mirror descent (MD) algorithm (Beck & Teboulle, 2003) is a proximal convex optimization method that minimizes a linear approximation of the objective together with a proximity term, defined in terms of a Bregman divergence between the old and new solution estimates. In our analysis we choose the Bregman divergence to be the Kullback Leibler (KL) divergence, d KL. If {fk}K k=1 is a sequence of convex functions fk : Rd R, and C is a constraints set, the k-th iterate of MD is the following: xk+1 arg min x C {t K fk(xk), x xk + d KL(x||xk)}, where t K is a stepsize. In our case, C is the unit simplex , and thus the optimization problem has a closed-form solution, i [d], xk+1(i) = xk(i) exp( t K ifk(xk)) P j xk(j) exp( t K jfk(xk)). The MD algorithm ensures Regret(K ) = PK k=1 f(xk) minx f(x) O( K) for all K [K]. 3. Related Work Approximate Policy Optimization: A large body of work addresses the convergence properties of policy optimization algorithms from an optimization perspective. In Kakade & Langford (2002), the authors analyzed the Conservative Policy Iteration (CPI) algorithm, an RL variant of the Frank-Wolfe algorithm (Scherrer & Geist, 2014; Vieillard et al., 2019), and showed it approximately converges to the global optimal solution. Recently, Liu et al. (2019) established the convergence of TRPO when neural networks are being used as the function approximators. Furthermore, Shani et al. (2019) showed that TRPO (Schulman et al., 2015) is in fact a natural RL adaptation of the MD algorithm, and established convergence guarantees. In (Agarwal et al., 2019), the authors obtained convergence results for policy gradient based algorithms. However, all of the aforementioned works rely on the strong assumption of a finite concentrability coefficient, i.e., maxπ,s,h dπ h (s; p)/dπ h(s; p) < . This assumption bypasses the need to address exploration (Kakade & Langford, 2002), and leads to global guarantees based on the local nature of the policy gradients (Scherrer & Geist, 2014). Mirror Descent in Adversarial Reinforcement Learning: There are two different methodologies for using MD updates in RL. The first and more practical one, is using MD-like updates directly on the policy. The second is based on optimizing over the space of state-action occupancy measures, that is, visitation frequencies for state-action pairs. An occupancy measure represents a policy implicitly. For convenience, previous results for regret minimization using MD approaches are summarized in Table 1. Following the policy optimization approach, and assuming bandit feedback and known dynamics, Neu et al. (2010b) (OMDP-BF) established O(K2/3) regret for the average reward criteria. Alternatively, by assuming full information on the reward functions, unknown dynamics and further assuming both the reward and transition dynamics are linear in some d-dimensional features, Cai et al. (2019) established O( d3H4K) regret for their OPPO algorithm. The tabular case is a specific setting of the latter for d = SA. Instead of directly optimizing the policy, Zimin & Neu (2013) proposed optimizing over the space of state-action occupancy measures with the Relative Entropy Policy Search Optimistic Policy Optimization with Bandit Feedback 4 (O-REPS) algorithm. The O-REPS algorithm implicitly learns a policy by solving an MD optimization problem on the primal linear programming formulation of the MDP (Altman, 1999; Neu et al., 2017). Considering full information and unknown transitions, Rosenberg & Mansour (2019b) obtained O( S2AH4K) regret for their UC-O-REPS algorithm. Later, Rosenberg & Mansour (2019a) extended the algorithm to bandit feedback and obtained a regret of O(K3/4). Recently, by considering an optimistically biased importance sampling estimator, Jin et al. (2019) established O( S2AH4K) for their UOB-REPS algorithm . The OREPS variants updates constitute solving a convex optimization problem with HS2A variables on each episode, instead of the closed form solution updates of the direct policy optimization variants. Value-based Regret Minimization in Episodic RL: As opposed to Policy-based methods, there is an extensive literature about regret minimization in episodic MDPs using value-based methods. The works of (Azar et al., 2017; Dann et al., 2017; Jin et al., 2018; Zanette & Brunskill, 2019; Efroni et al., 2019) use the optimism in face of uncertainty principle to achieve near-optimal regret bounds. Jin et al. (2018) also establish a lower bound of Ω( 4. Mirror Descent for MDPs Algorithm 1 POMD with Known Model Require: t K, π1 is the uniform policy. for k = 1, .., K do # Policy Evaluation for h = H, H 1, .., 1 do for s, a S A do Qπk h (s, a) = ch(s, a) + ph( | s, a)V πk h+1 end for end for # Policy Improvement for s, a, h S A [H] do πk+1 h (a|s)= πk h(a|s) exp( t KQ πk h (s,a)) P a πk h(a |s) exp( t KQ πk h (s,a )) end for end for The empirical success of TRPO (Schulman et al., 2015) and SAC (Haarnoja et al., 2018) had motivated recent study of MD-like update rules for solving MDPs (Geist et al., 2019) when the model of the environment is known. Although not explicitly discussed in (Geist et al., 2019), such an algorithm can also provide guarantees by similar proof technique Note that in Jin et al. (2019), the regret of UOB-REPS is O( S2AH2K). However, this is due to the loop-free assumption. To remove this assumption, one needs to multiply the number of states by a factor of H. for the case where the cost function is adversarially chosen on each episode. Policy Optimization by Mirror Descent (POMD) (see Algorithm 1) is conceptually similar to the Policy Iteration (PI) algorithm (Puterman, 2014). It alternates between two stages: (i) policy evaluation, and (ii) policy improvement. Furthermore, much alike PI, POMD updates its policy on the entire state space, given the evaluated Q-function. However, as oppose to PI, the policy improvement stage is soft . Instead of updating according to the greedy policy, the algorithm applies soft update that keeps the next policy close to the current one due to the KL-divergence term. Similarly to standard analysis of the MD algorithm, Geist et al. (2019) established O( K) bound on the regret of Algorithm 1. In the next sections, we apply the same approach to problems with unknown model and bandit feedback. 5. Extended Value Difference Lemma The analysis of both stochastic and adversarial cases is built upon a central lemma which we now review. The lemma is a variant of (Cai et al., 2019)[Lemma 4.2], which generalizes classical value difference lemmas. Rewriting it in the following form, enables us to establish our results (proof in Appendix D). Lemma 1 (Extended Value Difference). Let π, π be two policies, and M = (S, A, {ph}H h=1, {ch}H h=1) and M = (S, A, {p h}H h=1, {c h}H h=1) be two MDPs. Let ˆQπ,M h (s, a) be an approximation of the Q-function of policy π on the MDP M for all h, s, a, and let ˆV π,M h (s) = D ˆQπ,M h (s, ), πh( | s) E . Then, ˆV π,M 1 (s1) V π ,M 1 (s1) = h=1 E h D ˆQπ,M h (sh, ), πh( | sh) π h( | sh) E | s1, π , p i + h=1 E h ˆQπ,M h (sh,ah) c h(sh, ah) p h( |sh, ah) ˆV π,M h+1 |s1,π ,p i where V π ,M 1 is the value function of π in the MDP M . This lemma generalizes existing value difference lemmas. For example, in (Kearns & Singh, 2002; Dann et al., 2017) the term V π,M 1 (s) V π,M 1 (s) is analyzed, whereas in (Kakade & Langford, 2002) the term V π,M 1 (s) V π ,M 1 (s) is analyzed. In next sections, we will demonstrate how Lemma 1 results in a simple analysis of the POMD algorithm. Importantly, the resulting regret bounds do not depend on concentrability coefficients (see Section 3) nor on any other structural assumptions. Optimistic Policy Optimization with Bandit Feedback 5 6. Policy Optimization in Stochastic MDPs We are now ready to analyze the optimistic version of POMD for stochastic environments (see Algorithm 2). Instead of using the known model as in POMD, in Algorithm 2 we use the empirical model to estimate the Q-function of an empirical optimistic MDP, with the empirical transition function p and an optimistic cost function ˆc. The empirical transition function p and empirical cost function c are computed by averaging the observed transitions and costs, respectively, that is, pk h(s | s, a) = Pk k =1 1 sk h = s, ak h = a, sk h+1 = s Pk k =1 1 sk h = s, ak h = a 1 ck h(s, a) = Pk k =1 Ck h (s, a)1 sk h = s, ak h = a Pk k =1 1 sk h = s, ak h = a 1 , for every s, a, s , h, k. Algorithm 2 Optimistic POMD for Stochastic MDPs Require: t K, π1 is the uniform policy. for k = 1, ..., K do Rollout a trajectory by acting πk # Policy Evaluation s S, V k H+1(s) = 0 for h = H, .., 1 do for s, a S A do ˆck 1 h (s, a) = ck 1 h (s, a) bk 1 h (s, a), Eq. (6.1) Qk h(s, a)=ˆck 1 h (s, a)+ pk 1 h ( |s, a)V k h+1 Qk h(s, a) = max Qk h(s, a), 0 end for for s S do V k h (s) = Qk h(s, ), πk h( | s) end for end for # Policy Improvement for h, s, a [H] S A do πk+1 h (a|s)= πk h(a|s) exp( t KQk h(s,a)) P a πk h(a |s) exp( t KQk h(s,a )) end for Update counters and empirical model, nk, ck, pk The optimistic cost function ˆc is obtained by adding a bonus term which drives the algorithm to explore, i.e., ˆck 1 h (s, a) = ck 1 h (s, a) bk 1 h (s, a), and we set bk 1 h (s, a) = bc,k 1 h (s, a) + bp,k 1 h (s, a). (6.1) The two bonus terms compensate on the lack of knowledge of the true costs and transition model, and are bc,k 1 h (s, a) = O nk 1 h (s, a) bp,k 1 h (s, a) = O nk 1 h (s, a) The following theorem bounds the regret of Algorithm 2. A full proof is found in Appendix B.2. Theorem 1. For any K [K], setting t K = O H 1K 1/2 the regret of Algorithm 2 is bounded by Regret(K ) O Proof Sketch. We start by decomposing the regret into three terms according to Lemma 1, and then bound each term separately to get our final regret bound. For any π, Regret(K ) = k=1 V πk 1 (sk 1) V π 1 (sk 1) k=1 V πk 1 (sk 1) V k 1 (sk 1) + k=1 V k 1 (sk 1) V π 1 (sk 1) k V πk 1 (s1) V k 1 (s1) k,h E Qk h(sh, ), πk h( | sh) πh( | sh) | s1, π, p | {z } (ii) k,h E Qk h(sh,ah) ch(sh,ah) ph( |sh,ah)V k h+1 |s1,π,p | {z } (iii) Term (i): Bias of V k. Term (i) is the bias between the estimated and true value of πk, V k and V πk, respectively. Applying Lemma 1, while using E[X(sh, ah) | s1, πk, p] = E X(sk h, ak h) | Fk 1 for any Fk 1-measurable function X RS A, we bound Term (i) by X k,h E ck 1 h (sk h,ak h)+H pk 1 h ( |sk h, ak h) 1 |Fk 1 k,h E h bc,k 1 h (sk h, ak h) + bp,k 1 h (sk h, ak h) | Fk 1 i . Here ck 1 h (s, a) = ch(s, a) ck 1 h (s, a) and pk 1 h ( | s, a) = ph( | s, a) pk 1 h ( | s, a), are the differences Optimistic Policy Optimization with Bandit Feedback 6 between the true cost and transition model to the empirical cost and transition model. Applying Hoeffding s bound and L1 deviation bound (Weissman et al., 2003) we get that w.h.p. for any s, a nk 1 h (s, a) = br h(s, a), ph( | s, a) 1 O nk 1 h (s, a) = bp h(s, a). Thus, w.h.p., we get nk 1 h (sk h, ak h) | Fk 1 which can be bounded by O S2AH4K using standard techniques (e.g., Dann et al. (2017)). Term (ii): OMD Analysis. Term (ii) is the linear approximation used in MD optimization procedure. We bound it using an analysis of OMD. By applying usual OMD analysis (see Lemma 16) we have that for any policy π and s, h, Qk h( | s), πk h( | s) πh( | s) a πk h(a | s)(Qk h(s, a))2. We plug this back to Term (ii) and use the fact that 0 Qk h(s, a) H, to obtain Term (ii) = Qk h(sh, ), πk h( |sh) πh( |sh) | s1, π, p t K + t KH3K By choosing t K = p 2 log A/(H2K), we obtain Term (ii) p 2H4K log A. Term (iii): Optimism. We choose our exploration bonuses in Eq. (6.2) such that Term (iii) is non-positive. Specifically, we choose the bonus such that Qk h(s,a) ch(s,a) ph( |s,a)V k h+1 0 for any s, a, which implies that Term(iii) 0. Remark 6.1. The choice of the bonus term bp,k h (s, a) is smaller than in (Cai et al., 2019) by a factor of translates to an improved regret bound by this factor, although (Cai et al., 2019) assumes full-information feedback on the cost function. Remark 6.2 (Bonus vs. Optimistic Model). Instead of using the additive exploration bonus bp which compensate on the lack of knowledge of transition model one can use an optimistic model approach, as in UCRL2 (Jaksch et al., 2010). Following analogous analysis as of Theorem 1 one can establish the same guarantee O( S2AH4K). However, the additive bonus approach results in an algorithm with reduced computational cost. Remark 6.3 (Optimism of POMD). Unlike value-based algorithms (e.g., Jaksch et al. (2010)) V k, the value-function by which POMD improves upon, is not necessarily optimistic relatively to V . Instead, it is optimistic relatively to the value of πk, i.e., V k V πk. 7. Policy Optimization in Adversarial MDPs Algorithm 3 Optimistic POMD for Adversarial MDPs Require: t K, γ, π1 is the uniform policy. for k = 1, ..., K do Rollout a trajectory by acting πk for all h, s do Compute uk h(s) by πk, Pk 1, Eq. (7.1) end for # Policy Evaluation s S, V k H+1(s) = 0 for h = H, .., 1 do for s, a S A do ˆck h(s, a) = ck h(s,a)1{s=sk h,a=ak h} uk h(s)πk h(a|s)+γ ˆpk h( |s, a) arg min ˆph( |s,a) Pk 1 h (s,a) ˆph( |s, a)V k h+1 Qk h(s, a) = ˆck h(s, a) + ˆpk h( |s, a)V k h+1 end for for s S do V k h (s) = Qk h(s, ), πk h( | s) end for end for # Policy Improvement for h, s, a [H] S A do πk+1 h (a|s)= πk h(a|s) exp( t KQk h(s,a)) P a πk h(a |s) exp( t KQk h(s,a )) end for Update counters and model, nk, pk In this section, we turn to analyze an optimistic version of POMD for adversarial environments (Algorithm 3). Similarly to the stochastic case, Algorithm 3 follows the POMD scheme, and alternates between policy evaluation, and, soft policy improvement, based on MD-like updates. Optimistic Policy Optimization with Bandit Feedback 7 Unlike POMD for stochastic environments, the policy evaluation stage of Algorithm 3 uses different estimates of the instantaneous cost and model. The instantaneous cost is evaluated by a biased importance-sampling estimator, originally suggested by (Neu, 2015), and recently generalized to adversarial RL settings by (Jin et al., 2019), ˆck h(s, a) = ck h(s, a)1 s = sk h, a = ak h uk h(s)πk h(a | s) + γ , where uk h(s) = max ˆp Pk 1 dπk h (s; ˆp). (7.1) Here Pk 1 is the set of transition functions obtained by using confidence intervals around the empirical model (see Appendix C.1.2). In Algorithm 3 of Jin et al. (2019), the authors suggest a computationally efficient dynamic programming based approach for calculating uk h(s) for all h, s. The motivation for such an estimate lies in the EXP3 algorithm (Auer et al., 2002) for adversarial bandits, which uses an unbi- ased importance-sampling estimator ˆc(a) = ck(a)1{a=ak} πk(a) . Later, Neu (2015) showed that an optimistically biased esti- mator ˆc(a) = ck(a)1{a=ak} πk(a)+γ that motivates exploration can also be used in this setting. Generalizing the latter estimator to the adversarial RL setting requires to use the estimator ˆck h(s, a) = ck h(s,a)1{s=sk h,a=ak h} d πk h (s;p)πk h(a|s)+γ . However, since the model is unknown, Jin et al. (2019) uses uk h(s) as an upper bound on dπk h (s; p) which further drives exploration. Instead of using the empirical model and subtracting a bonus term, Algorithm 3 uses an optimistic model (Jaksch et al., 2010) for the policy evaluation stage. The model by which Qk is evaluated is the one which results in the smallest loss among possible models, ˆpk h( |s, a) arg min ˆph( |s,a) Pk 1 h (s,a) ˆph( |s, a)V k h+1. The solution to this optimization problem can be computed efficiently (see, e.g., Jaksch et al. (2010)). Remark 7.1 (Optimistic Model vs. Additive Exploration Bonus). Replacing the optimistic model with additive bonuses, we were able to establish O(K3/4) regret bound. It is not clear whether this approach can attain a O(K2/3) regret bound, as achieved when using an optimistic model. The following theorem bounds the regret of Algorithm 3. A full proof is found in Appendix C.2. Theorem 2. For any K [K], setting γ = O(A 1/2K 1/3) and t K = O(H 1K 2/3), the regret of Algorithm 3 is bounded by Regret(K ) O H2S A(K2/3 + SAK1/3) . Central to the analysis are the following claims, formally established in Appendix C. The first is proved in (Jin et al., 2019)[Lemma 11], based upon (Neu, 2015)[Lemma 1]. Claim 1 (Jin et al. (2019), Lemma 11). Let α1, .., αK be a sequence of Fk 1 measurable functions such that αk [0, 2γ]S A. Then, for any h and K [K], with high probability, PK k=1 P s,a αk(s, a) ˆck h(s, a) ck h(s, a) O(1). Claim 2. Let α1, .., αK be a sequence of Fk 1 measurable functions such that αk [0, 2γ]. For any s, h and K [K], with high probability, PK k=1 αk V k h (s) V πk h (s) O(H). Claim 2 (see Lemma 7 in the appendix) allows us to derive improved upper bound on PK k=1 V k h (s) which is crucial to derive the O(K2/3) regret bound. Naively, we can bound V k h (s) by recalling it is a value function of an MDP with costs bounded by 1/γ. This leads to the naive bound k=1 V k h (s) K H/γ. (7.2) However, a tighter upper bound can be obtained by applying Claim 2 with αk = 2γ for all k [K ]. We have that k=1 V k h (s) k=1 V πk h (s) + H where in the last relation we used the fact that for any s, h, V πk h (s) H. In the following proof sketch we apply the later upper bound and demonstrate its importance. Proof Sketch. We decompose the regret as in Theorem 1 to (i) Bias term, (ii) OMD term, and (iii) Optimism term. We bound both the Bias and Optimism terms in the appendix while relying on both Claim 1 and Claim 2. Term (ii): OMD Analysis. Similarly to the stochastic case, we utilize the usual OMD analysis (Lemma 16), which ensures that for any policy π and s, h, Qk h( | s), πk h( | s) πh( | s) a πk h(a | s)(Qk h(s, a))2 a πk h(a | s)Qk h(s, a) | {z } =V k h (s) Optimistic Policy Optimization with Bandit Feedback 8 where the second relation holds since 0 Qk h(s, a) H γ , and the third relation holds by applying Eq. (7.3). Plugging this in Term (ii) we get Term (ii) = Qk h(sh, ), πk h( |sh) πh( |sh) | s1, π, p t K + t KH2 8. Discussion On-policy vs. Off-policy. There are two prevalent approaches for policy optimization in practice, on-policy and off-policy. On-policy algorithms, e.g., TRPO (Schulman et al., 2015), update the policy based on data gathered following the current policy. This results in updating the policy only in observed states. However, in terms of theoretical guarantees, the convergence analysis of this approach requires the strong assumption of finite concentrability coefficient (Kakade & Langford, 2002; Scherrer & Geist, 2014; Agarwal et al., 2019; Liu et al., 2019; Shani et al., 2019). The assumption arises from the need to acquire global guarantees from the local nature of policy gradients. The approach taken in this work, is fundamentally different than such on-policy approaches. In each episode, instead of updating the policy only at visited states, the policy is updated over the entire state space, by using all the historical data (in the form of the empirical model). Thus, the analyzed approach bears resemblance to off-policy algorithms, e.g., SAC (Haarnoja et al., 2018). There, the authors i) estimate the Q-function of the current policy by sampling from a buffer, which contains historical data, and ii) apply an MDlike policy update to states sampled from the buffer. The uniform updates of policy-based methods analyzed in this work are in stark contrast to value-based algorithms, such as in (Jin et al., 2018; Efroni et al., 2019), where only observed states are updated. It remains an important open question, whether such updates could also be implemented in a provable policy based algorithm. In the case of stochastic POMD, this may be achieved by using optimistic Q-function estimates, instead of estimating the model with UCB-bonus, similarly to (Jin et al., 2019). There, the authors keep the estimates optimistic with respect to the optimal Q-function, Q . However, in approximate policy optimization, the policy improvement is done with respect to Qπk, as described in Algorithm 1. Therefore, differently than in (Jin et al., 2019), such off-policy version would require learning an optimistic Qπk estimator, instead of Q . Policy vs. State-Action Occupancy Optimization. In our work, we proposed algorithms which directly optimize the policy. In this scenario, the policy is updated independently at each time step h and state s. That is, an optimization problem is solved over the action space in each h, s. Therefore, this method requires solving HS optimization problems of size A, where each has a closed form solution in the tabular setting. Alternatively, algorithms based on the O-REPS framework (Zimin & Neu, 2013), follow a different approach and optimize over the state-action occupancy measures instead of directly on policies. In the case of unknown transition model, taking such an approach requires solving a constrained convex optimization problem, later relaxed to a convex optimization problem with only non-negativity constraints (Rosenberg & Mansour, 2019b) of size HS2A, in each episode. Unlike the policy optimization approach, this optimization problem does not have a closed form solution. Thus, the computational cost of optimizing over the stateaction occupancy measures is much worse than the policy optimization one. Another significant shortcoming in applying the O-REPS framework is the difficulty to scale it to the function approximation setting. Specifically, in case the state-action occupancy measure is represented by a non-linear function, it is unclear how to solve the constrained optimization problem as defined in (Rosenberg & Mansour, 2019b). Differently than the O-REPS framework, the policy optimization approach scales naturally to the function approximation setting, e.g., Haarnoja et al. (2018). In this important aspect, policy optimization is preferable. Interestingly, our work establishes O( K) regret when using POMD for the stochastic case, suggesting that policybased methods are sufficient for solving stochastic MDPs, and thus preferable, compared to the O-REPS framework, as they also enjoy better computational properties. However, in the adversarial case, Jin et al. (2019) recently established O( K) regret for the UOB-REPS algorithm, where the adversarial variant of POMD only obtains O K2/3 regret. Hence, it is of importance to understand whether it is possible to bridge this gap between policy and occupancy measure based methods, or alternatively to show that this gap is in fact a true drawback of policy optimization methods in the adversarial case. 9. Acknowledgments We thank the anonymous reviewers for providing us with very helpful comments. This work was partially funded by the Israel Science Foundation under ISF grant number 1380/16. Optimistic Policy Optimization with Bandit Feedback 9 Abbasi-Yadkori, Y., Bartlett, P., Bhatia, K., Lazic, N., Szepesvari, C., and Weisz, G. Politex: Regret bounds for policy iteration using expert prediction. In International Conference on Machine Learning, pp. 3692 3702, 2019. Agarwal, A., Kakade, S. M., Lee, J. D., and Mahajan, G. Optimality and approximation with policy gradient methods in markov decision processes. ar Xiv preprint ar Xiv:1908.00261, 2019. Altman, E. Constrained Markov decision processes, volume 7. CRC Press, 1999. Auer, P., Cesa-Bianchi, N., Freund, Y., and Schapire, R. E. The nonstochastic multiarmed bandit problem. SIAM journal on computing, 32(1):48 77, 2002. Azar, M. G., Osband, I., and Munos, R. Minimax regret bounds for reinforcement learning. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, pp. 263 272, 2017. Beck, A. and Teboulle, M. Mirror descent and nonlinear projected subgradient methods for convex optimization. Operations Research Letters, 31(3):167 175, 2003. Bhandari, J. and Russo, D. Global optimality guarantees for policy gradient methods. ar Xiv preprint ar Xiv:1906.01786, 2019. Cai, Q., Yang, Z., Jin, C., and Wang, Z. Provably efficient exploration in policy optimization. ar Xiv preprint ar Xiv:1912.05830, 2019. Dann, C., Lattimore, T., and Brunskill, E. Unifying pac and regret: Uniform pac bounds for episodic reinforcement learning. In Advances in Neural Information Processing Systems, pp. 5713 5723, 2017. Deisenroth, M. and Rasmussen, C. E. Pilco: A model-based and data-efficient approach to policy search. In Proceedings of the 28th International Conference on machine learning (ICML-11), pp. 465 472, 2011. Efroni, Y., Merlis, N., Ghavamzadeh, M., and Mannor, S. Tight regret bounds for model-based reinforcement learning with greedy policies. In Advances in Neural Information Processing Systems, pp. 12203 12213, 2019. Geist, M., Scherrer, B., and Pietquin, O. A theory of regularized markov decision processes. In International Conference on Machine Learning, pp. 2160 2169, 2019. Gu, S., Holly, E., Lillicrap, T., and Levine, S. Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates. In 2017 IEEE international conference on robotics and automation (ICRA), pp. 3389 3396. IEEE, 2017. Haarnoja, T., Zhou, A., Abbeel, P., and Levine, S. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In International Conference on Machine Learning, pp. 1861 1870, 2018. Jaksch, T., Ortner, R., and Auer, P. Near-optimal regret bounds for reinforcement learning. Journal of Machine Learning Research, 11(Apr):1563 1600, 2010. Jin, C., Allen-Zhu, Z., Bubeck, S., and Jordan, M. I. Is q-learning provably efficient? In Advances in Neural Information Processing Systems, pp. 4863 4873, 2018. Jin, C., Jin, T., Luo, H., Sra, S., and Yu, T. Learning adversarial markov decision processes with bandit feedback and unknown transition. ar Xiv preprint ar Xiv:1912.01192, 2019. Kakade, S. and Langford, J. Approximately optimal approximate reinforcement learning. In ICML, volume 2, pp. 267 274, 2002. Kakade, S. M. A natural policy gradient. In Advances in neural information processing systems, pp. 1531 1538, 2002. Kearns, M. and Singh, S. Near-optimal reinforcement learning in polynomial time. Machine learning, 49(2-3):209 232, 2002. Levine, S., Finn, C., Darrell, T., and Abbeel, P. End-toend training of deep visuomotor policies. The Journal of Machine Learning Research, 17(1):1334 1373, 2016. Lillicrap, T. P., Hunt, J. J., Pritzel, A., Heess, N., Erez, T., Tassa, Y., Silver, D., and Wierstra, D. Continuous control with deep reinforcement learning. ar Xiv preprint ar Xiv:1509.02971, 2015. Liu, B., Cai, Q., Yang, Z., and Wang, Z. Neural proximal/trust region policy optimization attains globally optimal policy. ar Xiv preprint ar Xiv:1906.10306, 2019. Maurer, A. and Pontil, M. Empirical bernstein bounds and sample variance penalization. stat, 1050:21, 2009. Munos, R. Error bounds for approximate policy iteration. In Fawcett, T. and Mishra, N. (eds.), Machine Learning, Proceedings of the Twentieth International Conference (ICML 2003), August 21-24, 2003, Washington, DC, USA, pp. 560 567. AAAI Press, 2003. Neu, G. Explore no more: Improved high-probability regret bounds for non-stochastic bandits. In Advances in Neural Information Processing Systems, pp. 3168 3176, 2015. Optimistic Policy Optimization with Bandit Feedback 10 Neu, G., Antos, A., György, A., and Szepesvári, C. Online markov decision processes under bandit feedback. In Advances in Neural Information Processing Systems, pp. 1804 1812, 2010a. Neu, G., György, A., and Szepesvári, C. The online loopfree stochastic shortest-path problem. In COLT, volume 2010, pp. 231 243. Citeseer, 2010b. Neu, G., Jonsson, A., and Gómez, V. A unified view of entropy-regularized markov decision processes. ar Xiv preprint ar Xiv:1705.07798, 2017. Orabona, F. A modern introduction to online learning. ar Xiv preprint ar Xiv:1912.13213, 2019. Peters, J. and Schaal, S. Policy gradient methods for robotics. In 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 2219 2225. IEEE, 2006. Peters, J. and Schaal, S. Reinforcement learning of motor skills with policy gradients. Neural networks, 21(4):682 697, 2008. Puterman, M. L. Markov decision processes: discrete stochastic dynamic programming. John Wiley & Sons, 2014. Rosenberg, A. and Mansour, Y. Online stochastic shortest path with bandit feedback and unknown transition function. In Advances in Neural Information Processing Systems, pp. 2209 2218, 2019a. Rosenberg, A. and Mansour, Y. Online convex optimization in adversarial markov decision processes. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, pp. 5478 5486, 2019b. Scherrer, B. Approximate policy iteration schemes: a comparison. In International Conference on Machine Learning, pp. 1314 1322, 2014. Scherrer, B. and Geist, M. Local policy search in a convex space and conservative policy iteration as boosted policy search. In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 35 50. Springer, 2014. Schulman, J., Levine, S., Abbeel, P., Jordan, M., and Moritz, P. Trust region policy optimization. In International Conference on Machine Learning, pp. 1889 1897, 2015. Shani, L., Efroni, Y., and Mannor, S. Adaptive trust region policy optimization: Global convergence and faster rates for regularized mdps. ar Xiv preprint ar Xiv:1909.02769, 2019. Sutton, R. S., Mc Allester, D. A., Singh, S. P., and Mansour, Y. Policy gradient methods for reinforcement learning with function approximation. In Advances in neural information processing systems, pp. 1057 1063, 2000. Vieillard, N., Pietquin, O., and Geist, M. On connections between constrained optimization and reinforcement learning. ar Xiv preprint ar Xiv:1910.08476, 2019. Wei, C.-Y., Jafarnia-Jahromi, M., Luo, H., Sharma, H., and Jain, R. Model-free reinforcement learning in infinitehorizon average-reward markov decision processes. ar Xiv preprint ar Xiv:1910.07072, 2019. Weissman, T., Ordentlich, E., Seroussi, G., Verdu, S., and Weinberger, M. J. Inequalities for the l1 deviation of the empirical distribution. Hewlett-Packard Labs, Tech. Rep, 2003. Zanette, A. and Brunskill, E. Tighter problem-dependent regret bounds in reinforcement learning without domain knowledge using value function bounds. In International Conference on Machine Learning, pp. 7304 7312, 2019. Zimin, A. and Neu, G. Online learning in episodic markovian decision processes by relative entropy policy search. In Advances in neural information processing systems, pp. 1583 1591, 2013.