# reward_is_enough_for_convex_mdps__7c964c9c.pdf Reward is Enough for Convex MDPs Tom Zahavy Deep Mind, London tomzahavy@deepmind.com Brendan O Donoghue Deep Mind, London bodonoghue@deepmind.com Guillaume Desjardins Deep Mind, London gdesjardins@deepmind.com Satinder Singh Deep Mind, London baveja@deepmind.com Maximising a cumulative reward function that is Markov and stationary, i.e., defined over state-action pairs and independent of time, is sufficient to capture many kinds of goals in a Markov decision process (MDP). However, not all goals can be captured in this manner. In this paper we study convex MDPs in which goals are expressed as convex functions of the stationary distribution and show that they cannot be formulated using stationary reward functions. Convex MDPs generalize the standard reinforcement learning (RL) problem formulation to a larger framework that includes many supervised and unsupervised RL problems, such as apprenticeship learning, constrained MDPs, and so-called pure exploration . Our approach is to reformulate the convex MDP problem as a min-max game involving policy and cost (negative reward) players , using Fenchel duality. We propose a meta-algorithm for solving this problem and show that it unifies many existing algorithms in the literature. 1 Introduction In reinforcement learning (RL), an agent learns how to map situations to actions so as to maximize a cumulative scalar reward signal. The learner is not told which actions to take, but instead must discover which actions lead to the most reward [64]. Mathematically, the RL problem can be written as finding a policy whose state occupancy has the largest inner product with a reward vector [55], i.e., the goal of the agent is to solve RL: max dπ K s,a r(s, a)dπ(s, a), (1) where dπ is the state-action stationary distribution induced by policy π and K is the set of admissible stationary distributions (see Definition 1). A significant body of work is dedicated to solving the RL problem efficiently in challenging domains [45, 62]. However, not all decision making problems of interest take this form. In particular we consider the more general convex MDP problem, Convex MDP: min dπ K f(dπ), (2) where f : K R is a convex function. Sequential decision making problems that take this form include Apprenticeship Learning (AL), pure exploration, and constrained MDPs, among others; see Table 1. In this paper we prove the following claim: We can solve Eq. (2) by using any algorithm that solves Eq. (1) as a subroutine. In other words, any algorithm that solves the standard RL problem can be used to solve the more general convex MDP problem. More specifically, we make the following contributions. 35th Conference on Neural Information Processing Systems (Neur IPS 2021). Firstly, we adapt the meta-algorithm of Abernethy and Wang [3] for solving Eq. (2). The key idea is to use Fenchel duality to convert the convex MDP problem into a two-player zero-sum game between the agent (henceforth, policy player) and an adversary that produces rewards (henceforth, cost player) that the agent must maximize [3, 6]. From the agent s point of view, the game is bilinear, and so for fixed rewards produced by the adversary the problem reduces to the standard RL problem with non-stationary reward (Fig. 1). History Cost Player Environment Figure 1: Convex MDP as an RL problem Secondly, we propose a sample efficient policy player that uses a standard RL algorithm (eg, [35, 60]), and computes an optimistic policy with respect to the non-stationary reward at each iteration. In other words, we use algorithms that were developed to achieve low regret in the standard RL setup, to achieve low regret as policy players in the min-max game we formulate to solve the convex MDP. Our main result is that the average of the policies produced by the policy player converges to a solution to the convex MDP problem (Eq. (2)). Inspired by this principle, we also propose a recipe for using deep-RL (DRL) agents to solve convex MDPs heuristically: provide the agent non-stationary rewards from the cost player. We explore this principle in our experiments. Finally, we show that choosing specific algorithms for the policy and cost players unifies several disparate branches of RL problems, such as apprenticeship learning, constrained MDPs, and pure exploration into a single framework, as we summarize in Table 1. Convex objective f Cost player Policy player Application λ dπ FTL RL (Standard) RL with λ as stationary reward function ||dπ d E||2 2 FTL Best response Apprenticeship learning (AL) [1, 75] dπ log(dπ) FTL Best response Pure exploration [30] ||dπ d E|| OMD Best response AL [66, 65] Ec [λc (dπ d E(c))] OMD Best response Inverse RL in contextual MDPs [10] λ1 dπ, s.t. λ2 dπ c OMD RL Constrained MDPs [7, 67, 12, 68, 18, 16, 11] dist(dπ, C) OMD Best response Feasibility of convex-constrained MDPs [44] minλ1,...,λk dk π λk OMD RL Adversarial Markov Decision Processes [57] maxλ Λ λ (dπ d E) OMD RL Online AL [61],Wasserstein GAIL [73, 78] KL(dπ||d E) FTL RL GAIL [31], state marginal matching [41], Ez KL(dz π||Ekdk π) FTL RL Diverse skill discovery [26, 20, 27, 21, 69, 4] Table 1: Instances of Algorithm 1 in various convex MDPs. as well as other KL divergences. c is a context variable. C is a convex set. f is concave. See Sections 4 & 6 for more details. 2 Reinforcement Learning Preliminaries In RL an agent interacts with an environment over a number of time steps and seeks to maximize its cumulative reward. We consider two cases, the average reward case and the discounted case. The Markov decision process (MDP) is defined by the tuple (S, A, P, R) for the average reward case and by the tuple (S, A, P, R, γ, d0) for the discounted case. We assume an infinite horizon, finite state-action problem where initially, the state of the agent is sampled according to s0 d0, then at each time t the agent is in state st S, selects action at A according to some policy π(st, ), receives reward rt R(st, at) and transitions to new state st+1 S according to the probability distribution P( , st, at). The two performance metrics we consider are given by Javg π = lim T 1 T E t=1 rt, Jγ π = (1 γ)E t=1 γtrt, (3) for the average reward case and discounted case respectively. The goal of the agent is to find a policy that maximizes Javg π or Jγ π. Any stationary policy π induces a state-action occupancy measure dπ, which measures how often the agent visits each state-action when following π. Let Pπ(st = ) be the probability measure over states at time t under policy π, then davg π (s, a) = lim T 1 T E t=1 Pπ(st = s)π(s, a), dγ π(s, a) = (1 γ)E t=1 γt Pπ(st = s)π(s, a), for the average reward case and the discounted case respectively. With these, we can rewrite the RL objective in Eq. (3) in terms of the occupancy measure using the following well-known result, which for completeness we prove in Appendix B. Proposition 1. For both the average and the discounted case, the agent objective function Eq. (3) can be written in terms of the occupancy measure as Jπ = P s,a r(s, a)dπ(s, a). Given an occupancy measure it is possible to recover the policy by setting π(s, a) = dπ(s, a)/ P a dπ(s, a) if P a dπ(s, a) > 0, and π(s, a) = 1/|A| otherwise. Accordingly, in this paper we shall formulate the RL problem using the state-action occupancy measure, and both the standard RL problem (Eq. (1)) and the convex MDP problem (Eq. (2)) are convex optimization problems in variable dπ. For the purposes of this manuscript we do not make a distinction between the average and discounted settings, other than through the convex polytopes of feasible occupancy measures, which we define next. Definition 1 (State-action occupancy s polytope [55]). For the average reward case the set of admissible state-action occupancies is Kavg = {dπ | dπ 0, X s,a dπ(s, a) = 1, X a dπ(s, a) = X s ,a P(s, s , a )dπ(s , a ) s S}, and for the discounted case it is given by Kγ = {dπ | dπ 0, X a dπ(s, a) = (1 γ)d0(s) + γ X s ,a P(s, s , a )dπ(s , a ) s S}. We note that being a polytope implies that K is a convex and compact set. The convex MDP problem is defined for the tuple (S, A, P, f) in the average cost case and (S, A, P, f, γ, d0) in the discounted case. This tuple is defining a state-action occupancy s polytope K (Definition 1), and the problem is to find a policy π whose state occupancy dπ is in this polytope and minimizes the function f (Eq. (2)). 3 A Meta-Algorithm for Solving Convex MDPs via RL To solve the convex MDP problem (Eq. (2)) we need to find an occupancy measure dπ (and associated policy) that minimizes the function f. Since both f : K R and the set K are convex this is a convex optimization problem. However, it is a challenging one due to the nature of learning about the environment through stochastic interactions. In this section we show how to reformulate the convex MDP problem (Eq. (2)) so that standard RL algorithms can be used to solve it, allowing us to harness decades of work on solving vanilla RL problems. To do that we will need the following definition. Definition 2 (Fenchel conjugate). For a function f : Rn R { , }, its Fenchel conjugate is denoted f : Rn R { , } and defined as f (x) := supy x y f(y). Remark 1. The Fenchel conjugate function f is always convex (when it exists) even if f is not. Furthermore, the biconjugate f := (f ) equals f if and only if f is convex and lower semicontinuous. Using this we can rewrite the convex MDP problem (Eq. (2)) as f OPT = min dπ K f(dπ) = min dπ K max λ Λ (λ dπ f (λ)) = max λ Λ min dπ K (λ dπ f (λ)) (4) where Λ is the closure of (sub-)gradient space { f(dπ)|dπ K}, which is a convex set [3, Theorem 4]. As both sets are convex, this is a convex-concave saddle-point problem and a zero-sum two-player game [54, 49], and we were able to swap the order of minimization and maximization using the minimax theorem [71]. With this we define the Lagrangian as L(dπ, λ) := λ dπ f (λ). For a fixed λ Λ, minimizing the Lagrangian is a standard RL problem of the form of Eq. (1), i.e., equivalent to maximizing a reward r = λ. Thus, one might hope that by producing an optimal dual variable λ we could simply solve d π = argmindπ K L( , λ ) for the optimal occupancy measure. However, the next lemma states that this is not possible in general. Lemma 1. There exists an MDP M and convex function f for which there is no stationary reward r RS A such that arg maxdπ K dπ r = arg mindπ K f(dπ). To see this note that for any reward r there is a deterministic policy that optimizes the reward [55], but for some choices of f no deterministic policy is optimal, eg, when f is the negative entropy function. This result tells us that even if we have access to an optimal dual-variable we cannot simply use it to recover the stationary distribution that solves the convex MDP problem in general. To overcome this issue we develop an algorithm that generates a sequence of policies {πk}k N such that the average converges to an optimal policy for Eq. (2), i.e., (1/K) PK k=1 dk π d π arg mindπ K f(dπ). The algorithm we develop is described in Algorithm 1 and is adapted from the meta-algorithm described in Abernethy and Wang [3]. It is referred to as a meta-algorithm since it relies on supplied sub-routine algorithms Algπ and Algλ. The reinforcement learning algorithm Algπ takes as input a reward vector and returns a state-action occupancy measure dπ. The cost algorithm Algλ can be a more general function of the entire history. We discuss concrete examples of Algπ and Algλ in Section 4. Algorithm 1: meta-algorithm for convex MDPs 1: Input: convex-concave payoff L : K Λ R, algorithms Algλ, Algπ, K N 2: for k = 1, . . . , K do 3: λk = Algλ(d1 π, . . . , dk 1 π ; L) 4: dk π = Algπ( λk) 5: end for 6: Return d K π = 1 K PK k=1 dk π, λK = 1 K PK k=1 λk In order to analyze this algorithm we will need a small detour into online convex optimization (OCO). In OCO, a learner is presented with a sequence of K convex loss functions ℓ1, ℓ2, . . . , ℓK : K R and at each round k must select a point xk K after which it suffers a loss of ℓk(xk). At time period k the learner is assumed to have perfect knowledge of the loss functions ℓ1, . . . , ℓk 1. The learner wants to minimize its average regret, defined as k=1 ℓk(xk) min x K In the context of convex reinforcement learning and meta-algorithm 1, the loss functions for the cost player are ℓk λ = L( , λk), and for the policy player are ℓk π = L(dk π, ), with associated average regrets Rπ K and Rλ K. This brings us to the following theorem. Theorem 1 (Theorem 2, [3]). Assume that Algπ and Algλ have guaranteed average regret bounded as Rπ K ϵK and Rλ K δK, respectively. Then Algorithm 1 outputs d K π and λK satisfying mindπ K L(dπ, λK) f OPT ϵK δK and maxλ Λ L( d K π , λ) f OPT + ϵK + δK. This theorem tells us that so long as the RL algorithm we employ has guaranteed low-regret, and assuming we choose a reasonable low-regret algorithm for deciding the costs, then the meta-algorithm will produce a solution to the convex MDP problem (Eq. (2)) to any desired tolerance, this is because f OPT f( d K π ) = maxλ L( d K π , λ) f OPT +ϵK +δK. For example, we shall later present algorithms that have regret bounded as ϵK = δK O(1/ K), in which case we have f( d K π ) f OPT O(1/ Non-Convex f. Remark 1 implies that the game maxλ Λ mindπ K (λ dπ f (λ)) is concaveconvex for any function f, so we can solve it with Algorithm 1, even for a non-convex f. From weak duality the value of the Lagrangian on the output of Algorithm 1, L( dπ, λ), is a lower bound on the optimal solution f OPT. In addition, since f(dπ) is always an upper bound on f OPT we have both an upper bound and a lower bound on the optimal value: L( dπ, λ) f OPT f( dπ). 4 Policy and Cost Players for Convex MDPs In this section we present several algorithms for the policy and cost players that can be used in Algorithm 1. Any combination of these algorithms is valid and will come with different practical and theoretical performance. In Section 6 we show that several well known methods in the literature correspond to particular choices of cost and policy players and so fall under our framework. In addition, in this section we assume that λmax = max λ Λ max s,a |λ(s, a)| < , which holds when the set Λ is compact. One way to guarantee that Λ is compact is to consider functions f with Lipschitz continuous gradients (which implies bounded gradients since the set K is compact). For simplicity, we further assume that λmax 1. By making this assumption we assure that the non stationary rewards produced by the cost player are bounded by 1 as is usually done in RL. 4.1 Cost Player Follow the Leader (FTL) is a classic OCO algorithm that selects λk to be the best point in hindsight. In the special case of convex MDPs, as defined in Eq. (4), FTL has a simpler form: λk = arg max λ Λ j=1 L(dj π, λ) = arg max λ Λ j=1 dj π Kf (λ) = f( dk 1 π ), (6) where dk 1 π = Pk 1 j=1 dj π and the last equality follows from the fact that ( f ) 1 = f [56]. The average regret of FTL is guaranteed to be RK c/ K under some assumptions [29]. In some cases, and specifically when the set K is a polytope and the function f is strongly convex, FTL can enjoy logarithmic or even constant regret; see [32, 29] for more details. Online Mirror Descent (OMD) uses the following update [47, 9]: λk = arg max λ Λ (λ λk 1) λL(dk 1 π , λk 1) + αk Br(λ, λk 1) , where αk is a learning rate and Br is a Bregman divergence [14]. For Br(x) = 0.5||x||2 2, we get online gradient descent [79] and for Br(x) = x log(x) we get multiplicative weights [23] as special cases. We also note that OMD is equivalent to a linearized version of Follow the Regularized Leader (FTRL) [43, 28]. The average regret of OMD is RK c/ K under some assumptions, see, for example [28]. 4.2 Policy Players 4.2.1 Best Response In OCO, the best response is to simply ignore the history and play the best option on the current round, which has guaranteed average regret bound of RK 0 (this requires knowledge of the current loss function, which is usually not applicable but is in this case). When applied to Eq. (4), it is possible to find the best response dk π using standard RL techniques since dk π = arg min dπ K Lk(dπ, λk) = arg min dπ K dπ λk f (λk) = arg max dπ K dπ ( λk), which is an RL problem for maximizing the reward ( λk). In principle, any RL algorithm that eventually solves the RL problem can be used to find the best response, which substantiates our claim in the introduction. For example, tabular Q-learning executed for sufficiently long and with a suitable exploration strategy will converge to the optimal policy [72]. In the non-tabular case we could parameterize a deep neural network to represent the Q-values [45] and if the network has sufficient capacity then similar guarantees might hold. We make no claims on efficiency or tractability of this approach, just that in principle such an approach would provide the best-response at each iteration and therefore satisfy the required conditions to solve the convex MDP problem. 4.2.2 Approximate Best Response The caveat in using the best response as a policy player is that in practice, it can only be found approximately by executing an RL algorithm in the environment. This leads to defining an approximate best response via the Probably Approximately Correct (PAC) framework. We say that a policy player is PAC(ϵ, δ), if it finds an ϵ-optimal policy to an RL problem with probability of at least 1 δ. In addition, we say that a policy π is ϵ-optimal if its state occupancy d π is such that max dπ K dπ ( λk) d π ( λk) ϵ. For example, the algorithm in [40] can find an ϵ-optimal policy to the discounted RL problem after seeing O SA (1 γ)3ϵ2 log( 1 δ ) samples; and the algorithm in [36] can find an ϵ-optimal policy for the average reward RL problem after seeing O t2 mix SA δ ) samples, where tmix is the mixing time (see, eg, [42, 76] for a formal definition). The following Lemma analyzes the sample complexity of Algorithm 1 with an approximate best response policy player for the average reward RL problem [36]. The result can be easily extended to the discounted case using the algorithm in [40]. Other relaxations to the best response for specific algorithms can be found in [65, 44, 33, 30]. Lemma 2 (The sample complexity of approximate best response in convex MDPs with average occupancy measure). For a convex function f, running Algorithm 1 with an oracle cost player with regret Rλ K = O(1/K) and an approximate best response policy player that solves the average reward RL problem in iteration k to accuracy ϵk = 1/k returns an occupancy measure d K π that satisfies f( d K π ) f OPT ϵ with probability 1 δ after seeing O(t2 mix SA log(2K/ϵδ)/ϵ3δ3) samples. Similarly, for Rλ K = O(1/ K), setting ϵk = 1/ k requires O(t2 mix SA log(2K/ϵδ)/ϵ4δ4) samples. 4.2.3 Non-Stationary RL Algorithms We now discuss a different type of policy players; instead of solving an MDP to accuracy ϵ, these algorithms perform a single RL update to the policy, with cost λk. In our setup the reward is known and deterministic but non-stationary, while in the standard RL setup it is unknown, stochastic, and stationary. We conjecture that any RL algorithm can be adapted to the known non-stationary reward setup we consider here. In most cases both Bayesian [51, 48] and frequentist [8, 35] approaches to the stochastic RL problem solve a modified (eg, by adding optimism) Bellman equation at each time period and swapping in a known but non-stationary reward is unlikely to present a problem. To support this conjecture we shall prove that this is exactly the case for UCRL2 [35]. UCRL2 is an RL algorithm that was designed and analyzed in the standard RL setup, and we shall show that it is easily adapted to the non-stationary but known reward setup that we require. To make this claim more general, we will also discuss a similar result for the MDPO algorithm [61] that was given in a slightly different setup. UCRL2 is a model based algorithm that maintains an estimate of the reward and the transition function as well as confidence sets about those estimates. In our case the reward at time k is known, so we only need to consider uncertainty in the dynamics. UCRL2 guarantees that in any iteration k, the true transition function is in a confidence set with high probability, i.e., P Pk for confidence set Pk. If we denote by JP,R π the value of policy π in an MDP with dynamics P and reward R then the optimistic policy is πk = arg maxπ max P Pk JP , λk π . Acting according to this policy is guaranteed to attain low regret. In the following results for UCRL2 we will use the constant D, which denotes the diameter of the MDP, see [35, Definition 1] for more details. In the supplementary material (Appendix E), we provide a proof sketch that closely follows [35]. Lemma 3 (Non stationary regret of UCRL2). For an MDP with dynamics P, diameter D, an arbitrary sequence of known and bounded rewards ri : maxs,a |ri(s, a)| 1 K i=1 , such that the optimal average reward at episode k, with respect to P and rk is J k, then with probability at least 1 δ, the average regret of UCRL2 is at most RK = 1 K PK k=1 J k J πk k O(DS p A log(K/δ)/K). Next, we give a PAC(ϵ, δ) sample complexity result for the mixed policy πK, that is produced by running Algorithm 1 with UCRL2 as a policy player. Lemma 4 (The sample complexity of non-stationary RL algorithms in convex MDPs). For a convex function f, running Algorithm 1 with an oracle cost player with regret Rλ K c0/ K and UCRL2 as a policy player returns an occupancy measure d K π that satisfies f( d K π ) f OPT ϵ with probability 1 δ after K = O D2S2A δ2ϵ2 log( 2DSA δϵ ) steps. MDPO. Another optimistic algorithm is Mirror Descent Policy Optimization [60, MDPO]. MDPO is a model free RL algorithm that is very similar to popular DRL algorithms like TRPO [58] and MPO [2]. In [24, 59, 5], the authors established the global convergence of MDPO and in [15, 60], the authors showed that MDPO with optimistic exploration enjoys low regret. The analysis for MDPO is given in a finite horizon MDP with horizon H, which is not the focus of our paper. Nevertheless, to support our conjecture that any stochastic RL algorithm can be adapted to the known non-stationary reward setup, we quickly discuss the regret of MDPO in this setup. We also note that MDPO is closer to practical DRL algorithms [70]. In a finite horizon MDP with horizon H and known, non-stationary and bounded rewards, the regret of MDPO is bounded by RK O(H2S p A/K) [61, Lemma 4] with high probability. To compare this result with UCRL2, we refer to a result from [57], which analyzed UCRL2 in the adversarial setup, that includes our setup as a special case. In a finite horizon MDP with horizon H it was shown that setting δ = SA/K with probability 1 δ its regret is bounded by RK O(HS p A log(K)/K) [57, Corollary 5], which is better by a factor of H than MDPO. Discussion. Comparing the results in Lemma 4 with Lemma 2 suggests that using an RL algorithm with non stationary reward as a policy player requires O(1/ϵ2) samples to find an ϵ optimal policy, while using an approximate best response requires O(1/ϵ3). In first glance, this results also improves the previously best known result of Hazan et al. [30] for approximate Frank-Wolfe (FW) that requires O(1/ϵ3) samples. However, there are more details that have to be considered as we now discuss. Firstly, Lemma 4 and Lemma 2 assume access to an oracle cost player with some regret and do not consider how to implement such a cost player. The main challenge is that the cost player does not have access to the true state occupancy and must estimate it from samples. If we do not reuse samples from previous policies to estimate the state occupancy of the current policy we will require O(1/ϵ3) trajectories overall [30]. A better approach would use the samples from previous episodes to learn the transition function. Then, given the estimated transition function and the policy, we can compute an approximation of the state occupancy. We conjecture that such an approach would lead to a O(1/ϵ2) sample complexity, closing the gap with standard RL. Secondly, while our focus is on the dependence in ϵ, our bound Lemma 4 is not tight in δ, i.e., it scales with 1/δ2 where it should be possible to achieve a log(1/δ) scaling. Again we conjecture an improvement in the bound is possible; see, eg, [38, Appendix F.]. 5 Convex Constraints We have restricted the presentation so far to unconstrained convex problems, in this section we extend the above results to the constrained case. The problem we consider is min dπ K f(dπ) subject to gi(dπ) 0, i = 1, . . . m, where f and the constraint functions gi are convex. Previous work focused on the case where both f and gi are linear [7, 67, 12, 68, 18, 16, 11]. We can use the same Fenchel dual machinery we developed before, but now taking into account the constraints. Consider the Lagrangian L(dπ, µ) = f(dπ) + Xm i=1 µigi(dπ) = max ν (ν dπ f (ν)) + Xm i=1 µi max vi (dπvi g i (vi)) . over dual variables µ 0, with new variables vi and ν. At first glance this does not look convexconcave, however we can introduce new variables ζi = µivi to obtain L(dπ, µ, ν, ζ1, . . . , ζm) = ν dπ f (ν) + Xm i=1 (dπζi µig i (ζi/µi)) . (7) This is convex (indeed affine) in dπ and concave in (ν, µ, ζ1, . . . , ζm), since it includes the perspective transform of the functions gi [13]. The Lagrangian involves a cost vector, ν + Pm i=1 ζi, linearly interacting with dπ, and therefore we can use the same policy players as before to minimize this cost. For the cost player, it is possible to use OMD on Eq. (7) jointly for the variables ν, µ and ζ. It is more challenging to use best-response and FTL for the cost-player variables as the maximum value of the Lagrangian is unbounded for some values of dπ. Another option is to treat the problem as a three-player game. In this case the policy player controls dπ as before, one cost player chooses (ν, ζ1, . . . , ζm) and can use the algorithms we have previously discussed, and the other cost player chooses µ with some restrictions on their choice of algorithm. Analyzing the regret in that case is outside the scope of this paper. In this section we explain how existing algorithms can be seen as instances of the meta-algorithm for various choices of the objective function f and the cost and policy player algorithms Algλ and Algπ. We summarized the relationships in Table 1. 6.1 Apprenticeship Learning In apprenticeship learning (AL), we have an MDP without an explicit reward function. Instead, an expert provides demonstrations which are used to estimate the expert state occupancy measure d E. Abbeel and Ng [1] formalized the AL problem as finding a policy π whose state occupancy is close to that of the expert by minimizing the convex function f(dπ) = ||dπ d E||. The convex conjugate of f is given by f (y) = y d E if ||y|| 1 and otherwise, where || || denotes the dual norm. Plugging f into Eq. (4) results in the following game: min dπ K ||dπ d E|| = min dπ K max ||λ|| 1 λ dπ λ d E. (8) Inspecting Eq. (8), we can see that the norm in the function f that is used to measure the distance from the expert induces a constraint set for the cost variable, which is a unit ball in the dual norm. Algλ=OMD, Algπ=Best Response/RL. The Multiplicative Weights AL algorithms [65, MWAL] was proposed to solve the AL problem with f(dπ) = ||dπ d E|| . It uses the best response as the policy player and multiplicative weights as the cost player (a special case of OMD). MWAL has also been used to solve AL in contextual MDPs [10] and to find feasible solutions to convex-constrained MDPs [44]. We note that in practice the best response can only be solved approximately, as we discussed in Section 4. Instead, in online AL [61] the authors proposed to use MDPO as the policy player, which guarantees a regret bound of RK c/ K. They showed that their algorithm is equivalent to Wasserstein GAIL [73, 78] and in practice tends to perform similarly to GAIL. Algλ=FTL, Algπ=Best Response. When the policy player plays the best response and the cost player plays FTL, Algorithm 1 is equivalent to the Frank-Wolfe algorithm [22, 3] for minimizing f (Eq. (2)). Pseudo-code for this is included in the appendix (Algorithm 3). The algorithm finds a point dk π K that has the largest inner-product (best response) with the negative gradient (i.e., FTL). Abbeel and Ng [1] proposed two algorithms for AL, the projection algorithm and the max margin algorithm. The projection algorithm is essentially a FW algorithm, as was suggested in the supplementary [1] and was later shown formally in [75]. Thus, it is a projection free algorithm in the sense that it avoids projecting dπ into K, despite the name. In their case the gradient is given by f(dπ) = dπ d E. Thus, finding the best response is equivalent to solving an MDP whose reward is d E dπ. In a similar fashion, FW can be used to solve convex MDPs more generally [30]. Specifically, in [30], the authors considered the problem of pure exploration, which they defined as finding a policy that maximizes entropy. Fully Corrective FW. The FW algorithm has many variants (see [33] for a survey) some of which enjoy faster rates of convergence in special cases. Concretely, when the constraint set is a polytope, which is the case for convex MDPs (Definition 1), some variants achieve a linear rate of convergence [34, 75]. One such variant is the Fully corrective FW, which replaces the learning rate update (see line 4 of Algorithm 3 in the supplementary), with a minimization problem over the convex hull of occupancy measures at the previous time-step. This is guaranteed to be at least as good as the learning rate update. Interestingly, the second algorithm of Abbeel and Ng [1], the max margin algorithm, is exactly equivalent to this fully corrective FW variant. This implies that the max-margin algorithm enjoys a better theoretical convergence rate than the projection variant, as was observed empirically in [1]. 6.2 GAIL and DIAYN: Algλ=FTL, Algπ=RL We now discuss the objectives of two popular algorithms, GAIL [31] and DIAYN [20], which perform AL and diverse skill discovery respectively. Our analysis suggests that GAIL and DIAYN share the same objective function. In GAIL, this objective function is minimized, which is a convex MDP, however, in DIAYN it is maximized, which is therefore not a convex MDP. We start the discussion with DIAYN and follow with a simple construction showing the equivalence to GAIL. DIAYN. Discriminative approaches [26, 20] rely on the intuition that skills are diverse when they are entropic and easily discriminated by observing the states that they visit. Given a probability space (Ω, F, P), state random variables S : Ω S and latent skills Z : Ω Z with prior p, the key term of interest being maximized in DIAYN [20] is the mutual information: I(S; Z) = Ez p;s dzπ[log p(z|s) log p(z)], (9) where dz π is the stationary distribution induced by the policy π(a | s, z). For each skill z, this corresponds to a standard RL problem with (conditional) policy π(a | s, z) and reward function r(s|z) = log p(z|s) log p(z). The first term encourages the policy to visit states for which the underlying skill has high-probability under the posterior p(z | s), while the second term ensures a high entropy distribution over skills. In practice, the full DIAYN objective further regularizes the learnt policy by including entropy terms log π(a | s, z). For large state spaces, p(z|s) is typically intractable and Eq. 9 is replaced with a variational lower-bound, where the true posterior is replaced with a learned discriminator qφ(z|s). Here, we focus on the simple setting where z is a categorical distribution over |Z| outcomes, yielding |Z| policies πz, and qφ is a classifier over these |Z| skills with parameters φ. We now show that a similar intrinsic reward can be derived using the framework of convex MDPs. We start by writing the true posterior as a function of the per-skill state occupancy dz π = p(s | z), and using Bayes rules, p(z|s) = dz π(s)p(z) P k dkπ(s)p(k). Combing this with Eq. (9) yields: Ez p(z),s dzπ[log p(z|s) p(z)] = X s dz π(s) log dz π(s)p(z) P k dkπ(s)p(k) z p(z)KL(dz π|| X k p(k)dk π) = Ez KL(dz π||Ekdk π), (10) where KL denotes the Kullback Leibler divergence [39]. Intuitively, finding a set of policies π1, . . . , πz that minimize Eq. (10) will result in finding policies that visit similar states, measured using the KL distance between their respective state occupancies d1 π, . . . , dz π. This is a convex MDP because the KL-divergence is jointly convex in both arguments [13, Example 3.19]. We will soon show that this is the objective of GAIL. On the other hand, a set of policies that maximize Eq. (10) is diverse, as the policies visit different states, measured using the KL distance between their respective state occupancies d1 π, . . . , dz π. We follow on with deriving the FTL player for the convex MDP in Eq. (10). We will then show that this FTL player is producing an intrinsic reward that is equivalent to the intrinsic reward used in GAIL and DIAYN (despite the fact that DIAYN is not a convex MDP). According to Eq. (6), the FTL cost player will produce a cost λk at iteration k given by dzπKL(dz π|| X k p(k)dk π) = E z p(z) h log dz π P k dkπp(k) + 1 dz πp(z) P h log(p(z|s)) log(p(z)) | {z } Mutual Information +1 p(z|s) | {z } Gradient correction where the equality follows from writing the posterior as a function of the per-skill state occupancy dz π = p(s | z), and using Bayes rules, p(z|s) = dz π(s)p(z) P k dkπ(s)p(k). Replacing the posterior p(z|s) with a learnt discriminator qφ(z|s) recovers the mutual-information rewards of DIAYN, with additional terms 1 p(z | s) which we refer to as gradient correction terms. Inspecting the common scenario of a uniform prior over the latent variables, p(z) = 1/|Z|, we get that the expectation of the gradient correction term P z p(z)(1 p(z|s)) = 1 1/|Z| in each state. From the perspective of the policy player, adding a constant to the reward does not change the best response policy, nor the optimistic policy. Therefore, the gradient correction term does not have an effect on the optimization under a uniform prior, and we retrieved the reward of DIAYN. These algorithms differ however for more general priors p(z), which we explore empirically in Appendix F. GAIL. We further show how Eq. (10) extends to GAIL [31] via a simple construction. Consider a binary latent space of size |Z| = 2, where z = 1 corresponds to the policy of the agent and z = 2 corresponds to the policy of the expert which is fixed. In addition, consider a uniform prior over the latent variables, i.e., p(z = 1) = 1 2. By removing the constant terms in Eq. (11), one retrieves the GAIL [31] algorithm. The cost log(p(z|s)) is the probability of the discriminator to identify the agent, and the policy player is MDPO (which is similar to TRPO in GAIL). 7 Discussion In this work we reformulated the convex MDP problem as a convex-concave game between the agent and another player that is producing costs (negative rewards) and proposed a meta-algorithm for solving it. We observed that many algorithms in the literature can be interpreted as instances of the metaalgorithm by selecting different pairs of subroutines employed by the policy and cost players. The Frank-Wolfe algorithm, which combines best response with FTL, was originally proposed for AL [1, 75] but can be used for any convex MDP problem as was suggested in [30]. Zhang et al. [77], unified the problems of RL, AL, constrained MDPs with linear constraints and maximum entropy exploration under the framework of convex MDPs. We extended the framework to allow convex constraints (Section 5) and explained the objective of GAIL as a convex MDP (Section 6.2). We also discussed non convex objectives (Section 3) and analyzed unsupervised skill discovery via the maximization of mutual information (Section 6.2) as a special case. Finally, we would like to point out a recent work by Geist et al. [25], which was published concurrently to ours, and studies the convex MDP problem from the viewpoint of mean field games. There are also algorithms for convex MDPs that cannot be explained as instances of Algorithm 1. In particular, Zhang et al. [77] proposed a policy gradient algorithm for convex MDPs in which each step of policy gradient involves solving a new saddle point problem (formulated using the Fenchel dual). This is different from our approach since we solve a single saddle point problem iteratively, and furthermore we have much more flexibility about which algorithms the policy player can use. Moreover, for the convergence guarantee [77, Theorem 4.5] to hold, the saddle point problem has to be solved exactly, while in practice it is only solved approximately [77, Algorithm 1], which hinders its sample efficiency. Fenchel duality has also been used in off policy evaluation (OPE) in [46, 74]. The difference between these works and ours is that we train a policy to minimize an objective, while in OPE a target policy is fixed and its value is estimated from data produced by a behaviour policy. In order to solve a practical convex MDP problem in a given domain it would be prudent to use an RL algorithm that is known to be high performing for the vanilla RL problem as the policy player. From the theoretical point of view this could be MDPO or UCRL2, which we have shown come with strong guarantees. From the practical point of view using a high performing DRL algorithm, which may be specific to the domain, will usually yield the best results. For the cost player using FTL, i.e., using the gradient of the objective function, is typically the best choice. Acknowledgments and Disclosure of Funding We would like to thank Yasin Abbasi-Yadkorie, Vlad Mnih, Jacob Abernethy, Lior Shani and Doina Precup for their comments and discussion on this work. Work done at Deep Mind, the authors received no specific funding for this work. [1] P. Abbeel and A. Y. Ng. Apprenticeship learning via inverse reinforcement learning. In Proceedings of the twenty-first international conference on Machine learning, page 1. ACM, 2004. [2] A. Abdolmaleki, J. T. Springenberg, Y. Tassa, R. Munos, N. Heess, and M. Riedmiller. Maximum a posteriori policy optimisation. ar Xiv preprint ar Xiv:1806.06920, 2018. [3] J. D. Abernethy and J.-K. Wang. On frank-wolfe and equilibrium computation. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL https: //proceedings.neurips.cc/paper/2017/file/7371364b3d72ac9a3ed8638e6f0be2c9-Paper.pdf. [4] J. Achiam, H. Edwards, D. Amodei, and P. Abbeel. Variational option discovery algorithms. ar Xiv preprint ar Xiv:1807.10299, 2018. [5] A. Agarwal, S. M. Kakade, J. D. Lee, and G. Mahajan. Optimality and approximation with policy gradient methods in markov decision processes. In Conference on Learning Theory, pages 64 66. PMLR, 2020. [6] S. Agrawal and N. R. Devanur. Fast algorithms for online stochastic convex programming. In Proceedings of the twenty-sixth annual ACM-SIAM symposium on Discrete algorithms, pages 1405 1424. SIAM, 2014. [7] E. Altman. Constrained Markov decision processes, volume 7. CRC Press, 1999. [8] M. G. Azar, I. Osband, and R. Munos. Minimax regret bounds for reinforcement learning. In International Conference on Machine Learning, pages 263 272, 2017. [9] A. Beck and M. Teboulle. Mirror descent and nonlinear projected subgradient methods for convex optimization. Operations Research Letters, 31:167 175, 2003. [10] S. Belogolovsky, P. Korsunsky, S. Mannor, C. Tessler, and T. Zahavy. Inverse reinforcement learning in contextual mdps. Machine Learning, 2021. [11] S. Bhatnagar and K. Lakshmanan. An online actor critic algorithm with function approximation for constrained markov decision processes. Journal of Optimization Theory and Applications, 153(3):688 708, 2012. [12] V. S. Borkar. An actor-critic algorithm for constrained markov decision processes. Systems & control letters, 54(3):207 213, 2005. [13] S. Boyd and L. Vandenberghe. Convex optimization. Cambridge university press, 2004. [14] L. M. Bregman. The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming. USSR computational mathematics and mathematical physics, 7(3):200 217, 1967. [15] Q. Cai, Z. Yang, C. Jin, and Z. Wang. Provably efficient exploration in policy optimization. In International Conference on Machine Learning, pages 1283 1294. PMLR, 2020. [16] D. A. Calian, D. J. Mankowitz, T. Zahavy, Z. Xu, J. Oh, N. Levine, and T. Mann. Balancing constraints and rewards with meta-gradient d4{pg}. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=TQt98Ya7UMP. [17] C. Dann and E. Brunskill. Sample complexity of episodic fixed-horizon reinforcement learning. In Advances in Neural Information Processing Systems, pages 2818 2826, 2015. [18] Y. Efroni, S. Mannor, and M. Pirotta. Exploration-exploitation in constrained mdps. ar Xiv preprint ar Xiv:2003.02189, 2020. [19] L. Espeholt, H. Soyer, R. Munos, K. Simonyan, V. Mnih, T. Ward, Y. Doron, V. Firoiu, T. Harley, I. Dunning, S. Legg, and K. Kavukcuoglu. IMPALA: Scalable distributed deep-RL with importance weighted actor-learner architectures. In Proceedings of the 35th International Conference on Machine Learning, 2018. [20] B. Eysenbach, A. Gupta, J. Ibarz, and S. Levine. Diversity is all you need: Learning skills without a reward function. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=SJx63j Rq Fm. [21] C. Florensa, Y. Duan, and P. Abbeel. Stochastic neural networks for hierarchical reinforcement learning. In International Conference on Learning Representations, 2016. [22] M. Frank and P. Wolfe. An algorithm for quadratic programming. Naval research logistics quarterly, 3(1-2):95 110, 1956. [23] Y. Freund and R. E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. Journal of computer and system sciences, 55(1):119 139, 1997. [24] M. Geist, B. Scherrer, and O. Pietquin. A theory of regularized markov decision processes. In International Conference on Machine Learning, pages 2160 2169. PMLR, 2019. [25] M. Geist, J. Pérolat, M. Laurière, R. Elie, S. Perrin, O. Bachem, R. Munos, and O. Pietquin. Concave utility reinforcement learning: the mean-field game viewpoint. ar Xiv preprint ar Xiv:2106.03787, 2021. [26] K. Gregor, D. J. Rezende, and D. Wierstra. Variational intrinsic control. International Conference on Learning Representations, Workshop Track, 2017. URL https://openreview.net/forum? id=Skc-Fo4Yg. [27] K. Hausman, J. T. Springenberg, Z. Wang, N. Heess, and M. Riedmiller. Learning an embedding space for transferable robot skills. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=rk07ZXZRb. [28] E. Hazan. Introduction to online convex optimization. Foundations and Trends in Optimization, 2(3-4):157 325, 2016. [29] E. Hazan, A. Agarwal, and S. Kale. Logarithmic regret algorithms for online convex optimization. Machine Learning, 69(2-3):169 192, 2007. [30] E. Hazan, S. Kakade, K. Singh, and A. Van Soest. Provably efficient maximum entropy exploration. In International Conference on Machine Learning, pages 2681 2691. PMLR, 2019. [31] J. Ho and S. Ermon. Generative adversarial imitation learning. ar Xiv preprint ar Xiv:1606.03476, 2016. [32] R. Huang, T. Lattimore, A. György, and C. Szepesvári. Following the leader and fast rates in linear prediction: Curved constraint sets and other regularities. In Advances in Neural Information Processing Systems, pages 4970 4978, 2016. [33] M. Jaggi. Revisiting frank-wolfe: Projection-free sparse convex optimization. In Proceedings of the 30th international conference on Machine learning. ACM, 2013. [34] M. Jaggi and S. Lacoste-Julien. On the global linear convergence of frank-wolfe optimization variants. Advances in Neural Information Processing Systems, 28, 2015. [35] T. Jaksch, R. Ortner, and P. Auer. Near-optimal regret bounds for reinforcement learning. Journal of Machine Learning Research, 11(Apr):1563 1600, 2010. [36] Y. Jin and A. Sidford. Efficiently solving mdps with stochastic mirror descent. In International Conference on Machine Learning, pages 4890 4900. PMLR, 2020. [37] Y. Jin and A. Sidford. Towards tight bounds on the sample complexity of average-reward mdps. ar Xiv preprint ar Xiv:2106.07046, 2021. [38] E. Kaufmann, P. Ménard, O. D. Domingues, A. Jonsson, E. Leurent, and M. Valko. Adaptive reward-free exploration. In Algorithmic Learning Theory, pages 865 891. PMLR, 2021. [39] S. Kullback. Information theory and statistics. Courier Corporation, 1997. [40] T. Lattimore and M. Hutter. Pac bounds for discounted mdps. In International Conference on Algorithmic Learning Theory, pages 320 334. Springer, 2012. [41] L. Lee, B. Eysenbach, E. Parisotto, E. Xing, S. Levine, and R. Salakhutdinov. Efficient exploration via state marginal matching. ar Xiv preprint ar Xiv:1906.05274, 2019. [42] D. Levin, Y. Peres, and E. Wilmer. Markov Chains and Mixing Times. American Mathematical Society, 2017. [43] B. Mc Mahan. Follow-the-regularized-leader and mirror descent: Equivalence theorems and l1 regularization. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, pages 525 533. JMLR Workshop and Conference Proceedings, 2011. [44] S. Miryoosefi, K. Brantley, H. Daumé III, M. Dudík, and R. Schapire. Reinforcement learning with convex constraints. ar Xiv preprint ar Xiv:1906.09323, 2019. [45] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves, M. Riedmiller, A. K. Fidjeland, G. Ostrovski, et al. Human-level control through deep reinforcement learning. nature, 518(7540):529 533, 2015. [46] O. Nachum, Y. Chow, B. Dai, and L. Li. Dualdice: Behavior-agnostic estimation of discounted stationary distribution corrections. ar Xiv preprint ar Xiv:1906.04733, 2019. [47] A. S. Nemirovskij and D. B. Yudin. Problem complexity and method efficiency in optimization. In Wiley-Interscience, 1983. [48] B. O Donoghue. Variational Bayesian reinforcement learning with regret bounds. ar Xiv preprint ar Xiv:1807.09647, 2018. [49] B. O Donoghue, T. Lattimore, and I. Osband. Stochastic matrix games with bandit feedback. ar Xiv preprint ar Xiv:2006.05145, 2020. [50] B. O Donoghue, I. Osband, and C. Ionescu. Making sense of reinforcement learning and probabilistic inference. In International Conference on Learning Representations, 2020. [51] I. Osband, D. Russo, and B. Van Roy. (More) efficient reinforcement learning via posterior sampling. In Advances in Neural Information Processing Systems, pages 3003 3011, 2013. [52] I. Osband, C. Blundell, A. Pritzel, and B. V. Roy. Deep exploration via bootstrapped dqn. In Proceedings of the 30th International Conference on Neural Information Processing Systems, pages 4033 4041, 2016. [53] I. Osband, Y. Doron, M. Hessel, J. Aslanides, E. Sezener, A. Saraiva, K. Mc Kinney, T. Lattimore, C. Szepesvari, S. Singh, et al. Behaviour suite for reinforcement learning. In International Conference on Learning Representations, 2019. [54] M. J. Osborne and A. Rubinstein. A course in game theory. MIT press, 1994. [55] M. L. Puterman. Markov decision processes: discrete stochastic dynamic programming. John Wiley & Sons, 1984. [56] R. T. Rockafellar. Convex analysis. Princeton university press, 1970. [57] A. Rosenberg and Y. Mansour. Online convex optimization in adversarial markov decision processes. In International Conference on Machine Learning, pages 5478 5486. PMLR, 2019. [58] J. Schulman, S. Levine, P. Abbeel, M. Jordan, and P. Moritz. Trust region policy optimization. In International conference on machine learning, pages 1889 1897. PMLR, 2015. [59] L. Shani, Y. Efroni, and S. Mannor. Adaptive trust region policy optimization: Global convergence and faster rates for regularized mdps. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 5668 5675, 2020. [60] L. Shani, Y. Efroni, A. Rosenberg, and S. Mannor. Optimistic policy optimization with bandit feedback. In International Conference on Machine Learning, pages 8604 8613. PMLR, 2020. [61] L. Shani, T. Zahavy, and S. Mannor. Online apprenticeship learning. ar Xiv preprint ar Xiv:2102.06924, 2021. [62] D. Silver, J. Schrittwieser, K. Simonyan, I. Antonoglou, A. Huang, A. Guez, T. Hubert, L. Baker, M. Lai, A. Bolton, et al. Mastering the game of go without human knowledge. nature, 550 (7676):354 359, 2017. [63] A. L. Strehl and M. L. Littman. An analysis of model-based interval estimation for markov decision processes. Journal of Computer and System Sciences, 74(8):1309 1331, 2008. [64] R. S. Sutton and A. G. Barto. Reinforcement learning: An introduction. MIT press, 2018. [65] U. Syed and R. E. Schapire. A game-theoretic approach to apprenticeship learning. In Advances in neural information processing systems, pages 1449 1456, 2008. [66] U. Syed, M. Bowling, and R. E. Schapire. Apprenticeship learning using linear programming. In Proceedings of the 25th international conference on Machine learning, pages 1032 1039. ACM, 2008. [67] C. Szepesvári. Constrained mdps and the reward hypothesis, 2020. URL https://readingsml. blogspot.com/2020/03/constrained-mdps-and-reward-hypothesis.html. [68] C. Tessler, D. J. Mankowitz, and S. Mannor. Reward constrained policy optimization. In International Conference on Learning Representations, 2019. URL https://openreview.net/ forum?id=Skfrvs A9FX. [69] D. Tirumala, A. Galashov, H. Noh, L. Hasenclever, R. Pascanu, J. Schwarz, G. Desjardins, W. M. Czarnecki, A. Ahuja, Y. W. Teh, et al. Behavior priors for efficient reinforcement learning. ar Xiv preprint ar Xiv:2010.14274, 2020. [70] M. Tomar, L. Shani, Y. Efroni, and M. Ghavamzadeh. Mirror descent policy optimization. ar Xiv preprint ar Xiv:2005.09814, 2020. [71] J. Von Neumann. Zur theorie der gesellschaftsspiele. Mathematische annalen, 100(1):295 320, 1928. [72] C. J. Watkins and P. Dayan. Q-learning. Machine learning, 8(3-4):279 292, 1992. [73] H. Xiao, M. Herman, J. Wagner, S. Ziesche, J. Etesami, and T. H. Linh. Wasserstein adversarial imitation learning. ar Xiv preprint ar Xiv:1906.08113, 2019. [74] M. Yang, O. Nachum, B. Dai, L. Li, and D. Schuurmans. Off-policy evaluation via the regularized lagrangian. ar Xiv preprint ar Xiv:2007.03438, 2020. [75] T. Zahavy, A. Cohen, H. Kaplan, and Y. Mansour. Apprenticeship learning via frank-wolfe. AAAI, 2020, 2020. [76] T. Zahavy, A. Cohen, H. Kaplan, and Y. Mansour. Average reward reinforcement learning with unknown mixing times. The Conference on Uncertainty in Artificial Intelligence (UAI), 2020. [77] J. Zhang, A. Koppel, A. S. Bedi, C. Szepesvari, and M. Wang. Variational policy gradient method for reinforcement learning with general utilities. ar Xiv preprint ar Xiv:2007.02151, 2020. [78] M. Zhang, Y. Wang, X. Ma, L. Xia, J. Yang, Z. Li, and X. Li. Wasserstein distance guided adversarial imitation learning with reward shape exploration. In 2020 IEEE 9th Data Driven Control and Learning Systems Conference (DDCLS), pages 1165 1170. IEEE, 2020. [79] M. Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In Proceedings of the 20th international conference on machine learning (icml-03), pages 928 936, 2003.