# robust_phidivergence_mdps__dc74f46d.pdf Robust ϕ-Divergence MDPs Chin Pang Ho City University of Hong Kong clint.ho@cityu.edu.hk Marek Petrik University of New Hampshire mpetrik@cs.unh.edu Wolfram Wiesemann Imperial College London ww@imperial.ac.uk In recent years, robust Markov decision processes (MDPs) have emerged as a prominent modeling framework for dynamic decision problems affected by uncertainty. In contrast to classical MDPs, which only account for stochasticity by modeling the dynamics through a stochastic process with a known transition kernel, robust MDPs additionally account for ambiguity by optimizing in view of the most adverse transition kernel from a prescribed ambiguity set. In this paper, we develop a novel solution framework for robust MDPs with s-rectangular ambiguity sets that decomposes the problem into a sequence of robust Bellman updates and simplex projections. Exploiting the rich structure present in the simplex projections corresponding to ϕ-divergence ambiguity sets, we show that the associated s-rectangular robust MDPs can be solved substantially faster than with state-of-the-art commercial solvers as well as a recent first-order solution scheme, thus rendering them attractive alternatives to classical MDPs in practical applications. 1 Introduction Markov decision processes (MDPs) are a flexible and popular framework for dynamic decisionmaking problems and reinforcement learning [39, 48]. A practical limitation of the standard MDP model is that it assumes the model parameters, such as transition probabilities and rewards, to be known exactly. In reinforcement learning and other applications, these parameters must be estimated from sampled data, which introduces estimation errors. Optimal MDP solutions, referred to as policies, are well known to be sensitive to errors and may fail catastrophically when deployed [25, 56]. Robust MDPs (RMDPs) mitigate the sensitivity of MDPs to estimation errors by computing a policy that is optimal for the worst plausible realization of the transition probabilities. This set of plausible transition probabilities is known as the ambiguity set. Most prior work considers ambiguity sets that are rectangular. In this work, we focus on s-rectangular ambiguity sets, which assume that the worst transition probabilities are chosen independently in each state [25, 56]. While several other models of rectangularity have been studied [9, 13, 21, 28], s-rectangular ambiguity sets are popular due to their generality and their performance both in-sample and out-of-sample [56]. However, even though polynomial-time algorithms based on dynamic programming concepts have been developed for s-rectangular sets, those algorithms may be too slow in practice. Solving RMDPs requires the solution of a convex optimization problem in every step of value or policy iteration, which can become prohibitively slow even in moderatly sized problems with 100s of states [5, 9, 14, 19]. Motivated by the difficulty of solving RMDPs, several fast algorithms have been proposed for srectangular RMDPs [5, 9, 14, 19]. The preponderance of the earlier work has focused on ambiguity sets defined in terms of L1and L -norms. These ambiguity sets are polyhedral, and they can be analyzed using linear programming techniques which offer fruitful avenues to exploit the structure inherent to those sets. However, recent statistical studies point to the superior solution quality offered by nonlinear ambiguity sets defined in terms of the Kullback-Leibler (KL) divergence, the L2-norm and other metrics [17]. Linear optimization solvers are not applicable to RMDPs with s-rectangular ambiguity sets defined in terms of non-polyhedral ambiguity sets, as the corresponding optimization 36th Conference on Neural Information Processing Systems (Neur IPS 2022). problems are in general convex conic programs (e.g., exponential cone program in the case of KL divergence); thus, they are currently solved using first-order methods [14] or general convex conic solvers such as MOSEK [3], which tend to be complex, closed-source and slow. As our main contribution, we propose a new suite of fast algorithms for solving RMDPs with ϕdivergence constrained s-rectangular ambiguity sets. ϕ-divergences, also known as f-divergences, constitute a generalization of the KL divergence that encompasses the Burg entropy as well as the L1and weighted L2-norms as special cases [4, 6]. Moreover, ϕ-divergence ambiguity sets benefit from rigorous statistical performance guarantees, and they are optimal among all (known and unknown) data-driven optimization paradigms for certain types of worst-case out-of-sample performance guarantees [35]. The radii of ϕ-divergence ambiguity sets can be selected either via crossvalidation or via statistical bounds [26, 32, 55]. Robust MDPs with ϕ-divergence sets are challenging and unexplored for both (s, a)-rectangular and s-rectangular ambiguity sets. Solving ϕ-divergence RMDPs using value iteration requires the solution of seemingly unstructured min-max problems. Our main insight is that these min-max problems can be reduced to a small number of highly structured projection problems onto a probability simplex. We use this insight to develop tailored solution schemes for the projection problems corresponding to several popular ϕ-divergence ambiguity sets, which in turn give rise to efficient solution methods for the respective RMDPs. Ignoring tolerances, our algorithms achieve an overall O(S2 A log A) or O(S2 log S A) time complexity to compute the robust Bellman operator, where S and A denote the numbers of states and actions, respectively. Since the evaluation of a non-robust Bellman operator requires a runtime of O(S2 log A), our algorithms only incur an additional logarithmic overhead to account for robustness in the transition probabilities. This computational complexity compares favorably with the larger time complexity of a recent first-order solution scheme for KL divergence-constrained s-rectangular RMDPs (which we will elaborate on later in the paper) as well as a minimum complexity of O(S4.5 A) for the naïve solution with state-of-the-art interior-point algorithms. Our framework is general enough to readily accommodate for ϕ-divergences that have not been studied previously in the context of s-rectangular ambiguity sets, such as the Burg entropy and the χ2-distance. For other ϕ-divergences, such as the L1-norm, our framework results in the same complexity at substantially simplified proofs. The algorithms developed in this paper speed up the computation of robust Bellman updates, and so they can be used in combination with a variety of RMDP solution schemes. In particular, they can be used to accelerate the standard robust value iteration, policy iteration, modified policy iteration [22] and partial policy iteration [19]. They can also be combined with a first order gradient method [14] that has been introduced recently. In addition, fast algorithms for computing the Bellman operator also play a crucial role when scaling robust algorithms to value function approximation [50], model-free reinforcement learning [33, 42], and robust policy gradients [49]. In this paper, we focus on the model-based setting, which is currently under active study [24, 29, 31] and has many important reallife applications [12, 20, 58]; moreover, it also serves as an important building block to constructing model-free algorithms. While this paper focuses on the s-rectangular ambiguity sets, the proposed algorithms in this paper can also be applied to the special case of (s, a)-rectangular ambiguity sets. The remainder of the paper proceeds as follows. Section 2 reviews relevant prior work and Section 3 describes our basic RMDP setting. Then, Section 4 shows how the robust Bellman operator for a large class of ambiguity sets can be reduced to a sequence of structured projections onto a simplex. We describe novel algorithms for efficiently computing the simplex projections for several ϕ-divergences in Section 5. Finally, Section 6 presents experimental results that compare the runtime of our algorithms with general conic solvers as well as a recent first-order optimization algorithm [14]. Notation. We denote by e the vector of all ones, whose context determines its dimension. We refer to the probability simplex in Rn by n = {p Rn + : e p = 1}. For x Rn, we let min{x} = min{xi : i = 1, . . . , n} (similar for the maximum operator), and we define [x]+ Rn + component-wise as ([x]+)i = max{xi, 0}, i = 1, . . . , n. We refer to the conjugate of a function f : Rn R by f (y) = sup{y x f(x) : x Rn}. Random variables are indicated by a tilde. 2 Related Work While RMDPs have been studied since the seventies [45], they have witnessed significant recent interest due to their widespread adoption in applications ranging from assortment optimization [43], medical decision-making [12, 62] and hospital operations management [16], production planning [58] and energy systems [20] to model predictive control [10], aircraft collision avoidance [23], wireless communications [57] and the robustification against approximation errors in aggregated MDPs [37]. Efficient implementations of the robust value iteration have been first proposed by [11, 21, 32] for RMDPs with (s, a)-rectangular ambiguity sets, where the worst transition probabilities are considered separately for each state and action. The authors study ambiguity sets that bound the distance of the transition probabilities to some nominal distribution in terms of finite scenarios, interval matrix bounds, ellipsoids, the relative entropy, the KL divergence and maximum a posteriori models. Subsequently, similar methods have been developed by [57] for interval matrix bounds as well as likelihood uncertainty models, by [37] for 1-norm ambiguity sets as well as by [62] for interval matrix bounds intersected with a budget constraint. All of these contributions have in common that they focus on (s, a)-rectangular ambiguity sets where the existence of optimal deterministic policies is guaranteed, and it is not clear how they could be extended to the more general class of s-rectangular ambiguity sets where all optimal policies may be randomized. In contrast to (s, a)-rectangular ambiguity sets, s-rectangular ambiguity sets restrict the conservatism among transition probabilities corresponding to different actions in the same state, which tends to lead to a superior performance in data-driven settings. [56] solve the subproblems arising in the robust value iteration of an s-rectangular RMDP as linear or conic optimization problems using commercial off-the-shelf solvers. Despite their polynomial-time complexity, general-purpose solvers cannot exploit the structure present in these subproblems, which renders them suitable primarily for small problem instances. More efficient tailored solution methods for s-rectangular RMDPs have subsequently been developed by [5, 18, 19]. [18] develop a homotopy continuation method for RMDPs with (s, a)-rectangular and s-rectangular weighted 1-norm ambiguity sets, while [5] adapt the algorithm of [18] to unweighted -norm ambiguity sets. [19] embed the algorithms of [18] in a partial policy iteration, which generalizes the robust modified policy iteration proposed by [22] for (s, a)-rectangular RMDPs to s-rectangular RMDPs. While the present paper focuses on the robust value iteration for ease of exposition, we note that our algorithms can also be combined with the partial policy iteration of [19] to obtain further speedups. [9] establish a relationship between s-rectangular RMDPs and twice regularized MDPs, which they subsequently use to propose efficient Bellman updates for a modified policy iteration. While their approach can solve RMDPs in almost the same time as a classical non-robust MDPs, the obtained policies can be conservative as the worst-case transition probabilities are not restricted to reside in a probability simplex and, therefore, may be negative and/or add up to more or less than 1. Finally, [14] propose a first-order framework for RMDPs with s-rectangular KL and spherical ambiguity sets that interleaves primal-dual first-order updates with approximate value iteration steps. The authors show that their algorithms outperform a robust value iteration that solves the emerging subproblems using state-of-the-art commercial solvers. We compare our solution method for KL ambiguity sets with the approach proposed by [14] in terms of its theoretical complexity and numerical runtimes. While this paper exclusively studies s-rectangular uncertainty sets, alternative generalizations of (s, a)-rectangular ambiguity sets have been proposed in the literature as well. For example, [28] consider k-rectangular ambiguity sets where the transition probabilities of different states can be coupled, [13] study factor ambiguity model ambiguity sets where the transition probabilities depend on a small number of underlying factors, and [51] construct ambiguity sets that bound marginal moments of state-action features defined over entire MDP trajectories. Moving beyond the modelbased setting, there is also an active line of research on robust reinforcement learning, such as least squares policy iteration [33], analysis on sample complexity [34], robust Q-learning and robust TDC algorithms [42, 53], and robust policy gradient [54]. We also note the papers [7, 15, 60] which study the related problem of distributionally robust MDPs whose transition probabilities are themselves regarded as random objects that are drawn from distributions which are only partially known. The connections between RMDPs and multi-stage stochastic programs as well as distributionally robust problems are explored further by [44, 46, 47]. 3 Preliminaries Robust MDPs We study RMDPs with a finite state space S = {1, . . . , S} and a finite action space A = {1, . . . , A}. We assume an infinite planning horizon, but all of our results immediately extend to a finite time horizon. Without loss of generality, we assume that every action a A is admissible in every state s S. The RMDP starts in a random initial state s0 that follows the known probability distribution p0 from the probability simplex S in RS. If action a A is taken in state s S, then the RMDP transitions randomly to the next state according to the conditional probability distribution psa S. We condense the transition probabilities psa to the tensor p ( S)S A. The transition probabilities are only known to reside in a non-empty, compact ambiguity set P ( S)S A. For a transition from state s S to state s S under action a A, the decision maker receives an expected reward of rsas R+. As with the transition probabilities, we condense these rewards to the tensor r RS A S + . Without loss of generality, we assume that all rewards are non-negative. We denote by Π = ( A)S the set of all stationary (i.e., time-independent) randomized policies. A policy π Π takes action a A in state s S with probability πsa. The transition probabilities p P and the policy π Π induce a stochastic process {( st, at)} t=0 on the space (S A) of sample paths. We refer by Ep,π to expectations with respect to this process. The decision maker is risk-neutral but ambiguity-averse and wishes to maximize the worst-case expected total reward under a discount factor λ (0, 1), max π Π min p P Ep,π " X t=0 λt r st, at, st+1 s0 p0 # Note that the maximum and minimum in (1) are both attained by the Weierstrass theorem since Π and P are non-empty and compact, while the objective function is finite since λ < 1. Rectangular Ambiguity Sets For general ambiguity sets P, evaluating the inner minimization in (1) is NP-hard even if the policy π Π is fixed [56]. For these reasons, much of the research on RMDPs and their applications has focused on rectangular ambiguity sets. Among the most general rectangular ambiguity sets are the s-rectangular ambiguity sets P satisfying P = p ( S)S A : ps Ps s S , where Ps ( S)A, s S, see [25, 56, 59, 61]. In contrast to the simpler class of (s, a)-rectangular ambiguity sets, s-rectangular ambiguity sets restrict the choice of transition probabilities ps1, . . . , ps A corresponding to different actions a applied in the same state s. This limits the conservatism of the resulting RMDP (1) and typically leads to a better performance of the optimal policy [56]. Although Bellman s optimality principle extends to s-rectangular RMDPs and there is always an optimal stationary policy, all optimal policies of an s-rectangular RMDP may be randomized. We study a new general class of s-rectangular ambiguity sets that can be expressed as ps ( S)A : X a A da(psa, psa) κ where κ R+ is the uncertainty budget and the distance functions da(psa, psa), a A, are ϕ-divergences (also known as f-divergences) satisfying da(psa, psa) = X psas ϕ psas Here, ϕ: R+ R+ is a convex function satisfying ϕ(1) = 0. Intuitively, a ϕ-divergence measures the distance between two probability distributions. With an appropriate choice of ϕ, it generalizes other metrics including the KL divergence, the Burg entropy, L1and L2-norms and others [4, 6]. Table 1 reports some popular ϕ-divergences that we study in this paper. The variation distance coincides with the L1-based s-rectangular ambiguity sets studied in earlier work [18, 19]. Note that although we assume that ϕ is the same for different state-action pairs, the proposed approach also works for the more general case where da(psa, psa) = P s S psas ϕsas (psas /psas ). Robust Value Iteration A standard approach for computing the optimal value and the optimal policy of an RMDP (1) is the robust value iteration [21, 32, 25, 56]: Starting with an initial estimate v0 RS of the state-wise optimal value to-go, we conduct robust Bellman iterations of the form vt+1 J(vt), t = 0, 1, . . ., where the robust Bellman operator J is defined component-wise as [J(v)]s = max πs A min ps Ps a A πsa psa (rsa + λv) s S. (3) Divergence da(psa, psa) ϕ(t) Complexity of J Prior Art Kullback-Leibler P s psas log psas psas t log t t + 1 O(S2A log A) O(ℓ2S2A) Burg Entropy P log t + t 1 O(S2A log A) no poly-time Variation Distance P s |psas psas | |t 1| O(S2 log SA) O(S2 log SA) χ2-Distance P s (psas psas )2 psas (t 1)2 O(S2 log SA) O(S4.5A) Table 1: Summary of the ϕ-divergences studied in this paper, together with the complexity of our robust Bellman operator J (applied across all states s S) as well as the best known results from the literature. The complexity estimates omit constants and tolerances that are reported in Section 5 of the paper. ℓ , where present, refers to the number of Bellman iterations conducted so far. This yields the optimal value p0 v , where the limit v = limt vt is approached component-wise at a geometric rate. The optimal policy π Π, finally, is recovered state-wise via π s arg max πs A min ps Ps a A πsa psa (rsa + λv ) s S. 4 Robust Bellman Updates via Simplex Projections In this section, we show that the robust Bellman operator J reduces to a generalized projection problem. This reduction is important because it underlies our fast algorithms for computing J. At the core of the robust value iteration is the solution of the max-min problem (3). By applying the minimax theorem, the right-hand side of (3) equals to minimize max a A psa (rsa + λv) subject to X a A da(psa, psa) κ The above optimization problem can be solved via bisection on its objective value; that is, we seek the lowest possible β such that psa (rsa + λv) β, for each a A where ps satisfies the constraints in (4). For any given β in the bisection method, we check whether such ps exists by solving the following generalized da-projection problem of the nominal transition probabilities psa: P(psa; b, β) = minimize da(psa, psa) subject to b psa β psa S Here, psa S are the decision variables and psa S, b RS + and β R+ are parameters. For any fixed β, we compute P a A P(psa; rsa + λv, β) and construct ps ( S)A where psa is the optimal solution of the associated projection problem P(psa; b, β). We can then distinguish between the following two cases: a A P(psa; rsa + λv, β) κ, then the constructed ps is a feasible solution to (5). Therefore, β upper bounds the optimal objective value. a A P(psa; rsa + λv, β) > κ, then there is no feasible ps ( S)A such that the objective value attains β or less. Therefore, β lower bounds the optimal objective value. As a consequence of the cases above, one can compute the robust Bellman update (3) efficiently by bisection if the projection problem (5) can be solved efficiently. The proof of Theorem 1 describes further details needed to implement this algorithm, including the initial upper and lower bounds on β and bounds on the precision needed when solving the projection problems. Note that problem (5) is infeasible if and only if min{b} > β. Moreover, problem (5) is trivially solved by psa with an optimal objective value of 0 whenever b psa β. To avoid these trivial cases, Figure 1: The generalized da-projection problem (5) in S = 3 dimensions (a) and two-dimensional projections for the variation distance (b), the χ2-distance (c) and the KL divergence (d). The gray shaded areas represent the probability simplex S, the red dashed lines show the boundary of the intersection of the halfspace b psa β with the probability simplex, and the white shapes illustrate contour lines centered at the nominal transition probabilities psa. we assume throughout the paper that min{b} β and b psa > β. We illustrate the feasible region and optimal solution to problem (5) for different ϕ-divergences in Figure 1. Our generalized da-projection (5) relates to the rich literature on projections onto simplices, which we review in the next section. In fact, our algorithms in the next section solve a variant of the simplex projection problem that is restricted by an additional inequality constraint. We therefore believe that our algorithms may find additional applications outside the RMDP literature. In the following, we say that for a given estimate vt RS of the optimal value function, the robust Bellman iteration (3) is solved to ϵ-accuracy by any vt+1 RS satisfying vt+1 J(vt) ϵ. We seek ϵ-optimal solutions because our ambiguity sets are nonlinear and hence the exact Bellman iterate J(vt) may be irrational even if vt is rational. To simplify the exposition, we define R = [1 λ] 1 max{rsas : s, s S, a A} as an upper bound on all [J(v)]s, v v and s S. For divergence-based ambiguity sets, the projection problem (5) is generically nonlinear and can hence not be expected to be solved to exact optimality. To account for this additional complication, we say that for a given psa S, b RS + and β R+, the generalized da-projection P(psa; b, β) is solved to δ-accuracy by any pair (d, d) R2 satisfying P(psa; b, β) [d, d] and d d δ. Theorem 1. Assume that the generalized da-projection (5) can be computed to any accuracy δ > 0 in time O(h(δ)). Then the robust Bellman iteration (3) can be computed to any accuracy ϵ > 0 in time O(AS h(ϵκ/[2AR + Aϵ]) log[R/ϵ]). Theorem 1 reduces the evaluation of the robust Bellman iterator J, which involves the solution of a max-min optimization problem over an s-rectangular ambiguity set that couples all actions a A, to a sequence of much simpler and highly structured projection problems that are no longer coupled across different actions a A. The next section describes efficient solution schemes for the projection problem (5) in the context of several ϕ-divergence ambiguity sets. The runtimes of these solution schemes are summarized in Table 1. Note that the evaluation of a non-robust Bellman operator requires a runtime of O(S2 log A), which implies that our algorithms only incur an additional logarithmic overhead to account for robustness in the transition probabilities. 5 Fast Projections on ϕ-Divergence Simplices We next describe fast algorithms for computing generalized projections onto the probability simplex. Combined with the results from Section 4, these algorithms can be used to efficiently compute the robust Bellman operator. Note that some ϕ-divergences, such as the KL divergence and the χ2-distance, imply that if psas = 0 for some s, s S and a A, then psas = 0 for all psa S with da(psa, psa) < , and thus we can remove indices s with psas = 0. For other ϕ-divergences, such as the Burg entropy and the variation distance, one can readily verify that our results remain valid no matter whether psa > 0 or not, but the formulations and proofs require additional case distinctions and/or limit arguments. To simplify the exposition, we therefore assume that psa > 0. Proposition 1. For the distance function da(psa, psa) = P s S psas ϕ psas psas , the optimal value of the projection problem (5) equals the optimal value of the bivariate convex problem maximize βα + ζ X psas ϕ ( αbs + ζ) subject to α R+, ζ R. (6) Proposition 1 reduces the S-dimensional projection problem (5) to a two-dimensional optimization problem over the dual variables α and ζ. In the following, we show that for the ϕ-divergences from Table 1, problem (6) can be further simplified to univariate convex optimization problems that can be solved efficiently via bisection, binary search or sorting. 5.1 Kullback-Leibler Divergence We first show that for the KL divergence ϕ(t) = t log t t + 1, the reduced projection problem (6) can be further simplified to a univariate convex optimization problem. Proposition 2. For the KL divergence ϕ(t) = t log t t + 1, the optimal value of the projection problem (5) equals the optimal value of the univariate convex problem maximize α R+ βα log psas e αbs ! We next show that the univariate optimization problem (2) admits an efficient solution via bisection. Theorem 2. If β min{b} + ω for some ω > 0, then the projection problem (5) can be solved to any δ-accuracy in time O(S log[max{b} log(min{p} 1)/(δω)]). Note that the projection problem (5) is infeasible whenever β < min{b}. The condition in the statement of Theorem 2 can thus be interpreted as a strict feasibility requirement. It is worth contrasting the result of Theorem 2 with the solution of the projection problem (5) as an exponential cone program. The latter would result in a practical complexity of O(S3), assuming that which is often observed in practice the number of iterations of the employed interior-point solver does not grow with the problem dimensions. A theoretically guaranteed complexity, on the other hand, does not seem to be available at present as the commercial state-of-the-art solvers for exponential conic programs are not proven to terminate in polynomial time. Corollary 1. The robust Bellman iteration (3) over a KL divergence ambiguity set can be computed to any accuracy ϵ > 0 in time O(S2 A log A log[R 2 log(min{p} 1)/(ϵ2κ)] log[R/ϵ]). [14] propose a first-order framework for RMDPs over s-rectangular KL divergence ambiguity sets whose robust Bellman update enjoys a complexity of O(ℓ2 S2 A log(ϵ 1)), where ℓis the iteration number. A careful analysis results in an overall convergence rate for the optimal MDP policy of O(S3 A2 ϵ 1 log[ϵ 1]). In contrast, the convergence rate of our robust value iteration amounts to O(S2 A log A log[R 2 log(min{p} 1)/(ϵ2κ)] log[R/ϵ] log[ϵ 1]). Treating the problem parameters R, p and κ as constants, our convergence rate simplifies to O(S2 A log A log[ϵ 2] log2[ϵ 1]), which compares favourably against the convergence rate of the first-order scheme. Our numerical results in Section 6 show that this theoretical difference appears to carry over to a favourable empirical performance on test instances as well. We finally note the related work [1], which optimizes a linear function over the intersection of a probability simplex with a constraint on the KL divergence to a nominal distribution. While one could in principle modify that algorithm to solve our projection problem (5), the resulting algorithm would require an additional bisection and would thus be significantly slower than ours. 5.2 Burg Entropy Similar to the KL divergence, the reduced projection problem (6) can be further simplified to a univariate convex optimization problem for the Burg entropy ϕ(t) = log t + t 1. Proposition 3. For the Burg entropy ϕ(t) = log t + t 1, if β > min{b}, then the optimal value of the projection problem (5) equals the optimal value of the univariate convex problem maximize α [0,1] psas log 1 + α bs β β min{b} Similar to the KL divergence, the univariate optimization problem (8) can be solved efficiently. Theorem 3. If β min{b} + ω for some ω > 0, then the projection problem (5) can be solved to any δ-accuracy in time O(S log[max{b}/(δω)]). As with the KL divergence, the projection problem (5) corresponding to the Burg entropy can be solved in a practical complexity of O(S3) as an exponential cone program, whereas we are not aware of any state-of-the-art solvers equipped with theoretical guarantees. To our best knowledge, RMDPs with s-rectangular Burg entropy ambiguity sets have not been studied previously in the literature. Corollary 2. The robust Bellman iteration (3) over a Burg entropy ambiguity set can be computed to any accuracy ϵ > 0 in time O(S2 A log A log[R 2/(ϵ2κ)] log[R/ϵ]). Similar to the previous subsection, we note that the related paper [1] optimizes a linear function over the intersection of a probability simplex with a bound on the Burg entropy to a nominal distribution. While that algorithm could in principle be employed to solve our projection problem (5), the resulting solution scheme would not be competitive due to the inclusion of an additional bisection. 5.3 Variation Distance We first provide an equivalent univariate optimization problem for the reduced projection problem (6) corresponding to the variation distance ϕ(t) = |t 1|. Proposition 4. For the variation distance ϕ(t) = |t 1|, the optimal value of the projection problem (5) equals the optimal value of the univariate convex problem maximize α R+ 2 + α(min{b} β) X psas [2 + α (min{b} bs )]+ . (9) Once more, the univariate optimization problem (9) admits an efficient solution. Theorem 4. The projection problem (5) can be solved exactly in time O(S log S). Note that in contrast to the previous results, Theorem 4 employs a binary search and thus offers an exact solution to the projection problem (5). Our result of Theorem 4 matches the complexity of the homotopy continuation method proposed by [19]. The correctness and runtime of their algorithm, however, relies on lengthy ad hoc arguments, whereas Theorem 4 relies on the groundwork laid by Theorem 1 and Proposition 1. Problem (5) can also be solved as a linear program with a practical complexity of O(S3) and a theoretical complexity of O(S3.5). Corollary 3. The robust Bellman iteration (3) over a variation distance ambiguity set can be computed to any accuracy ϵ > 0 in time O(S2 log S A log[R/ϵ]). [40] study the related problem of optimizing a linear function over the intersection of a probability simplex with an unweighted 1-norm constraint, and they identify structural properties of the optimal solutions. Since the linear function and the norm constraint are in different places of the optimization problem, however, their findings are not directly applicable to our setting. 5.4 χ2-Distance In contrast to the previous subsections, we directly solve the bivariate problem (6) for the χ2-distance ϕ(t) = (t 1)2 without first formulating an associated univariate optimization problem. Theorem 5. For the χ2-distance ϕ(t) = (t 1)2, the optimal value of the projection problem (5) can be computed exactly in time O(S log S). Theorem 5 splits the bivariate piecewise quadratic optimization problem (6) corresponding to the χ2-distance into S + 1 bivariate quadratic problems by sorting the components of b. Each of these S + 1 problems can be reduced to the solution of 3 univariate quadratic problems that themselves admit analytical solutions. S MOSEK fast MOSEK/fast 1,000 47.56 0.20 243.23 1,500 71.00 0.29 241.92 2,000 89.57 0.39 231.46 2,500 114.82 0.48 239.11 3,000 139.13 0.58 241.86 S = A MOSEK fast MOSEK/fast 100 3,383.35 22.32 151.56 150 15,353.40 51.67 297.17 200 46,049.30 83.86 549.10 250 104,709.00 130.34 803.35 300 215,871.00 176.35 1,224.05 Table 2: Comparison of our algorithms ( fast ) vs. MOSEK for the projection problem (left) and the Bellman update (right) on KL-divergence constrained ambiguity sets. Runtimes are reported in ms. S = A f-o (3 its) f-o (5 its) fast f-o/fast (3 its) f-o/fast (5 its) 100 175.80 489.66 22.32 7.88 21.94 150 397.84 1,103.53 51.67 7.70 21.36 200 704.93 1,955.05 83.86 8.41 23.31 250 1,051.25 2,921.71 130.34 8.07 22.42 300 1,542.03 4,283.27 176.36 8.74 24.29 Table 3: Comparison of our algorithms ( fast ) vs. the first-order method of [14] (after ℓ= 3, 5 its.) for the Bellman update on KL-divergence constrained ambiguity sets. Runtimes are reported in ms. Corollary 4. The robust Bellman iteration (3) over a χ2-distance ambiguity set can be computed to any accuracy ϵ > 0 in time O(S2 log S A log[R/ϵ]). The projection problem (5) for the χ2-distance ambiguity set can be solved as a quadratic program with a practical complexity of O(S3) as well as a theoretical complexity of O(S3.5). The first-order framework of [14] also applies to RMDPs over s-rectangular spherical uncertainty sets. In that case, the robust Bellman update enjoys a complexity of O(ℓ2 S2 A log2(ϵ 1)), where ℓis the iteration number. A careful analysis results in an overall convergence rate for the optimal MDP policy of O(S3 log S A2 ϵ 1 log[ϵ 1]). In contrast, the convergence rate of our robust value iteration amounts to O(S2 log S A log[R/ϵ] log[ϵ 1]). Treating the parameter R as a constant, our convergence rate simplifies to O(S2 log S A log2[ϵ 1]), which compares favourably against the convergence rate of [14]. We remark, however, that the spherical ambiguity sets of [14] differ from the χ2-distance ambiguity sets studied here, and as such the two methods are not directly comparable. We also note that our χ2-distance ambiguity sets enjoy a strong statistical justification [4, 6]. Computing unweighted 2-norm projections of points onto S-dimensional probability simplices has manifold applications in image processing, finance, optimization and machine learning [1, 8]. [30] proposes one of the earliest algorithms that computes this projection in time O(S2) by iteratively reducing the dimension of the problem using Lagrange multipliers. The minimum complexity of O(S) is achieved, among others, by [27] through a linear-time median-finding algorithm and by [36] through a filtered bucket-clustering method. Note, however, that these algorithms do not account for the weights and the additional inequality constraint present in our generalized projection (5). The unweighted 2-norm projection of a point onto the intersection of the S-dimensional probability simplex with an axis-parallel hypercube is computed by [52] through a sorting-based method and by [2] through Newton s method, respectively. [38] optimize a linear function over the intersection of a probability simplex with an unweighted 2-norm constraint through an iterative dimension reduction scheme. [1], finally, study algorithms that optimize linear functions over the intersection of a probability simplex and a bound on the unweighted 2-norm distance to a nominal distribution. 6 Numerical Results We compare our fast suite of algorithms with the state-of-the-art solver MOSEK 9.3 [3] (commercial) and the first-order method of [14]. All experiments are implemented in C++, and they are run on a 3.6 GHz 8-Core Intel Core i9 CPU with 32 GB 2667 MHz DDR4 main memory. The source code is available at https://sites.google.com/view/clint-chin-pang-ho. S MOSEK fast MOSEK/fast 1,000 49.94 0.05 981.91 1,500 76.61 0.08 945.99 2,000 93.13 0.11 854.39 2,500 123.16 0.14 879.72 3,000 153.06 0.17 917.18 S = A MOSEK fast MOSEK/fast 100 148.63 2.59 57.40 150 353.58 6.01 58.85 200 687.14 11.78 58.34 250 1,212.30 19.38 62.54 300 1,908.92 25.84 73.87 Table 4: Comparison of our algorithms ( fast ) vs. MOSEK for the projection problem (left) and the Bellman update (right) on χ2-distance constrained ambiguity sets. Runtimes are reported in ms. For our experiments, we synthetically generate random RMDP instances as follows. For the projection problem, we sample each component of b uniformly at random between 0 and 1. Similarly, we sample each component of psa uniformly at random between 0 and 1 and subsequently scale psa so that its elements sum up to 1. The parameter β, finally, is uniformly distributed between min{b} + 10 8 and p sab 10 8 to adhere to the assumptions of our paper. For the robust Bellman update, all vectors bsa and all transition probabilities psa, s S and a A, are generated according to the above procedure. The parameter κ is also sampled from a uniform distribution supported on [0, 1]. Tables 2 4 report average computation times over 50 randomly generated test instances for the KL-divergence and the χ2-distance based ambiguity sets and show that the proposed algorithms outperform other methods. The tables reveal that our algorithms are about two orders of magnitude faster than MOSEK in solving the projection problem (5). For computing the robust Bellman update J, our algorithms are also orders of magnitude faster than MOSEK, and this ratio increases with the problem size. The results also show that our algorithm outperforms the first-order method of [14]. The difference becomes more significant with an increasing number of robust Bellman updates. While our algorithms outperform the first-order scheme by a factor of 8 in the third Bellman iteration, they outperform it by a factor of about 20 in the fifth Bellman iteration. Since first-order methods are known to require many iterations for convergence, we conclude that our algorithm compares favorably in this experiment as well. 7 Conclusion We study s-rectangular robust MDPs with ϕ-divergence ambiguity sets. We develop efficient algorithms for computing robust Bellman updates for several important special cases of this ambiguity set. Our experimental results indicate that the proposed algorithms outperform MOSEK. Future work should address extensions to scalable model-free algorithms. Acknowledgments This work was supported, in part, by the Engineering and Physical Sciences Research Council (EPSRC) grant EP/W003317/1, by the City U Start-Up Grant (Project No. 9610481), by the National Natural Science Foundation of China (Project No. 72032005), by the Chow Sang Sang Group Research Fund sponsored by Chow Sang Sang Holdings International Limited (Project No. 9229076), and by the NSF grant No. 1815275. Any opinion, finding, conclusion, or recommendation expressed in this material are those of the authors and do not necessarily reflect the views of the Engineering and Physical Sciences Research Council and the National Natural Science Foundation of China. [1] L. Adam and V. Mácha. Projections onto the canonical simplex with additional linear inequalities. Optimization Methods & Software, Available online first, 2020. [2] M. S. Ang, J. Ma, N. Liu, K. Huang, and Y. Wang. Fast projection onto the capped simplex with applications to sparse regression in bioinformatics. In Advances in Neural Information Processing Systems, volume 34, 2021. [3] MOSEK Ap S. MOSEK Fusion API for C++ 9.3.20, 2019. URL https://docs.mosek.com/ latest/cxxfusion/index.html. [4] G. Bayraksan and D. K. Love. Data-driven stochastic programming using phi-divergences. In D. M. Aleman and A. C. Thiele, editors, INFORMS Tut ORials in Operations Research, pages 1 19. 2015. [5] B. Behzadian, M. Petrik, and C. P. Ho. Fast algorithms for l -constrained s-rectangular robust MDPs. In Advances in Neural Information Processing Systems, volume 34, 2021. [6] A. Ben-Tal, D. den Hertog, A. de Waegenaere, B. Melenberg, and G. Rennen. Robust solutions of optimization problems affected by uncertain probabilities. Management Science, 59(2): 341 357, 2013. [7] Z. Chen, P. Yu, and W. B. Haskell. Distributionally robust optimization for sequential decisionmaking. Optimization, 68(12):2397 2426, 2019. [8] L. Condat. Fast projection onto the simplex and the l1 ball. Mathematical Programming, 158 (1 2):575 585, 2016. [9] E. Derman, M. Geist, and S. Mannor. Twice regularized MDPs and the equivalence between robustness and regularization. In Advances in Neural Information Processing Systems, volume 35, pages (Pre Proceedings), 2021. [10] M. Diehl and J. Bjornberg. Robust dynamic programming for min-max model predictive control of constrained uncertain systems. IEEE Transactions on Automatic Control, 49(12):2253 2257, 2004. [11] R. Givan, S. Leach, and T. Dean. Bounded-parameter Markov decision processes. Artificial Intelligence, 122(1):71 109, 2000. [12] J. Goh, M. Bayati, S. A. Zenios, S. Singh, and D. Moore. Data uncertainty in Markov chains: Application to cost-effectiveness analyses of medical innovations. Operations Research, 66(3): 697 715, 2018. [13] V. Goyal and J. Grand-Clément. Robust Markov decision process: Beyond rectangularity. Available on ar Xiv, 2018. [14] J. Grand-Clément and C. Kroer. Scalable first-order methods for robust MDPs. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 12086 12094, 2021. [15] J. Grand-Clément and C. Kroer. First-order methods for Wasserstein distributionally robust MDPs. In Proceedings of Machine Learning Research, volume 139, pages 2010 2019, 2021. [16] J. Grand-Clément, C. W. Chan, V. Goyal, and G. Escobar. Robust policies for proactive ICU transfers. Working Paper, 2019. [17] Vishal Gupta. Near-optimal Bayesian ambiguity sets for distributionally robust optimization. Management Science, 65(9):4242 4260, 2019. [18] C. P. Ho, M. Petrik, and W. Wiesemann. Fast Bellman updates for robust MDPs. In Proceedings of the 35th International Conference on Machine Learning, pages 979 1988, 2018. [19] C. P. Ho, M. Petrik, and W. Wiesemann. Partial policy iteration for l1-robust Markov decision processes. Journal of Machine Learning Research, 22:1 46, 2021. [20] Q. Huang, Q.-S. Jia, and X. Guan. Robust scheduling of EV charging load with uncertain wind power integration. IEEE Transactions on Smart Grid, 9(2):1043 1054, 2018. [21] G. N. Iyengar. Robust dynamic programming. Mathematics of Operations Research, 30(2): 257 280, 2005. [22] D. L. Kaufman and A. J. Schaefer. Robust modified policy iteration. INFORMS Journal on Computing, 25(3):396 410, 2013. [23] M. J. Kochenderfer and J. P. Chryssanthacopoulos. Robust airborne collision avoidance through dynamic programming. Project Report ATC-371 for the Federal Aviation Administration, 2011. [24] M.A.S. Kolarijani, G.F. Max, and P. Mohajerin Esfahani. Fast approximate dynamic programming for infinite-horizon markov decision processes. In Advances in Neural Information Processing Systems, volume 34, pages 23652 23663, 2021. [25] Y. Le Tallec. Robust, Risk-Sensitive, and Data-driven Control of Markov Decision Processes. Ph D thesis, Massachusetts Institute of Technology, 2007. [26] D. Love and G. Bayraksan. Phi-divergence constrained ambiguous stochastic programs for data-driven optimization. Technical report, 2015. [27] N. Maculan and G. G. de Paula Jr. A linear-time median-finding algorithm for projecting a vector on the simplex of Rn. Operations Research Letters, 8(4):219 222, 1989. [28] S. Mannor, O. Mebel, and H. Xu. Robust MDPs with k-rectangular uncertainty. Mathematics of Operations Research, 41(4):1484 1509, 2016. [29] A. Al Marjani, A. Garivier, and A. Proutiere. Navigating to the best policy in markov decision processes. In Advances in Neural Information Processing Systems, volume 34, pages 25852 25864, 2021. [30] C. Michelot. A finite algorithm for finding the projection of a point onto the canonical simplex of Rn. Journal of Optimization Theory and Applications, 50(1):195 200, 1986. [31] A. Nie, E. Brunskill, and C. Piech. Play to grade: Testing coding games as classifying markov decision process. In Advances in Neural Information Processing Systems, volume 34, pages 1506 1518, 2021. [32] A. Nilim and L. El Ghaoui. Robust control of Markov decision processes with uncertain transition matrices. Operations Research, 53(5):780 798, 2005. [33] K. Panaganti and D. Kalathil. Robust reinforcement learning using least squares policy iteration with provable performance guarantees. In Proceedings of the 38th International Conference on Machine Learning, pages 511 520, 2021. [34] K. Panaganti and D. Kalathil. Sample complexity of robust reinforcement learning with a generative model. In International Conference on Artificial Intelligence and Statistics, pages 9582 9602. PMLR, 2022. [35] B. P.G. Van Parys, P. Mohajerin Esfahani, and D. Kuhn. From data to decisions: Distributionally robust optimization is optimal. Management Science, 67(6):3387 3402, 2021. [36] G. Perez, M. Barlaud, L. Fillatre, and J.-C. Régin. A filtered bucket-clustering method for projection onto the simplex and the l1 ball. Mathematical Programming, 182(1 2):445 464, 2020. [37] M. Petrik and D. Subramanian. RAAM: The benefits of robustness in approximating aggregated MDPs in reinforcement learning. In Advances in Neural Information Processing Systems, volume 27, pages 1979 1987, 2014. [38] A. Philpott, V. de Matos, and L. Kapelevich. Distributionally robust SDDP. Computational Management Science, 15(3 4):431 454, 2018. [39] M. L. Puterman. Markov Decision Processes: Discrete Stochastic Dynamic Programming. John Wiley & Sons, 1994. [40] H. Rahimian, G. Bayraksan, and T. Homem-de-Mello. Identifying effective scenarios in distributionally robust stochastic programs with total variation distance. Mathematical Programming, 173(1 2):393 420, 2019. [41] R. T. Rockafellar. Convex Analysis. Princeton University Press, 1997. [42] A. Roy, H. Xu, and S. Pokutta. Reinforcement learning under model mismatch. In Advances in Neural Information Processing Systems, volume 30, 2017. [43] P. Rusmevichientong and H. Topaloglu. Robust assortment optimization under the multinomial logit choice model. Operations Research, 60(4):865 882, 2012. [44] A. Ruszczy nski. Risk-averse dynamic programming for Markov decision processes. Mathematical Programming, 125(2):235 261, 2010. [45] J. K. Satia and R. E. Lave Jr. Markovian decision processes with uncertain transition probabilities. Operations Research, 21(3):728 740, 1973. [46] A. Shapiro. Rectangular sets of probability measures. Operations Research, 64(2):528 541, 2016. [47] A. Shapiro. Distributionally robust optimal control and MDP modeling. Operations Research Letters, 49(3):809 814, 2021. [48] R. S. Sutton and A. G. Barto. Reinforcement Learning: An Introduction. The MIT Press, second edition, 2018. [49] A. Tamar, Y. Glassner, and S. Mannor. Policy gradients beyond expectations: Conditional value-at-risk. Available on ar Xiv, 2014. [50] A. Tamar, S. Mannor, and H. Xu. Scaling up robust MDPs using function approximation. In Proceedings of the 31st International Conference of Machine Learning, 2014. [51] A. Tirinzoni, X. Chen, M. Petrik, and B. D. Ziebart. Policy-conditioned uncertainty sets for robust Markov decision processes. In Advances in Neural Information Processing Systems, volume 31, pages 8953 8963, 2018. [52] W. Wang and C. Lu. Projection onto the capped simplex. Available on ar Xiv, 2015. [53] Y. Wang and S. Zou. Online robust reinforcement learning with model uncertainty. In Advances in Neural Information Processing Systems, volume 34, pages 7193 7206, 2021. [54] Yue Wang and Shaofeng Zou. Policy gradient method for robust reinforcement learning. In Proceedings of the 39th International Conference on Machine Learning, pages 23484 23526, 2022. [55] T. Weissman, E. Ordentlich, G. Seroussi, S. Verdu, and M. J. Weinberger. Inequalities for the l1 deviation of the empirical distribution. Technical Report, https://www.hpl.hp.com/research/info theory/papers/HPL-2003-97R1Web.pdf, 2003. [56] W. Wiesemann, D. Kuhn, and B. Rustem. Robust Markov decision processes. Mathematics of Operations Research, 38(1):153 183, 2013. [57] H. Xiao, K. Yang, and X. Wang. Robust power control under channel uncertainty for cognitive radios with sensing delays. IEEE Transactions on Wireless Communications, 12(2):646 655, 2013. [58] L. Xin and D. A. Goldberg. Distributionally robust inventory control when demand is a martingale. Available on ar Xiv, 2018. [59] H. Xu and S. Mannor. Distributionally robust Markov decision processes. In Advances in Neural Information Processing Systems, volume 23, pages 2505 2513, 2010. [60] H. Xu and S. Mannor. Distributionally robust Markov decision processes. Mathematics of Operations Research, 37(2):288 300, 2012. [61] P. Yu and H. Xu. Distributionally robust counterpart in Markov decision processes. IEEE Transactions on Automatic Control, 61(9):2538 2543, 2016. [62] Y. Zhang, L. N. Steimle, and B. T. Denton. Robust Markov decision processes for medical treatment decisions. Available on Optimization Online, 2017. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope?[Yes] The abstract well explains the main contributions of the paper. (b) Did you describe the limitations of your work? [Yes] We took great care to ensure that our numerical results provide an objective and unbiased assessment of our solution approach. (c) Did you discuss any potential negative societal impacts of your work? [N/A] There are no new negative societal impacts that we can identify. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] We have carefully read those guidelines, and to our best understanding our paper fully complies with them. 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] All of our results state the full set of assumptions, with exception of the blanket assumptions that are assumed to hold throughout the paper and that are clearly marked as such. (b) Did you include complete proofs of all theoretical results? [Yes] All proofs are contained in the appendix. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] All code and instructions for our experimental results are published online. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] All details are mentioned either in the numerical results section or in the appendix. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] Included in the appendix. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] Included in the numerical results section. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] We use C++ and MOSEK, all of which are cited in the text. (b) Did you mention the license of the assets? [Yes] All licenses are mentioned. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] All code and instructions for our experimental results are published online. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] We use synthetic data in our experiments. (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] We use synthetic data in our experiments. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] We use synthetic data in our experiments. (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] We use synthetic data in our experiments. (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A] We use synthetic data in our experiments.