# rhopomdps_have_lipschitzcontinuous_epsilonoptimal_value_functions__d39c4963.pdf ρ-POMDPs have Lipschitz-Continuous ϵ-Optimal Value Functions Mathieu Fehr1, Olivier Buffet2, Vincent Thomas2, Jilles Dibangoye3 1 École Normale Supérieure de la rue d Ulm, Paris, France 2 Université de Lorraine, CNRS, Inria, LORIA, Nancy, France 3 Université de Lyon, INSA Lyon, Inria, CITI, Lyon, France mathieu.fehr@ens.fr, olivier.buffet@loria.fr, vincent.thomas@loria.fr, jilles.dibangoye@inria.fr Many state-of-the-art algorithms for solving Partially Observable Markov Decision Processes (POMDPs) rely on turning the problem into a fully observable problem a belief MDP and exploiting the piece-wise linearity and convexity (PWLC) of the optimal value function in this new state space (the belief simplex ). This approach has been extended to solving ρ-POMDPs i.e., for informationoriented criteria when the reward ρ is convex in . General ρ-POMDPs can also be turned into fully observable problems, but with no means to exploit the PWLC property. In this paper, we focus on POMDPs and ρ-POMDPs with λρ-Lipschitz reward function, and demonstrate that, for finite horizons, the optimal value function is Lipschitz-continuous. Then, value function approximators are proposed for both upperand lower-bounding the optimal value function, which are shown to provide uniformly improvable bounds. This allows proposing two algorithms derived from HSVI which are empirically evaluated on various benchmark problems. 1 Introduction Many state-of-the-art algorithms for solving Partially Observable Markov Decision Processes (POMDPs) rely on turning the problem into a fully observable problem namely a belief MDP and exploiting the piece-wise linearity and convexity (PWLC) of the optimal value function [Sondik, 1971, Smallwood and Sondik, 1973] in this new problem s state space (here the belief space ). State of the art off-line algorithms [Pineau et al., 2006, Smith and Simmons, 2004] maintain approximators that (i) are upper or lower bounds, and (ii) have generalization capabilities: a local update at b improves the bound in a surrounding region of b. This approach has been extended to solving ρ-POMDPs as belief MDPs i.e., problems whose performance criterion depends on the belief (e.g., active information gathering) when the reward ρ is convex in [Araya-López et al., 2010].1 Yet, it does not extend to problems with non-convex ρ e.g., (i) if a museum monitoring system is rewarded for each visitor located with enough certainty (i.e., using a threshold function), or (ii) if collecting data regarding patients while preserving their privacy by discarding information that could harm anonymity. Generalizing value function approximators are also an important topic in (fully observable, monoagent) reinforcement learning, as recently with Deep RL [Mnih et al., 2013]. To allow for errorbounded approximations in continuous settings, some works have built on the hypothesis that the dynamics and the reward function were Lipschitz-continuous (LC), which leads to Lipschitzcontinuous value functions [Laraki and Sudderth, 2004, Hinderer, 2005, Fonteneau et al., 2009, 1And also to solving Decentralized POMDPs (Dec-POMDPs) as occupancy MDPs (o MDPs) i.e., when designing multiple collaborating controllers [Dibangoye et al., 2013, 2016]. 32nd Conference on Neural Information Processing Systems (Neur IPS 2018), Montréal, Canada. Rachelson and Lagoudakis, 2010, Dufour and Prieto-Rumeau, 2012]. Ieong et al. [2007] also considered exploiting the LC property in heuristic search settings. These approaches cannot be applied to the aforementioned partially observable problems as the dynamics of the induced MDPs are a priori not LC. This paper shows that, for ρ-POMDPs with λρ-Lipschitz reward function (and thus for any POMDP) and for finite horizons, the optimal value function is still LC, a property that shall replace the PWLC property. Yet, to allow for better approximators and tighter theoretical bounds, we use an extended definition of Lipschitz-continuity where (i) the Lipschitz constant is a vector rather than a scalar, and (ii) we consider local rather than uniform LC. From there, value function approximators are proposed for both upperand lower-bounding the optimal value function. Following Smith [2007], these approximators are shown to provide uniformly improvable bounds [Zhang and Zhang, 2001] for use with point-based algorithms like HSVI, which is then guaranteed to converge to an ϵ-optimal solution. This allows proposing two algorithms derived from HSVI: (i) one that uses guaranteed/safe Lipschitz constants, but at the cost of overly pessimistic error bounds, and (ii) one that searches for good Lipschitz constants, but then losing optimality guarantees. This work is also a step towards solving partially observable stochastic games as continuous-space SGs with LC approximators. The paper is organized as follows. Section 2 discusses related work on inforation-oriented control. Sec. 3 presents background on POMDPs, ρ-POMDPs and Lipschitz continuity. Sec. 4 demonstrates that, for finite horizons, the optimal value function is Lipschitz-continuous, Sec. 5 proposes value function approximators and two point-based algorithms (based on HSVI). Sec. 6 evaluates them empirically. Proofs are provided as supplementary material. 2 Related Work Early research on information-oriented control (IOC) involved problems formalized either (i) as POMDPs (as Egorov et al. [2016] did recently, since an observation-dependent reward can be trivially recast as a state-dependent reward), or (ii) with belief-dependent rewards (and mostly ad-hoc solution techniques). ρ-POMDPs allow easily formalizing many if not most IOC problems. Araya-López et al. [2010] show that a ρ-POMDP with convex belief-dependent reward ρ can be solved with modified point-based POMDP solvers exploiting the PWLC property (with error bounds that depend on the quality of the PWLC-approximation of ρ). The POMDP-IR framework [Spaan et al., 2015] allows describing IOC problems with linear rewards thus, a subclass of PWLC ρ-POMDPs (i.e., when ρ is PWLC). Yet, as Satsangi et al. [2015] showed that a PWLC ρ-POMDP can be turned into a POMDP-IR, both classes are in fact equivalent. In both cases the proposed solution techniques are modified POMDP solvers, and it seems (to us) that an algorithm proposed in one framework should apply with limited changes in the other framework. For its part, the general ρ-POMDP framework allows formalizing more problems e.g., directly specifying an entropy-based criterion. While Spaan et al. [2015] obtain better empirical results with their POMDP-IR-based method than with a ρ-POMDP-based method, this probably says more about particular solutions applied on a particular problem than about the frameworks themselves (as discussed above). The case of non-convex ρ (including information-averse scenarios) may have been mostly avoided up to now because no satisfying solution technique existed. The present work analyzes the optimal value function s properties when ρ is Lipschitz-continuous, which leads to a prototype solution algorithm. This is a first step towards proposing new tools for solving a wider class of information-oriented POMDPs than currently feasible. Future work will thus be more oriented towards practical applications, possibly with evaluations on surveillance problems which are only motivating scenarios in the present paper. Note that Egorov et al. [2016] propose solutions dedicated to surveillance (with an adversarial setting), not for general IOC problems. Regarding adversarial settings, another very promising direction is exploiting the Lipschitz continuity in a similar manner to solve (zero-sum) Partially Observable Stochastic Games. 3 Background Notations: We denote: ˆx = x/ x 1 the normalization of a vector x; |x| a component-bycomponent (Cb C) absolute value operator; maxxf(x) a Cb C maximum operator for vector-valued function f(x); and 1 a row vector of 1s. A POMDP [Astrom, 1965] is defined by a tuple S, A, Z, P, r, γ, b0 , where S, A and Z are finite sets of states, actions and observations; Pa,z(s, s ) gives the probability of transiting to state s and observing observation z when applying action a in state s (Pa,z is an S S matrix); r(s, a) R is the reward associated to performing action a in state s; γ [0; 1) is a discount factor; and b0 is the initial belief state i.e., the initial probability distribution over possible states. The objective is then to find a policy π that prescribes actions depending on past actions and observations so as to maximize the expected discounted sum of rewards (here with an infinite temporal horizon). To that end, a POMDP is often turned into a belief MDP , A, T, r, γ, b0 where is the simplex of possible belief states, A is the same action set, and T(b, a, b ) = P(b |b, a) and r(b, a) = P s b(s)r(s, a) are the induced transition and reward functions. This setting allows considering policies π: A, each being associated to its value function V π(b) .= E[P t=0 γtr(bt, π(bt)|b0 = b]. Optimal policies maximize V π in all belief states reachable from b0. Their value function V is the fixed point of Bellman s optimality operator (H) [Bellman, 1957] HV : b 7 maxa[r(b, a) + γ P z Pa,zb 1V (ba,z)], and acting greedily with respect to V provides such a policy. V being piece-wise linear and convex (PWLC)2 for any finite horizon [Sondik, 1971, Smallwood and Sondik, 1973] allows to approximate it from below by an upper-envelope U of hyperplanes, and from above by a lower-envelope L of points. A local update at belief state b then allows improving U or L not only at b but in its vicinity. This generalization allows for error-bounded approximations using a finite number of belief points [Pineau et al., 2006], and for more efficient branch pruning in heuristic search approaches [Smith, 2007]. All this led to current off-line point-based algorithms such as PBVI [Pineau et al., 2003, 2006], HSVI [Smith and Simmons, 2004, 2005, Smith, 2007], SARSOP [Kurniawati et al., 2008], Gap Min [Poupart et al., 2011], and PGVI [Zhang et al., 2014]. We shall consider in particular HSVI (Heuristic Search Value Iteration, see Algorithm 1) as it is a prototypical algorithm maintaining both U and L, and providing performance guarantees by stopping when U(b0) L(b0) is below an ϵ threshold. HSVI decides on where to perform updates by generating trajectories picking (i) actions greedily w.r.t. to U and (ii) observations so as to reduce the gap between U and L. Importantly, the HSVI framework is based on uniformly improvable bounds (cf. Sec. 5.1) and applicable beyond POMDPs with PWLC approximations. 3.2 ρ-POMDPs ρ-POMDPs [Araya-López et al., 2010] differ from POMDPs in their reward function ρ(b, a) rather than r(s, a) that allows defining not only control-oriented criteria, but also information-oriented ones, thus generalizing POMDPs. Such problems are met regularly, but often modeled and addressed with ad-hoc techniques [Fox et al., 1998, Mihaylova et al., 2006]. Araya-López et al. [2010] have shown that, (i) if ρ is PWLC, previously described techniques can still be applied with similar error bounds, and (ii) if ρ is convex and either Lipschitz-continuous or α-Hölder (as Shannon s entropy), then a PWLC approximation of ρ can be used to obtain error-bounded solutions again. While many problems can be modeled with convex ρ, this leaves us with a number of problems that cannot be solved with similar approximations. Here, we will exploit Lipschitz-continuous reward functions ρ to solve more general ρ-POMDPs with similar algorithmic schemes. As an example, in the museum monitoring scenario, with X the random variable for a visitor s location and b X the corresponding belief, then ρX(b, a) .= σ(α( b X β)) with σ( ) the sigmoid function is a smooth threshold function (thus non-convex) whose Lipschitz constant depends on α > 0 and rewarding preferably distributions whose maximum probability is greater than β [0, 1]. 2It is thus also Lipschitz-continuous. Algorithm 1: Heuristic Search Value Iteration & Inc-lc-HSVI 1 Fct HSVI (ϵ) 2 Initialize L and U 3 while (U(b0) L(b0)) > ϵ do 4 Recursively Try (b0, d = 0) 6 Fct Recursively Try (b, d) 7 if (U(b) L(b)) > γ dϵ then 8 Update (b) 9 a arg maxa A{r(b, a) + γ P z Pa,zb 1U(ba,z)} 10 z arg maxz Z{ Pa ,zb 1 (U(ba ,z) L(ba ,z) γ dϵ)} 11 Recursively Try (ba ,z , d + 1) 12 Update (b) 14 Fct Update (b) 15 L Update (L, b) 16 U Update (U, b) /* Note: Vanilla HSVI for POMDPs uses PWLC approximators. lc-HSVI is HSVI with LC approximators. */ /* Below: Main loop of Incremental lc-HSVI (see Sec. 5.2). */ 17 Fct inc-lc-HSVI (ϵ, λ0) 18 λ λ0 19 while fails(lc-HSVI (ϵ, λ)) do 21 return L 3.3 Lipschitz-Continuities (in normed spaces) Let f : X Y be a function, where X and Y are normed spaces. f is uniformly Lipschitzcontinuous if there exists λf R+ such that, for all (x, x ) X2, f(x) f(x ) λf x x . f is locally Lipschitz-continuous if, for each x, there exists λf(x) R+ such that, for all x X, f(x) f(x ) λf(x) x x . The former definition is more common and induces uniform continuity of f, but we will rely on the later (omitting locally ), which induces local continuity of f, to handle more problems and obtain tighter bounds. We propose another generalization using vector rather than scalar Lipschitz constants, again to allow for tighter bounds: f is Lipschitzcontinuous if, for each x, there exists a row vector λf(x) (R+)dim(X) such that, for all x X, f(x) f(x ) λf(x) |x x | (scalar product equivalent to a weighted L1-norm). Note that Lipschitz-continuity is here always relative to the simplex , not R|S|. and A being both compact, properties that hold in the local and vector setting also hold in the uniform and/or scalar setting (but bounds are looser). 4 Lipschitz-continuity of V Assuming a ρ-POMDP with local and vector Lipschitz-continuous reward function with constant λρ(b, a) in (b, a), this section first states that Bellman s optimality operator preserves the LC property, which then allows proving that V is LC for finite horizons, but not necessarily for infinite ones. Proposition 1 (H preserves Lipschitz-Continuity). Given a ρ-POMDP with λρ( , )-LC reward function, and a λV ( )-LC value function V , then HV is (at least) λHV ( )-LC with, in each belief b, λHV (b) = maxa h λρ(b, a) + γ X z [(|V (ba,z)| + λV (ba,z)ba,z) 1 + λV (ba,z)] Pa,z i . (1) Proof (sketch). A key point here is to show that κ(w) .= w 1V ( ˆw) is LC (see supplementary material), which relies on the triangle (in)equality. Then the rest consists essentially in some algebra using this property and other LC properties. As can be observed in Eq. (1), the resulting update formula of the value function s Lipschitz constant exploits both the locality (dependence on the belief b) and the use of a vector rather than a scalar. The dependence of the update formula on |V (ba,z)| may seem surprising since adding constant kr R to ρ(b, a) should induce adding a related constant k V to V without changing local Lipschitz constants. This dependence is due to approximations made in the proof. |V (ba,z)| can in fact be replaced by |V (ba,z) + k V | in Equation 1 with a tunable k V . This induces the multi-objective task of minimizing the components of the (vector) Lipschitz constant through k V , a problem we address by setting k V = maxz V (ba,z)+minz V (ba,z) Theorem 1 (Local Lipschitz-continuity of V for finite T). Given a ρ-POMDP with λρ(b, a)-LC ρ, for any finite time horizon T, the optimal value function is (locally+vector) LC. Proof. The value function for T = 0 is trivially 0-LC. By induction, as Bellman s optimality operator preserves the LC property (Prop. 1), the optimal value function is LC for any finite T. Asymptotic Behavior Previous results do not tell whether the resulting Lipschitz constant tends to a limit value when T goes to infinity. This issue is considered here for a scalar and uniform constant λ. λρ(b, a) and λt(b) (the Lipschitz constant for Vt, not indicating the horizon T) respectively become λρ and λt. Corollary 1. For a ρ-POMDP with (uniform) λρ-LC ρ, the optimal value function verifies, for all t and all b1, b2, |V t (b1) V t (b2)| λt b1 b2 1, where λt = λρ + γV lim 1 (2γ)T t 1 (2γ) if γ = 1 2, λρ + γV lim (T t) if γ = 1 with V lim .= 1 1 γ maxb,a|r(b, a)|. In the common case γ 1 2, λ0 diverges when T . Hence, V may not be LC in the infinite horizon setting. Yet, it will suffice to compute finite-horizon LC approximations of V (as usual with PWLC approximations), as explained in the next section. 5 Approximating V This section shows how to define, initialize, update and prune LC upperand lower-bounding approximators of V , and then derive an ϵ-optimal variant of the HSVI algorithm. 5.1 Upperand Lower-Bounding V The upper-bounding LC approximator is defined as a finite set of downward-pointing L1-cones (see Figure 1 (left)), where an upper-bounding cone c U β = β, u, λ located at belief β, with summit value u and slope vector λ induces a function Uβ(b) = u + λ |β b|. The upper bound is thus defined as the lower envelope of a set of cones CU = {c U β }β BU i.e., U(b) = minβ BU Uβ(b). Respectively, for the lower-bounding approximator: a lower-bounding (upward-pointing) cone c L β = β, l, λ induces a function Lβ(b) = l λ |β b|; and the lower bound is defined as the upper envelope of a set of cones CL = {c L β}β BL i.e., L(b) = maxβ BL Lβ(b). We now (i) show how the (pointwise) update of the upperor lower-bound preserves this representation; (ii) verify that the properties required for HSVI to converge to an ϵ-optimal solution still hold; and (iii) discuss their initialization. Updating (Upper and Lower) Bounds The following proposition and its counterpart state that, for both U and L, a pointwise update results in adding a new cone with its own Lipschitz constant. Theorem 2 (Updating U). Let us assume that (i) ρ is λρ(b, a)-LC for each (b, a), and (ii) the upper bound U is described by a set of upper cones CU. Then, for any b, an improved upper bound is obtained by adding a cone in b, with value and Lipschitz constant: u(b) = [HU](b) = max a ρ(b, a) + γ X z Pa,zb 1U(ba,z) λ(b) = maxa λρ(b, a ) + γ X h λ(βa ,z) + |u(βa ,z)| + λ(βa ,z)βa ,z 1 i Pa ,z where βa,z is the current point in B that best approximates U in ba,z (see Fig. 1 (left)). Figure 1: (left) An optimal V surrounded by its upper and lower bounds (2 cones for U in red, and 3 cones for L in blue). The value of U at ba,z is approximated by the cone located at βa,z. (right) Grid environment of the grid-info problem (Sec. 6.1). u3, λ3 u2, λ3 Figure 2: Bellman update of U (left) and L (right) at b resulting in 1 (dashed) cone per action. (right) The upper envelope of 3 cones is approximated by (solid) cone b, u2, λ3 . (left) The upper envelope of 3 cones is approximated by (solid) cone b, l2, λ2 . The operator performing this update at b is noted KU b and the updated upper bound is thus KU b U. Intuitively (see Fig. 2 (left)), each action induces one cone, which may be preferred depending on the point b where the evaluation is done. Yet, rather than adding the upper enveloppe of this set of cones, a single upper-bounding cone is employed using the maximum u(b) and λ(b). A pointwise update of U in b consists in computing (i) u(b) by Bellman update, in Θ(|A| |Z| (|S|2 + |BU| |S|)) (memoizing points βa,z that optimize U(ba,z)) and (ii) λ(b) by a Bellman-like update, in Θ(|A| |Z| |S|2). The latter computation searches for the worst Lipschitz constant component by component, thus harming the generalization capabilities. Better Lipschitz constants could be obtained by first pruning cones that are dominated by other cones. For the lower bound L, an update requires adding at least one of the cones with maximum l, but we even add each cone induced by each action. The complexity is unchanged (replacing BU by BL). Pruning Cones The LC setting requires a procedure for pruning cones. Without loss of generality, we consider U. A cone c U β must be maintained if there exists b such that Uβ(b) < minβ BU\{β} Uβ (b), which is equivalent to finding a strictly negative value of φU β (b) .= Uβ(b) minβ BU\{β} Uβ (b). This could be done by applying a minimization procedure on φU β until a negative value is found. A more pragmatic approach is to search, for cone c U β , if another cone c U β exists that completely dominates it i.e., (i) u(β) u(β ) + λ(β )|β β |, and (ii) λ(β) λ(β ). This can be improved by comparing not the Lipschitz constants, but the value at the corners of the simplex, to make sure that dominance is checked inside . The resulting process is conservative and cheap, but not complete. Preservation of HSVI s Convergence Properties Ensuring finite time convergence of HSVI (even beyond POMDPs) to an ϵ-optimal solution requires using (a) a uniformly improvable (UI) lower bound L, i.e., HL L, where H is Bellman optimality operator; (b) respectively a uniformly improvable (UI) upper bound U; (c) a strong pointwise update operator for the lower bound L, KL , i.e., for each b where it is applied and any L, (i) (HL)(b) = (KL b L)(b) and (ii) (HL)(b ) (KL b L)(b ) in any other point b ; and (d) resp. a strong pointwise update operator for U, KU . Trivially, our proposed operators are strong, thus conservative [Smith, 2007, Def. 3.24, 3.25]. Also, any conservative update operator preserves UI [Smith, 2007, Th. 3.29]. We thus essentially need to ensure that initializations induce UI bounds (see next sub-section). Initialization For a usual POMDP (ρ = r), initializations described by Smith [2007] are UI by construction. For a ρ-POMDP, similar constructions seem difficult to obtain. Another option is to go back to a POMDP with reward (linear in b) ru upper-bounding (resp. rl lower-bounding) ρ. We can then (i) employ the associated POMDP initialization, or (ii) solve the resulting POMDPs. In each case, the resulting bounds can be used as UI LC (with infinitely many cones). Going further, ρ could even be better upper- (resp. lower-) bounded by a lower (resp. upper) envelope of linear reward functions, which would lead to better initializations of U (resp. L) by taking lower (resp. upper) envelopes of independent bounds. 5.2 Algorithms We will distinguish HSVI variants depending on the approximators at hand: pwlc-HSVI, lc-HSVI and pw-HSVI respectively depend on the classical PWLC approximators, the LC approximators previously described, and non-generalizing pointwise (PW) approximators (equivalent to cones with an infinite Lipschitz constant). In each case, HSVI s convergence guarantees hold. Incremental Variant lc-HSVI computes (using Eq.(1)) upper bounds on the true (local and vector) Lipschitz constants i.e., the smallest constants for which the Lipschitz property holds. Yet, these upper bounds are often very pessimistic, which leads to (i) a poor generalization capability of U and L, and, (ii) as a consequence, a very slow convergence. To circumvent the resulting pessimistic bounds and to assess how much is lost due to this pessimism, we also propose another algorithm that incrementally searches for a valid (global and scalar) Lipschitz constant λ. The intuition is that, despite the search process, the resulting planning process could be more efficient due to (i) quickly detecting when a constant is invalid, and (ii) quickly converging to a solution when a valid constant is found. One issue is that the algorithm may terminate with an invalid solution. We first need to define lc-HSVI(λ), a variant of lc-HSVI where the Lipschitz constant is uniformly constrained to (scalar) value λ. As a consequence, (i) adding a new cone at β only requires computing u(β) or l(β), and (ii) the pragmatic pruning process is complete. If λ is not large enough, the algorithm may fail due to (LXU) L and U crossing each other at an update point b, (NUI) L or U being not uniformly improvable i.e., an update leads to a worse value than expected or, (UR) unstable results i.e., two consecutive runs with values λt and λt+1 verify |Lt(b0) Lt+1(b0)| > ϵ. Then, lc-HSVI(λ) is incorporated in an incremental algorithm, inc-lc-HSVI (see Alg. 1, fct. inc-lc-HSVI), that starts with some initial λ and runs lc-HSVI(λ) with geometrically increasing values of λ until lc-HSVI(λ) returns with no (LXU/NUI/UR) failure. As already mentioned, this process does not guarantee that a large enough λ has been found for L and U to be proper bounds. In practice, we use λ0 = 1, but problem-dependent values should be considered to avoid being sensitive to affine transforms of the reward function. 6 Experiments 6.1 Benchmark Problems To evaluate the various algorithms at hand, we consider both POMDP and ρ-POMDP benchmark problems. The former problems a diverse set taken from Cassandra s POMDP page3 allow comparing the proposed algorithms against the standard pwlc-HSVI. The ρ-POMDP problems are all based on a grid environment as described below. grid-info ρ-POMDP We consider an agent moving on a 3 3 toric grid with white and black cells (see Fig. 1 (right)). Each cell is indicated by its coordinates (x, y) ( {1, 2, 3}). The agent is initially placed uniformly at random. Moves (n,s,e,w) succeed with probability .8, otherwise the agent stays still. The current cell s color is observed with probability 1. γ is set to 0.95. Let bx (resp. by) be the belief over the x (resp. y) coordinate. Then, ρ(b) = + bx 1 31 1 (resp. bx 1 31 1) rewards knowing x (kx) (resp. not knowing x ( kx)). And replacing bx by by allows rewarding knowing y (ky) and not knowing y ( ky). 6.2 Experiments We run x-HSVI (x {pwlc, pw, lc, inc-lc}) on all benchmark problems with the exception of pwlc-HSVI not being run on ρ-POMDPs setting ϵ = 0.1 and a timeout of 600s. In inc-lc-HSVI, λ is initially set to 1. L and U are initialized (i) for POMDPs, using HSVI1 s blind estimate and MDP estimate, and (ii) for ρ-POMDPs, using Rmin 1 γ and Rmax 1 γ . The Java program4 is run on an i5 CPU M540 at 2.53GHz. Experimental results are presented in Table 1. When convergence is not achieved, we look at the final L(b0) and U(b0) values to assess the efficiency of an algorithm. Note that, for inc-lc-HSVI, log2(λ) gives the number of restarts. inc-lc-HSVI s Restart Criteria We first look at the effect of the three restart criteria in inc-lc-HSVI through the top sub-table. The first two columns are similar, showing that not testing that L and U cross each other (no LXU) has little influence. Looking at execution traces, the LXU criterion is in fact only triggered when not checking for uniform improvability (no NUI). The time to converge is notably sped up by not testing for unstable results (no UR), with only one case of convergence to bad values in the Tiger problem (tiger70). More speed improvement is obtained by not testing uniform improvability (no NUI), in which case the LXU rule is triggered more often. As a result, we take as our default configuration the no NUI setting which only uses the LXU and UR stopping criteria. Comparing Approximators and Algorithms We now compare the four algorithms at hand through the bottom sub-table. pwlc-HSVI (when applicable) dominates overall the experiments, except on a few cases where inc-lc-HSVI converges in less time. As can be observed, the Lipschitz constants obtained by inc-lc-HSVI are of the same order of magnitude as the ones derived in pwlc HSVI from the final lower bounds L. inc-lc-HSVI(no NUI) would be a satisfying solution on the benchmarks at hand (when not using the PWLC property) if not lacking theoretical guarantees. For its part, lc-HSVI ends up with worst-case constants orders of magnitude larger in many cases, which suggests that its bounds have little generalization capabilities, as in pw-HSVI. pw-HSVI is obviously faster than lc-HSVI due to much cheaper updates. 7 Discussion This work shows that, for finite horizons, the optimal value function of a ρ-POMDP with Lipschitz ρ is Lipschitz-continuous (LC). The Lipschitz-continuity here is not uniform with a scalar constant, but local with a vector constant, which allows for more efficient updates. While the PWLC property (of V ) provides useful generalizing lower and upper bounds for POMDPs and ρ-POMDPs with convex ρ, the LC property provides similar bounds for POMDPs and ρ-POMDPs with LC ρ where V may not be convex. These bounds are envelopes of either upwardor downward-pointing cones , and, 3http://www.pomdp.org/examples/ 4Full code available here: https://gitlab.inria.fr/buffet/lc-hsvi-nips18 Table 1: Comparison of (top) inc-lc-HSVI with all 3 stopping criteria, or with 1 of them disabled and (bottom) x-HSVI algorithms (for x {pwlc, pw, lc, inc-lc}), in terms of (i) CPU time (timeout 600s), (ii) number of trajectories, (iii-iv) width (gap) at b0, and (v) Lipschitz constant x-HSVI inc-lc inc-lc(no LXU) inc-lc(no NUI) inc-lc(no UR) t (s) (#it) gap(b0) λ t (s) (#it) gap(b0) λ t (s) (#it) gap(b0) λ t (s) (#it) gap(b0) λ 4x3.95 1 (254) 0.10 1 1 (254) 0.10 1 1 (254) 0.10 1 1 (254) 0.10 1 4x4.95 0 (125) 0.10 1 0 (125) 0.10 1 0 (125) 0.10 1 0 (125) 0.10 1 cheese.95 0 (69) 0.10 1 0 (69) 0.10 1 0 (69) 0.10 1 0 (69) 0.10 1 cit 600 (44) 0.67 1 600 (48) 0.67 1 600 (41) 0.67 1 600 (44) 0.67 1 hallway 600 (645) 1.11 1 600 (642) 1.11 1 600 (611) 1.12 1 600 (622) 1.12 1 hallway2 600 (687) 1.00 1 600 (686) 1.00 1 600 (668) 1.00 1 600 (694) 1.00 1 milos-aaai97 600 (1331) 49.25 512 600 (1300) 49.45 512 600 (1797) 43.78 64 600 (1287) 49.47 512 mit 600 (46) 0.69 1 600 (47) 0.69 1 600 (47) 0.69 1 600 (46) 0.69 1 network 600 (8572) 0.44 512 600 (8462) 0.47 512 34 (2819) 0.10 128 258 (6849) 0.10 256 paint.95 0 (84) 0.10 1 0 (84) 0.10 1 0 (84) 0.10 1 0 (84) 0.10 1 pentagon 600 (62) 0.73 1 600 (56) 0.73 1 600 (60) 0.73 1 600 (56) 0.73 1 shuttle.95 0 (47) 0.08 4 0 (47) 0.08 4 0 (47) 0.08 4 0 (47) 0.08 4 tiger85 0 (15) 0.07 64 0 (15) 0.07 64 0 (15) 0.07 64 0 (16) 0.07 32 tiger-grid 600 (455) 3.95 8 600 (449) 3.95 8 600 (626) 0.78 1 600 (456) 3.95 8 grid-info kx 12 (242) 0.95 64 12 (242) 0.95 64 1 (279) 0.10 4 8 (350) 0.10 32 grid-info ky 27 (395) 0.10 512 26 (395) 0.10 512 1 (279) 0.10 4 26 (395) 0.10 512 grid-info kx 600 (1482) 0.83 128 600 (1535) 0.72 128 4 (695) 0.10 4 151 (1572) 0.10 32 grid-info ky 600 (757) 9.19 2048 600 (717) 9.40 2048 4 (707) 0.10 4 600 (773) 9.14 2048 x-HSVI pwlc pw lc inc-lc(no NUI) t (s) (#it) gap(b0) λ t (s) (#it) gap(b0) t (s) (#it) gap(b0) λ t (s) (#it) gap(b0) λ 4x3.95 1 (134) 0.10 1.19 600 (447) 0.94 600 (214) 3.27 1.9e+05 1 (254) 0.10 1 4x4.95 1 (120) 0.10 0.66 1 (134) 0.10 6 (134) 0.10 2.4e+05 0 (125) 0.10 1 cheese.95 0 (59) 0.10 1.15 0 (69) 0.10 1 (69) 0.10 4.8e+04 0 (69) 0.10 1 cit 600 (19) 0.13 1.39 600 (126) 0.84 601 (34) 0.84 3.0e+01 600 (41) 0.67 1 hallway 600 (414) 0.35 0.70 600 (683) 1.30 600 (203) 1.33 1.8e+00 600 (611) 1.12 1 hallway2 600 (385) 0.67 0.63 600 (690) 1.04 600 (208) 1.06 4.0e+00 600 (668) 1.00 1 milos-aaai97 600 (1152) 29.55 89.49 600 (1725) 49.05 600 (595) 52.87 2.3e+03 600 (1797) 43.78 64 mit 600 (21) 0.12 1.81 600 (236) 0.87 600 (64) 0.88 1.9e+17 600 (47) 0.69 1 network 498 (7703) 0.10 168.22 600 (3021) 453.62 600 (941) 510.76 5.0e+16 34 (2819) 0.10 128 paint.95 2 (143) 0.10 1.00 600 (3695) 3.55 600 (1008) 4.37 1.7e+281 0 (84) 0.10 1 pentagon 601 (27) 0.31 1.00 600 (89) 0.83 600 (32) 0.83 2.8e+01 600 (60) 0.73 1 shuttle.95 0 (23) 0.10 22.77 0 (42) 0.09 0 (42) 0.09 7.5e+00 0 (47) 0.08 4 tiger85 0 (15) 0.09 55.00 0 (15) 0.08 0 (15) 0.08 2.2e+02 0 (15) 0.07 64 tiger-grid 600 (264) 0.51 11.45 600 (1563) 16.74 600 (375) 17.23 4.5e+01 600 (626) 0.78 1 grid-info kx ( ) 600 (709) 0.24 600 (358) 1.82 3.6e+01 1 (279) 0.10 4 grid-info ky ( ) 27 (432) 0.10 344 (426) 0.10 4.0e+01 1 (279) 0.10 4 grid-info kx ( ) 600 (2319) 9.15 600 (889) 13.74 3.3e+01 4 (695) 0.10 4 grid-info ky ( ) 600 (1393) 6.66 600 (604) 8.52 3.1e+01 4 (707) 0.10 4 with appropriate initializations, are uniformly improvable. Two algorithms are proposed: HSVI used with these LC bounds which preserves HSVI s convergence properties , and an incremental algorithm that searches for a (uniform) scalar Lipschitz constant allowing for fast computations with no guarantees that the bounds are valid. The experiments show that there lc-HSVI s pessimistic constants are far from inc-lc-HSVI s guesses. This encourages searching for better (safe) Lipschitz constants possibly using a particular norm such that the dynamics of the b MDP are LC,5 as Platzman [1977] did for sub-rectangular b MDPs (a restrictive class of problems) but also improving the initialization of L and U, and possibly inc-lc-HSVI s restart and stopping criteria (ideally guaranteeing that a valid constant is found). We also aim at exploiting the Lipschitz continuity to solve partially observable stochastic games (POSGs) [Hansen et al., 2004]. Indeed, while the PWLC property allows efficiently solving not only POMDPs, but also Dec-POMDPs turned into occupancy MDPs [Dibangoye et al., 2013, 2016], the LC property may allow to provide generalizing bounds for POSGs turned into occupancy SGs, starting with 2-player 0-sum scenarios. Acknowledgments Let us thank David Reboullet for helping with some proofs (reminding us about the power of the triangle (in)equality), and the anonymous reviewers for their insightful comments. 5Previously cited works on LC MDPs rely on LC dynamics. M. Araya-López, O. Buffet, V. Thomas, and F. Charpillet. A POMDP extension with belief-dependent rewards. In Advances in Neural Information Processing Systems 23 (NIPS-10), 2010. K. Astrom. Optimal control of Markov processes with incomplete state information. Journal of Mathematical Analysis and Applications, 10(1), 1965. ISSN 0022-247X. R. Bellman. A Markovian decision process. Journal of Mathematics and Mechanics, 6(5), 1957. J. Dibangoye, C. Amato, O. Buffet, and F. Charpillet. Optimally solving Dec-POMDPs as continuousstate MDPs. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence (IJCAI-13), 2013. J. Dibangoye, C. Amato, O. Buffet, and F. Charpillet. Optimally solving Dec-POMDPs as continuousstate MDPs. Journal of Artificial Intelligence Research, 55, 2016. URL http://www.jair.org/ papers/paper4623.html. F. Dufour and T. Prieto-Rumeau. Approximation of Markov decision processes with general state space. Journal of Mathematical Analysis and Applications, 388(2), 2012. ISSN 0022-247X. doi: https://doi.org/10.1016/j.jmaa.2011.11.015. URL http://www.sciencedirect.com/science/ article/pii/S0022247X11010353. M. Egorov, M. J. Kochenderfer, and J. J. Uudmae. Target surveillance in adversarial environments using POMDPs. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI-16), 2016. R. Fonteneau, S. Murphy, L. Wehenkel, and D. Ernst. Inferring bounds on the performance of a control policy from a sample of trajectories. In Proceedings of the IEEE Symposium on Approximate Dynamic Programming and Reinforcement Learning (ADPRL-09), 2009. D. Fox, W. Burgard, and S. Thrun. Active Markov localization for mobile robots. Robotics and Autonomous Systems, 25(3 4), 1998. doi: http://dx.doi.org/10.1016/S0921-8890(98)00049-9. E. Hansen, D. Bernstein, and S. Zilberstein. Dynamic programming for partially observable stochastic games. In Proceedings of the Nineteenth National Conference on Artificial Intelligence (AAAI-04), 2004. K. Hinderer. Lipschitz continuity of value functions in Markovian decision processes. Mathematical Methods of Operations Research, 62(1), Sep 2005. doi: 10.1007/s00186-005-0438-1. URL https://doi.org/10.1007/s00186-005-0438-1. S. Ieong, N. Lambert, Y. Shoham, and R. Brafman. Near-optimal search in continuous domains. In Proceedings of the National Conference on Artificial Intelligence (AAAI-07), 2007. H. Kurniawati, D. Hsu, and W. Lee. SARSOP: Efficient point-based POMDP planning by approximating optimally reachable belief spaces. In Robotics: Science and Systems IV, 2008. R. Laraki and W. D. Sudderth. The preservation of continuity and Lipschitz continuity by optimal reward operators. Mathematics of Operations Research, 29(3), 2004. L. Mihaylova, T. Lefebvre, H. Bruyninckx, and J. D. Schutter. NATO Science Series on Data Fusion for Situation Monitoring, Incident Detection, Alert and Response Management, volume 198, chapter Active Robotic Sensing as Decision Making with Statistical Methods. 2006. V. Mnih, K. Kavukcuoglu, D. Silver, A. Graves, I. Antonoglou, D. Wierstra, and M. Riedmiller. Playing atari with deep reinforcement learning. In NIPS Deep Learning Workshop, 2013. J. Pineau, G. Gordon, and S. Thrun. Point-based value iteration: An anytime algorithm for POMDPs. In Proceedings of the Eighteenth International Joint Conference on Artificial Intelligence (IJCAI03), 2003. J. Pineau, G. Gordon, and S. Thrun. Anytime point-based approximations for large POMDPs. Journal of Artificial Intelligence Research (JAIR), 27, 2006. L. K. Platzman. Finite Memory Estimation and Control of Finite Probabilistic Systems. Ph D thesis, Massachusetts Institute of Technology (MIT), 1977. P. Poupart, K.-E. Kim, and D. Kim. Closing the gap: Improved bounds on optimal POMDP solutions. In Proceedings of the Twenty-First International Conference on Automated Planning and Scheduling (ICAPS-11), 2011. E. Rachelson and M. Lagoudakis. On the locality of action domination in sequential decision making. In Proc. of the International Symposium on Artificial Intelligence and Mathematics (ISAIM-10), 2010. Y. Satsangi, S. Whiteson, and M. T. J. Spaan. An analysis of piecewise-linear and convex value functions for active perception POMDPs. Technical Report IAS-UVA-15-01, IAS, Universiteit van Amsterdam, 2015. R. Smallwood and E. Sondik. The optimal control of partially observable Markov decision processes over a finite horizon. Operation Research, 21, 1973. T. Smith. Probabilistic Planning for Robotic Exploration. Ph D thesis, The Robotics Institute, Carnegie Mellon University, 2007. T. Smith and R. Simmons. Heuristic search value iteration for POMDPs. In Proceedings of the Annual Conference on Uncertainty in Artificial Intelligence (UAI-04), 2004. T. Smith and R. Simmons. Point-based POMDP algorithms: Improved analysis and implementation. In Proceedings of the Twenty-First Conference on Uncertainty in Artificial Intelligence (UAI-05), 2005. E. Sondik. The Optimal Control of Partially Observable Markov Decision Processes. Ph D thesis, Stanford University, 1971. M. T. Spaan, T. S. Veiga, and P. U. Lima. Decision-theoretic planning under uncertainty with information rewards for active cooperative perception. Autonomous Agents and Multi-Agent Systems, 29(6), 2015. N. L. Zhang and W. Zhang. Speeding up the convergence of value iteration in partially observable Markov decision processes. Journal of Artificial Intelligence Research, 14, 2001. URL http: //dx.doi.org/10.1613/jair.761. Z. Zhang, D. Hsu, and W. S. Lee. Covering number for efficient heuristic-based pomdp planning. In Proceedings of the 31st International Conference on Machine Learning (ICML-14), 2014. URL http://jmlr.org/proceedings/papers/v32/zhanga14.pdf.