# futuredependent_valuebased_offpolicy_evaluation_in_pomdps__5e5f51c1.pdf Future-Dependent Value-Based Off-Policy Evaluation in POMDPs Masatoshi Uehara Genentech uehara.masatoshi@gene.com Haruka Kiyohara Cornell University hk844@cornell.edu Andrew Bennett Morgan Stanley Andrew.Bennett@morganstanley.com Victor Chernozhukov MIT vchern@mit.edu Nan Jiang UIUC nanjiang@illinois.edu Nathan Kallus Cornell University kallus@cornell.edu Chengchun Shi LSE c.shi7@lse.ac.uk Wen Sun Cornell University ws455@cornell.edu We study off-policy evaluation (OPE) for partially observable MDPs (POMDPs) with general function approximation. Existing methods such as sequential importance sampling estimators suffer from the curse of horizon in POMDPs. To circumvent this problem, we develop a novel model-free OPE method by introducing future-dependent value functions that take future proxies as inputs and perform a similar role to that of classical value functions in fully-observable MDPs. We derive a new off-policy Bellman equation for future-dependent value functions as conditional moment equations that use history proxies as instrumental variables. We further propose a minimax learning method to learn future-dependent value functions using the new Bellman equation. We obtain the PAC result, which implies our OPE estimator is close to the true policy value under Bellman completeness, as long as futures and histories contain sufficient information about latent states. Our code is available at https://github.com/aiueola/neurips2023-future-dependent-ope. 1 Introduction Reinforcement learning (RL) has demonstrated success when it is possible to interact with the environment and collect data in an adaptive manner. However, for domains such as healthcare, education, robotics, and social sciences, online learning can be risky and unethical [e.g., KHS+15, KL19, TDHL19, SRRH21]. To address this issue, a variety of offline RL methods have recently been developed for policy learning and evaluation using historical data in these domains [see LKTF20, for an overview]. In this paper, we focus on the off-policy evaluation (OPE) problem, which concerns estimating the value of a new policy (called the evaluation policy) using offline log data that was Co-first author. This work was done at Cornell University Co-first author. This work was done at Tokyo Institute of Technology This work was done at Cornell University 37th Conference on Neural Information Processing Systems (Neur IPS 2023). generated under a different policy (called the behavior policy) [TB16, JL16, TDHL19, CQ22]. OPE is especially useful in the aforementioned high-stakes domains. We focus on off-policy evaluation (OPE) in partially observable Markov decision processes (POMDPs). Partial observability is a common phenomenon in practical applications [Cas98, KLC98, XLSS20], and it poses a serious challenge for sample-efficient learning. Most existing OPE methods are developed for MDPs and their statistical learning guarantees rely crucially on the Markov assumption. To extend these methods to POMDPs, one may use the entire history of observations as a state to satisfy Markovanity. This allows us to apply sequential importance sampling [SIS, Pre00, JSLZ19, HW21] or its variant [e.g., sequential doubly robust (SDR) methods RRS98, Mvd LRG01, JL16, TB16, FCG18, BMVVDL19] to the transformed data for valid policy evaluation. An alternative approach is to employ value-based methods by constructing a history-dependent Q-function that takes the entire history as input. These value functions can be estimated via fitted-Q evaluation [FQE, EGW05, MS08, LVY19] or minimax methods [ASM08, CJ19, FLL19, NCDL19, UHJ20, UIJ+21, ZW22]. However, all aforementioned estimators suffer from the curse of horizon, as their estimation errors grow exponentially with respect to the (effective) horizon and become prohibitively large in long-horizon settings [LLTZ18, KU20, KU22]. The goal of this paper is to devise practical and efficient OPE methods in large partially observable environments, while breaking the curse of horizon. As a first step, we restrict our attention to the evaluation of short memory-based policies that take several recent observations instead of the entire history as inputs. Short memory-based policies are widely used in practice, since maintaining the whole previous observation leads to a well-known computational challenge dubbed as the curse of history [PGT06, GMR22]. For example, in the deep Q-network (DQN) algorithm, four game frames are stacked together to train the optimal policy in Atari [MKS+13], while Open AI Five uses a window of length 16 [BBC+19]. Even when considering the evaluation of policies with shortmemory or memory-less characteristics (i.e., policies that depend solely on current observations), na ıve methods such as SIS, SDR, and FQE still suffer from the curse of horizon4. Our proposal takes a model-free perspective and introduces a new concept called future-dependent value functions . The proposed value function does not involve latent states which are unobservable in POMDPs. Nor does it relies on the entire data history. It performs the role of standard value functions by incorporating future observations as proxies for latent states. We demonstrate that these future-dependent value functions satisfy a new off-policy Bellman equation for POMDPs, which involves past observations to approximate latent states. To estimate these future-dependent value functions from offline data using the new Bellman equation, we propose minimax learning methods that accommodate various function approximation, including deep neural networks, RKHS, and linear models. Notably, our method extends the classical LSTD [LP03] developed in MDPs to the POMDP setting when using linear models. To summarize, our contributions are as follows: We propose a novel model-free method for OPE that leverages future-dependent value functions. Our key idea is to utilize future and historical observations as proxies for latent states. We derive a new Bellman equation for learning future-dependent value functions. The proposed estimator can accommodate various types of function approximations. We provide PAC guarantees to demonstrate that our method can address the curse of horizon, and conduct numerical experiments to showcase its superiority over existing methods. Note that similar concepts of future-dependent value functions have recently been introduced in the context of OPE in confounded POMDPs [SUHJ22] and online RL in POMDPs [USL+22]. However, we focus on OPE in non-confounded POMDPs. Further detailed discussions can be found in Section 1.1. 1.1 Related Literature OPE in confounded POMDPs. OPE with unmeasured confounders has been actively studied [ZB16, NKYB20, TSM20, WYW20, KZ20, LFY+21, NJ21, GCZ+22, SZY+22, XZS+22, LMWY22, CWWY23, BSZ23]. Among them, [TSM20] adopted the POMDP model to formulate the OPE problem in the presence of unmeasured confounders. They borrowed ideas from the causal inference literature on double negative control adjustment [MGT18, MST18, CPS+20, 4Notice that even when restricting to short-memory policies, it is necessary to use the entire history instead of the short memory as the state to meet the Markov property. KMU21, SMNTT20] to derive consistent value estimators in tabular settings. Later, [BKLM21] and [SUHJ22] extend their proposal by incorporating confounding bridge functions, thereby enabling general function approximation. (see the difference between these bridge functions and the proposed future-dependent value function in Section B). However, these methods do not apply to our unconfounded POMDP setting. This is because these methods require the behavior policy to only depend on the latent state to ensure certain conditional independence assumptions. These assumptions are violated in our setting, where the behavior policy may depend on the observation a common scenario in practical applications. In addition, we show that in the unconfounded setting, it is feasible to leverage multi-step future observations to relax specific rank conditions in their proposal, which is found to be difficult in the confounded setting [NJ21]. Refer to Example 1. Online learning in POMDPs. In the literature, statistically efficient online learning algorithms with polynomial sample complexity have been proposed in tabular POMDPs [GDB16, ALA16, JKKL20, LCSJ22], linear quadratic Gaussian setting (LQG) [LAHA21, SSH20], latent POMDPs [KECM21] and reactive POMDPs/PSRs [KAL16, JKA+17]. All the methods above require certain model assumptions. To provide a more unified framework, researchers have actively investigated online learning in POMDPs with general function approximation, as evidenced by a vast body of work [ZUSL22, LNSJ22, CBM22, ZXZ+22]. As the most relevant work, by leveraging future-dependent value functions, [USL+22] propose an efficient PAC RL algorithm in the online setting. They require the existence of future-dependent value functions and a low-rank property of the Bellman loss. Instead, in our approach, while we similarly require the existence of future-dependent value functions, we do not require the low-rank property of Bellman loss. Instead, we use the invertibility condition in Theorem 1, i.e., we have informative history proxies as instrumental variables. As a result, the strategies to ensure efficient PAC RL in POMDPs are quite different in the offline and online settings. Learning dynamical systems via spectral learning. There is a rich literature on POMDPs by representing them as predictive state representations (PSRs) [Jae00, LS01, RGT04, SJR04]. PSRs are models of dynamical systems that represent the state as a vector of predictions about future observable events. They are appealing because they can be defined directly from observable data without inferring hidden variables and are more expressive than Hidden Markov Models (HMMs) and POMDPs [SJR04]. Several spectral learning algorithms have been proposed for PSRs [BSG11, HKZ12, BGG13], utilizing conditional mean embeddings or Hilbert space embeddings. These approaches provide closed-form learning solutions, which simplifies computation compared to EM-type approaches prone to local optima and non-singularity issues. While these methods have demonstrated success in real-world applications [BSG11] and have seen subsequent improvements [KJS15, HDG15, VSH+16, DHB+17], it is still unclear how to incorporate more flexible function approximations such as neural networks in a model-free end-to-end manner with a valid PAC guarantee. Our proposed approach not only offers a new model-free method for OPE, but it also incorporates these model-based spectral learning methods when specialized to dynamical system learning with linear models (see Section F). Planning in POMDPs. There is a large amount of literature on planning in POMDPs. Even if the models are known, exact (or nearly optimal) planning in POMDPs is known to be NP-hard in the sense that it requires exponential time complexity with respect to the horizon [PT87, BDRS96]. This computational challenge is often referred to as the curse of history [PGT06]. A natural idea to mitigate this issue is to restrict our attention to short-memory policies [AYA18, MY20, KY22, GMR22]. Practically, a short-memory policy can achieve good empirical performance, as mentioned earlier. This motivates us to focus on evaluating short-memory policies in this paper. 2 Preliminaries In this section, we introduce the model setup, describe the offline data and present some notations. Model setup. We consider an infinite-horizon discounted POMDP M = S, A, O, r, γ, O, T where S denotes the state space, A denotes the action space, O denotes the observation space, O : S (O) denotes the emission kernel (i.e., the conditional distribution of the observation given the state), T : S A (S) denotes the state transition kernel (i.e., the conditional distribution of the next state given the current state-action pair), r : S A R denotes the reward function, and γ [0, 1) is the discount factor. All the three functions T, r, O are unknown to the learner. For simplicity, we consider the evaluation of memory-less policies π : O (A) that solely depend on the current observation O in the main text. The extension to policies with short-memory is discussed in Section C. O0 A0 O1 A1 O2 A 1 O 1 Figure 1: Case with MH = 1, MF = 2. An action A is generated depending on O. The extension to memory-based policy is discussed in Section C. Next, given a memory-less evaluation policy πe, we define the parameter of interest, i.e., the policy value. Following a policy πe, the data generating process can be described as follows. First, S0 is generated according to some initial distribution νS (S). Next, the agent observes O0 O( | S0), executes the initial action A0 πe( | O0), receives a reward r(S0, A0), the environment transits to the next state S1 T( | S0, A0), and this process repeats. Our objective is to estimate J(πe) := Eπe [P t=0 γt Rt] , where the expectation Eπe is taken by assuming the data trajectory follows the evaluation policy πe. Offline data. We convert the trajectory data generated by a behavior policy πb : O (A), into a set of history-observation-action-reward-future transition tuples (denoted by Dtra) and a set of initial observations (denoted by Dini). The first data subset enables us to learn the reward, emission and transition kernels whereas the second data subset allows us to learn the initial observation distribution. Specifically, the dataset Dtra consists of N data tuples {(H(i), O(i), A(i), R(i), F (i))}N i=1. We use (H, O, A, R, F ) to denote a generic history-observation-action-reward-future tuple where H denotes the MH-step historical observations obtained prior to the observation O and F denotes the MF -step future observations after (O, A) for some integers MH, MF 1. Specifically, at a given time step t in the data trajectory, we use (O, A, R) to denote (Ot, At, Rt), and set H = (Ot MH:t 1, At MH:t 1) and F = (Ot+1:t+MF , At+1:t+MF 1). We additionally set F = (Ot:t+MF 1, At:t+MF 2). Note we use the prime symbol to represent the next time step. See Figure 1 for details when we set t = 0. Throughout this paper, we use uppercase letters such as (H, S, O, A, R, S , F ) to denote random variables in the offline data, and lowercase letters such as (h, s, o, a, r, s , f ) to denote their realizations, unless stated otherwise. To simplify the presentation, we assume the stationarity of the environment, i.e., the marginal distributions of (H, S, F) and (H , S , F ) are identical. The dataset Dini consists of N data tuples {O(i) 0:MF 1, A(i) 0:MF 2}N i=1 which is generated as follows: S0 νS, O0 O( | S0), A0 πb( | O0), S1 T( | S0, A0), , until we observe O(i) MF 1 and A(i) MF 1. We denote its distribution over F = OMF AMF 1 by νF( ). Remark 1 (Standard MDPs). Consider the setting where S = O and MH = 0, MF = 1. In that case, we set H to S instead of histories. Then, Dtra = {O(i), A(i), R(i), O (i)}. We often assume νO (νF in our setting) is known. This yields the standard OPE setting in MDPs [CJ19, UHJ20]. Notation. We streamline the notation as follows. We define a state value function under πe: V πe(s) := Eπe[P k=0 γk Rk | S0 = s] for any s S. Let dπe t ( ) be the marginal distribution of St under the policy πe. Then, we define the discounted occupancy distribution dπe( ) := (1 γ) P t=0 γtdπe t ( ). We denote the domain of F, H by F = (O A)MF 1 O, H = (O A)MH, respectively. The notations E, ED (without any subscripts) represent the population or sample average over the offline data D = Dtra Dini, respectively. We denote the distribution of offline data by Pπb( ). We define the marginal density ratio wπe(S) := dπe(S)/Pπb(S). Given a matrix C, we denote its Moore-Penrose inverse by C+ and its smallest singular value by σmin(C). For a given integer m > 0, let Im denote an m m identity matrix. Let denote outer product and [T] = {0, , T} for any integer T > 0. 3 Identification via Future-dependent Value Functions In this section, we present our proposal to identify policy values under partial observability by introducing future-dependent value functions. We remark that the target policy s value is identifiable from the observed data via SIS or SDR. Nonetheless, as commented earlier in Section 1, these methods suffer from the curse of horizon. Here, we propose an alternative identification approach that can possibly circumvent the curse of horizon. It serves as a building block to motivate the proposed estimation methods in Section 4. In fully observable MDPs, estimating a policy s value is straightforward using the value functionbased method J(πe) = Es νS[V πe(s)]. However, in partial observable environments where the latent state is inaccessible, the state value function V πe(s) is unidentifiable. To address this challenge, we propose the use of future-dependent value functions that are defined based on observed variables and serve a similar purpose to state value functions in MDPs. Definition 1 (Future-dependent value functions). Future-dependent value functions g V [F R] are defined such that the following holds almost surely, E[g V (F) | S] = V πe(S). Recall that the expectation is taken with respect to the offline data generated by πb. Crucially, the future-dependent value functions mentioned above may not always exist, and they need not be unique. Existence is a vital assumption in our framework, although we don t insist on uniqueness. We will return to the topics of existence and uniqueness after demonstrating their relevance in offline policy evaluation. Hereafter, we explain the usefulness of future-dependent value functions. Future-dependent value functions are defined as embeddings of value functions on latent states onto multi-step futures. Notice that J(πe) = Es νS[V πe(s)] = Ef νF [g V (f)]. These future-dependent value functions are useful in the context of OPE as they enable us to accurately estimate the final policy value. However, the future-dependent value function itself cannot be identified since its definition relies on unobserved states. To overcome this challenge, we introduce a learnable counterpart called the learnable futuredependent value function. This learnable version is defined based on observed quantities and thus can be identified. Definition 2 (Learnable future-dependent value functions). Define µ(O, A) := πe(A | O)/πb(A | O). Learnable future-dependent value functions b V [F R] are defined such that the following holds almost surely, 0 = E [µ(O, A){R + γb V (F )} b V (F) | H] . (1) Recall that the expectation is taken with respect to the offline data generated by πb. We denote the set of solutions by BV . To motivate this definition, we recall that the off-policy Bellman equation in MDPs [DNP+14] can be expressed as follows: V πe(S) = E h µ(O, A)(R + γV πe(S )) | S i . (2) Then, by the definition of future-dependent value functions and certain conditional independence relations (F (O, A) | S ), we obtain that 0 = E [µ(O, A) {R + γg V (F )} g V (F) | S] . (3) Since H (O, A, F ) | S, taking another conditional expectation given H on both sides yields that 0 = E [µ(O, A) {R + γg V (F )} g V (F) | H] . (4) Therefore, (4) can be seen as an off-policy Bellman equation for future-dependent value functions, analogous to the Bellman equation (2) in MDPs. Based on the above discussion, we present the following lemma. Lemma 1. Future-dependent value functions are learnable future-dependent value functions. Remark 2 (Non-stationary case). When the offline data is non-stationary, i.e, the pdfs of (H, S) and (H , S ) are different, we need to additionally require E[g V (F ) | S ] = V πe(S ) in the definition. Remark 3 (Comparisons with related works). Similar concepts have been recently proposed in the context of confounded POMDPs [see e.g., SUHJ22, BK21]. However, our proposal significantly differs from theirs. First, their confounded setting does not cover our unconfounded setting because their behavior policies πb is not allowed to depend on current observations as mentioned in Section 1.1. Secondly, their proposal heir proposal overlooks the significant aspect of incorporating multi-step future observations, which plays a pivotal role in facilitating the existence of future-dependent value functions, as will be discussed in Example 1. For a detailed discussion, refer to Section B. Algorithm 1 Minimax OPE on POMDPs Require: Dataset D, function classes Q [F R], Ξ [H R], hyperparameter λ 0 1: ˆb V = argminq Q maxξ Ξ EDtra[{µ(A, O){R + γq(F )} q(F)}ξ(H) λξ2(H)]. 2: return ˆJVM = EDini[ˆb V (f)] Finally, we present a theorem to identify the policy value. Theorem 1 (Identification). Suppose (1a) the existence of learnable future-dependent value functions (need not be unique); (1b) the invertiblity condition, i.e., any g : S R that satisfies E[g(S) | H] = 0 must also satisfy g(S) = 0 ( i.e., g(s) = 0 for almost every s that belongs to the support of S), (1c) the overlap condition wπe(S) := dπe(S)/Pπb(S) < , µ(O, A) < . Then, for any b V BV , J(πe) = Ef νF [b V (f)]. (5) We assume three key conditions: the observability condition (i.e., BV = ), the invertibility condition, and the overlap condition. We call Condition (1a) the observability condition since it is reduced to the well-known concept of observability in the LQG control theory, as we will see in Section D. While Condition (1a) itself is concerned with learnable future-dependent value functions, it is implied by the existence of unlearnable future-dependent value functions according to Lemma 1 which can be verified using Picard s theorem in functional analysis [CFR07]. In general, the observability condition requires the future proxy F to contain sufficient information about S and is likely to hold when F consists of enough future observations. We will see more interpretable conditions in the tabular POMDPs (Example 1) and POMDPs with Hilbert space embeddings (HSE-POMDPs) where the underlying dynamical systems have linear conditional mean embeddings (Example 5 in Section D). The invertibility condition is imposed to ensure that a learnable future-dependent value function b V satisfies (3) (note that right hand side of Eq. 3 is a function of Z, S instead of H). Again, we will present more interpretable conditions in the tabular and linear settings below and in Section D. Roughly speaking, it requires H to retain sufficient information about S. In that sense, the history proxy serves as an instrumental variable (IV), which is widely used in economics [Hor11, New13] 5. Finally, the overlap condition is a standard assumption in OPE [UIJ+21]. Example 1 (Tabular Setting). In the tabular case, abstract assumptions in Theorem 1 are reduced to certain rank assumptions. We first define Sb = {s S : Pπb(s) > 0}. We define a matrix Prπb(F | Sb) R|F| |S| whose (i, j)-th element is Prπb(F = xi | S = x j) where xi is the ith element in F, and x j is the jth element in Sb. We similarly define another matrix Prπb(Sb, H) whose (i, j)-th element is Prπb(S = x i, H = x j ) where x j denotes the jth element in H. Lemma 2 (Sufficient conditions for observability and invertibility ). (a) When rank(Prπb(F | Sb)) = |Sb|, future-dependent value functions exist. Then, from Lemma 1, learnable future-dependent value functions exist. (b) The invertiblity is satisfied when rank(Prπb(Sb, H)) = |Sb|. The first two conditions require that the cardinalities of F and H must be greater than or equal to Sb, respectively. The proof of Lemma 2 is straightforward, as integral equations reduce to matrix algebra in the tabular setting. We note that similar conditions have been assumed in the literature on HMMs and POMDPs [SBS+10, BSG11, BGG13]. In particular, Prπb(O | Sb) = |Sb| has been imposed in previous works [NJ21, SUHJ22] in the context of confounded POMDPs. Nonetheless, our condition Prπb(F | Sb) = |Sb| is strictly weaker when F includes multi-step future observations, demonstrating the advantage of incorporating multi-step future observations compared to utilizing only the current observation. A more detailed discussion can be found in Appendix D.1. 4 Estimation with General Function Approximation In this section, we demonstrate how to estimate the value of a policy based on the results presented in Section 3. We begin by outlining the proposed approach for estimating b V ( ). The key observation is 5The observation that history can serve as an instrumental variable in POMDPs is mentioned in [HDG15, VSH+16]. However, these works aim to learn the system dynamics instead of policy evaluation or learning. Hence, concepts like future-dependent value functions do not appear in these works. that it satisfies E[L(b V , ξ)] = 0 for any ξ : H R where L(q, ξ) is defined as L(q, ξ) := {µ(A, O){R + γq(F )} q(F)}ξ(H) for q : F R and ξ : H R. Given some constrained function classes Q [F R] and Ξ [H R] and a hyperparameter λ 0, the estimator is computed according to Line 1 of Algorithm 1. When the realizability BV Q = holds and Ξ is unconstrained, we can easily show that the population-level minimizers argminq Q maxξ Ξ E[L(q, ξ) 0.5λξ2(H)] are all learnable future-dependent value functions. This is later formalized in Theorem 2. We can use any function classes such as neural networks, RKHS, and random forests to parameterize Q and Ξ. Here, the function class Ξ plays a critical role in measuring how q deviates from the ground truth b V . The hyperparameter λ is introduced to obtain a fast convergence rate. We call it a stabilizer instead of a regularizer since λ does not need to shrink to zero as n approaches infinity. Note that regularizers are needed when we penalize the norms of q and ξ, i.e., ˆb V = argmin q Q max ξ Ξ ED[L(q, ξ) 0.5λξ2(H)] + 0.5α q 2 Q 0.5α ξ 2 Ξ, for certain function norms Q and Ξ and hyperparameters α , α > 0. In the remainder of this section, we present three concrete examples using linear models, RKHSs and neural networks. Let Q and Ξ denote L2-norms when we use linear models and RKHS norms when using RKHSs. Example 2 (Linear models). Suppose Q, Ξ are linear, i.e., G = {θ ϕF( ) | θ Rd F }, Ξ = {θ ϕH( ) | θ Rd H} with features ϕF : F Rd F , ϕH : H Rd H. Then, ˆb V ( ) = ϕ F( ) M 2 {αId H + λM3} 1M2 + α Id F 1 M 2 {αId H + λM3} 1M1, M1 = ED[µ(O, A)RϕH(H)], M2 = ED[ϕH(H){ϕ F(F) γµ(O, A)ϕ F(F )}], M3 = ED[ϕH(H)ϕ H(H)]. When α = 0, α = 0, λ = 0, the value estimators boil down to ˆb V ( ) = ϕ F( )M+ 2 M1, ˆJVM = Ef νF [ϕ F(f)]M+ 2 M1. The above estimators are closely related to the LSTD estimators in MDPs [LP03]. Specifically, the off-policy LSTD estimator for state-value functions [DNP+14, UHJ20] is given by Es νS[ϕ S (s)]ED[ϕS(S){ϕ S (S) γµ(S, A)ϕ S (S )}]+ED[µ(S, A)RϕS(S)]. (6) Our new proposed estimator ˆJVM is Ef νF [ϕ F(f)]ED[ϕH(H){ϕ F(F) γµ(O, A)ϕ F(F )}]+ED[µ(O, A)RϕH(H)], which is very similar to (6). The critical difference lies in that we use futures (including current observations) and histories as proxies to infer the latent state S under partial observability. Example 3 (RKHSs). Let Q, Ξ be RKHSs with kernels k F( , ) : F F R, k H( , ) : H H R. Then, ˆb V ( ) = k F( ) {K F} {αIn + KH} 1K F + α In 1 {K F} K1/2 H {αIn + KH} 1K1/2 H Y where Y Rn, k( ) Rn, KH Rn n, KF Rn n such that {KH}(i,j) = k H(H(i), H(j)), {KF}(i,j) = k F(F (i), F (j)), {Y}i = µ(O(i), A(i))R(i), {K F}(i,j) = k F(F (i), F (j)) γk F(F (i), F (j)), {k F( )}i = k F( , F (i)). Example 4 (Neural Networks). We set Q to a class of neural networks and recommend to set Ξ to a linear or RKHS class so that the minimax optimization is reduced to single-stage optimization. The resulting optimization problem can be solved by off-the-shelf stochastic gradient descent (SGD) methods. Specifically, when we set Ξ to a linear model, ˆb V is reduced to ˆb V = argmin q Q Z(q) {αI + λED[ϕH(H)ϕ H(H)]} 1Z(q), Z(q) := ED[µ(A, O){R + γq(F )} q(F)}ϕH(H)]. When we set Ξ to be an RKHS, ˆb V is reduced to ˆb V = argmin q Q {Z (q)} K1/2 H {αIn + λKH} 1K1/2 H Z(q), {Z (q)}i := µ(A(i), O(i)){R(i) + γq(F (i))} q(F (i)). Remark 4 (Comparison with minimax methods for MDPs). Our proposed methods are closely related to the minimax learning approach (a.k.a. Bellman residual minimization) for RL [ASM08]. More specifically, in MDPs where the evaluation policy is memory-less, [UIJ+21, TFL+20] proposed minimax learning methods based on the following equations, E[h(S){µ(S, A){γV πe(S ) + R} V πe(S)}] = 0, h : S R. These methods are no longer applicable in POMDPs since we are unable to directly observe the latent state. Although substituting the latent state with the entire history can still provide statistical guarantees, it is susceptible to the curse of horizon. Remark 5 (Modeling of system dynamics). Our minimax estimators can be extended to learning system dynamics. In particular, for tabular POMDPs, these results are closely related to the literature on spectral learning in HMMs and POMDPs [SBS+10, BSG11, BGG13]. Refer to Section F,G. Remark 6 (Finite horizon setting). The extension to the finite horizon setting is discussed in Section E 5 Finite Sample Results We study the finite sample convergence rate of the proposed estimators in this section. To simplify the technical analysis, we impose three assumptions. First, we assume the function classes are bounded, i.e., Q CQ, Ξ CΞ for some CQ, CΞ > 0. Second, we assume that the offline data are i.i.d.6 Third, we assume Q, Ξ are finite hypothesis classes. Meanwhile, our results can be extended to infinite hypothesis classes using the global/local Rademacher complexity theory; see e.g., [UIJ+21]. To simplify the presentation, following standard convention in OPE, we suppose the initial distribution is known, i.e., |Dini| = . Accuracy of ˆb V . We first demonstrate that ˆb V consistently estimates the learnable future-dependent value bridge functions. To formalize this, we introduce the following Bellman operators. Definition 3 (Bellman operators). The Bellman residual operator onto the history is defined as T : [F R] q( ) 7 E[µ(O, A){R + γq(F )} q(F) | H = ], and the Bellman residual error onto the history is defined as E[(T q)2(H)]. Similarly, the Bellman residual operator onto the latent state, T S is defined as T S : [F R] q( ) 7 E[µ(O, A){R + γq(F )} q(F) | S = ], and the Bellman residual error onto the latent state is defined as E[{T S(q)}2(S)]. The conditional expectations equal to zero when h and s lie outside the support of H and S, respectively. The Bellman residual error onto the history is zero for any learnable future-dependent value function, i.e., E[(T b V )2(H)] = 0. Thus, this is a suitable measure to assess how well value function-based estimators approximate the true learnable future-dependent value functions. Theorem 2 (Finite sample property of ˆb V ). Set λ > 0. Suppose (2a) BV Q = 0 (realizability) and (2b) T Q Ξ (Bellman completeness). With probability 1 δ, we have E[(T ˆb V )2(H)]1/2 c{1/λ + λ} max(1, CQ, CΞ) q ln(|Q||Ξ|c/δ) n where c is some universal constant. Note (2a) and (2b) are commonly assumed in the literature on MDPs [CJ19, UIJ+21] as well. In particular, Bellman completeness means that the function class Ξ is sufficiently rich to capture the Bellman update of functions in Q. For instance, these assumptions are naturally met in HSEPOMDPs. We require λ > 0 in the statement of Theorem 2 to obtain a parametric convergence rate. When λ = 0, although we can obtain a rate of Op(n 1/4), it is unclear whether we can achieve Op(n 1/2). Accuracy of ˆJVM. We derive the finite sample guarantee for ˆJVM. 6Without the independence assumption, similar results can be similarly established by imposing certain mixing conditions; see e.g., [SZLS20, LQM20, KU22]. Theorem 3 (Finite sample property of ˆJVM). Set λ > 0. Suppose (2a), (2b), (2c) any element in q Q that satisfies E[{T S(q)}(S) | H] = 0 also satisfies T S(q)(S) = 0. (2d) the overlap µ(O, A) < and any element in q Q that satisfies T S(q)(S) = 0 also satisfies T S(q)(S ) = 0 where S dπe(s). With probability 1 δ, we have |J(πe) ˆJVM| c(1/λ + λ) (1 γ)2 max(1, CQ, CΞ)IV1(Q)Dr Q[dπe, Pπb] ln(|Q||Ξ|c/δ) IV2 1(Q) := sup {q Q;E[{T (q)(H)}2] =0} E[{T S(q)(S)}2] E[{T (q)(H)}2] , (8) Dr2 Q[dπe, Pπb] := sup {q Q;Es Pπb [{T S(q)(s)}2] =0} Es dπe[{T S(q)(s)}2] Es Pπb [{T S(q)(s)}2]. (9) On top of (2a) and (2b) that are assumed in Theorem 2, when IV1(Q) < , Dr Q(dπe, Pπb) < , (2c) and (2d) hold, we have the non-vacuous PAC guarantee. The condition Dr Q(dπe, Pπb) < and (2d) are the overlap conditions, which are adaptive to the function class Q and are weaker than (1c). These are also used in MDPs [XCJ+21]. Here, Dr Q(dπe, Pπb) is a refined version of the density ratio and satisfies Dr Q(dπe, Pπb) wπe(S). The condition IV1(Q) < and (2c) are characteristic conditions in POMDPs that quantify how much errors are properly translated from on H to on S. Similar assumptions are frequently imposed in IV literature [DLMS20, CP12]. The upper error bound in (7) does not have explicit exponential dependence on the effective horizon 1/(1 γ). In particular, as shown in Section D, for tabular POMDPs and HSE-POMDPs, the terms Dr Q(dπe, Pπb) and κ(Q) can be reduced to certain condition numbers associated with covariance matrices spanned by feature vectors; see (15) and (16) in Appendix D. Hence, unlike SIS-based methods, we are able to break the curse of horizon. The numbers of future and history proxies included in F and H represents a tradeoff. Specifically, if F contains enough observations, it is likely that (2a) will hold. Meanwhile, if H contains enough observations, it is more likely that (2b) will hold. These facts demonstrate the benefits of including a sufficient number of observations in F and H. However, the statistical complexities ln(|Q||Ξ|) will increase with the number of observations in F and H. Lastly, it s worth noting that while our methods effectively address the curse of horizon, they may incur exponential growth concerning the number of future proxies used. This also applies to history proxies, which should be longer than the length of short memory policies. Here, we focus on the explanation of future proxies. For instance, in the tabular case, log |Ω| might scale with |O|MF when considering Ωas the set of all functions on OMF . However, this situation differs significantly from the curse of horizon, a challenge that naive methods like replacing states with the entire history encounter. These methods would necessitate the entire history to achieve Markovianity, whereas we only require a shorter length of future observations to establish the conditions outlined in Theorem 1 (existence), which can be much shorter. Specific examples are provided throughout the paper, including Example 1, which discusses the tabular setting and demonstrates that we essentially need as many future proxies as states, as long as there is sufficient statistical dependence between them. 6 Experiment This section empirically evaluates the performance of the proposed method on a synthetic dataset.7. We use the Cart Pole environment provided by Open AI Gym [BCP+16], which is commonly employed in other OPE studies [SUHJ22, FCG18]. By default, this non-tabular environment consists of 4dimensional states, which are fully observable. Following [SUHJ22], we create partial observability by adding independent Gaussian noises to each dimension of the state as O(j) = S(j)(1 + N(1 + 0.32)), 1 j 4. To define behavior and evaluation policies, we first train an expert policy using DDQN [VHGS16] on latent states S. Subsequently, we apply Behavior Cloning (BC) on two datasets, one containing pairs of latent state and action (S, A) and the other containing pairs of observation and action (O, A), respectively. Then, we use the base policy obtained by BC on the state-action 7Our code is available at https://github.com/aiueola/neurips2023-future-dependent-ope Figure 2: Logarithms of relative biases (left) and MSEs (right) of the proposed and the baseline estimators for various values of ϵ, which specify the evaluation policy. The confidence intervals are obtained through 100 Monte Carlo simulations. pairs (S, A) to define an ϵ-greedy behavior policy, where we set ϵ = 0.3. 8 Similarly, the evaluation policy is also an ϵ-greedy policy, based on the base policy obtained by BC on the observation-action pairs (O, A), with different values of ϵ [0.1, 0.3, 0.5, 0.7]. We conduct the experiment with 100 random seeds, and for each simulation, we collect logged data consisting of 1000 trajectories, each containing 100 steps. We compare our proposal with Sequential Importance Sampling (SIS) [Pre00] and the naive minimax OPE [UHJ20], which is designed for fully-observable MDPs and does not account for partial observability. The naive minimax estimator is defined as if the environment was fully observable, replacing H and F in Algorithm 1 with the current observation O. In contrast, our proposed method uses a 3-step history as H and a one-step future as F to address partial observability. Both our proposed method and the naive approach use two-layer neural networks for the function Q and RKHSs for Ξ, as detailed in Example 4. We present the results in Figure 2, which demonstrate the superior accuracy of our proposed estimator compared to the baselines (SIS and naive minimax estimator) in terms of mean square errors (MSEs). Additional experimental details and ablation results, including the variations in the length of H, F, and the choice of RKHSs, can be found in Appendix H.3. 7 Conclusion We present a novel approach for OPE of short-memory policies in POMDPs. Our method involves introducing future-dependent value functions and the associated off-policy Bellman equations, followed by proposing a minimax estimator based on these equations. This is the first model-free method that allows for general function approximation and mitigates the curse of horizon. Our proposal is grounded in three interpretable key assumptions: observability, which asserts the presence of (short) future observations retaining adequate information about latent states, invertibility, which posits the existence of (short) histories preserving ample information about latent states; and the overlap between evaluation policies and behavior policies. We have several avenues for enhancing our proposals. Firstly, automatically determining the appropriate lengths of futures and histories holds practical significance. Additionally, exploring recent attention mechanisms that extend beyond selecting the most recent history proxies or the nearest future proxies shows promise. Secondly, while we establish Bellman equations for POMDPs and use a simple minimax approach akin to [DLMS20], we may benefit from leveraging more refined methods introduced in recent research for solving conditional moment equations [BKM+23c, BKM+23a, BKM+23b]. Acknowledgements This material is based upon work supported by the National Science Foundation under Grant Nos. IIS 2112471, IIS 2141781, IIS 1846210, IIS 2154711 and an Engineering and Physical Sciences Research Council grant EP/W014971/1. 8When collecting offline data from the behavior policy, the behavior policy takes the observation O as input, noting that the dimension of S and O is the same in our experimental setting. [ALA16] Kamyar Azizzadenesheli, Alessandro Lazaric, and Animashree Anandkumar. Reinforcement learning of pomdps using spectral methods. In Conference on Learning Theory, pages 193 256. PMLR, 2016. [ASM08] Andr as Antos, Csaba Szepesv ari, and R emi Munos. Learning near-optimal policies with bellman-residual minimization based fitted policy iteration and a single sample path. Machine Learning, 71(1):89 129, 2008. [AYA18] Kamyar Azizzadenesheli, Yisong Yue, and Animashree Anandkumar. Policy gradient in partially observable environments: Approximation and convergence. ar Xiv preprint ar Xiv:1810.07900, 2018. [BBC+19] Christopher Berner, Greg Brockman, Brooke Chan, Vicki Cheung, Przemyslaw Debiak, Christy Dennison, David Farhi, Quirin Fischer, Shariq Hashme, Chris Hesse, et al. Dota 2 with large scale deep reinforcement learning. ar Xiv preprint ar Xiv:1912.06680, 2019. [BCP+16] Greg Brockman, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, and Wojciech Zaremba. Openai gym. ar Xiv preprint ar Xiv:1606.01540, 2016. [BDRS96] Dima Burago, Michel De Rougemont, and Anatol Slissenko. On the complexity of partially observed markov decision processes. Theoretical Computer Science, 157(2):161 183, 1996. [Ber12] Dimitri Bertsekas. Dynamic programming and optimal control: Volume I, volume 1. Athena scientific, 2012. [BGG13] Byron Boots, Geoffrey Gordon, and Arthur Gretton. Hilbert space embeddings of predictive state representations. ar Xiv preprint ar Xiv:1309.6819, 2013. [BK21] Andrew Bennett and Nathan Kallus. Proximal reinforcement learning: Efficient off-policy evaluation in partially observed markov decision processes. 2021. [BKLM21] Andrew Bennett, Nathan Kallus, Lihong Li, and Ali Mousavi. Off-policy evaluation in infinite-horizon reinforcement learning with latent confounders. In International Conference on Artificial Intelligence and Statistics, pages 1999 2007. PMLR, 2021. [BKM+23a] Andrew Bennett, Nathan Kallus, Xiaojie Mao, Whitney Newey, Vasilis Syrgkanis, and Masatoshi Uehara. Inference on strongly identified functionals of weakly identified functions. In The Thirty Sixth Annual Conference on Learning Theory, pages 2265 2265. PMLR, 2023. [BKM+23b] Andrew Bennett, Nathan Kallus, Xiaojie Mao, Whitney Newey, Vasilis Syrgkanis, and Masatoshi Uehara. Minimax instrumental variable regression and l2 convergence guarantees without identification or closedness. In Gergely Neu and Lorenzo Rosasco, editors, Proceedings of Thirty Sixth Conference on Learning Theory, volume 195 of Proceedings of Machine Learning Research, pages 2291 2318, 2023. [BKM+23c] Andrew Bennett, Nathan Kallus, Xiaojie Mao, Whitney Newey, Vasilis Syrgkanis, and Masatoshi Uehara. Source condition double robust inference on functionals of inverse problems. ar Xiv preprint ar Xiv:2307.13793, 2023. [BKS19] Andrew Bennett, Nathan Kallus, and Tobias Schnabel. Deep generalized method of moments for instrumental variable analysis. In Advances in Neural Information Processing Systems 32, pages 3564 3574. 2019. [BMVVDL19] Aurelien Bibaut, Ivana Malenica, Nikos Vlassis, and Mark Van Der Laan. More efficient off-policy evaluation through regularized targeted learning. In International Conference on Machine Learning, pages 654 663. PMLR, 2019. [BSG11] Byron Boots, Sajid M Siddiqi, and Geoffrey J Gordon. Closing the learning-planning loop with predictive state representations. The International Journal of Robotics Research, 30(7):954 966, 2011. [BSZ23] David Bruns-Smith and Angela Zhou. Robust fitted-q-evaluation and iteration under sequentially exogenous unobserved confounders. ar Xiv preprint ar Xiv:2302.00662, 2023. [Cas98] Anthony R Cassandra. A survey of pomdp applications. In Working notes of AAAI 1998 fall symposium on planning with partially observable Markov decision processes, volume 1724, 1998. [CBM22] Fan Chen, Yu Bai, and Song Mei. Partially observable rl with b-stability: Unified structural condition and sharp sample-efficient algorithms. ar Xiv preprint ar Xiv:2209.14990, 2022. [CFR07] Marine Carrasco, Jean-Pierre Florens, and Eric Renault. Linear inverse problems in structural econometrics estimation based on spectral decomposition and regularization. Handbook of econometrics, 6:5633 5751, 2007. [CJ19] Jinglin Chen and Nan Jiang. Information-theoretic considerations in batch reinforcement learning. In International Conference on Machine Learning, pages 1042 1051. PMLR, 2019. [CKPW18] Zhiyong Cui, Ruimin Ke, Ziyuan Pu, and Yinhai Wang. Deep bidirectional and unidirectional lstm recurrent neural network for network-wide traffic speed prediction. ar Xiv preprint ar Xiv:1801.02143, 2018. [CNS18] Victor Chernozhukov, Whitney K Newey, and Rahul Singh. Learning l2 continuous regression functionals via regularized riesz representers. ar Xiv preprint ar Xiv:1809.05224, 2018. [CNSS20] Victor Chernozhukov, Whitney Newey, Rahul Singh, and Vasilis Syrgkanis. Adversarial estimation of riesz representers. ar Xiv preprint ar Xiv:2101.00009, 2020. [CO20] Sayak Ray Chowdhury and Rafael Oliveira. No-regret reinforcement learning with value function approximation: a kernel embedding approach. ar Xiv preprint ar Xiv:2011.07881, 2020. [CP12] Xiaohong Chen and Demian Pouzo. Estimation of nonparametric conditional moment models with possibly nonsmooth generalized residuals. Econometrica, 80(1):277 321, 2012. [CPS+20] Yifan Cui, Hongming Pu, Xu Shi, Wang Miao, and Eric Tchetgen Tchetgen. Semiparametric proximal causal inference. ar Xiv preprint ar Xiv:2011.08411, 2020. [CQ22] Xiaohong Chen and Zhengling Qi. On well-posedness and minimax optimal rates of nonparametric q-function estimation in off-policy evaluation. ar Xiv preprint ar Xiv:2201.06169, 2022. [CWWY23] Siyu Chen, Yitan Wang, Zhaoran Wang, and Zhuoran Yang. A unified framework of policy learning for contextual bandit with confounding bias and missing observations. ar Xiv preprint ar Xiv:2303.11187, 2023. [DHB+17] Carlton Downey, Ahmed Hefny, Byron Boots, Geoffrey J Gordon, and Boyue Li. Predictive state recurrent neural networks. Advances in Neural Information Processing Systems, 30, 2017. [DJW20] Yaqi Duan, Zeyu Jia, and Mengdi Wang. Minimax-optimal off-policy evaluation with linear function approximation. In International Conference on Machine Learning, pages 2701 2709. PMLR, 2020. [DLMS20] Nishanth Dikkala, Greg Lewis, Lester Mackey, and Vasilis Syrgkanis. Minimax estimation of conditional moment models. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 12248 12262, 2020. [DNP+14] Christoph Dann, Gerhard Neumann, Jan Peters, et al. Policy evaluation with temporal differences: A survey and comparison. Journal of Machine Learning Research, 15:809 883, 2014. [EGW05] Damien Ernst, Pierre Geurts, and Louis Wehenkel. Tree-based batch mode reinforcement learning. Journal of Machine Learning Research, 6:503 556, 2005. [FCG18] Mehrdad Farajtabar, Yinlam Chow, and Mohammad Ghavamzadeh. More robust doubly robust off-policy evaluation. In International Conference on Machine Learning, pages 1447 1456. PMLR, 2018. [FLL19] Yihao Feng, Lihong Li, and Qiang Liu. A kernel loss for solving the bellman equation. Advances in Neural Information Processing Systems, 32, 2019. [FS19] Dylan J Foster and Vasilis Syrgkanis. Orthogonal statistical learning. ar Xiv preprint ar Xiv:1901.09036, 2019. [GCZ+22] Hongyi Guo, Qi Cai, Yufeng Zhang, Zhuoran Yang, and Zhaoran Wang. Provably efficient offline reinforcement learning for partially observable markov decision processes. In International Conference on Machine Learning, pages 8016 8038. PMLR, 2022. [GDB16] Zhaohan Daniel Guo, Shayan Doroudi, and Emma Brunskill. A pac rl algorithm for episodic pomdps. In Artificial Intelligence and Statistics, pages 510 518. PMLR, 2016. [GMR22] Noah Golowich, Ankur Moitra, and Dhruv Rohatgi. Planning in observable pomdps in quasipolynomial time. ar Xiv preprint ar Xiv:2201.04735, 2022. [HDG15] Ahmed Hefny, Carlton Downey, and Geoffrey J Gordon. Supervised learning for dynamical system learning. Advances in neural information processing systems, 28:1963 1971, 2015. [HDL+21] Botao Hao, Yaqi Duan, Tor Lattimore, Csaba Szepesv ari, and Mengdi Wang. Sparse feature selection makes batch reinforcement learning more sample efficient. In International Conference on Machine Learning, pages 4063 4073. PMLR, 2021. [HKZ12] Daniel Hsu, Sham M Kakade, and Tong Zhang. A spectral algorithm for learning hidden markov models. Journal of Computer and System Sciences, 78(5):1460 1480, 2012. [Hor11] Joel L Horowitz. Applied nonparametric instrumental variables estimation. Econometrica, 79(2):347 394, 2011. [HW17] David A Hirshberg and Stefan Wager. Augmented minimax linear estimation. ar Xiv preprint ar Xiv:1712.00038, 2017. [HW21] Yuchen Hu and Stefan Wager. Off-policy evaluation in partially observed markov decision processes. ar Xiv preprint ar Xiv:2110.12343, 2021. [Jae00] Herbert Jaeger. Observable operator models for discrete stochastic time series. Neural computation, 12(6):1371 1398, 2000. [JKA+17] Nan Jiang, Akshay Krishnamurthy, Alekh Agarwal, John Langford, and Robert E Schapire. Contextual decision processes with low bellman rank are pac-learnable. In International Conference on Machine Learning, pages 1704 1713. PMLR, 2017. [JKKL20] Chi Jin, Sham M Kakade, Akshay Krishnamurthy, and Qinghua Liu. Sample-efficient reinforcement learning of undercomplete pomdps. ar Xiv preprint ar Xiv:2006.12484, 2020. [JL16] Nan Jiang and Lihong Li. Doubly robust off-policy value evaluation for reinforcement learning. In International Conference on Machine Learning, pages 652 661. PMLR, 2016. [JSLZ19] Binyan Jiang, Rui Song, Jialiang Li, and Donglin Zeng. Entropy learning for dynamic treatment regimes. Statistica Sinica, 29(4):1633, 2019. [KAL16] Akshay Krishnamurthy, Alekh Agarwal, and John Langford. Pac reinforcement learning with rich observations. Advances in Neural Information Processing Systems, 29, 2016. [KB14] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. [KECM21] Jeongyeol Kwon, Yonathan Efroni, Constantine Caramanis, and Shie Mannor. Rl for latent mdps: Regret guarantees and a lower bound. Advances in Neural Information Processing Systems, 34, 2021. [KHS+15] Predrag Klasnja, Eric B Hekler, Saul Shiffman, Audrey Boruvka, Daniel Almirall, Ambuj Tewari, and Susan A Murphy. Microrandomized trials: An experimental design for developing just-in-time adaptive interventions. Health Psychology, 34(S):1220, 2015. [KJS15] Alex Kulesza, Nan Jiang, and Satinder Singh. Spectral learning of predictive state representations with insufficient statistics. In Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015. [KL19] Michael R Kosorok and Eric B Laber. Precision medicine. Annual review of statistics and its application, 6:263 286, 2019. [KLC98] Leslie Pack Kaelbling, Michael L Littman, and Anthony R Cassandra. Planning and acting in partially observable stochastic domains. Artificial intelligence, 101(1-2):99 134, 1998. [KMU21] Nathan Kallus, Xiaojie Mao, and Masatoshi Uehara. Causal inference under unmeasured confounding with negative controls: A minimax learning approach. ar Xiv preprint ar Xiv:2103.14029, 2021. [KU20] Nathan Kallus and Masatoshi Uehara. Double reinforcement learning for efficient offpolicy evaluation in markov decision processes. J. Mach. Learn. Res., 21(167):1 63, 2020. [KU22] Nathan Kallus and Masatoshi Uehara. Efficiently breaking the curse of horizon in off-policy evaluation with double reinforcement learning. Operations Research, 2022. [KY22] Ali Devran Kara and Serdar Yuksel. Near optimality of finite memory feedback policies in partially observed markov decision processes. Journal of Machine Learning Research, 23(11):1 46, 2022. [KZ20] Nathan Kallus and Angela Zhou. Confounding-robust policy evaluation in infinitehorizon reinforcement learning. ar Xiv preprint ar Xiv:2002.04518, 2020. [LAHA21] Sahin Lale, Kamyar Azizzadenesheli, Babak Hassibi, and Anima Anandkumar. Adaptive control and regret minimization in linear quadratic gaussian (lqg) setting. In 2021 American Control Conference (ACC), pages 2517 2522. IEEE, 2021. [LCSJ22] Qinghua Liu, Alan Chung, Csaba Szepesv ari, and Chi Jin. When is partially observable reinforcement learning not scary? ar Xiv preprint ar Xiv:2204.08967, 2022. [LFY+21] Luofeng Liao, Zuyue Fu, Zhuoran Yang, Yixin Wang, Mladen Kolar, and Zhaoran Wang. Instrumental variable value iteration for causal offline reinforcement learning. ar Xiv preprint ar Xiv:2102.09907, 2021. [LKTF20] Sergey Levine, Aviral Kumar, George Tucker, and Justin Fu. Offline reinforcement learning: Tutorial, review, and perspectives on open problems. ar Xiv preprint ar Xiv:2005.01643, 2020. [LLTZ18] Qiang Liu, Lihong Li, Ziyang Tang, and Dengyong Zhou. Breaking the curse of horizon: Infinite-horizon off-policy estimation. In Advances in Neural Information Processing Systems 31, pages 5356 5366. 2018. [LMWY22] Miao Lu, Yifei Min, Zhaoran Wang, and Zhuoran Yang. Pessimism in the face of confounders: Provably efficient offline reinforcement learning in partially observable markov decision processes. ar Xiv preprint ar Xiv:2205.13589, 2022. [LNSJ22] Qinghua Liu, Praneeth Netrapalli, Csaba Szepesvari, and Chi Jin. Optimistic mle a generic model-based algorithm for partially observable sequential decision making. ar Xiv preprint ar Xiv:2209.14997, 2022. [LP03] Michail G Lagoudakis and Ronald Parr. Least-squares policy iteration. The Journal of Machine Learning Research, 4:1107 1149, 2003. [LQM20] Peng Liao, Zhengling Qi, and Susan Murphy. Batch policy learning in average reward markov decision processes. ar Xiv preprint ar Xiv:2007.11771, 2020. [LS01] Michael Littman and Richard S Sutton. Predictive representations of state. Advances in neural information processing systems, 14, 2001. [LVY19] Hoang Le, Cameron Voloshin, and Yisong Yue. Batch policy learning under constraints. In International Conference on Machine Learning, pages 3703 3712. PMLR, 2019. [MGT18] Wang Miao, Zhi Geng, and Eric J Tchetgen Tchetgen. Identifying causal effects with proxy variables of an unmeasured confounder. Biometrika, 105(4):987 993, 2018. [MKS+13] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. Playing atari with deep reinforcement learning. ar Xiv preprint ar Xiv:1312.5602, 2013. [MS08] R emi Munos and Csaba Szepesv ari. Finite-time bounds for fitted value iteration. Journal of Machine Learning Research, 9(5), 2008. [MST18] Wang Miao, Xu Shi, and Eric Tchetgen Tchetgen. A confounding bridge approach for double negative control inference on causal effects. ar Xiv preprint ar Xiv:1808.04945, 2018. [Mvd LRG01] Susan A Murphy, Mark J van der Laan, James M Robins, and Conduct Problems Prevention Research Group. Marginal mean models for dynamic regimes. Journal of the American Statistical Association, 96(456):1410 1423, 2001. [MY20] Curtis Mc Donald and Serdar Y uksel. Exponential filter stability via dobrushin s coefficient. Electronic Communications in Probability, 25:1 13, 2020. [NCDL19] Ofir Nachum, Yinlam Chow, Bo Dai, and Lihong Li. Dualdice: Behavioragnostic estimation of discounted stationary distribution corrections. ar Xiv preprint ar Xiv:1906.04733, 2019. [New13] Whitney K Newey. Nonparametric instrumental variables estimation. American Economic Review, 103(3):550 56, 2013. [NJ21] Yash Nair and Nan Jiang. A spectral approach to off-policy evaluation for pomdps. ar Xiv preprint ar Xiv:2109.10502, 2021. [NKYB20] Hongseok Namkoong, Ramtin Keramati, Steve Yadlowsky, and Emma Brunskill. Off-policy policy evaluation for sequential decisions under unobserved confounding. ar Xiv preprint ar Xiv:2003.05623, 2020. [PGT06] Joelle Pineau, Geoffrey Gordon, and Sebastian Thrun. Anytime point-based approximations for large pomdps. Journal of Artificial Intelligence Research, 27:335 380, 2006. [Pre00] Doina Precup. Eligibility traces for off-policy policy evaluation. Computer Science Department Faculty Publication Series, page 80, 2000. [PT87] Christos H Papadimitriou and John N Tsitsiklis. The complexity of markov decision processes. Mathematics of operations research, 12(3):441 450, 1987. [RGT04] Matthew Rosencrantz, Geoff Gordon, and Sebastian Thrun. Learning low dimensional predictive representations. In Proceedings of the twenty-first international conference on Machine learning, page 88, 2004. [RRS98] Andrea Rotnitzky, James M Robins, and Daniel O Scharfstein. Semiparametric regression for repeated outcomes with nonignorable nonresponse. Journal of the american statistical association, 93(444):1321 1339, 1998. [SBS+10] Le Song, Byron Boots, Sajid M Siddiqi, Geoffrey Gordon, and Alex Smola. Hilbert space embeddings of hidden markov models. In Proceedings of the 27th International Conference on International Conference on Machine Learning, pages 991 998, 2010. [SHSF09] Le Song, Jonathan Huang, Alex Smola, and Kenji Fukumizu. Hilbert space embeddings of conditional distributions with applications to dynamical systems. In Proceedings of the 26th Annual International Conference on Machine Learning, pages 961 968, 2009. [SJR04] Satinder Singh, Michael James, and Matthew Rudary. Predictive state representations: A new theory for modeling dynamical systems. In Proceedings of the 20th Conference on Uncertainty in Artificial Intelligence, 2004. [SMNTT20] Xu Shi, Wang Miao, Jennifer C Nelson, and Eric J Tchetgen Tchetgen. Multiply robust causal inference with double-negative control adjustment for categorical unmeasured confounding. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 82(2):521 540, 2020. [SRRH21] Adish Singla, Anna N Rafferty, Goran Radanovic, and Neil T Heffernan. Reinforcement learning for education: Opportunities and challenges. ar Xiv preprint ar Xiv:2107.08828, 2021. [SSH20] Max Simchowitz, Karan Singh, and Elad Hazan. Improper learning for non-stochastic control. In Conference on Learning Theory, pages 3320 3436. PMLR, 2020. [SUHJ22] Chengchun Shi, Masatoshi Uehara, Jiawei Huang, and Nan Jiang. A minimax learning approach to off-policy evaluation in confounded partially observable markov decision processes. In International Conference on Machine Learning, pages 20057 20094. PMLR, 2022. [SZLS20] Chengchun Shi, Sheng Zhang, Wenbin Lu, and Rui Song. Statistical inference of the value function for reinforcement learning in infinite horizon settings. ar Xiv preprint ar Xiv:2001.04515, 2020. [SZY+22] Chengchun Shi, Jin Zhu, Shen Ye, Shikai Luo, Hongtu Zhu, and Rui Song. Off-policy confidence interval estimation with confounded markov decision process. Journal of the American Statistical Association, pages 1 12, 2022. [TB16] Philip Thomas and Emma Brunskill. Data-efficient off-policy policy evaluation for reinforcement learning. In International Conference on Machine Learning, pages 2139 2148. PMLR, 2016. [TDHL19] Anastasios A Tsiatis, Marie Davidian, Shannon T Holloway, and Eric B Laber. Dynamic Treatment Regimes: Statistical Methods for Precision Medicine. CRC press, 2019. [TFL+20] Ziyang Tang, Yihao Feng, Lihong Li, Dengyong Zhou, and Qiang Liu. Harnessing infinitehorizon off-policy evaluation: Double robustness via duality. ICLR 2020, 2020. [TSM20] Guy Tennenholtz, Uri Shalit, and Shie Mannor. Off-policy evaluation in partially observable environments. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 10276 10283, 2020. [UHJ20] Masatoshi Uehara, Jiawei Huang, and Nan Jiang. Minimax weight and q-function learning for off-policy evaluation. In International Conference on Machine Learning, pages 9659 9668. PMLR, 2020. [UIJ+21] Masatoshi Uehara, Masaaki Imaizumi, Nan Jiang, Nathan Kallus, Wen Sun, and Tengyang Xie. Finite sample analysis of minimax offline reinforcement learning: Completeness, fast rates and first-order efficiency. ar Xiv preprint ar Xiv:2102.02981, 2021. [US21] Masatoshi Uehara and Wen Sun. Pessimistic model-based offline reinforcement learning under partial coverage. ar Xiv preprint ar Xiv:2107.06226, 2021. [USL+22] Masatoshi Uehara, Ayush Sekhari, Jason D Lee, Nathan Kallus, and Wen Sun. Provably efficient reinforcement learning in partially observable dynamical systems. ar Xiv preprint ar Xiv:2206.12020, 2022. [VHGS16] Hado Van Hasselt, Arthur Guez, and David Silver. Deep reinforcement learning with double q-learning. In Proceedings of the AAAI conference on artificial intelligence, volume 30, 2016. [VSH+16] Arun Venkatraman, Wen Sun, Martial Hebert, J Bagnell, and Byron Boots. Online instrumental variable regression with applications to online linear system identification. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 30, 2016. [WYW20] Lingxiao Wang, Zhuoran Yang, and Zhaoran Wang. Provably efficient causal reinforcement learning with confounded observational data. ar Xiv preprint ar Xiv:2006.12311, 2020. [XCJ+21] Tengyang Xie, Ching-An Cheng, Nan Jiang, Paul Mineiro, and Alekh Agarwal. Bellman-consistent pessimism for offline reinforcement learning. Advances in neural information processing systems, 34, 2021. [XLSS20] Zekun Xu, Eric Laber, Ana-Maria Staicu, and Emanuel Severus. Latent-state models for precision medicine. ar Xiv preprint ar Xiv:2005.13001, 2020. [XZS+22] Yang Xu, Jin Zhu, Chengchun Shi, Shikai Luo, and Rui Song. An instrumental variable approach to confounded off-policy evaluation. ar Xiv preprint ar Xiv:2212.14468, 2022. [ZB16] Junzhe Zhang and Elias Bareinboim. Markov decision processes with unobserved confounders: A causal approach. Technical report, Technical Report R-23, Purdue AI Lab, 2016. [ZLKB20] Andrea Zanette, Alessandro Lazaric, Mykel Kochenderfer, and Emma Brunskill. Learning near optimal policies with low inherent bellman error. In International Conference on Machine Learning, pages 10978 10989. PMLR, 2020. [ZUSL22] Wenhao Zhan, Masatoshi Uehara, Wen Sun, and Jason D Lee. Pac reinforcement learning for predictive state representations. ar Xiv preprint ar Xiv:2207.05738, 2022. [ZW22] Andrea Zanette and Martin J Wainwright. Bellman residual orthogonalization for offline reinforcement learning. ar Xiv preprint ar Xiv:2203.12786, 2022. [ZWB21] Andrea Zanette, Martin J Wainwright, and Emma Brunskill. Provable benefits of actor-critic methods for offline reinforcement learning. Advances in neural information processing systems, 34, 2021. [ZXZ+22] Han Zhong, Wei Xiong, Sirui Zheng, Liwei Wang, Zhaoran Wang, Zhuoran Yang, and Tong Zhang. A posterior sampling framework for interactive decision making. ar Xiv preprint ar Xiv:2211.01962, 2022.