# parameterized_projected_bellman_operator__b83df8a1.pdf Parameterized Projected Bellman Operator Th eo Vincent12*, Alberto Maria Metelli3, Boris Belousov1, Jan Peters1245, Marcello Restelli3, Carlo D Eramo246 1German Research Center for AI (DFKI), Germany 2Department of Computer Science, TU Darmstadt, Darmstadt, Germany 3Department of Electronics, Computer Science, and Bioengineering, Politecnico di Milano, Milano, Italy 4Hessian.ai, Germany 5Centre for Cognitive Science, TU Darmstadt, Darmstadt, Germany 6Center for Artificial Intelligence and Data Science, University of W urzburg, W urzburg, Germany Approximate value iteration (AVI) is a family of algorithms for reinforcement learning (RL) that aims to obtain an approximation of the optimal value function. Generally, AVI algorithms implement an iterated procedure where each step consists of (i) an application of the Bellman operator and (ii) a projection step into a considered function space. Notoriously, the Bellman operator leverages transition samples, which strongly determine its behavior, as uninformative samples can result in negligible updates or long detours, whose detrimental effects are further exacerbated by the computationally intensive projection step. To address these issues, we propose a novel alternative approach based on learning an approximate version of the Bellman operator rather than estimating it through samples as in AVI approaches. This way, we are able to (i) generalize across transition samples and (ii) avoid the computationally intensive projection step. For this reason, we call our novel operator projected Bellman operator (PBO). We formulate an optimization problem to learn PBO for generic sequential decision-making problems, and we theoretically analyze its properties in two representative classes of RL problems. Furthermore, we theoretically study our approach under the lens of AVI and devise algorithmic implementations to learn PBO in offline and online settings by leveraging neural network parameterizations. Finally, we empirically showcase the benefits of PBO w.r.t. the regular Bellman operator on several RL problems. Introduction Value-based reinforcement learning (RL) is a popular class of algorithms for solving sequential decision-making problems with unknown dynamics (Sutton and Barto 2018). For a given problem, value-based algorithms aim at obtaining the most accurate estimate of the expected return from each state, i.e., a value function. For instance, the well-known value-iteration algorithm computes value functions by iterated applications of the Bellman operator (Bellman 1966), of which the true value function is the fixed point. Although the Bellman operator can be applied in an exact way in dynamic programming, it is typically estimated from samples at each application when dealing with problems with unknown models, i.e., empirical Bellman operator (Watkins *Correspondence to: theo.vincent@dfki.de Copyright 2024, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 1989; Bertsekas 2019). Intuitively, the dependence of the empirical version of value iteration on the samples has an impact on the efficiency of the algorithms and on the quality of the obtained estimated value function, which becomes accentuated when solving continuous problems that require function approximation, e.g., approximate value iteration (AVI) (Munos 2005; Munos and Szepesv ari 2008). Moreover, in AVI approaches, costly function approximation steps are needed to project the output of the Bellman operator back to the considered value function space. In this paper, we introduce the projected Bellman operator (PBO), which consists of a function Λ : Ω Ωdefined on the parameters ω Ωof the value function approximator Qω. Contrary to the standard (empirical) Bellman operator Γ, which acts on the value functions Qk to compute targets that are then projected to obtain Qk+1, our PBO Λ acts on the parameters of the value function to directly compute updated parameters ωk+1 = Λ(ωk) (Figure 1). The advantages of our approach are twofold: (i) PBO is applicable for an arbitrary number of iterations without using further samples, and (ii) the output of PBO always belongs to the considered value function space as visualized in Figure 2, thus avoiding the costly projection step which is required when using the Bellman operator coupled with function approximation. We show how to estimate PBO from transition samples by leveraging a parametric approximation which we call parameterized PBO, and we devise two algorithms for offline and online RL to learn it. Starting from initial parameters ω0, AVI approaches obtain consecutive approximations of the value function Qωk by applying the Bellman operator iteratively over samples (Figure 2b). Instead, we make use of the samples to learn the PBO only. Then, starting from initial parameters ω0, PBO can produce a chain of value function parameters of arbitrary length (as shown with the blue lines in Figure 2a) without requiring further samples. This means an Figure 1: PBO Λ (left) operates on value function parameters as opposed to AVI (right), that uses the empirical Bellman operator Γ followed by the projection operator Πproj. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) (a) Projected Bellman Operator (PBO) (ours) (b) Approximate Value Iteration (AVI) Figure 2: Behavior of our PBO and AVI in the parametric space of value functions Q. Here Q , Qω , and QΛ , are respectively the optimal value function, its projection on the parametric space, and the fixed point of PBO. Contrary to the regular Bellman operator, PBO can be applied for an arbitrary number of steps (blue lines) without requiring additional samples. accurate approximation of PBO can compute optimal value function parameters starting from any initial parameters in the chosen space without requiring additional samples. Related Work Several papers in the literature proposed variants of the standard Bellman operator to induce certain behaviors. We review these approaches, noting that they all act on the space of action-value functions, thus needing a costly projection step onto the considered function space. Conversely, to the best of our knowledge, our work is the first attempt to obtain an alternative Bellman operator that avoids the projection step by directly acting on the parameters of value functions. Bellman operator variations. Variants of the Bellman operator are widely studied for entropy-regularized MDPs (Neu, Jonsson, and G omez 2017; Geist, Scherrer, and Pietquin 2019; Belousov and Peters 2019). The softmax (Haarnoja et al. 2017; Song, Parr, and Carin 2019), mellowmax (Asadi and Littman 2017), and optimistic (Tosatto et al. 2019) operators are all examples of variants of the Bellman operator to obtain maximum-entropy exploratory policies. Besides favoring exploration, other approaches address the limitations of the standard Bellman operator. For instance, the consistent Bellman operator (Bellemare et al. 2016) is a modified operator that addresses the problem of inconsistency of the optimal action-value functions for suboptimal actions. The distributional Bellman operator (Bellemare, Dabney, and Munos 2017) enables operating on the whole return distribution instead of its expectation, i.e., the value function (Bellemare, Dabney, and Rowland 2023). Furthermore, the logistic Bellman operator uses a logistic loss to solve a convex linear programming problem to find optimal value functions (Bas-Serrano et al. 2021). Finally, the Bayesian Bellman operator is employed to infer a posterior over Bellman operators centered on the true one (Fellows, Hartikainen, and Whiteson 2021). Note that our PBO can be seamlessly applied to any of these variants of the standard Bellman operator. Finally, we recognize that learning an approximation of the Bellman operator shares some similarities with learning a reward-transition model in reinforcement learning. However, we point out that our approach is profoundly different, as we map action-value parameters to other action-value parameters, in contrast to model-based reinforcement learning, which maps states and actions to rewards and next states. Operator learning. Literature in operator learning is mostly focused on supervised learning, with methods for learning operators over vector spaces (Micchelli and Pontil 2005) and parametric approaches for learning non-linear operators (Chen and Chen 1995), with a resurgence of recent contributions in deep learning. For example, Kovachki et al. (2021, 2023) learn mappings between infinite function spaces with deep neural networks, or Kissas et al. (2022) apply an attention mechanism to learn correlations in the target function for efficient operator learning. We note that our work on the learning of the Bellman operator in reinforcement learning is orthogonal to methods for operator learning in supervised learning, and could potentially benefit from advanced techniques in the literature. Hyper Networks. Our approach shares similarities with Hyper Networks (Ha, Dai, and Le 2016; Sarafian, Keynan, and Kraus 2021; Beck et al. 2023). In the classification proposed in Chauhan et al. (2023), PBO distinguishes itself by taking as input the parameters generated at the previous forward pass. In this sense, PBO can be seen as a recurrent variant of Hyper Networks that generates a series of parameters representing the consecutive Bellman iterations. Preliminaries We consider discounted Markov decision processes (MDPs) defined as M = S, A, P, R, γ , where S is a measurable state space, A is a finite, measurable action space, P : S A (S)1 is the transition kernel of the dynamics of the system, R : S A R is a reward function, and γ [0, 1) is a discount factor (Puterman 1990). A policy π : S A is a function mapping a state to an action, inducing a value function V π(s) E h P+ t=0 γt R(St, π(St))|S0 = s i representing the expected cumulative discounted reward starting in state s and following policy π thereafter. Similarly, the action-value function Qπ(s, a) E h P+ t=0 γt R(St, At)|S0 = s, A0 = a, At = π(St) i is the expected discounted cumulative reward executing action a in state s, following policy π thereafter. RL aims to find 1 (X) denotes the set of probability measures over the set X. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) an optimal policy π yielding the optimal value function V (s) maxπ:S A V π(s) for every s S (Puterman 1990). The (optimal) Bellman operator Γ is a fundamental tool in RL for obtaining optimal policies, defined as: (Γ Q)(s, a) R(s, a) + γ Z S P(ds |s, a) max a A Q(s , a ), (1) for all (s, a) S A. It is well-known that Bellman operators are γ-contraction mappings in L -norm, such that their iterative application leads to the unique fixed point Γ Q = Q in the limit (Bertsekas 2015). We consider using function approximation to represent value functions and denote Ωas the space of their parameters. Thus, we define QΩ= {Qω : S A R|ω Ω} as the set of value functions representable by parameters of Ω. Projected Bellman Operator The application of the Bellman operator in RL requires transition samples (Equation 1) and a costly projection step onto the used function space (Munos 2003, 2005; Munos and Szepesv ari 2008). We are interested in obtaining an operator that overcomes these issues while emulating the behavior of the regular Bellman operator. Hence, we introduce the projected Bellman operator (PBO), which we define as follows. Definition 1. Let QΩ= {Qω : S A R|ω Ω} be a function approximation space for the action-value function, induced by the parameter space Ω. A projected Bellman operator (PBO) is a function Λ : Ω Ω, such that Λ arg min Λ:Ω Ω E(s,a) ρ,ω ν Γ Qω(s, a) QΛ(ω)(s, a) 2 , (2) for state-action and parameter distributions ρ and ν. Note that Γ is the regular optimal Bellman operator on QΩ. Conversely, PBO is an operator Λ acting on actionvalue function parameters ω Ω. In other words, this definition states that a PBO is the mapping Ω Ωthat most closely emulates the behavior of the regular optimal Bellman operator Γ . Thus, by acting on the parameters ω of action-value functions Qω, PBO can be applied for an arbitrary number of steps starting from any initial parameterizations ω0, without using additional transition samples. Moreover, being a mapping between parameters ωk to parameters ωk+1, PBO does not require the costly projection step needed by the regular Bellman operator. Learning Projected Bellman Operators The PBO is unknown and has to be estimated from samples. We propose to approximate PBO with a parameterized PBO Λϕ differentiable w.r.t. its parameters ϕ Φ, enabling the use of gradient descent on a loss function.2 We do so by formulating an empirical version of the problem (2) that can be optimized via gradient descent by using a given dataset of parameters ω W ν and transitions (s, a, r, s ) D ρ. Crucially, we can leverage PBO to augment the dataset of 2For ease of presentation, we use PBO and parameterized PBO interchangeably whenever clear from the context. Algorithm 1: Projected FQI & *Projected DQN* 1: Inputs: - samples D = { sj, aj, rj, s j }J j=1; - parameters W = {ωl}L l=1; - #Bellman iterations K; - initial parameters ϕ of parameterized PBO Λϕ; - #Epochs E. 2: for e {1, . . . , E} do 3: ϕ = ϕ 4: for some training steps do 5: *Collect samples D with policy given by QΛK ϕ (ω)* 6: *D D D * 7: Gradient descent over parameters ϕ minimizing loss (3) or (4) using D *or a batch of D* and W. 8: end for 9: end for 10: Return: Parameters ϕ of parameterized PBO Λϕ parameters ω W with sequences generated by PBO at no cost of additional samples. The resulting loss is (s,a) D ω W ΓQΛk 1 ϕ (ω)(s, a) QΛk ϕ(ω)(s, a) 2 , (3) where K is an arbitrary number of optimal Bellman operator iterations. Note that for K = 1 we obtain the empirical version of the optimization problem (2). An additional idea is to add a term that corresponds to an infinite number of iterations, i.e., the fixed point, LΛ ϕ = LΛϕ + X ΓQΛ ϕ (s, a) QΛ ϕ (s, a) 2 | {z } Fixed point term where Λ ϕ is the fixed point of the parameterized PBO. Note that the addition of the fixed-point term is only possible for classes of parameterized PBOs where the fixed point can be computed, and the result is differentiable w.r.t. the parameters ϕ, as we describe in the following section. We can now devise two algorithms to learn PBO in both offline and online RL. Our algorithms can be seen as variants of the offline algorithm fitted Q-iteration (FQI) (Ernst, Geurts, and Wehenkel 2005) and online algorithm deep Qnetwork (DQN) (Mnih et al. 2015); thus, we call them projected fitted Q-iteration (Pro FQI) and projected deep Qnetwork (Pro DQN). Algorithm 1 compactly describes both Pro FQI and Pro DQN, highlighting the additional steps required for the online setting (i.e., Pro DQN). Both Pro FQI and Pro DQN are given initial randomly sampled datasets of transitions and parameters of PBO. As an online algorithm, Pro DQN periodically adds new transitions to the dataset by executing a policy derived from the action-value function obtained by applying the current approximation of PBO for K times. For stability reasons, we perform gradient descent steps only on the parameters ϕ of QΛk ϕ in the loss The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) (a) Pro FQI (ours) Figure 3: Behavior of Pro FQI and FQI in the function space QΩfor one iteration. The ability to apply PBO for an arbitrary number of times enables Pro FQI to generate a sequence of action-value functions QΛk ϕ(ω0) that can be used to enrich the loss function to learn PBO (see red lines). On the contrary, one iteration of FQI corresponds to a single application of the Bellman operator followed by the projection step onto the function space. (Equation 3), excluding the ones corresponding to the target ΓQΛk 1 ϕ , as commonly done in semi-gradient methods (Sutton and Barto 2018). Similar to Mnih et al. (2015), the target parameters are updated periodically after an arbitrary number of iterations. Soft updates, e.g., Polyak averaging (Lillicrap et al. 2015), can also be used. As also illustrated by Figure 3, Pro FQI and FQI behave substantially differently in the space of action-value functions QΩ. Pro FQI aims to learn PBO to subsequently use it to generate a sequence (Qω0, QΛϕ(ω0), QΛ2 ϕ(ω0), . . . ) of actionvalue parameters starting from any parameters ω0; on the contrary, FQI generates a sequence of action-value functions (Qω0, Qω1, . . . ) that need to be projected onto the function space at every iteration. Moreover, Pro FQI can apply PBO multiple times to form a richer loss than the one for FQI, which can only consider one application of the Bellman operator at a time (see red lines in Figure 3a and 3b). Analysis of Projected Bellman Operator Besides the practical advantages of being independent of samples and not needing a projection step, PBO plays an interesting role when investigated theoretically in the AVI framework. In particular, the ability of PBO to iterate for an arbitrary number of times enables us to theoretically prove its benefit in terms of approximation error at a given timestep K by leveraging the following established results in AVI. Theorem 2. (See Theorem 3.4 of Farahmand (2011)) Let K N , ρ, ν two distribution probabilities over S A. For any sequence (Qk)K k=0 B (S A, Rγ) where Rγ depends on reward function and discount factor, we have Q QπK 1,ρ CK,γ,Rγ (5) + inf r [0,1] F(r; K, ρ, γ) P ro F QI z }| { K X k=1 α2r k Γ Qk 1 Qk 2 2,ν | {z } F QI where αk and CK,γ,Rγ do not depend on the sequence (Qk)K k=0. F(r; K, ρ, γ) relies on the concentrability coefficients of the greedy policies w.r.t. the value functions. This theorem shows that the distance between the value function and the optimal value function (left-hand side of Equation (5)) is upper-bounded by a quantity proportional to the approximation errors Γ Qk 1 Qk 2 2,ν at each iteration k. As highlighted in Equation (5), at every iteration k, the loss of FQI contains only one term of the sum. Conversely, the loss of Pro FQI contains the entire sum of approximation errors from iteration k = 1 to K. This means that while FQI minimizes the approximation error for a single iteration at a time, Pro FQI minimizes the approximation errors for multiple iterations, thus effectively reducing the upper bound on the approximation error at iteration K. In the following, we empirically showcase this theoretical advantage and we further analyze the properties of PBO for two representative classes of RL problems, namely (i) finite MDPs and (ii) linear quadratic regulators (Bradtke 1992; Pang and Jiang 2021). Proofs of the following results and additional analysis of PBO in low-rank MDPs (Agarwal et al. 2020; Sekhari et al. 2021) can be found in the appendix. Finite Markov Decision Processes Let us consider finite state and action spaces of cardinality N and M, respectively, and a tabular setting with Ω= RN M. Given the use of tabular approximation, it is intuitive that each entry of the table can be modeled with a different single parameter, i.e., there is a bijection between QΩand Ω, which allows us to write, for ease of notation, the parameters of the action-value function as Q instead of ω. Proposition 3. The PBO exists, and it is equal to the optimal Bellman operator Λ (Q) = R + γP max a A Q( , a). (6) Note that PBO for finite MDPs is a γ-contraction mapping for the L -norm, like an optimal Bellman operator. As an example of finite MDP, we consider the chain-walk environment in Figure 5, with a chain of length N = 20. We parameterize value functions as tables to leverage our theoretical results on finite MDPs. In Figure 4, we show the ℓ2-norm of the difference between the optimal action-value function and the action-value function computed with FQI The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) FQI Pro FQI Pro FQI Pro FQIchain PBO 0 6 ||Q * Qk||2 k = 2 k = 6 k = 11 (a) K = 2. FQI Pro FQI Pro FQI Pro FQIchain PBO 0 6 ||Q * Qk||2 k = 5 k = 9 k = 14 (b) K = 5. FQI Pro FQI Pro FQI Pro FQIchain PBO 0 6 ||Q * Qk||2 k = 15 k = 19 k = 24 (c) K = 15. Figure 4: ℓ2-norm of the difference between the optimal action-value function and the approximated action-value function on chain-walk. Here K is the number of iterations included in the training loss (3) of PBO, while k is the number of PBO applications after training. Results are averaged over 20 seeds. Note that PBO enables using k K compared to FQI, which results in better convergence for increasing k for each fixed K {2, 5, 15}. Figure 5: Test task: chain-walk from Munos (2003). The reward is 0 in all states except green states where it equals 1. and Pro FQI. We consider three different values of Bellman iterations, namely K {2, 5, 15}. For FQI, the K iterations are the regular iterations where the empirical Bellman operator is applied to the current approximation of the actionvalue function. For Pro FQI, K is the number of iterations included in the PBO loss. This ensures a fair comparison, given that both methods have access to the same number of Bellman iterations. Once PBO is trained, we apply it for different numbers of iterations k K, on given actionvalue function parameters. In Figure 4, Pro FQI uses a linear approximation of PBO trained with the loss (3), Pro FQI uses a linear approximation of PBO trained with the loss (4), Pro FQIchain uses the closed-form PBO in Equation (6) considering R and P as unknown parameters to learn, and PBO indicates the use of the same closed-form solution assuming known R and P. For the three variants of Pro FQI, we observe that the approximation error decreases as the number of PBO iterations increases, evincing that PBO accurately emulates the true Bellman operator. In the case of K = 2 and K = 5, we see that Pro FQI and Pro FQIchain obtain a better approximation of the action-value function compared to Pro FQI, thanks to, respectively, the inclusion of fixed point in the loss and the use of the closed-form solution. Interestingly, in the case of K = 15, Pro FQI obtains a slightly worse approximation than Pro FQI. We explain this behavior with the fact that when the linear approximation is inadequate for modeling PBO, adding the fixed point in the loss could harm the estimate. Linear Quadratic Regulation Now, we consider the continuous MDPs class of linear quadratic regulator (LQR) with S = A = R. The transition model P(s, a) = As + Ba is deterministic and the reward function R(s, a) = Qs2 + 2Ssa + Ra2 is quadratic, where A, B, Q, S and R, are constants inherent to the MDP. We choose to parameterize the action-value functions with a 2-dimensional parameter vector QΩ= {(s, a) 7 Gs2 + 2Isa + Ma2|(G, I) R2} where M is a chosen constant, for visualization purposes. Proposition 4. PBO exists, and for any ω R2, its closed form is given by:3 Λ : ω = G I " Q + A2(G I2 M ) S + AB(G I2 We leverage our theoretical analysis for LQR (Bradtke 1992; Pang and Jiang 2021) by parameterizing value functions accordingly, and we conduct a similar analysis to the one for chain-walk. This time, we evaluate the distance between the optimal action-value function parameters, which can be computed analytically, and the estimated ones. We use the closed-form solution obtained in Equation 7 assuming the parameters (A, B, Q, S) known (PBO), and unknown (Pro FQILQR). Figure 6 confirms the pattern observed on the chain-walk. Pro FQI and Pro FQI obtain a better approximation than FQI, which is reasonably worse than Pro FQILQR and PBO. We also observe that Pro FQILQR obtains a significantly better approximation than the other variants for a large number of iterations (blue bars), confirming the advantages of exploiting the closed-form solution of PBO for finite MDPs. We additionally show the sequence of parameters corresponding to the iterations of FQI, Pro FQILQR, and Pro FQI, in Figure 7. The training is done with K = 2. The sequence starts from the chosen initial parameters (G, I) = (0, 0) for all algorithms and proceeds towards the optimal parameters, which are computed analytically. Both Pro FQI and Pro FQILQR apply the PBO learned 3Under a mild assumption over the sample distribution, see in the Proofs section in the appendix. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) FQI Pro FQI Pro FQI Pro FQILQR PBO 10 7 k = 2 k = 5 k = 8 (a) K = 2. FQI Pro FQI Pro FQI Pro FQILQR PBO 10 7 k = 3 k = 6 k = 9 (b) K = 3. FQI Pro FQI Pro FQI Pro FQILQR PBO 10 7 k = 4 k = 7 k = 10 (c) K = 4. Figure 6: ℓ2-norm of the difference between the optimal action-value function parameters and the estimated ones on LQR. Here K is the number of iterations included in the training loss (3) of PBO, while k is the number of PBO applications after training. Results are averaged over 20 seeds. Similar to Figure 4, PBO shows improved convergence for k K compared to FQI. Figure 7: Behavior of FQI and Pro FQI in the space of actionvalue function parameters (G, I) for LQR. FQI is performed with 2 iterations, and Pro FQI uses a PBO trained with K = 2, for the sake of fair comparison. Notably, Pro FQI can leverage PBO by applying it for 8 iterations, being able to get closer to the optimal parameters (black star) than FQI. after the K = 2 iterations, for 8 iterations; thus, the sequence of parameters for both algorithms is composed of 8 points each, while FQI has 2. It is clear that the parameters found by FQI are considerably further to the target parameters than the ones found by both Pro FQI and Pro FQILQR. In particular, the latter gets the closest to the target, in line with the results in Figure 6, again evincing the accuracy of the learned PBO and the benefit of performing multiple applications of it, as expected from Theorem 2. Experiments We consider both Pro FQI and Pro DQN, comparing their performance with their regular counterparts4. To handle the increased complexity of the input space of the considered problems, we leverage neural network regression to model our PBO. We consider an offline setting, where we use Pro FQI on car-on-hill (Ernst, Geurts, and Wehenkel 2005), and an online setting, where we use Pro DQN on bicycle balancing (Randlov and Alstrøm 1998), and lunar lander (Brockman et al. 2016). We want to answer the following research question: does PBO enable moving toward the fixed point more effectively than the empirical Bellman operator? 4The code is available at https://github.com/theovincent/PBO (a) Pro FQI (ours). Figure 8: Policies on car-on-hill using K = 9 and 9 application of PBO, averaged over 20 seeds. The purple and green colors refer to the two discrete actions (move left or right). Our Pro FQI approximates the optimal policy more closely. Projected Fitted Q-Iteration We initially evaluate Pro FQI on the car-on-hill problem. As done in Ernst, Geurts, and Wehenkel (2005), we measure performance by generating roll-outs starting from different states on a grid of size 17 17, and accounting for the fact that the dataset D does not contain every starting state of the grid, by weighting the obtained performance from each starting state by the number of times it occurs in the dataset (see appendix). First, the benefit of PBO is observable in the quality of the policy computed by FQI and Pro FQI (Figure 8). After only 9 training iterations, Pro FQI obtains a policy that is very close to the optimal one of car-on-hill (see the well-known shape of the optimal car-on-hill policy in Figure 8a (Ernst, Geurts, and Wehenkel 2005)), while FQI is significantly more inaccurate. The consequences of this are reflected in the performance shown in Figure 9, that are obtained with FQI and Pro FQI for three different values of Bellman iterations K (black vertical line). Again, for FQI, K is the number of regular iterations consisting of an application of the empirical Bellman operator and the projection step; for Pro FQI, K is the number of applications of PBO that are used in the training loss (Equation 3). The iterations on the x-axis in Figure 9 are the regular iterations for FQI and the applications of the trained PBO for Pro FQI. Thus, The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Average return Average return Average return (c) K = 15. Figure 9: Discounted cumulative reward on car-on-hill. The black vertical line indicates the number of iterations considered for FQI and the number of iterations K considered to train PBO. Results are averaged over 20 seeds with 95% confidence intervals. Note that Pro FQI achieves higher performance than FQI and remains stable after K iterations. Average return (a) Bicycle balancing. Average return (b) Lunar lander. Figure 10: Discounted cumulative reward on bicycle and lunar lander. The black vertical line indicates the number of target network updates performed by DQN and the number of iterations K considered to train PBO. Results are averaged over 40 seeds with 95% confidence intervals. the iterations on the right side of the line can only be reached with Pro FQI. The purpose of this analysis is to evaluate the different quality of trajectories toward the fixed point obtained by the empirical Bellman operator in FQI, and the trained PBO in Pro FQI, for the same amount of transition samples. We observe that Pro FQI obtains better performance than FQI consistently. This evinces that PBO learns an accurate estimate of the true Bellman operator, which allows obtaining a satisfactory action-value function faster than the standard empirical Bellman operator used by FQI. It is interesting to note that the performance of Pro FQI remains stable for subsequent applications of PBO after the number of iterations used for training. Projected Deep Q-Network We also evaluate PBO in an online setting, using our Pro DQN algorithm and comparing against DQN (Mnih et al. 2015). We consider a bicycle balancing (Randlov and Alstrøm 1998) problem and the lunar lander environment (Brockman et al. 2016). We set the number of Bellman iterations K = 8 for bicycle balancing and K = 10 for the lunar lander. Similar to the offline setting, K is the number of updates of the target network for DQN, and the number of applications of PBO in the training loss (Equation 3) for Pro DQN. Due to the need to collect new samples while learning a policy, training PBO in an online setting needs a slightly different treatment than the offline case. Recalling Algorithm 1, we point out that new samples are collected after each gradient descent step, by using the action-value function obtained after K applications of the current PBO. We find this choice to work well in practice; however, we can envision multiple possibilities for effective exploration strategies based on PBO, that we postpone to future works. Figure 10 shows the discounted cumulative reward collected by DQN and Pro DQN on both bicycle balancing and lunar lander, for different numbers of iterations. For DQN, each iteration corresponds to an update of the target network, while for Pro DQN it indicates an application of the trained PBO. Figure 10a shows that contrary to car-on-hill, on the bicycle balancing task Pro DQN improves slower than DQN, but it keeps increasing performance after K iterations thanks to PBO. This behavior illustrates that PBO can be used after the training to move the parameters in a favorable direction (see Figure 2a (blue trajectory)). Similarly, Figure 10b shows that, on lunar lander, Pro DQN keeps increasing ever so slightly, but it is also stronger than DQN overall. Discussion and Conclusion We introduced the novel idea of an operator that directly maps parameters of action-value functions to others, as opposed to the regular Bellman operator that requires costly projections steps onto the space of action-value functions. This operator called the projected Bellman operator can generate a sequence of parameters that can progressively approach the ones of the optimal action-value function. We formulated an optimization problem and an algorithmic implementation to learn PBO in offline and online RL. One limitation of our method in its current form is that, given that the size of input and output spaces of PBO depends on the number of parameters of the action-value function, it is challenging to scale to problems that learn action-value functions with deep neural networks with millions of parameters (Mnih et al. 2015). Nevertheless, recent advances in deep learning methods for learning operators (Kovachki et al. 2021, 2023) can provide a promising future direction. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Acknowledgments This work was funded by the German Federal Ministry of Education and Research (BMBF) (Project: 01IS22078). This work was also funded by Hessian.ai through the project The Third Wave of Artificial Intelligence 3AI by the Ministry for Science and Arts of the state of Hessen and by the grant Einrichtung eines Labors des Deutschen Forschungszentrum f ur K unstliche Intelligenz (DFKI) an der Technischen Universit at Darmstadt . This paper was also supported by FAIR (Future Artificial Intelligence Research) project, funded by the Next Generation EU program within the PNRR-PE-AI scheme (M4C2, Investment 1.3, Line on Artificial Intelligence). Agarwal, A.; Kakade, S.; Krishnamurthy, A.; and Sun, W. 2020. Flambe: Structural complexity and representation learning of low rank mdps. Advances in neural information processing systems, 33: 20095 20107. Asadi, K.; and Littman, M. L. 2017. An alternative softmax operator for reinforcement learning. In International Conference on Machine Learning, 243 252. PMLR. Bas-Serrano, J.; Curi, S.; Krause, A.; and Neu, G. 2021. Logistic Q-learning. In International Conference on Artificial Intelligence and Statistics, 3610 3618. PMLR. Beck, J.; Jackson, M. T.; Vuorio, R.; and Whiteson, S. 2023. Hypernetworks in Meta-Reinforcement Learning. In Proceedings of The 6th Conference on Robot Learning, volume 205 of Proceedings of Machine Learning Research. PMLR. Bellemare, M. G.; Dabney, W.; and Munos, R. 2017. A distributional perspective on reinforcement learning. In International Conference on Machine Learning, 449 458. PMLR. Bellemare, M. G.; Dabney, W.; and Rowland, M. 2023. Distributional Reinforcement Learning. MIT Press. http: //www.distributional-rl.org. Bellemare, M. G.; Ostrovski, G.; Guez, A.; Thomas, P.; and Munos, R. 2016. Increasing the action gap: New operators for reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 30. Bellman, R. 1966. Dynamic programming. Science, 153(3731): 34 37. Belousov, B.; and Peters, J. 2019. Entropic regularization of markov decision processes. Entropy, 21(7): 674. Bertsekas, D. 2019. Reinforcement learning and optimal control. Athena Scientific. Bertsekas, D. P. 2015. Dynamic Programming and Optimal Control 4 th Edition , Volume II. Athena Scientific. Bradtke, S. 1992. Reinforcement learning applied to linear quadratic regulation. Advances in neural information processing systems, 5. Brockman, G.; Cheung, V.; Pettersson, L.; Schneider, J.; Schulman, J.; Tang, J.; and Zaremba, W. 2016. Openai gym. ar Xiv preprint ar Xiv:1606.01540. Chauhan, V. K.; Zhou, J.; Lu, P.; Molaei, S.; and Clifton, D. A. 2023. A Brief Review of Hypernetworks in Deep Learning. ar Xiv preprint ar Xiv:2306.06955. Chen, T.; and Chen, H. 1995. Universal approximation to nonlinear operators by neural networks with arbitrary activation functions and its application to dynamical systems. IEEE Transactions on Neural Networks, 6(4): 911 917. Ernst, D.; Geurts, P.; and Wehenkel, L. 2005. Tree-Based Batch Mode Reinforcement Learning. JMLR, 6: 503 556. Farahmand, A.-m. 2011. Regularization in reinforcement learning. Fellows, M.; Hartikainen, K.; and Whiteson, S. 2021. Bayesian Bellman Operators. Advances in Neural Information Processing Systems, 34: 13641 13656. Geist, M.; Scherrer, B.; and Pietquin, O. 2019. A theory of regularized markov decision processes. In International Conference on Machine Learning, 2160 2169. PMLR. Ha, D.; Dai, A.; and Le, Q. 2016. Hyper Networks. In International Conference on Learning Representations. Haarnoja, T.; Tang, H.; Abbeel, P.; and Levine, S. 2017. Reinforcement learning with deep energy-based policies. In International conference on machine learning, 1352 1361. PMLR. Kissas, G.; Seidman, J. H.; Guilhoto, L. F.; Preciado, V. M.; Pappas, G. J.; and Perdikaris, P. 2022. Learning operators with coupled attention. Journal of Machine Learning Research, 23(215): 1 63. Kovachki, N.; Li, Z.; Liu, B.; Azizzadenesheli, K.; Bhattacharya, K.; Stuart, A.; and Anandkumar, A. 2021. Neural operator: Learning maps between function spaces. ar Xiv preprint ar Xiv:2108.08481. Kovachki, N.; Li, Z.; Liu, B.; Azizzadenesheli, K.; Bhattacharya, K.; Stuart, A.; and Anandkumar, A. 2023. Neural operator: Learning maps between function spaces with applications to PDEs. Journal of Machine Learning Research, 24(89): 1 97. Lillicrap, T. P.; Hunt, J. J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; and Wierstra, D. 2015. Continuous control with deep reinforcement learning. ar Xiv preprint ar Xiv:1509.02971. Micchelli, C. A.; and Pontil, M. 2005. On learning vectorvalued functions. Neural computation, 17(1): 177 204. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A. A.; Veness, J.; Bellemare, M. G.; Graves, A.; Riedmiller, M.; Fidjeland, A. K.; Ostrovski, G.; et al. 2015. Human-level control through deep reinforcement learning. nature, 518(7540): 529 533. Munos, R. 2003. Error bounds for approximate policy iteration. In ICML, volume 3, 560 567. Munos, R. 2005. Error bounds for approximate value iteration. In Proceedings of the National Conference on Artificial Intelligence, volume 20, 1006. Menlo Park, CA; Cambridge, MA; London; AAAI Press; MIT Press; 1999. Munos, R.; and Szepesv ari, C. 2008. Finite-Time Bounds for Fitted Value Iteration. Journal of Machine Learning Research, 9(5). The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24) Neu, G.; Jonsson, A.; and G omez, V. 2017. A unified view of entropy-regularized markov decision processes. ar Xiv preprint ar Xiv:1705.07798. Pang, B.; and Jiang, Z.-P. 2021. Robust reinforcement learning: A case study in linear quadratic regulation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, 9303 9311. Puterman, M. L. 1990. Markov decision processes. Handbooks in operations research and management science, 2: 331 434. Randlov, J.; and Alstrøm, P. 1998. Learning to Drive a Bicycle Using Reinforcement Learning and Shaping. In Proceedings of the 15th International Conference on Machine Learning, 463 471. Sarafian, E.; Keynan, S.; and Kraus, S. 2021. Recomposing the Reinforcement Learning Building Blocks with Hypernetworks. In Proceedings of the 38th International Conference on Machine Learning, Proceedings of Machine Learning Research, 9301 9312. PMLR. Sekhari, A.; Dann, C.; Mohri, M.; Mansour, Y.; and Sridharan, K. 2021. Agnostic reinforcement learning with lowrank mdps and rich observations. Advances in Neural Information Processing Systems, 34: 19033 19045. Song, Z.; Parr, R.; and Carin, L. 2019. Revisiting the softmax bellman operator: New benefits and new perspective. In International conference on machine learning, 5916 5925. PMLR. Sutton, R. S.; and Barto, A. G. 2018. Reinforcement learning: An introduction. MIT press. Tosatto, S.; D Eramo, C.; Pajarinen, J.; Restelli, M.; and Peters, J. 2019. Exploration driven by an optimistic bellman equation. In 2019 International Joint Conference on Neural Networks (IJCNN), 1 8. IEEE. Watkins, C. J. C. H. 1989. Learning from delayed rewards. The Thirty-Eighth AAAI Conference on Artificial Intelligence (AAAI-24)