# proximal_point_imitation_learning__3a88fc2e.pdf Proximal Point Imitation Learning Luca Viano LIONS, EPFL Lausanne, Switzerland luca.viano@epfl.ch Angeliki Kamoutsi ETH Zurich Zurich, Switzerland kamoutsa@ethz.ch Gergely Neu Universitat Pompeu Fabra Barcelona, Spain gergely.neu@gmail.com Igor Krawczuk LIONS, EPFL Lausanne, Switzerland igor.krawczuk@epfl.ch Volkan Cevher LIONS, EPFL Lausanne, Switzerland volkan.cevher@epfl.ch This work develops new algorithms with rigorous efficiency guarantees for infinite horizon imitation learning (IL) with linear function approximation without restrictive coherence assumptions. We begin with the minimax formulation of the problem and then outline how to leverage classical tools from optimization, in particular, the proximal-point method (PPM) and dual smoothing, for online and offline IL, respectively. Thanks to PPM, we avoid nested policy evaluation and cost updates for online IL appearing in the prior literature. In particular, we do away with the conventional alternating updates by the optimization of a single convex and smooth objective over both cost and Q-functions. When solved inexactly, we relate the optimization errors to the suboptimality of the recovered policy. As an added bonus, by re-interpreting PPM as dual smoothing with the expert policy as a center point, we also obtain an offline IL algorithm enjoying theoretical guarantees in terms of required expert trajectories. Finally, we achieve convincing empirical performance for both linear and neural network function approximation. 1 Introduction This work is concerned with the prototypical setting of imitation learning (IL) where 1. An expert provides demonstrations of state-action pairs in an environment. The expert could be optimal or suboptimal with respect to an unknown cost/reward function. 2. The learner chooses distance measure between its policy to be learned and the expert empirical distribution estimated from demonstrations. 3. The learner employs an algorithm, which additionally may or may not use interactions with the environment, to minimize the chosen distance. In IL, the central goal of the learner is to recover a policy competitive with expert with respect to the underlying unknown cost function. IL is important for several real world applications like driving [62], robotics [88], and economics/finance [27] at the expense of following resources: (R1) expert demonstrations, (R2) (optional) interactions with the environment where the expert collected the demonstrations, and (R3) computational resources for solving the problem template. Interestingly, while there is a vast amount of literature using optimization ideas on the IL problem template, i.e. Lagrangian duality [51, 38, 59, 63, 64], resource guarantees are still widely missing since the optimization literature focuses on the resource (R3) where IL literature mainly focuses on 36th Conference on Neural Information Processing Systems (Neur IPS 2022). the first two resources (R1) and (R2). Our work leverages deeper connections between optimization tools and IL by showing how classical optimization tools can be applied in a linear programming formulation of IL problem guaranteeing efficiency in all (R1), (R2), (R3). Our contributions: This work aims at designing an algorithm enjoying both theoretical guarantees and convincing empirical performance. Our methodology is rooted in classical optimization tools and the LP approach to MDPs. More precisely, the method uses the recently repopularized overparameterization technique to obtain the Q-function as a Lagrangian multiplier [77, 14] and solves the associated program using a PPM update with appropriately chosen Bregman divergences. This results to an actor-critic algorithm, with the key feature that the policy evaluation step involves optimization of a single concave and smooth objective over both cost and Q-functions. In this way, we avoid instability or poor convergence due to adversarial training [51, 122, 70, 105], and can also recover an explicit cost along with Q-function. We further account for potential optimization errors, presenting an error propagation analysis that leads to rigorous guarantees for both online and offline setting. For the context of linear MDPs [14, 121, 55, 22, 116, 7, 84], we provide explicit convergence rates and error bounds for the suboptimality of the learned policy, under mild assumptions, significantly weaker than those found in the literature until now. To our knowledge, such guarantees in this setting are provided for the first time. Finally, we demonstrate that our approach achieves convincing empirical performance for both linear and neural network function approximation. Related Literature. The first algorithm addressing the imitation learning problem is behavioral cloning [93]. Due to the covariate shift problem [98, 99], it has low efficiency in terms of expert trajectories (R1). To address this issue, [100, 87, 4, 95, 111, 85, 123, 5, 68, 69] proposed to cast the problem as inverse reinforcement learning (IRL). IRL improves the efficiency in terms of expert trajectories, at the cost of introducing the need of running reinforcement learning (RL) repetitively, which can be prohibitive in terms of environment samples (R2) and computation (R3). A successive line of work started with [112] highlights that repeated calls to an RL routine can be avoided. This work inspired generative adversarial imitation learning (GAIL) [51] and other follow-up works [38, 59, 63, 64] that leveraged optimization tools like primal-dual algorithms but did not try to deepen the optimization connections to derive efficiency guarantees in terms of all (R1),(R2),(R3). Finally, a recent line of work [40, 57] in IL bypasses the need of optimizing over cost functions and thus avoids instability due to adversarial training. Although these algorithms achieve impressive empirical performance in challenging high dimensional benchmark tasks, they are hampered by limited theoretical understanding. This is the fundamental difference from our work, which enjoys both favorable practical performance and strong theoretical guarantees. Existing model-free IL theoretical papers with global convergence guarantees assume either a finite horizon episodic MDP setting [70], or tabular MDPs [105], or the infinite horizon case but with restrictive assumptions, such as linear quadratic regulator setting [21], continuous kernelized nonlinear regulator [26, 56], access to a generative model and coherence assumption on the choice of features [58, 14], bounded strong concentrability coefficients [122] or a linear transition law that can be completely specified by a finite-dimensional matrix [70]. On the other hand, we provide convergence guarantees and error bounds for the context of linear MDPs [14, 121, 55, 22, 116, 7, 84] under a mild feature excitation condition assumption. Despite being linear, the transition law can still have infinite degrees of freedom. To our knowledge, such guarantees in this setting are provided for the first time. Our work applies the technique known as regularization in the online learning literature [6, 103] and Bregman proximal-point or smoothing in optimization literature [97, 82] to the LP formulation for MDPs [73, 35, 36, 17, 48, 49, 33, 34, 102, 91, 92, 1, 65, 30, 79, 115, 67, 13, 31, 55, 106]. From this perspective, we can see Deep Inverse Q-Learning [57] and IQ-Learn [40] that consider entropy regularization in the objective as smoothing using uniform distribution as center point. In our case, we instead use as center point the previous iteration of the algorithm (for the online case) or the expert (for the offline case). From the technical point of view, the most important related works are the analysis of REPS/QREPS [90, 14, 89] and O-REPS [124] that first pointed out the connection between REPS and PPM. We build on their techniques with some important differences. In particular, while in the LP formulation of RL, PPM and mirror descent [15, 47] are equivalent, recognizing that they are not equivalent in IL is critical for stronger empirical performance. As an independent interest, our techniques can be used to improve upon the best rate for REPS in the tabular setting [89] and to extend the guarantees to linear MDPs. In order to discuss in more detail our research questions and situate them among prior related theoretical and practical works, we provide in Appendix A an extended literature review. 2 Background 2.1 Markov Decision Processes The RL environment and its underlying dynamics are typically abstracted as an MDP given by a tuple (S, A, P, ν0, c, γ), where S is the state space, A is the action space, P : S A S is the transition law, ν0 S is the initial state distribution, c [0, 1]|S||A| is the cost, and γ (0, 1) is the discount factor. For simplicity, we focus on problems where S and A are finite but too large to be enumerated. A stationary Markov policy π: S A interacts with the environment iteratively, starting with an initial state s0 ν0. At round t, if the system is at state st, an action at π( |st) is sampled and applied to the environment. Then a cost c(s, a) is incurred, and the system transitions to the next state st+1 P( |s, a). The goal of RL is to solve the optimal control problem ρ c minπ ρc(π), where ρc(π) (1 γ) ν0, Vπ c is the normalized total discounted expected cost of π. The state value function Vπ c R|S| of π, given cost c, is defined by V π c (s) Eπ s h P t=0 γtc(st, at) i , where Eπ s denotes the expectation with respect to the trajectories gen- erated by π starting from s0 = s. The optimal value function V c R|S| is defined by V c (s) minπ V π c (s). The optimal state-action value function Q c R|S||A|, given by Q c(s, a) c(s, a) + γ P s V c (s )P(s |s, a), is known to characterize optimal behaviors. Indeed V c is the unique solution to the Bellman optimality equation V c (s) = mina Q c(s, a). In addition, any deterministic policy π c(s) = arg mina Q c(s, a) is known to be optimal. For every policy π, we define the normalized state-action occupancy measure µπ S A, by µπ(s, a) (1 γ) P t=0 γt Pπ ν0 [st = s, at = a] , where Pπ ν0[ ] denotes the probability of an event when following π starting from s0 ν0. The occupancy measure can be interpreted as the discounted visitation frequency of state-action pairs. This allows us to write ρc(π) = µπ, c . 2.2 Imitation Learning Similarly to RL, the IL problem is posed in the MDP formalism, with the critical difference that the true cost ctrue is unknown. Instead, we have access to a finite set of truncated trajectories sampled i.i.d. by executing an expert policy πE in the environment. The goal is to learn a policy that performs better than πE with respect to the unknown ctrue. To this end, we adopt the apprenticeship learning formalism [4, 112, 50, 51, 105], which carries the assumption that ctrue belongs to a class of cost functions C. We then seek an apprentice policy πA that outperforms the expert across C by solving the following optimization problem ζ min π d C(π, πE), (1) where d C(π, πE) maxc C ρc(π) ρc(πE) defines the C-distance between π and πE [51, 28, 122, 70]. Then, πA satisfies the goal of IL, since it holds that ρctrue(πA) ρctrue(πE) ζ 0. Intuitively, the cost class C distinguishes the expert from other policies. The maximization in (1) assigns high total cost to non-expert policies and low total cost to πE [51], while the minimization aims to find the policy that matches the expert as close as possible with respect to d C. By writing d C in its dual form d C(µπ, µπE) maxc C µπ, c µπE, c , it can be interpreted as an integral probability metric [80, 60] between the occupancy measures µπ and µπE. Depending on how C is chosen, d C turns to a different metric of probability measures like the 1-Wasserstein distance [117, 32] for C = Lip1(S A), the total variation for C = {c | c 1}, or the maximum mean discrepancy for C = {c | c H 1}, where Lip1(S A) denotes the space of 1-Lipschitz functions on S A, and H denotes the norm of a reproducing kernel Hilbert space H [104]. In our theoretical analysis, we focus on linearly parameterized cost classes [111, 112, 51, 70, 105] of the form C {cw Pm i=1 wiϕi | w W}, where {ϕi}m i=1 R|S||A| + are fixed feature vectors, such that ϕi 1 1 for all i [m], and W is a a convex constraint set for the cost weights w. This assumption is not necessarily restrictive as usually in practice the true cost depends on just a few key properties, but the desirable weighting that specifies how different desiderata should be traded-off is unknown [4]. Moreover, the cost features can be complex nonlinear functions that can be obtained via unsupervised learning from raw state observations [20, 29]. The matrix Φ [ϕ1 . . . ϕm] gives rise a feature expectation vector (FEV) ρΦ(π) (ρϕ1(πE), . . . , ρϕm(πE))T Rm of a policy π. Then, by choosing W to be the ℓ2 unit ball Bm 1 {w Rm | w 2 1} [4], we get a feature expectation matching objective d C(π, ππE) = ρΦ(π) ρΦ(πE) 2, while for W being the probability simplex [m] [111, 112] we have a worst-case excess cost objective d C(π, ππE) = maxi [m] ρϕi(π) ρϕi(πE) . For clarity, we will replace c by w in the notation of the quantities defined in Section 2.1. 3 A Q-Convex-Analytic Viewpoint Our methodology builds upon the convex-analytic approach to AL, first introduced by [112], with the key difference that we consider a different convex formulation that introduces Q-functions as slack variables. This allows to design a practical scalable model-free algorithm with theoretical guarantees. Let F {µ R|S||A| | (B γP) µ = (1 γ)ν0, µ 0} be the state-action polytope, where P is the vector form of P, i.e., P(s,a),s P(s |s, a), and B is a binary matrix defined by B(s,a),s 1 if s = s , and B(s,a),s 0 otherwise. The linear constraints that define the set F, also known as Bellman flow constraints, precisely characterize the set of state-action occupancy measures. Proposition 1 (94). We have that µ F if and only if there exists a unique stationary Markov policy π such that µ = µπ. If µ F then the policy πµ(a|x) µ(x,a) P a A µ(x,a ) has occupancy measure µ. Using Proposition 1 and the dual form of the C-distance d C(µ, µπE) = maxw W µ µπE, cw , it follows that (1) is equivalent to the primal convex program ζ = minµ{ d C(µ, µπE) | µ F}. In particular for W = [m] and by using an epigraphic transformation, we end up with an LP program [112], while for W = Bm 1 we get a quadratic objective with linear constraints [4]. A slight variation of the above reasoning is to introduce a mirror variable d and split the Bellman flow constraints in the definition of F. We then get the primal convex program ζ = min (µ,d){ d C(µ, µπE) | (µ, d) M}, (Primal) where the new polytope is given by M {(µ, d) | B d = γP µ + (1 γ)ν0, µ = d, d 0}. This overparameterization trick has been first introduced by Mehta and Meyn [76] and has been recently revisited by [14, 84, 67, 83, 77, 71]. A salient feature of this equivalent formulation is that it introduces a Q-function as Lagrange multiplier to the equality constraint d = µ, and so lends itself to data-driven algorithms. To motivate further this new formulation, in Appendix C, we shed light to its dual and provide an interpretation of the dual optimizers. In particular, when W = Bm 1 , we show that (V wtrue, Q wtrue, wtrue) is a dual optimizer. For our theoretical analysis we focus on the linear MDP setting [55], i.e., we assume that the transition law is linear in the feature mapping. We denote by ϕ(s, a) the (s, a)-th row of Φ. Assumption 1 (Linear MDP). There exists a collection of m probability measures ω = (ω1, . . . , ωm) on S, such that P( |s, a) = ω( ), ϕ(s, a) , for all (s, a). Moreover ϕ(s, a) [m], for all (s, a). Assumption 1 essentialy says that the transition matrix P has rank at most m, and P = ΦM for some matrix M Rm |S|. It is worth noting that in the case of continuous MDPs, despite being linear, the transition law P( |s, a) can still have infinite degrees of freedom. This is a substantial difference from the recent theoretical works on IL [70, 105] which consider either a linear quadratic regulator, or a transition law that can be completely specified by a finite-dimensional matrix such that the degrees of freedom are bounded. Assumption 1 enables us to consider a relaxation of (Primal). In particular, we aggregate the constraints µ = d by imposing Φ µ = Φ d instead, and introduce a variable λ = Φ µ. It follows that λ lies in the m-dimensional simplex [m]. Then, we get the following convex program ζ = min (λ,d){max w W λ, w µπE, cw | (λ, d) MΦ}, (Primal ) where MΦ {(λ, d) | B d = γM λ + (1 γ)ν0, λ = Φ d, λ [m], d S A}. As shown in [84, 14, 83], for linear MDPs, the set of occupancy measures F can be completely characterized by the set MΦ (c.f., Proposition 2). While the number of constraints and variables in (Primal ) is intractable for large scale MDPs, in the next paragraph, we show how this problem can be solved using a proximal point scheme. 4 Proximal Point Imitation Learning By using a Lagrangian decomposition, we have that (Primal ) is equivalent to the following bilinear saddle-point problem min x X max y Y y, Ax + b , (SPP) where A R(2m+|S|) (m+|S||A|), and b R(m+|S|+|S||A|) are appropriately defined (see Appendix D), x [λ , d ] , y [w , V , θ ] , X [m] S A, and Y W R|S| Rm. Since in practice we do not have access to the whole policy πE, but instead can observe a finite set of i.i.d. sample trajectories DE {(x(l) 0 , a(l) 0 , x(l) 1 , a(l) 1 , . . . , x(l) H , a(l) H )}n E l=1 πE, we define the vector bb by replacing ρΦ(πE) with its empirical counterpart ρΦ(c πE) (by taking sample averages) in the definition of b. We then consider the empirical objective f(x) maxy Y D y, Ax + bb E and apply PPM on the decision variable x. For the λ-variable we use the relative entropy D(λ||λ ) Pm i=1 λ(i) log λ(i) λ (i), while for the occupancy measure d we use the conditional relative entropy s,a d(s, a) log πd(a|s) πd (a|s). With this choice we can rewrite the PPM update as (λk+1, dk+1) = arg min λ [m],d S A max y Y η D(λ||Φ dk) + 1 αH(d||dk), (2) where we used primal feasibility to replace λk with Φ dk as the center point of the relative entropy. PPM is implicit, meaning that it requires the evaluation of the gradient at the next iterate xk+1. Such a requirement makes it not implementable in general. However, in the following, we describe a procedure to apply proximal point to our specific f(x). The following Proposition summarizes the result. Proposition 2. For a parameter θ Rm, we define the logistic state-action value function Qθ R|S||A| by Qθ Φθ, and the k-step logistic state value function Vk θ R|S| by V k θ (s) 1 a πdk 1(a|s)e αQθ(s,a) ! Moreover, we define the k-step reduced Bellman error function δk w,θ Rm by δk w,θ w+γMVk θ θ. Then, the PPM update (λ k, d k) in 2 is given by λ k(i) (Φ dk 1)(i) e ηδk w k,θ k (i), (3) πd k(a|s) πdk 1(a|s) e αQθ k (s,a), (4) where (w k, θ k) is the maximizer over W Rm of the k-step logistic policy evaluation objective i=1 (Φ dk 1)(i)e ηδk w,θ(i) + (1 γ) ν0, Vk θ ρΦ(c πE), w . (5) Moreover, it holds that Gk(w k, θ k) = λ k, w k ρΦ(c πE), w k + 1 ηD(λ k||Φ λk 1) + 1 αH(d k||dk 1). If in addition Assumption 1 holds, then d k is a valid occupancy measure, i.e., d k F and so d k = µπd k . The proof of Proposition 2 is broken down into a sequence of lemmas and is presented in Appendix E. It employs an analytical-oracle g : Y X given by g(y; xk) arg min λ [m],d S A η D(λ||Φ dk) + 1 and a max-oracle h : X Y given by h(x) arg maxy Y y, Ag(y; x) + 1 τ DΩ(g(y; x)||x), where we used DΩto compact the two divergences. By noting that the PPM update Equation (2) can be rewritten as xk+1 = g(h(xk); xk), its analytical computation is reduced to the characterization of the two aforementioned oracles. In particular, the updates (3) (4) come from the analytical-oracle while (5) is the objective of the max-oracle. The choice of conditional entropy as Bregman divergence for the λ variable living in the probability simplex is standard in the optimization literature and is known to mitigate the effect of dimension. In particular, as noted in [85], the classic REPS algorithm [90] can be seen as mirror descent with relative entropy regularization. On the other hand, the choice of conditional entropy as Bregman divergence for the d variable is less standard and has been popularized by Q-REPS [14]. Such particular divergence leads to an actor-critic algorithm that comes with several merits. By Proposition 2, it is apparent that we get analytical softmin updates for the policy πd rather than the occupancy measure d. Moreover, these softmin updates are expressed in terms of the logistic Q-function and do not involve the unknown transition matrix P. Consequently, we avoid the problematic occupancy measure approximation and the restrictive coherence assumption on the choice of features needed in [13, 58], as well as the biased policy updates appearing in REPS [90, 89]. In addition, the newly introduced logistic policy evaluation objective Gk(w, θ) has several desired properties. It is concave and smooth in (w, θ) and has bounded gradients. Therefore, it does not suffer from the pathologies of the squared Bellman error [78] and does not require heuristic gradient clipping techniques. Moreover, unlike [58] it allows a model-free implementation without the need for a generative model (see Section 4.1) We stress the fact that the max-oracle of our proximal point scheme performs the cost update and policy evaluation phases jointly. This is a rather novel feature of our algorithm that differs from the separate cost update and policy evaluation step used in recent theoretical imitation learning works [122, 105, 70]. Our joint optimization over cost and Q-functions avoids instability due to adversarial training and can also recover an explicit cost along with the Q-function without requiring knowledge or additional interaction with the environment (see Section 5). It is worth noting that application of primal-dual mirror descent to (SPP) does not have this favorable property. While in the standard MDP setting, proximal point and mirror descent coincide because of the linear objective, in imitation learning proximal point optimization makes a difference. In Appendix K, we include a more detailed discussion and numerical comparison between PPM and mirror descent updates. 4.1 Practical Implementation Exact optimization of the logistic policy evaluation objective is infeasible in practical scenarios, due to unknown dynamics and limited computation power. In this section, we design a practical algorithm that uses only sample transitions by obtaining stochastic (albeit biased) gradient estimators. Proposition 2 gives rise to Proximal Point Imitation Learning (P2IL), a model-free actor-critic IRL algorithm described in Algorithm 1. The key feature of P2IL is that the policy evaluation step involves optimization of a single smooth and concave objective over both cost and state-action value function parameters. In this way, we avoid instability or poor convergence in optimization due to nested policy evaluation and cost updates, as well as the undesirable properties of the widely used squared Bellman error. In particular, the kth iteration of P2IL consists of the following two steps : (i) (Critic Step) Computation of an approximate maximizer (wk, θk) arg maxw,θ Gk(w, θ) of the concave logistic policy evaluation objective, by using a biased stochastic gradient ascent subroutine; (ii) (Actor Step) Soft-min policy update πk(a|s) πk 1(a|s) e αQθk (s,a) expressed in terms of the logistic Q-function. The domain Θ in Algorithm 1 is the ℓ -ball with appropriately chosen radius D to be specified later (see Proposition 3). Moreover, ΠΘ(x) arg miny Θ x y 2 (resp. ΠW(w)) denotes the Euclidean projection of x (resp. w) onto Θ (resp. W). In order to estimate the gradients θ Gk(w, θ) and w Gk(w, θ) we invoke the Biased Stochastic Gradient Estimator subroutine (BSGE) (Algorithm 2) given in Appendix H. By using the linear MDP Assumption 1 and leveraging ridge regression and plug-in estimators, the proposed stochastic gradients can be computed via simple linear algebra with computational complexity poly(m, n(t)), independent of the size of the state space. Algorithm 1 Proximal Point Imitation Learning: P2IL(Φ, DE, K, η, α) Input: Feature matrix Φ, expert demonstrations DE, number of iterations K, step sizes η and α, number of SGD iterations T, SGD learning rates β = {βt}T 1 t=0 , number-of-samples function n : N N Initialize π0 as uniform distribution over A Compute the empirical FEV ρΦ(c πE) using expert demonstrations DE for k = 1, . . . K do // Critic-step (policy evaluation) Initialize θk,0 = 0 and wk,0 = 0 Run πk 1 and collect i.i.d. samples Bk = {(s(n) k 1, a(n) k 1, s (n) k 1)}n(T ) n=1 such that (s(n) k 1, a(n) k 1) µπk 1 and s (n) k 1 P( |s(n) k 1, a(n) k 1) for t = 0, . . . T 1 do Compute biased stochastic gradient estimators b w Gk(wk,t, θk,t), b θGk(wk,t, θk,t) = BSGE k, wk,t, θk,t, n(t) wk,t+1 = ΠW wk,t + βt b w Gk(wk,t, θk,t) θk,t+1 = ΠΘ θk,t + βt b θGk(wk,t, θk,t) end for (wk, θk) = ( 1 T PT t=1 wk,t, 1 T PT t=1 θk,t) // Actor-step (policy update) Policy update: πk(a|s) πk 1(a|s) e αQθk (s,a) end for Output: Mixed policy bπK of {πk}k [K] 4.2 Theoretical Analysis The first step in our theoretical analysis is to study the propagation of optimization errors made by the algorithm on the true policy evaluation objective. In particular at each iteration step k, the ideal policy evaluation update (w k, θ k) and the ideal policy update π k are given by (w k, θ k) = arg maxw,θ Gk(w, θ), and π k(a|s) = πk 1(a|s)e α(Qθ k (s,a) V k θ k (s)). On the other hand, consider the realised policy evaluation update (wk, θk) such that Gk(w k, θ k) Gk(wk, θk) = ϵk, the corre- sponding policy πk given by πk = πk 1(a|s)e α(Qθk (s,a) V k θk (s)), and let dk µπk. We denote by bπK the extracted mixed policy of {πk}K k=1. We are interested in upper-bounding the suboptimality gap d C(bπK, πE) of Algorithm 1 as a function of εk. To this end, we need the following assumption. Assumption 2. It holds that λmin(E(s,a) dk ϕ(s, a)ϕ(s, a)T) β, for all k [K]. Assumption 2 states that every occupancy measure dk induces a positive definite feature covariance matrix, and so every policy πk explores uniformly well in the feature space. This assumption is common in the RL theory literature [2, 46, 37, 66, 3, 7]. It is also related to the condition of persistent excitation from the control literature [81]. The following proposition ensures that maxw,θ W Rm Gk(w, θ) = maxw,θ W Θ Gk(w, θ). Therefore, this constraint does not change the problem optimality, but will considerably accelerate the convergence of the algorithm by considering smaller domains. Proposition 3. There exists a maximizer θ k such that θ k 1+|log β| We can now state our error propagation theorem. Theorem 1. Let bπK be the output of running Algorithm 1 for K iterations, with n E 2 log( 2m δ ) ε2 expert trajectories of length H 1 1 γ log( 1 ε). Let C 1 βη q 2α 1 γ + 8η + q 18α 1 γ . Then, with probability at least 1 δ, it holds that d C(bπK, πE) 1 K D(λ ||ΦTd0) η + H(d ||d0) By Theorem 1, whenever the policy evaluation errors εk, as well as the estimation error ε can be kept small, Algorithm 1 ouputs a policy bπK with small suboptimality gap ρctrue(bπK) ρctrue(πE). 0 15 30 45 Samples (x 100) (a) River Swim 0 250 500 750 1000 Samples (x 100) (b) Cart Pole 0 25 50 75 100 Samples (x 100) (c) Double Chain 0 60 120 180 240 Samples (x 100) (d) Gridworld 0 4 8 12 Samples (x 100) (e) Acrobot Proximal Point IQLearn AIRL GAIL AIRL Linear GAIL Linear Figure 1: Online IL Experiments. We show the total returns vs the number of env steps. Notably, there is no direct dependence on the size of the state space or the dimension of the feature space. In the ideal case, where εk = 0 for all k, the convergence rate is O(1/K). The provided error propagation analysis still holds with general function approximation, i.e., in the context of deep RL. Indeed, by choosing Φ = I, Assumption 1 is trivially satisfied and the θ variable in the objective Gk is replaced by a Q-function. In practice, the estimation error ε can be made arbitrary small, by increasing the number of expert demonstrations n E. Moreover, the next theorem ensures that under Assumptions 1 and 2 the biased stochastic gradient ascent (BSGA) subroutine has sublinear convergence rate. Theorem 2. Let (wk, θk) be the output of the BSGA subroutine in Algorithm 1 for T iterations, with n(t) max O γ2m Dt (η+α)2β log T m δ , O mt (η+α)2β log T m δ sample transitions, and learning rates t). Then, ϵk = Gk(w k, θ k) Gk(wk, θk) O( max{η,1}m D T ), with probability 1 δ. Corollary 1 (Resource guarantees). Choose η = α = 1 and let K = Ω ϵ 1 , T = Ω ϵ 4 . Then for Ω(KT) = Ω ϵ 5 sample transitions, Ω ε 2 expert trajectories and approximately solving Ω ϵ 1 concave maximization problems, we can ensure d C(bπ, πE) O(ϵ+ε), with high probability. Offline Setting. Finally, we notice that using Φ µπE as the reference distribution for the relative entropy we can obtain an offline algorithm that does not require environment interactions. By reinterpreting smoothing [82] as one step of proximal point, and using similar arguments as in the proof of Theorem 1, we can provide similar theoretical guarantees for the offline setting. The formal statement of the theoretical result as well as the optimization of the empirical policy evaluation objective are presented in Appendix J (see Theorems 4 and 6). 5 Experiments In this section, we demonstrate that our approach achieves convincing empirical performance in both online and offline IL settings on several environments.1 The precise setting is detailed in Appendix L. Online Setting. We first present results in various tabular environments where we can implement our algorithm without any practical relaxation outperforming GAIL [51], AIRL [38] and IQ-Learn [40]. Results are given in Figure 1. Good performance but inferior to IQ-Learn is observed also for continuous states environments (Cart Pole and Acrobot) where we used neural networks function approximation. Offline Setting. Figures 2a to 2c shows that our method is competitive with the state-of-the-art offline IL methods IQLearn [40] and AVRIL [25] that recently showed performances superior to other methods like [54][64]. We also tried our algorithm in the complex image-based Pong task from the Atari suite. Figure 2d shows that the algorithm reaches the expert level after observing 2e5 expert samples. We did not find AVRIL competitive in this setting, and skip it for brevity. In these settings, we verified that the algorithmic performance is convincing even for costs parameterized by neural networks. Continuous control experiments. We attain the expert performance also in 2 Mu Jo Co environments: Ant, Half Cheetah, Hopper, and Walker (see Figures 2e to 2h). The additional difficulty in implementing the algorithm in continuous control experiments is that the analytical form of the policy 1The code is available at the following link https://github.com/lviano/P2IL. improvement step is no longer computationally tractable because this would require to compute an integral over the continuous action space. Therefore, we approximated this update using the Soft Actor Critic (SAC) [44] algorithm. SAC requires environment samples making the algorithm online. The good empirical result opens the question of analyzing policy improvement errors as in [41]. 0 3 6 9 12 15 Expert Trajectories (a) Acrobot 0 3 6 9 12 15 Expert Trajectories (b) Cart Pole 0 3 6 9 12 15 Expert Trajectories (c) Lunar Lander 0 100 200 300 400 Expert Samples (x 100) Proximal Point IQLearn AVRIL 0 500 1000 1500 2000 Samples (x 100) (e) Half Cheetah 0 500 1000 1500 2000 Samples (x 100) 0 1000 2000 3000 4000 Samples (x 100) 0 1000 2000 3000 4000 Samples (x 100) (h) Walker2d Proximal Point IQLearn AIRL GAIL Figure 2: Neural function approximation experiments. Figures 2a to 2c show the total returns vs the number of expert trajectories. Figures 2e to 2h show the total returns vs the number of env steps. Figure 2d shows the total return vs the number of expert state-action pairs. Recovered Costs. A unique algorithmic feature of the proposed methodology is that we can explicitly recover a cost along with the Q-function without requiring adversarial training. In Figure 3, we visualize our recovered costs in a simple 5x5 Gridworld. Most importantly, we verify that the recovered costs induce nearly optimal policies w.r.t. the unknown true cost function. Compared to IQ-Learn [40], we do not require knowledge or further interaction with the environment. Therefore, the recovered cost functions show promising transfer capability to new dynamics. Action index State Index Action index State Index (c) -V ctrue (d) -V c K Figure 3: Recovered Costs in Gridworld. Comparison between the true cost ctrue and the cost c K recovered by P2IL. We notice that the optimal value functions V ctrue and V c K present the same pattern. Hence, the optimal policy with respect to c K is nearly optimal with respect to ctrue. Cost Transfer Setting. We experimented with a transfer cost setting on a Gridworld (Figure 4). We consider two different Gridworld MDP environments, say M and f M, with opposite action effects. This means that action Down in f M corresponds to action Left in M and vice versa. Similarly, the effects of Up and Right are swapped between f M and M. We denote by Vπ f M,ctrue (resp. V f M,ctrue) the value function of policy π (resp. optimal value function) in the MDP environment f M with cost function ctrue. Moreover, we denote by π M,c the optimal policy in the MDP environment M under cost function c. Figure (a) gives the corresponding optimal value function. Figure (b) presents the value function of the expert policy πE = π M,ctrue used as target by P2IL. Figure (d) shows the value function of the learned imitating policy πK from P2IL. Finally, Figure (b) depicts the value function of the optimal policy π f M,c K for the environment f M endowed with the recovered cost function c K by P2IL (with access to samples from M). We conclude that the policy π f M,c K is optimal in f M with cost ctrue. By contrast, the expert policy πE = π M,ctrue used as target by P2IL performs poorly and as a consequence also the imitating policy πK does so. All in all, we notice that the recovered cost induces an optimal policy for the new dynamics while the imitating policy fails. Albeit, cost transfer is successful in this experiment we do not expect this fact to be true in general because we do not tackle the issue of cost shaping [87]. (a) V f M,ctrue (b) V π M,ctrue f M,ctrue (c) V π f M,c K f M,ctrue (d) VπK f M,ctrue Figure 4: Cost Transfer Experiment in Gridworld. We compare the performance of several policies in the new MDP environment f M with cost function ctrue. We notice that the recovered cost induces an optimal policy for the new dynamics while the imitating policy fails. 6 Discussion and Outlook In this work, we studied a Proximal Point Imitation Learning (P2IL) algorithm with both theoretical guarantees and convincing empirical performance. Our methodology is rooted in classical optimization tools and the LP approach to MDPs. The most significant merits of P2IL are the following: (i) It optimizes a convex and smooth logistic Bellman evaluation objective over both cost and Q-functions. In particular, it avoids instability due to adversarial training and can also recover an explicit cost along with Q function; (ii) In the context of linear MDPs, it comes with efficient resource guarantees and error bounds for the suboptimality of the learned policy (Theorem 2 and Corollary 1). In particular, given poly(1/ε, log(1/δ), m) many samples , it recovers an ε-optimal policy, with probability 1 δ. Notably, the bound is independent of the size of the state-action space; (iii) Beyond the linear MDP setting, it can be implemented in a model-free manner, for both online and offline setups, with general function approximation without losing its theoretical specifications. This is justified by providing an error propagation analysis (Theorems 1 and 4), guaranteeing that small optimization errors lead to high-quality output policy; (iv) It enjoys not only strong theoretical guarantees but also favorable empirical performance. At the same time, our newly introduced methods bring challenges and open questions. One interesting question is whether one can accelerate the PPM updates and improve the convergence rate. Another direction for future work is to provide rigorous arguments for the near-optimality of the recovered cost function. On the practical side, we plan to conduct experiments in more challenging environments than Mu Jo Co and Atari. We hope our new techniques will be useful to future algorithm designers and lay the foundations for overcoming current limitations and challenges. In Appendix B, we point out in detail a few interesting future directions. Acknowledgements The authors would like to thank the anonymous reviewer for their suggestions to improve the presentation and for motivating us to inspect the recovered cost function. This work has received funding from the Enterprise for Society Center (E4S), the European Research Council (ERC) under the European Union s Horizon 2020 research and innovation programme grant agreement OCAL, No. 787845, the European Research Council (ERC) under the European Union s Horizon 2020 research and innovation programme (grant agreement n 725594 - time-data), the Swiss National Science Foundation (SNSF) under grant number 200021_205011. Gergely Neu was supported by the European Research Council (ERC) under the European Union s Horizon 2020 research and innovation programme (Grant agreement No. 950180). Luca Viano acknowledges travel support from ELISE (GA no 951847). [1] Y. Abbasi-Yadkori, P. L. Bartlett, and A. Malek. Linear programming for large-scale Markov decision problems. In International Conference on Machine Learning (ICML), 2014. [2] Y. Abbasi-Yadkori, P. Bartlett, K. Bhatia, N. Lazic, C. Szepesvari, and G. Weisz. Politex: Regret bounds for policy iteration using expert prediction. In International Conference on Machine Learning (ICML), 2019. [3] Y. Abbasi-Yadkori, N. Lazic, C. Szepesvari, and G. Weisz. Exploration-enhanced politex. ar Xiv:1908.10479, 2019. [4] P. Abbeel and A. Y. Ng. Apprenticeship learning via inverse reinforcement learning. In International Conference on Machine Learning (ICML), 2004. [5] P. Abbeel, D. Dolgov, A. Y. Ng, and S. Thrun. Apprenticeship learning for motion planning with application to parking lot navigation. In IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2008. [6] J. D. Abernethy, E. Hazan, and A. Rakhlin. Competing in the dark: An efficient algorithm for bandit linear optimization. In Annual Conference on Learning Theory (COLT), 2008. [7] A. Agarwal, S. Kakade, A. Krishnamurthy, and W. Sun. Flambe: Structural complexity and representation learning of low rank MDPs. Advances in neural information processing systems (Neur IPS), 2020. [8] A. Ayoub, Z. Jia, C. Szepesvari, M. Wang, and L. Yang. Model-based reinforcement learning with value-targeted regression. In International Conference on Machine Learning (ICML), 2020. [9] J. A. Bagnell and J. G. Schneider. Covariant policy search. In International Joint Conference on Artificial Intelligence (IJCAI), 2003. [10] G. Banjac and J. Lygeros. A data-driven policy iteration scheme based on linear programming. In IEEE Conference on Decision and Control (CDC), 2019. [11] P. Barde, J. Roy, W. Jeon, J. Pineau, C. Pal, and D. Nowrouzezahrai. Adversarial soft advantage fitting: Imitation learning without policy optimization. In Advances in Neural Information Processing Systems (Neur IPS), 2020. [12] A. G. Barto, R. S. Sutton, and C. W. Anderson. Neuronlike adaptive elements that can solve difficult learning control problems. IEEE transactions on systems, man, and cybernetics, pages 834 846, 1983. [13] J. Bas-Serrano and G. Neu. Faster saddle-point optimization for solving large-scale Markov decision processes. In Conference on Learning for Dynamics and Control (L4DC), 2020. [14] J. Bas-Serrano, S. Curi, A. Krause, and G. Neu. Logistic Q-learning. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2021. [15] A. Beck and M. Teboulle. Mirror descent and nonlinear projected subgradient methods for convex optimization. Operations Research Letters, 31(3):167 175, 2003. [16] P. N. Beuchat, A. Georghiou, and J. Lygeros. Performance guarantees for model-based approximate dynamic programming in continuous spaces. IEEE Transactions on Automatic Control, 65(1):143 158, 2020. [17] V. S. Borkar. A convex analytic approach to Markov decision processes. Probability Theory and Related Fields, 78(4):583 602, 1988. [18] S. Boucheron, G. Lugosi, and P. Massart. Concentration inequalities: A nonasymptotic theory of independence. Oxford university press, 2013. [19] G. Brockman, V. Cheung, L. Pettersson, J. Schneider, J. Schulman, J. Tang, and W. Zaremba. Open AI Gym. ar Xiv:1606.01540, 2016. [20] D. S. Brown, R. Coleman, R. Srinivasan, and S. Niekum. Safe imitation learning via fast Bayesian reward inference from preferences. In International Conference on Machine Learning (ICML), 2020. [21] Q. Cai, M. Hong, Y. Chen, and Z. Wang. On the global convergence of imitation learning: a case for linear quadratic regulator. ar Xiv:1901.03674, 2019. [22] Q. Cai, Z. Yang, C. Jin, and Z. Wang. Provably efficient exploration in policy optimization. In International Conference on Machine Learning (ICML), 2020. [23] Y. Carmon, Y. Jin, A. Sidford, and K. Tian. Variance reduction for matrix games. In Advances in Neural Information Processing Systems (Neur IPS), 2019. [24] N. Cesa-Bianchi and G. Lugosi. Prediction, learning, and games. Cambridge university press, 2006. [25] A. J. Chan and M. van der Schaar. Scalable Bayesian inverse reinforcement learning. ar Xiv:2102.06483, 2021. [26] J. Chang, M. Uehara, D. Sreenivas, R. Kidambi, and W. Sun. Mitigating covariate shift in imitation learning via offline data with partial coverage. Advances in Neural Information Processing Systems (Neuri PS), 2021. [27] A. Charpentier, R. Elie, and C. Remlinger. Reinforcement learning in economics and finance. ar Xiv:20031004, 2020. [28] M. Chen, Y. Wang, T. Liu, Z. Yang, X. Li, Z. Wang, and T. Zhao. On computation and generalization of generative adversarial imitation learning. International Conference on Learning Representations (ICLR), 2020. [29] T. Chen, S. Kornblith, M. Norouzi, and G. Hinton. A simple framework for contrastive learning of visual representations. In International Conference on Machine Learning (ICML), 2020. [30] Y. Chen, L. Li, and M. Wang. Scalable bilinear π learning using state and action features. In International Conference on Machine Learning (ICML), 2018. [31] C.-A. Cheng, R. T. des Combes, B. Boots, and G. Gordon. A reduction from reinforcement learning to no-regret online learning. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2020. [32] R. Dadashi, L. Hussenot, M. Geist, and O. Pietquin. Primal Wasserstein imitation learning. In International Conference on Learning Representations (ICLR), 2021. [33] D. P. De Farias and B. Van Roy. The linear programming approach to approximate dynamic programming. Operations Research, 51(6):850 865, 2003. [34] D. P. De Farias and B. Van Roy. On constraint sampling in the linear programming approach to approximate dynamic programming. Mathematics of Operations Research, 29(3):462 478, 2004. [35] G. T. De Ghellinck and G. D. Eppen. Linear programming solutions for separable Markovian decision problems. Management Science, 13(5):371 394, 1967. [36] E. V. Denardo. On linear programming in a Markov decision problem. Management Science, 16(5):281 288, 1970. [37] Y. Duan, Z. Jia, and M. Wang. Minimax-optimal off-policy evaluation with linear function approximation. In International Conference on Machine Learning (ICML), 2020. [38] J. Fu, K. Luo, and S. Levine. Learning robust rewards with adverserial inverse reinforcement learning. In International Conference on Learning Representations (ICLR), 2018. [39] T. Furmston and D. Barber. Variational methods for reinforcement learning. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2010. [40] D. Garg, S. Chakraborty, C. Cundy, J. Song, and S. Ermon. IQ-learn: Inverse soft-Q learning for imitation. In Advances in Neural Information Processing Systems (Neu RIPS), 2021. [41] M. Geist, B. Scherrer, and O. Pietquin. A Theory of Regularized Markov Decision Processes. In International Conference on Machine Learning (ICML), 2019. [42] A. Geramifard, C. Dann, R. H. Klein, W. Dabney, and J. P. How. RLPy: A value-functionbased reinforcement learning framework for education and research. Journal of Machine Learning Research, 16(46):1573 1578, 2015. [43] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. In Advances in Neural Information Processing Systems (Neur IPS), 2014. [44] T. Haarnoja, A. Zhou, P. Abbeel, and S. Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In International Conference on Machine Learning (ICML), 2018. [45] F. Hanzely, P. Richtarik, and L. Xiao. Accelerated bregman proximal gradient methods for relatively smooth convex optimization. Computational Optimization and Applications, 79(2): 405 440, 2021. [46] B. Hao, T. Lattimore, C. Szepesvári, and M. Wang. Online sparse reinforcement learning. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2021. [47] E. Hazan. Introduction to online convex optimization. Foundations and Trends in Optimization, 2(3-4):157 325, 2016. [48] O. Hernández-Lerma and J. B. Lasserre. Discrete-Time Markov Control Processes: Basic Optimality Criteria. Springer-Verlag New York, 1996. [49] O. Hernández-Lerma and J. B. Lasserre. Further Topics on Discrete-Time Markov Control Processes. Springer-Verlag New York, 1999. [50] J. Ho and S. Ermon. Generative adversarial imitation learning. In Advances in Neural Information Processing Systems (Neur IPS), 2016. [51] J. Ho, J. K. Gupta, and S. Ermon. Model-free imitation learning with policy optimization. In International Conference on Machine Learning (ICML), 2016. [52] S. R. Howard, A. Ramdas, J. Mc Auliffe, and J. Sekhon. Time-uniform, nonparametric, nonasymptotic confidence sequences. The Annals of Statistics, 49(2), 2021. [53] D. Hsu, S. M. Kakade, and T. Zhang. Random design analysis of ridge regression. In Conference on Learning Theory (COLT), 2012. [54] D. Jarrett, I. Bica, and M. van der Schaar. Strictly batch imitation learning by energy-based distribution matching. ar Xiv:2006.14154, 2021. [55] Y. Jin and A. Sidford. Efficiently solving MDPs with stochastic mirror descent. In International Conference on Machine Learning (ICML), 2020. [56] S. Kakade, A. Krishnamurthy, K. Lowrey, M. Ohnishi, and W. Sun. Information theoretic regret bounds for online nonlinear control. Advances in Neural Information Processing Systems (Neur IPS), 2020. [57] G. Kalweit, H. Maria, M. Werling, and J. Boedecker. Deep inverse Q-learning with constraints. In Advances in Neural Information Processing Systems (Neur IPS), 2020. [58] A. Kamoutsi, G. Banjac, and J. Lygeros. Efficient performance bounds for primal-dual reinforcement learning from demonstrations. In International Conference on Machine Learning (ICML), 2021. [59] L. Ke, S. Choudhury, M. Barnes, W. Sun, G. Lee, and S. Srinivasa. Imitation learning as f-divergence minimization. In International Workshop on the Algorithmic Foundations of Robotics (WAFR), 2020. [60] C. Kent, J. Li, J. Blanchet, and P. Glynn. Modified Frank Wolfe in probability space. In Advances in Neural Information Processing Systems (Neur IPS), 2021. [61] D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), 2015. [62] W. B. Knox, A. Allievi, H. Banzhaf, F. Schmitt, and P. Stone. Reward (mis)design for autonomous driving, 2021. [63] I. Kostrikov, K. K. Agrawal, D. Dwibedi, S. Levine, and J. Tompson. Discriminator-actorcritic: Addressing sample inefficiency and reward bias in adversarial imitation learning. In International Conference on Learning Representations (ICLR), 2019. [64] I. Kostrikov, O. Nachum, and J. Tompson. Imitation learning via off-policy distribution matching. In International Conference on Learning Representations (ICLR), 2020. [65] C. Lakshminarayanan, S. Bhatnagar, and C. Szepesvári. A linearly relaxed approximate linear program for Markov decision processes. IEEE Transactions on Automatic Control, 63(4): 1185 1191, 2018. [66] N. Lazic, D. Yin, M. Farajtabar, N. Levine, D. Gorur, C. Harris, and D. Schuurmans. A maximum-entropy approach to off-policy evaluation in average-reward MDPs. Advances in Neural Information Processing Systems (Neur IPS), 2020. [67] D. Lee and N. He. Stochastic primal-dual Q-learning algorithm for discounted MDPs. In American Control Conference (ACC), 2019. [68] S. Levine, Z. Popovi c, and V. Koltun. Feature construction for inverse reinforcement learning. In Advances in Neural Information Processing Systems (Neur IPS), 2010. [69] S. Levine, Z. Popovi c, and V. Koltun. Nonlinear inverse reinforcement learning with Gaussian processes. In Advances in Neural Information Processing Systems (Neur IPS), 2011. [70] Z. Liu, Y. Zhang, Z. Fu, Z. Yang, and Z. Wang. Learning from demonstration: Provably efficient adversarial policy imitation with linear function approximation. In International Conference on Machine Learning (ICML), 2022. [71] F. Lu, P. G. Mehta, S. P. Meyn, and G. Neu. Convex q-learning. In 2021 American Control Conference (ACC), pages 4749 4756, 2021. doi: 10.23919/ACC50511.2021.9483244. [72] Y. Malitsky and M. K. Tam. A forward-backward splitting method for monotone inclusions without cocoercivity. SIAM Journal on Optimization, 30(2):1451 1472, 2020. [73] A. Manne. Linear programming and sequential decisions. Management Science, 6(3):259 267, 1960. [74] A. Martinelli, M. Gargiani, and J. Lygeros. Data-driven optimal control with a relaxed linear program. ar Xiv:2003.08721, 2020. [75] C. Mc Diarmid. Concentration, pages 195 248. Springer Berlin Heidelberg, 1998. [76] P. Mehta and S. Meyn. Q-learning and pontryagin s minimum principle. In IEEE Conference on Decision and Control (CDC), 2009. [77] P. G. Mehta and S. P. Meyn. Convex Q-learning, Part 1: Deterministic optimal control. ar Xiv:2008.03559, 2020. [78] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare, A. Graves, M. Riedmiller, A. K. Fidjeland, G. Ostrovski, S. Petersen, C. Beattie, A. Sadik, I. Antonoglou, H. King, D. Kumaran, D. Wierstra, S. Legg, and D. Hassabis. Human-level control through deep reinforcement learning. Nature, 518(7540):529 533, 2015. [79] P. Mohajerin Esfahani, T. Sutter, D. Kuhn, and J. Lygeros. From infinite to finite programs: explicit error bounds with applications to approximate dynamic programming. SIAM Journal on Optimization, 28(3):1968 1998, 2018. [80] A. Müller. Integral probability metrics and their generating classes of functions. Advances in Applied Probability, 29(2):429 443, 1997. [81] K. S. Narendra and A. M. Annaswamy. Persistent excitation in adaptive systems. International Journal of Control, 45(1):127 160, 1987. [82] Y. Nesterov. Smooth minimization of nonsmooth functions. Math. Programming, 103:127 152, 2005. [83] G. Neu and J. Olkhovskaya. Online learning in mdps with linear function approximation and bandit feedback. In Advances in Neural Information Processing Systems (Neur IPS), 2021. [84] G. Neu and C. Pike-Burke. A unifying view of optimism in episodic reinforcement learning. In Advances in Neural Information Processing Systems (Neur IPS), 2020. [85] G. Neu and C. Szepesvári. Apprenticeship learning using inverse reinforcement learning and gradient methods. In Conference on Uncertainty in Artificial Intelligence (UAI), 2007. [86] G. Neu, A. Jonsson, and V. Gómez. A unified view of entropy-regularized Markov decision processes. ar Xiv:1705.07798, 2017. [87] A. Y. Ng and S. J. Russell. Algorithms for inverse reinforcement learning. In International Conference on Machine Learning (ICML), 2000. [88] T. Osa, J. Pajarinen, G. Neumann, J. Bagnell, P. Abbeel, and J. Peters. An algorithmic perspective on imitation learning. Foundations and Trends in Robotics, 2018. [89] A. Pacchiano, J. Lee, P. Bartlett, and O. Nachum. Near optimal policy optimization via REPS. In Advances in Neural Information Processing Systems (Neur IPS), 2021. [90] J. Peters, K. Mülling, and Y. Altun. Relative entropy policy search. In National Conference on Artificial Intelligence (AAAI), 2010. [91] M. Petrik and S. Zilberstein. Constraint relaxation in approximate linear programs. In International Conference on Machine Learning (ICML), pages 809 816, 2009. [92] M. Petrik, G. Taylor, R. Parr, and S. Zilberstein. Feature selection using regularization in approximate linear programs for Markov decision processes. In International Conference on International Conference on Machine Learning (ICML), 2010. [93] D. A. Pomerleau. Efficient training of artificial neural networks for autonomous navigation. Neural Computation, 3(1):88 97, 1991. [94] M. L. Puterman. Markov Decision Processes: Discrete Stochastic Dynamic Programming. John Wiley & Sons, Inc., USA, 1st edition, 1994. [95] N. D. Ratliff, J. A. Bagnell, and M. A. Zinkevich. Maximum margin planning. In International Conference on Machine Learning (ICML), 2006. [96] S. Reddy, A. D. Dragan, and S. Levine. SQIL: imitation learning via regularized behavioral cloning. ar Xiv:1905.11108, 2019. [97] R. T. Rockafellar. Monotone operators and the proximal point algorithm. SIAM Journal on Control and Optimization, 14(5):877 898, 1976. [98] S. Ross and D. Bagnell. Efficient reductions for imitation learning. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2010. [99] S. Ross, G. Gordon, and D. Bagnell. A reduction of imitation learning and structured prediction to no-regret online learning. In International Conference on Artificial Intelligence and Statistics (AISTATS), 2011. [100] S. Russell. Learning agents for uncertain environments (extended abstract). In Annual Conference on Computational Learning Theory (COLT), 1998. [101] J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov. Proximal policy optimization algorithms. ar Xiv:1707.06347, 2017. [102] P. J. Schweitzer and A. Seidmann. Generalized polynomial approximations in Markovian decision processes. Journal of Mathematical Analysis and Applications, 110(2):568 582, 1985. [103] S. Shalev-Shwartz. Online learning and online convex optimization. Foundations and Trends in Machine Learning, 4(2):107 194, 2012. [104] S. Shalev-Shwartz and S. Ben-David. Understanding Machine Learning: From Theory to Algorithms. Cambridge University Press, 2014. [105] L. Shani, T. Zahavy, and S. Mannor. Online apprenticeship learning. ar Xiv:2102.06924, 2021. [106] R. Shariff and C. Szepesvári. Efficient planning in large MDPs with weak linear function approximation. In Advances in Neural Information Processing Systems (Neur IPS), 2020. [107] M. Sion. On general minimax theorems. Pacific Journal of Mathematics, 8(1):171 176, 1958. [108] A. L. Strehl and M. L. Littman. An analysis of model-based interval estimation for markov decision processes. Journal of Computer and System Sciences, 74(8):1309 1331, 2008. [109] T. Sutter, A. Kamoutsi, P. E. Esfahani, and J. Lygeros. Data-driven approximate dynamic programming: A linear programming approach. In IEEE Conference on Decision and Control (CDC), 2017. [110] R. S. Sutton and A. G. Barto. Reinforcement learning: An introduction. MIT press, second edition, 2018. [111] U. Syed and R. E. Schapire. A game-theoretic approach to apprenticeship learning. In Advances in Neural Information Processing Systems (Neur IPS), 2007. [112] U. Syed, M. Bowling, and R. Schapire. Apprenticeship learning using linear programming. In International Conference on Machine Learning (ICML), 2008. [113] E. Todorov, T. Erez, and Y. Tassa. Mujoco: A physics engine for model-based control. In IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 5026 5033, 2012. [114] N. Vieillard, T. Kozuno, B. Scherrer, O. Pietquin, R. Munos, and M. Geist. Leverage the average: an analysis of kl regularization in reinforcement learning. Advances in Neural Information Processing Systems, 33:12163 12174, 2020. [115] M. Wang. Randomized linear programming solves the Markov decision problem in nearly linear (sometimes sublinear) time. Mathematics of Operations Research, 45(2):517 546, 2020. [116] R. Wang, S. S. Du, L. Yang, and R. R. Salakhutdinov. On reward-free reinforcement learning with linear function approximation. Advances in Neural Information Processing Systems (Neur IPS), 2020. [117] H. Xiao, M. Herman, J. Wagner, S. Ziesche, J. Etesami, and T. H. Linh. Wasserstein adversarial imitation learning. ar Xiv:1906.08113, 2019. [118] T. Xu, Z. Li, and Y. Yu. Error bounds of imitating policies and environments. In Advances in Neural Information Processing Systems (Neur IPS), 2020. [119] S. Yan and N. He. Bregman augmented lagrangian and its acceleration. ar Xiv preprint ar Xiv:2002.06315, 2020. [120] L. Yang and K.-C. Toh. Bregman proximal point algorithm revisited: a new inexact version and its variant. ar Xiv preprint ar Xiv:2105.10370, 2021. [121] L. Yang and M. Wang. Sample-optimal parametric Q-learning using linearly additive features. In International Conference on Machine Learning (ICML), 2019. [122] Y. Zhang, Q. Cai, Z. Yang, and Z. Wang. Generative adversarial imitation learning with neural network parameterization: global optimality and convergence rate. In International Conference on Machine Learning (ICML), 2020. [123] B. D. Ziebart, A. Maas, J. A. Bagnell, and A. K. Dey. Maximum entropy inverse reinforcement learning. In National Conference on Artificial Intelligence (AAAI), 2008. [124] A. Zimin and G. Neu. Online learning in episodic Markovian decision processes by relative entropy policy search. Advances in neural information processing systems (Neur IPS), 2013. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] We reflect the contribution described in the introduction with the theoretical results in Section 4, Section 4.2 (b) Did you describe the limitations of your work? [Yes] We state the assumptions needed for the analysis in Sections 3 and 4.2 (c) Did you discuss any potential negative societal impacts of your work? [N/A] The work is mainly theoretical, we do not foresee potential negative impacts on the society. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] We acknowledge habing read the review guidelines. 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] We state the assumptions needed for the analysis in Section 3 Section 4.2 (b) Did you include complete proofs of all theoretical results? [Yes] The main results are stated in Section 4.2 and the proofs are included as supplementary material. 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] The code is included in the supplementary material. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] Training details are provided in the Appendix. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] We averaged multiple seeds whenever possible computationally. We ran a single seed for the computational expensive Pong environment. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] We specified the resource in the Supplementary material. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] We cite [40] for using their codebase and their expert data. (b) Did you mention the license of the assets? [Yes] We mention the license in the Supplementary. (c) Did you include any new assets either in the supplemental material or as a URL? [Yes] We attach the code to the supplementary. (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [Yes] We discuss this in the supplemnetary material (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] We do not use personal data. 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] We did not involve human partecipants. (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] We did not involve human partecipants. (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A] We did not involve human partecipants.