# rize_adaptive_regularization_for_imitation_learning__6d7724b4.pdf Published in Transactions on Machine Learning Research (11/2025) RIZE: Adaptive Regularization for Imitation Learning Adib Karimi adibkarimi23@aut.ac.ir Department of Computer Engineering Amirkabir University of Technology Mohammad Mehdi Ebadzadeh ebadzadeh@aut.ac.ir Department of Computer Engineering Amirkabir University of Technology Reviewed on Open Review: https: // openreview. net/ forum? id= a6DWq XJZCZ We propose a novel Inverse Reinforcement Learning (IRL) method that mitigates the rigidity of fixed reward structures and the limited flexibility of implicit reward regularization. Building on the Maximum Entropy IRL framework, our approach incorporates a squared temporal-difference (TD) regularizer with adaptive targets that evolve dynamically during training, thereby imposing adaptive bounds on recovered rewards and promoting robust decision-making. To capture richer return information, we integrate distributional RL into the learning process. Empirically, our method achieves expert-level performance on complex Mu Jo Co and Adroit environments, surpassing baseline methods on the Humanoid-v2 task with limited expert demonstrations. Extensive experiments and ablation studies further validate the effectiveness of the approach and provide insights into reward dynamics in imitation learning. Our source code is available at https://github.com/adibka/RIZE. 1 Introduction Designing effective reward functions remains a fundamental challenge in Reinforcement Learning (RL). While dense, well-shaped rewards can facilitate learning, they often require laborious task-specific tuning and domain expertise, limiting their scalability (Amodei et al., 2016; Peng et al., 2020). To circumvent these limitations, researchers have explored alternative approaches, including sparse rewards upon task completion (Silver et al., 2016), learning from expert trajectories (Schaal, 1996; Ng & Russell, 2000; Osa et al., 2018), human preference-based reward modeling (Christiano et al., 2017; Lee et al., 2021; Hejna & Sadigh, 2023), and intrinsically motivated RL (Oudeyer et al., 2007; Schmidhuber, 2010; Colas et al., 2022). Among these, Inverse Reinforcement Learning (IRL) (Abbeel & Ng, 2004) offers a compelling alternative by inferring reward functions directly from expert demonstrations, bypassing manual reward engineering. IRL has driven breakthroughs in robotics (Osa et al., 2018), autonomous driving (Knox et al., 2023), and drug discovery (Ai et al., 2024). A prominent framework in IRL is Maximum Entropy (Max Ent) IRL (Ziebart, 2010), which underpins many state-of-the-art (SOTA) IL methods. Prior works have combined Max Ent IRL with adversarial training (Ho & Ermon, 2016; Fu et al., 2018) to minimize divergences between agent and expert distributions. However, these adversarial methods often suffer from instability during training. To address this, recent research has introduced implicit reward regularization, which indirectly represents rewards via Q-values by inverting the Bellman equation. For instance, IQ-Learn (Garg et al., 2021) unifies reward and policy representations using Q-functions with an L2-norm regularization on rewards, while LSIQ (Al-Hafez et al., 2023) minimizes the chisquared divergence between expert and mixture distributions, resulting in a squared temporal difference (TD) error objective analogous to SQIL (Reddy et al., 2020). Despite its effectiveness, this method has limitations: Corresponding author. Published in Transactions on Machine Learning Research (11/2025) LSIQ assigns fixed targets for implicit rewards (e.g., +1 for expert samples and -1 for agent samples), which constrains flexibility by treating all tasks and state-action pairs uniformly, limiting performance and requiring additional gradient steps for convergence. We propose an extension of implicit reward regularization under the Max Ent IRL framework, introducing two key advancements. Adaptive Targets: We enhance prior TD-error regularization by introducing learnable targets λπE and λπ that dynamically adjust during training. These targets replace static constraints with context-sensitive reward alignment, preventing rewards from over-increasing or over-decreasing. Crucially, our theoretical analysis reveals that these adaptive bounds constrain implicit rewards to well-defined ranges 1 2c + min {λπE, λπ} , 1 2c + max {λπE, λπ} , where c is the regularization coefficient. Since implicit rewards derive from Q-values, this regularization indirectly stabilizes policy training. Distributional RL Integration: We incorporate return distributions Zπ(s, a) (Bellemare et al., 2017) to capture richer uncertainty information in returns, while using their expectations for policy optimization. Though distributional RL has shown success in adversarial IRL (Zhou et al., 2023), it remains unexplored in non-adversarial settings. Our work bridges this gap, demonstrating its efficacy in Max Ent IRL while preserving theoretical guarantees. Unifying these advances, our framework outperforms IL baselines on Mu Jo Co (Todorov et al., 2012) and Adroit (Rajeswaran et al., 2018) benchmarks. Notably, our approach shows clear benefits on complex tasks like Humanoid-v2 and Hammer-v1, where adaptive targets and distributional learning improve training stability, as confirmed by our ablations. Our contributions are threefold. First, we introduce adaptive targets for implicit reward regularization, enabling dynamic reward bounds that enhance stability and alignment during training. Second, we integrate return distributions into implicit reward frameworks, capturing richer uncertainty information while preserving theoretical consistency. Third, we empirically validate our approach through extensive experiments on Mu Jo Co and Adroit tasks, demonstrating superior performance in complex environments. 2 Related Work Imitation learning (IL) and inverse reinforcement learning (IRL) (Watson et al., 2023) are foundational paradigms for training agents to mimic expert behavior from demonstrations. Behavioral Cloning (BC) (Pomerleau, 1991), the simplest IL approach, treats imitation as a supervised learning problem by directly mapping states to expert actions. While computationally efficient, BC is prone to compounding errors (Ross & Bagnell, 2011) due to covariate shift during deployment. The Maximum Entropy IRL framework (Ziebart, 2010) addresses this limitation by probabilistically modeling expert behavior as reward maximization under an entropy regularization constraint, establishing a theoretical foundation for modern IRL methods. The advent of adversarial training marked a pivotal shift in IL methodologies. Ho & Ermon (2016) introduced Generative Adversarial Imitation Learning (GAIL), which formulates imitation learning as a generative adversarial game (Goodfellow et al., 2014) where an agent learns a policy indistinguishable from the expert s by minimizing the Jensen Shannon divergence between their state action distributions. This framework was generalized by Ghasemipour et al. (2019) in f-GAIL, which replaces the Jensen Shannon divergence with arbitrary (f)-divergences to broaden applicability. Concurrently, Kostrikov et al. (2019) proposed Discriminator Actor Critic (DAC), improving sample efficiency via off-policy updates and terminal-state reward modeling while mitigating reward bias. Most recently, Chang et al. (2024) cast adversarial imitation as a boosting procedure: AILBoost maintains an ensemble of weighted policies and trains the discriminator against a weighted replay buffer approximating the ensemble s occupancy, enabling fully off-policy training and reporting gains over DAC on Deep Mind Control tasks (Tassa et al., 2018). Recent advances have shifted toward methods that bypass explicit reward function estimation. Kostrikov et al. (2020) introduced Value DICE, an offline IL method that leverages an inverse Bellman operator to avoid adversarial optimization. Similarly, Garg et al. (2021) developed IQ-Learn, which circumvents the challenges of Max Ent IRL by optimizing implicit rewards derived directly from expert Q-values. A parallel research direction simplifies reward engineering by assigning fixed rewards to expert and agent samples. Reddy et al. (2020) pioneered this approach with Soft Q Imitation Learning (SQIL), which assigns binary rewards to transitions from expert and agent trajectories. Most recently, Al-Hafez et al. (2023) proposed Least Squares Inverse Q-Learning (LSIQ), enhancing regularization by minimizing chi-squared divergence between expert Published in Transactions on Machine Learning Research (11/2025) and mixture distributions while explicitly managing absorbing states through critic regularization. In the same spirit of avoiding adversarial training, Coherent Soft Imitation Learning (CSIL) (Watson et al., 2023) inverts the soft policy update to derive an explicit shaped coherent reward from a behavior-cloned policy and then fine-tunes the policy with standard RL using online or offline data. Orthogonally, Jain et al. (2025) introduce Successor Feature Matching, a non-adversarial IRL method that forgoes explicit reward learning by directly matching expert and learner successor features, and notably supports state-only demonstrations. Complementing these, Wu et al. (2025) propose a diffusion-based framework that learns score functions on expert and learner states and optimizes a score-difference cost (a diffusion score divergence), offering a nonadversarial alternative that also works with state-only demos. Our work builds on IQ-Learn (Garg et al., 2021) and LSIQ (Al-Hafez et al., 2023) by applying a squared TD regularizer with adaptive targets and by employing an Implicit Quantile Network (IQN) critic (Dabney et al., 2018a). 3 Background 3.1 Preliminary We consider a Markov Decision Process (MDP) (Puterman, 2014) to model policy learning in Reinforcement Learning (RL). The MDP framework is defined by the tuple S, A, p0, P, R, γ , where S denotes the state space, A the action space, p0 the initial state distribution, P : S A S [0, 1] the transition kernel with P( | s, a) specifying the likelihood of transitioning from state s given action a, R : S A R the reward function, and γ [0, 1] the discount factor which tempers future rewards. A stationary policy π Π is characterized as a mapping from states s S to distributions over actions a A. The primary objective in RL (Sutton & Barto, 2018) is to maximize the expected sum of discounted rewards, expressed as Eπ [P t=0 γt R(st, at)]. Furthermore, the occupancy measure ρπ(s, a) for a policy π Π is given by (1 γ)π(a | s) P t=0 γt P(st = s | π). The corresponding measure for an expert policy, πE, is similarly denoted by ρE. In Imitation Learning (IL), the expert policy πE is typically unknown, and only a finite set of expert demonstrations is available, rather than explicit reward feedback from the environment. 3.2 Distributional Reinforcement Learning Maximum Entropy (Max Ent) RL (Haarnoja et al., 2018) focuses on addressing the stochastic nature of action selection by maximizing the entropy of the policy, while Distributional RL (Bellemare et al., 2017) emphasizes capturing the inherent randomness in returns. Combining these perspectives, the distributional soft value function Z : S A Z (Ma et al., 2020) for a policy π Π encapsulates uncertainty in both rewards and actions, with Z representing the space of return distributions. It is formally defined as: t=0 γt[R(st, at) + αH(π( | st))], (1) where H(π) = Eπ[ log π(a | s)] denotes the entropy of the policy, and α > 0 balances entropy with reward. The distributional soft Bellman operator Bπ D : Z Z for a given policy π is introduced as (Bπ DZ)(s, a) D= R(s, a) + γ[Z(s , a ) α log π(a | s )], where s P( | s, a), a π( | s ), and D= signifies equality in distribution. Notably, this operator exhibits contraction properties under the p-Wasserstein metric, ensuring convergence to a unique fixed point, the distributional soft return function for the given policy. A practical approach to approximating the return distribution Z involves modeling its quantile function F 1 Z (τ), evaluated at specific quantile levels τ [0, 1] (Dabney et al., 2018b). The quantile function is defined as F 1 Z (τ) = inf{z R : τ FZ(z)}, where FZ(z) = P(Z z) is the cumulative distribution function of Z. For simplicity, we denote the quantile-based representation as Zτ(s, a) := F 1 Z (τ). To discretize this representation, we define a sequence of quantile levels, denoted as {τi}i=0,...,N 1, where 0 = τ0 < ... < τN 1 = 1. These quantiles partition the unit interval into N fractions. For uniformly sampled quantiles τ U(0, 1), Zτ(s, a) denotes the τ-quantile (scalar) of the return distribution. Published in Transactions on Machine Learning Research (11/2025) 3.3 Inverse Reinforcement Learning Given expert trajectory data, Maximum Entropy (Max Ent) Inverse RL (Ziebart, 2010) aims to infer a reward function R(s, a) from the family R = RS A. Instead of assuming a deterministic expert policy, this method optimizes for stochastic policies π Π that maximize R while matching expert behavior. GAIL (Ho & Ermon, 2016) extends this framework by introducing a convex reward regularizer ψ : RS A R, leading to the adversarial objective: max R R min π Π L(π, R) = EρE[R(s, a)] Eρπ[R(s, a)] H(π) ψ(R) . (2) IQ-Learn (Garg et al., 2021) departs from adversarial training by implicitly representing rewards through Q-functions Q Ω(Piot et al., 2014). It leverages the inverse soft Bellman operator T π, defined as: (T πQ)(s, a) = Q(s, a) γEs P ( |s,a),a π( |s )[Q(s , a ) α log π(a |s )]. (3) For a fixed policy π, T π is bijective, ensuring a one-to-one correspondence between Q-values and rewards: T πQ = R and Q = (T π) 1R. This allows reframing the Max Ent IRL objective (2) in Q-policy space as max Q Ωminπ Π J (π, Q). IQ-Learn simplifies the problem by defining the implicit reward RQ(s, a) = T πQ(s, a) and applying an L2 regularizer ψ(RQ). The final objective becomes: max Q Ωmin π Π J (π, Q) = EρE[RQ(s, a)] Eρπ[RQ(s, a)] αH(π) c h EρE[RQ(s, a)2] + Eρπ[RQ(s, a)2] i . (4) 4 Methodology This section introduces a framework that integrates Distributional Reinforcement Learning with Inverse RL. We first show how return distributions can replace point-estimate critics. We then propose an adaptive regularization technique for implicit rewards and analyze its properties. Finally, we derive RIZE, which combines distributional critics with bounded-reward imitation learning. 4.1 Distributional Value Integration Our approach departs from traditional imitation learning by explicitly using return distributions as critics within an actor-critic framework (Zhou et al., 2023). We argue that learning the soft return distribution Z(s, a) in Equation (1), rather than relying solely on point estimates like Q(s, a), enhances decision-making by capturing uncertainty in complex environments. This view is consistent with recent neuroscience findings suggesting that decision-making in the prefrontal cortex relies on learning distributions over outcomes rather than only their expectations (Muller et al., 2024). Moreover, access to the full return distribution enables the computation of statistical moments most notably the expectation which we optimize both policy and critic using the expectation of Z (Bellemare et al., 2017; Dabney et al., 2018b;a), yielding a more robust learning signal while remaining compatible with IQ-Learn. We compute Q as the expectation of the soft return distribution: i=0 (τi+1 τi) Zτi(s, a) , (5) where {τi} are quantile fractions and Zτi(s, a) denotes the corresponding quantile values (see Lemma A.1). 4.2 Implicit Reward Regularization In this section, we propose a regularizer for inverse RL that refines existing implicit reward formulations (Garg et al., 2021). The implicit reward is defined as: RQ(s, a) = Q(s, a) γEP,π [Q(s , a ) α log π(a |s )] . (6) Published in Transactions on Machine Learning Research (11/2025) Previous works typically regularize implicit rewards either using L2-norms (Garg et al., 2021) or by treating them as squared-TD errors between rewards and fixed targets (Reddy et al., 2020; Al-Hafez et al., 2023). While we adopt a similar squared-TD setting, we introduce adaptive targets λπE (for the expert πE) and λπ (for the imitation policy π) to construct our convex regularizer Γ : RS A R: Γ(RQ, λ) = EρE (RQ(s, a) λπE)2 + Eρπ (RQ(s, a) λπ)2 . (7) These targets self-update through a feedback loop where reward estimates continuously adapt to match moving targets: min λπE EρE h (RQ(s, a) λπE)2i , min λπ Eρπ h (RQ(s, a) λπ)2i . (8) Substituting Γ(RQ, λ) for the L2 term in Equation (4) and using Equation (5) to compute Q, we obtain: L(π, Q) = EρE[RQ(s, a)] Eρπ[RQ(s, a)] αH(π) c h EρE (RQ(s, a) λπE)2 + Eρπ (RQ(s, a) λπ)2 i , (9) where c is the regularization coefficient. As in IQ-Learn and LSIQ, our method seeks behavior indistinguishable from expert demonstrations in a non-adversarial implicit-reward setting. Prior work analyzes optimal implicit rewards, drawing on Max Min analyses from GANs (Al-Hafez et al., 2023; Goodfellow et al., 2014). Because we bind rewards to adaptive targets via Γ(RQ, λ) (Equation (7)), we carry out an analogous analysis, stated below. Proposition 4.1. Let RQ(s, a) = (T πQ)(s, a) denote the implicit reward derived from point-estimate Qvalues, where Q(s, a) = E[Z(s, a)]. Let ρE(s, a) and ρπ(s, a) denote occupancy measures under πE and π, respectively. For fixed π, the optimal TD-regularized reward satisfies: R Q(s, a) = ρE(s, a) ρπ(s, a) (2c) (ρE(s, a) + ρπ(s, a)) + ρE(s, a)λπE + ρπ(s, a)λπ ρE(s, a) + ρπ(s, a) . (10) Proof. Differentiating L(π, Q) in Equation (9) with respect to RQ(s, a) (holding π fixed) and setting the result to zero yields 0 = ρE ρπ 2c ρE(RQ λπE) + ρπ(RQ λπ) , (11) from which Equation (10) follows. Corollary 4.2. The optimal implicit reward satisfies: R Q(s, a) 1 2c + λmin, 1 where λmin := min {λπE, λπ} and λmax := max {λπE, λπ}. The coefficient c and adaptive targets bound rewards within a well-defined range. Proof. Considering the optimal reward Equation (10), the first term lies in 1 2c, 1 2c since ρE, ρπ 0 (achieved when either ρπ 0 or ρE 0). The second term is a convex combination of λπE and λπ, thus in [λmin, λmax]. Combining intervals yields the result. In practice, we use λ [5, 10] and c [0.1, 0.5], which we found to promote stable training (see Appendix C.7). Moreover, when the occupancy measures match, the optimal reward simplifies significantly, as stated in Corollary A.2: R Q(s, a) = λπE = λπ . (13) Recalling that we represent rewards through Q-values, the boundedness of the optimal reward ensures that critic updates remain constrained and because the policy directly depends on these critic values promotes stable policy optimization. However, as previously noted (Al-Hafez et al., 2023; Viano et al., 2022), the convergence guarantee originally stated for IQ-Learn (Garg et al., 2021) does not extend to the χ2-regularizer used in practice. Consequently, a formal proof for the resulting alternating SAC updates remains an open question and is left for future work. Published in Transactions on Machine Learning Research (11/2025) 0.25 0.50 0.75 1.00 RIZE IQ SQIL LSIQ 0.25 0.50 0.75 1.00 0.2 0.4 0.6 0.8 0.2 0.4 0.6 0.8 Optimality Gap 0.4 0.6 0.8 1.0 RIZE IQ SQIL LSIQ 0.25 0.50 0.75 1.00 0.4 0.6 0.8 1.0 0.2 0.4 0.6 Optimality Gap Figure 1: RLiable (Agarwal et al., 2021) plots for RIZE vs. BC, LSIQ, SQIL, CSIL, and IQ-Learn on six Mu Jo Co/Adroit tasks. For each setting (3 demos; 10 demos), we report aggregate Median, IQM, Mean, and Optimality Gap with 95% confidence intervals computed via percentile bootstrap stratified over tasks and five seeds. Scores are normalized to expert performance. Higher is better for Median, IQM, and Mean; lower is better for Optimality Gap. 4.3 Practical Algorithm We now introduce RIZE: Adaptive Regularization for Imitation Learning (Algorithm 1). We approximate Z and π with neural networks, optimizing via Q = E[Z]. Following Distributional SAC (Ma et al., 2020), we employ target policies for next-action sampling and a double-critic architecture with target networks for stability. Lower learning rates proved critical for robust policy updates, with a four-layer MLP policy network necessary for complex tasks. Targets λπE and λπ are optimized to match expected rewards. We initialize λπE higher than λπ and update λπ with lower learning rates due to its sensitivity. Algorithm 1 RIZE 1: Initialize Zϕ, πθ, λπE, and λπ 2: for step t in {1, . . . , N} do 3: Calculate Q(s, a) = E[Zϕ(s, a)] using Eq. 5 4: Update Zϕ using Eq. 9 5: ϕt+1 ϕt βZ ϕ[ L(ϕ)] 6: Update πθ (like SAC) θt+1 θt + βπ θE s D, a πθ( |s) [ min k=1,2 Qk(s, a) α log πθ(a|s)] 8: Update λπ and λπE using Eq. 8 9: λπ t+1 λπ t βλπ λπΓ(RQ, λ) 10: λπE t+1 λπE t βλπE λπE Γ(RQ, λ) 11: end for 5 Experiments We study continuous-control imitation learning from state action expert samples, evaluating our algorithm on five Mu Jo Co (Todorov et al., 2012) benchmarks (Half Cheetah-v2, Walker2d-v2, Ant-v2, Humanoid-v2, Hopper-v2) and one Adroit Hand task (Hammer-v1). We compare against state-of-the-art baselines IQ-Learn (Garg et al., 2021), LSIQ (Al-Hafez et al., 2023), SQIL (Reddy et al., 2020), CSIL (Watson et al., 2023) and Behavior Cloning (BC) (Pomerleau, 1991). All experiments use five random seeds (Henderson et al., 2018). We assess each method with three and ten expert trajectories. We report mean 95% confidence intervals (CI) across five seeds. Episode returns are normalized by expert performance. For implementation details, additional experimental results, and visualizations, see Appendix B and C. Main Results. Figure 1 reports aggregate metrics computed with RLiable (Agarwal et al., 2021), which provides statistically principled evaluation via Median, Interquartile Mean (IQM), Mean, and Optimality Gap with stratified bootstrap confidence intervals; we adopt this protocol throughout, consistent with recent IL practice (e.g., AILBoost (Chang et al., 2024); SFM (Jain et al., 2025)). Across both 3and 10-demonstration settings, RIZE achieves higher Median, IQM, and Mean and a lower Optimality Gap than the baselines; the Published in Transactions on Machine Learning Research (11/2025) 3 10 Trajectories Half Cheetah-v2 3 10 Trajectories Walker2d-v2 3 10 Trajectories 3 10 Trajectories 3 10 Trajectories Humanoid-v2 3 10 Trajectories BC LSIQ SQIL IQLearn CSIL RIZE Expert Figure 2: Normalized returns on Mu Jo Co and Adroit tasks for RIZE and baselines. We first compute, per seed, the average episodic return over the final third of training steps; bars show the mean across five seeds and error bars denote the 95% confidence interval. Returns are normalized to expert performance and reported for both 3 and 10 expert demonstrations. 100 200 300 Half Cheetah-v2 100 200 300 Walker2d-v2 100 200 300 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) Humanoid-v2 100 300 500 Steps ( 103) BC LSIQ SQIL IQLearn CSIL RIZE Expert Figure 3: Learning curves on Mu Jo Co and Adroit tasks with 10 expert demonstrations. Lines show the mean normalized return across five seeds; shaded regions denote 95% confidence intervals. only competitive methods overall are CSIL and IQ-Learn. Increasing the number of expert trajectories from 3 to 10 consistently improves RIZE on all RLiable aggregates. Task-wise, RIZE is the only method that solves Humanoid-v2 (all baselines fail), while on Hammer-v1 RIZE, CSIL, and BC show higher returns than the remaining methods. In terms of sample efficiency, CSIL often attains strong returns early benefiting from behavior-cloning initialization but when BC is ineffective (e.g., Humanoid-v2), CSIL underperforms whereas RIZE ultimately succeeds. These trends are corroborated by the bar plot in Figure 2 (average of the final third of training) and the learning curves in Figure 3. For learning curves with three demonstrations, see Figure 6. Recovered Rewards. In this section, we assess whether our choice of regularizer Γ(RQ, λ) in Equation (7) with adaptive targets can effectively bound the implicit rewards. By Corollary 4.2 together with Equation (12), the optimal implicit reward is confined to the interval 1 2c + λmin, 1 2c + λmax . Since expert rewards are supposed to be higher than those of the learner during training (based on IRL objective Published in Transactions on Machine Learning Research (11/2025) 100 200 300 Half Cheetah-v2 100 200 300 Walker2d-v2 100 200 300 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) Humanoid-v2 100 300 500 Steps ( 103) Expert reward Policy reward Expert + 1 2c Policy 1 2c Figure 4: Implicit reward curves for expert and policy samples on Mu Jo Co and Adroit tasks with 10 expert demonstrations. Each subplot reports the mean across five seeds, with shaded regions showing the 95% confidence interval. Theoretical upper and lower bounds derived in this work are overlaid as separate curves in each subplot. loss 4), and the adaptive targets track these levels, we empirically observe λπE λπ; hence λmax = λπE and λmin = λπ. In practice, we should see rewards confined to 1 2c + λπ, 1 2c + λπE . Figure 4 depicts the recovered reward trajectories together with the theoretical upper and lower bounds. Across tasks, the curves remain within the predicted band, indicating that our regularizer with adaptive targets effectively bounds both expert and policy rewards. We also observe that the position and width of the band vary with the task and data regime, reflecting the flexibility of the adaptive targets to adjust to different dynamics. In contrast, fixed (non-adaptive) targets lack this ability, leading either to overly loose bounds or to target reward mismatch. Overall, these results support the role of Γ(RQ, λ) and the adaptive targets in producing stable, well-calibrated implicit rewards that align with Corollary 4.2. For the 3-demonstration setting, Figure 7 shows reward trajectories with their theoretical bounds; for cross-method comparisons of reward trajectories, see Figures 9, 8, and 10a. Ablation on Critic Architecture. To assess the effect of modeling the return distribution, we replace the IQN-based critic Z(s, a) (Dabney et al., 2018a) in RIZE with a point-estimate Q(s, a) critic and compare against the original implementation. Here, Z(s, a) denotes the full distribution of discounted returns whose expectation yields Q(s, a). As shown in Figure 5, the IQN critic consistently outperforms the Q-network across all tasks, exhibiting lower variance and greater sample efficiency. Notably, on Humanoid-v2 and Hammer-v1, the Q-network fails to match expert performance, whereas the IQN critic maintains expert-level returns. See Figure 13 for reward trajectories with theoretical bounds when using a Q-network in place of the IQN critic, and Figures 11, 12, and 10b for critic value estimation curves on Mu Jo Co and Adroit tasks. 6 Conclusion We propose a novel IRL framework that overcomes the limitations of fixed reward mechanisms through dynamic reward adaptation and context-sensitive regularization. Our approach ensures bounded implicit rewards and stable value function updates, leading to robust policy optimization. By integrating distributional Published in Transactions on Machine Learning Research (11/2025) 100 200 300 Half Cheetah-v2 100 200 300 Walker2d-v2 100 200 300 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) Humanoid-v2 100 300 500 Steps ( 103) RIZE (with Z(s,a)) RIZE (with Q(s,a)) Expert Figure 5: Ablation on critic architecture: Z(s, a) via Implicit Quantile Networks (IQN) (Dabney et al., 2018a) versus classic Q(s, a). We report expert normalized returns across all Mu Jo Co and Adroit tasks using three expert demonstrations; metrics show the mean over five seeds with 95% confidence intervals. RL with implicit reward learning, we capture richer return dynamics while preserving theoretical guarantees. Empirical results on Mu Jo Co and Adroit benchmarks show expert-like proficiency on the Humanoid-v2 and Hammer-v1 tasks with three expert demonstrations. Ablation studies confirm the important role of our regularization mechanism. This work unifies implicit reward regularization with distributional return representations, offering a scalable and sample-efficient solution for complex decision-making. Future directions include extending this framework to offline IL and risk-sensitive robotics, where TD regularizer with adaptive targets and uncertainty-aware return distributions can further improve robustness and generalization. P. Abbeel and A. Y. Ng. Apprenticeship learning via inverse reinforcement learning. In International Conference on Machine Learning, 2004. Rishabh Agarwal, Max Schwarzer, Pablo Samuel Castro, Aaron Courville, and Marc G Bellemare. Deep reinforcement learning at the edge of the statistical precipice. In A. Beygelzimer, Y. Dauphin, P. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, 2021. URL https: //openreview.net/forum?id=uqv8-U4l KBe. Cheng Ai, Hong Yang, Xinyu Liu, Ruijie Dong, Ying Ding, and Fei Guo. Mtmol-gpt: De novo multi-target molecular generation with transformer-based generative adversarial imitation learning. PLo S Computational Biology, 20(6):e1012229, 2024. doi: 10.1371/journal.pcbi.1012229. URL https://doi.org/10. 1371/journal.pcbi.1012229. Firas Al-Hafez, Davide Tateo, Oleg Arenz, Guoping Zhao, and Jan Peters. LS-IQ: Implicit reward regularization for inverse reinforcement learning. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=o3Q4m8jg4BR. Dario Amodei, Chris Olah, Jacob Steinhardt, Paul F. Christiano, John Schulman, and Dan Mané. Concrete problems in AI safety. Co RR, abs/1606.06565, 2016. URL http://arxiv.org/abs/1606.06565. Alex Ayoub, Kaiwen Wang, Vincent Liu, Samuel Robertson, James Mcinerney, Dawen Liang, Nathan Kallus, and Csaba Szepesvari. Switching the loss reduces the cost in batch reinforcement learning. In Ruslan Published in Transactions on Machine Learning Research (11/2025) Salakhutdinov, Zico Kolter, Katherine Heller, Adrian Weller, Nuria Oliver, Jonathan Scarlett, and Felix Berkenkamp (eds.), Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 2135 2158. PMLR, 21 27 Jul 2024. URL https: //proceedings.mlr.press/v235/ayoub24a.html. Marc G. Bellemare, Will Dabney, and Rémi Munos. A distributional perspective on reinforcement learning. In International Conference on Machine Learning, 2017. URL https://api.semanticscholar.org/ Corpus ID:966543. Greg Brockman, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, and Wojciech Zaremba. Open AI Gym. ar Xiv preprint ar Xiv:1606.01540, 2016. Jonathan Daniel Chang, Dhruv Sreenivas, Yingbing Huang, Kianté Brantley, and Wen Sun. Adversarial imitation learning via boosting. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=Du Qkq Se9en. Paul F Christiano, Jan Leike, Tom Brown, Miljan Martic, Shane Legg, and Dario Amodei. Deep reinforcement learning from human preferences. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL https://proceedings.neurips.cc/paper_files/paper/ 2017/file/d5e2c0adad503c91f91df240d0cd4e49-Paper.pdf. Cédric Colas, Tristan Karch, Olivier Sigaud, and Pierre-Yves Oudeyer. Autotelic agents with intrinsically motivated goal-conditioned reinforcement learning: A short survey. Journal of Artificial Intelligence Research, 74, July 2022. doi: 10.1613/jair.1.13554. URL https://hal.science/hal-03901771. Will Dabney, Georg Ostrovski, David Silver, and Remi Munos. Implicit quantile networks for distributional reinforcement learning. In Proceedings of the 35th International Conference on Machine Learning, Proceedings of Machine Learning Research, pp. 1096 1105. PMLR, 2018a. Will Dabney, Mark Rowland, Marc G. Bellemare, and Remi Munos. Distributional reinforcement learning with quantile regression. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, AAAI 18/IAAI 18/EAAI 18. AAAI Press, 2018b. ISBN 978-1-57735-800-8. Jesse Farebrother, Jordi Orbay, Quan Vuong, Adrien Ali Taiga, Yevgen Chebotar, Ted Xiao, Alex Irpan, Sergey Levine, Pablo Samuel Castro, Aleksandra Faust, Aviral Kumar, and Rishabh Agarwal. Stop regressing: Training value functions via classification for scalable deep RL. In Forty-first International Conference on Machine Learning, 2024. URL https://openreview.net/forum?id=d Vp FKfq F3R. Justin Fu, Katie Luo, and Sergey Levine. Learning robust rewards with adversarial inverse reinforcement learning. In International Conference on Learning Representations, 2018. Justin Fu, Aviral Kumar, Ofir Nachum, George Tucker, and Sergey Levine. D4rl: Datasets for deep datadriven reinforcement learning, 2021. URL https://arxiv.org/abs/2004.07219. Divyansh Garg, Shuvam Chakraborty, Chris Cundy, Jiaming Song, and Stefano Ermon. Iq-learn: Inverse soft-q learning for imitation. In Advances in Neural Information Processing Systems, volume 34, 2021. Seyed Kamyar Seyed Ghasemipour, Richard Zemel, and Shixiang Gu. A divergence minimization perspective on imitation learning methods. In Conference on Robot Learning, 2019. Ian J Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial networks. In Advances in Neural Information Processing Systems, volume 27, 2014. Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In International Conference on Machine Learning, 2018. Published in Transactions on Machine Learning Research (11/2025) Joey Hejna and Dorsa Sadigh. Inverse preference learning: Preference-based RL without a reward function. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview. net/forum?id=g AP52Z2dar. Peter Henderson, Riashat Islam, Philip Bachman, Joelle Pineau, Doina Precup, and David Meger. Deep reinforcement learning that matters. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, AAAI 18/IAAI 18/EAAI 18. AAAI Press, 2018. ISBN 978-1-57735-800-8. Jonathan Ho and Stefano Ermon. Generative adversarial imitation learning. In Advances in Neural Information Processing Systems, volume 29, 2016. Arnav Kumar Jain, Harley Wiltzer, Jesse Farebrother, Irina Rish, Glen Berseth, and Sanjiban Choudhury. Non-adversarial inverse reinforcement learning via successor feature matching. In The Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id= Lv RQgsvd5V. W. Bradley Knox, Alessandro Allievi, Holger Banzhaf, Felix Schmitt, and Peter Stone. Reward (mis)design for autonomous driving. Artif. Intell., 316(C), March 2023. ISSN 0004-3702. doi: 10.1016/j.artint.2022. 103829. URL https://doi.org/10.1016/j.artint.2022.103829. Ilya Kostrikov, Kumar Krishna Agrawal, Debidatta Dwibedi, Sergey Levine, and Jonathan Tompson. Discriminator-actor-critic: Addressing sample inefficiency and reward bias in adversarial imitation learning. In International Conference on Learning Representations, 2019. Ilya Kostrikov, Ofir Nachum, and Jonathan Tompson. Imitation learning via off-policy distribution matching. In International Conference on Learning Representations, 2020. Kimin Lee, Laura Smith, and Pieter Abbeel. Pebble: Feedback-efficient interactive reinforcement learning via relabeling experience and unsupervised pre-training. In Lee, Kimin and Smith, Laura and Abbeel, Pieter, 2021. Xiaoteng Ma, Qiyuan Zhang, Li Xia, Zhengyuan Zhou, Jun Yang, and Qianchuan Zhao. Dsac: Distributional soft actor critic for risk-sensitive reinforcement learning. ar Xiv, 2020. Timothy H. Muller, James L. Butler, Sebastijan Veselic, Bruno Miranda, Joni D. Wallis, Peter Dayan, Timothy E. J. Behrens, Zeb Kurth-Nelson, and Steven W. Kennerley. Distributional reinforcement learning in prefrontal cortex. Nature Neuroscience, 27(3):403 408, mar 2024. ISSN 1546-1726. doi: 10.1038/ s41593-023-01535-w. URL https://doi.org/10.1038/s41593-023-01535-w. Andrew Y. Ng and Stuart J. Russell. Algorithms for inverse reinforcement learning. In International Conference on Machine Learning, 2000. Takayuki Osa, Joni Pajarinen, Gerhard Neumann, J. Andrew Bagnell, Pieter Abbeel, and Jan Peters. An algorithmic perspective on imitation learning. Foundations and Trends in Robotics, 2018. Pierre-Yves Oudeyer, Frdric Kaplan, and Verena V. Hafner. Intrinsic motivation systems for autonomous mental development. IEEE Transactions on Evolutionary Computation, 11(2):265 286, 2007. doi: 10. 1109/TEVC.2006.890271. Xue Bin Peng, Erwin Coumans, Tingnan Zhang, Tsang-Wei Edward Lee, Jie Tan, and Sergey Levine. Learning agile robotic locomotion skills by imitating animals. Co RR, abs/2004.00784, 2020. URL https: //arxiv.org/abs/2004.00784. B. Piot, M. Geist, and O. Pietquin. Boosted and Reward-regularized Classification for Apprenticeship Learning. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems (AAMAS), 2014. Published in Transactions on Machine Learning Research (11/2025) D. A. Pomerleau. Efficient training of artificial neural networks for autonomous navigation. Neural Computation, 3(1):88 97, 1991. Martin L. Puterman. Markov Decision Processes: Discrete Stochastic Dynamic Programming. John Wiley & Sons, 2014. Aravind Rajeswaran, Vikash Kumar, Abhishek Gupta, Giulia Vezzani, John Schulman, Emanuel Todorov, and Sergey Levine. Learning complex dexterous manipulation with deep reinforcement learning and demonstrations, 2018. URL https://arxiv.org/abs/1709.10087. Siddharth Reddy, Anca D Dragan, and Sergey Levine. Sqil: Imitation learning via reinforcement learning with sparse rewards. In International Conference on Learning Representations, 2020. Stéphane Ross and J. Andrew Bagnell. A reduction of imitation learning and structured prediction to no-regret online learning. In International Conference on Artificial Intelligence and Statistics, 2011. Stefan Schaal. Learning from demonstration. In M.C. Mozer, M. Jordan, and T. Petsche (eds.), Advances in Neural Information Processing Systems, volume 9. MIT Press, 1996. URL https://proceedings. neurips.cc/paper_files/paper/1996/file/68d13cf26c4b4f4f932e3eff990093ba-Paper.pdf. Jürgen Schmidhuber. Formal theory of creativity, fun, and intrinsic motivation (1990 2010). IEEE Transactions on Autonomous Mental Development, 2(3):230 247, 2010. doi: 10.1109/TAMD.2010.2056368. David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, et al. Mastering the game of go with deep neural networks and tree search. Nature, 529(7587):484 489, 2016. doi: 10.1038/nature16961. URL https://doi.org/10.1038/nature16961. Richard S. Sutton and Andrew G. Barto. Reinforcement Learning: An Introduction. The MIT Press, Cambridge, Massachusetts; London, England, second edition, 2018. Yuval Tassa, Yotam Doron, Alistair Muldal, Tom Erez, Yazhe Li, Diego de Las Casas, David Budden, Abbas Abdolmaleki, Josh Merel, Andrew Lefrancq, Timothy Lillicrap, and Martin Riedmiller. Deepmind control suite, 2018. URL https://arxiv.org/abs/1801.00690. E. Todorov, T. Erez, and Y. Tassa. Mu Jo Co: A physics engine for model-based control. In 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 5026 5033, 2012. Luca Viano, Angeliki Kamoutsi, Gergely Neu, Igor Krawczuk, and Volkan Cevher. Proximal point imitation learning. In Advances in Neural Information Processing Systems 35 (Neur IPS 2022), 2022. Joe Watson, Sandy Huang, and Nicolas Heess. Coherent soft imitation learning. In Thirty-seventh Conference on Neural Information Processing Systems, 2023. URL https://openreview.net/forum?id=k CCD8d2a Eu. Runzhe Wu, Yiding Chen, Gokul Swamy, Kianté Brantley, and Wen Sun. Diffusing states and matching scores: A new framework for imitation learning. In The Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id=k WRKNDU6u N. Y. Zhou, M. Lu, X. Liu, Z. Che, Z. Xu, J. Tang, and Y Zhang. Distributional generative adversarial imitation learning with reproducing kernel generalization. Neural Networks, 165, 2023. Brian D Ziebart. Modeling purposeful adaptive behavior with the principle of maximum causal entropy. Ph D thesis, University of Washington, 2010. Published in Transactions on Machine Learning Research (11/2025) A Supporting Proofs A.1 Expectation of the Return Distribution Lemma A.1. The expectation of the distributional return satisfies E[Z(s, a)] = Q(s, a). Proof. From the soft return distribution definition t=0 γt [R(st, at) + αH(π( | st))] , (14) taking expectations yields: E[Z(s, a)] = E t=0 γt (R(st, at) + αH(π( | st))) t=0 γt E [R(st, at) + αH(π( | st))] = Q(s, a) (by soft Q-value definition). A.2 Convergence of the Optimal Reward Corollary A.2. For ρπ = ρE, the optimal reward satisfies R Q(s, a) = λπE = λπ. Proof. When ρπ = ρE, Proposition 4.1 simplifies to R Q(s, a) = λπE + λπ Substituting this expression into the loss functions for λπE and λπ in Equation (8) yields min λπE EρE 4 (λπ λπE)2 , min λπ Eρπ 4 (λπE λπ)2 . (16) Differentiating each objective with respect to its corresponding target and setting the derivatives to zero gives λπ = λπE. This result implies that the optimal targets coincide at convergence. Therefore, the optimal reward reduces to R Q(s, a) = λπE = λπ . (17) B Implementation Details Mu Jo Co Suite. We evaluate RIZE on five Gym (Brockman et al., 2016) Mu Jo Co (Todorov et al., 2012) locomotion tasks: Half Cheetah-v2, Walker2d-v2, Ant-v2, Humanoid-v2, and Hopper-v2. Expert trajectories for these tasks are taken from IQ-Learn (Garg et al., 2021) and were generated with Soft Actor Critic (Haarnoja et al., 2018); each trajectory contains 1,000 state action transitions. Episode returns are normalized by expert performance with the following expert evaluation returns: Half Cheetah (5,100), Walker2d (5,200), Ant (4,700), Humanoid (5,300), and Hopper (3,500). Adroit. For Hammer-v1 from the Adroit suite (Rajeswaran et al., 2018), we use the D4RL dataset (Fu et al., 2021) and filter the top 100 episodes from the original 5,000. The resulting expert subset has an average Published in Transactions on Machine Learning Research (11/2025) return of 16,800, and each episode comprises 200 state action pairs. We normalize returns in this domain by the average expert return of the selected subset. Baselines. We evaluate five baselines: IQ-Learn (Garg et al., 2021), LSIQ (Al-Hafez et al., 2023), SQIL (Reddy et al., 2020), CSIL (Watson et al., 2023), and Behavior Cloning (BC) (Pomerleau, 1991). For Mu Jo Co tasks, we use the authors original code and configurations for IQ-Learn, LSIQ, and CSIL; SQIL is implemented using the LSIQ codebase, and BC follows the CSIL implementation. For Hammer-v1, we use the original CSIL code and configs. For IQ-Learn on Hammer, we observe that it does not solve the task even after a small sweep over the entropy coefficient in {0.01, 0.03, 0.1} and the loss mode in {value, v0}; we report α=0.03 and loss=value. For LSIQ on Hammer, we search over α {0.01, 0.05, 0.1} and loss {value, v0} and select α=0.1, loss=value as the better-performing variant. For SQIL, we set α=0.2 consistent with our other tasks. Our architecture integrates components from Distributional SAC (DSAC)1 (Ma et al., 2020) and IQLearn2 (Garg et al., 2021), with hyperparameters tuned through search and ablation studies. Key configurations for experiments involving three and ten demonstrations are summarized in Table 1. All implementation details used in our experiments are publicly available at https://github.com/adibka/RIZE. Distributional SAC Components. The critic network is implemented as a three-layer multilayer perceptron (MLP) with 256 units per layer, trained using a learning rate of 3 10 4. The policy network is a four-layer MLP, also with 256 units per layer. To enhance training stability, we employ a target policy a delayed version of the online policy and sample next-state actions from this module. For return distribution training Zπ ϕ,τ, we adopt the Implicit Quantile Networks (IQN) (Dabney et al., 2018a) approach by sampling quantile fractions τ uniformly from U(0, 1). Additionally, dual critic networks with delayed updates are used, which empirically improve training stability. We use the following settings across tasks: replay buffer size 106, batch size 256, 24 quantile levels, and 10,000 pretraining steps. We evaluate every 104 steps, which takes 3.5 minutes. IQ-Learn Adaptations. Key adaptations from IQ-Learn include adjustments to the regularizer coefficient c and entropy coefficient α. Specifically, for the regularizer coefficient c, we find that c = 0.5 yields robust performance on the Humanoid task, while c = 0.1 works better for other tasks. For the entropy coefficient α, smaller values lead to more stable training. Unlike RL, where exploration is crucial, imitation learning relies less on entropy due to the availability of expert data. Across all tasks, we set initial target reward parameters as λπE = 10 and λπ = 5. Furthermore, we observe that lower learning rates for target rewards improve overall learning performance. Previous implicit reward methods such as IQLearn, Value DICE, and LSIQ3 have employed distinct modifications to the loss function. In our setup, two main loss variants are defined: value loss: L(π, Q) = EρE[Q(s, a) γV (s )] Eρ[V (s) γV (s )] c Γ(RQ, λ) L(π, Q) = EρE[Q(s, a) γV (s )] (1 γ)Ep0[V (s0)] c Γ(RQ, λ) Here, ρ is a mixture distribution, p0 denotes the initial distribution, RQ(s, a) is the implicit reward defined as RQ(s, a) = Q(s, a) γV (s ), the state-value function is given by V (s ) = Q(s , a ) α log π(a |s ), and lastly, our convex regularizer is expressed as Γ(RQ, λ) = EρE[(RQ λπE)2] + Eρπ[(RQ λπ)2]. The choice between v0 or value loss variants depends on environment complexity: we find that for a complex task like Humanoid-v2, the v0 variant demonstrates greater robustness. And, for Half Cheetah-v2, Walker2d-v2, Hopper-v2, Ant-v2, and Hammer-v1, the value variant performs better. 1https://github.com/xtma/dsac 2https://github.com/Div99/IQ-Learn 3https://github.com/robfiras/ls-iq/tree/main Published in Transactions on Machine Learning Research (11/2025) Table 1: Hyperparameters for 3 and 10 Demonstrations (merged where identical) Environment α (3 / 10) c lr π lr λπE lr λπ (3 / 10) Ant-v2 0.05 / 0.10 0.1 5 10 5 1 10 4 1 10 5 / 1 10 4 Half Cheetah-v2 0.05 / 0.10 0.1 5 10 5 1 10 4 1 10 5 / 1 10 4 Walker2d-v2 0.05 / 0.10 0.1 5 10 5 1 10 4 1 10 5 / 1 10 4 Hopper-v2 0.20 / 0.20 0.1 5 10 5 1 10 4 1 10 4 / 1 10 4 Humanoid-v2 0.05 / 0.10 0.5 1 10 5 1 10 4 5 10 5 / 1 10 5 Adroit Hand Hammer-v1 0.30 / 0.30 0.1 3 10 5 1 10 4 5 10 5 / 5 10 5 100 200 300 Half Cheetah-v2 100 200 300 Walker2d-v2 100 200 300 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) Humanoid-v2 100 300 500 Steps ( 103) BC LSIQ SQIL IQLearn CSIL RIZE Expert Figure 6: Learning curves on Mu Jo Co and Adroit tasks with 3 expert demonstrations. Lines show the mean normalized return across five seeds; shaded regions denote 95% confidence intervals. C Additional Experiments This section augments the main experimental results with additional experiments, including targeted ablations (critic architecture, regularization design, and loss choice) and a hyperparameter sensitivity analysis. C.1 Main Results (Extended) Figure 6 complements the main-text results by presenting learning curves for the three-demonstration regime. Learning curves mirror the 10-demo setting (cf. Figure 3): RIZE and CSIL performs better than the baselines with more stable learning across tasks; CSIL shows strong early progress due to behavior-cloning initialization; IQ-Learn is competitive yet below RIZE; and LSIQ/SQIL trail behind. Task-wise, Humanoid-v2 remains solved only by RIZE, while on Hammer-v1 RIZE and CSIL lead. As expected with fewer demonstrations, normalized returns are lower and variance is higher, but the relative ordering of methods and the sample-efficiency patterns are consistent with the RLiable aggregates reported in the main body (Figure 1). C.2 Recovered Reward (Extended) We complement the main-text analysis with additional plots in the 3 demonstration regime and cross-method comparisons. First, Figure 7 mirrors the main-body analysis for the 10 demo setting: it overlays recovered Published in Transactions on Machine Learning Research (11/2025) 100 200 300 Half Cheetah-v2 100 200 300 Walker2d-v2 100 200 300 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) Humanoid-v2 100 300 500 Steps ( 103) Expert reward Policy reward Expert + 1 2c Policy 1 2c Figure 7: Implicit reward curves for expert and policy samples on Mu Jo Co and Adroit tasks with 3 expert demonstrations. Each subplot reports the mean across five seeds, with shaded regions showing the 95% confidence interval. Theoretical upper and lower bounds derived in this work are overlaid as separate curves in each subplot. reward trajectories (expert and policy) with the theoretical bounds implied by our regularizer Γ(RQ, λ) and adaptive targets (Corollary 4.2). As in the main text, the curves remain within the predicted band, and the band s position/width adapts by task, reflecting how the adaptive targets track and limit expert/policy rewards in practice. Across all three figures Figures 9, 8, and 10a RIZE and LSIQ keep implicit rewards bounded, whereas SQIL and IQ-Learn exhibit drifting or unbounded rewards on the challenging Humanoid-v2 and Hammer-v1 tasks. In LSIQ, boundedness is likely due to its target clipping; notably, expert-sample rewards concentrate near zero. By contrast, SQIL fails to control reward scales on nearly all tasks, and IQ-Learn is particularly unstable on Humanoid-v2 and Hammer-v1. These observations support the view that adaptive, task-sensitive regularization is effective in maintaining calibrated implicit rewards. C.3 Critic Value-estimation We examine how critic estimates evolve in RIZE versus baselines. Unlike methods that train point-estimate Q-networks, RIZE employs an IQN critic that models the full return distribution Z(s, a) and optimizes losses using its expectation E[Z(s, a)] (Dabney et al., 2018a). Because all policies are updated by maximizing state action values, unbounded estimates can destabilize training and undermine robustness. Figures 10b, 11, and 12 show that RIZE and LSIQ maintain bounded values on the challenging Humanoid-v2 and Hammer-v1 tasks, whereas SQIL and IQ-Learn exhibit large, drifting estimates. For LSIQ, boundedness primarily arises from target clipping to [ 200, 200], which also explains why expert-sample values cluster near +200. Importantly, bounded critics alone do not guarantee expert-level control within our training budgets (3e5 steps for Mu Jo Co; 5e5 for Adroit): LSIQ remains below RIZE in Figures 3 and 6, indicating that value stability achieved by adaptive reward regularization can translate into better performance than rigid critic clipping with fixed reward targets. Published in Transactions on Machine Learning Research (11/2025) Half Cheetah-v2 Walker2d-v2 Humanoid-v2 Expert Policy Expert Policy Expert Policy 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 24 Expert Policy Figure 8: Implicit reward curves on Mu Jo Co tasks comparing RIZE with SQIL, IQ-Learn, and LSIQ using 10 expert demonstrations. Lines show the mean across five seeds; shaded regions denote 95% confidence intervals. C.4 Ablation on Critic Architecture (Extended) Figure 13 complements the main-body ablation by examining recovered rewards with theoretical bounds when RIZE uses a point-estimate Q(s, a) critic in place of the IQN return-distribution critic Z(s, a) (cf. Figure 5 for returns, and Figure 7 for bounded rewards under IQN). With the Q critic, implicit rewards are less well constrained in Humanoid-v2 and Hammer-v1 environments: we observe more excursions beyond the predicted band and larger oscillations, in contrast to the IQN setting where rewards remain tightly within the adaptive interval. This comparison suggests that both components contribute to reward stability: the squared-TD regularizer with adaptive targets provides principled bounds, and modeling the return distribution with IQN supplies steadier targets/updates that help keep rewards within those bounds. Together, these effects yield more reliable policy learning and align with the performance gap observed in the return curves. C.5 Ablation on Loss Choice (HL-Gauss in LSIQ) Recent work argues that some of the gains attributed to distributional RL may stem from the loss rather than from modeling the return distribution itself (Farebrother et al., 2024; Ayoub et al., 2024). Following this loss swapping view, we replace the mean-squared error (MSE) used in LSIQ s critic with the HL-Gauss classification loss (Farebrother et al., 2024), which discretizes the value range into bins and trains with a Gaussian-smoothed target over bins (turning value regression into calibrated classification). Setup. Among our baselines, LSIQ (and SQIL) are natural candidates because their critics use MSE-like objectives, whereas IQ-Learn and RIZE optimize Max Ent-style IRL objectives. We therefore apply the swap Published in Transactions on Machine Learning Research (11/2025) Half Cheetah-v2 Walker2d-v2 Humanoid-v2 Expert Policy Expert Policy Expert Policy 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 24 Expert Policy Figure 9: Implicit reward curves on Mu Jo Co tasks comparing RIZE with SQIL, IQ-Learn, and LSIQ using three expert demonstrations. Lines show the mean across five seeds; shaded regions denote 95% confidence intervals. to LSIQ and keep all other components unchanged. Following LSIQ s critic clipping in [ 200, 200], we set vmin = 200, vmax = 200, use num_bins=101, and Gaussian width σ = 8. We evaluate on Walker2d-v2, Ant-v2, and Hammer-v1 with three demonstrations. Results. As shown in Figure 14, HL-Gauss underperforms the original MSE-based LSIQ on all three tasks: it yields no improvement on Walker2d-v2 or Ant-v2, and remains near zero on Hammer-v1. A likely cause is structural: LSIQ s critic objective is the sum of two squared-error terms (expert and policy), which is integral to its least-squares/mixture formulation; swapping those squared losses for a categorical (multi-bin) classification surrogate alters the optimization geometry and weakens the intended expert policy mixture shaping. In contrast, standard Q-learning where loss swaps have shown benefits regresses to a single TD target, making the classification surrogate a closer drop-in for MSE. C.6 Ablations on Regularization Strategies We study how the regularizer design affects performance on Walker2d-v2, Ant-v2, and Hammer-v1 with three demonstrations. Our baseline uses a squared TD-error regularizer Γ in Equation (7) with separately optimized adaptive targets for expert and policy samples. We compare this to two alternatives: (i) a coupled target, where we replace the separate targets with a single shared λ (initialized to 0 or 10), and (ii) a plain L2-regularizer on rewards (no targets), akin to IQ-Learn s constraint. Figure 15 shows that squared TD with separate, adaptive targets yields the most robust and highest returns across tasks. Using a coupled target is brittle: λ=10 attains expert-like performance on Ant-v2, remains Published in Transactions on Machine Learning Research (11/2025) Hammer-v1 n=3 Hammer-v1 n=10 Expert Policy Expert Policy 15 Expert Policy 100 300 500 Steps ( 103) 100 300 500 Steps ( 103) Expert Policy Hammer-v1 n=3 Hammer-v1 n=10 Expert Policy Expert Policy Expert Policy 100 300 500 Steps ( 103) 100 300 500 Steps ( 103) Expert Policy Figure 10: Adroit Hammer-v1 results: (a) implicit reward curves and (b) value estimation curves, each reported for 3 and 10 expert demonstrations (n denotes number of trajectories). Lines show the mean across five seeds; shaded regions denote 95% confidence intervals. below the original result on Walker2d-v2, and fails on Hammer-v1, while λ=0 collapses across all tasks. Replacing squared TD with a plain L2 penalty also fails on all three tasks in our setup. We attribute this difference from IQ-Learn s reports to implementation and hyperparameter choices tailored to our setting: RIZE employs an IQN critic, target policy networks, different learning rates (notably for the policy), a higher entropy coefficient (0.05 vs. 0.01), and a smaller regularizer coefficient (0.1 vs. 0.5). These choices were tuned for squared TD with adaptive targets; swapping to an L2 penalty disrupts learning. Overall, the results indicate that adaptive, decoupled targets are crucial for stabilizing reward across tasks. C.7 Hyperparameter Tuning We present our analysis and comparison of important hyperparameters utilized in our algorithm. Plots depict mean over five seeds with 95% confidence intervals. Adaptive Targets. Selecting appropriate initial values and learning rates for the automatic fine-tuning of λπE and λπ is critical in our approach. First, we observe that a suitable learning rate is essential for the stable training of our imitation learning agent, as illustrated in Figure 16a. Our findings indicate that λπ must be optimized very slowly; using larger learning rates can destabilize training and hinder progress. In contrast, λπE demonstrates greater resilience when optimized with higher learning rates. Additionally, λπE remains robust even with varying initial values. However, as shown in Figure 16b, failing to select an appropriate initial value for λπ can negatively impact learning. Overall, Figures 16a and 16b highlight the need for careful selection of both the learning rate and initial value when optimizing λπ, while λπE exhibits considerable robustness in this regard. Regularization Coefficient. Our experiments on the regularizer coefficient c reveal that smaller values of c encourage expert-like performance, while larger values overly constrain rewards and targets, limiting learning. This finding highlights the critical role of selecting an appropriate c, as it directly impacts the Published in Transactions on Machine Learning Research (11/2025) Half Cheetah-v2 Walker2d-v2 Humanoid-v2 Expert Policy Expert Policy Expert Policy 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) Expert Policy Figure 11: Value-estimation curves on Mu Jo Co tasks comparing RIZE (IQN critic (Dabney et al., 2018a)) with SQIL, IQ-Learn, and LSIQ (point-estimate Q) using 10 expert demonstrations. Lines are means over five seeds; shaded regions denote 95% confidence intervals. balance between learning from expert data and regularization: higher values prioritize regularization at the cost of learning, whereas smaller values favor learning but reduce regularization (see Figure 17a). Entropy Coefficient. We observe that the entropy coefficient is a crucial hyperparameter in inverse reinforcement learning (IRL) problems. As shown in Figure 17b, IRL methods typically require small values for α, a point previously noted (Garg et al., 2021). With expert demonstrations available, an imitation learning (IL) policy does not need to explore for optimal actions, as these are provided by the demonstrations. Consequently, higher values of α can lead to training instability, ultimately resulting in policy collapse. Published in Transactions on Machine Learning Research (11/2025) Half Cheetah-v2 Walker2d-v2 Humanoid-v2 Expert Policy Expert Policy Expert Policy 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) Expert Policy Figure 12: Value-estimation curves on Mu Jo Co tasks comparing RIZE (IQN critic (Dabney et al., 2018a)) with SQIL, IQ-Learn, and LSIQ (point-estimate Q) using three expert demonstrations. Lines are means over five seeds; shaded regions denote 95% confidence intervals. Published in Transactions on Machine Learning Research (11/2025) 100 200 300 Half Cheetah-v2 100 200 300 Walker2d-v2 100 200 300 100 200 300 Steps ( 103) 100 200 300 Steps ( 103) Humanoid-v2 100 300 500 Steps ( 103) 1.0 1e7 Hammer-v1 Expert reward Policy reward Expert + 1 2c Policy 1 2c Figure 13: Implicit reward curves for expert and policy samples on Mu Jo Co and Adroit tasks using RIZE with a classic Q(s, a) critic. Each subplot shows the mean over five seeds with shaded 95% confidence intervals, using three expert demonstrations. Theoretical upper and lower bounds derived in this work are overlaid as separate curves in each subplot. 100 200 300 Steps ( 103) Walker2d-v2 100 200 300 Steps ( 103) 100 300 500 Steps ( 103) LSIQ (MSE) LSIQ (HL-Gauss) Expert Figure 14: Loss swap in LSIQ: MSE vs. HL-Gauss on Walker2d-v2, Ant-v2, and Hammer-v1 with three demonstrations. We keep vmin = 200, vmax = 200, num_bins=101, and σ = 8. HL-Gauss does not improve over MSE and remains near zero on Hammer-v1. 100 200 300 Steps ( 103) Walker2d-v2 100 200 300 Steps ( 103) 100 300 500 Steps ( 103) RIZE (main) RIZE (coupled, =0) RIZE (coupled, =10) RIZE (no TD) Expert Figure 15: Ablation of regularization strategies on selected tasks (Walker2d, Ant, Hammer) using three expert demonstrations. Results are expert-normalized and reported as the mean over five seeds with 95% confidence intervals. Published in Transactions on Machine Learning Research (11/2025) 100 200 300 Steps ( 103) RIZE (lr E =10 4, lr =10 5) RIZE (lr E =10 4, lr =10 3) RIZE (lr E =10 3, lr =10 5) 100 200 300 Steps ( 103) RIZE ( E =10, =5) RIZE ( E =10, =0) RIZE ( E =20, =5) Figure 16: Fine-tuning analysis of adaptive targets. (a) Learning rates: Turquoise represents our method s primary result with learning rates of 1e 4 for λπE and 1e 5 for λπ. Orange and blue lines indicate higher learning rates (e.g., 1e 3) for λπE and λπ, respectively. (b) Starting values: Turquoise shows the main result with initial values of 10 for λπE and 5 for λπ, while other lines explore different starting values. Three trajectories are used throughout the analysis. 100 200 300 Steps ( 103) RIZE (w/ C=0.1) RIZE (w/ C=0.6) 100 200 300 Steps ( 103) RIZE (w/ =0.05) RIZE (w/ =0.5) Figure 17: (a) Effect of the regularizer coefficient c. Turquoise shows the primary result of our method with c = 0.1, while gray represents a larger value (c = 0.6). (b) Effect of the temperature parameter α. Turquoise shows the result with α = 0.05, and gray corresponds to a larger value (α = 0.5). Three trajectories are used throughout the analysis.