# qexponential_family_for_policy_optimization__c18bcb7c.pdf Published as a conference paper at ICLR 2025 q-EXPONENTIAL FAMILY FOR POLICY OPTIMIZATION Lingwei Zhu University of Tokyo lingwei4@ualberta.ca Haseeb Shah University of Alberta hshah1@ualberta.ca Han Wang University of Alberta han8@ualberta.ca Yukie Nagai University of Tokyo Martha White University of Alberta Policy optimization methods benefit from a simple and tractable policy parametrization, usually the Gaussian for continuous action spaces. In this paper, we consider a broader policy family that remains tractable: the q-exponential family. This family of policies is flexible, allowing the specification of both heavy-tailed policies (q > 1) and light-tailed policies (q < 1). This paper examines the interplay between q-exponential policies for several actor-critic algorithms conducted on both online and offline problems. We find that heavy-tailed policies are more effective in general and can consistently improve on Gaussian. In particular, we find the Student s t-distribution to be more stable than the Gaussian across settings and that a heavy-tailed q-Gaussian for Tsallis Advantage Weighted Actor-Critic consistently performs well in offline benchmark problems. In summary, we find that the Student s t policy a strong candidate for drop-in replacement to the Gaussian. Our code is available at https://github.com/lingweizhu/qexp. 1 INTRODUCTION Policy optimization methods optimize the parameters of a stochastic policy towards maximizing some performance measure (Sutton et al., 1999). These methods benefit from a simple and tractable policy functional. For discrete action spaces, the Boltzmann-Gibbs (BG) policy is often preferred (Mei et al., 2020; Cen et al., 2022); while the Gaussian policy is standard for the continuous case. For continuous action spaces, sampling the BG policy is computationally expensive due to the normalizing log-partition function. A Gaussian policy is often used as a tractable approximation. While there are other candidates such as the Beta policy (Chou et al., 2017), the Gaussian remains the most common choice for both online and offline policy optimization methods (Haarnoja et al., 2018; Neumann et al., 2023; Xiao et al., 2023). Figure 1: The policy parametrizations in this paper. Student s t is a q-exponential with q = 1 + 2/(ν + 1) 1.67. In this paper, we consider a broader policy family that remains tractable called the q-exponential family. The q-exponential family was proposed to study non-extensive system behaviors in the statistical physics (Naudts, 2010; Matsuzoe & Ohara, 2011), and has recently been exploited in transformers (Peters et al., 2019; Martins et al., 2022). By setting q = 1, it recovers the standard exponential family. With q > 1, we can obtain policies with heavier tails than the Gaussian, such as the Student s t-distribution (Kobayashi, 2019) or the L evy Process distribution (Simsekli et al., 2019; Bedi et al., 2024). Heavy-tailed distributions can preferable as they are more robust (Lange et al., 1989), can facilitate exploration and help escape local optima in the sparse reward context (Chakraborty et al., 2023). When q < 1, light-tailed (sparse) policies such as the q-Gaussian distribution can be recovered. The sparse q-Gaussian has finite support and indicates joint first authors. Published as a conference paper at ICLR 2025 Reference Scope of expq Explicit? Heavy & Sparse? Continuous? RL? Naudts (2010); Matsuzoe & Ohara (2011) q R - Martins et al. (2022) q < 1 - Lee et al. (2018); Chow et al. (2018a) q = 0 Lee et al. (2020); Zhu et al. (2023; 2024) q < 1 Li et al. (2023) q = 0 This paper q R Table 1: Existing works and their scopes. We are the first to consider the general q-exponential family for the parameterized policy in reinforcement learning. The family includes continuous heavy-tailed and sparse policies. Prior works in RL considered only the discrete case or continuous policy with a specific entropic index q. Further, in many cases they still used a Gaussian policy parameterization to approximate an implicit target distribution that is a q-exponential, rather than explicitly using the q-Gaussian as the policy parameterization. can serve as a continuous generalization of the discrete sparsemax. As a result, q-Gaussian helps alleviate safety concerns incurred by the infinite support Gaussian (Xu et al., 2023; Li et al., 2023). Such q-exponential families have been considered in reinforcement learning, with the existing work summarized in Table 1. Lee et al. (2018); Chow et al. (2018b) studied the discrete setting with q = 0, called the sparsemax. Li et al. (2023) similarly considered q = 0 policy parameterization for the continuous action setting. All other works, however, used a Gaussian policy parameterization to (implicitly) approximate an idealized target distribution that is q-Gaussian, and specifically for q < 1 Lee et al. (2020); Zhu et al. (2024). Such a choice is suboptimal, as Gaussians are used to approximate light-tailed (sparse) target policies. And in fact that choice was not strictly necessary as the policy parameterization need not have been chosen to be Gaussian: it could also have been a q-Gaussian. Though obvious in hindsight, this gap was likely due to simply not considering the use of general continuous q-exponential family for the policy parameterization, which is what we introduce in this work. 80 90 100 110 Performance relative to Squashed Gaussian (%) 100% of Squashed Gaussian Gaussian Student's t Heavy-Tailed q-Gaussian Overall Performance Across Figure 2: Performance relative to the Squashed Gaussian on the offline D4RL Mu Jo Co task, averaged across the selected algorithms and environments. In this paper, we empirically investigate the q-exponential family as a replacement for the Gaussian inside several existing policy optimization algorithms. Our contributions include the following. (1) We show how to use q-exponential family policy parameterizations inside a variety of existing actor-critic algorithms. (2) We provide comprehensive experiments on both online and offline problems showing that q-exponential family policies can improve on the Gaussian by a large margin. In particular, we find that the Student s t policy is more stable, performing well across algorithms and problems, shown in Figure 2. (3) We provide empirical evidence supporting the assumption that algorithms may prefer specific policies depending on the actor loss objective. In particular, we find that by replacing the Gaussian with a heavy-tailed q Gaussian, Tsallis Advantage Weighted Actor-Critic (Zhu et al., 2024) consistently performs better across offline benchmark problems. This outcome makes sense; as mentioned above, this algorithm implicitly has a target policy that is a q-Gaussian, so using a matching q-Gaussian parameterization should perform better. 2 BACKGROUND We focus on discounted Markov Decision Processes (MDPs) expressed by the tuple (S, A, P, µ, r, γ), where S and A denote state space and action space, respectively. Let (X) denote the set of probability distributions over X. P and µ denote the transition probability and initial state distribution, respectively. r(s, a) defines the reward associated with that transition. γ (0, 1) is the discount factor. Published as a conference paper at ICLR 2025 A policy π : S (A) is a mapping from the state space to distributions over actions. To assess the quality of a policy, we define the expected return as J(π) = R A π(a|s)r(s, a)dads, where ρπ(s) = P t=0 γt P(st = s) is the unnormalized state visitation frequency. The goal is to learn a policy that maximizes J(π). We also define the action value and state value as Qπ(s, a) = Eπ [P t=0 γtr(st, at)|s0 µ, a0 = a], V π(s) = Eπ [Qπ(s, a)]. For the ease of later notations, we write the dependence on state as subscript, e.g. Q(s, a) will be written as Qs(a). In practice, the policy is often parametrized by a vector of parameters θ Rn. The policy can then be optimized by adjusting its parameters to the high reward region utilizing its gradient information. The Policy Gradient Theorem (Sutton et al., 1999) featured by many policy gradient methods states that the gradient can be computed by: θJ(π; θ) = Es ρπ, a πθ [Qπ s (a) θ ln πs(a; θ)] . In practice, the expectation is approximated by sampling. When the state space is large, the action value function is also parametrized, leading to the Actor-Critic methods (Degris et al., 2012). In contrast to the study of policy gradient algorithms, the impact of specific policy parametrizations on performance remains a less studied topic. Researchers typically consider policy parametrizations that can be written as the following: πs(a; θ) = 1 Zs exp θ ϕs(a) = exp θ ϕs(a) Z s . (1) Here, ϕs(a) is a vector of statistics and θ Rn is a vector of parameters, Zs is the normalizing constant ensuring the policy is a valid distribution and Zs := exp (Z s). One immediate instance is the Boltzmann-Gibbs (BG) policy πBG,s(a; θ) = exp Qs(a) Zs , where Z s = ln R exp Qs(a) da is the log-partition function. In the discrete case, it is also called the softmax transformation (Cover & Thomas, 2006). BG policy has been studied extensively in RL for encouraging exploration and smoothing the optimization landscape, to name a few applications (Haarnoja et al., 2018; Ahmed et al., 2019; Cen et al., 2022). However, evaluating the log-partition function is in general intractable. 3 EXPONENTIAL AND q-EXPONENTIAL FAMILIES We first review the commonly used policy parametrizations. They permit an expression using the exponential function. We arrive at the more general q-exponential family by deforming the exponential. In Table 2, we summarize all policies presented in the paper. 3.1 THE EXPONENTIAL FAMILY POLICIES The Gaussian policy is one of the simplest distributions one can consider due to its omnipresence in statistics and parametric estimation as well as its widely available sampling procedure implementations. Since evaluating the log-partition function of BG is intractable, due to the aforementioned advantages many researchers consider the Gaussian policy instead: πs(a) = 1 2πσs exp (a µs)2 . For simplicity we drop the dependence on state s. To see it is a member of the exponential family, in Eq. (1) let θ = [ µ 2σ2 ] for µ ( , ), σ > 0; ϕs(a) = [a, a2] , and Zs = ln This amounts to setting Qs(a) = (a µ)2 2σ2 in the BG (Gu et al., 2016). We write a Gaussian policy as πN,s(a) = N(a; µ, σ2). The gradients of the Gaussian are µ ln πs(a) = (a µ) σ ln πs(a) = (a µ)2 σ. On one hand, the Gaussian policy is simple to implement. On the other hand, when σ becomes small, Gaussian can be unstable due to overly large gradients and can prematurely concentrate on a suboptimal action. As a result, it is susceptible to noise/outliers and does not encourage sufficient exploration due to its thin tails. This paper investigates location-scale alternatives within the generalized q-exponential family. Another interesting member is the Beta distribution (Chou et al., 2017): πBeta,s(a) = Γ(α+β) Γ(α)Γ(β)aα 1(1 a)β 1, a (0, 1), where Γ( ) is the gamma function. It can be retrieved from equation 1 by letting θ = [α, β] , ϕs(a) = [ln a, ln(1 a)] , Zs = Γ(α)Γ(β) Γ(α+β) . Since Beta distribution s support is bounded between (0, 1), Chou et al. (2017) argued that it might alleviate the Published as a conference paper at ICLR 2025 Family Policy Parameters θ Statistics ϕs(a) Normalization Zs ln πs(a) Gaussian [ µ 2σ2 ] [a, a2] 2πσ Eq. (13) Beta [α, β] [ln a, ln(1 a)] Γ(α)Γ(β) Student s t 2µ 2 ) Eq. (14) q-Gaussian (q < 1) [ µ 2σ2 ] [a, a2] π 1 q Γ( 1 1 q +1) Γ( 1 1 q + 3 2) Eq. (15) q-Gaussian (1 < q < 3) q π q 1 Γ( 1 q 1 1 Table 2: Policy parametrizations from the exp and q-exp families studied in this paper. We are primarily interested in the location-scale family. Their multivariate forms are shown in Appendix A. Figure 3: expq x and lnq x for q < 1 and q > 1. When q = 1 they respectively recover their standard counterpart. For q < 1 the q-exp can return zero values and hence q-exp policies may achieve sparsity. For q > 1, q-exp decays more slowly towards 0, resulting in heavy-tailed behaviors. The rightmost shows the q-Gaussian with different q. bias introduced by truncating Gaussian densities outside the action space bounds. The beta policy is the only non-location-scale family distribution in this paper. However, as we will show in the experiments, the Beta policy generally does not perform favourably against the Gaussian. 3.2 THE q-EXPONENTIAL FAMILY, HEAVY-TAILED AND LIGHT-TAILED POLICIES Generalizing the exponential family using the q-exponential function has been extensively discussed in statistical physics (Naudts, 2002; Tsallis, 2009; Naudts, 2010; Amari & Ohara, 2011). In the machine learning literature, the q-exponential generalization has attracted some attention since it allows for tuning the tail behavior by adjusting the value of q (Sears, 2008; Ding & Vishwanathan, 2010; Amid et al., 2019). The q-exponential and its unique inverse function q-logarithm are: ( exp x, q = 1 [1 + (1 q)x] 1 1 q + , q = 1 , lnqx := ( ln x, q = 1 x1 q 1 1 q , q = 1 (2) where [ ]+ := max{ , 0}. q-exp/log generalize exp/log since limq 1 expq x = exp x and limq 1 lnq x = ln x. Similar to exp, q-exp is an increasing and convex function for q > 0, satisfying expq(0) = 1. However, an important difference of q-exp is that expq(a + b) = expq(a) expq(b) unless q = 1. We visualize q-exp/log in Figure 3. We now define the q-exponential family as: πq,s(a; θ) = 1 Zq,s expq θ ϕs(a) = expq θ ϕs(a) Z q,s , (3) where θ, ϕs(a), Zq,s have similar meanings to equation 1. Note that Zq,s = expq Z q,s unless q = 1. The q-exponential family includes the q-Gaussian and Student s t distributions described in the next subsections. Published as a conference paper at ICLR 2025 3.2.1 q-GAUSSIAN As the counterpart of Gaussian in the q-exp family, q-Gaussian unifies both light-tailed and heavytailed policies by varying the entropic index q (Matsuzoe & Ohara, 2011): πNq,s(a) = 1 Zq,s expq where Zq,s = π 1 q Γ 1 1 q + 1 / Γ 1 1 q + 3 2 if < q < 1, π q 1 Γ 1 q 1 1 2 / Γ 1 q 1 if 1 < q < 3. It is heavy-tailed when 1 < q < 3 and light-tailed when q < 1. πNq,s(a) is no longer integrable for q 3 (Naudts, 2010). We visualize these q-Gaussians in Figure 3. Since popular libraries like the Py Torch (Paszke et al., 2019) do not have implementations of q Gaussians available, we discuss their sampling methods. It was shown by (Martins et al., 2022) that a sparse q-Gaussian (q < 1) random variable permits a stochastic representation µ + r Au, where u Unif SN is a random sample from the N 1 dimensional unit sphere. A is the scaled matrix |Σ| 1 2N+ 4 1 q Σ 1 2 . r is the radius of the distribution, and the ratio follows the Beta distribution r2/R2 Beta ((2 q)/(1 q), N/2) , where R is radius of the supporting sphere of the standard q-Gaussian Nq(0, I): Γ 2 q 1 q π N 1 q 2+(1 q)N Notice that R depends only on the dimensionality N and the entropic index q. This method provides low-variance samples, but unfortunately it does not extend to q > 1. Therefore, for 1 < q < 3 we adopt the Generalized Box-M uller Method (GBMM) (Thistleton et al., 2007) to transform uniform random variables u1, u2 Unif(0, 1)N by the following: 2 lnq (u1) cos (2πu2) , z2 = q 2 lnq (u1) sin (2πu2) . (6) Then each of z1, z2 is a standard q-Gaussian variable with new entropic index q = (3q 1)/(q + 1). Often we know the desired q in advance, in this case we simply let the q-log take on the index q = (q 1)/(3 q ). The desired random vector is given by µ + Σ 1 2 z. 3.2.2 STUDENT S T Algorithm 1: q-Gaussian sampling Input: q, N, µ, Σ if q < 1 then sample u Unif(SN) sample z Beta 2 q 1 q, N compute R per Eq. (5) compute A = |Σ| 1 2N+ 4 1 q Σ 1 2 z R2Au else if q > 1 then sample u1, u2 Unif(0, 1)N compute z by GBMM Eq. (6) return µ + Σ 1 2 z Heavy-tailed distributions like the Student s t are popular for robust modelling (Lange et al., 1989). The Student s t distribution is πSt,s(a) = Γ ν+1 2 1 + (a µ)2 (7) where ν > 0 is the degree of freedom. As ν , Student s t distribution approaches the Gaussian. Numerically, Student s t with ν 30 is considered to closely match the Gaussian. Therefore, ν can be an important learnable parameter in addition to its location µ and scale σ. It allows the policy to adaptively balance the exploration-exploitation tradeoff by interpolating the Gaussian and heavy-tailed policies. Now let q = 1 + 2 ν+1 and define Zq,s := πνσ Γ ν 2 , θ ϕs(a) := Zq 1 q,s (1 q) (a µ)2 πSt, s(a) = expq θ ϕs(a) ln2 q Zq,s , Published as a conference paper at ICLR 2025 which we see it is indeed a q-exp policy and Z q,s = ln2 q Zq,s. Student s t policy has been used in (Kobayashi, 2019) to encourage exploration and to escape local optima. Another related case is the Cauchy s distribution recovered when q = 2 (or ν = 1 from Student s t). Cauchy s distribution can be used as the starting point for learning Student s t. Note that Cauchy s distribution does not have valid mean, variance or any higher moments. 4 USING q-EXPONENTIAL FAMILIES FOR ACTOR-CRITIC ALGORITHMS In this section, we outline three key actor-critic algorithms we use in our study and the nuances of incorporating q-exp policies into them. For example, the q-exp policies may not have closed-form Shannon entropy. Therefore, approximations are needed for algorithms like SAC and Greedy AC. Moreover, though for the Gaussian evaluating the log-likelihood for off-policy/offline actions causes no problem, it raises a new issue for the light-tailed q-Gaussian, since these actions can fall outside of its support. Soft Actor-Critic. SAC (Haarnoja et al., 2018) encourages exploration by adding to reward the Shannon entropy. The actor minimizes the following KL loss LSAC(ϕ) := Es B [DKL(πϕ,s || πBG,s)] = Es B where states are sampled from replay buffer B. The parametrized policy πϕ is projected to be close to the BG policy. By default πϕ is chosen to be the Gaussian policy, but potentially a more exploring policy like the Student s t could be better. Depending on action values, BG can have multiple modes and heavy tails. The Gaussian may not be able to fully capture these characteristics. Greedy Actor-Critic. Greedy AC (Neumann et al., 2023) maintains an additional proposal policy for exploration by maximizing Shannon entropy augmented rewards. Its actor policy maximizes unbiased reward and learns from the high-quality actions generated by the proposal policy. To simplify notations, we use I(s) to denote the set of high quality actions given s. LGreedy AC, prop(ϕ) := E s B a I(s) [ ln πϕ,s H (πϕ,s)] , LGreedy AC, actor( ϕ) := E s B a I(s) Greedy AC maximizes log-likelihood of the actor and entropy-augmented likelihood for the proposal policy. Note that the when πϕ,s is a q-exp policy, it may not have a closed-form Shannon entropy expression. Therefore, we can use log-probabilities as a surrogate just like in SAC. Tsallis Advantage Weighted Actor-Critic. TAWAC (Zhu et al., 2024) proposed to use a light-tailed q-exp policy for offline learning. However, the light-tailed distribution was approximated with the Gaussian which is an infinite-support policy. Let πD denote the empirical behavior policy and D the offline dataset. TAWAC minimizes the following actor loss: L(ϕ) := Es D [DKL(πTKL,s || πϕ,s)] = E s D a πD expq Qs(a) Vs ln πϕ,s(a) , (9) where πTKL,s(a) πD,s(a) expq τ 1 (Qs(a) Vs) denotes the Tsallis KL regularized policy. πϕ,s mimics a TKL policy which can be sparse depending on q . In this case, it is natural to expect that a sparse policy parametrization may lead to better performance. Algorithm 2: Out-of-support action handling for the light-tailed q-Gaussian Input: out-of-support action a sample in-support actions {bi}i=1:K solve i = arg mini ||bi a||2 2 return bi Algorithms like TAWAC that sample from a behavior policy πD needs extra caution when using the q-exp policies. When the light-tailed q-Gaussian is used, numerical issues can be incurred since the action sampled may fall outside the support of πϕ, leading to undefined log-likelihood. To resolve this problem, we propose to sample from πϕ a batch of K actions and replace the out-of-support action with the in-support one with least L2 distance, see Alg. 2. Published as a conference paper at ICLR 2025 Figure 4: Learning curves on the classic control environments. Only the Gaussian and the best policy parametrization for each setting were shown with full opacity. The best policy is picked based on the total area under the curve (AUC). TAWAC(0) refers to TAWAC with entropic index q = 0 in Eq. (9). Despite tuning hyperparameters separately for each policy, Gaussian is the best policy in only 1/12 settings. In most other settings, the Gaussian policy performs significantly worse than the best. Figure 5: (Left) The percentage of times that each policy parametrization is better than the Gaussian across all algorithm-environment combinations based on total AUC. If the bar is above the 50% line, then it means that the policy parametrization is better than Gaussian on average. We see that Student s t and Light-tailed Gaussians are better than the Gaussian in 75% and 66% of the settings, respectively. (Right) Count of times where a policy parametrization performed the best across all algorithm-environment combinations based on AUC. We observe that the student-t policy performed the best in 5/12 settings, whereas the Gaussian policy performed the best only once. 5 EXPERIMENTS Our empirical study s primary goal is to understand better the performance differences under this broader class of policy parameterizations in both online and offline settings. We ran experiments Published as a conference paper at ICLR 2025 with different algorithms, to get a better sense of how conclusions about policy parameterization vary across different actor-critic algorithms. We parametrize Student-t s DOF parameter ν in addition to its location and scale. By contrast, the heavy-tailed q-Gaussian is fixed at q = 2, since its allowable range is 1 < q < 3. For the light-tailed q-Gaussian, we opt for the standard choice of q = 0. Since Student s t, heavy-tailed q-Gaussian, and Gaussian have unbounded support, we clipped the sampled action to fit the task s action space without modifying the density. We swept the hyperparameters using five random seeds, then increased the number of seeds to 10 for the best parameter setting. The hyperparameter sweep ranges and the best values are provided in Appendix D.2 and D.3. 5.1 ONLINE CLASSIC CONTROL Domains and Baselines. We used three classical control environments in the continuous action setting: Mountain Car (Sutton & Barto, 2018), Pendulum (Degris et al., 2012) and Acrobot (Sutton & Barto, 2018). We chose the cost-to-goal version of Mountain Car, which outputs 1 reward per time step to encourage reaching the goal early. We compared SAC, Greedy AC and two versions of TAWAC, q = 0 and q = 2. Results. Figure 4 shows the learning curves of all algorithm-environment combinations. Only the Gaussian and the environment-specific best policy are shown with full opacity, computed based on area under curve (AUC). One immediate observation is that, though all three algorithms by default choose the Gaussian policy, it was seldom the best policy parametrization. Environment-wise, on Mountain Car the Gaussian did not rank the best for any of the algorithms. By contrast, the Beta policy attained the first place with SAC, as was the light-tailed q-Gaussian with TAWAC. The same trend for the Gaussian holds in Acrobot and Pendulum as well, with exception only on TAWAC(0) Acrobot, where its curve closely resembled that of the light-tailed q-Gaussian. Algorithm-wise, three observations are to be made: (i) on Mountain Car the Beta policy performed significantly better than others. This could be due to its flexibility in maintaining a skewed distribution shape that matches the BG policy more closely in contrast to the other location scale family members. (ii) The q-Gaussians in general outperformed the Gaussian on TAWAC(0) and TAWAC(2) whose actor explicitly mimics a q-exp policy. (iii) Student s t has ranked the top involving all three algorithms. Figure 5 LHS summarizes the percentage of each policy parametrization outperforming the Gaussian. The Student s t and light-tailed Gaussian went above 50%, suggesting potentially greater applicability. The RHS shows out of 12 total combinations, how many times each policy parametrization has ranked the top. The result shows that the Student s t attained 5 times, contrasting the 1 time of the Gaussian. Steps (x104) Gaussian Heavy-tailed q-Gaussian Light-tailed q-Gaussian Greedy-AC Policy Evolution on Mountain Car Figure 6: Policy evolution of Greedy AC on Mountain Car. The Gaussian collapsed into a delta-like policy after only 10% of the learning horizon. In Figure 6 we visualized the evolution of Gaussian and q-Gaussian policies on the starting state over the first 4 104 steps (10% of the entire learning horizon). Note that the allowed action range is [ 1, 1] but the plot shows [ 2, 2] for better visualization. Gaussian tends to quickly concentrate like a delta policy. This can be detrimental to algorithms like SAC and Greedy AC which demand stochasticity to generate diverse samples. By contrast, both lightand heavytailed q-Gaussians tend to be more stochastic. In Figure 11 we show the Manhattan plot of SAC with all swept hyperparameters on all environments. Though there is no a definitive winner for all cases, it is visible that the Student s t and Gaussian have a similar behavior to hyperparameter changes. Therefore, if we are tackling a problem where the Gaussian works, the Studentt is very likely to work. And judging from Fig. 5, we know that Student s t is 75% more likely to perform better than Gaussian given the same hyperparameter sweeping range. Published as a conference paper at ICLR 2025 TAWAC AWAC IQL In AC TAWAC AWAC IQL In AC TAWAC AWAC IQL In AC 0 Normalized Return Half Cheetah Hopper Walker2d Medium-Replay Data Squashed Gaussian Student's t Heavy-Tailed q-Gaussian Figure 7: Normalized scores on Medium-Replay level datasets from the Mu Jo Co suite. The black bar shows the median. Boxes and whiskers are 1 and 1.5 interquartile ranges, respectively. See Figure 15 for full comparison. Environment-wise, TAWAC with heavy-tailed q-Gaussian is often the top performer. Algorithm-wise, Student s t consistently outperforms Squashed Gaussian. 5.2 OFFLINE D4RL MUJOCO TAWAC AWAC IQL In AC 40% Medium-Expert Data TAWAC AWAC IQL In AC Medium Data TAWAC AWAC IQL In AC Medium-Replay Data Proportion to Squashed Gaussian Performance Gaussian Beta Student's t Heavy-Tailed q-Gaussian Figure 8: Relative improvement to the Squashed Gaussian policy, averaged over multiple environments in the Mu Jo Co suite. The Student s t consistently outperforms the Gaussian with all the chosen algorithms. The heavy-tailed q-Gaussian with TAWAC and IQL also achieved significant improvement. The improvement can reach up to 20%. Black vertical lines at the top indicate one standard error. Domains and Baselines. We used the standard benchmark Mu Jo Co suite from D4RL to evaluate algorithm-policy combinations (Fu et al., 2020). The following algorithms are compared: TAWAC, Advantage Weighetd Actor-Critic (AWAC) (Nair et al., 2021), Implicit Q-Learning (IQL) (Kostrikov et al., 2022), In-sample Actor-Critic (In AC) (Xiao et al., 2023). For TAWAC, we fixed its leading q -exp with q = 0. In Appendix C.2 we detailed the compared algorithms. We also included a popular variant of the Gaussian known as the Squashed Gaussian for comparison. Being able to evaluate the offline log-probability is critical to the tested algorithms, we found that light-tailed q-Gaussian leads to poor performance even with random online sampling, hence we do not show them here. Results. Figure 7 compared the normalized scores on the Medium-Replay datasets. It can be seen that environment-wise, TAWAC + heavy-tailed q-Gaussian was the top performer, and could improve on the Squashed Gaussian by a non-negligible margin. On Half Cheetah, heavy-tailed q-Gaussian attained the best score with every algorithm. Algorithm-wise, the heavy-tailed q-Gaussian or/and Student s t were better or equivalent to the Squashed Gaussian, except with AWAC on Hopper. Student s t was stable across algorithms, including these with which heavy-tailed q-Gaussian performed poorly (e.g., In AC). This demonstrates the value of the learnable DOF parameter that allows it interpolates the Gaussian. In Appendix E we provided comparison on other policies and datasets. Figure 8 summarized the relative improvement over the Squashed Gaussian across environments. Several observations can be made: (i) though the Squashed Gaussian outperformed the Gaussian in general, it was seldom the best performer. (ii) the Student s t could consistently perform better Published as a conference paper at ICLR 2025 Figure 9: Policy evolution of all actions dimensions of TAWAC on Walker2d Medium Replay. Student s t was flexible in that on some dimensions it had lighter tails like the Gaussian by having large DOF (e.g. 4th), and with heavier tails on the others by having smaller DOF (e.g. 3rd, 6th). The peaks at the edges were caused by clipping actions into the allowed range. than the Gaussian, the improvement can sometimes reach up to 20%. The same holds for the heavy-tailed q-Gaussian with TAWAC and IQL. (iii) though there was no single winner for all cases, choosing the Student s t for the actors with exponential loss functions (AWAC, IQL, In AC), or the heavy-tailed q-Gaussian for q-exponential actor losses (e.g. TAWAC) are generally effective. Figure 9 visualized the policy evolution of the Squashed Gaussian and the two heavy-tailed policies, learned with TAWAC on Medium-Replay Walker2D. Squashed Gaussian tended to converge slower here. Since the offline Mu Jo Co environments are fully deterministic, a wide distribution indicates failure of finding the mode of the optimal action and therefore can be detrimental to learning performance. The Squashed Gaussian converged slower than the heavy-tailed (performed the best) and the Student s t. Student s t was flexible in that it beared lighter tails like the Gaussian in some dimensions by having a large DOF, for example in the 4th and 5th dimensions. On the other hand, it could take heavy tails by having a small DOF like in the 3rd and 6th dimensions. 6 CONCLUSION The Gaussian policy is standard for policy optimization algorithms on continuous action spaces. In this paper we considered a broader family of policies that remains tractable, called the q-exponential family. We empirically investigated their utility as a promising alternative to the Gaussian. Specifically, we looked at the Student s t, lightand heavy-tailed q-Gaussian policies. Extensive experiments on both online and offline tasks with various actor-critic methods showed that heavy-tailed policies are in general effective. In summary, we found the Student s t policy to be generally more performant and stable than the Gaussian and could be used as a drop-in replacement. By contrast, the Heavy-tailed q-Gaussian seemed to favor especially Tsallis regularization and outperformed the baselines. We acknowledge that the paper has limitations. Perhaps the greatest is the inherent dilemma of the light-tailed q-Gaussian evaluating out-of-support actions. Off-policy/offline algorithms require evaluating actions from some behavior policy and the actions can fall outside the support of the sparse q-Gaussian. Na ıvely discarding these samples results extremely slow or no learning. In this paper we proposed to alleviate this issue by replacing them with the in-support sampled action with the least L2 distance. Nonetheless, this method did not help much in offline experiments. We envision a potential solution that is left to future investigation: projecting the out-of-support actions precisely to the boundary of the q-Gaussian. Published as a conference paper at ICLR 2025 Zafarali Ahmed, Nicolas Le Roux, Mohammad Norouzi, and Dale Schuurmans. Understanding the impact of entropy on policy optimization. In Proceedings of 36th International Conference on Machine Learning, volume 97, pp. 151 160, 2019. Shun-ichi Amari and Atsumi Ohara. Geometry of q-exponential family of probability distributions. Entropy, 13(6):1170 1185, 2011. Ehsan Amid, Manfred K. Warmuth, and Sriram Srinivasan. Two-temperature logistic regression based on the tsallis divergence. In Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, volume 89, pp. 2388 2396, 2019. Amrit Singh Bedi, Anjaly Parayil, Junyu Zhang, Mengdi Wang, and Alec Koppel. On the sample complexity and metastability of heavy-tailed policy search in continuous control. Journal of Machine Learning Research, 25(39):1 58, 2024. Boris Belousov and Jan Peters. Entropic regularization of markov decision processes. Entropy, 21(7), 2019. Shicong Cen, Chen Cheng, Yuxin Chen, Yuting Wei, and Yuejie Chi. Fast global convergence of natural policy gradient methods with entropy regularization. Operations Research, 70(4): 2563 2578, 2022. Souradip Chakraborty, Amrit Singh Bedi, Kasun Weerakoon, Prithvi Poddar, Alec Koppel, Pratap Tokekar, and Dinesh Manocha. Dealing with sparse rewards in continuous control robotics via heavy-tailed policy optimization. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 989 995, 2023. Po-Wei Chou, Daniel Maturana, and Sebastian Scherer. Improving stochastic policy gradients in continuous control with deep reinforcement learning using the beta distribution. In Proceedings of the 34th International Conference on Machine Learning, pp. 834 843, 2017. Yinlam Chow, Ofir Nachum, and Mohammad Ghavamzadeh. Path consistency learning in Tsallis entropy regularized MDPs. In International Conference on Machine Learning, pp. 979 988, 2018a. Yinlam Chow, Nachum Ofir, Edgar Duenez-guzman, and Mohammad Ghavamzadeh. A Lyapunovbased Approach to Safe Reinforcement Learning. In Annual Conference on Neural Information Processing Systems (NIPS), pp. 1 10, 2018b. Thomas M. Cover and Joy A. Thomas. Elements of Information Theory (Wiley Series in Telecommunications and Signal Processing). Wiley-Interscience, USA, 2006. Thomas Degris, Martha White, and Richard S. Sutton. Off-policy actor-critic. In Proceedings of the 29th International Coference on International Conference on Machine Learning, pp. 179 186, 2012. Nan Ding and S.v.n. Vishwanathan. t-logistic regression. In Advances in Neural Information Processing Systems, volume 23, 2010. Justin Fu, Aviral Kumar, Ofir Nachum, George Tucker, and Sergey Levine. D4rl: Datasets for deep data-driven reinforcement learning, 2020. Scott Fujimoto and Shixiang Shane Gu. A minimalist approach to offline reinforcement learning. In Thirty-Fifth Conference on Neural Information Processing Systems, 2021. S. Furuichi, K. Yanagi, and K. Kuriyama. Fundamental properties of tsallis relative entropy. Journal of Mathematical Physics, 45(12):4868 4877, 2004. Shigeru Furuichi. On the maximum entropy principle and the minimization of the fisher information in tsallis statistics. Journal of Mathematical Physics, 50:013303, 01 2010. Peter Gr unwald and Alexander Dawid. Game theory, maximum entropy, minimum discrepancy and robust bayesian decision theory. Annals of Statistics, 32, 2004. Published as a conference paper at ICLR 2025 Shixiang Gu, Timothy Lillicrap, Ilya Sutskever, and Sergey Levine. Continuous deep q-learning with model-based acceleration. In Proceedings of The 33rd International Conference on Machine Learning, pp. 2829 2838, 2016. Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proceedings of the 35th International Conference on Machine Learning, pp. 1861 1870, 2018. E. T. Jaynes. Information theory and statistical mechanics. Phys. Rev., 106:620 630, 1957. Taisuke Kobayashi. Student-t policy in reinforcement learning to acquire global optimum of robot control. Applied Intelligence, 49:4335 4347, 2019. Ilya Kostrikov, Ashvin Nair, and Sergey Levine. Offline reinforcement learning with implicit q-learning. In International Conference on Learning Representations, 2022. Kenneth L. Lange, Roderick J. A. Little, and Jeremy M. G. Taylor. Robust statistical modeling using the t distribution. Journal of the American Statistical Association, 84:881 896, 1989. Kyungjae Lee, Sungjoon Choi, and Songhwai Oh. Sparse markov decision processes with causal sparse tsallis entropy regularization for reinforcement learning. IEEE Robotics and Automation Letters, 3:1466 1473, 2018. Kyungjae Lee, Sungyub Kim, Sungbin Lim, Sungjoon Choi, Mineui Hong, Jae In Kim, Yong-Lae Park, and Songhwai Oh. Generalized tsallis entropy reinforcement learning and its application to soft mobile robots. In Robotics: Science and Systems XVI, pp. 1 10, 2020. Yuhan Li, Wenzhuo Zhou, and Ruoqing Zhu. Quasi-optimal reinforcement learning with continuous actions. In The Eleventh International Conference on Learning Representations, 2023. Andr A F. T. Martins, Marcos Treviso, Ant A³nio Farinhas, Pedro M. Q. Aguiar, M A rio A. T. Figueiredo, Mathieu Blondel, and Vlad Niculae. Sparse continuous distributions and fenchelyoung losses. Journal of Machine Learning Research, 23(257):1 74, 2022. Hiroshi Matsuzoe and Atsumi Ohara. Geometry of q-exponential families. In Recent Progress in Differential Geometry and Its Related Fields, pp. 55 71, 2011. Jincheng Mei, Chenjun Xiao, Csaba Szepesvari, and Dale Schuurmans. On the global convergence rates of softmax policy gradient methods. In Proceedings of the 37th International Conference on Machine Learning, volume 119, pp. 6820 6829, 2020. Ashvin Nair, Murtaza Dalal, Abhishek Gupta, and Sergey Levine. {AWAC}: Accelerating online reinforcement learning with offline datasets, 2021. Jan Naudts. Deformed exponentials and logarithms in generalized thermostatistics. Physica Astatistical Mechanics and Its Applications, 316:323 334, 2002. Jan Naudts. The q-exponential family in statistical physics. Journal of Physics: Conference Series, pp. 012003, 2010. Samuel Neumann, Sungsu Lim, Ajin George Joseph, Yangchen Pan, Adam White, and Martha White. Greedy actor-critic: A new conditional cross-entropy method for policy improvement. In The Eleventh International Conference on Learning Representations, 2023. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32, pp. 8024 8035, 2019. Ben Peters, Vlad Niculae, and Andr e F. T. Martins. Sparse sequence-to-sequence models. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 1504 1519, 2019. Published as a conference paper at ICLR 2025 Kaare Brandt Petersen and Michael Syskind Pedersen. The matrix cookbook. 2012. Timothy Sears. Generalized Maximum Entropy, Convexity and Machine Learning. Ph D thesis, The Australian National University and Computer Science Laboratory, Research School of Information Sciences and Engineering, 2008. Umut Simsekli, Levent Sagun, and Mert Gurbuzbalaban. A tail-index analysis of stochastic gradient noise in deep neural networks. In Proceedings of the 36th International Conference on Machine Learning, Proceedings of Machine Learning Research, pp. 5827 5837, 2019. Richard S. Sutton and Andrew G. Barto. Reinforcement Learning: An Introduction. A Bradford Book, Cambridge, MA, USA, 2018. Richard S. Sutton, David Mc Allester, Satinder Singh, and Yishay Mansour. Policy gradient methods for reinforcement learning with function approximation. In Advances in Neural Information Processing Systems (NIPS), pp. 1057 1063, 1999. H. Suyari and M. Tsukada. Law of error in tsallis statistics. IEEE Transactions on Information Theory, 51(2):753 757, 2005. William J. Thistleton, John A. Marsh, Kenric Nelson, and Constantino Tsallis. Generalized box m Uller method for generating q-gaussian random deviates. IEEE Transactions on Information Theory, 53:4805 4810, 2007. C. Tsallis. Introduction to Nonextensive Statistical Mechanics: Approaching a Complex World. Springer New York, 2009. ISBN 9780387853581. Chenjun Xiao, Han Wang, Yangchen Pan, Adam White, and Martha White. The in-sample softmax for offline reinforcement learning. In The Eleventh International Conference on Learning Representations, 2023. Haoran Xu, Li Jiang, Jianxiong Li, and Xianyuan Zhan. A policy-guided imitation approach for offline reinforcement learning. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho (eds.), Advances in Neural Information Processing Systems, 2022. Haoran Xu, Li Jiang, Jianxiong Li, Zhuoran Yang, Zhaoran Wang, Victor Wai Kin Chan, and Xianyuan Zhan. Offline RL with no OOD actions: In-sample learning via implicit value regularization. In The Eleventh International Conference on Learning Representations, 2023. Lingwei Zhu, Zheng Chen, Matthew Schlegel, and Martha White. Generalized munchausen reinforcement learning using tsallis kl divergence. In Advances in Neural Information Processing Systems (Neur IPS), 2023. Lingwei Zhu, Matthew Schlegel, Han Wang, and Martha White. Offline reinforcement learning with tsallis regularization. Transactions on Machine Learning Research, 2024. Brian D. Ziebart. Modeling Purposeful Adaptive Behavior with the Principle of Maximum Causal Entropy. Ph D thesis, Carnegie Mellon University, Carnegie Mellon University, 2010. Published as a conference paper at ICLR 2025 The Appendix is organized into the following sections. In section A we summarize the multivariate form of q-exp policies and derive gradients of their log-likelihood. In section C we discuss the connection between the q-exp family and the entropy regularization literature. Based on this, we further discuss how different algorithms may prefer specific policies depending on its actor loss. We then provide implementation details including hyperparameters and how to sample from q-Gaussian in section D. Lastly we provide additional experimental results in section E. A Multivariate q-exp Policies and Log-likelihood B Connection to Entropy Regularization C Actor Losses D Implementation Details E Additional Results A MULTIVARIATE DENSITY OF q-EXP POLICIES Policy Density ln πs(a) 2 |Σ| 1 2 exp 1 2 (a µ) Σ 1 (a µ) Eq. (13) Student s t Γ( N+ν ν (a µ) Σ 1 (a µ) i N+ν q-Gaussian (q < 1) (1 q) N 2 |Σ| 1 2 expq 1 2 (a µ) Σ 1 (a µ) q-Gaussian (1 < q < 3) (q 1) N 2 Γ( 3 q 2(q 1) + N Γ( 3 q 2(q 1))π N 2 |Σ| 1 2 expq 1 2 (a µ) Σ 1 (a µ) Table 3: Multivariate q-exp policies and gradients of log-likelihood. In Table 3 we show multivariate density of the q-exp policies introduced in the main text. Note that multivariate Student s t is constructed based on the assumption that a diagonal Σ leads to independent action dimensions, same as the Gaussian policy. On the other hand, for q-Gaussian this is no longer true, since a diagonal Σ does not lead to product of univariate densities. In the main text we showed their one-dimensional cases for simplicity. For experiments the multivariate densities were used for experiments. We now derive their gradients of log-likelihood with respect to parameters. The following equations will be used frequently (Petersen & Pedersen, 2012): µ (a µ) Σ 1(a µ) = 2Σ 1(a µ), (10) Σ ln |Σ| = Σ 1 , (11) Σ(a µ) Σ 1(a µ) = Σ 1(a µ)(a µ) Σ 1. (12) With these tools in hand, the following gradient expressions can be readily derived. Published as a conference paper at ICLR 2025 A.1 GAUSSIAN Being a member of the exponential family, the gradient of Gaussian log-likelihood allows straightforward derivation by using Eq. (10)-Eq. (12): ln πs(a) = N 2(a µ) Σ 1(a µ) µ ln πs(a) = Σ 1(a µ), Σ ln πs(a) = 1 2 Σ 1 Σ 1(a µ)(a µ) Σ 1 . A.2 STUDENT S T In addition to µ, Σ, Student s t policy has an additional learnable parameter degree of freedom ν. Recall that ν = 1 corresponds to the Cauchy s distribution, while numerically with ν 30 it can be seen as a Gaussian distribution. ln πs(a) = ln Γ N + ν 2 ln |Σ| N + ν ν (a µ) Σ 1(a µ) µ ln πs(a) = N + ν ν Σ 1(a µ) 1 + 1 ν (a µ) Σ 1(a µ), Σ ln πs(a) = 1 Σ 1 (N + ν)Σ 1(a µ)(a µ) Σ 1 ν + (a µ) Σ 1(a µ) ν ln πs(a) = ψ N + ν ν (a µ) Σ 1(a µ) 1 ν (a µ) Σ 1(a µ) ν + (a µ) Σ 1(a µ), (14) where ψ( ) is the digamma function. For µ and Σ we again leveraged Eq. (10)-Eq. (12). A.3 q-GAUSSIAN Since we do not parametrize the entropic index q, the gradients of log-likelihood with respect to µ, Σ are the same for both heavyand light-tailed q-Gaussian. Therefore, we focus on the light-tailed case q < 1 and absorb into the constant C the terms only related to q. ln πs(a) = ln C 1 2 ln |Σ| + 1 1 q ln 1 1 q 2 (a µ) Σ 1(a µ) µ ln πs(a) = 1 1 q (1 q)Σ 1(a µ) 1 1 q 2 (a µ) Σ 1(a µ) + = Σ 1(a µ) 2(a µ) Σ 1(a µ) 1 q , Σ ln πs(a) = 1 Σ 1 Σ 1(a µ)(a µ) Σ 1 2(a µ) Σ 1(a µ) 1 q (15) It is interesting to see that the gradients of q-Gaussian log-likelihood can be seen as the Gaussian counterparts scaled by the reciprocal of expq( )1 q. Since expq can take on zero values when q < 1, the gradients as well as the log-likelihood function may be undefined outside the support. However, this does not happen for heavy-tailed q-Gaussian 1 < q < 3. To make these policies suitable for deep reinforcement learning, we discuss in Appendix D how to parametrize the policies using neural networks. B CONNECTION TO ENTROPY REGULARIZATION The q-exp family provides a general class of stochastic policies. But perhaps more importantly, they can be derived as solutions to the maximum Tsallis entropy principle (Suyari & Tsukada, 2005; Published as a conference paper at ICLR 2025 Furuichi, 2010), generalizing the maximum Shannon entropy principle (Jaynes, 1957; Gr unwald & Dawid, 2004; Ziebart, 2010). We discuss both principles in equation 16. For notational convenience, we define the inner product for any two functions F1, F2 R|S| |A| over actions as F1, F2 R|S|. We write Fs to express the function s dependency F on state s. Often Fs R|A|, whenever its component is of concern, we denote it by Fs(a). B.1 BOLTZMANN-GIBBS REGULARIZATION Consider a regularized policy as the solution to the following regularization problem: πΩ,s = arg max πs A πs, Qs Ω(πs), (16) where Ωis a proper, lower semi-continuous, strictly convex function. We can absorb the regularization coefficient τ > 0 into Ωby Ω:= τ Ω. It is a classic result that at the limit τ 0 the unregularized optimal action is recovered: limτ 0 πτ Ω,s = 1{a = a }, i.e., a that maximizes Qs. One of the most well-studied regularizers is the negative Shannon entropy Ω(πs) = πs, ln πs , which leads to the Boltzmann-Gibbs policy πBG,s(a) = exp Qs(a) Zs . Another popular choice is the KL divergence Ω(πs) = πs, ln πs ln µs for some reference policy µs. The regularized policy is πKL,s(a) = µs(a) exp Qs(a) Zs . Notice that it is also a member of the exponential family by writing πKL,s(a) = exp Qs(a) Zs + ln µs(a) . B.2 TSALLIS REGULARIZATION Originally, the deformed logarithm function was introduced in the statistical physics to generalize the Shannon entropy by deforming the logarithm contained in it (Naudts, 2010). Consider replacing Shannon entropy in equation 16 with the negative Tsallis entropy Ωq(πs) = 1 q 1 ( 1, πq s 1). It has been shown that Ωq(πs) leads to following regularized policy: πΩq,s(a) = exp2 q Qs(a) Z q,s . (17) We see that when q = 2, it recovers the sparsemax policy introduced in Section 3.2. As indicated by (Zhu et al., 2023), the effect of different q ( , 1) lies in the extent of thresholding. One can also consider regularization by the Tsallis KL divergence Dq KL(πs || µs) := D πs, lnq µs πs (Furuichi et al., 2004). Likewise to the KL case, µ is typically taken to be the last policy, in which the regularized policy is the product of two q-exp functions. It is worth noting that there are other regularization functionals that can induce q-exp policies. One of the prominent examples is the α-entropy/divergence, which can be defined by simply letting p = 1 q in Ωq(πs) (Peters et al., 2019; Belousov & Peters, 2019). It is shown in (Xu et al., 2022; 2023) that when α = 1 it induces the sparsemax policy. Therefore, q-exp policies can also be viewed as solutions to the α regularization. B.3 TSALLIS ADVANTAGE WEIGHTED ACTOR CRITIC An advantage of q-exp (resp. exp) policies is it may improve the consistency of algorithms that explicitly mimics a q-exp (resp. exp) policy. For example, Tsallis Advantage Weighted Actor Critic (TAWAC) proposed to use a light-tailed q-exp policy for offline learning (Zhu et al., 2024). However, TAWAC was implemented with Gaussian, which amounts to approximating a light-tailed distribution using one with infinite support. Let πD denote the empirical behavior policy and D the offline dataset. TAWAC minimizes the following actor loss, where we ignore the parametrization of value functions: L(ϕ) := Es D [DKL(πTKL,s || πϕ,s)] = E s D a πD expq Qs(a) Vs ln πϕ,s(a) , (18) where πTKL,s(a) πD,s(a) expq τ 1 (Qs(a) Vs) denotes the Tsallis KL regularized policy. We can generalize TAWAC to online learning by simply changing the expectation to be w.r.t. arbitrary behavior policy. It is clear that depending on q , choosing Gaussian as πϕ may incur inconsistency with the theory. A q-exp policy would be more suitable and could improve the performance. As evidenced by our experimental results, heavy tailed policies indeed further improve the performance of TAWAC by a large margin. Published as a conference paper at ICLR 2025 C ACTOR LOSSES To help understand when exp-family policies (resp. q-exp) may be more preferable, we compare the actor loss functions of the algorithms in the experiment section. C.1 ONLINE ALGORITHMS Soft Actor-Critic. SAC minimizes the following KL loss for the actor LSAC(ϕ) := Es B [DKL(πϕ( |s) || πBG( |s))] = Es B exp τ 1Q(s, ) where states are sampled from replay buffer B. The parametrized policy πϕ is projected to be close to the BG policy, therefore it is reasonable to expect that choosing πϕ from the exp-family may be more preferable. Depending on action values, BG can be skewed, multi-modal. Therefore, the symmetric, unimodal Gaussian may not be able to fully capture these characteristics. Greedy Actor-Critic. Greedy AC maintains an additional proposal policy besides the actor. The proposal policy is responsible for producing actions from which the top k% of actions are used to update the actor. The proposal policy itself is updated similarly but with an entropy bonus encouraging exploration. To simplify notations, we use I(s) to denote the set containing top k% actions given s. LGreedy AC, prop(ϕ) := E s B a I(s) [ ln πϕ(a|s) H (πϕ( |s))] , LGreedy AC, actor( ϕ) := E s B a I(s) ln π ϕ(a|s) . Greedy AC maximizes log-likelihood of the actor and proposal policy. These policies impose no constraints on the functional form of π. Online Tsallis AWAC. Online TAWAC is extended to condition on the behavior policy that collects experiences πtheory(a|s) πbehavior(a|s) expq Q(s,a) V (s) LTAWAC(ϕ) : = Es B [DKL(πtheory( |s) || πϕ( |s))] = E s B a π ϕ Q(s, a) V (s) ln πϕ(a|s) , where the condition a π ϕ is because the target policy is used to sample actions. Since Tsallis AWAC explicitly minimizes KL loss to a q-exp policy, which can be light-tailed/heavy-tailed depending on q. Therefore, choosing a q-exp πϕ could lead to better performance. C.2 OFFLINE ALGORITHMS AWAC. Advantage Weighted Actor-Critic (AWAC) is the basis of many algorithms. AWAC minimizes the following actor loss: LAWAC(ϕ) := E s D a πD exp Q(s, a) V (s) ln πϕ(a|s) , which is derived as the result of minimizing KL loss DKL(πD || πϕ) and applying the trick in Eq. 18, i.e., πtheory(a|s) πD(a|s) exp Q(s,a) V (s) τ = exp Q(s,a) V (s) τ ln πD(a|s) . However, the shape of this policy can be multi-modal and skewed depending on the values and πD. It is visible from experimental results that Beta and Squashed Gaussian have similar performance. IQL. In contrast to AWAC, Implicit Q-Learning (IQL) does not have an explicit actor learning procedure and uses LAWAC(ϕ) as a means for policy extraction from the learned value functions. The exponential advantage function acts simply as weights. Therefore, IQL does not assume the functional form of πϕ. In AC. In-Sample Actor-Critic (In AC) proposed to impose an in-sample constraint on the entropyregularized BG policy. As such, the dependence on the behavior policy is moved into the exponentialadvantage weighting function: LIn AC(ϕ) := E s D a πD exp Q(s, a) V (s) τ ln πD(a|s) ln πϕ(a|s) . Published as a conference paper at ICLR 2025 Figure 10: Beta distribution with α < 1, β < 1 takes on a bowl shape rather than a bell shape. The shape can also be skewed as well as symmetric. As a result, In AC is not as sensitive to the advantage weighting as AWAC does, which implies that In AC may favor an exp πϕ but less than AWAC. Offline Tsallis AWAC. The offline case of Tsallis AWAC is same as the online case except the change of expectation: LTAWAC(ϕ) := E s D a πD Q(s, a) V (s) ln πϕ(a|s) . Same with the online case, offline Tsallis AWAC may theoretically prefer a q-exp πϕ. TD3BC. In Appendix E we include additional results of TD3BC (Fujimoto & Gu, 2021), whose actor loss is obtained by simply augmenting the TD3 loss with a behavior cloning term: LTD3BC(ϕ) := E s D a πD h λQ(s, π(s)) (π(s) a)2i . The behavior cloning term is simply minimizing the L2 distance to actions in the dataset. Though another interpretation by (Xiao et al., 2023) is that this term can be understood as applying KL regularization to Gaussian policy. D IMPLEMENTATION DETAILS Details of our implementation is provided in this section. Specifically, we detail our design choices hyperparameters and network architectures. D.1 POLICIES We discuss how to parametrize Beta, Student s t and q-Gaussian policies. Specifically, we parametrize α, β for Beta policy; µ, Σ for q-Gaussian. In additional to location and scale, Student s t has an additional learnable parameter ν. For Student s t policy, we initialized a base DOF ν0 = 1 and learn ν by the softplus function. The Student s t policy therefore always has DOF ν > 1, which is equivalent to starting as the Cauchy s distribution. For Beta policy, we similarly constrain α, β to be the output of softplus function plus 1. This is because when α < 1, β < 1 the Beta policy takes on a bowl shape rather than a bell shape, see Figure 10. For Gaussian and q-Gaussian policies, we follow the standard practice to parametrize mean by the tanh activation and scale by the log-std transform. In the tested off-policy/offline algorithms, it is necessary to evaluate log-probability for offpolicy/offline actions stored in the buffer. For light-tailed q-Gaussian this can cause numerical issues since the evaluated actions may fall outside the support, incurring for log-probability. To avoid this issue, we sample a batch of on-policy actions from the q-Gaussian and replace the out-of-support actions with the nearest action in the L2 sense. Published as a conference paper at ICLR 2025 In our experiments, all environments had bounded action space. Squashed Gaussian and light-tailed q-Gaussian provide bounded output. However, Student s t, heavy tailed q-Gaussian and Gaussian have unbounded support. For these distributions, we clipped the sampled action to fit the action space of the task, without further modification on the density. The mean value is constrained using tanh function in distributions with unbounded support, except the standard Gaussian in offline learning. D.2 ONLINE EXPERIMENTS We used three classical control environments in the continuous action setting: Mountain Car (Sutton & Barto, 2018), Pendulum (Degris et al., 2012) and Acrobot (Sutton & Barto, 2018). All episodes are truncated at 1000 time steps. In Mountain Car, the action is the force applied to the car in [ 1, 1], and the agent receives a reward of -1 at every time step. In Pendulum, the action is the torque applied to the base of the pendulum in [ 2, 2] and the reward is defined by r = (θ2 +0.1 ( dθ dt )2 +0.001 a2) where θ denotes the angle, dθ dt is the derivative of time and a the torque applied. Finally, in acrobot, the action is the torque applied on the joint between two links in [ 1, 1] and the agent receives a reward of 1 per time step. Experiment settings: When sweeping different hyperparameter configurations, we pause the training every 10,000 time steps and then evaluate the learned policy by averaging the total reward over 3 episodes. However, when running the best hyperparameter configuration, we evaluate by freezing the policy every 1000 time steps and then computing the total reward obtained for 1 episode. Parameter sweeping: We sweep the hyperparameters with 5 independent runs and then evaluate the run configuration for 30 seeds. We select the best hyperparameters based on the overall area under curve. When running the best hyperparameter configurations, we discard the original 5 seeds used for the hyperparameter sweep in order to avoid the bias caused by hyperparameter selection. Details regarding the fixed and swept hyperparameters are provided in Table 4. Agent learning: We used a 2-layer network with 64 nodes on each layer and Re LU non-linearities. The batch size was 32. Agents used a target network for the critic, updated with polyak averaging with α = 0.01. Hyperparameter Value Critic Learning rate Swept in {1 10 2, 1 10 3, 1 10 4, 1 10 5} Critic learning rate multiplier for actor Swept in {0.1, 1, 10} Temperature Swept in {0.01, 0.1, 1} Discount rate 0.99 Hidden size of Value network 64 Hidden layers of Value network 2 Hidden size of Policy network 64 Hidden layers of Policy network 2 Minibatch size 32 Adam.β1 0.9 Adam.β2 0.999 Number of seeds for sweeping 10 Number of seeds for the best setting 30 Table 4: Default hyperparameters and sweeping choices for online experiments. D.3 OFFLINE EXPERIMENTS We use the Mu Jo Co suite from D4RL (Apache-2/CC-BY licence) (Fu et al., 2020) for offline experiments. The D4RL offline datasets all contain 1 million samples generated by a partially trained SAC agent. The name reflects the level of the trained agent used to collect the transitions. The Medium dataset contains samples generated by a medium-level (trained halfway) SAC policy. Medium-expert mixes the trajectories from the Medium level and that produced by an expert agent. Medium-replay consists of samples in the replay buffer during training until the policy reaches the Published as a conference paper at ICLR 2025 medium level of performance. In summary, the ranking of levels is Medium-expert > Medium > Medium-replay. Experiment settings: We conducted the offline experiment using 9 datasets provided in D4RL: halfcheetah-medium-expert, halfcheetah-medium, halfcheetah-medium-replay, hopper-mediumexpert, hopper-medium, hopper-medium-replay, walker2d-medium-expert, walker2d-medium, and walker2d-medium-replay. We run 5 agents: TAWAC, AWAC, IQL, In AC, and TD3BC. The results of TD3BC are posted in the appendix. For each agent, we tested 5 distributions: Gaussian, Squashed Gaussian, Beta, Student s t, and Heavy-tailed q-Gaussian. As offline learning algorithms usually require a distribution covering the whole action space, Light-tailed q-Gaussian is not considered in offline learning experiments. Each agent was trained for 1 106 steps. The policy was evaluated every 1000 steps. The score was averaged over 5 rollouts in the real environment; each had 1000 steps. Parameter sweeping: All results shown in the paper were generated by the best parameter setting after sweeping. We list the parameter setting in Table 5. Learning rate and temperature in TAWAC + medium datasets were swept as the experiments in their publication did not include the medium dataset. The best learning rates are reported in Table 6, and the temperatures are listed in Table 7. Hyperparameter Value Learning rate Swept in {3 10 3, 1 10 3, 3 10 4, 1 10 4} See the best setting in Table 6 Temperature Same as the number reported in the publication of each algorithm. Except in TAWAC + medium datasets, the value was swept in {1.0, 0.5, 0.01}. See the setting in Table 7 IQL Expectile 0.7 Discount rate 0.99 Hidden size of Value network 256 Hidden layers of Value network 2 Hidden size of Policy network 256 Hidden layers of Policy network 2 Minibatch size 256 Adam.β1 0.9 Adam.β2 0.99 Number of seeds for sweeping 5 Number of seeds for the best setting 10 Table 5: Default hyperparameters and sweeping choices for offline experiments. Agent learning: We used a 2-layer network with 256 nodes on each layer. The batch size was 256. Agents used a target network for the critic, updated with polyak averaging with α = 0.005. The discount rate was set to 0.99. Sampling. To give an intuition for sampling time, we drew 105 samples from a randomly initialized actor on two environments: Half Cheetah with 17-dim state and 6-dim action. The sparse q Gaussian, heavy-tailed q-Gaussian and Gaussian respectively cost (107.12, 72.09, 27.94) seconds. We confirmed that the methods in Alg. 1 were on the same magnitude to the Gaussian, but the sparse q-Gaussian cost more than the heavy-tailed due to more computation to produce low-variance samples. This is further confirmed by Hopper with 11-dim state, 3-dim action, where they costed (98.13, 65.17, 25.17) seconds. E FURTHER RESULTS Figure 11 shows the Manhattan plot of Soft-Actor-Critic (SAC) with all swept hyperparameters on the online classic control environments. Student-t and Gaussian both seem to have a similar behavior to hyperparameters. Although there is no definitive winner here, we can safely conclude that if we have a problem where Gaussian works, Student-t is very likely to work. Additionally, give the results Published as a conference paper at ICLR 2025 Dataset Distribution TAWAC AWAC IQL In AC TD3BC Half Cheetah-Medium-Expert Heavy-Tailed q-Gaussian 0.001 0.001 0.001 0.001 0.0003 Half Cheetah-Medium-Expert Squashed Gaussian 0.001 0.0003 0.0003 0.001 0.0003 Half Cheetah-Medium-Expert Gaussian 0.0003 0.0001 0.0003 0.0003 0.0003 Half Cheetah-Medium-Expert Beta 0.001 0.0003 0.001 0.001 0.001 Half Cheetah-Medium-Expert Student s t 0.001 0.0003 0.0003 0.0003 0.001 Half Cheetah-Medium-Replay Heavy-Tailed q-Gaussian 0.001 0.001 0.001 0.001 0.001 Half Cheetah-Medium-Replay Squashed Gaussian 0.001 0.0003 0.0003 0.001 0.003 Half Cheetah-Medium-Replay Gaussian 0.001 0.0001 0.0003 0.001 0.001 Half Cheetah-Medium-Replay Beta 0.001 0.0003 0.0003 0.001 0.001 Half Cheetah-Medium-Replay Student s t 0.001 0.0003 0.0003 0.0003 0.003 Half Cheetah-Medium Heavy-Tailed q-Gaussian 0.001 0.001 0.001 0.001 0.0003 Half Cheetah-Medium Squashed Gaussian 0.001 0.0003 0.001 0.001 0.0003 Half Cheetah-Medium Gaussian 0.0003 0.0001 0.0003 0.001 0.001 Half Cheetah-Medium Beta 0.001 0.001 0.001 0.001 0.0003 Half Cheetah-Medium Student s t 0.001 0.0003 0.001 0.001 0.001 Hopper-Medium-Expert Heavy-Tailed q-Gaussian 0.001 0.001 0.001 0.001 0.0001 Hopper-Medium-Expert Squashed Gaussian 0.001 0.001 0.001 0.001 0.0001 Hopper-Medium-Expert Gaussian 0.0003 0.0003 0.001 0.001 0.0001 Hopper-Medium-Expert Beta 0.001 0.001 0.001 0.003 0.0003 Hopper-Medium-Expert Student s t 0.001 0.0003 0.001 0.003 0.0001 Hopper-Medium-Replay Heavy-Tailed q-Gaussian 0.001 0.0001 0.001 0.0001 0.001 Hopper-Medium-Replay Squashed Gaussian 0.0001 0.0003 0.001 0.0003 0.001 Hopper-Medium-Replay Gaussian 0.0003 0.0003 0.001 0.0003 0.001 Hopper-Medium-Replay Beta 0.0001 0.0003 0.0003 0.003 0.003 Hopper-Medium-Replay Student s t 0.0003 0.0003 0.0003 0.0003 0.001 Hopper-Medium Heavy-Tailed q-Gaussian 0.003 0.001 0.001 0.001 0.0001 Hopper-Medium Squashed Gaussian 0.001 0.0003 0.001 0.0003 0.0001 Hopper-Medium Gaussian 0.001 0.001 0.0003 0.001 0.001 Hopper-Medium Beta 0.001 0.001 0.003 0.001 0.001 Hopper-Medium Student s t 0.001 0.001 0.001 0.001 0.0001 Walker2d-Medium-Expert Heavy-Tailed q-Gaussian 0.0003 0.001 0.001 0.001 0.0003 Walker2d-Medium-Expert Squashed Gaussian 0.001 0.001 0.0003 0.001 0.0003 Walker2d-Medium-Expert Gaussian 0.0003 0.0001 0.0003 0.001 0.001 Walker2d-Medium-Expert Beta 0.001 0.0003 0.001 0.001 0.001 Walker2d-Medium-Expert Student s t 0.001 0.0003 0.0003 0.0003 0.0003 Walker2d-Medium-Replay Heavy-Tailed q-Gaussian 0.0003 0.0003 0.003 0.0003 0.001 Walker2d-Medium-Replay Squashed Gaussian 0.001 0.0003 0.0003 0.001 0.001 Walker2d-Medium-Replay Gaussian 0.001 0.0003 0.0003 0.001 0.003 Walker2d-Medium-Replay Beta 0.001 0.0003 0.0003 0.001 0.0003 Walker2d-Medium-Replay Student s t 0.0003 0.0003 0.0003 0.001 0.001 Walker2d-Medium Heavy-Tailed q-Gaussian 0.003 0.001 0.001 0.001 0.0001 Walker2d-Medium Squashed Gaussian 0.001 0.001 0.001 0.001 0.0001 Walker2d-Medium Gaussian 0.001 0.0001 0.001 0.001 0.0001 Walker2d-Medium Beta 0.001 0.0003 0.003 0.001 0.0001 Walker2d-Medium Student s t 0.001 0.0003 0.001 0.001 0.0001 Table 6: Best learning rates for offline experiments. in the main text, Student s t more likely to perform better given the same hyperparameter sweeping range. Our additional offline results include all algorithm-policy combination on all environments. We also include TD3BC (Fujimoto & Gu, 2021) for comparison. Figure 12 shows the overall comparison with TD3. It is clear that Squashed Gaussian performs well and Beta can show slight improvements in some cases. Though it is visible that no much difference is shown except on the Medium-Replay data. We conjecture that the better performance of Squashed Gaussian and Beta could be due to the TD3BC behavior cloning loss. It is encouraged that policy closely approximates the actions from the dataset. Therefore, policies like Beta that can concentrate faster may be more advantageous. Figures 13 to 15 display boxplots of the combinations on environments of each level. Consistent observations to that in the main text can be drawn from these plots, but with the exception that in Figure 14 the environment-wise best combination is TAWAC + Student s t. TD3BC does not exhibit strong sensitivity to the choice of policy. Published as a conference paper at ICLR 2025 Figure 11: Manhattan plot of Soft-Actor-Critic (SAC) with all swept hyperparameters on the online classic control environments. The rewards on the y-axis are averaged over the final 10% of the total steps. Since different policy parameterizations have different numbers of runs in the sweep, we oversampled the smaller sweeps with replacement. From the plot of Acrobot, we observe that Studentt and Gaussian both respond similarly to changing hyper-parameters. Therefore, we hypothesize that if we have an environment where the Gaussian policy works, Student-t is also very likely to work. Additionally, from Figure 5 (left), we know that student-t is 75% more likely to outperform the Gaussian given the same hyperparameter sweeping range. Published as a conference paper at ICLR 2025 Dataset Distribution TAWAC AWAC IQL In AC TD3BC Half Cheetah-Medium-Expert Heavy-Tailed q-Gaussian 1.00 1.00 0.33 0.10 2.50 Half Cheetah-Medium-Expert Squashed Gaussian 1.00 1.00 0.33 0.10 2.50 Half Cheetah-Medium-Expert Gaussian 1.00 1.00 0.33 0.10 2.50 Half Cheetah-Medium-Expert Beta 1.00 1.00 0.33 0.10 2.50 Half Cheetah-Medium-Expert Student s t 1.00 1.00 0.33 0.10 2.50 Half Cheetah-Medium-Replay Heavy-Tailed q-Gaussian 0.01 1.00 0.33 0.50 2.50 Half Cheetah-Medium-Replay Squashed Gaussian 0.01 1.00 0.33 0.50 2.50 Half Cheetah-Medium-Replay Gaussian 0.01 1.00 0.33 0.50 2.50 Half Cheetah-Medium-Replay Beta 0.01 1.00 0.33 0.50 2.50 Half Cheetah-Medium-Replay Student s t 0.01 1.00 0.33 0.50 2.50 Half Cheetah-Medium Heavy-Tailed q-Gaussian 0.01 0.50 0.33 0.33 2.50 Half Cheetah-Medium Squashed Gaussian 0.01 0.50 0.33 0.33 2.50 Half Cheetah-Medium Gaussian 0.01 0.50 0.33 0.33 2.50 Half Cheetah-Medium Beta 0.01 0.50 0.33 0.33 2.50 Half Cheetah-Medium Student s t 0.01 0.50 0.33 0.33 2.50 Hopper-Medium-Expert Heavy-Tailed q-Gaussian 0.50 1.00 0.33 0.01 2.50 Hopper-Medium-Expert Squashed Gaussian 0.50 1.00 0.33 0.01 2.50 Hopper-Medium-Expert Gaussian 0.50 1.00 0.33 0.01 2.50 Hopper-Medium-Expert Beta 0.50 1.00 0.33 0.01 2.50 Hopper-Medium-Expert Student s t 0.50 1.00 0.33 0.01 2.50 Hopper-Medium-Replay Heavy-Tailed q-Gaussian 0.50 0.50 0.33 0.50 2.50 Hopper-Medium-Replay Squashed Gaussian 0.50 0.50 0.33 0.50 2.50 Hopper-Medium-Replay Gaussian 0.50 0.50 0.33 0.50 2.50 Hopper-Medium-Replay Beta 0.50 0.50 0.33 0.50 2.50 Hopper-Medium-Replay Student s t 0.50 0.50 0.33 0.50 2.50 Hopper-Medium Heavy-Tailed q-Gaussian 0.50 0.50 0.33 0.10 2.50 Hopper-Medium Squashed Gaussian 0.50 0.50 0.33 0.10 2.50 Hopper-Medium Gaussian 0.50 0.50 0.33 0.10 2.50 Hopper-Medium Beta 0.50 0.50 0.33 0.10 2.50 Hopper-Medium Student s t 0.01 0.50 0.33 0.10 2.50 Walker2d-Medium-Expert Heavy-Tailed q-Gaussian 0.01 0.10 0.33 0.10 2.50 Walker2d-Medium-Expert Squashed Gaussian 0.01 0.10 0.33 0.10 2.50 Walker2d-Medium-Expert Gaussian 0.01 0.10 0.33 0.10 2.50 Walker2d-Medium-Expert Beta 0.01 0.10 0.33 0.10 2.50 Walker2d-Medium-Expert Student s t 0.01 0.10 0.33 0.10 2.50 Walker2d-Medium-Replay Heavy-Tailed q-Gaussian 0.50 0.10 0.33 0.50 2.50 Walker2d-Medium-Replay Squashed Gaussian 0.50 0.10 0.33 0.50 2.50 Walker2d-Medium-Replay Gaussian 0.50 0.10 0.33 0.50 2.50 Walker2d-Medium-Replay Beta 0.50 0.10 0.33 0.50 2.50 Walker2d-Medium-Replay Student s t 0.50 0.10 0.33 0.50 2.50 Walker2d-Medium Heavy-Tailed q-Gaussian 0.01 0.10 0.33 0.33 2.50 Walker2d-Medium Squashed Gaussian 1.00 0.10 0.33 0.33 2.50 Walker2d-Medium Gaussian 1.00 0.10 0.33 0.33 2.50 Walker2d-Medium Beta 1.00 0.10 0.33 0.33 2.50 Walker2d-Medium Student s t 0.01 0.10 0.33 0.33 2.50 Table 7: Temperature settings for offline experiments. Table 8 examined the accumulated probabilities that fell on each It can be seen that the Student s t and the Gaussian tended to increasingly put more densities on the boundaries. This is in sheer contrast to the heavy-tailed q-Gaussian that put the majority of probability density within the boundary. This may explain the better performance of TAWAC + heavy-tailed q-Gaussian. Lastly, for all of the results shown above, their learning curves are shown in Figures 16 to 20. We smoothed the curves with window size 10 for better visualization. Published as a conference paper at ICLR 2025 Policy # Updates 0 100 200 300 400 Heavy-tailed q-Gaussian (24.39, 13.19) (45.23, 2.36) (45.49, 2.04) (45.52, 1.98) (45.54, 1.89) Student s t (148.43, 71.23) (198.89, 45.30) (205.04, 37.04) (207.15, 32.96) (207.00, 33.84) Gaussian (190.96, 65.89) (206.92, 53.05) (211.77, 39.08) (213.57, 33.71) (214.39, 31.26) Table 8: The summation of probability density accumulated on the left and the right edge in Figure 9 before clipping. Each pair indicates the left and right edge. The Student s t and the Gaussian increasing put more densities on the edges as compared to the heavy-tailed q-Gaussian. TAWAC AWAC IQL In AC TD3BC 40% Medium-Expert Data TAWAC AWAC IQL In AC TD3BC Medium Data TAWAC AWAC IQL In AC TD3BC Medium-Replay Data Proportion to Squashed Gaussian Performance Gaussian Beta Student's t Heavy-Tailed q-Gaussian Figure 12: Relative improvement to the Squashed Gaussian policy, averaged over environments. Black vertical lines at the top indicate one standard error. For TD3BC, Beta policy outperforms the Squashed Gaussian on Medium-Expert and Medium-Replay. Normalized Return Half Cheetah Hopper Walker2d Medium-Expert Data Gaussian Squashed Gaussian Beta Student's t Heavy-Tailed q-Gaussian Figure 13: Normalized scores on Medium-Expert level datasets. The black bar shows the median. Boxes and whiskers show 1 and 1.5 interquartile ranges, respectively. Fliers are not plotted for uncluttered visualization. Environment-wise, In AC with heavy-tailed q-Gaussian is the top performer. Algorithm-wise, heavy-tailed or/and Student s t can improve or match the performance of the Squashed Gaussian except AWAC. With TD3BC no significant difference between policies is observed. Published as a conference paper at ICLR 2025 Normalized Return Half Cheetah Hopper Walker2d Medium Data Gaussian Squashed Gaussian Beta Student's t Heavy-Tailed q-Gaussian Figure 14: Normalized scores on Medium level datasets. The black bar shows the median. Boxes and whiskers show 1 and 1.5 interquartile ranges, respectively. Fliers are not plotted for uncluttered visualization. Environment-wise, In AC with heavy-tailed q-Gaussian is the top performer. Algorithmwise, heavy-tailed q-Gaussian has observed significant performance drop with AWAC and In AC on Hopper and Walker2d. With TD3BC no significant difference between policies is observed. Normalized Return Half Cheetah Hopper Walker2d Medium-Replay Data Gaussian Squashed Gaussian Beta Student's t Heavy-Tailed q-Gaussian Figure 15: Normalized scores on Medium-Replay level datasets. The black bar shows the median. Boxes and whiskers show 1 and 1.5 interquartile ranges, respectively. Fliers are not plotted for uncluttered visualization. Environment-wise, TAWAC + heavy-tailed q-Gaussian is the best performer. Algorithm-wise, Student s t is stable and can match or improve on the performance of (Squashed) Gaussian. Published as a conference paper at ICLR 2025 Figure 16: TAWAC learning curves in all datasets. Columns show different environments and rows are the levels of the environments. x-axis denotes the number of steps ( 104), and y-axis is the normalized score. Each curve was smoothed with window size 10. Figure 17: AWAC learning curves in all datasets. Columns show different environments and rows are the levels of the environments. x-axis denotes the number of steps ( 104), and y-axis is the normalized score. Each curve was smoothed with window size 10. Published as a conference paper at ICLR 2025 Figure 18: IQL learning curves in all datasets. Columns show different environments and rows are the levels of the environments. x-axis denotes the number of steps ( 104), and y-axis is the normalized score. Each curve was smoothed with window size 10. Figure 19: In AC learning curves in all datasets. Columns show different environments and rows are the levels of the environments. x-axis denotes the number of steps ( 104), and y-axis is the normalized score. Each curve was smoothed with window size 10. Published as a conference paper at ICLR 2025 Figure 20: TD3+BC learning curves in all datasets. Columns show different environments and rows are the levels of the environments. x-axis denotes the number of steps ( 104), and y-axis is the normalized score. Each curve was smoothed with window size 10.