# distributional_paretooptimal_multiobjective_reinforcement_learning__0ac2a9d5.pdf Distributional Pareto-Optimal Multi-Objective Reinforcement Learning Xin-Qiang Cai1 , Pushi Zhang2 , Li Zhao2, Jiang Bian2, Masashi Sugiyama3,1 Ashley J. Llorens2 1 The University of Tokyo, Tokyo, Japan 2 Microsoft Research Asia, Beijing, China 3 RIKEN AIP, Tokyo, Japan Multi-objective reinforcement learning (MORL) has been proposed to learn control policies over multiple competing objectives with each possible preference over returns. However, current MORL algorithms fail to account for distributional preferences over the multi-variate returns, which are particularly important in realworld scenarios such as autonomous driving. To address this issue, we extend the concept of Pareto-optimality in MORL into distributional Pareto-optimality, which captures the optimality of return distributions, rather than the expectations. Our proposed method, called Distributional Pareto-Optimal Multi-Objective Reinforcement Learning (DPMORL), is capable of learning distributional Pareto-optimal policies that balance multiple objectives while considering the return uncertainty. We evaluated our method on several benchmark problems and demonstrated its effectiveness in discovering distributional Pareto-optimal policies and satisfying diverse distributional preferences compared to existing MORL methods. 1 Introduction Apart from most Reinforcement Learning (RL) works that consider the scalar reward of a task [1, 2, 3, 4], Multi-Objective Reinforcement Learning (MORL) has recently received extensive attention due to its adeptness at managing intricate decision-making issues with multiple conflicting objectives [5, 6, 7, 8]. In many multi-objective tasks, the relative preferences of users over different objectives are typically indeterminate a priori. Consequently, MORL s primary aim is to learn a variety of optimal policies under different preferences to approximate the Pareto frontier of optimal solutions. It has been demonstrated that MORL can significantly reduce the reliance on scalar reward design for objective combinations and dynamically adapt to the varying preferences of different users. However, numerous real-world situations involve not only unknown relative preferences across multiple objectives, but also uncertain preferences in return distributions. These may include preference in risk [9, 10], safety conditions [11], and non-linear user s utility [12]. Consider, for instance, autonomous driving scenarios where agents need to strike a balance between safety and efficiency objectives. Different users might possess varying levels of risk tolerance. Some may demand high safety performance, tolerating lower expected efficiency, while others may seek a more balanced performance between safety and efficiency. Current MORL methods, by focusing exclusively on expected values with linear preferences, may not adequately capture multi-objective risk-sensitive preferences, hence being unable to deliver diverse policies catering to users with varied risk preferences. In this work, we broaden the concept of Pareto-optimality in MORL to encompass distributional Pareto-optimality, which prioritizes the optimality of return distributions over mere expectations, as Equal contribution. Work done during an internship at Microsoft Research Asia. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). (a) MORL s Pareto-front (b) Distributional MORL s Paretofront Figure 1: The comparison of learning targets between the traditional MORL tasks and distributional MORL tasks, in which f1 and f2 are two (conflicting) objectives. depicted in Figure 1. From a theoretical perspective, we define Distributional Pareto-Optimal (DPO) policies, which capture the optimality of multivariate distributions through stochastic dominance [13, 14]. Distributional Pareto-Optimality delineates the set of policies with optimal return distributions, which we formally establish as an extension of Pareto-optimality in MORL. On the practical side, we propose a novel method named Distributional Pareto-Optimal Multi Objective Reinforcement Learning (DPMORL)2, which aims to learn a set of DPO policies. We demonstrate that a policy achieving the highest expected utility for a given strictly increasing utility function is a DPO policy. To this end, we first learn diverse non-linear utility functions and then optimize policies under them [15]. Experimental outcomes on several benchmark problems attest to DPMORL s effectiveness in optimizing policies that meet preferences on multivariate return distributions. We posit that our proposed method significantly addresses the challenge of managing multiple conflicting objectives with unknown preferences on multivariate return distributions in complex decision-making situations. Our main contributions are listed as follows: 1. We introduce the concept of Distributional Pareto-Optimal (DPO) policies. This expands the notion of Pareto-optimality in MORL to include preferences over the entire distribution of returns, not just their expected values. This enables agents to express more nuanced preferences over policies and align their decisions more closely with their actual objectives. 2. We propose DPMORL, a new algorithm for learning DPO policies under non-linear utility functions. This algorithm accommodates a broad range of distributional preferences, thus offering a more flexible and expressive approach to MORL. 3. We execute extensive experiments on various MORL tasks, demonstrating the effectiveness of our approach in learning DPO policies. Our findings show that our algorithm consistently surpasses existing MORL methods in terms of optimizing policies for multivariate expected and distributional preferences, underscoring its practical benefits. 2 Related Work Multi-Objective Reinforcement Learning. MORL has emerged as a vibrant research area in the reinforcement learning community owing to its adeptness in managing intricate decision-making scenarios involving multiple conflicting objectives [5, 6]. A plethora of MORL algorithms have been put forth in the literature. For instance, [16] proposed the utilization of Generalized Policy Improvement (GPI) to infer a set of policies, employing Optimistic Linear Support (OLS) for dynamic reward weight exploration [17]. This GPI-based learning process was further enhanced by [7] through the incorporation of a world model to augment sample efficiency. [8] proposed an evolutionary approach to learn policies with diverse weights. However, our proposed approach diverges from these by leveraging utility functions to guide policy learning [18, 19]. The utility-based paradigm [6] accentuates the user s utility in decision-making problems, capitalizing on known information about the user s utility function and permissible policy types. Such methods scalarize the multidimensional objectives into a single reward function, enabling traditional RL algorithms to infer 2The code is available on https://github.com/zpschang/DPMORL. desirable policies, while also respecting axiomatic principles where necessary. Several studies have utilized non-linear utility functions to guide policy learning [20, 12], catering to more intricate preferences compared to linear ones but still in view of the expectation of the Pareto front. In contrast, our work extends the Pareto-optimality to distributional Pareto-optimality by introducing a set of non-linear utility functions with distributional attributes to guide policy learning. Some of the recent works [21, 22, 23] in MORL defines Distributional Undominated Set or ESR set which is both very close to Distributionally Pareto-Optimal policies in this paper. These works propose novel designs to enhance policy improvement in distributional reinforcement learning algorithms to find the Distributional Undominated Set. Compared to these works, we build novel theoretical results on the conditions of DPO policies and a novel algorithm (DPMORL) of learning DPO policies based on reward shaping that can be built on the top of any online RL algorithm. Distributional Reinforcement Learning. Distributional RL extends traditional RL by modeling the entire distribution of returns, rather than just their expected values [24, 25]. This approach has been shown to improve both learning efficiency and policy quality in a variety of single-objective RL tasks [26, 27]. Key algorithms in this area include Categorical DQN (C51) [24], Quantile Regression DQN (QR-DQN) [26], and Distributional MPO [27]. Despite the success of these distributional RL algorithms in single-objective settings, their direct extension to MORL has been limited due to the added complexity of handling multiple conflicting objectives and expressing preferences over the distribution of returns. On the other hand, the insights and techniques developed in distributional RL can provide a valuable foundation for incorporating distributional preferences into MORL, as demonstrated by our proposed approach. Risk-Sensitive and Safe Reinforcement Learning. Risk-sensitive and safe RL are special cases of MORL that accentuate specific facets of reward distributions [6]. Risk-sensitive RL primarily considers reward variance, striving to optimize policies that negotiate the trade-off between expected return and risk [28, 29, 30]. Conversely, safe RL prioritizes constraints on the agent s behavior, ensuring adherence to specified safety criteria throughout the learning process [31, 32, 33]. Some studies have proposed the integration of constraints into the state space, constructing new Markov Decision Processes [11]. Others have explored the distributional aspects of rewards, investigating the implications of distributional RL on risk-sensitive and safety-oriented decision-making [26, 24]. However, our proposed setting is more general in terms of objectives, as it considers a broader range of user preferences and captures the entire distribution of rewards, rather than focusing solely on specific aspects such as risk, variance, or constraints. By extending MORL to incorporate distributional properties, our approach enables the learning of distributional Pareto-optimal policies that cater to diverse user preferences and offer better decision-making in a wide range of real-world applications. 3 Preliminaries: Multi-Objective Reinforcement Learning Similar to the procedure of RL [34, 35], in MORL, the agent interacts with an environment modeled as a Multi-Objective Markov Decision Process (MOMDP) with multi-dimensional reward functions. A Multi-Objective MDP is defined by a tuple (S, A, P, P0, R, γ, T), where S is the state space, A is the action space, P is the state transition function, P0(s) is the initial state distribution, R : S A S RK is a vectored reward function and K is the number of objectives, γ is the discount factor, T is the total timesteps3. The goal of the agent is to learn a set of Pareto-optimal policies, which represent the best possible trade-offs among the different objectives. One popular methodology for MORL problems is the utility-based method, which combines the multi-dimensional reward functions into a single scalar reward function using a weighted sum or another aggregation method [36]. The intuition is to map the agent s preferences over different objectives to a scalar value for the agent training. Given a weight vector w = (w1, w2, . . . , w K), with wi representing the importance of the i-th objective, the scalarized reward function is defined as rscalar(s, a) = PK i=1 wiri(s, a). The agent then solves the scalarized MDP by optimizing its policy to maximize the expected scalar reward, using standard reinforcement learning algorithms like Q-learning or policy gradient methods. This approach can be straightforward to implement and has been shown to be effective in various MORL settings under expected preferences [37]. However, such a linear combination of each dimension of the reward cannot deal with the preferences with distributional considerations. 3We assume that current time t is a part of state to accommodate for the finite horizon setting. 4 Distributionally Pareto-Optimal Multi-Objective Reinforcement Learning In this section, we introduce our proposed theoretical framework and algorithms for extending MORL to handle distributional preferences. 4.1 Distributionally Pareto-Optimal Policies In this work, we consider the reward as a K-dimensional vector, where each element represents the reward for a specific objective. Given the MOMDP and a policy π, we define the random variable of multi-objective return as Z(π) = PT t=0 γtrt, where rt is the multi-dimensional reward at step t, and the states s0, s1, , s T and actions a0, a1, , a T are sampled from the MOMDP and the policy π respectively. The return distribution of policy π, denoted as µ(π), represents the joint distribution of the returns Z(π) following policy π. The utility function f is a non-decreasing function that maps a K-dimensional return into a scalar utility value for the agent [38, 39, 40], capturing the user s distributional preferences over the different objectives. The expected utility of policy π under the utility function f, represented as Ez µ(π)f(z), is the expected value of applying f to the return distribution µ(π). Our goal is to learn a set of policies that are distributionally Pareto-optimal, which means that their return distributions of each policy cannot dominate that of another. To measure such a relationship, here we first introduce the concept of stochastic dominance: Definition 1 (Stochastic Dominance for Multivariate Distribution). A multivariate distribution µ1 dominates another distribution µ2, denoted as µ1 SD µ2, if and only if µ1 = µ2 and for any non-decreasing utility function f : RK R, µ1 has greater expected utility than µ2, i.e. Ez µ1f(z) Ez µ2f(z). Definition 2 (Stochastic Dominance for Policies). A policy π1 stochastically dominates another policy π2, denoted as π1 SD π2, if and only if µ(π1) SD µ(π2), indicating π1 has greater expected utility than π2 under any non-decreasing utility function f. 1 0 1 2 3 4 5 6 z0 (a) Stochastic Dominance Example 1 5.0 2.5 0.0 2.5 5.0 7.5 10.0 12.5 z0 1 and 2 cannot dominate each other (b) Stochastic Dominance Example 2 2 0 2 4 6 8 10 z0 1 and 2 cannot dominate each other (c) Stochastic Dominance Example 3 Figure 2: Three 2D examples of stochastic dominance given synthetic returns of two policies. (a) π2 SD π1, since any non-decreasing utility function satisfies Ef(Z(π2)) Ef(Z(π1)). (b) π1 and π2 cannot dominate each other. For example, there exists f(z) = Re Lu(z0 5.0), such that Ef(Z(π2)) > Ef(Z(π1)), thus π1 cannot dominate π2. Similarly, π2 cannot dominate π1. Thus there is no dominant relationship between the two policies. (c) π1 and π2 cannot dominate each other, which is similar to (b). The definition of stochastic dominance extends univariate first-order stochastic dominance [13] into multivariate cases. Figure 2 provides some examples of stochastic dominance given 2D returns of two policies. The definition of stochastic dominance in policies allows for comparing the optimality of distributions between policies, which allows us to define the Distributionally Pareto-optimal policy: Definition 3 (Distributionally Pareto-Optimal Policy). Formally, π1 is Distributionally Pareto Optimal Policy if there does not exist a policy π2 such that µ(π2) SD µ(π1). In other words, π1 is considered Distributionally Pareto-Optimal (DPO) if it is not stochastically dominated by any other policy. DPO policies are essential in our framework as they represent the most desirable policies for users with different distributional preferences. To find such a policy, we have the following theorem: Theorem 1. If a policy π has optimal expected utility under some non-decreasing utility function f, where either utility function f is strictly increasing, or π is the only optimal policy with greatest expected utility under utility function f, then π is a Distributionally Pareto-Optimal policy. Also, any Pareto-Optimal policy is a Distributionally Pareto-Optimal policy. The proof of Theorem 1 is provided in Appendix A which utilizes a form of optimal transport theory for stochastic dominance. In practice, this theorem guarantees that the optimal policy of a strictly increasing non-linear utility function is a DPO policy, making it a suitable candidate for deployment in MORL problems. The strict increasing property of utility function f in Theorem 1 is crucial in the theoretical analysis for proving the distributional pareto-optimality. In the next section, we base on the result of this theorem to find optimal policies for a diverse set of utility functions, in order to learn the set of DPO policies. This result also shows that our definition of Distributional Pareto-Optimal policies is an extension of Pareto-optimal policies, allowing for a more diverse set of optimal policies to be captured. We also formally prove that optimal policies for risk-sensitive and safe constraint objectives belong to the set of Distributional Pareto-Optimal policies. This proves that DPO policies can successfully cover policies with diverse distributional preferences. The detailed proof is provided by Theorem 3 in Appendix A. 4.2 Distributionally Pareto-Optimal Multi-Objective Reinforcement Learning We now present our proposed algorithm, Distributionally Pareto-Optimal Multi-Objective Reinforcement Learning (DPMORL). The main idea of DPMORL is to learn a set of non-linear utility functions that can guide the agent to discover distributional Pareto-optimal policies. The algorithm proceeds in two stages: (1) generating the utility functions and (2) training the policies. 4.2.1 Utility Function Generation with Diversity-based Objective The first component of our algorithm focuses on generating a diverse set of plausible utility functions, upon which we find the optimal policies to find a diverse set of optimal policies. This is essential to ensure our method can accommodate various distributional preferences and adapt to different problem settings. To achieve this, we propose (1) Non-decreasing Neural Network for parameterizing a diverse set of non-linear utility functions (2) an objective function of minimum distance which encourages generating a diverse set of utility functions. Non-decreasing Neural Network. We employ non-decreasing neural network to parameterize the utility function. The use of a neural network allows us to represent complex, non-linear, and arbitrary continuous utility functions, while the non-decreasing constraint ensures that the utility function satisfies the desired properties for multi-objective problems. We ensure the non-decreasing property in neural networks by constraining the weight matrix to be non-negative and the activation function to be non-decreasing following existing work in convex neural networks [41] and QMIX [42], which can approximate any multivariate non-decreasing function with arbitrary small errors [43]. The implementation details of the Non-decreasing Neural Network are provided in Appendix B. Diversity-based Objective Function. We propose an objective function based on diversity for learning a diverse set of plausible utility functions. Specifically, we define fθ1, fθ2, , fθM to be the set of candidate utility functions, parameterized by θ1, , θM. For any given utility function fθi, the learning objective is defined as Jvalue(θi) = min j =i Ez U([0,1]K)[fθi(z) fθj(z)]2 (1) Jgrad(θi) = min j =i Ez1,z2 U([0,1]K) fθi(z2) fθi(z1) z2 z1 fθj(z2) fθj(z1) J(θi) = αJvalue(θi) + (1 α)Jgrad(θi) (3) Optimizing this objective function can encourage diversity among the generated utility functions within the normalized range [0, 1]K in both value and derivative, leading to more comprehensive coverage of potential user preferences. We minimize the objective function by gradient descent on non-decreasing neural networks to generate a set of N utility functions. To make the learned utility function strictly increasing for satisfying the constraints in Theorem 1, we add an additional term to the learned utility function as fθi(z) = fθi(z) + α K PK k=1 zk in our experiment with α = 0.01, and use fθi(z), i = 1, , M as the strictly increasing utility function to learn DPO policies. 4.2.2 Optimizing Policy with Utility-based Reinforcement Learning Once we have generated a diverse set of utility functions, the second component of our algorithm focuses on optimizing policies to maximize the expected utility. This process, which we call utilitybased RL, leverages the generated utility functions to guide the optimization of policies. By focusing on the expected utility, our method can efficiently balance the trade-offs between multiple objectives and distributions, ultimately generating policies that are more likely to align with user preferences. We show that the following utility-based reinforcement learning algorithm can effectively optimize the policy with respect to a given utility function. Algorithm 1 Utility-based Reinforcement Learning Input: policy π, an environment M = (S, A, P, P0, R, γ, T), utility function f Output: new policy π 1: Augment state space with S = S Z, where Z is the space of cumulative multivariate returns. 2: Let transition P0( ) and P( |(st, zt), at) with s0 P(s0), z0 = 0, and st+1 P( |st, at), zt+1 = zt + γtrt. 3: Let scalar reward function R as R((st, zt), at, (st+1, zt+1)) = γ t[f(zt+1) f(zt)]. 4: Optimize policy π under environment M = ( S, A, P, P0, R, γ, T) under off-the-shelf RL algorithm (such as PPO or SAC) to π . Briefly, Algorithm 1 augments the state space with the cumulative multi-objective returns, and transforms the multi-dimensional rewards into a scalar reward by the difference in the utility function f. The following result shows that the new scalar-reward environment M generated by Algorithm 1 has the same optimal policy as the optimal policy under utility function f: Theorem 2. The optimal policy π under environment M = ( S, A, P, P0, R, γ, T), with scalar reward function R((st, zt), at, (st+1, zt+1)) = γ t[f(zt+1) f(zt)] is the optimal policy under the utility function f, i.e. Ez µ(π ) [f(z)] = max π Ez µ(π) [f(z)] . An advantage of Algorithm 1 is that we can directly utilize off-the-shelf RL algorithms, such as PPO [44] and SAC [45] without any modification, which makes the algorithm easy to implement using widespread existing implementations of online RL algorithms. It is also important to note that our algorithm simplifies to optimizing the weighted sum of rewards in MORL when the utility function is linear. This implies that our method is a generalization of the linear utility function MORL approaches by accommodating a wide range of non-linear utility functions. This flexibility makes our algorithm particularly suited for problems where the user s preferences may not be adequately captured by a linear utility function. In the next experimental section, DPMORL only undergoes a single iteration: we initially generate a set of N utility functions as per the methodology detailed in Section 4.2.1. Subsequently, it optimizes a set of N policies using these generated utility functions, as outlined in Section 4.2.2. 5 Experiments In this section, we conducted several experiments under the setting of MORL. Through the experiments, we want to investigate the following questions: 1. Can Utility-based RL effectively learn policies with diverse distributional preferences? 2. Can DPMORL generate a set of diverse non-linear utility functions? 3. Can DPMORL obtain promising performance compared with state-of-the-art MORL methods in view of expected and distributional preferences? 5.1 Case Study of Utility Functions To answer the first question, we train policies with a diverse set of utility functions on Diverse Goal, Deep Sea Treasure and Reacher. Here, we focus on showing the effectiveness of the Utility-based RL algorithm (Algorithm 1). We select several different non-linear functions f, each in favor of distributions at one target. We train policy πi with utility function fi with Algorithm 1 for each target, and show the policy s trajectory and return samples. (a) The map of the Diverse Goal environment. (b) Reward distribution within each goal position. Figure 3: Illustrations of the Diverse Goal environment. Diverse Goal is a MORL environment in Figure 3 with multiple goals that the agent can take multiple steps to reach, where each goal has its unique reward function. Upon reaching a particular goal, the agent secures a 2D reward. This reward is sampled from the specific normal distribution associated with that goal, as illustrated in Figure 3b. Conversely, if the agent reaches the boundary of the map, it incurs a negative 2-dimensional reward. The environment requires the agent to navigate the trade-offs between various reward distributions, underscoring the complexity and nuance inherent to the task. In this study, we investigate if the learned policy can reach the goal that yields the highest expected utility under the utility function. The results are illustrated in Figure 4. Under different types of utility functions and return distributions, the utility-based RL algorithm is able to find the optimal policy with the highest expected utility, which shows the effectiveness of the utility-based RL algorithm. Deep Sea Treasure and Reacher are among the most widely used MORL environments. Under the Deep Sea Treasure environment and Reacher environment, we show the return distributions of the policies optimized with utility-based RL under different utility functions. In Figure 5 and Figure 6, we use utility functions generated by DPMORL in Section 4.2.1 and a set of manually constructed utility functions for different types of distributional preferences respectively. From the result in Figure 5 and Figure 6, we can see that our algorithm effectively finds policies that maximize the expected utility for different non-linear utility functions. Meanwhile, the Deep Sea Treasure environment aligns well (a) Utility functions in Diverse Goal environment used for training policies. (b) Return distribution of policy to maximize expected utility for each utility function. (c) The trajectory of policies learned by maximizing the expected utility for each utility function. Figure 4: The case study results of DPMORL under Diverse Goal environment. (a) Utility functions generated by our method in Section 4.2.1 under Deep Sea Treasure environment. (b) Return distribution of policy learned by DPMORL to maximize expected utility for each utility function. Figure 5: The case study results of DPMORL under Deep Sea Treasure environment. (a) Utility functions with different distributional preferences under Reacher environment. (b) Return distribution of policy learned by DPMORL to maximize expected utility for each utility function. Figure 6: The case study results of DPMORL under Reacher environment. with safe RL settings, where the two returns are time penalty and task reward respectively. With our method, users can provide different utility functions derived from safety constraints, and optimize the expected utility via DPMORL for obtaining policies that satisfy the constraint. 5.2 Main Experiment In this section, we illustrate the performance of DPMORL on five environments based on MOGymnasium [46] to answer the latter two questions. Environments. We conducted experiments across five environments based on MO-Gymnasium [46] to evaluate the performance of our proposed method, DPMORL. These environments represent a diverse range of tasks, from simple toy problems to more complex continuous control tasks, and cover various aspects of multi-objective reinforcement learning: Deep Sea Treasure: A classic MORL benchmark that requires exploration in a gridworld to find treasures with different values and depths. Fruit Tree: A multi-objective variant of the classic gridworld problem, where an agent has to collect different types of fruits with varying rewards and penalties. Hal Cheetah, Hopper, Mountain Car: Three continuous control tasks that require controlling different agents for task solving and minimizing energy usage. More details about the environments are gathered in Appendix C.1. Baselines. We compare DPMORL with four state-of-the-art baselines in the context of distributional preferences of MORL: Optimistic Linear Support (OLS) [47, 16]; Prediction-Guided Multi-Objective Reinforcement Learning (PGMORL) [8]; Generalized Policy Improvement with Linear Support (GPI-LS) [7]; Generalized Policy Improvement with Prioritized Dyna (GPI-PD) [7]. Training Details. For all methods, including DPMORL and the baselines, each policy was trained for 1 107 steps. We learn a set of N = 20 policies for DPMORL and all of the baselines to ensure a fair comparison. Finally, we use the learned N = 20 policies in each method for evaluations. Figure 7: Illustration of 2D utility functions learned by our methods in Section 4.2.1. Implementation Details. We use PPO algorithms implemented in Stable Baselines 3 [48] in Algorithm 1. We use a normalization technique by linearly mapping the return into scale [0, 1]K without modifying the optimal policies. Detailed implementations are provided in Appendix B. Evaluation Metrics. To thoroughly evaluate the performance of DPMORL and compare it with the baseline methods, we employed four distinct metrics. These comprise two existing MORL metrics, Hyper Volume and Expected Utility, in addition to two novel metrics developed specifically for assessing distributional preference, i.e., Constraint Satisfaction and Variance Objective. The latter two are designed to underscore the optimality of multivariate distributions associated with the learned policies. In terms of Constraint Satisfaction, we randomly generate M = 100 constraints for the policy set produced. The Constraint Satisfaction metric is then computed by considering the highest probability within the policy set that satisfies each individual constraint. The Variance Objective metric, on the other hand, involves generating M = 100 random linear weights. These weights are applied to both the expected returns and the standard deviations of returns in each dimension. This objective encourages attaining greater expected returns while simultaneously reducing variance, thereby catering to dynamic preferences. Further details about the implementation of these evaluation metrics are provided in Appendix C.2 and Appendix C.3. 5.3 Results Generation of Utility Functions. In accordance with the methodology detailed in Section 4.2.1, we employ non-decreasing neural networks in conjunction with diversity-focused objectives to generate a diverse assortment of utility functions. The generated functions are visually represented in Figure 7. The outcomes clearly indicate that optimizing diversity-based objective functions allows for generating a broad range of non-linear utility functions, thereby encompassing an expansive array of preferences with respect to returns. Subsequently, we utilize the first N = 20 utility functions depicted in Figure 7 to train an equivalent number of policies under DPMORL. Standard MORL Metrics. The results under standard MORL metrics are shown in Table 1. Focusing on the Expected Utility, the performance of DPMORL is the highest in the Hopper, Half Cheetah, Mountain Car, and Deep Sea Treasure environments (4/5), indicating that our method outperforms others in terms of expected utility. While for Fruit Tree environment, DPMORL also obtains consistent performance with the best contender, GPI-LS. In terms of the Hyper Volume, DPMORL also performs the best in the Hopper, Half Cheetah, Mountain Car, and Fruit Tree environments (4/5). The results show that DPMORL can yield better performance across both metrics and most environments (Hopper, Half Cheetah, Mountain Car, and Deep Sea Treasure), meanwhile delivering robust and consistent results on the other one (Fruit Tree). Since PGMORL works on continuous action spaces, here we omit the results of PGMORL on environments with discrete action space, Deep Sea Treasure and Fruit Tree. Distributional MORL Metrics. The results under distributional metrics are shown in Table 2. DPMORL outperforms other methods in most environments on both Constraint Satisfaction (5/5) and Variance Objective (3/5), indicating its strong ability to handle the distributional multi-objective reinforcement learning problem. For the rest environments, DPMORL also obtained comparable results compared with the best contender, GPI-LS. The results show that the policies learned by Table 1: Experimental results of each method on each standard MORL metric on all five environments. EU stands for Expected Utility, and HV stands for Hypervolume. Environment Hopper Half Cheetah Mountain Car Deep Sea Treasure Fruit Tree Metric EU HV EU HV EU HV EU HV EU HV GPI-PD 2374.44 5033458.34 412.62 1083227.57 -55.44 7367.59 5.93 9.75 6.98 0.14 GPI-LS 1398.00 1446705.10 98.31 2839051.60 -37.37 8052.58 5.04 9.36 3.84 1.49 OLS 175.07 3298.13 -580.38 1420270.13 -470.00 2.45 4.64 10.28 6.27 7.49 PGMORL 300.19 32621.25 111.30 383681.25 -429.48 94.45 - - - - DPMORL (Ours) 3492.93 12154967.99 1189.68 8593769.62 -29.89 8663.80 6.70 8.68 6.89 16.39 Table 2: Experimental results of each method on Distributional metric on all five environments. Constraints stands for Constraints Satisfaction, and Var stands for Variance Objective. Environment Hopper Half Cheetah Mountain Car Deep Sea Treasure Fruit Tree Metric Constraints Var Constraint Var Constraint Var Constraint Var Constraint Var GPI-PD 0.47 979.74 0.64 83.49 1.00 -31.15 0.85 2.59 0.65 3.42 GPI-LS 0.25 607.47 0.60 51.67 1.00 -22.73 0.85 1.99 0.33 1.35 OLS 0.05 75.48 0.43 -311.27 0.05 -244.21 0.80 1.86 0.50 2.98 PGMORL 0.05 98.47 0.50 45.48 0.11 -249.69 - - - - DPMORL (Ours) 0.76 1645.89 0.82 431.26 1.00 -16.32 0.90 2.54 0.67 3.21 DPMORL have a higher probability of satisfying randomly generated constraints, and can better balance the trade-off between expectations and variances. Return Distributions of Learned Policies by DPMORL. We provide visualizations of the multidimensional return distributions of each policy learned by DPMORL and four baseline methods (PGMORL, OLS, GPI-LS and GPI-PD) on different MORL environments, as shown in Appendix C.4. DPMORL s learned policy set demonstrates higher diversity with respect to return distributions compared to baseline methods. DPMORL on More Than 2-dimensional Return Space. We run experiments with 3 objectives under Hopper and Fruit Tree environment, and show the learned return distribution and standard MORL evaluation metrics in Appendix C.5. The results demonstrate that DPMORL can effectively learn DPO policies under more than 3-dimensional reward functions, and achieves better performance in different MORL metrics compared to baseline methods under 3-dimensional reward functions. Ablation Analysis with Variable Number of Policies. We provide the ablation studies of DPMORL with a varying amount of learned policies N in Appendix C.6. As the number of policies increases, both expected utility and hypervolume increase, which means DPMORL can indeed obtain better performance as the number of policies increases. On the other hand, DPMORL has obtained promising results when there exist only 5 policies, which verify the effectiveness of DPMORL in all environments. 6 Conclusion In this work, we initialized the study of distributional MORL, specifically when preferences over different objectives and their return distributions are uncertain. We introduced the concept of Distributional Pareto-Optimal (DPO) policies with rigorous theoretical analysis, which extend the notion of Pareto-optimality in MORL to include preferences over the entire distribution of returns, not just their expected values. To obtain such desirable policies, we proposed a new algorithm, DPMORL, designed to learn DPO policies with non-linear utility functions. DPMORL allows for expressing a wide range of distributional preferences, providing a flexible and expressive approach to MORL. Experiment results showed that DPMORL consistently outperformed existing MORL methods in terms of optimizing policies for multivariate expected and distributional preferences. Acknowledgments This research was supported by the Institute for AI and Beyond, UTokyo; JST SPRING, Grant Number JPMJSP2108; and Microsoft Research Asia D-CORE program. The authors would like to thank Soichiro Nishimori, Kazuki Ota, and the anonymous reviewers for their insightful comments and suggestions. [1] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. Playing atari with deep reinforcement learning. ar Xiv preprint ar Xiv:1312.5602, 2013. [2] Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT press, 2018. [3] Yunke Wang, Chang Xu, Bo Du, and Honglak Lee. Learning to weight imperfect demonstrations. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pages 10961 10970. PMLR, 2021. [4] Yunke Wang, Chang Xu, and Bo Du. Robust adversarial imitation learning via adaptivelyselected demonstrations. In Zhi-Hua Zhou, editor, Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence, IJCAI 2021, Virtual Event / Montreal, Canada, 19-27 August 2021, pages 3155 3161. ijcai.org, 2021. [5] Diederik M Roijers, Peter Vamplew, Shimon Whiteson, and Richard Dazeley. A survey of multiobjective sequential decision-making. Journal of Artificial Intelligence Research, 48:67 113, 2013. [6] Conor F. Hayes, Roxana Radulescu, Eugenio Bargiacchi, Johan Källström, Matthew Macfarlane, Mathieu Reymond, Timothy Verstraeten, Luisa M. Zintgraf, Richard Dazeley, Fredrik Heintz, Enda Howley, Athirai A. Irissappane, Patrick Mannion, Ann Nowé, Gabriel de Oliveira Ramos, Marcello Restelli, Peter Vamplew, and Diederik M. Roijers. A practical guide to multi-objective reinforcement learning and planning. Auton. Agents Multi Agent Syst., 36(1):26, 2022. [7] Lucas Nunes Alegre, Ana L. C. Bazzan, Diederik M. Roijers, Ann Nowé, and Bruno C. da Silva. Sample-efficient multi-objective learning via generalized policy improvement prioritization. Co RR, abs/2301.07784, 2023. [8] Jie Xu, Yunsheng Tian, Pingchuan Ma, Daniela Rus, Shinjiro Sueda, and Wojciech Matusik. Prediction-guided multi-objective reinforcement learning for continuous robot control. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pages 10607 10616. PMLR, 2020. [9] Anirban Majumdar and Claire J Tomlin. Risk-sensitive sequential decision making: Uncertaintybased performance metrics and risk-sensitive control. Foundations and Trends in Systems and Control, 4(4):265 365, 2017. [10] Shiau Hong Lim and Ilyas Malik. Distributional reinforcement learning for risk-sensitive policies. In Advances in Neural Information Processing Systems, 2022. [11] Aivar Sootla, Alexander I Cowen-Rivers, Taher Jafferjee, Ziyan Wang, David H Mguni, Jun Wang, and Haitham Ammar. Sauté rl: Almost surely safe reinforcement learning using state augmentation. In International Conference on Machine Learning, pages 20423 20443. PMLR, 2022. [12] Mathieu Reymond, Conor F Hayes, Denis Steckelmacher, Diederik M Roijers, and Ann Nowé. Actor-critic multi-objective reinforcement learning for non-linear utility functions. Autonomous Agents and Multi-Agent Systems, 37(2):23, 2023. [13] Haim Levy. Stochastic dominance and expected utility: Survey and analysis. Management science, 38(4):555 593, 1992. [14] Moshe Shaked and J George Shanthikumar. Stochastic orders. Springer Science Business Media, 2007. [15] Patrick Mannion, Jim Duggan, and Enda Howley. Dynamic treatment regimes in mobile health. In 2016 International Joint Conference on Neural Networks (IJCNN), pages 526 533. IEEE, 2016. [16] Lucas Nunes Alegre, Ana L. C. Bazzan, and Bruno C. da Silva. Optimistic linear support and successor features as a basis for optimal policy transfer. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvári, Gang Niu, and Sivan Sabato, editors, International Conference on Machine Learning, ICML 2022, 17-23 July 2022, Baltimore, Maryland, USA, volume 162 of Proceedings of Machine Learning Research, pages 394 413. PMLR, 2022. [17] Axel Abels, Diederik M. Roijers, Tom Lenaerts, Ann Nowé, and Denis Steckelmacher. Dynamic weights in multi-objective deep reinforcement learning. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pages 11 20. PMLR, 2019. [18] Matthieu Geist, Julien Pérolat, Mathieu Laurière, Romuald Elie, Sarah Perrin, Olivier Bachem, Rémi Munos, and Olivier Pietquin. Concave utility reinforcement learning: The mean-field game viewpoint. In Piotr Faliszewski, Viviana Mascardi, Catherine Pelachaud, and Matthew E. Taylor, editors, 21st International Conference on Autonomous Agents and Multiagent Systems, AAMAS 2022, Auckland, New Zealand, May 9-13, 2022, pages 489 497. International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS), 2022. [19] Mridul Agarwal, Vaneet Aggarwal, and Tian Lan. Multi-objective reinforcement learning with non-linear scalarization. In Piotr Faliszewski, Viviana Mascardi, Catherine Pelachaud, and Matthew E. Taylor, editors, 21st International Conference on Autonomous Agents and Multiagent Systems, AAMAS 2022, Auckland, New Zealand, May 9-13, 2022, pages 9 17. International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS), 2022. [20] Conor F Hayes, Mathieu Reymond, Diederik M Roijers, Enda Howley, and Patrick Mannion. Distributional monte carlo tree search for risk-aware and multi-objective reinforcement learning. In Proceedings of the 20th international conference on autonomous agents and multiagent systems, pages 1530 1532, 2021. [21] Conor F Hayes, Diederik M Roijers, Enda Howley, and Patrick Mannion. Multi-objective distributional value iteration. In Adaptive and Learning Agents Workshop (AAMAS 2022), 2022. [22] Conor F Hayes, Timothy Verstraeten, Diederik M Roijers, Enda Howley, and Patrick Mannion. Expected scalarised returns dominance: a new solution concept for multi-objective decision making. Neural Computing and Applications, pages 1 21, 2022. [23] Willem Röpke, Conor F. Hayes, Patrick Mannion, Enda Howley, Ann Nowé, and Diederik M. Roijers. Distributional multi-objective decision making. In Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI 2023, 19th-25th August 2023, Macao, SAR, China, pages 5711 5719. ijcai.org, 2023. [24] Marc G Bellemare, Will Dabney, and R emi Munos. A distributional perspective on reinforcement learning. In Proceedings of the 34th International Conference on Machine Learning Volume 70, pages 449 458. JMLR. org, 2017. [25] Pushi Zhang, Xiaoyu Chen, Li Zhao, Wei Xiong, Tao Qin, and Tie-Yan Liu. Distributional reinforcement learning for multi-dimensional reward functions. In Marc Aurelio Ranzato, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, and Jennifer Wortman Vaughan, editors, Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, Neur IPS 2021, December 6-14, 2021, virtual, pages 1519 1529, 2021. [26] Will Dabney, Mark Rowland, Marc G Bellemare, and Remi Munos. Distributional reinforcement learning with quantile regression. In Thirty-Second AAAI Conference on Artificial Intelligence, 2018. [27] Gabriel Barth-Maron, Matthew W. Hoffman, David Budden, Will Dabney, Dan Horgan, Dhruva TB, Alistair Muldal, Nicolas Heess, and Timothy P. Lillicrap. Distributed distributional deterministic policy gradients. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. Open Review.net, 2018. [28] Vivek S Borkar. Risk-sensitive reinforcement learning. SIAM Journal on Control and Optimization, 40(3):818 834, 2002. [29] Aviv Tamar, Yonathan Glassner, and Shie Mannor. Optimality and approximation in online and risk-averse markov decision processes. The Journal of Machine Learning Research, 16(1):2553 2592, 2015. [30] L A Prashanth and Shalabh Bhatnagar. Risk-sensitive reinforcement learning. ar Xiv preprint ar Xiv:1604.00147, 2016. [31] Javier Garc ia and Fernando Fern andez. Safe exploration in markov decision processes. ar Xiv preprint ar Xiv:1205.1111, 2012. [32] Joshua Achiam, David Held, Aviv Tamar, and Pieter Abbeel. Constrained policy optimization. ar Xiv preprint ar Xiv:1705.10528, 2017. [33] Yinlam Chow, Ofir Nachum, Edgar A Duenez-Guzman, and Mohammad Ghavamzadeh. Lyapunov-based safe policy optimization for continuous control. ar Xiv preprint ar Xiv:1805.08711, 2018. [34] Zi-Xuan Chen, Xin-Qiang Cai, Yuan Jiang, and Zhi-Hua Zhou. Anomaly guided policy learning from imperfect demonstrations. In Piotr Faliszewski, Viviana Mascardi, Catherine Pelachaud, and Matthew E. Taylor, editors, 21st International Conference on Autonomous Agents and Multiagent Systems, AAMAS 2022, Auckland, New Zealand, May 9-13, 2022, pages 244 252. [35] Yunke Wang, Xiyu Wang, Anh-Dung Dinh, Bo Du, and Charles Xu. Learning to schedule in diffusion probabilistic models. In Ambuj K. Singh, Yizhou Sun, Leman Akoglu, Dimitrios Gunopulos, Xifeng Yan, Ravi Kumar, Fatma Ozcan, and Jieping Ye, editors, Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD 2023, Long Beach, CA, USA, August 6-10, 2023, pages 2478 2488. ACM, 2023. [36] Zsolt Gabor and Zoltan Koles. Multi-objective reinforcement learning: A comprehensive overview. Proceedings of the IEEE-INNS-ENNS International Joint Conference on Neural Networks, 3:2030 2035, 1998. [37] Kristof Van Moffaert, Madalina M Drugan, and Ann Nowé. Multi-objective deep reinforcement learning. ar Xiv preprint ar Xiv:1610.02707, 2016. [38] Xin-Qiang Cai, Yao-Xiang Ding, Yuan Jiang, and Zhi-Hua Zhou. Imitation learning from pixel-level demonstrations by hashreward. In Frank Dignum, Alessio Lomuscio, Ulle Endriss, and Ann Nowé, editors, AAMAS 21: 20th International Conference on Autonomous Agents and Multiagent Systems, Virtual Event, United Kingdom, May 3-7, 2021, pages 279 287. ACM, 2021. [39] Xin-Qiang Cai, Yao-Xiang Ding, Zi-Xuan Chen, Yuan Jiang, Masashi Sugiyama, and Zhi-Hua Zhou. Seeing differently, acting similarly: Heterogeneously observable imitation learning. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. Open Review.net, 2023. [40] Yunke Wang, Bo Du, and Chang Xu. Unlabeled imperfect demonstrations in adversarial imitation learning. In Brian Williams, Yiling Chen, and Jennifer Neville, editors, Thirty-Seventh AAAI Conference on Artificial Intelligence, AAAI 2023, Thirty-Fifth Conference on Innovative Applications of Artificial Intelligence, IAAI 2023, Thirteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2023, Washington, DC, USA, February 7-14, 2023, pages 10262 10270. AAAI Press, 2023. [41] Brandon Amos, Lei Xu, and J Zico Kolter. Input convex neural networks. In International Conference on Machine Learning, pages 146 155. PMLR, 2017. [42] Tabish Rashid, Mikayel Samvelyan, Christian Schroeder De Witt, Gregory Farquhar, Jakob Foerster, and Shimon Whiteson. Monotonic value function factorisation for deep multi-agent reinforcement learning. The Journal of Machine Learning Research, 21(1):7234 7284, 2020. [43] Charles Dugas, Yoshua Bengio, François Bélisle, Claude Nadeau, and René Garcia. Incorporating functional knowledge in neural networks. Journal of Machine Learning Research, 10(6), 2009. [44] John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. Proximal policy optimization algorithms. Co RR, abs/1707.06347, 2017. [45] Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Jennifer G. Dy and Andreas Krause, editors, Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018, volume 80 of Proceedings of Machine Learning Research, pages 1856 1865. PMLR, 2018. [46] Lucas N. Alegre, Florian Felten, El-Ghazali Talbi, Grégoire Danoy, Ann Nowé, Ana L. C. Bazzan, and Bruno C. da Silva. MO-Gym: A library of multi-objective reinforcement learning environments. In Proceedings of the 34th Benelux Conference on Artificial Intelligence BNAIC/Benelearn 2022, 2022. [47] Diederik M Roijers. Multi-objective decision-theoretic planning. AI Matters, 2(4):11 12, 2016. [48] Antonin Raffin, Ashley Hill, Adam Gleave, Anssi Kanervisto, Maximilian Ernestus, and Noah Dormann. Stable-baselines3: Reliable reinforcement learning implementations. Journal of Machine Learning Research, 22(268):1 8, 2021. [49] Takashi Kamihigashi and John Stachurski. Partial stochastic dominance via optimal transport. Operations Research Letters, 48(5):584 586, 2020. [50] Cédric Villani et al. Optimal transport: old and new, volume 338. Springer, 2009. [51] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In Yoshua Bengio and Yann Le Cun, editors, 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. Lemma 1. If a policy π is the only optimal policy under some non-decreasing utility function f, then it is a Distributionally Pareto-Optimal policy. Proof. If π is the only optimal policy under non-decreasing utility function f, then for any policy policy π , Ez µ(π)f(z) > Ez µ(π )f(z). (4) From Definition 1 and Definition 2, if any policy π stochastically dominates π, then for any utility function f, Ez µ(π ) Ez µ(π)f(z), which makes a contradition. Therefore, π is not stochastically dominated by any other policies, and π is a Distributional Pareto-Optimal policy. Lemma 2. If a policy π is one of the optimal policy under some strictly increasing utility function f, then it is a Distributionally Pareto-Optimal policy. The utility function f is strictly increasing if and only if x y and x = y f(x) > f(y). Proof. Following the Equation (2) and (3) of [49] which is derived from Kantorovich duality (Theorem 5.10 in [50]), we have µ1 SD µ2 if and only if µ1 = µ2 and max γ Π(µ1,µ2) E(x,y) γ[1x y] = 1 (5) where Π(µ1, µ2) is the set of all couplings of µ1 and µ2, which is the set of distributions γ(x, y) such that Z γ(x, y)dy = µ1(x), Z γ(x, y)dx = µ2(y). Suppose there exists a policy π stochastically dominate π, then there exists γ(x, y) Π(µ(π ), µ(π)), such that E(x,y) γ[1x y] = 1. Since π is one of the optimal policy under strictly increasing utility function f, we have Ex µ(π )f(x) Ey µ(π)f(y) 0 Z f(x)µ(π )(x)dx Z f(y)µ(π)(y)dy 0 ZZ [f(x) f(y)]γ(x, y)dxdy 0 Since E(x,y) γ[1x y] = 1, i.e. RR x y γ(x, y)dxdy = 1, we have ZZ x y [f(x) f(y)]γ(x, y)dxdy = 0 x y [f(x) f(y)]γ(x, y)dxdy 0 There are two cases. Case 1: there exists c 0 such that RR x y+c γ(x, y)dxdy > 0. Since f is an strictly increasing function, we have If x y and x = y, then f(x) > f(y). Then, we have: x y [f(x) f(y)]γ(x, y)dxdy x y+c [f(x) f(y)]γ(x, y)dxdy + ZZ y+c x y [f(x) f(y)]γ(x, y)dxdy x y+c [f(y + c) f(y)]γ(x, y)dxdy Since c 0, f(y + c) f(y) > 0. Since RR x y+c γ(x, y) > 0 and γ(x, y) is continuous almost everywhere, we have RR x y+c[f(y + c) f(y)]γ(x, y)dxdy > 0, which yields a contradiction. Case 2: there does not c 0 such that RR x y+c γ(x, y)dxdy > 0. Then, we have E(x,y) γ(x,y)1[x = y] = 1, which contradicts with the assuption that µ(π ) = µ(π). Therefore, we have proved that there does not exist π that stochastically dominate π, when π has optimal expected utility under some strictly increasing utility function f. Lemma 3. Any Pareto-Optimal policy is a Distributionally Pareto-Optimal policy. Proof. We first show that if distribution µ stochastically dominates µ, then the expectation Ez µ [z] dominates Ez µ[z]. Following Kantorovich duality, there exists a coupling γ of µ and µ that E(x,y) γ[1x y] = 1. Then, Ez µ [z] Ez µ[z] = E(x,y) γ[x y] E(x,y) γ[0] = 0 When E(x,y) γ[x y] = 0, it requires that x = y holds where γ(x, y) > 0, which makes µ = µ that contradicts with the fact that µ SD µ. Therefore, the expectation Ez µ [z] dominates the expectation Ez µ[z]. It follows that if the expectation Ez µ[z] is not dominated by the expectation Ez µ [z], then the distribution µ is not stochastically dominated by µ . If policy π is Pareto-optimal, then its multidimensional expected returns E[Z(π)] are not dominated by E[Z(π )] of any other policies π . Then, the return distribution µ(π) of policy π is not stochastically dominated by the return distribution µ(π ) of any other policies π . Therefore, policy π is a Distributionally Pareto-Optimal policy. To conclude, any Pareto-Optimal policy is a Distributionally Pareto-Optimal policy. Theorem 1. If a policy π has optimal expected utility under some non-decreasing utility function f, where either 1. utility function f is strictly increasing 2. π is the only optimal policy with greatest expected utility under utility function f then π is a Distributionally Pareto-Optimal policy. Also, any Pareto-Optimal policy is a Distributionally Pareto-Optimal policy. Proof. The theorem directly follows from Lemma 1, Lemma 2 and Lemma 3. The following theorem proves that the optimal policy over Value at Risk (Va R) and over Conditional Value at Risk (CVa R) belong to the set of Distributionally Pareto-Optimal policies. Here, we use the random variable µ(π) to represent the negative risk, where smaller value represents larger risks, in order to be aligned with the definition of Pareto-Optimality. Definition 4. Given a distribution µ, the Value at Risk (Va R) is defined as Va Rα(µ) = min{z0|Ez µ1[z z0] α} Definition 5. Given a distribution µ, the Conditional Value at Risk (CVa R) is defined as CVa Rα(µ) = Ez µ[z|z Va Rα(µ)] Theorem 3. When K = 1, there exists policy with optimal Value at Risk (Va R), and policy with optimal Condition Value at Risk (CVa R), which belongs to the set of Distributional Pareto-Optimal Policies. Proof. We let the policy set ΠVa R as the policy set with optimal value at risk, which z0 = Va Rα(µ(π)) for any π ΠVa R. We let Π Va R to be the policy set with non-optimal policy, with Va Rα(µ(π )) < z0 for any π Π Va R. Suppose all Distributionally Pareto-Optimal policy belongs to Π Va R. We define a utility function as f(z) = 1, z z0 0, z < z0 , which is non-decreasing. Then, we have Ez µ(π)f(z) α for any π ΠVa R, while Ez µ(π )f(z) < α for any π Π Va R. Therefore, any policy π ΠVa R is not stochastically dominated by any policy π Π Va R. Therefore, there must exists a Distributional Pareto-Optimal Policy in Π, which results in a contradiction. Therefore, there must exists a policy with optimal value at risk, which is Distributionally Pareto-Optimal. Similarly to the case in Va R, we let the policy set ΠCVa R as the policy set with optimal conditional value at risk, which z0 = CVa Rα(µ(π)) for any π ΠCVa R. We let Π CVa R to be the policy set with non-optimal CVa R, with CVa Rα(µ(π )) < z0 for any π Π CVa R. Then we can prove that any policy Π CVa R does not stocastically dominate the policy in ΠCVa R. Suppose the opposite case where π ΠCVa R is dominated by π Π CVa R. We let f(z), f (z) to be the cumulative distribution function of µ(π) and µ(π ), and f(z) f (z) must be satisfied to make π SD π . Therefore, z0 = CVa R(µ(π)) = Ez U(0,α)f 1(z) Ez U(0,α)f 1(z) = CVa R(µ(π )) which results in a contradiction. Therefore, any policy Π CVa R does not stocastically dominate the policy in ΠCVa R. We thus proved that there exists a Distributionally Pareto-Optimal policy in ΠCVa R, which is an optimal policy under conditional value at risk. B Implementation Details Non-Decreasing Neural Network. We used Non-Decreasing Neural Network to parameterize the utility function. The network architecture consisted of three fully connected layers, with our designed non-decreasing activation activation functions applied to the hidden layers: f(x) = concat(max(x, 0.5), min(x, 0.5), clip(x, 0.5, 0.5)) This design of activation function is an extension to Re LU activation function, which empirically make the learned utility function more diverse, covering more types of decision boundaries. After each gradient update during training, we clip all the trainable weights in the network to be non-negative to ensure that the resulting utility function was non-decreasing. The network s input dimension corresponded to the number of objectives in the environment, and the output dimension was a scalar representing the utility value. Return Normalization. We use a normalization technique by linearly mapping the return into scale [0, 1]K without modifying the optimal policies. Specifically, we keep track of the minimum and maximum multivariate returns that the policies have currently achieved, denote by zmin and zmax. We let zmid = (zmin + zmax)/2, and d = maxi[zmax,i zmin,i]. Before the any multivariate return z is fed into the utility function, we normalize it by d(znorm zmid) + 1 This maps z linearly to znorm, where znorm [0, 1]K, without affecting the scale of each dimension of returns. In Algorithm 1, the return zt is normalized before feeding into the utility function to compute the scalar reward. DPMORL Algorithm. For the implementation of our DPMORL algorithm, we used the Proximal Policy Optimization (PPO) algorithm [44] as the basic RL algorithm. The learning rate was set to 3 10 4, with a discount factor γ of 0.99. For PPO, we used a clipping parameter of 0.2. The batch size was set to 256 for all environments and algorithms, with updates performed every 2,048 steps. The utility functions were trained using the Adam optimizer [51], with a learning rate of 3 10 4. C Experimental Details C.1 Environments The details of the environments used in the main paper are listed below: Deep Sea Treasure: A classic MORL benchmark that requires exploration in a gridworld to find treasures with different values and depths. Fruit Tree: A multi-objective variant of the classic gridworld problem, where an agent has to collect different types of fruits with varying rewards and penalties. Hal Cheetah: A continuous control task where a simulated cheetah must run and jump over obstacles while minimizing energy consumption. Hopper: A continuous control task that involves balancing and hopping a simulated onelegged robot while minimizing energy usage. Mountain Car: A classic reinforcement learning problem, extended to include multiple objectives, such as minimizing the time to reach the goal and reducing energy consumption. In the Appendix, we additionally add one environment: Reacher: A discrete control task where a simulated robotic arm must reach a target while minimizing energy usage. C.2 Standard MORL Metrics The details of the standard MORL metrics used in the main paper are listed below: Expected Utility is similar to R-Metrics used in Multi-Objective Optimization (MOO). This metric is concerned with the expected utility of the policies on the Pareto front for various weights. It provides a means to assess the quality of the policies derived by the algorithm in terms of their expected performance across different utility functions. A higher Expected Utility indicates that the policies offer a better trade-off between the objectives, thereby leading to more desirable outcomes under a variety of user preferences: EU(Π) = Ew W[max π Π fw(z)], (6) where fw denotes the user s utility function linearly parameterized by w, and z denotes the returns of one episode. Hypervolume is a prominent measure used in Multi-Objective Optimization (MOO) and Multi-Objective Reinforcement Learning (MORL). This metric calculates the volume in the objective space dominated by the Pareto front with respect to a reference point. The Hypervolume gives a scalar measure of the quality of a Pareto front, with higher values indicating a better Pareto front. A Pareto front with a larger Hypervolume signifies that the algorithm can yield a wider range of policies, thus dominating more of the objective space, which indicates a better exploration of the policy space: HV(Π, r) = Leb(y RK | π Π : z SD y SD r), (7) where r is the reference point in the objective space and Leb denotes the Lebesgue measure, which in this context refers to the volume in the objective space. The reference points used in the experiment of the main paper are calculated as the minimal mean returns of each method under evaluation. C.3 Distributional MORL Metrics We propose two novel metrics for assessing distributional preference: Constraint Satisfaction and Variance Objective. For Constraint Satisfaction, we define u(p, π) to be the probability of return random variance z Z(π) that satisfy the constraint p, representing how a policy satisfy the given constraint. We randomly generate each constraint defined by n linear constraints represented by {z| i, w T i z ci}, where wi is uniformly sampled from the surface wi 0, 1T wi = 1, and ci is sampled from the distribution of U(minz Z w T i z, maxz Z w T i z), where Z is the set of all return samples of the final policies learned by DPMORL. For Variance Objective, we define u(p, π) to be a weighted sum of expected return Ez µ(π)z and standard deviation of return p Varz µ(π)[z]. We randomly generate wi R2K, where wi 0 and Figure 8: Return distribution of the set of policies learned by DPMORL on each environment. 1T wi = 1. We let wi = [w1 i , w2 i ], where w1 i , w2 i RK, and use w1 i and -w2 i as the weight for expected return and standard derivation of return respectively. For both of the metric, we define u(p, π) to be the satisfaction score of policy π under preference p. We generate M = 100 number of preference p1, p2, , p M. For both of the Constraint Satisfaction and Variance Objective metric, the mean score is defined as i=1 max j {1,2, ,N} u(pi, πj) The mean score in the above equation measures the best performance of satisfying the preference of the user, averaged in M = 100 preferences. Higher mean score represents better ability of meeting different preferences. In Table 2 of the main experiment, all of the numbers are calculated with the above equation. C.4 Return Distributions of Learned Policies by DPMORL and Baseline Methods We demonstrate the return distribution of the learned policies by DPMORL in Figure 8, which demonstrates that DPMORL is able to learn a set of policies with optimal return distributions. We also visualize the return distribution of policies learned by four baseline methods, which is shown in Figure 9. DPMORL s learned policy set has higher diversity with respect to return distributions compared to baseline methods. C.5 DPMORL on More Than 2-Dimensional Return Space We run experiments with 3 objectives under Hopper and Fruit Tree environment, where we use the first three dimension of the reward in the MO-Hopper and Fruit-Tree environment the MO-Gymnasium. We show the 3-dimensional return distribution of the set of policies learned by DPMORL and baseline methods in Figure 10. The results demonstrate that DPMORL can effectively learn DPO policies under more than two-dimensional reward functions. Also, DPMORL learns better 3-dimensional return distributions than GPI-PD, a competitive MORL baseline method. We also show quantitative evaluation metrics under Hopper and Fruit Tree environments with 3 objectives in Table 3. The results demonstrate that DPMORL achieves better performance in different MORL metrics compared to baseline methods under 3-dimensional reward functions. Figure 9: Return distribution of the policy set learned by four baseline methods: PGMORL, OLS, GPI-LS, and GPI-PD. Figure 10: DPMORL can effectively learn DPO policies under more than two-dimensional reward functions. Left: 3-dimensional return distribution of the set of policies learned by DPMORL under Hopper and Fruit Tree environment. Right: 3-dimensional return distribution under the same environment learned by GPI-PD. Table 3: Experimental results of each method on Hopper and Fruit Tree environments with 3dimensional reward function. Environment Hopper Fruit Tree Metric Expected Utility CVa R Constraint Satisfaction Variance Objective Expected Utility CVa R Constraint Satisfaction Variance Objective GPI-LS 1037.77 1286.61 0.15 480.94 4.19 4.55 0.27 1.60 GPI-PD 1511.40 1339.67 0.36 664.02 4.44 6.76 0.20 2.30 OLS 441.62 194.56 0.08 78.77 4.72 6.96 0.35 2.36 DPMORL 2107.26 2840.44 0.67 949.96 5.32 7.25 0.38 2.57 C.6 Ablation Analysis with Variable Number of Policies We also provide the ablation studies of DPMORL with a varying amount of learned policies N, as shown in Table 4. We can see that as the number of policies increases, both expected utility and hypervolume increase, which indicates that DPMORL can indeed obtain better performance as the number of policies increases. On the other hand, DPMORL has obtained promising results when there exist only 5 policies, which verify the effectiveness of DPMORL in all environments. Table 4: Experimental results of DPMORL on each standard MORL metric on five environments with various number of policies. EU stands for Expected Utility, and HV stands for Hypervolume. Environment Hopper Half Cheetah Mountain Car Deep Sea Treasure Fruit Tree Metric EU HV EU HV EU HV EU HV EU HV DPMORL (N = 5) 3274.36 9917663.14 817.61 5676147.04 -29.94 7462.10 5.82 8.93 6.22 18.74 DPMORL (N = 10) 3309.52 10512291.02 1244.48 8497055.37 -29.94 7465.59 5.86 14.88 6.65 21.64 DPMORL (N = 15) 3377.75 11453079.82 1245.63 8497055.37 -29.94 7465.59 5.90 15.02 6.68 23.80 DPMORL (N = 20) 3377.75 11453079.82 1749.27 13417685.27 -29.94 7465.59 6.53 19.68 6.68 23.80