# inferring_lexicographicallyordered_rewards_from_preferences__415479b2.pdf Inferring Lexicographically-Ordered Rewards from Preferences Alihan H uy uk1, William R. Zame2, Mihaela van der Schaar1 2 3 1University of Cambridge 2University of California, Los Angeles 3The Alan Turing Institute ah2075@cam.ac.uk, zame@econ.ucla.edu, mv472@cam.ac.uk Modeling the preferences of agents over a set of alternatives is a principal concern in many areas. The dominant approach has been to find a single reward/utility function with the property that alternatives yielding higher rewards are preferred over alternatives yielding lower rewards. However, in many settings, preferences are based on multiple often competing objectives; a single reward function is not adequate to represent such preferences. This paper proposes a method for inferring multi-objective reward-based representations of an agent s observed preferences. We model the agent s priorities over different objectives as entering lexicographically, so that objectives with lower priorities matter only when the agent is indifferent with respect to objectives with higher priorities. We offer two example applications in healthcare one inspired by cancer treatment, the other inspired by organ transplantation to illustrate how the lexicographically-ordered rewards we learn can provide a better understanding of a decision-maker s preferences and help improve policies when used in reinforcement learning. Introduction Modeling the preferences of agents over a set of alternatives plays an important role in many areas, including economics, marketing (Singh, Hansen, and Gupta 2005), politics (Br auninger, M uller, and Stecker 2016), and healthcare (M uhlbacher and Johnson 2016). One common approach to modeling preferences is to find a utility function over alternatives with the property that alternatives with higher utility are preferred over the ones with lower utility. This approach has been studied extensively in the machine learning (ML) literature as well although the ML literature uses the term reward function rather than utility function. In reinforcement learning particularly, inferring reward function from the observed behavior of an agent (viz. inverse reinforcement learning) has proved an effective method of replicating their policy (Ng and Russell 2000; Abbeel and Ng 2004). Moreover, as Brown et al. (2019); Brown, Goo, and Niekum (2019) have recently shown, if reward functions are inferred from the preferences of an expert instead, then policies can even be improved. In many circumstances, agent behavior is based on multiple often competing objectives. Healthcare in particular is replete with such circumstances. In treating cancer Copyright 2022, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. r1 ! " ?> r1 ! " r1 ! " ?= r1 ! " r2 ! " ?> r2 ! " N pairs of alternatives Observed preferences P Agent makes lexicographic preferences. Figure 1: Setting of LORI. We model the preferences of agents lexicographically through reward functions r1, r2, ..., rk while allowing for errors and indifference to small differences. LORI aims to infer these reward functions from a dataset of observed preferences over pairs of alternatives. (and many other diseases), clinicians aim for treatment that is the most effective but also has the fewest harmful sideeffects. This is especially true in radiation therapy (Wilkens et al. 2007; Jee, Mc Shan, and Fraass 2007), where high doses are needed to be effective against tumors but also cause damage to surrounding tissue. In organ transplantation, clinicians aim to make the best match between the organ and the patient but also to give priority to patients who have been waiting the longest and/or have the most urgent need (Coombes and Trotter 2005; Schaubel et al. 2009). In the allocation of resources in a pandemic, clinicians hope to minimize the spread of infection but also to safeguard the most vulnerable populations (Koyuncu and Erol 2010; Gutjahr and Nolz 2016). In these situations, and many others, the preferences of decision-makers reflect the priorities they assign to various criteria. This paper provides a method for inferring multi-objective reward-based representations of a decision-maker s observed preferences. Such representations provide a better understanding of the decision-maker s preferences and promote reinforcement learning. We model priorities over different objectives as entering lexicographically, so that objectives that have lower priority matter only when the decision-maker is indifferent with respect to objective that have higher priority. While lexicographic ordering is certainly not the only way an agent might prioritize different objectives, it is a prevalent one: there is plenty of evidence showing that humans use The Thirty-Sixth AAAI Conference on Artificial Intelligence (AAAI-22) Method Prototype Given Inferred Ordinal PL Boutilier et al. (2004) P Ordinal lexicographic PL Yaman et al. (2008) P, Unordered {φi RX} Lex.-ordered {φi RX} Cardinal single-dimensional PL Chu and Ghahramani (2005) P r RX Conventional IRL Ziebart et al. (2008) D r RS A Preference-based IRL Brown et al. (2019) P r RS A IRL w/ specifications Vazquez-Chanlatte et al. (2018) D r {0, 1}H IRL w/ constraints Scobee and Sastry (2020) D, r2 RS A r1 {0, 1}S A LORI Ours P Lex.-ordered {ri RX} Table 1: Comparison with related work. LORI is the only method that can infer lexicographically-ordered (lex.-ordered) reward functions solely from observed preferences. Preferences, demonstrations, features/attributes, and reward/utility functions are denoted by P, D, {φi} and r; the space of alternatives, states, actions, and state-action histories are denoted by X, S, A, and H. lexicographic reasoning when making decisions (Kohli and Jedidi 2007; Slovic 1975; Tversky, Sattah, and Slovic 1988; Colman and Stirk 1999; Drolet and Luce 2004; Yee et al. 2007). In modeling priorities lexicographically, we take into account that the decision-maker may be indifferent to small differences. We allow for the possibility that the decisionmaker may be an expert consultant, a treating clinician, the patient, or a population of experts, clinicians or patients. As we shall see, these considerations shape our model. Contributions We introduce a new stochastic preference model based on multiple lexicographically-ordered reward functions. We formulate Lexicographically-Ordered Reward Inference (LORI) as the problem of identifying such models from observed preferences of an agent and provide a Bayesian algorithm to do so. We offer two examples one inspired by cancer treatment, the other inspired by organ transplantation to illustrate how the lexicographicallyordered reward functions we learn can be used to interpret and understand the preferences of a decision-maker, and demonstrate how inferring lexicographically-ordered reward functions from preferences of an expert can help improve policies when used in reinforcement learning. Related Work As a method for learning reward-based representations of lexicographic preferences, LORI is related to preference learning (PL), and as a tool for discovering lexicographically-ordered reward functions for reinforcement learning purposes, it is related to inverse reinforcement learning (IRL). Preference Learning Preference learning is the problem of finding a model that best explains a given set of observed preferences over a set of alternatives. It can be tackled either with an ordinal approach, where a binary preference relation between alternatives is learned directly (e.g. Boutilier et al. 2004), or with a cardinal/numerical approach, where such relations are induced through reward functions (e.g. Chu and Ghahramani 2005; Brochu, Freitas, and Ghosh 2008; Bonilla, Guo, and Sanner 2010; Salvatore et al. 2013). Lexicographic preferences have been primarily considered from an ordinal perspective. Schmitt and Martignon (2006); Dombi, Imreh, and Vincze (2007); Yaman et al. (2008); Kohli, Boughanmi, and Kohli (2019) assume that each alternative has an unordered set of attributes (i.e. features in ML literature) and preferences are made by comparing those attributes lexicographically. They aim to learn in what order the attributes are compared. Kohli and Jedidi (2007); Jedidi and Kohli (2008) consider variations of this approach including binary-satisficing lexicographic preferences, which allows indifference when comparing attributes and relaxes the assumption that attributes are compared one at a time. Booth et al. (2010); Bra uning and H ullermeyer (2012); Liu and Truszczynski (2015); Fernandes, Cardoso, and Palacios (2016); Bra uning et al. (2017); Fargier, Gimenez, and Mengin (2018) generalize this framework and learn lexicographic preference trees instead, where the priority order of attributes is not assumed to be static but allowed to change dynamically depending on the outcome of pairwise comparisons in higherpriority attributes. All of these methods are purely ordinal rather than cardinal. We take a different approach to modeling lexicographic preferences. In our model, each alternative has an associated multi-dimensional reward and a preference relation over alternatives is induced by the standard lexicographic preference relation between their associated rewards. The goal of LORI is to infer the reward functions that determine the reward vector of each alternative. Remember that existing methods for inferring lexicographic relations assume the set of relevant attributes to be specified beforehand (except for their priority order, which need to be inferrred). Inferring the latent reward functions in our model can be conceptualized as learning those attributes from scratch. Inverse Reinforcement Learning Given the demonstrated behavior of an agent, IRL aims to find a reward function that makes the demonstrated behavior appear optimal (Abbeel and Ng 2004; Ramachandran and Amir 2007; Ziebart et al. 2008; Boularias, Kober, and Peters 2011; Levine, Popovic, and Koltun 2011; Finn, Levine, and Abbeel 2016). When the demonstrated behavior is in fact optimal, the learned reward function can guide (forward) reinforcement learning to reproduce optimal policies. However, agents do not always behave optimally according to the judgement of others or even according to their own judgement. In that case, conventional IRL (followed by reinforcement learning) can, at best, lead to policies that mimic the demonstrated suboptimal behavior. Preference-based IRL is an alternative approach to conventional IRL, where a reward function is inferred from the preferences of an expert over various demonstrations instead (Sugiyama, Meguro, and Minami 2012; Wirth, F urnkranz, and Neumann 2016; Christiano et al. 2017; Ibarz et al. 2018). Recently, Brown et al. (2019) showed that this alternative preference-based approach can identify the intended reward function of the expert and lead to optimal policies even when the demonstrations provided are suboptimal. Brown, Goo, and Niekum (2019) showed that these expert preferences can be generated synthetically in scenarios where it is possible to interact with the environment. Both conventional and preference-based IRL methods focus almost exclusively on inferring a single reward function to represent preferences. However, as we have discussed in the introduction, many important tasks are not readily evaluated in terms of a single reward function. Task representations that go beyond single reward functions has been considered in IRL. Most notably, Vazquez-Chanlatte et al. (2018) propose non-Markovian and Boolean specifications to describe more complex tasks, and Chou, Berenson, and Ozay (2018); Scobee and Sastry (2020) infer the constraints that a task might have when a secondary goal is also provided. (We view tasks with constraints as the special case of lexicographically-prioritized objectives in which the reward function describing the secondary goal is only maximized when all constraints are satisfied.) To the best of our knowledge, LORI is the first reward inference method that can learn general lexicographically-ordered reward functions solely from observed preferences (as can be seen in Table 1). Problem Formulation We assume that we are given observations about the preferences of a decision-maker or a set of decision-makers in the form of a pair P = (X, n), where X is a set of alternatives and n : X X Z+ is a function. We interpret n(x , x ) as the number of times that alternative x was observed to be preferred to alternative x . Note that n(x , x ) may be zero because x was never observed to be preferred to x . We allow for the possibility that both n(x , x ) > 0 and n(x , x ) > 0, either because we are observing the preferences of a single decision-maker who is not completely consistent or because we are observing the preferences of a population of decision-makers who do not entirely agree. Write A = {(x , x ) X X : n(x , n ) > 0}; this is the set of pairs (x , x ) for which x is observed to be preferred to x at least once. If (x , x ) A, we often write x x . (Note that we do not assume that is a preference relation in the sense used in economics; in particular, we do not assume that the preference relation is asymmetric or transitive.) We seek to explain these observations in terms of an ordered set of k reward functions {ri RX}k i=1 (numbered so that r1 is prioritized over r2, r2 is prioritized over r3, and so on), in the sense that x x tends to be observed if the reward vector r(x ) = {r1(x ), . . . , rk(x )} lexicographically dominates the reward vector r(x ), in which case we write r(x ) >lex r(x ). Formally, r(x ) >lex r(x ) if and only if there exists i {1, . . . , k} such that ri(x ) > ri(x ) and rj(x ) = rj(x ) for all j < i. Because we allow for the possibility that x x and x x , we incorporate randomness; we also allow for indifference in the presence of small differences. Our objective can be summarized as: Objective Infer reward functions {ri}k i=1 from the observed preferences P of a decision-maker. It should be emphasized that LORI does not assume that there are reward functions and a lexicographic ordering that represent the given preferences exactly; it simply attempts to find reward functions and a lexicographic ordering that represents the given preferences as accurately as possible. Lexicographically-ordered Reward Inference A substantial literature views x x as a random event and models the probability of this event in terms of a noisy comparison between the rewards of x and x . This idea is the foundation of the logistic choice model originated by Mc Fadden (1973); recent ML work includes Brown et al. (2019); Brown, Goo, and Niekum (2019). Formally, given a reward function r RX and a parameter α 0, this literature defines P(x x |r) = eαr(x ) eαr(x ) + eαr(x ) = 1 1 + e α(r(x ) r(x )) . (1) The parameter α represents the extent to which preference is random. (Randomness of the preference may reflect inconsistency on the part of the decision-maker, or reflect the aggregate preferences of some population of decision-makers experts or clinicians or patients.) Note that if α = 0 then preference is uniformly random; at the opposite extreme, as α then preference becomes perfectly deterministic. Given a reward function r, parameter α, observed preferences P = (X, n) and alternatives x , x X, we can ask how likely it is that r, α would generate the same relationship between x , x that we observe in P. By definition, n(x , x ) is the number of times that x is observed to be preferred to x and n(x , x ) is the number of times that x is observed to be preferred to x , so N(x , x ) = n(x , x ) + n(x , x ) is the total number of times that x , x are observed to be compared. The probability that the preference generated by r agrees with the observations in P with respect to x , x is P(Px ,x |r) = N(x ,x ) n(x ,x ) P(x x |r)n(x ,x ) P(x x |r)n(x ,x ) . Hence, the probability that the preference generated by r agrees with P (i.e. the likelihood of r with respect to P) is L(r; P) = P(P|r) = Q (x ,x ) X XP(Px ,x |r) 1/2 . (3) (The exponent 1/2 is needed because each of the terms P(Px ,x |r) appears twice: once indexed by (x , x ) and again indexed by (x , x ).) If the true reward function is known/assumed to belong to a given family R, it can be inferred by finding the maximum likelihood estimate (MLE) ˆr = argmaxr R L(r; P). 2 1 0 1 r1(x ) r1(x ) r2(x ) r2(x ) (a) Baseline preferences (α1 = α2 = 3, ε1 = ε2 = 1) 2 1 0 1 r1(x ) r1(x ) r2(x ) r2(x ) (b) Less consistent preferences (α1 = α2 = 1.5, ε1 = ε2 = 1) 2 1 0 1 r1(x ) r1(x ) r2(x ) r2(x ) (c) Less tolerant preferences (α1 = α2 = 3, ε1 = ε2 = 0.5) Figure 2: Probability distribution P(x x |r1, r2) over preferences induced by reward functions r1, r2. Smaller αi s lead to preferences that are less consistent while smaller εi s lead to ones less tolerant of differences. Lexicographic Setting Adapting this probabilistic viewpoint to our context requires two changes. The first is that we make probabilistic comparisons for multiple reward functions. The second is that we allow for the possibility that the rewards r(x ), r(x ) may not be exactly equal but the difference between them may be regarded as negligible. We therefore begin by defining: P(x i x |ri) = 1 1 + e αi(ri(x ) ri(x ) εi) , P(x i x |ri) = 1 1 + e αi(ri(x ) ri(x ) εi) , (4) P(x i x |ri) = 1 P(x i x |ri) P(x i x |ri) . Respectively, these are the probabilities that x is regarded as significantly better than, significantly worse than or not significantly different from x when measured in terms of the reward function ri. As before, the parameter αi 0 represents the extent to which the reward is random. The parameter εi 0 measures the extent to which the rewards of x and x must differ for the difference to be regarded as significant. Our model is then given by P(x x |{ri}k i=1) P(x i x |ri) j=1 P(x j x |rj) . (5) Figure 2 offers a visual description of how the parameters αi and εi control the properties of this preference distribution. As before, the probability that the preference generated by {ri} agrees with P with respect to x , x is P(Px ,x |{ri}) = N(x ,x ) n(x ,x ) P(x x |{ri})n(x ,x ) P(x x |{ri})n(x ,x ) . Hence, the probability that the preference generated by {ri} agrees with P (i.e. the likelihood of {ri} w.r.t. P) is L({ri}; P) = P(P|{ri}) = Q (x ,x ) X XP(Px ,x |{ri}) 1/2 . (7) And, as before, if the true reward functions are known/assumed to belong to a given family R, they can be inferred by finding the maximum likelihood estimate {ˆri} = argmax{ri R} L({ri}; P). Now, suppose we consider a parameterized family of reward functions such that ri = rθi for θi Θ, where Θ is the space of parameters. Then, the (approximate) MLE of the parameters {θi}k i=1 can simply be found via gradient descent using the negative log-likelihood λ = log L({rθi}; P) as the loss function. This is because λ is differentiable with respect to {θi} as long as rθ is a differentiable parameterization with respect to θ (which we show in Appendix A). In our exposition so far, we have chosen to treat αi s and εi s as hyperparameters for simplicity. However, in practice, they can easily be treated as free variables and learned alongside with parameters {θi}, which we will be doing in all of our experiments. (We show that λ is differentiable with respect to {αi} and {εi} in Appendix A as well.) Finally, it needs to be emphasized that LORI is inherently capable of identifying the priority order of reward functions as well as which reward functions are relevant to modeling preferences. This is because different permutations of a given set of parameters {θi}k i=1 are all present in our search space Θk. In contrast, ordinal models of lexicographic preferences assume relevant attributes to be specified beforehand and only learn the priority order of those attributes. Analysis of LORI Our model satisfies the following desirable properties: (i) Setting k = 1 and ε1 = 0 reproduces the logistic model given in (1), (ii) Taking αi , εi 0, and αiεi yields the no-errors, deterministic case, (iii) The parameters εi have natural interpretations. They are the thresholds above which a reward difference is considered significant: P(x i x |ri) > 1/2 if and only if ri(x ) ri(x ) > εi. It is worth noting that lexicographic representations using multiple reward functions are not only convenient, but (often) necessary. For the simplest example, consider the ordinary lexicographic preference relation on R2: (x1, x2) (x 1, x 2) if x1 > x 1 or x1 = x 1 and x2 > x 2. It is well-known that there does not exist any reward function r : R2 R that represents . That is, there is no reward function r with the property that (x1, x2) (x 1, x 2) if and only if r(x1, x2) > r(x 1, x 2). Similarly, the ordinary lexicographic ordering on Rk+1 cannot be represented by a lexicographic ordering involving only k reward functions. When we consider probabilistic orderings and allow for errors, we can again find simple situations in which preferences that employ k + 1 reward functions cannot be represented in terms of k reward functions. For example, consider X = R2; let the reward functions r1, r2 : R2 R be the coordinate projections and take α1 = α2 = 1 and ε1 = ε2 = 1. The probabilistic preference relation defined by these reward functions and parameters is intransitive. For example, consider the points x, x , x R2 defined by: x = ( 0.6, 2), x = (0, 0), x = (0.6, 2). Direct calculation shows that P(x x ) > 0.5, P(x x ) > 0.5, and P(x x) > 0.5. On the other hand, suppose we are given a reward function r : R2 R, and parameters α 0 and ε 0. If we define to be the probabilistic preference relation defined by r, α, ε, then is necessarily transitive. To see this, note in order that if P(x x ) > 0.5, we must have α(r(x ) r(x ) ε) > 0; in particular, we must have r(x ) r(x ) > ε. Hence, if P(x x ) > 0.5 and P(x x ) > 0.5 then we must have r(x) r(x ) > ε and r(x ) r(x ) > ε, whence r(x) r(x ) > 2ε > ε and P(x x ) > 0.5. Finally, a reasonable question to ask is how to determine the number of reward functions k when using LORI. For lexicographic models, even when the number of potential criteria considered by the agent is large, the number of criteria that actually enter into preference is likely to remain small. Remember that, in a lexicographic model, the i-th most important criterion will only be relevant for a particular decision if the i 1 more important criteria are all deemed equivalent. For most decision, this would be unlikely for even moderately large i. This means using a lexicographic model that employs only a few criteria (i.e. a model with small k) would be enough in most cases; increasing k any further would have little to no benefit in terms of accuracy. We demonstrate this empirically in Appendix B. Illustrative Examples in Healthcare Inferring lexicographically-ordered reward functions from preferences can be used either to improve decision-making behavior or to understand it. Here, we give examples of each in healthcare. However, it is worth noting that applications of LORI are not limited to the healthcare setting; it can be applied in any decision-making environment where multiple objectives are at play. Improving Behavior: Cancer Treatment Consider the problem of treatment planning for cancer patients. For a given patient, write at A = {0, 1} for the binary application of a treatment such as chemotherapy, zt Z = R for tumor volume (as a measure of the treatment s efficacy), and wt W = R for the white blood cell (WBC) count (as a measure of the treatment s toxicity) at time t {1, 2, . . .}. In our experiments, we will simulate the tumor volume zt and the WBC count wt of a patient given a treatment plan at according to the pharmacodynamic models of Iliadis and Barbolosi (2000): zt+1 = zt + 0.003zt log(1000/zt) 0.15ztat + νt wt+1 = wt + 1.2 0.15wt 0.4wtat + ηt , (8) where νt N(0, 0.52) and ηt N(0, 0.52) are noise variables, z1 N(30, 52), and w1 = 8. Notice that both the tumor volume and the WBC count decrease when the treatment is applied and increase otherwise. We assume that clinicians aim to minimize the tumor volume while keeping the average WBC count above a threshold of five throughout the treatment. We define the set of alternatives to be all possible treatment trajectories: X = S τ=1(A Z W)τ. Then, the treatment objective can be represented in terms of lexicographic reward functions: r1(a1:τ, z1:τ, w1:τ) = min 5, 1 τ Pτ t=1wt , r2(a1:τ, z1:τ, w1:τ) = 1 τ Pτ t=1zt . (9) A policy π (A)Z W is a function from the features of a patient at time t to a distribution over actions such that at π( |zt, wt). By definition, the optimal policy is given by π = argmaxlex,π E[r(a1:τ, z1:τ, w1:τ)]. We do not assume that clinicians follow the optimal policy, but rather follow some policy that approximates the optimal policy: πb(a|z1:t, w1:t) = (1 ϵ)π (a|z1:t, w1:t) + ϵ/|A| for some ϵ > 0, which we will call the behavior policy. This means that the policy actually followed by clinicians is improvable. Now, suppose we want to improve the policy followed by clinicians using reinforcement learning but we do not have access to their preferences explicitly in the form of reward functions {r1, r2} so we cannot compute the optimal policy π directly. Instead, we have access to some demonstrations D X generated by clinicians as they follow the behavior policy πb and we query an expert or panel of experts (which might consist of the clinicians themselves) over the demonstrations/alternatives in D to obtain a set of observed preferences P. We use LORI to estimate the reward functions {r1, r2} from the observed preferences P. Setup For each experiment, we take ϵ = 0.5 and generate 1000 trajectories with τ = 20 to form the demonstration set D. Then, we generate preferences by sampling 1000 trajectory pairs from D and evaluating according to the groundtruth reward functions {r1, r2} and the model given in (5), where ε1 = ε2 = 0.1 and α1 = α2 = 10 log(9) (ties are broken uniformly at random). These form the set of expert preferences P. We infer k = 2 reward functions from the expert preferences P using LORI; for comparison, we infer a single reward function using T-REX (Brown et al. 2019), which is the single-dimensional counterpart of LORI (with k = 1), and another single reward function from demonstrations D instead of preferences P using Bayesian IRL (Ramachandran and Amir 2007). (Keep in mind that LORI infers lexicographic reward functions but the two alternatives infer only a Alg. RMSE Accuracy BIRL 0.323 0.005 88.1% 0.95% T-REX 0.214 0.007 89.1% 0.61% LORI 0.103 0.089 92.4% 2.71% Table 2: Comparison of reward functions based on preference prediction performance. LORI performs the best. Behavior BC BIRL T-REX LORI BC 55.7 1.3 BIRL 65.6 1.1 57.3 2.1 T-REX 65.7 1.2 57.5 2.3 49.8 0.9 LORI 71.5 3.8 65.6 5.0 69.4 15 68.9 15 Optimal 75.1 1.3 70.0 1.6 82.5 0.2 82.4 0.5 62.0 16 Table 3: Comparison of policies based on how often the row policies are preferred to the column policies. Note that reported values represent percentages. LORI is the most frequently preferred to the behavior policy and its performance is comparable to that of the optimal policy. single reward function.) LORI also infers parameters ε1 and ε2 together with r1, r2. (We set α1 = α2 = 1; there is no loss of generality because the other variables simply scale.) Then, using the estimated reward functions, we train optimal policies using reinforcement learning. In the case of LORI, we use the algorithm proposed in G abor, Kalm ar, and Szapesv ari (1998), which is capable of optimizing thresholded lexicographic objectives. We also learn a policy directly from the demonstration set D by performing behavioral cloning (BC), where actions are simply regressed on patient features. The resulting policy aims to mimic the behavior policy πb as closely as possible. Details regarding the implementation of algorithms can be found in Appendix C. We repeat each experiment five times to obtain error bars. Results To evaluate the quality of reward functions learned by LORI, T-REX, and BIRL, we compare the prediction accuracy of the preference models they induce on a test set of 1000 additional preferences; the results are shown in Table 2. We see that LORI performs better than either T-REX and BIRL. This is because LORI is the only method to capture the multi-objective nature of the clinicians goal (and the expert s preferences). BIRL performs the worst since the demonstrations in D are suboptimal. Table 3 compares the performance of various policies: the behavior policy, the policy that is learned via BC, the policies that are trained on the basis of reward functions learned by BIRL, T-REX and LORI, and the true optimal policy. Each entry in Table 3 shows the frequency with which the row policy is preferred to the column policy. Letting x π be the distribution of trajectories x generated by following policy π, the frequency that policy π is preferred to the policy π is defined to be P(π π ) = Ex π ,x π P(x x |r1, r2). (This is estimated by sampling 1000 trajectories from both policies.) We use the frequency with which one policy is preferred to another as the measure for the improvement pro- vided by the first policy over the second. (Note that because there is no single ground-truth reward function, it is not feasible to compare the values of the two policies, which would have been the usual measure used in reinforcement learning.) As can be seen, LORI improves on other methods, improves on the behavioral policy more often than other methods, and performs comparably to the true optimal policy. Appendix B includes additional experiments where the ground-truth preferences are generated by a single reward function (rather than r1, r2 defined in (9)). Since our preference model is strictly a generalization of single-dimensional models, LORI (with k = 2) still performs the best but it does not improve over T-REX as much. Understanding Behavior: Organ Transplantation Consider the organ allocation problem for transplantation. Write P for the space of all patients (or patient characteristics) and O for the set of organs. At a given time t T, there is a set of patients Pt P who are waiting for an organ transplant. When an organ ot O becomes available at time t, an allocation policy matches that organ with a particular patient pt Pt. In effect, the allocation policy expresses a preference relation: letting X = P O be the set of alternatives, the pairing (pt, ot) X is preferred over alternative pairings {(p , ot)}p Pt\{pt} that were also possible at time t. From these observations, we can infer a lexicographic reward representation {ri}k i=1 that explains the decisions made by the allocation policy. (Notice that if we view elements of P as patient characteristics, then the observed preferences need not be consistent over time.) Setup We investigate the liver allocation preferences in the US in terms of the benefit a transplantation could provide and the need of patients for a transplant. These two objectives are at odds with each other since the patient in greatest need might not be the patient who would benefit the most from an available organ. As a measure of benefit, we consider the estimated additional number of days a patient would survive if they were to receive the available organ, and as a measure of need, we consider the estimated number of days they would survive without a transplant. We estimate both benefit and need via Cox models following the same methodology as Transplant Benefit, which is used in the current allocation policy of the UK (Neuberger et al. 2008). Our analysis is based on the Organ Procurement and Transplantation Network (OPTN) data for liver transplantations as of December 4, 2020. From the OPTN data, we only consider patients who were in the waitlist during 2019 and for whom benefit and need can be estimated from the recorded features. This leaves us with 4070 patients, 3942 organ arrivals, and 2,450,718 observations about the allocation policy s preferences over patient-organ pairings. We use LORI (with k = 2) and T-REX (Brown et al. 2019), which is the single-dimensional counterpart of LORI (k = 1, ε1 = 0), to infer reward functions that are linear with respect to benefit and need. (The restriction to linear reward functions means that preferences depend only on the benefit and need differences between the two pairings.) In the case of LORI, we also infer ε1 and ε2 to determine what amount of benefit 0 30 60 90 Need difference (days) Benefit improvement (days) required to compansate the (negative) need difference Figure 3: Trade-off between benefit and need according to T-REX and LORI. For both T-REX and LORI, there is a tradeoff of benefit against need. For T-REX, this trade-off occurs at a constant rate. For LORI, the trade-off occurs at an almost constant rate when the need difference is below 60 days, but above 60 days the required trade-off of benefit against need increases sharply, and above 65 days it is almost impossible for benefit to compensate for the need difference. or need is considered to be significant by the allocation policy. For T-REX, we set α = 1, and for LORI, we set α1 = α2 = 1 (again without any loss of generality). Results The single-dimensional reward function learned by T-REX is r = 0.0086 Benefit + 0.0137 Need (10) while the lexicographically-ordered reward functions learned by LORI are r1 = 0.0001 Benefit + 0.0139 Need r2 = 0.0562 Benefit + 0.0002 Need (11) with ε1 = 0.8944 and ε2 = 1.8830. As Figure 3 shows, LORI reveals that need is prioritized over benefit. This can be seen by the weights in the primary reward function r1 and secondary reward function r2, but more specifically in the finding that a need difference greater than 65 days cannot be outweighed by any benefit difference; guaranteeing that the patient with greater need will receive the organ. By contrast, there is no such finding for T-REX: any need difference can be outweighed by a sufficiently large benefit difference. This finding is consistent with current allocation practices in the US. When an organ becomes available for transplantation, it is offered to a patient based on their MELD score (Wiesner et al. 2003), which is strictly a measure of how sick the patient is, and so considers only the patient s need and not the benefit they will obtain. However after an offer is made, clinicians might still choose to reject the offer in order to wait for a future offer that would be more beneficial for their patient. Since offers are first made on the basis of need and then accepted/rejected on the basis of benefit, it is reasonable for need to have priority over benefit in the representation learned by LORI. Figure 4 depicts visually the preferences induced by the two alternative representations. As pointed out above, in -180 -90 0 90 180 Need difference (days) Benefit difference (days) (a) Preferences learned by T-REX -180 -90 0 49 90 180 Need difference (days) Benefit difference (days) (b) Preferences learned by LORI Figure 4: Preferences with respect to benefit and need differences between a candidate pairing and an alternative pairing. Red (solid) means the candidate pairing while blue (striped) means the alternative pairing is likely to be preferred with probability at least 0.5. For LORI, white denotes a region of indifference, where the probability of preferring either pairing is less than 0.5. the preferences learned by LORI, a need difference greater than 65 days cannot be outweighed by any benefit difference, while in the preferences learned by T-REX, any need difference can be outweighed by a sufficiently large benefit difference (and the rate at which benefit trades-off for need is constant). Moreover, LORI identifies a region (indicated by white in Figure 4b), where differences in benefit and need become insignificantly small and preferences are made mostly at random (at least with respect to benefit and need). Conclusion We proposed LORI, a method for inferring lexicographicallyordered reward functions from observed preferences of an agent. Through examples from healthcare, we showed that learning such reward functions can be useful in either improving or understanding behavior. While we have modeled priorities over different objectives lexicographically, which happens to be the case in many decision-making scenarios including our healthcare examples, there are numerous other ways an agent might reason about multiple objectives. Future work can focus on inferring preference representations based on alternative notions of multi-objective decision-making. Acknowledgements This work was supported by the US Office of Naval Research (ONR) and the National Science Foundation (NSF, grant number 1722516). Part of our experimental results are based on the liver transplant data from OPTN, which was supported in part by Health Resources and Services Administration contract HHSH250-2019-00001C. References Abbeel, P.; and Ng, A. Y. 2004. Apprenticeship learning via inverse reinforcement learning. In Proc. 23rd AAAI Conf. Artif. Intell. Bonilla, E. V.; Guo, S.; and Sanner, S. 2010. Gaussian process preference elicitation. In Proc. 24th Conf. Neural Inf. Process. Syst. Booth, R.; Chevaleyre, Y.; Lang, J.; Mengin, J.; and Sombattheera, C. 2010. Learning conditionally lexicographic preference relations. In Proc. 19th Eur. Conf. Artif. Intell. Boularias, A.; Kober, J.; and Peters, J. 2011. Relative entropy inverse reinforcement learning. In Proc. 14th Int. Conf. Artif. Intell. Statist. Boutilier, C.; Brafman, R. I.; Domshlak, C.; and Hoss, H. H. 2004. CP-nets: a tool for representing and reasoning with conditional ceteris paribus preference statements. J. Artif. Intell. Res., 21: 135 191. Bra uning, M.; and H ullermeyer, E. 2012. Learning conditional lexicographic preference trees. In Proc. ECAI-12 Workshop Preference Learn.: Problems Appl. AI. Bra uning, M.; H ullermeyer, E.; Keller, T.; and Glaum, M. 2017. Lexicographic preferences for predictive modeling of human decision making: a new machine learning method with an application in accounting. Eur. J. Oper. Res., 258(1): 295 306. Br auninger, T.; M uller, J.; and Stecker, C. 2016. Modeling preferences using roll call votes in parliamentary systems. Political Anal., 24: 189 210. Brochu, E.; Freitas, N.; and Ghosh, A. 2008. Active preference learning with discrete choice data. In Proc. 22nd Conf. Neural Inf. Process. Syst. Brown, D.; Goo, W.; and Niekum, S. 2019. Better-thandemonstrator imitation learning via automatically-ranked demonstrations. In Proc. 3rd Conf. Robot Learn. Brown, D.; Goo, W.; Prabhat, N.; and Niekum, S. 2019. Extrapolating beyond suboptimal demonstrations via inverse reinforcement learning from observations. In Proc. 36th Int. Conf. Mach. Learn. Chou, G.; Berenson, D.; and Ozay, N. 2018. Learning constraints from demonstrations. In Proc. 13th Int. Workshop Algorithmic Found. Robot. Christiano, P. F.; Leike, J.; Brown, T.; Martic, M.; Legg, S.; and Amodei, D. 2017. Deep reinforcement learning from human preferences. In Proc. 31st Conf. Neural Inf. Process. Syst. Chu, W.; and Ghahramani, Z. 2005. Preference learning with Gaussian processes. In Proc. 22nd Int. Conf. Mach. Learn. Colman, A. M.; and Stirk, J. A. 1999. Singleton bias and lexicographic preferences among equally valued alternatives. J. Econom. Behav. Organ., 40(4): 337 351. Coombes, J. M.; and Trotter, J. F. 2005. Development of the allocation system for deceased donor liver transplantation. Clin. Med. Res., 3(2): 87 92. Dombi, J.; Imreh, C.; and Vincze, N. 2007. Learning lexicographic orders. Eur. J. Oper. Res., 183: 748 756. Drolet, A.; and Luce, M. F. 2004. Cognitive load and tradeoff avoidance. J. Consumer Res., 31: 63 77. Fargier, H.; Gimenez, P.-F.; and Mengin, J. 2018. Learning lexicographic preference trees from positive examples. In Proc. 32nd AAAI Conf. Artif. Intell. Fernandes, K.; Cardoso, J. S.; and Palacios, H. 2016. Learning and ensembling lexicographic preference trees with multiple Kernels. In Int. Joint Conf. Neural Netw. Finn, C.; Levine, S.; and Abbeel, P. 2016. Guided cost learning: deep inverse optimal control via policy optimization. In Proc. 33rd Int. Conf. Mach. Learning. G abor, Z.; Kalm ar, Z.; and Szapesv ari, C. 1998. Multi-criteria reinfrocement learning. In Proc. 15th Int. Conf. Mach. Learn. Gutjahr, W. J.; and Nolz, P. C. 2016. Multicriteria optimization in humanitarian aid. Eur. J. Oper. Res., 252(2): 351 366. Ibarz, B.; Leike, J.; Pohlen, T.; Irving, G.; Legg, S.; and Amodei, D. 2018. Reward learning from human preferences and demonstrations in Atari. In Proc. 32nd Conf. Neural Inf. Process. Syst. Iliadis, A.; and Barbolosi, D. 2000. Optimizing drug regimens in cancer chemotherapy by an efficacy-toxicity mathematical model. Comput. Biomed. Res., 33(3): 211 226. Jedidi, K.; and Kohli, R. 2008. Inferring latent class lexicographic rules from choice data. J. Math. Psychol., 52: 241 240. Jee, K.-W.; Mc Shan, D. L.; and Fraass, B. A. 2007. Lexicographic ordering: intuitive multicriteria optimization for IMRT. Phys. Medicine Biol., 52(7): 1845 1861. Kohli, R.; Boughanmi, K.; and Kohli, V. 2019. Randomized algorithms for lexicographic inference. Oper. Res., 67(2): 357 375. Kohli, R.; and Jedidi, K. 2007. Representation and inference of lexicographic preference models and their variants. Marketing Sci., 26(3): 380 399. Koyuncu, M.; and Erol, R. 2010. Optimal resource allocation model to mitigate the impact of pandemic influenza: a case study for Turkey. J. Med. Syst., 34(1): 61 70. Levine, S.; Popovic, Z.; and Koltun, V. 2011. Nonlinear inverse reinforcement learning with Gaussian processes. In Proc. 24th Conf. Neural Inf. Process. Syst. Liu, X.; and Truszczynski, M. 2015. Learning partial lexicographic preference trees over combinatorial domains. In Proc. 29th AAAI Conf. Artif. Intell. Mc Fadden, D. 1973. Conditional logit analysis of qualitative choice behavior. In Zarembka, P., ed., Frontiers in Econometrics. New York: Academic Press. M uhlbacher, A.; and Johnson, F. R. 2016. Choice experiments to quantify preferences for health and healthcare: state of the practice. Appl. Health Econ. Health Policy, 14(3): 253 266. Neuberger, J.; Gimson, A.; Davies, M.; Akyol, M.; O Grady, J.; Burroughs, A.; and Hudson, M. 2008. Selection of patients for liver transplantation and allocation of donated livers in the UK. Gut, 57(2): 252 257. Ng, A. Y.; and Russell, S. J. 2000. Algorithms for inverse reinforcement learning. In Proc. 17th Int. Conf. Mach. Learn. Ramachandran, D.; and Amir, E. 2007. Bayesian inverse reinforcement learning. In Proc. 20th Int. Joint Conf. Artif. Intell. Salvatore, C.; Salvatore, G.; Kadzi nski, M.; and Słowi nski, R. 2013. Robust ordinal regression in preference learning and ranking. Mach. Learn., 93(2): 381 422. Schaubel, D. E.; Guidinger, M. K.; Biggins, S. W.; Kalbfleisch, J. D.; Pomfret, E. A.; Sharma, P.; and Merion, R. M. 2009. Survival benefit-based deceased-donor liver allocation. Amer. J. Transplantation, 9: 970 981. Schmitt, M.; and Martignon, L. 2006. On the complexity of learning lexicographic strategies. J. Mach. Learn. Res., 7: 55 83. Scobee, D. R. R.; and Sastry, S. S. 2020. Maximum likelihood constraint inference for inverse reinforcement learning. In Proc. 8th Int. Conf. Learn. Representations. Singh, V. P.; Hansen, K. T.; and Gupta, S. 2005. Modeling preferences for common attributes in multicategory brand choice. J. Marketing Res., 42: 195 209. Slovic, P. 1975. Choice between equally values alternatives. J. Experiment. Psych.: Human Perception Performance, 1: 280 287. Sugiyama, H.; Meguro, T.; and Minami, Y. 2012. Preferencelearning based inverse reinforcement learning for dialog control. In Proc. 13th Annu. Conf. Int. Speech Commun. Assoc. Tversky, A.; Sattah, S.; and Slovic, P. 1988. Contingent weighting in judgement and choice. Psych. Rev., 95: 317 384. Vazquez-Chanlatte, M.; Jha, S.; Tiwari, A.; Ho, M. K.; and Seshia, S. A. 2018. Learning task specifications from demonstrations. In Proc. 32nd Conf. Neural Inf. Process. Syst. Wiesner, R.; Edwards, E.; Freeman, R.; Harper, A.; Kim, R.; Kamath, P.; Kremers, W.; Lake, J.; Howard, T.; Merion, R. M.; Wolfe, R. A.; Krom, R.; and The United Network for Organ Sharing Liver Disease Severity Score Committe. 2003. Model for end-stage liver disease (MELD) and allocation of donor livers. Gastroenterology, 124(1): 91 96. Wilkens, J. J.; Alaly, J. R.; Zakarian, K.; Thorstad, W. L.; and Deasy, J. O. 2007. IMRT treatment planning based on prioritizing prescription goals. Phys. Medicine and Biol., 52(6): 1675 1692. Wirth, C.; F urnkranz, J.; and Neumann, G. 2016. Model-free preference-based reinforcement learning. In Proc. 30th AAAI Conf. Artif. Intell. Yaman, F.; Walsh, T. J.; Littman, M. L.; and Desjardins, M. 2008. Democratic approximation of lexicographic preference models. In Proc. 25th Int. Conf. Mach. Learn. Yee, M.; Dahan, E.; Hauser, J. R.; and Orlin, J. 2007. Greedoid-based noncompansatory consideration-then-choice inference. Marketing Sci., 26(4): 532 549. Ziebart, B. D.; Maas, A.; Bagnell, J. A.; and Day, A. K. 2008. Maximum entropy inverse reinforcement learning. In Proc. 23rd AAAI Conf. Artif. Intell.