# perseval_assessing_personalization_in_text_summarizers__536aa330.pdf Per SEval: Assessing Personalization in Text Summarizers Sourish Dasgupta1,*, Ankush Chander1, Parth Borad1, Isha Motiyani1, Tanmoy Chakraborty2,* 1Dhirubhai Ambani Institute of Information & Communication Technology, India 2Indian Institute of Technology Delhi, India Corresponding authors: sourish_dasgupta@daiict.ac.in, tanchak@iitd.ac.in Personalized summarization models cater to individuals subjective understanding of saliency, as represented by their reading history and current topics of attention. Existing personalized text summarizers are primarily evaluated based on accuracy measures such as BLEU, ROUGE, and METEOR. However, a recent study argued that accuracy measures are inadequate for evaluating the degree of personalization of these models and proposed EGISES, the first metric to evaluate personalized text summaries. It was suggested that accuracy is a separate aspect and should be evaluated standalone. In this paper, we challenge the necessity of an accuracy leaderboard, suggesting that relying on accuracy-based aggregated results might lead to misleading conclusions. To support this, we delve deeper into EGISES, demonstrating both theoretically and empirically that it measures the degree of responsiveness, a necessary but not sufficient condition for degree-of-personalization. We subsequently propose Per SEval, a novel measure that satisfies the required sufficiency condition. Based on the benchmarking of ten SOTA summarization models on the PENS dataset, we empirically establish that (i) Per SEval is reliable w.r.t human-judgment correlation (Pearson s r = 0.73; Spearman s ρ = 0.62; Kendall s τ = 0.42), (ii) Per SEval has high rank-stability, (iii) Per SEval as a rankmeasure is not entailed by EGISES-based ranking, and (iv) Per SEval can be a standalone rank-measure without the need of any aggregated ranking. 1 Introduction With the incessant rise of information deluge, it has become even more imperative to develop efficient and accurate models to summarize the salient information in long documents for faster consumption, eliminating the irrelevant and supporting faster decision-making Ter Hoeve et al. (2022). However, the notion of saliency can be highly subjective in many use cases, particularly for documents containing multiple aspects and topics. This calls for summarizers that must be personalized to the users preferences as depicted by their reading behavior and current topic(s) of attention (Ghodratnama et al., 2021; Ao et al., 2021). This calls for robust and reliable measures for evaluating the degree-of-personalization in them. Dearth of Personalization Evaluation. Major studies on evaluating text summaries focus on accuracy measurement and include the proposal of a multitude of measures such as the ROUGE variants (e.g., ROUGE-n/L/SU4 etc.) (Lin, 2004), BLEU (Papineni et al., 2002), METEOR (Banerjee & Lavie, 2005), BERTScore (Zhang et al., 2020), PYRAMID (Gao et al., 2019) and the more recently proposed ones such as SUPERT (Gao et al., 2020), WIDAR (Jain et al., 2022), and Info LM (Colombo et al., 2022b). Other studies acknowledged the need for more qualitative aspects such as consistency, coherence, and fluency (Yuan et al., 2021; Deng et al., 2021; Tam et al., 2023; Jain et al., 2023; Zhong et al., 2022). Recently, Vansh et al. (2023) established, both theoretically and empirically, that personalization is a different aspect than accuracy. The authors proposed a measure for degree-of-personalization for the first time and called it EGISES . The EGISES Paradox. To put it succinctly, EGISES measures the average ratio of the (normalized) deviation between model-generated summaries and their corresponding user-expected summaries, capturing Figure 1: EGISES Personalization-Accuracy Paradox. The absurd case of high personalization (thereby high user-experience), yet low accuracy. the strong notion of a model s degree of insensitivity to users subjective expectations. However, we argue that this degree of insensitivity does not truly measure personalization but rather a related and necessary aspect, which can be understood as the responsiveness of a model. A high degree of personalization must imply a very high-quality user experience (UX). However, one cannot expect high UX while having low accuracy. At the same time, we demonstrate that a model can have high accuracy but a poor EGISES score. We resolve this apparent paradox by establishing both theoretically and empirically that EGISES, in reality, accounts for only the degree of responsiveness of models. In other words, we mathematically prove that a model can have a very good EGISES score but may still fall short of being personalized simply because of low accuracy performance. P-Accuracy is not a solution. As an extension to standard accuracy measures, Vansh et al. (2023) proposed a measure called P-Accuracy (P-Acc) that penalizes accuracy by a factor that is a function of the EGISES score. Although P-Acc might apparently seem to address the EGISES paradox, in this paper, we prove that P-Acc fails to address the issue for two reasons. First, it has not been designed to evaluate the degree-of-personalization of a model but rather to get a more realistic, human-evaluator adjustable accuracy score. This makes P-Acc unable to distinguish the two models, where one exhibits low accuracy but high EGISES while the other exhibits high accuracy but low EGISES. Secondly, we prove that P-Acc is unstable for models with relatively low accuracy and may lead to anomalous results if we measure personalization using it. This motivates us to propose a novel consolidated personalization measurement framework for text summarizers, called Per SEval (Personalized Summarizer Evaluator) that builds on the design principles of EGISES and bridges the UX-gap". Per SEval Design Principle. The underlying design principle of Per SEval entails that higher accuracy performance should not obfuscate the original EGISES score of a model. Otherwise, a model can be considered highly personalized simply because of its high accuracy, which is misleading, as was proven in (Vansh et al., 2023). However, the converse should not be true; hence, a lower accuracy score should penalize the original EGISES score. Based on these design objectives, we propose a penalty factor, called EDP (Effective DEGRESS Penalty) that can be injected into EGISES to form Per SEval to measure the true degree of personalization. EDP incorporates the inconsistency of accuracy of model w.r.t its best accuracy performance and how much that is off from the maximum achievable accuracy (which is 0 for a normalized metric). In the best case, the EDP factor comes to 1 (i.e., the penalty is 0), while in the worst case, it tends to 0 (i.e., the penalty is 1). Observations and Insights. We first empirically establish that the EGISES-based rank does not simply entail the Per SEval-based rank. In other words, models rank differently when EDP is applied. Therefore, we can conclude that the EGISES-paradox not just theoretically exists but has real evidence. We then show that Per SEval provides a much more reliable ranking of models with significantly higher human-judgment correlation in terms of Pearson s r (0.73), Spearman s ρ (0.62), and Kendall s τ (0.42). For fair comparisons, we consider the same top ten state-of-the-art news summarization models and the same PENS test dataset (Ao et al., 2021) that Vansh et al. (2023) considered to evaluate EGISES. We also take a step beyond (Vansh et al., 2023) and demonstrate that the accuracy leaderboard is not only insufficient but is at best redundant and at worst can be misleading for evaluating personalized summarizers. This is established by showing that the Borda-Kendall consensus-based aggregated ranking (Cook & Seiford, 1982) of the models does not have a better human-judgement correlation when compared to the correlation of Per SEval alone.1 2 Background 2.1 EGISES is Not Enough: The Personalization-Accuracy Paradox As proposed in (Vansh et al., 2023), the degree-of-personalization is a quantitative measure of how much a summarization model fine-tuned for personalization is adaptive to a user s (i.e., reader s) subjective expectation. This also implies that it measures how accurately a model can capture the user s "evolving" profile reflected through reading history (i.e., a temporal span of the reading and skipping actions of a user on a sequence of documents that is interleaved by the actions of generating and reading summaries). This is because the subjective expectation is a function of the reading history. A low degree of personalization, by definition, implies poor user experience. If a model does not efficiently capture the user s profile, it may lead to summaries that contain (a) additional irrelevant information, (b) are not coherent with the reading history, and (c) do not cover all the topics of interest within the reading history. In this situation, poor UX would mean that the user would have to spend more time getting to the information he/she is interested in or suffer from information overload and fatigue. To illustrate this, we borrow the example given by Vansh et al. (2023) where we have two readers, Alice and Bob, having two very different reading histories over the same time span. Alice has been primarily reading news articles mostly on the broad topic of "civilian distress" in the Hamas-Israeli conflict, while Bob is following up on articles related to "war-front battles". In a situation where both Bob and Alice come up with an article covering both topics but rather wish to read the summary of the article instead, their respective expected summaries would be quite different. A non-personalized summarization model can be considered 100% accurate w.r.t recall and F-score-based measures (e.g., ROUGE-variants and METEOR, respectively) if it generates a summary that is a union of the expected summaries of Alice and Bob. On the other hand, a model can be considered 100% accurate w.r.t precision-based measures (eg., BLEU) if it covers the intersection between the expected summaries of Alice and Bob, which will not be much since they are quite different profiles. However, while the first situation would imply that Alice and Bob experience significantly redundant content in the generated summary that would drop their overall UX, the second case would entail that both Alice and Bob will have poor UX in terms of significant lack of content coverage. Hence, the utility of any summarization model as experienced by any user is directly associated with the subjective UX and cannot be traded in for accuracy, which is a rather clinical scoring mechanism that does not truly reflect the UX of a user in terms of redundancy and coverage (among other subjective aspects such as correctness and fluency). This phenomenon was established theoretically and empirically by Vansh et al. (2023) for the first time, thereby emphasizing the fact that accuracy measures can mislead an evaluator to select a model that has poor UX. To address this, they proposed a novel measure, called EGISES, for personalization evaluation in summarizers. However, in this paper we establish both theoretically and empirically that if EGISES is used for personalization evaluation (i.e., a measure to understand a model s capacity to engage readers in terms of UX), then we can come to a rather paradoxical possibility where a model can have a high degree of personalization (i.e., acceptable EGISES score) but low accuracy, and yet, by definition, that would entail high UX (see Figure 1). In other words, although high accuracy can lead to poor UX, the inverse (i.e., low accuracy leading to high UX) is absurd. We term this as the personalization-accuracy paradox and attribute it to the incorrect interpretation or usage of EGISES. In this paper, we propose Per SEval as a corrective measure of EGISES that resolves this paradox2. In the following section, we show that EGISES measures responsiveness, a necessary yet distinct attribute to personalization. 1Code: https://github.com/KDM-LAB/Perseval-TMLR 2Per SEval should not be understood as an alternative "improved" measure, and therefore, is not comparable to EGISES. 2.2 Personalization vs. Responsiveness In this section, we first distinguish responsiveness from personalization. Informally, responsiveness is the capacity of a model to discern the differences in the profiles (i.e., reading histories) of two readers quantitatively and accurately predict the dissimilarity in their corresponding expected summaries that is proportionate to their profile difference. However, there can be scenarios where a model exhibits high responsiveness at the cost of losing accuracy. To illustrate this, we continue with the example from the previous section. Suppose we observe an arbitrary model to generate two different summaries for a given news article, one focusing on "Israeli Prime Minister" and the other on "Jewish protests on war", skipping the article s content on civilian distress and war-front battle information. In that case, we have to conclude that the model apparently discerned the difference between Alice and Bob s profiles, thereby predicting the proportionate dissimilarity in the expected summaries but not the expected summaries themselves. Thus, the model is inaccurate and yet responsive. Therefore, interpreting such responsiveness as personalization leads us to the personalization-accuracy paradox. We prove this formally in Section 3. We establish that EGISES measures how sensitive (or insensitive) a model is to the differences in the readers subjective expectations (i.e., responsiveness) but not personalization. Therefore, EGISES can give a fairly good score to the model in the example. To elucidate this, we first define an Oracle personalized summarization model as follows: Definition 1. Personalized Summarization Oracle. A summarization model Mθ,h (parameterized with θ) is an Oracle if for specific j-th reader profile hj (i.e., reading history) it generates an optimal summary s (di,hj) of the document di (i.e., Mθ,h : (di, hj) 7 s (di,hj)), where s (di,hj) s uij uij; uij is the j-th reader s subjective expected summary of di and is determined by hj. We now recall the notion of insensitivity-to-subjectivity, the foundation of EGISES, as in Vansh et al. (2023): Definition 2. Weak Insensitivity-to-Subjectivity. A summarization model Mθ,h is (weakly) Insensitiveto-Subjectivity w.r.t a given document di and corresponding readers j and k, if (uij, uik), (σ(uij, uik) τ U max) (σ(suij, suik) > τ S max), where σ is an arbitrary distance metric defined on the metric space M, where d, u and s are defined3, τ U max is the maximum limit for ui, uj to be mutually indistinguishable, and τ S max is the maximum limit for sui, suj to be mutually indistinguishable. Definition 3. Strong Insensitivity-to-Subjectivity. A summarization model Mθ,h is (strongly) Insensitive-to-Subjectivity w.r.t a given document di and corresponding readers j and k, if (uij, uik), Mθ,h satisfies: (i) the condition of weak insensitivity, and (ii) (σ(uij, uik) > τ U max) (σ(suij, suik) τ S max). Based on this notion, Vansh et al. (2023) defined (summary-level) "deviation" of a model Mθ,h. We generalize this to our notion of summary-level Degree-of-Responsiveness (DEGRESS), the measure for responsiveness, as follows: Definition 4. Summary-level DEGRESS. Given a document di and j-th reader s expected summary uij, the summary-level responsiveness of a personalized model Mθ,h, (denoted by DEGRESS(suij|(di, uij))), is defined as the " proportional divergence" between model-generated summary suij of di for j-th user from all other user-specific summary versions w.r.t a corresponding divergence of uij from all other user-profiles. DEGRESS(suij|(di, uij)) is formulated as follows: DEGRESS(suij|(di, uij)) = 1 |Udi| min(Xijk, Yijk) + ϵ max(Xijk, Yijk) + ϵ Xijk = exp(w(uij|uik)) l=1 exp(w(uij|uil)) σ(uij, uik); Yijk = exp(w(suij|suik)) l=1 exp(w(suij|suil)) σ(suij, suik) w(uij|uik) = σ(uij, uik) σ(uij, di) ; w(suij|suik) = σ(suij, suik) σ(suij, di) 3σ(ui, ui) = 0; σ(ui, uj) [0, 1]; σ satisfies positivity, reflexive, maximality, symmetry, and the triangle inequality. Figure 2: Existence of EGISES Personalization-Accuracy Paradox: High (and same; (a)) DEGRESS, yet low accuracy (red line; (b)). Here, |D| is the total number of documents in the evaluation dataset, |U| is the total number of users who created gold-reference summaries that reflect their expected summaries (and thereby, their subjective preferences), and |Udi| (= |Sdi|) is the number of users who created gold-references for document di. w is the divergence of the model-generated summary suij (and the corresponding expected summary uij) from document di itself in comparison to all the other versions. It helps to determine how much percentage (therefore, the softmax function) of the divergence (i.e., σ(suij, suik) should be considered for the calculation of DEGRESS. If suij is farther than suik w.r.t di then DEGRESS(suij|(di, uij)) < DEGRESS(suik|(di, uik)), implying that Mθ,h is more responsive to the k-th reader. A lower value of DEGRESS(suij|(di, uij)) indicates that while reader-profiles are different, the generated summary suij is very similar to other reader-specific summaries (or vice versa), and hence, is not responsive at the summary-level. The system-level DEGRESS and EGISES have been formulated as follows: DEGRESS(Mθ,h) = j=1 DEGRESS(suij |(di,uij)) |D| ; EGISES(Mθ,h) = 1 DEGRESS(Mθ,h) (2) EGISES measures the degree of insensitivity-to-subjectivity for relative benchmarking of how much models lack personalization (i.e., a lower score is better within the range [0, 1]) instead of assigning an absolute goodness score. As can be noted, the EGISES formalism does not enforce any penalty on accuracy drop. Here, accuracy would be an inverse function of σ(suij, uij)|di for the same metric distance σ that DEGRESS uses. Hence, EGISES (and DEGRESS) should be interpreted as a measure of responsiveness (i.e., proportionate divergence) and not personalization. 3 EGISES Personalization-Accuracy Paradox: Formal Proof In this section, we mathematically prove the existence of the condition that, for sufficiently high DEGRESS (and thereby EGISES), there exists low accuracy. Theorem 1. The accuracy f 1(σ(su, u)) of a model Mθ,h on the metric space M with distance function σ( , ) can vary keeping DEGRESS(su|(d, u)) constant. Proof. We follow the same triangulation proof technique as in (Vansh et al., 2023). Let d, u and s be triangulated as per Figure 2. Keeping d and u fixed, we can perform an arbitrary rotation operation with d as center (rot( , d, δ); δ: angle of rotation) on suij and suik s.t. rot( , d, δ) is a closure operator in M. Now, (p, q) M, s.t. max p σ(rot(suij, di, δp), uij) > min p σ(rot(suij, uij, δp), uij); And similarly, max q σ(rot(suik, di, δq), uik) > min q σ(rot(suik, uik, δq), uik) If p = q, then DEGRESS(sui |(di, ui )) (and therefore, EGISES) remains unchanged for a given di. However, due to the existence of a total ordering of accuracy f 1(σ(s, u)), for any arbitrary α, the accuracy, therefore, can be varied by changing δ from a minima to a maxima (f 1(σ(s, u)) [0, 1]). The proof establishes the theoretical existence of the personalization-accuracy paradox if we interpret EGISES as a measure of personalization instead of responsiveness. It sets the motivation to design a corrective measure for EGISES that truly measures personalization. As discussed in the introduction, Vansh et al. (2023) proposed a follow-up measure called P-Accuracy (P-Acc), that might seem to solve this problem. However, some major limitations of P-Acc prevent it from being used for evaluating degree-of-personalization. We elucidate that in the following section. 4 Limitations of P-Accuracy (P-Acc) as a Solution to the EGISES Paradox In this section, we prove that P-Acc, as proposed in Vansh et al. (2023), cannot be used as a solution to the EGISES paradox. We begin with a recap of the formulation of P-Acc as follows: P-Acc(Mθ,h) = Score Acc(Mθ,h) Unit P -Acc where: Unit P -Acc = 1 [α (fsig(β EGISES(Mθ,h)) Score Acc(Mθ,h) )]; Score Acc : Acc. score w.r.t a chosen accuracy measure Here, β (0, 1] calibrates the importance to be given to lack of personalization in a model while α [0, 1] is the control that helps the human-evaluator to regulate the final penalty to accuracy due to lack of personalization. According to P-Acc, it apparently seems that there cannot be a situation where if both accuracy and EGISES scores are low (i.e., DEGRESS is high), the final score will be high, thereby preventing the EGISES Paradox. However, three limiting properties of P-Acc render it unusable for personalization evaluation. We prove their existence as follows: Theorem 2 (Limitation 1: Lack of Interpretability of Score). There exists infinite possible model collections Θ = {Θi | Θi = {Mθi ,h}} where (k, l) pair : P-Acc(Mθik,h) = P-Acc(Mθil,h) when (i) Score Acc(Mθik,h) Score Acc(Mθil,h), (ii) Score Acc(Mθik,h) Score Acc(Mθil,h), and (iii) Score Acc(Mθik,h) = Score Acc(Mθil,h). P-Acc(Mθ,h) = Score Acc(Mθ,h) α fsig(β EGISES(Mθ,h)), {By definition} (k, l) pair : P-Acc(Mθik,h) = P-Acc(Mθil,h) = Score Acc(Mθik,h) α fsig(β EGISES(Mθik,h)) = Score Acc(Mθil,h) α fsig(β EGISES(Mθil,h)) = Score Acc(Mθik,h) Score Acc(Mθil,h) = α (fsig(β EGISES(Mθik,h)) fsig(β EGISES(Mθil,h))) Case 1/2: Let us assume that a sufficiently large k exists s.t. Score Acc(Mθik,h) = k Score Acc(Mθil,h) fsig(β EGISES(Mθik,h)) fsig(β EGISES(Mθil,h)) = k 1 α Score Acc(Mθil,h) = fsig(β EGISES(Mθik,h)) fsig(β EGISES(Mθil,h)) k 1 { α [0, 1]} = EGISES(Mθik,h) EGISES(Mθil,h) k 1 = EGISES(Mθik,h) = k + EGISES(Mθil,h) which is a possible condition. Hence, the assumption is valid. Case 3: Let us assume Score Acc(Mθil,h, h) = Score Acc(Mθil,h) {k = 1} EGISES(Mθik,h, h) EGISES(Mθil,h, h) which is a possible condition. Hence, the assumption is valid. The proof entails that P-Acc scores cannot discriminate the three cases. Theorem 3 (Limitation 2: Lack of Interpretability of Lower Bound). There exists a condition when P-Acc(Mθi ,h) < 0 even when Score Acc(Mθi ,h) 0. P-Acc(Mθ,h) = Score Acc(Mθ,h) α fsig(β EGISES(Mθ,h)), {By definition} Let Score Acc(Mθ,h) = k for sufficiently large k Let us assume that P-Acc(Mθ,h) < 0 Score Acc(Mθ,h) (α fsig(β EGISES(Mθ,h))) < 0 = fsig(β EGISES(Mθ,h)) = δ+ + Score Acc(Mθ,h) α where δ+ is a positive quantity = EGISES(Mθ,h) = 1 1 (δ+ + Score Acc(Mθ,h) δ+ + Score Acc(Mθ,h) α, β, δ+, s.t. Score Acc(Mθ,h) 0 The proof implies that P-Acc can take a negative value even for a reasonably high accuracy model. As an example, for α = 0.5, β = 1 (i.e., full importance to lack of personalization), EGISES(Mθ,h) 0.9 if Score Acc(Mθ,h) 0.35 then P-Acc < 0. Theorem 4 (Limitation 3: Anomalous Behavior). There exists anomalous condition when P-Acc(Mθik,h) < P-Acc(Mθil,h) even when EGISES(Mθik,h) < EGISES(Mθil,h) (NT: lower is better). (k, l) pair : P-Acc(Mθik,h) < P-Acc(Mθil,h) < 0 {From theorem 3} Score Acc(Mθik,h) Score Acc(Mθil,h) < α (fsig(β EGISES(Mθik,h) fsig(β EGISES(Mθil,h)) = fsig(β EGISES(Mθik,h) fsig(β EGISES(Mθil,h) > Score Acc(Mθik,h) Score Acc(Mθil,h) = EGISES(Mθik,h) EGISES(Mθil,h) > Score Acc(Mθik,h) Score Acc(Mθil,h) (EGISES(Mθik,h), EGISES(Mθil,h)) > 1 α (Score Acc(Mθik,h), Score Acc(Mθil,h)) which is a possible condition. A bounding case of the proof condition is when for Score Acc(Mθik,h) = Score Acc(Mθil,h) = 0, EGISES(Mθik,h) = 0 (best case) = P-Acc(Mθik,h) = 1 (worst case), and EGISES(Mθil,h) = 1 (worst case) = P-Acc(Mθil,h) = 0.37 (better case). Hence, P-Acc is not suitable for models that have relatively low accuracy scores. The fact that P-Acc is useful to regulate accuracy (and not measure personalization) is evident from Table 8 in Appendix C.2 where we observe that, unlike Per SEval-scores as outlined in section 8.1, P-Acc scores are heavily dominated by the evaluated PENS models accuracy scores and therefore, do not differ significantly even though they lack responsiveness (i.e., low EGISES scores; see Table 7 in Appendix C.2). As a remedy, we propose Per SEval (Personalized Summarizer Evaluator) in the next section. 5 Per SEval: Measure for Personalization The design objective of Per SEval is two-fold: (i) to ensure that a model is penalized for poor accuracy performance, but at the same time, (ii) to ensure that the evaluation of responsiveness (i.e., DEGRESS) is not obfuscated by high accuracy (since high accuracy does not entail high responsiveness as proved in Vansh et al. (2023)). In other words, the underlying maxim behind the design of Per SEval is accuracy is not a reward, but lack of it is surely a penalty! We term this penalty as Effective DEGRESS Penalty Factor (EDP). As per the design maxim, if accuracy is 100%, then there will be no EDP applied, and the Per SEval score will be the same as the DEGRESS score. In the subsequent sections, we develop the motivation and formulation of EDP. 5.1 Accuracy-drop Penalty (ADP) In this section, we introduce the first component of EDP - Accuracy-drop Penalty (ADP). ADP is a document-level penalty due to a drop in accuracy for the best-case scenario where a model-generated summary of document di (suij) is closest to the corresponding reader s expected summary uij. In this case, we denoted suij as sui . We define ADP as follows: Definition 5. Accuracy-drop Penalty. Given document di and user-generated summaries Udi, the document-level ADP of a model Mθ,h, denoted by ADP(sui*|(di, ui*)), is the relative deviation of the best performance (σ (sui , ui )|di) of Mθ,h σ(sui , ui )|di from the best possible performance (i.e., 0) w.r.t its proximity to the worst possible performance (i.e., 1). Document-level ADP is formulated as follows: ADP(sui*|(di, ui*)) = 1 1 + 10γ 4 exp 10 σ (sui ,ui )|di 0 (1 σ (sui ,ui )|di)+ϵ where, σ (sui , ui )|di = |Udi| min j=1 σ(suij, uij)|di; {ϵ : An infinitesimally small number (0, 1)} Here, ADP is defined as a shifted sigmoid function where γ 4 helps to bring the minimum penalty to zero, while the factor of 10 in the exponentiation ensures that the maximum penalty reaches 1 when the ratio of the difference in the best case to the worst case is around 1.5 before it starts exploding (i.e., over-penalization). ADP ensures that even if the DEGRESS score is acceptable, a penalty due to accuracy drop can still be imposed as a part of EDP. ADP, however, fails to address the scenario where the best-case scenario is acceptable (i.e., accuracy is fairly high) but is rather an outlier case i.e., for most of the other model-generated summary versions, there is a considerable accuracy drop. To address this issue, we introduce a second penalty component within EDP called Accuracy-inconsistency Penalty (ACP). 5.2 Accuracy-inconsistency Penalty (ACP) ACP accounts for outlier conditions of the best performance, as explained previously. It evaluates how a model performs w.r.t accuracy for a specific generated summary compared to its average performance. More formally, it is defined as follows: Definition 6. Accuracy-inconsistency Penalty. Given document di and user-generated summaries Udi, the summary-level ACP of a model Mθ,h, denoted by ACP(suij|(di, uij)), is defined as the relative deviation of the summary performance σ(suij, uij)|di of Mθ,h from its best performance σ (sui*, ui*)|di as compared to the deviation of its average performance σ(sui , ui )|di from σ (sui*, ui*)|di. It is to be noted that, unlike ADP, ACP is a summary-level measure. This penalty evaluates if the model consistently performs w.r.t accuracy and therefore, conversely, does not inject any additional penalty to DEGRESS when a model is consistent. The summary-level ACP is formulated as: ACP(suij|(di, uij)) = 1 1 + 10γ 4 exp 10 σ(suij ,uij)|di σ (sui ,ui )|di (σ(sui ,ui )|di σ (sui ,ui )|di)+ϵ where, σ(sui , ui )|di = 1 |Udi| j=1 σ(suij, uij)|di 5.3 Per SEval: Formulation We now lay the design of the Per SEval framework as an extension to DEGRESS (i.e., 1 EGISES). A multiplicative injection of EDP (0, 1] should be such that the best accuracy (i.e., ADP = 0) with no inconsistency (i.e., ACP = 0) would lead to an EDP value of 1, and thereby, DEGRESS remains unobfuscated (which is the desired objective). The following formulation of Per SEval guarantees these properties: Per SEval(suij|(di, uij)) = DEGRESS(suij|(di, uij)) EDP(suij|(di, uij)) where, EDP(suij|(di, uij)) = 1 1 1 + 10α 3 exp (10β 1 DGP(suij|(di, uij))) and, DGP(suij|(di, uij)) = ADP(sui*|(di, ui*)) + ACP(suij|(di, uij)) The system-level Per SEval score is as follows: Per SEval(Mθ,h) = j=1 Per SEval(suij |(di,uij)) We design EDP as an inverse sigmoid function of the overall DEGRESS Penalty (DGP), which sums up ADP and ACP. α and β are hyper-parameters that help to control the shape of EDP. α 3 ensures that EDP is 1 (i.e., 1 0) when there is no penalty, thereby making Per SEval equivalent to DEGRESS. β 1 ensures that the function does not drop sharply, thereby over-penalizing (and hence, dampening) an otherwise fairly good DEGRESS score (i.e., responsiveness). Since β may significantly affect the overall human-judgment correlation, we did an ablation study (Fig 3; Section 8.1) to find the optimal value (which was observed to be 1.7). The system-level Per SEval [0, 1] and is bounded by the system-level DEGRESS score. 6 Benchmarking of SOTA Summarization Models w.r.t Per SEval 6.1 Model Benchmarking Dataset Microsoft PENS Dataset (News Domain). Our study, as in (Vansh et al., 2023), assesses models using test data from the PENS dataset provided by Microsoft Research (Ao et al., 2021)4. This dataset pairs news headlines with articles, serving as concise summaries. The test set creation involved two phases: initially, 103 English speakers selected 50 articles of their interest from a pool of 1000, sorted based on exposure time. In the second phase, participants generated preferred headlines (gold references) for 200 articles without knowledge of the originals. The assignment ensured an average of four gold-reference summaries per article. The PENS dataset was chosen because it is the only one that contains the users reading history, i.e., the temporal sequence of interactions (clicking, reading, and user-generated gold summaries), making it ideal for evaluating five SOTA personalized summarization models that require user reading history as an input. Open AI CNN/Daily Mail Dataset (News Domain). To understand the applicability of Per SEval on mainstream gold-standard news datasets, we design an indirect evaluation methodology with the Open AI CNN/DM dataset (validation and test) released by Stiennon et al. (2020). The test dataset contains 639 articles processed by 19 policies, resulting in 5515 summaries rated by the same annotators on accuracy, coherence, coverage, and overall quality using a 7-point Likert scale. However, unlike the PENS dataset, this dataset lacks the temporal span of human evaluator interactions forming the reading history required by the PENS summarization models. Additionally, we do not have gold-reference summaries (but only ratings).5 Hence, to conduct our experiments, we evaluated the personalization of 4 (out of 19) Open AI policies, focusing on RLHF-tuned policies, as base-model and SFT-based policies are non-personalized. Since the policy-generated summaries are not specific to any human annotator as the policies are not trained on any reading history, we do not observe multiple user-specific summaries from the same policy - a necessary condition for evaluating personalization. To address this, we concatenate all summaries generated by different policies (say, k such summaries) that were rated above 6 (out of rmax = 7) by a specific annotator (say, Alice) into one combined gold-reference summary as per Alice (u Alice). After this 4We comply with the Microsoft Research License Terms. 5It is to be noted that human-written gold reference summaries were not always selected or rated as highest by the human annotators, clearly indicating the subjectivity of saliency. step, the content of each of all the policy-generated summaries (s) that received some rating ([1, rmax]) from Alice (r Alice(s) r(s,Alice)) is then concatenated (k + (rmax r(s,Alice)))-times to create a policy-generated surrogate personalized summary (s Alice) for Alice. Ideally, this surrogate should match Alice s combined gold-reference u Alice. If r(s,Alice) is much lower than rmax, then the surrogate should get penalized. The injection of redundant, unmatchable content helps to achieve this. We repeat this process for all annotators and generate the Per SEval scores of each policy across 27 (out of 639) articles. Since k and r(s, ) can differ between annotators, the penalty (and thereby length) of the surrogates will subjectively differ as well. Open AI TL;DR (Reddit) Dataset (Open Domain). To understand the broader applicability of Per SEval, we also appropriated the Open AI TL;DR dataset Stiennon et al. (2020). This dataset is a collection of 123,169 Reddit posts adopted from the dataset by Völske et al. (2017) covering 29 subreddits (i.e., domains) of varied types, their corresponding Open AI policy-generated summaries, and the human-written gold-reference summaries. A subset of the validation dataset comprises 1038 posts that were fed into 13 policies to generate 7713 summaries. Like the Open AI CNN/DM dataset, 32 human annotators rated the summaries in the same format as in the previous Open AI CNN/DM dataset. As a part of the appropriation, we carry the same method as that with CNN/DM leading to the evaluation of 4 RLHF-tuned policies (out of 13) and 57 (out of 1038) annotated articles. However, we would like to emphasize that this indirect benchmarking is neither an equivalent evaluation methodology to the PENS dataset based one, nor is a simulation of the required setup. There is so far no other dataset comparable to PENS for evaluating personalization, and hence, all benchmarking results in the outlined method should be seen as baseline within the scope of the method. 6.2 SOTA Summarization Models Evaluated Personalized Summarization Models (Microsoft PENS dataset). We study ten SOTA summarization models for comparative benchmarking as in (Vansh et al., 2023). Five of them are specifically trained personalized models and follow the PENS framework (Ao et al., 2021): PENS-NRMS Injection-Type-1 (T1)/Type-2 (T2), PENS-NAML T1, PENS-EBNR T1 and PENS-EBNR T2. The others are generic SOTA models - BRIO (Liu et al., 2022), Sim CLS (Liu & Liu, 2021), Big Bird-Pegasus (Zaheer et al., 2020), Prophet Net (Qi et al., 2020), and T5-base (Orzhenovskii, 2021). The selections are based on their consistent top-5 ranking over the preceding four years on the CNN/Daily Mail news dataset Hermann et al. (2015). Appendix A contains model descriptions. Non-personalized Summarization Models (Microsoft PENS dataset). For the non-personalized models, we follow the evaluation setup used by Vansh et al. (2023) by augmenting the documents with the reference summaries of each reader as document titles (i.e., cues). This results in subjective document versions corresponding to each reader. The models ideally should pick up the cues and generate them back as an output, thereby inducing an apparent sense of personalization. This injection process provides robust baselines for comparative evaluation. RLHF-tuned Open AI Summarization Policies (Open AI CNN/DM & TL;DR (Reddit) datasets). We analyze five Open AI policies (parameter size 1.3B and 6.7B) that have been RLHF-tuned using TL;DR (Reddit) training datasets. The base pre-trained model architecture is similar to that of GPT-3. The base model is then further fine-tuned in a supervised setup on a filtered set of the Reddit posts to generate SFT-models (denoted "sup"). The Reward Model (denoted "rm") is a linear head that predicts the rating. The Reinforcement Learning of the policy model is done via the PPO algorithm Zheng et al. (2023b). 6.3 Baseline Distance Metrics and Scores Per SEval is a generic measurement framework (like EGISES) where the specific metric space M on which σ is defined should be appropriately selected such that we achieve the best human-judgment (HJ) correlation. In this paper, we choose seven summarization accuracy metrics that are defined on standard algebraic spaces and plug them in Per SEval in isolation as distance metrics (i.e., σ) : (i) ROUGE (RG)-L, (ii) ROUGE (RG)-SU4, Basline-I: Microsoft PENS Test Dataset with Cue Injection Models PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Big Bird-Pegasus 0.205 0.143 0.168 0.156 0.253 0.320 0.195 Sim CLS 0.079 0.032 0.016 0.014 0.157 0.286 0.18 Prophet Net 0.054 0.027 0.018 0.016 0.097 0.228 0.135 T5 (Base) 0.035 0.022 0.011 0.016 0.073 0.207 0.124 BRIO 0.030 0.021 0.008 0.013 0.107 0.242 0.154 Personalization Benchmarking: Microsoft PENS Test Dataset with Human Reading History PENS-NAML T1 0.018 0.013 0.019 0.010 0.025 0.074 0.051 PENS-NRMS T1 0.016 0.011 0.016 0.008 0.022 0.062 0.045 PENS-EBNR T1 0.009 0.008 0.010 0.005 0.015 0.037 0.029 PENS-EBNR T2 0.002 0.005 0.003 0.002 0.006 0.008 0.013 PENS-NRMS T2 0.002 0.004 0.003 0.002 0.006 0.007 0.013 Basline-II: Open AI CNN/DM Test and TL;DR (Reddit) Validation Datasets with Human-Feedback (i.e., Rating) Models (i.e., Policies) PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ sup4/rm4 (6.7B) 0.943/0.719 0.941/0.730 0.906/0.525 0.887/0.395 0.943/0.732 0.998/0.719 0.999/0.836 sup4/rm4-t.7 (6.7B) NA/0.711 NA/0.648 NA/0.593 NA/0.465 NA/0.737 NA/0.790 NA/0.818 sup4/rm4 (1.3B) 0.617/0.523 0.617/0.502 0.413/0.357 0.612/0.163 0.617/0.566 0.994/0.795 0.999/0.605 sup4/rm4-kl14 (6.7B) 0.783/NA 0.783/NA 0.562/NA 0.592/NA 0.783/NA 0.884/NA 0.783/NA sup4/rm4-t.7 (1.3B) 0.505/0.481 0.5/0.414 0.267/0.417 0.266/0.323 0.502/0.523 0.92/0.776 0.646/0.602 Table 1: SOTA model-benchmarking on Microsoft PENS and Open AI CNN/DM & TL;DR (Reddit) datasets w.r.t Per SEval (PSE) (NA: not applied on the dataset); Per SEval hyper-parameters: α = 3, β = 1.7 (optimal β; see 3), and γ = 4. Observation 1: Disagreement between keyword-based (RG-L, RG-SU4, METEOR, BLEU) and non-keyword-based PSE variants (JSD, BScore, Info LM) in terms of non-personalized baseline leaderboard from rank-3 onward; Observation 2: Info LM-αβ has more reliable discriminatory performance with less sharp changes for PENS dataset; Observation 3: Specialized personalized models lack personalization w.r.t all variants of Per SEval; Observation 4: RLHF-tuned LLM-based policies exhibit much higher personalization baseline (i.e, baseline-II) compared to non-personalized specialized summarization models (i.e., baseline-I) within the limited scope of the surrogate-based indirect evaluation; Observation 5: PSE-scores are very high for top-performing Open AI policy due to most surrogate summaries being highly rated and very similar to the combined gold-references. (iii) BLEU-1, (iv) METEOR,6 (v) Bert Score (BScore) defined on embedding space, (vi) Jenson-Shannon Distance (Menéndez et al., 1997) on probability space, and (vii) Info LM-αβ (α = 1; β = 1) on probability space generated from the embedding space of a masked-LM. RG is chosen because it has a very high HJ (Pearson & Kendall) correlation (> 0.7) in most standard datasets such as CNN/DM and TAC-2008 (for RG-L) as reported in (Bhandari et al., 2020; Zhang et al., 2024), and DUC-2001/2002 (for RG-L/SU-4) (Lin, 2004). For the same reason, the αβ variant of Info LM is chosen. Comprehensive benchmark results w.r.t optimal Per SEval hyperparameters (i.e., α = 3, β = 1.7, γ = 4; see Figure 3 for ablation results) for each of the variants are given in Table 1. We observe that most non-personalized models, such as Big Bird-Pegasus, produce significantly stronger baselines across most Per SEval variants. However, keyword-based PSE variants (RG-L, RG-SU4, METEOR, BLEU) disagree with non-keyword-based PSE variants (JSD, BScore, Info LM) in terms of the non-personalized baseline leaderboard from rank-3 onward. One reason for this is that JSD, BScore, and Info LM being embedding-based can handle out-of-distribution (OOD) situations leading to lesser penalty for keyword-level mismatch that happens in abstraction summarization. We also find that the Info LM-αβ variant shows "smoother discrimination" (i.e., no sharp jump in the performance) when compared to RG-SU4, BLEU, and JSD. At the same time, the discriminatory performance of Info LM-αβ and JSD variants is much better than BScore, METEOR, and RG variants, leading to a more reliable leaderboard. The results on the Open AI datasets should be seen within the context that: (a) the models are essentially policy variants that differs w.r.t size and specific PPO implementation, and (b) the indirect benchmarking method is limited (as outlined in section 6.2). 6First four defined as similarity functions on the string space while BScore is defined as similarity function (as opposed to distance function) on high dimensional embedding space; see Appendix B.1 for details on each of the seven measures. 7 Meta-evaluation of Per SEval: Experiment Design In this section, we lay down the experiment design that forms the foundation to establish (i) the reliability of Per SEval w.r.t human-judgment correlation, (ii) the stability of Per SEval, (iii) Per SEval as a rank-measure is not entailed by DEGRESS-rank (i.e., the EGISES paradox is empirically existent), and (iv) Per SEval can be a standalone rank-measure without the need of any aggregated ranking. 7.1 Meta-evaluation of Per SEval Reliability: Survey-based Direct Method Meta-evaluation Objective. As pointed by Vansh et al. (2023), unlike the meta-evaluation of accuracy measures, direct meta-evaluation of personalization is not a feasible task since that would require human evaluators to work as a team and to understand how their subjective assessment of the Per SEval scores (i.e., to what extent they agree with the scores) corresponds with the differences in the model-generated summaries and their own subjective expected summaries. As an alternative, we propose a survey-based evaluation methodology to simulate this scenario. We argue that the meta-evaluation of Per SEval should have two objectives: (i) whether human evaluators would judge the responsiveness of models in the same "ratio-way" as what DEGRESS does, and (ii) whether human evaluators would apply the accuracy drop to the responsiveness in the same "factor-way" as what Per SEval does. In other words, the central objective is to validate to what extent human evaluators agree with the design principles of Per SEval at a cognitive level. Participants. We opened the survey to a selected pool of graduate students demonstrating fair English comprehension. The pool consisted of students from five different backgrounds: (i) computer science, (ii) electrical engineering, (iii) humanities & social sciences, (iv) mathematics, and (v) physics. The survey was promoted in ongoing courses, student associations, and social media groups. 169 students ( 45% male, 55% female) within the age group of 25-40 completed the survey. No other personal details were asked. Survey Procedure. Each participant was shown a pair of gold-reference summaries (corresponding to two user profiles) for a specific news article from the PENS dataset. Along with this, five pairs of model-generated summaries were shown, each pair corresponding to five of the ten models studied in this paper (i.e., two participant responses covered all the ten models for a given document). To eliminate response bias, the participants were unaware of which of the six pairs was a gold-reference, and the model names were also not revealed. The same set (shuffled) was shown to two random participants to get an average. Each participant was asked to provide similarity ratings between 1 (low) and 6 (very high) for the summary pairs. A sample snapshot questionnaire is provided in Appendix D (figure 4). Modeling Human-judgment of Personalization (Per SEval-HJ). We model a "human version" of DEGRESS (termed DEGRESS-HJ) using the normalized rating as the divergences (i.e., σ(uij, uik) and σ(suij, suik) for document di and gold-references uij and uik). As mentioned above, if Per SEval needs to be reliable, then the necessary condition is that DEGRESS should have a high correlation with DEGRESS-HJ, failing which it can be concluded that human evaluators do not interpret divergences in the ratio-way of DEGRESS. We use the standard Pearson s coefficient (r), Spearman s ρ, and Kendall s τ rank coefficients for this. As a sufficient part of the reliability test, we model the "human version" of EDP using standard accuracy measures (i.e., RG-L, RG-SU4, METEOR, BLEU, and Info LM-αβ) as surrogates. The motivation behind this is that such measures are known to have high HJ-correlation. The objective was to check whether such a surrogate, if used as a factor (just as in Per SEval) with Per SEval-HJ, shows a high correlation with Per SEval, failing which implies that human evaluators do not cognitively resonate with this factor-styled discounting of DEGRESS 7.2 Meta-evaluation w.r.t Stability Per SEval, being a rank measure, is said to be stable if, for any random sample of document (and corresponding gold-references) selected from the evaluation dataset, the rank of the evaluated models does not change. This meta-measure is important because it objectively checks if Per SEval can be relied on any arbitrary personalization evaluation dataset, unlike the Per SEval-HJ based reliability evaluation, which is subjective and indirect in nature. In order to establish stability, we need to define it w.r.t a specific sampling method. Sampling Method for Stability Meta-evaluation. To understand the stability of Per SEval, we create random sample collections Ck (D,S,U), where k = {80%, 60%, 40%, 20%} of the PENS dataset and N = |(D, S, U)|. Figure 3: PSE-ILM Ablation: Effect of β on HJ-Corr; Optimal performance at β = 1.7 across all three standard correlation measures (Pearson r, Spearman ρ, Kendall τ). Each collection has ten random sample sets nk i N; i = [1 : 10] (with replacement). We benchmark the models on all 40 sample sets to obtain corresponding leaderboards. We compare that with the rank obtained from the entire dataset. We formalize this notion of stability of a rank measure as follows: Definition 7. Weakly Stable Rank Measure. A rank measure is ϵ-weakly stable if the maximum rankcorrelation (w.r.t stat. τ) between the measure-generated model-ranking on each nk i and model-ranking on the entire dataset N is ϵ. Definition 8. Strongly Stable Rank Measure. A rank measure is δ-strongly stable if (i) it is ϵ-weakly stable, (ii) the bias over C(D,S,U) w.r.t the mean score of each Ck (D,S,U) for each evaluated model is δb, and (iii) the variance over C(D,S,U) w.r.t the expected variance of the scores of each Ck (D,S,U) for each model is δvar; δ = max(δb, δvar). 8 Observations and Insights In this section, we provide empirical support for the reliability and stability of Per SEval and show that accuracy leaderboards may be misleading (or, at best, redundant) for personalization analysis. 8.1 Meta-evaluation of Per SEval: Results Reliability of Per SEval. We compute the HJ-correlation of the seven variants of Per SEval w.r.t to each of the five Per SEval-HJ variants (RG-L, RG-SU4, METEOR, BLEU, and Info LM-αβ; see Table 2 for the results7). An 11-point hyper-parameter ablation study shows that the optimal correlation is at β = 1.7 (Figure 3). We observe that Per SEvalβ=1.7-Info LM-αβ has the overall best correlation across all five Per SEval-HJ variants. We further observed that: (a) Per SEval-HJInfo LM αβ as a human-judgment estimate has the best performance of each Per SEval-variants (Pearson s r = 0.79; Spearman s ρ = 0.68; Kendall s τ = 0.47 w.r.t PSE-HJInfo LM-αβ), and (b) Per SEval-Info LM-αβ performs the best across. The high performance of Per SEvalβ=1.7-Info LM-αβ can attributed to Info LM-αβ, as an accuracy (and thereby, distance measure) having a very high system-level HJ-correlation (Pearson s τ 0.95; Kendall s τ 0.85) when compared to the other measures on the CNN/DM dataset, as reported in Colombo et al. (2022b). Indirect Evaluation Method is Inadequate for Meta-evaluation. As discussed in section 6.2, the method of evaluating personalization of RLHF-tuned policies is indirect and limited in scope, since these policies are trained on annotator profile agnostic ratings and not on their reading history. Hence, although the Open AI dataset may be used for generating baselines in this limited setup, they are not suitable for meta-evaluation of any personalization measure. To validate this point, we design an Open AI datasetoriented Per SEval-HJ variant where we model the divergence between policy-generated surrogate summaries 7Per SEval-HJ: Human judgment est.; Stat. Significance of Corr. (Strong, Moderate, Low, None): p-value < 0.01. Microsoft PENS dataset/Open AI CNN-DM dataset/Open AI Tl;DR (Reddit) dataset Per SEval-HJRG-L HJ-Corr. PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Pearson s r 0.23/ 0.90/-0.89 0.34/ 0.89/-0.81 0.49/ 0.89/-0.83 0.43/ 0.67/-0.63 -0.05/ 0.90/-0.92 -0.41/0.02/0.07 0.79/ 0.21/-0.71 Spearman s ρ -0.25/ 0.80/-0.80 -0.27/ 0.80/-0.80 0.15/ 0.80/-0.80 -0.28/ 0.40/-0.80 -0.26/ 0.80/-1.00 -0.28/0.20/0 0.69/ 0.40/-0.80 Kendall s τ -0.20/ 0.67/-0.67 -0.24/ 0.67/-0.67 0.16/ 0.67/-0.67 -0.24/ 0.33/-0.67 -0.20/ 0.67/-1.00 -0.24/0/0 0.47/ 0.33/-0.67 Per SEval-HJRG-SU4 HJ-Corr. PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Pearson s r 0.91/ 0.90/-0.89 0.96/ 0.89/-0.81 0.98/ 0.67/-0.82 0.98/ 0.89/-0.62 0.77/ 0.89/-0.92 0.49/0.03/0.07 0.73/ 0.21/-0.70 Spearman s ρ 0.07/ 0.80/-0.80 0.05/ 0.80/-0.80 -0.10/ 0.80/-0.80 -0.02/ 0.40/-0.80 0.16/ 0.80/-1.00 0.15/0.20/0 0.77/ 0.40/-0.80 Kendall s τ 0.11/ 0.67/-0.67 0.07/ 0.67/-0.67 -0.07/ 0.33/-0.67 -0.02/ 0.67/-0.67 0.20/ 0.67/-1.00 0.16/0/0 0.6/ 0.33/-0.67 Per SEval-HJMETEOR HJ-Corr. PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Pearson s r 0.06/ 0.90/-0.90 0.18/ 0.90/-0.81 0.36/ 0.89/-0.86 0.29/ 0.67/-0.67 -0.23/ 0.90/-0.92 -0.56/0.02/0.08 0.77/ 0.21/-0.72 Spearman s ρ -0.28/ 0.80/-0.80 -0.33/ 0.80/-0.80 0.12/ 0.80/-0.80 -0.27/ 0.40/-0.80 -0.32/ 0.80/-1.00 -0.37/0.20/0 0.71/ 0.40/-0.80 Kendall s τ -0.29/ 0.67/-0.67 -0.33/ 0.67/-0.67 0.07/ 0.67/-0.67 -0.24/ 0.33/-0.67 -0.38/ 0.67/-1.00 -0.42/0/0 0.51/ 0.33/-0.67 Per SEval-HJBLEU HJ-Corr. PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Pearson s r 0.84/ 0.90/-0.89 0.91/ 0.89/-0.81 0.95/ 0.89/-0.82 0.95/ 0.67/-0.62 0.67/ 0.89/-0.92 0.36/0.03/0.07 0.63/ 0.21/-0.71 Spearman s ρ -0.13/ 0.80/-0.80 -0.12/ 0.80/-0.80 -0.08/ 0.80/-0.80 -0.19/ 0.40/-0.80 -0.03/ 0.80/-1.00 -0.02/0.20/0 0.65/ 0.40/-0.80 Kendall s τ -0.16/ 0.67/-0.67 -0.11/ 0.67/-0.67 -0.07/ 0.67/-0.67 -0.20/ 0.33/-0.67 -0.07/ 0.67/-1.00 -0.02/0/0 0.51/ 0.33/-0.67 Per SEval-HJJSD HJ-Corr. PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Pearson s r 0.54/ 0.90/-0.89 0.59/ 0.90/-0.80 0.64/ 0.89/-0.86 0.62/ 0.67/-0.68 0.43/ 0.90/-0.92 0.12/0.01/0.06 0.43/ 0.21/-0.72 Spearman s ρ 0.02/ 0.80/-0.80 -0.02/ 0.80/-0.80 -0.07/ 0.80/-0.80 -0.11/ 0.40/-0.80 0.11/ 0.80/-1.00 0.07/0.20/0 0.50/ 0.40/-0.80 Kendall s τ 0.01/ 0.67/-0.67 -0.04/ 0.67/-0.67 -0.08/ 0.67/-0.67 -0.15/ 0.33/-0.67 0.14/ 0.67/-1.00 0.09/0/0 0.38/ 0.33/-0.67 Per SEval-HJBScore HJ-Corr. PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Pearson s r 0.65/ 0.93/-0.76 0.59/ 0.92/-0.66 0.47/ 0.89/-0.73 0.51/ 0.70/-0.52 0.81/ 0.93/-0.80 0.83/-0.04/-0.17 0.56/ 0.24/-0.50 Spearman s ρ 0.60/0.80/-0.80 0.56/ 0.80/-0.80 0.24/ 0.80/-0.80 0.47/ 0.40/-0.80 0.69/ 0.80/-1.00 0.64/0.20/0 0.84/ 0.40/-0.40 Kendall s τ 0.72/ 0.67/-0.67 0.71/ 0.67/-0.67 0.25/ 0.67/-0.67 0.66/ 0.33/-0.67 0.82/ 0.67/-1.00 0.81/0/0 0.73/ 0.33/-0.33 Per SEval-HJInfo LM αβ HJ-Corr. PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Pearson s r 0.49/ 0.91/-0.76 0.60/ 0.90/-0.71 0.66/ 0.88/-0.61 0.58/ 0.67/-0.35 0.78/ 0.90/-0.79 0.51/-0.04/-0.10 0.76/ 0.20/-0.67 Spearman s ρ 0.21/ 0.80/-0.40 0.60/ 0.80/-0.40 0.50/ 0.80/-0.40 0.55/ 0.40/-0.40 0.62/ 0.80/-0.80 0.47/ 0.20/-0.60 0.61/ 0.40/-0.40 Kendall s τ 0.24/ 0.67/-0.33 0.51/ 0.67/-0.33 0.42/ 0.67/-0.33 0.47/ 0.33/-0.33 0.45/ 0.67/-0.67 0.38/ 0/-0.33 0.42/ 0.33/-0.33 Table 2: Per SEval (PSE) Reliability: Human-judgment (HJ) corr. between PSEβ=1.7-X and Per SEval-HJX on Microsoft PENS/Open AI CNN-DM/Open AI TL;DR (Reddit) datasets. Observation 1: Per SEval Info LM-αβ has maximum highest HJ correlation across all Per SEval-HJ variants on PENS dataset; Observation 2: Per SEval-HJ-Info LM-αβ and Per SEval-HJ-BScore are optimal meta-evaluation measures since they are most closely aligned to all the seven Per SEval variants; Observation 3: Difference in human annotator ratings does not serve as reliable surrogate for estimating DEGRESS-HJ (and thereby, Per SEval-HJ) as evident from contradictory results on structurally equivalent Open AI CNN/DM and TL;DR (Reddit) datasets; Observation 4: As a follow-up, the indirect evaluation method can serve for baseline generation but not meta-evaluation. (σ(suij, suik)) with the difference in the ratings given by the annotators (i.e., |r(suij) r(suik)|) and that between combined gold-references (σ(uij, uik)) with the difference in the average ratings of all summaries included in the combined gold reference (i.e., | r(uij) r(uik)|). For calculating penalty due to EDP, we take the distance between gold-reference and generated summary to be |r(suij) r(uij)). We then create seven variants of Per SEval-HJ-X as in the PENS dataset-based evaluation, where X stands for the measure used for estimating the DEGRESS normalization factor (i.e., the distance of sui and ui from the document di; see equation 2). It is evident from Table 2 that the HJ-correlation so computed is unstable since the results are contradictory even though both the Open AI datasets are structured in exactly the same manner with four common policies. This shows that the difference in rating cannot be an alternative to the direct similarity judgment that was collected via the survey. Inter-Corr. EG-RG-L EG-RG-SU4 EG-METEOR EG-BLEU EG-JSD EG-BScore EG-Info LM-αβ Spearman s ρ 0.903 0.9758 0.8667 0.9636 0.997 0.8476 0.9879 Kendall s τ 0.8222 0.9111 0.7333 0.9111 0.9888 0.7047 0.9555 Table 3: EGISES (EG-X) Paradox: PSE-Xβ=1.7 rank-disagreement (< 1 inter-correlation) due to EDP. See Table 7 in Appendix C.2 for detailed EGISES scores. PENS Test Dataset Sample Set (Random Selection) Ranking Models 100% 80% 60% 40% 20% Bias Variance 1 Big Bird-Pegasus 0.15 0.153 0.1535 0.1558 0.1564 0.0024 5.81E-06 2 Sim CLS 0.133 0.1351 0.1356 0.1357 0.138 0.0018 3.08E-06 3 BRIO 0.112 0.1143 0.1151 0.116 0.1155 0.0013 1.77E-06 4 Prophet Net 0.098 0.1003 0.1018 0.1012 0.1031 0.0017 2.90E-06 5 T5 (Base) 0.088 0.0899 0.0905 0.091 0.0912 0.0012 1.33E-06 6 PENS-NAML T1 0.036 0.0364 0.0367 0.0376 0.039 0.0012 1.41E-06 7 PENS-NRMS T1 0.0315 0.0326 0.0327 0.0331 0.033 0.0006 3.26E-07 8 PENS-EBNR T1 0.0206 0.0209 0.021 0.0212 0.0228 0.0008 6.00E-07 9 PENS-NRMS T2 0.0103 0.0103 0.0102 0.0107 0.0111 0.0003 1.14E-07 10 PENS-EBNR T2 0.0096 0.0097 0.0097 0.0103 0.0107 0.0004 1.84E-07 Table 4: Per SEvalβ=1.7-Info LM-αβ Stability: 0.0024-strongly-stable w.r.t ϵ-Spearman = 1; ϵ-Kendall = 1. HJ-Corr. PSE-ILM-αβ BK(PSE-ILM-αβ, RG-L) BK(PSE-LM-αβ, ILM-αβ) BK(PSE-ILM-αβ, BLEU) Spearman ρ 0.6849 0.1656 -0.3447 0.632 Kendall τ 0.4667 0.0698 -0.3027 0.46 Table 5: Accuracy-leaderboards may mislead: HJ-Corr. of Borda-Kendall (BK) consensus-based aggregated rank vs. PSEβ=1.7-Info LM (ILM)-αβ. Evidence of the EGISES Paradox. We look at the inter-correlations between EGISES-rank and Per SEval-rank of the selected models on the PENS dataset. The values less than 1 across all the variants (see Table 3) suggest that Per SEval is not just an offset of EGISES and is not entailed by it (which would have otherwise been the case if the EGISES paradox was non-existent in reality). Stability of Per SEval. We compute ϵ-δ-stability of the best performing Per SEval-Info LM-αβ on PENS over the ten models as per the sampling method and stability definitions in Section 7.2. We observe Per SEval to be 0.00248-strongly-stable w.r.t Spearman-ϵ = 1 and Kendall-ϵ = 1 (see Table 4). This establishes a very high stability of Per SEval along with its reliability, making it robust. For detailed results, see Appendix C.1. 8.2 Accuracy Leaderboards may Mislead We demonstrate that accuracy leaderboards, at best, are redundant and Per SEval-rank is sufficient to capture personalization. For this, we generate the Borda-Kendall consensus-based aggregated rank (Colombo et al., 2022a) and compare the HJ-correlation with that of Per SEval-Info LM-αβ. We observe that the stand-alone HJ correlation has the same strength w.r.t the aggregated rank for accuracy measures like BLEU, thereby rendering them redundant in the context of personalization evaluation, while is significantly higher (Spearman ρ: 0.51+ ; Kendall τ: 0.40+ ) than that of measures such as RG-L, Info LM-αβ (see Table 5). This indicates that accuracy ranking can also inject noise. 8.3 Computational Complexity of Per SEval. We now show that the best performing Per SEval-αβ is linear time w.r.t to summary lengths - i.e., O(ls+lu) for a fixed evaluation dataset |D| where lsu and lu are lengths of model-generated summary su and user-generated reference summary u. 8Maximum of δ-bias and δ-variance over all ten models Theorem 5 (Per SEval Time Complexity). The best worst-case time complexity of Per SEval is O (|D| (lsu + lu)). Proof. Let tσX be the time-complexity of computing the distance using measure σX.9 O(DEGRESS(suij|(di, uij))) = O(|Udi| tσX ) = O(tσX ) { |Udi| = number of reference summaries per article constant C}; = O(DEGRESS) = |D| (|Udi| O(DEGRESS(suij|(di, uij)))) = |D| O(tσX ) Now, O(EDP(suij|(di, uij)) = (O(ADP(suij|(di, uij))) + O(ACP(suij|(di, uij)))) = O(Udi tσX ) = O(tσX ) = O(Per SEval) = |D| O(tσX ) = O(|D| tσX ) For σX {RG-L, RG-SU4, JSD, Info LM-αβ}: O(tσX ) = O(lsu + lu); {Suffix-tree based matching for RG;} {No. of layers, dimensions, and vocabulary size for a fixed MLM (BERT) used in Info LM can be taken as constant} For σX {BLEU, METEOR, BScore}: O(tσX ) = O(lsu lu); {No. of layers, dimensions, and vocabulary size for a fixed BERT model used in BScore can be taken as constant} O(Per SEval) = O (|D| (lsu + lu)) for the best-performing Per SEval-αβ For computing Per SEval, we observe an average runtime of 2:37 minutes for each of the evaluated ten models on the PENS dataset (3840 articles), which is reasonable for an offline procedure.10 9 Related Work Evaluation of Personalization Personalization evaluation has been well studied in recommendation systems (recsys) (Zangerle & Bauer, 2022), such as metrics based on the Jaccard Index, rank-order edit distance (Hannak et al., 2013), MAE/RMSE/Hit-Ratio (Li et al., 2024), and n DCG (normalized Discounted Cumulative Gain) (Matthijs & Radlinski, 2011). A comprehensive compilation of all recsys-oriented metrics and their applications can be found in (Zangerle & Bauer, 2022; Kuanr & Mohapatra, 2021). While relevant to recsys, these metrics are not pertinent for text summarization since they rely on human feedback (such as clicks and likes) on a rank list of potentially preferable items" a situation that does not exist for summarizers. A survey-based qualitative analysis of the usefulness of model-generated summaries was proposed by Ter Hoeve et al. (2022). Although this work establishes empirically that model-generated summary utility is subjective (as argued in this paper), yet to date, the only work on the formal quantitative evaluation of personalization in summarization models is EGISES (Vansh et al., 2023) which, however, can only capture responsiveness. There is a growing interest in learned summarization evaluation metrics, as opposed to designed ones Peyrard & Eckle-Kohler (2017); Peyrard & Gurevych (2018); Böhm et al. (2019); Liu et al. (2023). LLM-based automated summarization evaluation has also been studied Shen et al. (2023); Zheng et al. (2023a). However, to the best of our knowledge, these works have not focused on evaluating personalized summarization models where subjectivity in saliency needs to be respected (and not normalized). Any such automated evaluator needs to be fine-tuned on large volumes of user profiles having varied subjective expectations. This sort of dataset is hard to design, and such a project only makes sense if designed metrics are not able to achieve minimum human-judgment correlation. 9In this paper, X: RG-L, RG-SU4, METEOR, BLEU, JSD, BScore (we assume pre-computed embeddings), Info LM-αβ which uses BERT Base (uncased; 110M params) as pre-trained Masked Language Model. 10System specifications: Machine architecture: x86_64; CPU: Intel(R) Xeon(R) Silver 4216 CPU @ 2.10GHz; CPU Cores: 16; Thread(s) per core: 2. Personalized Summarization Aspect-based. An aspect-based summarization model generates summaries that are coherent with the aspects (i.e., themes/topics) therein (Narayan et al., 2018; Frermann & Klementiev, 2019; Tan et al., 2020; Hirsch et al., 2021; Hayashi et al., 2021; Meng et al., 2021; Soleimani et al., 2022). While explicit aspects can be restrictive and rather broad (e.g., the MA news training dataset where six broad aspects are identified: sport", health", travel", news", science technology", tv showbiz"), implicit aspect-based summarization implies augmenting the aspect query with concepts that are related to the predefined aspects. Although these models have specific use-cases, they are not trained to adapt to the reader s evolving profile (i.e., reading behavioral pattern) that constitutes discourse-level interest drift (and not just topic-level static interests). Also, the evaluation was w.r.t accuracy using standard ROUGE-variants. Interactive Human-feedback-based. One of the earliest interactive interface-based iterative personalized summarization frameworks was proposed by Yan et al. (2011) where users could click on specific sentences in the generated summary that are of their interest (implicit preference), and read the associated context (i.e., surrounding text) of the selected sentence before sending this preference as feedback to the model for a revised version. A similar interactive interface-based framework was proposed by PVS et al. (2018) where users can iteratively select sentences and the phrases therein that they prefer (and do not prefer as well), while the summarizer, an ILP-based model proposed by Boudin et al. (2015), updates the summary based on this feedback until the user is satisfied. On similar lines, Ghodratnama et al. (2021) introduced a personalized summarization method for extractive summarization. Extracted summary concepts are presented to readers for their feedback, which is used to iteratively fine-tune the summary until no further negative feedback is received. Bohn & Ling (2021) proposed a framework where the acceptance or rejection of summary sentences was made dynamic as the summary gets generated on-the-fly. However, all these works have been evaluated based on standard accuracy evaluations (such as ROUGE variants). User-preference Trained Reward Model-based. Another way of inducing personalization in models is to train a base model (pre-trained or supervised fine-tuning) as an agent within a reinforcement learning framework (usually policy-gradient based) using a reward model (RM) as the environment. The RM is trained on human preferences to predict human ratings for specific actions of the agent (i.e., selecting specific words/sentences), thereby providing the rewards (Stiennon et al., 2020; Nguyen et al., 2022). However, whether these models can explicitly remember" individual preferences (or even that of user groups with similar interests) is still to be probed and not quite clear. Nevertheless, the evaluations were on accuracy only. User-preference-history-based. There has been considerable work in user preference-history-based product review summarization. The user profile is usually modeled in terms of discrete attributes such as rating, user ID, and product-ID, and history text that represents user-written historical review-summaries. Some of the proposed models use user-specific vocabularies to predict user-preference words that in turn serve as a guide for generating summaries Ma et al. (2018); Li et al. (2019); Chan et al. (2020). Others have proposed models that learn user preference by jointly training these discrete attributes and the historical review-summaries Liu & Wan (2019); Xu et al. (2021; 2023), where the historical summaries provide information about the writing-style, purchasing preferences, and aspect-of-interest of the users. These works, however, do not fit into the general setup of personalized summarization because the summaries generated are not aligned to a prospective buyer s (i.e., a consumer of review-summary or reader in our parlance) preference behavior, but rather tuned to a different set of buyers who are active reviewers and who have provided gold-reference review summaries of their own reviews (i.e., the review-to-summary is a one-to-one mapping and not subjective). Nevertheless, the evaluation has been done using accuracy measures (ROUGE variants) only. So far, the only pertinent work incorporating the reader s history as preference is the proposed models that were designed using the PENS framework (Ao et al., 2021), which we studied extensively (see Section 6.2). It is clear from our study that they need significant improvement in terms of personalization (as measured by Per SEval). 10 Conclusion In this paper, we presented Per SEval, a corrective measure for EGISES (proposed by Vansh et al. (2023)), which, to the best of our knowledge, is the only known personalization measure for summarizers. We first introduced the concept of responsiveness, in contrast to personalization, as a measure to evaluate the capacity of a model to discern the differences in reader profiles (i.e., reading histories) and generate reader-specific summaries that maintain this difference proportionately. We then showed that EGISES measures the former. We thereby proved theoretically and empirically on the real-world PENS dataset that measuring responsiveness does not imply measuring personalization since there can be models that generate distinctly different summaries for different reader profiles (i.e., high responsiveness) but are quite off from the expected summaries (i.e., low accuracy and thereby low user-experience (UX)). We then formulated Per SEval (more specifically, DEGRESS) as a discounted EGISES where the discount factor is a penalty due to accuracy drop called EDP (Effective DEGRESS Penalty Factor). We analyzed the ten SOTA summarization models using seven variants of Per SEval and observed that the model leaderboard reliability depends on the chosen variant. We further observed that the variant Per SEval-Info LM-αβ performs best regarding rank-stability, a meta-evaluation measure we proposed in this paper. We also proposed a novel survey-based meta-evaluation protocol for human-judgment (HJ) to analyze the extent to which human annotators agree with the design principles of Per SEval at a cognitive level. We found that Per SEval-Info LM-αβ has the highest overall HJ-correlation (Pearson s r = 0.79; Spearman s ρ = 0.68; Kendall s τ = 0.47). We finally established that separate accuracy leaderboards for personalized summarizers can be misleading and Per SEval can serve as a unified measure, thereby emphasizing that personalization and accuracy are inseparable aspects of UX. Discussions & Future Directions User guideline: Per SEval as a measure is useful in summarization tasks where the user experience of the utility of the summaries generated by any model will be subjective and will depend upon the user s current preferences (which itself is a reflection of past activities). For example, executive summaries and Minutes-ofmeetings (Mo Ms) need to be personalized (as different participants might have different takeaways/todos from the meeting summary) without compromising on the accuracy of the proceedings. In the same way, news recommendation platforms are also likely to benefit from personalized summarization as it increases the likelihood of the article being clicked and read in full if the title or summary is more personalized as per the user profile. In such use cases, Per SEval should be adopted rather than accuracy measures or P-Accuracy. It is important to note that we do not rule out existing conventional accuracy measures and also P-Accuracy (as proposed in Vansh et al. (2023)) in favor of Per SEval. These measures have their own important role to play for different use cases. Specifically, summarization tasks that do not need personalization should be better evaluated using standard accuracy measures. An example would be the research paper summarization task where saliency, generally, should imply the core contributions of the paper, and hence, the user experience of such summary utility does not depend on specific user preference history. Likewise, there are use cases where P-Accuracy can be more useful, especially when accuracy is the primary objective, yet summarization models to be evaluated also claim to be personalized. An example would be the task of multi-document update summarization for very specific like-minded stakeholders (say, members of a government event-monitoring unit engaging in daily tweets) whose topics of preferences are already known. In this case, P-Accuracy will factor in the lack of personalization (if any) and provide a more realistic accuracy score (and, thereby, the accuracy leaderboard). Future directions: In this work, we analyze the effect of seven variants of Per SEval on SOTA summarization models, out of which the ROUGE-variants, BLEU, and METEOR are defined on the string space, JSD is defined on the probability space, BERTScore is defined on the embedding space, and Info LM is defined on the probability space that is generated from the embedding space using a masked-LM. Although these cover all the most common algebraic spaces on which Per SEval can be defined, it remains to be understood how other alternate measures on the same spaces, such as Bary Score (Colombo et al., 2021), Movers Score (Zhao et al., 2019), Depth Score (Staerman et al., 2022), and other variants of Info LM using multiple Csiszar f-divergences, will behave w.r.t HJ-correlation. Finally, we have only explored one method of estimating Per SEval-HJ (i.e., mimicking the human way of computing Per SEval using RG-L as the distance between model-generated summary and human-reference) for analyzing HJ-correlations. However, there can be other alternative methods of estimating Per SEval-HJ, including incorporating inter-annotator-agreement statistics (such as Kappa statistic). We also consider it a promising future direction to study the effectiveness of Per SEval as an objective function to steer the fine-tuning of the personalized summarization models under a Reward Model-driven reinforcement learning setup. In general, the current work can potentially open up directions of systematic evaluation studies of personalization capabilities in models, particularly LLMs with their In-context Learning capabilities, that apply to all AI challenges where the quality of the model output is subjective to the user s past preferences. We, therefore, encourage the research community to critically study our work and build on it. Ethics Statement We would like to declare that we used the PENS dataset prepared and released by Microsoft Research. Our human-judgment survey was conducted according to the norms set by the Institutional Review Board (IRB) and respects participant anonymity as per guidelines. Xiang Ao, Xiting Wang, Ling Luo, Ying Qiao, Qing He, and Xing Xie. PENS: A dataset and generic framework for personalized news headline generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 82 92. Association for Computational Linguistics, August 2021. doi: 10.18653/v1/2021.acl-long.7. URL https://aclanthology.org/2021.acl-long.7. Satanjeev Banerjee and Alon Lavie. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pp. 65 72, Ann Arbor, Michigan, June 2005. Association for Computational Linguistics. URL https://aclanthology.org/W05-0909. Manik Bhandari, Pranav Narayan Gour, Atabak Ashfaq, Pengfei Liu, and Graham Neubig. Re-evaluating evaluation in text summarization. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 9347 9359. Association for Computational Linguistics, November 2020. doi: 10.18653/v1/2020.emnlp-main.751. URL https://aclanthology.org/2020.emnlp-main.751. Florian Böhm, Yang Gao, Christian M. Meyer, Ori Shapira, Ido Dagan, and Iryna Gurevych. Better rewards yield better summaries: Learning to summarise without references. In Kentaro Inui, Jing Jiang, Vincent Ng, and Xiaojun Wan (eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 3110 3120, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1307. URL https://aclanthology.org/D19-1307. Tanner Bohn and Charles X. Ling. Hone as you read: A practical type of interactive summarization, 2021. URL https://arxiv.org/abs/2105.02923. Florian Boudin, Hugo Mougard, and Benoit Favre. Concept-based summarization using integer linear programming: From concept pruning to multiple optimal solutions. In Conference on Empirical Methods in Natural Language Processing (EMNLP) 2015, 2015. Hou Pong Chan, Wang Chen, and Irwin King. A unified dual-view model for review summarization and sentiment classification with inconsistency loss. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 20, pp. 1191 1200, New York, NY, USA, 2020. Association for Computing Machinery. ISBN 9781450380164. doi: 10.1145/3397271.3401039. URL https://doi.org/10.1145/3397271.3401039. Andrzej Cichocki, Sergio Cruces, and Shun-ichi Amari. Generalized alpha-beta divergences and their application to robust nonnegative matrix factorization. Entropy, 13:134 170, 2011. URL https://api. semanticscholar.org/Corpus ID:670108. Pierre Colombo, Guillaume Staerman, Chloé Clavel, and Pablo Piantanida. Automatic text evaluation through the lens of Wasserstein barycenters. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 10450 10466, 2021. Pierre Colombo, Nathan Noiry, Ekhine Irurozki, and Stéphan Clémençon. What are the best systems? new perspectives on nlp benchmarking. Advances in Neural Information Processing Systems, 35:26915 26932, 2022a. Pierre Jean A Colombo, Chloé Clavel, and Pablo Piantanida. Infolm: A new metric to evaluate summarization & data2text generation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pp. 10554 10562, 2022b. Wade D Cook and Lawrence M Seiford. On the borda-kendall consensus method for priority ranking problems. Management Science, 28(6):621 637, 1982. Mingkai Deng, Bowen Tan, Zhengzhong Liu, Eric Xing, and Zhiting Hu. Compression, transduction, and creation: A unified framework for evaluating natural language generation. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 7580 7605, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.emnlp-main.599. URL https://aclanthology.org/2021.emnlp-main.599. Abdul Ghafoor Etemad, Ali Imam Abidi, and Megha Chhabra. Fine-tuned t5 for abstractive summarization. International Journal of Performability Engineering, 17(10), 2021. Lea Frermann and Alexandre Klementiev. Inducing document structure for aspect-based summarization. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 6263 6273, Florence, Italy, July 2019. Association for Computational Linguistics. doi: 10.18653/v1/P19-1630. URL https://aclanthology.org/P19-1630. Yang Gao, Wei Zhao, and Steffen Eger. SUPERT: Towards new frontiers in unsupervised evaluation metrics for multi-document summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 1347 1354. Association for Computational Linguistics, July 2020. doi: 10.18653/v1/2020.acl-main.124. URL https://aclanthology.org/2020.acl-main.124. Yanjun Gao, Chen Sun, and Rebecca J Passonneau. Automated pyramid summarization evaluation. In Proceedings of the 23rd Conference on Computational Natural Language Learning (Co NLL), 2019. Samira Ghodratnama, Mehrdad Zakershahrak, and Fariborz Sobhanmanesh. Adaptive summaries: A personalized concept-based summarization approach by learning from users feedback. In Service-Oriented Computing ICSOC 2020 Workshops, pp. 281 293, Cham, 2021. Springer International Publishing. ISBN 978-3-030-76352-7. Aniko Hannak, Piotr Sapiezynski, Arash Molavi Kakhki, Balachander Krishnamurthy, David Lazer, Alan Mislove, and Christo Wilson. Measuring personalization of web search. In Proceedings of the 22nd international conference on World Wide Web, pp. 527 538, 2013. Hiroaki Hayashi, Prashant Budania, Peng Wang, Chris Ackerson, Raj Neervannan, and Graham Neubig. Wiki Asp: A Dataset for Multi-domain Aspect-based Summarization. Transactions of the Association for Computational Linguistics, 9:211 225, 03 2021. ISSN 2307-387X. doi: 10.1162/tacl_a_00362. URL https://doi.org/10.1162/tacl_a_00362. Karl Moritz Hermann, Tomáš Kočiský, Edward Grefenstette, Lasse Espeholt, Will Kay, Mustafa Suleyman, and Phil Blunsom. Teaching machines to read and comprehend. In Proceedings of the 28th International Conference on Neural Information Processing Systems - Volume 1, NIPS 15, pp. 1693 1701, Cambridge, MA, USA, 2015. MIT Press. Eran Hirsch, Alon Eirew, Ori Shapira, Avi Caciularu, Arie Cattan, Ori Ernst, Ramakanth Pasunuru, Hadar Ronen, Mohit Bansal, and Ido Dagan. i Facet Sum: Coreference-based interactive faceted summarization for multi-document exploration. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Association for Computational Linguistics, 2021. URL https://aclanthology.org/2021.emnlp-demo.33. Raghav Jain, Vaibhav Mavi, Anubhav Jangra, and Sriparna Saha. Widar-weighted input document augmented rouge. In Advances in Information Retrieval: 44th European Conference on IR Research, ECIR 2022, Stavanger, Norway, April 10 14, 2022, Proceedings, Part I, pp. 304 321. Springer, 2022. Sameer Jain, Vaishakh Keshava, Swarnashree Mysore Sathyendra, Patrick Fernandes, Pengfei Liu, Graham Neubig, and Chunting Zhou. Multi-dimensional evaluation of text summarization with in-context learning. In Findings of the Association for Computational Linguistics: ACL 2023, pp. 8487 8495, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.findings-acl.537. URL https://aclanthology.org/2023.findings-acl.537. Madhusree Kuanr and Puspanjali Mohapatra. Assessment methods for evaluation of recommender systems: A survey. Foundations of Computing and Decision Sciences, 46:393 421, 2021. URL https://api. semanticscholar.org/Corpus ID:245390443. Mike Lewis, Yinhan Liu, Naman Goyal, Marjan Ghazvininejad, Abdelrahman Mohamed, Omer Levy, Veselin Stoyanov, and Luke Zettlemoyer. BART: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7871 7880. Association for Computational Linguistics, July 2020. doi: 10.18653/v1/2020.acl-main.703. URL https://aclanthology.org/2020.acl-main.703. Junjie Li, Haoran Li, and Chengqing Zong. Towards personalized review summarization via user-aware sequence network. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence and Thirty-First Innovative Applications of Artificial Intelligence Conference and Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, AAAI 19/IAAI 19/EAAI 19. AAAI Press, 2019. ISBN 978-157735-809-1. doi: 10.1609/aaai.v33i01.33016690. URL https://doi.org/10.1609/aaai.v33i01.33016690. Shu Li, Yuan Zhao, Longjiang Guo, Meirui Ren, Jin Li, Lichen Zhang, and Keqin Li. Quantification and prediction of engagement: Applied to personalized course recommendation to reduce dropout in moocs. Information Processing & Management, 61(1):103536, 2024. Chin-Yew Lin. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pp. 74 81, Barcelona, Spain, July 2004. Association for Computational Linguistics. URL https://aclanthology.org/W04-1013. Chin-Yew Lin and Franz Josef Och. Automatic evaluation of machine translation quality using longest common subsequence and skip-bigram statistics. In Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics (ACL-04), pp. 605 612, 2004. Hui Liu and Xiaojun Wan. Neural review summarization leveraging user and product information. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, CIKM 19, pp. 2389 2392, New York, NY, USA, 2019. Association for Computing Machinery. ISBN 9781450369763. doi: 10.1145/3357384.3358161. URL https://doi.org/10.1145/3357384.3358161. Yang Liu, Dan Iter, Yichong Xu, Shuohang Wang, Ruochen Xu, and Chenguang Zhu. G-eval: NLG evaluation using gpt-4 with better human alignment. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp. 2511 2522, Singapore, December 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.emnlp-main.153. URL https://aclanthology.org/2023.emnlp-main. 153. Yixin Liu and Pengfei Liu. Sim CLS: A simple framework for contrastive learning of abstractive summarization. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp. 1065 1072. Association for Computational Linguistics, August 2021. doi: 10.18653/v1/2021.acl-short.135. URL https://aclanthology.org/2021.acl-short.135. Yixin Liu, Pengfei Liu, Dragomir Radev, and Graham Neubig. BRIO: Bringing order to abstractive summarization. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2890 2903, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.acl-long.207. URL https://aclanthology.org/2022.acl-long.207. Shuming Ma, Xu Sun, Junyang Lin, and Xuancheng Ren. A hierarchical end-to-end model for jointly improving text summarization and sentiment classification. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence. International Joint Conferences on Artificial Intelligence Organization, 2018. Nicolaas Matthijs and Filip Radlinski. Personalizing web search using long term browsing history. In Proceedings of the fourth ACM international conference on Web search and data mining, pp. 25 34, 2011. Rui Meng, Khushboo Thaker, Lei Zhang, Yue Dong, Xingdi Yuan, Tong Wang, and Daqing He. Bringing structure into summaries: a faceted summarization dataset for long scientific documents. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp. 1080 1089. Association for Computational Linguistics, August 2021. doi: 10.18653/v1/2021.acl-short.137. URL https://aclanthology.org/2021.acl-short.137. M.L. Menéndez, J.A. Pardo, L. Pardo, and M.C. Pardo. The jensen-shannon divergence. Journal of the Franklin Institute, 334(2):307 318, 1997. ISSN 0016-0032. doi: https://doi.org/10.1016/S0016-0032(96) 00063-4. URL https://www.sciencedirect.com/science/article/pii/S0016003296000634. Shashi Narayan, Shay B. Cohen, and Mirella Lapata. Don t give me the details, just the summary! topicaware convolutional neural networks for extreme summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 2018. URL https://aclanthology.org/D18-1206. Duy-Hung Nguyen, Nguyen Viet Dung Nghiem, Bao-Sinh Nguyen, Dung Tien Tien Le, Shahab Sabahi, Minh Tien Nguyen, and Hung Le. Make the most of prior data: A solution for interactive text summarization with preference feedback. In Findings of the Association for Computational Linguistics: NAACL 2022, pp. 1919 1930, Seattle, United States, July 2022. Association for Computational Linguistics. doi: 10.18653/v1/ 2022.findings-naacl.147. URL https://aclanthology.org/2022.findings-naacl.147. Shumpei Okura, Yukihiro Tagami, Shingo Ono, and Akira Tajima. Embedding-based news recommendation for millions of users. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD 17, pp. 1933 1942, New York, NY, USA, 2017. Association for Computing Machinery. ISBN 9781450348874. doi: 10.1145/3097983.3098108. URL https://doi.org/10.1145/3097983. 3098108. Mikhail Orzhenovskii. T5-long-extract at fns-2021 shared task. In Proceedings of the 3rd Financial Narrative Processing Workshop, pp. 67 69, 2021. Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pp. 311 318, 2002. Maxime Peyrard and Judith Eckle-Kohler. Supervised learning of automatic pyramid for optimizationbased multi-document summarization. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1084 1094, Vancouver, Canada, July 2017. Association for Computational Linguistics. doi: 10.18653/v1/P17-1100. URL https://aclanthology.org/ P17-1100. Maxime Peyrard and Iryna Gurevych. Objective function learning to match human judgements for optimization-based summarization. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pp. 654 660, New Orleans, Louisiana, June 2018. Association for Computational Linguistics. doi: 10.18653/v1/N18-2103. URL https://aclanthology.org/N18-2103. Avinesh PVS, Benjamin Hättasch, Orkan Özyurt, Carsten Binnig, and Christian M. Meyer. Sherlock: a system for interactive summarization of large text collections. Proc. VLDB Endow., 11(12):1902 1905, aug 2018. ISSN 2150-8097. doi: 10.14778/3229863.3236220. URL https://doi.org/10.14778/3229863.3236220. Weizhen Qi, Yu Yan, Yeyun Gong, Dayiheng Liu, Nan Duan, Jiusheng Chen, Ruofei Zhang, and Ming Zhou. Prophet Net: Predicting future n-gram for sequence-to-Sequence Pre-training. In Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 2401 2410. Association for Computational Linguistics, November 2020. doi: 10.18653/v1/2020.findings-emnlp.217. URL https://aclanthology.org/ 2020.findings-emnlp.217. GS Ramesh, Vamsi Manyam, Vijoosh Mandula, Pavan Myana, Sathvika Macha, and Suprith Reddy. Abstractive text summarization using t5 architecture. In Proceedings of Second International Conference on Advances in Computer Engineering and Communication Systems: ICACECS 2021, pp. 535 543. Springer, 2022. Chenhui Shen, Liying Cheng, Xuan-Phi Nguyen, Yang You, and Lidong Bing. Large language models are not yet human-level evaluators for abstractive summarization. In Findings of the Association for Computational Linguistics: EMNLP 2023, pp. 4215 4233, Singapore, December 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.findings-emnlp.278. URL https://aclanthology.org/ 2023.findings-emnlp.278. Amir Soleimani, Vassilina Nikoulina, Benoit Favre, and Salah Ait Mokhtar. Zero-shot aspect-based scientific document summarization using self-supervised pre-training. In Proceedings of the 21st Workshop on Biomedical Language Processing, pp. 49 62, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.bionlp-1.5. URL https://aclanthology.org/2022.bionlp-1.5. Guillaume Staerman, Pavlo Mozharovskyi, Pierre Colombo, Stéphan Clémençon, and Florence d Alché Buc. A pseudo-metric between probability distributions based on depth-trimmed regions, 2022. Nisan Stiennon, Long Ouyang, Jeffrey Wu, Daniel Ziegler, Ryan Lowe, Chelsea Voss, Alec Radford, Dario Amodei, and Paul F Christiano. Learning to summarize with human feedback. Advances in Neural Information Processing Systems, 33:3008 3021, 2020. Derek Tam, Anisha Mascarenhas, Shiyue Zhang, Sarah Kwan, Mohit Bansal, and Colin Raffel. Evaluating the factual consistency of large language models through news summarization. In Findings of the Association for Computational Linguistics: ACL 2023, pp. 5220 5255, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.findings-acl.322. URL https://aclanthology.org/ 2023.findings-acl.322. Bowen Tan, Lianhui Qin, Eric Xing, and Zhiting Hu. Summarizing text on any aspects: A knowledge-informed weakly-supervised approach. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 6301 6309. Association for Computational Linguistics, November 2020. doi: 10.18653/v1/2020.emnlp-main.510. URL https://aclanthology.org/2020.emnlp-main.510. T Tawmo, Mrinmoi Bohra, Pankaj Dadure, Partha Pakray, et al. Comparative analysis of t5 model for abstractive text summarization on different datasets. In Proceedings of the International Conference on Innovative Computing Communication (ICICC) 2022. SSRN, 2022. URL https://ssrn.com/abstract= 4096413orhttp://dx.doi.org/10.2139/ssrn.4096413. Maartje Ter Hoeve, Julia Kiseleva, and Maarten Rijke. What makes a good and useful summary? Incorporating users in automatic summarization research. In Marine Carpuat, Marie-Catherine de Marneffe, and Ivan Vladimir Meza Ruiz (eds.), Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 46 75, Seattle, United States, July 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.naacl-main.4. URL https://aclanthology.org/2022.naacl-main.4. Rahul Vansh, Darsh Rank, Sourish Dasgupta, and Tanmoy Chakraborty. Accuracy is not enough: Evaluating personalization in summarizers. In Findings of the Association for Computational Linguistics: EMNLP 2023, pp. 2582 2595, Singapore, December 2023. Association for Computational Linguistics. doi: 10.18653/ v1/2023.findings-emnlp.169. URL https://aclanthology.org/2023.findings-emnlp.169. Michael Völske, Martin Potthast, Shahbaz Syed, and Benno Stein. TL;DR: Mining Reddit to learn automatic summarization. In Lu Wang, Jackie Chi Kit Cheung, Giuseppe Carenini, and Fei Liu (eds.), Proceedings of the Workshop on New Frontiers in Summarization, pp. 59 63, Copenhagen, Denmark, September 2017. Association for Computational Linguistics. doi: 10.18653/v1/W17-4508. URL https://aclanthology. org/W17-4508. Chuhan Wu, Fangzhao Wu, Mingxiao An, Jianqiang Huang, Yongfeng Huang, and Xing Xie. Neural news recommendation with attentive multi-view learning. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence. International Joint Conferences on Artificial Intelligence Organization, 2019a. Chuhan Wu, Fangzhao Wu, Suyu Ge, Tao Qi, Yongfeng Huang, and Xing Xie. Neural news recommendation with multi-head self-attention. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 6389 6394, Hong Kong, China, November 2019b. Association for Computational Linguistics. doi: 10.18653/v1/D19-1671. URL https://aclanthology.org/D19-1671. Hongyan Xu, Hongtao Liu, Pengfei Jiao, and Wenjun Wang. Transformer reasoning network for personalized review summarization. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 21, pp. 1452 1461, New York, NY, USA, 2021. Association for Computing Machinery. ISBN 9781450380379. doi: 10.1145/3404835.3462854. URL https://doi.org/ 10.1145/3404835.3462854. Hongyan Xu, Hongtao Liu, Zhepeng Lv, Qing Yang, and Wenjun Wang. Pre-trained personalized review summarization with effective salience estimation. In Findings of the Association for Computational Linguistics: ACL 2023, pp. 10743 10754, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.findings-acl.684. URL https://aclanthology.org/2023.findings-acl. 684. Rui Yan, Jian-Yun Nie, and Xiaoming Li. Summarize what you are interested in: An optimization framework for interactive personalized summarization. In Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, pp. 1342 1351, Edinburgh, Scotland, UK., July 2011. Association for Computational Linguistics. URL https://aclanthology.org/D11-1124. Weizhe Yuan, Graham Neubig, and Pengfei Liu. Bartscore: Evaluating generated text as text generation. In Advances in Neural Information Processing Systems, volume 34, 2021. URL https://proceedings. neurips.cc/paper/2021/file/e4d2b6e6fdeca3e60e0f1a62fee3d9dd-Paper.pdf. Manzil Zaheer, Guru Guruganesh, Kumar Avinava Dubey, Joshua Ainslie, Chris Alberti, Santiago Ontanon, Philip Pham, Anirudh Ravula, Qifan Wang, Li Yang, and Amr Ahmed. Big bird: Transformers for longer sequences. In Advances in Neural Information Processing Systems, volume 33, pp. 17283 17297, 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/ c8512d142a2d849725f31a9a7a361ab9-Paper.pdf. Eva Zangerle and Christine Bauer. Evaluating recommender systems: Survey and framework. ACM Comput. Surv., 55(8), dec 2022. ISSN 0360-0300. doi: 10.1145/3556536. URL https://doi.org/10.1145/3556536. Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q. Weinberger, and Yoav Artzi. Bertscore: Evaluating text generation with bert. In International Conference on Learning Representations, 2020. URL https: //openreview.net/forum?id=Ske Hu CVFDr. Tianyi Zhang, Faisal Ladhak, Esin Durmus, Percy Liang, Kathleen Mc Keown, and Tatsunori B. Hashimoto. Benchmarking Large Language Models for News Summarization. Transactions of the Association for Computational Linguistics, 12:39 57, 01 2024. ISSN 2307-387X. doi: 10.1162/tacl_a_00632. URL https://doi.org/10.1162/tacl_a_00632. Wei Zhao, Maxime Peyrard, Fei Liu, Yang Gao, Christian M. Meyer, and Steffen Eger. Mover Score: Text generation evaluating with contextualized embeddings and earth mover distance. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 563 578, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1053. URL https: //aclanthology.org/D19-1053. Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, Hao Zhang, Joseph E. Gonzalez, and Ion Stoica. Judging LLM-as-ajudge with MT-bench and chatbot arena. In Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track, 2023a. URL https://openreview.net/forum?id=ucc HPGDlao. Rui Zheng, Shihan Dou, Songyang Gao, Yuan Hua, Wei Shen, Binghai Wang, Yan Liu, Senjie Jin, Qin Liu, Yuhao Zhou, Limao Xiong, Lu Chen, Zhiheng Xi, Nuo Xu, Wenbin Lai, Minghao Zhu, Cheng Chang, Zhangyue Yin, Rongxiang Weng, Wensen Cheng, Haoran Huang, Tianxiang Sun, Hang Yan, Tao Gui, Qi Zhang, Xipeng Qiu, and Xuanjing Huang. Secrets of rlhf in large language models part i: Ppo, 2023b. Ming Zhong, Yang Liu, Da Yin, Yuning Mao, Yizhu Jiao, Pengfei Liu, Chenguang Zhu, Heng Ji, and Jiawei Han. Towards a unified multi-dimensional evaluator for text generation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 2023 2038, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.emnlp-main.131. URL https://aclanthology.org/2022.emnlp-main.131. A Model Details We briefly introduce the SOTA summarization models that were analyzed to understand their degree-ofpersonalization below: 1. PENS-NRMS Injection-Type 1: The PENS framework (Ao et al., 2021) takes user embedding as input along with the news article to generate a personalized summary for that user. To generate user embedding, NRMS (Neural News Recommendation with Multi-Head Self-Attention) (Wu et al., 2019b) is used. It includes a news encoder that utilizes multi-head self-attentions to understand news titles. The user encoder learns user representations based on their browsing history and uses multi-head self-attention to capture connections between news articles. Additive attention is added to learning the news and user representations more effectively by selecting important words and articles. Here, Injection-Type 1 indicates that NRMS user embedding is injected into PENS by initializing the decoder s hidden state of the headline generator, which will influence the summary generation. 2. PENS-NRMS Injection-Type 2: To generate a personalized summary, NRMS user embedding is injected into attention values (Injection-Type 2) of PENS that helps to personalize attentive values of words in the news body. 3. PENS-NAML Injection-Type 1: NAML (Neural News Recommendation with Attentive Multi View Learning) Wu et al. (2019a) incorporates a news encoder that utilizes a multi-view (i.e., titles, bodies, and topic categories) attention model to generate comprehensive news representations. The user encoder is designed to learn user representations based on their interactions with browsed news. It also allows the selection of highly informative news during the user representation learning process. This user embedding is injected into the PENS model using Type-1 for personalization. 4. PENS-EBNR Injection-Type 1: EBNR (Embedding-based News Recommendation for Millions of Users) Okura et al. (2017) proposes a method for user representations by using an RNN model that takes browsing histories as input sequences. This user embedding is injected using Type 1 into the PENS model for personalization. 5. PENS-EBNR Injection-Type 2: This personalized model injects EBNR user embedding into PENS using type-2. 6. BRIO: Instead of a traditional MLE-based training approach, BRIO Liu et al. (2022) assumes a non-deterministic training paradigm that assigns probability mass to different candidate summaries according to their quality, thereby helping it to better distinguish between high-quality and low-quality summaries. 7. Sim CLS: Sim CLS (A Simple Framework for Contrastive Learning of Abstractive Summarization) Liu & Liu (2021) uses a two-stage training procedure. In the first stage, a Seq2Seq model (BART (Lewis et al., 2020)) is trained to generate candidate summaries with MLE loss. Next, the evaluation model, initiated with Ro BERTa is trained to rank the generated candidates with contrastive learning. 8. Big Bird-Pegasus: Big Bird Zaheer et al. (2020) is an extension of Transformer based models designed specifically for processing longer sequences. It utilizes sparse attention, global attention, and random attention mechanisms to approximate full attention. This enables Big Bird to handle longer contexts more efficiently and, therefore, can be suitable for summarization. 9. Prophet Net: Prophet Net Qi et al. (2020) is a sequence-to-sequence pre-trained model that employs n-gram prediction using the n-stream self-attention mechanism. Prophet Net optimizes n-step ahead prediction by simultaneously predicting the next n tokens based on previous context tokens, thus preventing overfitting on local correlations. 10. T5: T5 (Text-To-Text Transfer Transformer) is based on the Transformer-based Encoder-Decoder architecture that operates on the principle of the unified text-to-text task for any NLP problem, including summarization. Some recent analyses on the performance of T5 on summarization tasks can be found in Tawmo et al. (2022); Ramesh et al. (2022); Etemad et al. (2021). B Accuracy and Performance B.1 Accuracy Measures Compared 1. RG-L: ROUGE-L (Recall-Oriented Understudy for Gisting Evaluation) (Lin & Och, 2004) calculates the longest common subsequence between the generated summary and the reference summary and then measures the precision, recall, and F1 score based on this comparison. 2. RG-SU4: ROUGE-SU4 (Lin, 2004) was designed to consider skip-bigram matches as well, which allows for non-contiguous n-gram matches. 3. BLEU: BLEU (Bilingual Evaluation Understudy) (Papineni et al., 2002) is a popular evaluation metric that measures the precision of n-gram matches between the model-generated summaries and the reference summaries. BLEU computes a modified precision score for various n-gram lengths and then combines them using a geometric mean. 4. METEOR: METEOR (Metric for Evaluation of Translation with Explicit ORdering) (Banerjee & Lavie, 2005) matches unigrams based on surface forms, stemmed forms, and meanings and then calculates score using a combination of precision, recall, and the order-alignment of the matched words w.r.t reference summary. 5. Jensen-Shannon Distance: The Jensen-Shannon Distance (JSD) (Menéndez et al., 1997) is a metric used in summarization evaluation to measure the dissimilarity between probability distributions of words in a reference summary and a generated summary. It quantifies the information divergence and similarity, providing a nuanced assessment of the semantic content overlap between the two summaries. 6. Bert Score: Bert Score (BScore) (Zhang et al., 2020) is a metric for evaluating machine-generated summaries, emphasizing contextual embeddings from BERT to assess both word overlap and contextual relationships. It overcomes the limitations of keyphrase-based measures like ROUGE. 7. Info LM-αβ: Given a user-generated reference summary u and a model-generated summary su, Info LM (Colombo et al., 2022b) recursively masks each token position k of both u (denoted [u]k) and su (denoted [su]k) to obtain individual masked contexts of length lu and lsu respectively. For each masked context, it uses a pre-trained masked-language model to estimate the corresponding probability distribution over the vocabulary (i.e., pθ | [ ]k; Mθ,h ), resulting in two bags of distributions of size lu and lsu for u and su. The bags of distributions (for both masked u and masked su) are then averaged out, as follows11: p( | u; Mθ,h) k=1 γk pθ | [u]k; Mθ,h p( | su; Mθ,h) k=1 γk pθ | [su]k; Mθ,h Info LM then uses a chosen information measure I to compute the following: Info LM(u, su) I [p( | u), p( | su)] In our experiments, we chose I to be αβ-divergence (also called AB-Divergence; Dα,β AB) (Cichocki et al., 2011) where the divergence is defined as: Dα,β AB = 1 β(β + α) log X pβ+α i + 1 β + α log X qβ+α i 1 β log X pα i qβ i (8) B.2 Model Rank Aggregation & Agreement 1. Borda-Kendall Consensus based Rank Aggregation: The Borda-Kendall (BK) consensus entails aggregating a set of permutations, denoted as η1, . . . , ηL N, which represent the rankings of N models across L 1 tasks or instances (in our case, the pair of accuracy rank measure and the Per SEval-variant to be aggregated). This aggregation involves summing the ranks of each model and subsequently ranking the obtained sums. Formally: l=1 ηl n for every 1 n N, BK η1, . . . , ηL = argsort (sum1, . . . , sum T ) 2. Pearson s Correlation Coefficient (r): r = Pn i=1(xi x)(yi y) p Pn i=1(xi x)2 Pn i=1(yi y)2 where x, y are the means of the variables xi and yi ; n = the number of samples. 3. Spearman s ρ Coefficient: ρ = 1 6 P d2 i n(n2 1) where d = the pairwise distances of the ranks of the variables xi and yi ; n = the number of samples. 4. Kendall s τ Coefficient: τ = c d c + d = S n 2 = 2S n(n 1) where, c = the number of concordant pairs; d = the number of discordant pairs. 11γk are measures of the importance of the k-th token in u and su, respectively s.t. Plu k=1 γk = Plsu k=1 γk = 1. γk are computed using the corpus-level inverse document frequency (IDF) scores. PENS Test Dataset Sample Set (Random Selection) δ-Bias (in bold) of PSE-variants Models PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Big Bird-Pegasus 0.0009 0.0034 0.001 0.0031 0.0011 0.0015 0.0024 Sim CLS 0.0034 0.007 0.0049 0.0054 0.0019 0.0017 0.0018 BRIO 0.0041 0.0072 0.0055 0.0051 0.001 0.0017 0.0013 Prophet Net 0.0033 0.0059 0.0041 0.0052 0.0002 0.0015 0.0017 T5 (Base) 0.0035 0.0049 0.0047 0.0032 0.0004 0.0016 0.0012 PENS-NAML T1 0.0033 0.0011 0.0027 0.0003 0.0001 0.0016 0.0012 PENS-NRMS T1 0.0029 0.0003 0.0033 0.0008 0.0003 0.0016 0.0006 PENS-EBNR T1 0.0035 0.0004 0.0039 0.0002 0.0001 0.0016 0.0008 PENS-NRMS T2 0.0038 0.0002 0.0042 0.0003 0.0001 0.0016 0.0003 PENS-EBNR T2 0.0036 0.0007 0.0042 0.0002 0 0.0015 0.0004 δ-Variance (in bold) of PSE-variants Models PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ Big Bird-Pegasus 7.74E-07 1.17E-05 1.05E-06 9.43E-06 1.14E-06 2.36E-06 5.81E-06 Sim CLS 1.14E-05 4.86E-05 2.44E-05 2.92E-05 3.64E-06 2.82E-06 3.08E-06 BRIO 1.65E-05 5.25E-05 3.07E-05 2.61E-05 9.86E-07 2.75E-06 1.77E-06 Prophet Net 1.09E-05 3.47E-05 1.69E-05 2.71E-05 3.20E-08 2.39E-06 2.90E-06 T5 (Base) 1.21E-05 2.43E-05 2.17E-05 9.99E-06 1.54E-07 2.61E-06 1.33E-06 PENS-NAML T1 1.12E-05 1.13E-06 7.51E-06 8.96E-08 4.00E-09 2.43E-06 1.41E-06 PENS-NRMS T1 8.65E-06 9.84E-08 1.09E-05 6.14E-07 9.44E-08 2.43E-06 3.26E-07 PENS-EBNR T1 1.22E-05 1.58E-07 1.53E-05 5.20E-08 6.40E-09 2.53E-06 6.00E-07 PENS-NRMS T2 1.41E-05 5.44E-08 1.79E-05 1.02E-07 1.36E-08 2.41E-06 1.14E-07 PENS-EBNR T2 1.30E-05 4.94E-07 1.75E-05 3.76E-08 0 2.31E-06 1.84E-07 Summary Models PSE-RG-L PSE-RG-SU4 PSE-METEOR PSE-BLEU PSE-JSD PSE-BScore PSE-Info LM-αβ δ-stability 0.0041 0.0072 0.0055 0.0054 0.0019 0.0017 0.0024 ϵ-Spearman 1 1 1 1 1 1 1 ϵ-Kendall 1 1 1 1 1 1 1 Table 6: Per SEvalβ=1.7 Stability: 0.0072-strongly-stable w.r.t ϵ-Spearman = 1; ϵ-Kendall = 1 across variants C Detailed Experimental Results We provide a detailed analysis of all the seven Per SEval variants in terms of their stability performance in the following section. C.1 Per SEval Stability Results In this section, we provide a detailed analysis of the stability performance of all the seven Per SEval variants (in Section 8.1, we discussed that of the best performing Per SEval-Info LM variant only). We analyze the δ-bias and the δ-variance of each variant across all the ten SOTA models that have been studied. We observe that while the best-performing variant w.r.t bias is Per SEval-Bert Score and w.r.t variance is Per SEval-RG-L, the worst performances w.r.t both are pretty low with an overall 0.0072 δ-stability across all the variants (see Table 6). We also observed a consistent 100% rank-correlation (i.e., ϵ-stability) across all the variants, showing Per SEvalto be extremely stable. C.2 EGISES and P-Acc Results Models EG-RG-L EG-RG-SU4 EG-METEOR EG-BLEU EG-JSD EG-BScore EG-Info LM-αβ Big Bird-Pegasus 0.324 0.097 0.356 0.128 0.387 0.390 0.33 Sim CLS 0.429 0.268 0.501 0.376 0.512 0.472 0.28 BRIO 0.480 0.282 0.581 0.434 0.630 0.572 0.189 Prophet Net 0.548 0.431 0.602 0.501 0.608 0.577 0.289 T5 (Base) 0.592 0.474 0.642 0.551 0.641 0.619 0.225 PENS-NAML T1 0.869 0.702 0.866 0.798 0.883 0.863 0.575 PENS-NRMS T1 0.891 0.745 0.887 0.832 0.901 0.886 0.654 PENS-EBNR T1 0.935 0.820 0.931 0.892 0.938 0.933 0.793 PENS-EBNR T2 0.986 0.879 0.983 0.958 0.981 0.986 0.907 PENS-NRMS T2 0.988 0.908 0.986 0.965 0.983 0.989 0.824 Table 7: SOTA Responsiveness Benchmarking on Microsoft PENS Dataset w.r.t seven EGISES (EG) variants. Models P-Acc-RG-L P-Acc-RG-SU4 P-Acc-METEOR P-Acc-BLEU P-Acc-JSD P-Acc-BScore P-Acc-Info LM-αβ T5 (Base) 0.492 0.591 0.532 0.594 0.411 0.123 0.392 Prophet Net 0.481 0.580 0.524 0.588 0.393 0.121 0.368 BRIO 0.468 0.582 0.520 0.593 0.333 0.113 0.344 Sim CLS 0.446 0.558 0.500 0.577 0.340 0.107 0.301 Big Bird-Pegasus 0.412 0.517 0.438 0.508 0.304 0.119 0.382 PENS-NAML T1 0.391 0.523 0.404 0.518 0.350 0.106 0.522 PENS-NRMS T1 0.385 0.518 0.398 0.512 0.344 0.104 0.510 PENS-EBNR T1 0.383 0.514 0.398 0.508 0.341 0.103 0.512 PENS-NRMS T2 0.379 0.507 0.394 0.501 0.338 0.101 0.492 PENS-EBNR T2 0.376 0.506 0.390 0.500 0.336 0.102 0.504 Table 8: SOTA Personalized Accuracy Benchmarking on Microsoft PENS Dataset w.r.t P-Accuracy (P-Acc); P-Acc hyper-parameters: α = 0.5, β = 1 (i.e., complete importance to lack of personalization). D Survey Format: Human-Judgment Meta-evaluation of Per SEval In this section, we present the screenshot of the questionnaire designed for the survey for computing the human-judgment version of Per SEval (Per SEval-HJ). Two consecutive respondents evaluated the generated summary pairs of all ten benchmarked models. Figure 4: Sample Questionnaire: Six pairs of summaries for a specific document; five pairs are modelgenerated summaries (each user evaluates five of the ten models) for a specific document, while one pair is user-generated gold reference).