# towards_trustable_shap_scores__a76d43d9.pdf Towards Trustable SHAP Scores Olivier Létoffé1, Xuanxiang Huang2, Joao Marques-Silva3 1Univ. Toulouse, France 2CNRS@CREATE, Singapore 3ICREA, Univ. Lleida, Spain olivier.letoffe@orange.fr, xuanxiang.huang@cnrsatcreate.sg, jpms@icrea.cat SHAP scores represent the proposed use of the well-known Shapley values in e Xplainable Artificial Intelligence (XAI). Recent work has shown that the exact computation of SHAP scores can produce unsatisfactory results. Concretely, for some ML models, SHAP scores will mislead with respect to relative feature influence. To address these limitations, recently proposed alternatives exploit different axiomatic aggregations, all of which are defined in terms of abductive explanations. However, the proposed axiomatic aggregations are not Shapley values. This paper investigates how SHAP scores can be modified so as to extend axiomatic aggregations to the case of Shapley values in XAI. More importantly, the proposed new definition of SHAP scores avoids all the known cases where unsatisfactory results have been identified. The paper also characterizes the complexity of computing the novel definition of SHAP scores, highlighting families of classifiers for which computing these scores is tractable. Furthermore, the paper proposes modifications to the existing implementations of SHAP scores. These modifications eliminate some of the known limitations of SHAP scores, and have negligible impact in terms of performance. Introduction Shapley values for e Xplainable AI (XAI), i.e. SHAP scores (Lundberg and Lee 2017), are arguably among the most widely used explainability methods that target the attribution of (relative) feature importance, as exemplified by the success of the tool SHAP (Molnar 2023; Mishra 2023).1 Despite the massive popularity of SHAP scores, some works have identified limitations with their use (Young et al. 2019; Kumar et al. 2020; Sundararajan and Najmi 2020; Merrick and Taly 2020; Fryer, Strümke, and Nguyen 2021; Yan and Procaccia 2021; Mothilal et al. 2021; Afchar, Guigue, and Hennequin 2021; Watson et al. 2021; Kumar et al. 2021; Campbell et al. 2022). However, most of these limitations can be attributed to the results obtained with existing tools, and not necessarily with the theoretical foundations of SHAP scores. More recent work (Huang and Marques-Silva 2023; Huang and Marques-Silva 2024) uncovered examples of classifiers where exact SHAP scores assign manifestly unsatisfactory importance to features. Namely, features having no Copyright 2025, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 1See https://github.com/shap/shap. influence in a prediction can be assigned more importance than features having the most influence in the prediction. This recent evidence should be perceived as far more problematic, because it reveals apparent limitations with the theoretical foundations of SHAP scores, and not with concrete implementations. Nevertheless, Shapley values are of fundamental importance, not only in game theory (Chalkiadakis, Elkind, and Wooldridge 2012), but also in many other domains, namely because of their intrinsic properties (Shapley 1953). As a result, a natural question is whether the definitions of Shapley values in XAI can be changed, so as to avoid situations where the computed feature importance is problematic. Contributions. This paper argues that the key issue with SHAP scores is not the use of Shapley values in explainability per se, and shows that the identified shortcomings of SHAP scores can be solely attributed to the characteristic functions used in earlier works (Strumbelj and Kononenko 2010, 2014; Lundberg and Lee 2017; Janzing, Minorics, and Blöbaum 2020; Sundararajan and Najmi 2020; Arenas et al. 2021; Van den Broeck et al. 2021, 2022; Arenas et al. 2023). As noted in the recent past (Janzing, Minorics, and Blöbaum 2020; Sundararajan and Najmi 2020), by changing the characteristic function, one is able to produce different sets of SHAP scores.2 Motivated by these observations, the paper outlines fundamental properties that characteristic functions ought to exhibit in the context of XAI. Furthermore, the paper proposes several novel characteristic functions, which either respect some or all of the identified properties. In addition, the paper analyzes the impact of the novel characteristic functions on the computational complexity of computing SHAP scores, by building on recent work on the same topic (Van den Broeck et al. 2021; Arenas et al. 2021; Van den Broeck et al. 2022; Arenas et al. 2023). An indirect consequence of our work is that corrected SHAP scores can be safely used for feature attribution in XAI, while offering strong guarantees regarding known shortcomings. Furthermore, the results in the paper apply independently of the machine learning (ML) model considered. 2Unfortunately, this paper argues that past alternative proposals of characteristic functions (Janzing, Minorics, and Blöbaum 2020; Sundararajan and Najmi 2020) also exhibit critical limitations. The Thirty-Ninth AAAI Conference on Artificial Intelligence (AAAI-25) Related work. SHAP scores are ubiquitously used in XAI (Lundberg and Lee 2017; Molnar 2023; Mishra 2023). Recent work argues that the existing definitions of (exact) SHAP scores can yield unsatisfactory results (Huang and Marques-Silva 2023; Huang and Marques-Silva 2024). Motivated by these results, different works proposed alternative solutions to the use of SHAP scores (Yu, Ignatiev, and Stuckey 2023; Biradar et al. 2023; Yu et al. 2023; Biradar et al. 2024; Yu et al. 2024). Furthermore, one of these solutions (Biradar et al. 2024) investigates the use of power indices from the field of a priori voting power (Felsenthal and Machover 2015) as a solution for feature importance in XAI, covering several well-known power-indices. However, Shapley values are also at the core of the well-known Shapley-Shubik power index (Shapley and Shubik 1954), one of the best-known power indices, and which is not studied in (Biradar et al. 2024). Thus, an open question is how to extend the recent work on power indices for XAI (Biradar et al. 2024) to the case of the Shapley-Shubik index. Organization. The paper starts by introducing the notation and definitions used throughout the paper. Afterwards, the paper briefly dissects some of the recently reported shortcomings with SHAP scores (Marques-Silva and Huang 2024; Huang and Marques-Silva 2024; Huang and Marques-Silva 2023). Motivated by those shortcomings, the paper then proposes properties that characteristic functions should exhibit. The paper then proposes several novel characteristic functions, which are shown to correct some or all of the shortcomings of the characteristic functions used in earlier work. Next, the paper studies the complexity of computing SHAP scores given the novel characteristic functions proposed in this paper. The paper also outlines a simple modification to the SHAP tool (Lundberg and Lee 2017), which corrects some of the shortcomings of SHAP scores. Preliminaries Classification & regression problems. Let F = {1, . . . , m} denote a set of features. Each feature i F takes values from a domain Di. Domains can be categorical or ordinal. If ordinal, domains can be discrete or real-valued. Feature space is defined by F = D1 D2 . . . Dm. Throughout the paper domains are assumed to be discrete-valued.3 The notation x = (x1, . . . , xm) denotes an arbitrary point in feature space, where each xi is a variable taking values from Di. Moreover, the notation v = (v1, . . . , vm) represents a specific point in feature space, where each vi is a constant representing one concrete value from Di. A classifier maps each point in feature space to a class taken from K = {c1, c2, . . . , c K}. Classes can also be categorical or ordinal. However, and unless otherwise stated, classes are assumed to be ordinal. In the case of regression, each point in feature space is mapped to an ordinal value taken from a set C, e.g. C could denote Z or R. Therefore, a classifier 3The results in the paper can be generalized to continuous-valued features. For real-valued features, the only changes involve the definition of expected value and probability, respectively in (2) and (3). MC is characterized by a non-constant classification function κ that maps feature space F into the set of classes K, i.e. κ : F K. A regression model MR is characterized by a non-constant regression function ρ that maps feature space F into the set elements from C, i.e. ρ : F C. A classifier model MC is represented by a tuple (F, F, K, κ), whereas a regression model MR is represented by a tuple (F, F, C, ρ).4 When viable, we will represent an ML model M by a tuple (F, F, T, τ), with τ : F T, without specifying whether whether M denotes a classification or a regression model. A sample (or instance) denotes a pair (v, q), where v F and either q K, with q = κ(v), or q C, with q = ρ(v). Additional notation. An explanation problem is a tuple E = (M, (v, q)), where M can either be a classification or a regression model, and (v, q) is a target sample, with v F. (Observe that q = κ(v), with q K, in the case of a classification model, and q = ρ(v), with q C, in the case of a regression model.) Given x, v F, and S F, the predicate x S = v S is defined as follows: x S = v S := i S xi = vi The set of points for which x S = v S is defined by Υ(S; v) = {x F | x S = v S}. Distributions, expected value. Throughout the paper, it is assumed a uniform probability distribution on each feature, and such that all features are independent. Thus, the probability of an arbitrary point in feature space becomes: P(x) := 1/Πi F|Di| (1) That is, every point in the feature space has the same probability. The expected value of an ML model τ : F T is denoted by E[τ]. Furthermore, let E[τ(x) | x S = v S] represent the expected value of τ over points in feature space consistent with the coordinates of v dictated by S, which is defined as follows: E[τ(x) | x S = v S] := 1/|Υ(S; v)| X x Υ(S;v) τ(x) (2) Similarly, we define, P(π(x) | x S = v S) := 1/|Υ(S; v)| X x Υ(S;v) ITE(π(x), 1, 0) (3) where π : F {0, 1} is some predicate. Shapley values. Shapley values were proposed in the context of game theory in the early 1950s by L. S. Shapley (Shapley 1953). Shapley values were defined given some set N, and a characteristic function, i.e. a real-valued function defined on the subsets of N, υ : 2N R.5. It is well-known 4As shown in the extended version of this paper (Letoffe, Huang, and Marques-Silva 2024a), other ML models can also be represented with similar ideas. This is the case with the use of a softmax layer in neural networks (NNs). 5The original formulation also required super-additivity of the characteristic function, but that condition has been relaxed in more recent works (Dubey 1975; Young 1985). that Shapley values represent the unique function that, given N and υ, respects a number of important axioms. More detail about Shapley values is available in standard references (Shapley 1953; Dubey 1975; Young 1985; Roth 1988). Besides the recent uses in XAI, Shapley values have been used for assigning measures of relative importance in computational social choice (Chalkiadakis, Elkind, and Wooldridge 2012) (including a priori voting power (Shapley and Shubik 1954; Felsenthal and Machover 1998)), measurement of inconsistency in knowledge bases (Hunter and Konieczny 2006, 2010), and intensity of attacks in argumentation frameworks (Amgoud, Ben-Naim, and Vesic 2017). SHAP scores. In the context of explainability, Shapley values are most often referred to as SHAP scores (Strumbelj and Kononenko 2010, 2014; Lundberg and Lee 2017; Arenas et al. 2021, 2023), and consider a specific characteristic function υe : 2F R, which is defined by, υe(S; E) := E[τ(x) | x S = v S] (4) Thus, given a set S of features, υe(S; E) represents the expected value of the classifier over the points of feature space represented by Υ(S; v). The formulation presented in earlier work (Arenas et al. 2021, 2023) allows for different input distributions when computing the average values. For the purposes of this paper, it suffices to consider solely a uniform input distribution, and so the dependency on the input distribution is not accounted for. Independently of the distribution considered, it should be clear that in most cases υe( ) = 0; this is the case for example with boolean classifiers (Arenas et al. 2021, 2023). To simplify the notation, the following definitions are used, i(S; E, υ) := (υ(S {i}) υ(S)) (5) ς(|S|) := |S|!(|F| |S| 1)!/|F|! (6) (Observe that i is parameterized on E and υ.) Finally, let Sc E : F R, i.e. the SHAP score for feature i, be defined by,6. Sc E(i; E, υe) := X S (F\{i}) ς(|S|) i(S; E, υe) (7) Given a sample (v, q), the SHAP score assigned to each feature measures the contribution of that feature with respect to the prediction. From earlier work, it is understood that a positive/negative value indicates that the feature can contribute to changing the prediction, whereas a value of 0 indicates no contribution (Strumbelj and Kononenko 2010). Similarity predicate. Given an ML model and some input x, the output of the ML model is distinguishable with respect to the sample (v, q) if the observed change in the model s output is deemed sufficient; otherwise it is similar (or indistinguishable). This is represented by a similarity predicate (which will also be viewed as a boolean function) σ : F { , } (where signifies false, and signifies 6Throughout the paper, the definitions of i and Sc are explicitly associated with the characteristic function used in their definition. true).7 Concretely, σ(x; E) holds true iff the change in the ML model output is deemed insufficient and so no observable difference exists between the ML model s output for x and v.8 For regression problems, we write instead σ as the instantiation of a template predicate, i.e. σ(x; E) = Tσ(x; E, δ), where δ is an optional measure of output change, which can be set to 0.9 Given a change in the input from v to x, a change in the output is indistinguishable (i.e. the outputs are similar) if, σ(x; E) := Tσ(x; E, δ) := [|ρ(x) ρ(v)| δ] otherwise, it is distinguishable. For classification problems, similarity is defined to equate with not changing the predicted class. Given a change in the input from v to x, a change in the output is indistinguishable (i.e. the outputs are similar) if, σ(x; E) := [κ(x) = κ(v)] otherwise, it is distinguishable. (As shown in the remainder of this paper, σ allows abstracting away whether the underlying model implements classification or regression.) It will be helpful to list a few properties of σ. Observe that (A F).[E[σ(x; E) | x A] [0, 1]]. It is also plain to conclude that for A, B F, with A B, and given u {0, 1}, if E[σ(x; E) | x B] = u, then E[σ(x; E) | x A] = u, A few more properties of σ are apparent. For A F, u {0, 1}, (E[σ(x; E) | x A] = u) (x A).[σ(x; E) = u]. As a result, it is also the case that (E[σ(x; E) | x A] < 1) (x A).[σ(x; E) = 0]. Adversarial examples. Adversarial examples serve to reveal the brittleness of ML models (Szegedy et al. 2014; Goodfellow, Shlens, and Szegedy 2015). Adversarial robustness indicates the absence of adversarial examples. The importance of deciding adversarial robustness is illustrated by a wealth of competing alternatives (Brix et al. 2023). Given a sample (v, q), and a norm lp, a point x F is an adversarial example if the prediction for x is distinguishable from that for v. Formally, we write, AEx(x; E) := (||x v||p ϵ) σ(x; E) where the lp distance between the given point v and other points of interest is restricted to ϵ > 0. Moreover, we define a constrained adversarial example, such that the allowed set of points is given by the predicate x S = v S. Thus, AEx(x, S; E) := (||x v||p ϵ) (x S = v S) σ(x; E) Adversarial robustness is concerned with determining whether complex ML models do not exhibit adversarial examples for chosen samples. 7For simplicity, and with a minor abuse of notation, when σ is used in a scalar context, it is interpreted as a boolean function, i.e. σ : F {0, 1}, with 0 replacing and 1 replacing . 8Throughout the paper, parameterization are shown after the separator ; , and will be elided when clear from the context. 9Exploiting a threshold to decide whether there exists an observable change has been used in the context of adversarial robustness (Wu, Wu, and Barrett 2023). Furthermore, the relationship between adversarial examples and explanations is well-known (Ignatiev, Narodytska, and Marques-Silva 2019b; Wu, Wu, and Barrett 2023). Abductive and contrastive explanations (AXps/CXps). AXps and CXps are examples of formal explanations for classification problems (Ignatiev, Narodytska, and Marques Silva 2019a; Marques-Silva and Ignatiev 2022; Marques Silva 2022; Darwiche 2023). We adopt a generalization that encompasses regression problems (Marques-Silva 2024). (Although we define abductive/contrastive explanations in terms of probabilities, these can be rewritten using expected values. In addition, the proposed definitions are equivalent to the logic-based formulations proposed in other works (Marques Silva 2022).) A weak abductive explanation (WAXp) denotes a set of features S F, such that for every point in feature space the ML model output is similar to the given sample: (v, q). The condition for a set of features to represent a WAXp (which also defines a corresponding predicate WAXp) is as follows: WAXp(S; E) := P(σ(x; E) | x S = v S) = 1 Moreover, an AXp is a subset-minimal WAXp. A weak contrastive explanation (WCXp) denotes a set of features S F, such that there exists some point in feature space, where only the features in S are allowed to change, that makes the ML model output distinguishable from the given sample (v, q). The condition for a set of features to represent a WCXp (which also defines a corresponding predicate WCXp) is as follows: WCXp(S; E) := P(σ(x; E) | x F\S = v F\S) < 1 Moreover, a CXp is a subset-minimal WCXp. One immediate observation is that each WAXp is a hitting set (HS) of the set of WCXps, and each WCXp is a HS of the set of WAXps. Furthermore, one can prove that a set of features is an AXp iff it is a minimal hitting set (MHS) of the set of CXps, and vice-versa (Marques-Silva and Ignatiev 2022). (Although this result has been proved for classification problems, our proposed framework generalizes the result also to regression problems.) Given the previous definitions, it is plain the following result. Proposition 1. (x F). (AEx(x, F \ S; E)) iff WCXp(S; E), i.e. there exists a constrained adversarial example with the features F \ S iff the set S is a weak CXp. Feature (ir)relevancy. The set of features that are included in at least one (abductive) explanation is defined as follows: F(E) := {i X | X 2F AXp(X)} (8) where predicate AXp(X) holds true iff X is an AXp. (A wellknown result is that F(E) remains unchanged if CXps are used instead of AXps (Marques-Silva and Ignatiev 2022), in which case predicate CXp(X) holds true iff X is a CXp.) Finally, a feature i F is irrelevant, i.e. predicate Irrelevant(i) holds true, if i F(E); otherwise feature i is relevant, and predicate Relevant(i) holds true. Clearly, given some explanation problem E, (i F).Irrelevant(i) Relevant(i). Unsatisfactory SHAP Scores Recent work (Huang and Marques-Silva 2023; Marques Silva and Huang 2024; Huang and Marques-Silva 2024) reports examples of classifiers for which the SHAP scores are patently unsatisfactory. This section shows that similar unsatisfactory scores can be obtained with regression models.10 Case study an example regression model. Figure 1 shows a regression tree (RT) (Breiman et al. 1984) used as the running example for the remainder of the paper. (The RT is adapted from (James et al. 2017, Fig 8.1), with the attribute names simplified, and with the predicted values changed.) It is plain to conclude that F = {1, 2}, D1 = D2 = {0, 1}, F = {0, 1}2, C = R, and that the regression function is given either by the tabular representation (TR) or by the regression table (RT). In addition, the target instance is ((1, 1), 1). Figure 1(c) shows the values of υe for the possible values of set S. Furthermore, in the remainder of the paper, it is assumed that the value used to instantiate σ for this regression problem is δ = 0.5. Adversarial examples for the case study. As can be observed in Figure 2(a), for an input to result in a distinguishable output, feature 1 must be changed. In contrast, feature 2 need not be changed. Any subset-minimal set of features that must be changed to make the output distinguishable only includes feature 1. Explanations for the case study. The computation of abductive explanations is summarized in Figure 2(b). As can be observed, the set of AXps contains only {1}. Hence, by minimal-hitting set duality, the set of CXps is also {{1}}. The conclusions are that: i) if feature 1 is fixed, then the prediction cannot be changed; and ii) if the prediction is to be changed, then feature 1 must be changed. These conclusions are aligned with the conclusions obtained from analyzing the adversarial examples. SHAP scores for the case study. The computation of exact SHAP scores is summarized Figure 2(c), using the definitions introduced earlier in the paper. As can be observed, the SHAP score for feature 1 is 0, and the SHAP score for feature 2 is 1/4. Analysis. For the running example it is clear that reporting a non-zero SHAP score for feature 2 and a zero-value SHAP score for feature 1 is not only non-intuitive, but it also disagrees with the information provided both by the adversarial examples and the abductive explanations. Motivated by these situations where exact SHAP scores are manifestly unsatisfactory, there has been recent work on proposing alternatives to SHAP scores (Biradar et al. 2024; Yu et al. 2024). These recent alternatives to SHAP scores are based on power indices, studied for assessing voting power. However, there exists a power index that is based on Shapley values, namely the Shapley-Shubik index, which is not studied in (Biradar et al. 2024). This paper investigates how to change the definition of SHAP scores such that: i) the unsatisfactory results 10Additional evidence regarding the flaws of SHAP scores is available elsewhere (Letoffe, Huang, and Marques-Silva 2024b). row x1 x2 ρ(x) 1 0 0 1/2 2 0 1 3/2 3 1 0 1 4 1 1 1 (a) Tabular representation (TR) (b) Regression tree (RT) S rows(S) υe(S) 1, 2, 3, 4 3/4 {1} 3, 4 1 {2} 2, 4 5/4 {1, 2} 4 1 (c) Expected values Figure 1: Simple regression tree model, adapted from (James et al. 2017, Fig. 8.1). The target instance is ((1, 1), 1). x1 x2 AEx(x) || ||0 0 0 Y 2 0 1 Y 1 1 0 N 1 1 1 N 0 (a) Adversarial examples S WAXp(S) AXp(S) N {1} Y Y {2} N {1, 2} Y N (b) Abductive explanations S υe(S) υe(S {1}) 1(S) ς(S) ς(S) 1(S) 3/4 1 1/4 1/2 1/8 {2} 5/4 1 1/4 1/2 1/8 Sc E(1) = 0 S υe(S) υe(S {2}) 2(S) ς(S) ς(S) 2(S) 3/4 5/4 1/2 1/2 1/4 {1} 1 1 0 1/2 0 Sc E(2) = 1/4 (c) SHAP scores Figure 2: AExs, AXPs & SHAP scores for the regression tree from Figure 1 and target sample ((1, 1), 1). For simplicity, parameterizations are elided. of existing SHAP scores are eliminated; and ii) an instantiation of the Shapley-Shubik index is obtained. The envisioned approach is to maintain the definition of SHAP scores in terms of Shapley values, but change the characteristic function that has been widely used for defining SHAP scores, i.e. υe.11 The next section investigates several novel characteristic functions, all of which are based on the similarity predicate introduced earlier in the paper. Properties of Characteristic Functions Given the issues reported earlier in the paper, this section proposes properties that characteristic functions should respect. If characteristic functions fail to respect some of these properties, then the resulting SHAP scores can provide misleading information about relative feature importance. Weak value independence. Let M1 = (F, F, T1, τ1) be an ML model, with domain Di for each feature i F. Moreover, let M2 = (F, F, T2, τ2) be another ML model, with the same domains, and with |T1| = |T2|. In addition, let µ : T1 T2 be a surjective mapping from T1 to T2, such that for any x F, τ2(x) = µ(τ1(x)). Finally, let the target samples be (v, q), for M1, and (v, µ(q)) for M2, thus defining the explanation problems E1 = (M1, (v, q)) 11The Appendix shows that other well-known alternatives (Sundararajan and Najmi 2020) also yield unsatisfactory results. and E2 = (M2, (v, µ(q))). A characteristic function υt is weakly value-independent if, given surjective µ, (i F).[Sct(i; E1) = Sct(i; E2)] Strong value independence. Let M1 = (F, F, T1, τ1) be an ML model, with domain Di for each feature i F. Moreover, let M2 = (F, F, T2, τ2) be another classifier, with the same domains. In addition, let µ : K1 K2 be a mapping from T1 to T2, such that for q T1, and such that, (b T1).[(b = q) (µ(b) = µ(q))] Finally, let the target samples be (v, q), for M1, and (v, µ(q)) for M2, thus defining the explanation problems E1 = (M1, (v, q)) and E2 = (M2, (v, µ(q))). A characteristic function υt is strongly value-independent if, given µ, (i F).[Sct(i; E1) = Sct(i; E2)] Given the above, the following result holds12, Proposition 2. If a characteristic function is strongly valueindependent, then it is weakly value-independent. Compliance with feature (ir)relevancy. Characteristic functions should respect feature (ir)relevancy, i.e. a feature is irrelevant iff its (corrected) SHAP score is 0. Formally, a characteristic function υt is compliant with feature (ir)relevancy if, (i F).Irrelevant(i) (Sct(i) = 0) (9) In previous work (Huang and Marques-Silva 2023; Huang and Marques-Silva 2024; Marques-Silva and Huang 2024), 12All the proofs are included in the Appendix. SHAP scores are said to be misleading when compliance with feature (ir)relevancy is not respected. In the remainder of the paper, we assign the same meaning to the term misleading. Numerical neutrality. Existing definitions of SHAP scores are based on expected values and so require T (i.e. K or C) to be ordinal. However, classification problems often contemplate categorical classes. A characteristic function respects numerical neutrality if it can be used with both numerical and non-numerical T. Discussion. The properties proposed in this section target the issues reported in earlier work (Marques-Silva and Huang 2024; Huang and Marques-Silva 2024), where SHAP scores mislead with respect to relative feature importance. Additional properties might be devised to address other hypothetical issues. New Characteristic Functions This section proposes novel characteristic functions,13 most of which respect all of the target properties identified in the previous section. For each characteristic function υt, and given a fixed explanation problem E, the obtained SHAP scores will be unique. Some of these SHAP scores will not respect the properties proposed earlier in the paper, e.g. this is the case with υe, whereas some of the novel SHAP scores will respect all of those properties, in addition to the axioms proved by Shapley (Shapley 1953). Throughout this section, an explanation problem E = (M, (v, q)) is assumed, and it is used to parameterize the proposed characteristic functions. Defining the new characteristic functions. Given the definition of the similarity predicate, we now introduce the following main new characteristic functions. υs(S; E) := E[σ(x) | x S = v S] (10) υa(S; E) := 1 if υs(S; E) = 1 0 otherwise (11) υc(S; E) := 1 if υs(F \ S; E) < 1 0 otherwise (12) We will refer to characteristic functions υe (see (4)), υs, υa, υc, respectively as the expected value, the similarity, the WAXp-based, and the WCXp-based characteristic functions. Furthermore, we will introduce another characteristic function, which is shown to be tightly related with υa. υn(S; E) := 1 if υs(S; E) < 1 0 otherwise (13) (Observe that υn can be viewed as the complement of υa.) It is plain that for each characteristic υt, with t {a, c, n, s}, one can create a corresponding Shapley value Sc T , with T {A, C, N, S}. It suffices to replace the characteristic function υe used in (7), by υt, for the chosen t {a, c, n, s}. 13In the rest of the paper, the symbol υt will be used to denote some concrete characteristic function distinguished by the letter t. The SHAP scores obtained with such characteristic function υt will be denoted by Sc T (i.e. for scores we capitalize the corresponding letter, and so the parameterization is Sc T (S; E, υt)). Basic attributes of the new characteristic functions. We start by deriving some basic results regarding the characteristic functions υa, υc and υn. Throughout, it is assumed an explanation problem E. Proposition 3. Given the definition of υa, υc and υn, then Sc A(i; E, υa) 0, Sc C(i; E, υc) 0, and Sc N(i; E, υN) 0. Proposition 4. The following holds true: 1. (S F).[υa(S; E) = 1 WAXp(S; E)]. 2. (S F).[υn(S; E) = 1 WCXp(F \ S; E)]. 3. (S F).[υc(S; E) = 1 WCXp(S; E)]. Proposition 5. The following holds true: 1. (i F).[Sc A(i; E, υa) = Sc N(i; E, υn)]. 2. (i F).[Sc A(i; E, υa) = Sc C(i; E, υc)]. An immediate consequence of the results in Propositions 4 and 5, is that the complexity of computing SHAP scores Sc T is the same for T {A, C, N}. Furthermore, there are additional consequences to Proposition 5. Observe that υa and υc are defined in terms of predicates that are related by a hitting set relationship, i.e. as noted earlier in the paper, each WAXp is a hitting set of the set of WCXps, and each WCXp is a hitting set of the set of WAXps. We call such property hitting set duality. Thus, from Proposition 5, we obtain a stronger result. Consider two predicates Pβ and Pδ, mapping the powerset of F, i.e. 2F, to { , }, and such that Pβ and Pδ exhibit hitting set duality. Define υb such that υb(S) = 1 iff Pβ(S), and υd(S) = 1 iff Pδ(S). From υb and υd, we obtain the Shapley values Sc B and Sc D, respectively. As a result, Theorem 1. Given the definitions of υb and υd, it is the case that: (i F).Sc B(i) = Sc D(i). It is interesting to observe that Theorem 1 is a new result that finds application beyond XAI, in any practical uses of Shapley values, including those mentioned earlier in the paper. Properties of the new characteristic functions. We now assess which of the properties of characteristic functions proposed earlier are respected by which characteristic functions among those proposed in this section. It is plain that characteristic functions based on the similarity predicate respect numerical neutrality. Furthermore, another general result is that characteristic functions based on the similarity predicate guarantee strong (and so weak) value independence. Proposition 6. For t {s, a, c, n} and i F, it is the case that the characteristic function υt respects strong value independence. Proposition 7. For T {A, C, N}, then it is the case that (i F).Irrelevant(i) (Sc T (i) = 0). Finally, we observe that υs represents a boolean classifier, and so it exhibits the issues with SHAP scores uncovered for boolean classifiers based on υe in earlier work (Huang and Marques-Silva 2023; Huang and Marques-Silva 2024). Complexity of Computing SHAP Scores The previous section introduced novel characteristic functions that exhibit a number of desirable properties, which in turn ensure that SHAP scores will not produce misleading information (see Proposition 7). Another related question is how the novel characteristic functions impact the computational complexity of computing SHAP scores. This section starts the effort of mapping such computational complexity. Intractable cases. A number of intractability results have been obtained in recent years (Van den Broeck et al. 2021, 2022). As noted earlier in the paper, for boolean functions, the similarity predicate does not provide any difference with respect to the original classifier. The following result is clear. Proposition 8. For a boolean classifier, with κ(v) = 1, then (x F). (σ(x; E) = κ(x)). From Proposition 8 and Corollary 8 in (Van den Broeck et al. 2022), it is immediate that, Proposition 9. Computing SHAP scores Sc S is #P-hard for boolean classifiers in CNF or DNF. Clearly, given Proposition 9, then the computation of SHAP scores for more complex boolean classifiers is also #P-hard. Moreover, a key recent result regarding the computation of SHAP scores is that for the characteristic function υe there are polynomial-time algorithms for computing Sc E (Arenas et al. 2021, 2023). In contrast, for characteristic functions that build on WAXps/WCXps, the computation of SHAP scores becomes NP-hard, even for d-DNNF and DDBC classifiers.14 Proposition 10. For T {A, C, N}, the computation of the SHAP scores Sc T is NP-hard for d-DNNF & DDBC classifiers. Polynomial-time cases. As shown above, the most significant tractability result that is known for υe does not hold for υt, with t {a, c, n}. Nevertheless, some tractability results can be proved. For ML models represented by tabular representations (e.g. truth tables), it is simple to devise algorithms polynomial on the size of the classifier s representation (Huang and Marques-Silva 2023). As a result, it is the case that, Proposition 11. There exist polynomial-time algorithms for computing the SHAP scores Sc S, Sc A, Sc C for ML models represented by tabular representations. Given the recent results on the tractability of computing SHAP scores for deterministic and decomposable circuits (d-DNNFs) (Arenas et al. 2021, 2023), that consider boolean classifiers, then from Proposition 8 and (Arenas et al. 2023), it is the case that, Proposition 12. The computation of SHAP scores Sc S is in P for classifiers represented by non-boolean DDBCs. 14Deterministic Decomposable Boolean Circuits (DDBCs) and deterministic Decomposable Negation Normal (d-DNNF) form circuits denote well-known restrictions of boolean circuits, and are briefly overviewed in the supplemental materials. Similarity-Based SHAP This section outlines a first step towards addressing the issues with SHAP scores reported in earlier work, and observed in the tool SHAP (Lundberg and Lee 2017). Instead of running SHAP with the original training data and the original classifier, the similarity-based SHAP (referred to as s SHAP) replaces the original classifier by the similarity predicate, and reorganizes training data accordingly. In terms of running time complexity, the impact of the modifications to SHAP are negligible. More importantly, s SHAP will be approximating Sc S, since the underlying characteristic function is υs. In practice, s SHAP is built on top of the SHAP tool (Lundberg and Lee 2017). As noted earlier in the paper, the use of υs does not guarantee the non-existence of some of the issues reported in earlier work (Huang and Marques-Silva 2024), since it is known that even boolean classifiers can exhibit a number of issues related with the relative order of feature importance. Nevertheless, another question is whether υs can serve to correct SHAP scores (obtained with υe) in classifiers for which the reported issues rely on non-boolean classification. Difference in SHAP scores for example classifiers. To validate the improvements obtained with υs with respect to υe, we studied the non-boolean classifiers reported in (Huang and Marques-Silva 2024)15. For each classifier, each of the possible instances is analyzed, and the SHAP scores produced by the tools SHAP and s SHAP are recorded. If an irrelevant feature is assigned an absolute value larger than some other relevant feature, then a mismatch is declared. Table 1 summarizes the results obtained with the two tools, where columns SHAP-FRP mismatch shown the number of mismatches obtained with SHAP, and column s SHAP-FRP mismatch shows the number of mismatches obtained with s SHAP16. As can be concluded, SHAP produces several mismatches. In contrast, s SHAP produces no mismatch. It should be noted that both tools are approximating the SHAP scores given the respective characteristic functions, i.e. the computed scores are not necessarily the ones dictated by (7). Observe that υs consists of replacing the original classifier by a new boolean classifier. Hence, from (Huang and Marques-Silva 2024), such boolean classifiers can also produce misleading information. Nevertheless, given the results above and other experiments, in the cases where υs was used, we were unable to observe some of the issues proposed in 15From (Huang and Marques-Silva 2024), we consider (i) the two DTs of case study 2 (Fig. 3 in (Huang and Marques-Silva 2024)), referred to as cs02a and cs02b; (ii) the two DTs of case study 3 (Fig. 5 in (Huang and Marques-Silva 2024)), referred to as cs03a and cs03b; and (iii) the two DTs of case study 4 (Fig. 8 in (Huang and Marques-Silva 2024)), referred to as cs04a and cs04b. Moreover, for cs02a, cs02b, cs03a and cs03b there exist 16 instances, whereas for cs04a and cs04b there exist 24 instances (because of a discrete but non-boolean domain for one of the features.) 16It should be noted that s SHAP can replace SHAP in any application domain, ensuring similar performance. However, a more extensive assessment of the quality of the results of the two tools is unrealistic at present; we would have to be able to compute exact SHAP scores, and this is only computationally feasible for very simple ML models, e.g. restricted examples of DTs. DT SHAP-FRP mismatch s SHAP-FRP mismatch cs02a 11 0 cs02b 4 0 cs03a 5 0 cs03b 4 0 cs04a 15 0 cs04b 4 0 Table 1: Comparison between SHAP and s SHAP. The source code of s SHAP is available from https://github.com/ Xuanxiang Huang/aaai25_code earlier work (Marques-Silva and Huang 2024; Huang and Marques-Silva 2024). It is the subject of future work to decide whether such issues can occur for boolean classifiers. Conclusions Recent work demonstrated the existence of classifiers for which the exact SHAP scores are unsatisfactory. This paper argues that the issues identified with SHAP scores result from the characteristic functions used in earlier work. As a result, the paper devises several properties which characteristic functions must respect in order to compute SHAP scores that do not exhibit those issues. Complexity-wise, the paper argues that the proposed characteristic functions are as hard to compute as the characteristic functions used in earlier works studying the complexity of SHAP scores (Van den Broeck et al. 2021; Arenas et al. 2021; Van den Broeck et al. 2022; Arenas et al. 2023), or harder. Finally, the paper proposes simple modifications to the tool SHAP (Lundberg and Lee 2017), thereby obtaining SHAP scores that respect some of the proposed properties. Proofs Proposition 2. If a characteristic function is strongly valueindependent, then it is weakly value-independent. Proof. If a characteristic function is strongly value independent, it suffices to restrict the choices of µ to surjective functions to make it weakly value independent. Proposition 3. Given the definition of υa, υc and υn, then Sc A(i; E, υa) 0, Sc C(i; E, υc) 0, and Sc N(i; E, υN) 0. Proof. (Sketch) We only consider υa. (The proof for υc and υn follows from Proposition 5.) It is plain that i(S; E, υa) { 1, 0, 1}, given the possible values that υa can take. In fact, it is the case that i(S; E, υa) {0, 1}. If a set S F is a WAXp, then a proper superset is also a WAXp; hence it is never the case that i(S; E, υa) = 1. Since every i(S; E, υa) 0, then Sc A(i; E, υa) 0. Proposition 4. The following holds true: 1. (S F).[υa(S; E) = 1 WAXp(S; E)]. 2. (S F).[υn(S; E) = 1 WCXp(F \ S; E)]. 3. (S F).[υc(S; E) = 1 WCXp(S; E)]. Proof. We consider each case separately: 1. If υa(S; E) = 1, then, as noted earlier in the paper, σ(x; E) = 1 for all points x Υ(S), and so the ML model s prediction is indistinguishable from q for all points in Υ(S). Hence, by definition, S is a WAXp. Conversely, if S is an WAXp, then the prediction must be indistinguishable from q for all points x in Υ(S), (x Υ(S)).[σ(x; E)]. Thus, υa(S; E) = 1. 2. If υn(S; E) = 1, then σ(x; E) = 1 for some point(s) x Υ(S), and so the ML model s prediction is distinguishable from q for some point(s) in Υ(S). Hence, by definition, F \ S is a WCXp. Conversely, if F \ S is an WCXp, then the prediction must be distinguishable from q for some point(s) x in Υ(S), i.e. (x Υ(S)).[σ(x; E) = 1]. Thus, υn(S; E) = 1. 3. If υc(S; E) = 1, then σ(x; E) = 1 for some point(s) x Υ(F \ S), and so the ML model s prediction is distinguishable from q for some point(s) in Υ(F \ S). Hence, by definition, S is a WCXp. Conversely, if S is an WCXp, then the prediction must be distinguishable from q for some point(s) x in Υ(F \S), i.e. (x Υ(F \ S)).[σ(x; E) = 1]. Thus, υc(S; E) = 1. Proposition 5. The following holds true: 1. (i F).[Sc A(i; E, υa) = Sc N(i; E, υn)]. 2. (i F).[Sc A(i; E, υa) = Sc C(i; E, υc)]. Proof. We consider each case separately: 1. By definition, it is plain that υa(S; E) + υn(S; E) = 1, for any S F, because it must be the case that either υs(S; E) = 1 or υs(S; E) < 1, but not both. Given the values that υa(S; E) can take, it is also plain that i(S; E, υa) { 1, 0, 1}. Moreover, if i(S; E, υa) = 1, then i(S; E, υn) = 1. If i(S; E, υa) = 1, then i(S; E, υn) = 1. Also, if i(S; E, υa) = 0, then i(S; E, υn) = 0. Thus, for any i F and S F, i(S; E, υn) = i(S; E, υa). Hence, the result follows. 2. Since (S F).WCXp(F \ S; E) WAXp(S; E), by definition, then we have (i F), (S (F \ {i})), i(S; E, υa) = 1 WAXp(S; E) WAXp(S {i}; E) WCXp(F \ S; E) WCXp(F \ (S {i}); E) WCXp(F \ S; E) WCXp((F \ {i}) \ S; E) i((F \ {i}) \ S; E, υc) = 1 Now, let Φ(i) := {S (F \ {i}) | i(S; E, υa) = 1}. Then, by construction, Sc A(i; E, υa) = P S Φ(i) ς(|S|) (because i(S; E, υa) = 0 otherwise) and, by the equivalence above, Sc C(i; E, υc) = P S Φ(i) ς(|F \ {i}) \ S|). However, it is immediate to prove that ς(|S|) = ς(|F \ {i})\S|), and so the two sums are also equal. This proves the result. Proposition 6. For t {s, a, c, n} and i F, it is the case that the characteristic function υt respects strong value independence. Proof. For a characteristic function to respect strong value independence, the SHAP scores must not change if the values are mapped using some function µ. By hypothesis, for any point x F, the resulting ML models will predict µ(q) iff the original ML model predicts q. This means the resulting similarity predicates are the same for the two ML models, and so the SHAP scores Sc T , T {S, A, C, N}, remain unchanged. Proposition 7. For T {A, C, N}, then it is the case that (i F).Irrelevant(i) (Sc T (i) = 0). Proof. First, we consider υa. Let i F be an irrelevant feature. It is plain that i(S; E, υa) { 1, 0, 1}, given the possible values that υa can take. However, as argued above, i(S; E, υa) {0, 1}, since if a set S F is a WAXp, then a proper superset is also a WAXp; hence it is never the case that i(S; E, υa) = 1. We are interested in the sets S (F \ {i}) for which i(S; E, υa) = 1, since these are the only ones that contribute to making Sc A(i; E, υa) = 0. For i(S; E, υa) = 1, it must be the case that υa(S; E) = 0 and υa(S {i}; E) = 1. However, this would imply that i would be included in some AXp (Huang et al. 2023). But i is irrelevant, and so it is not included in any AXp. Hence, there exists no set S (F \ {i}) such that i(S; E, υa) = 1, and so Sc A(i; E, υa) = 0. Let Sc A(i; E, υa) = 0. An analysis similar to the above one allows concluding that there exist no sets S such that i(S; E, υa) = 1. Hence, it is never the case that υa(S; E) = 0 and υa(S {i}; E) = 1. Thus, i is not included in any AXp, and so it is irrelevant. For υc and υn, it suffices to invoke Proposition 5; hence, the features for which Sc A(i) = 0 are exactly the ones for which Sc C(i) = 0 and Sc N(i) = 0. This concludes the proof that Sc T , with T {A, C, N}, respects (9). Proposition 9. Computing SHAP scores Sc S is #P-hard for boolean classifiers in CNF or DNF. Proof. From (Van den Broeck et al. 2021, 2022), it is known that computing SHAP scores is polynomially equivalent to computing the expected value. In the boolean case, and so in the case of υs, this is polynomially equivalent to model counting. Furthermore, model counting for DNF and CNF formulas is #P-complete (Valiant 1979). Thus, computing the SHAP scores using υs is #P-hard. Proposition 10. For T {A, C, N}, the computation of the SHAP scores Sc T is NP-hard for d-DNNF & DDBC classifiers. Proof. We reduce the problem of feature relevancy to the problem of computing the SHAP scores Sc T , with T {A, C, N}. Since feature relevancy is NP-complete for d DNNF circuits (Huang et al. 2023), this proves that computing the SHAP scores Sc T , with T {A, C, N} is NP-hard. Given an explanation problem we can decide feature membership as follows. We compute the SHAP score for each feature i F. Moreover, since υt, t {a, c, n} are compliant with feature (ir)relevancy (see (9)), then Sc T (i; E, υt) = 0 iff feature i is irrelevant. Hence, if we could compute the SHAP scores in polynomial-time, then we could decide feature relevancy in polynomial-time, and so computing the SHAP-scores for d-DNNFs is NP-hard. Now, since DDBCs generalize d-DNNFs (Arenas et al. 2023), then computing the SHAP-scores for DDBCs is also NPhard. Limitations of SHAP Scores Based on Baselines We focus on BShap (Sundararajan and Najmi 2020); similar analyzes could be made for other baselines (Janzing, Minorics, and Blöbaum 2020; Sundararajan and Najmi 2020). Throughout this section, the baseline is a point w F. Furthermore, for each S F, let x S b be such that x S b,i = ITE(i S, vi, wi). Given w F, the BShap characteristic function υb is defined by υb(S) = κ(x S b ), for S F. Remarks about baselines. Analysis of the definition of BShap (Sundararajan and Najmi 2020) allows proving the following results. Proposition 13. The following holds: 1. BShap is only well-defined if all the domains are boolean, i.e. F = {0, 1}m. 2. BShap is only well-defined when w = v. Proof. By contradiction, let us consider i F, such that either |Di| > 2 or wi = vi. Then there exists a point z F such that zi {vi, wi}. By construction, for each S F, x S b is different from z. Thus, υb and so Scb do not depend on κ(z). Therefore, we can use the value of κ(z) to change the AXps (and CXps) without modifying the BShap scores. As there are at least 2(|F| 1) such points z, it is plain that constructing counterexample is simple. BShap also misleads. The following notation is used S F, let v S be defined by v S i = ITE(i S, vi, vi), with i F. For S F, then υb(S) = κ(v S). Proposition 14. υb misleads. Proof. Let κ(x1, x2) = ITE(x1 = 1, 1, 2x2), and instance (v, c) = ((1, 1), 1). It is plain that feature 1 influences both selecting the prediction 1 and changing the prediction to some other value. In contrast, feature 2 has not influence in either setting or changing the prediction of class 1. It is also plain that the set of AXps is {{1}}, and also that κ(x1, x2) = 1 iff x1 = 1. However, if we compute Scb, we get Scb(1) = 0 and Scb(2) = 1, which is of course misleading. To confirm the SHAP scores, we proceed as follows. υb( ) = κ(0, 0) = 0, υb({1}) = κ(1, 0) = 1, υb({2}) = κ(0, 1) = 2, and υb({1, 2}) = κ(1, 1) = 1. Thus, b(1, ) = υb({1}) υb( ) = 1, b(1, {2}) = υb({1, 2}) υb({2}) = 1, b(2, ) = υb({2}) υb( ) = 2, b(2, {1}) = υb({1, 2}) υb({2}) = 0. And finally, Scb(1) = ( b(1, {2}) + b(1, ))/2 = 0, Scb(2) = ( b(2, {1}) + b(2, ))/2 = 1. Acknowledgements We thank the anonymous reviewers for the helpful comments. This work was supported in part by the Spanish Government under grant PID2023-152814OB-I00, and by ICREA starting funds. This work was supported in part by the National Research Foundation, Prime Minister s Office, Singapore under its Campus for Research Excellence and Technological Enterprise (CREATE) programme. This work was also supported in part by the AI Interdisciplinary Institute ANITI, funded by the French program Investing for the Future PIA3 under Grant agreement no. ANR-19-PI3A-0004. Afchar, D.; Guigue, V.; and Hennequin, R. 2021. Towards Rigorous Interpretations: a Formalisation of Feature Attribution. In ICML, 76 86. Amgoud, L.; Ben-Naim, J.; and Vesic, S. 2017. Measuring the Intensity of Attacks in Argumentation Graphs with Shapley Value. In IJCAI, 63 69. Arenas, M.; Barceló, P.; Bertossi, L. E.; and Monet, M. 2021. The Tractability of SHAP-Score-Based Explanations for Classification over Deterministic and Decomposable Boolean Circuits. In AAAI, 6670 6678. Arenas, M.; Barceló, P.; Bertossi, L. E.; and Monet, M. 2023. On the Complexity of SHAP-Score-Based Explanations: Tractability via Knowledge Compilation and Non Approximability Results. J. Mach. Learn. Res., 24: 63:1 63:58. Biradar, G.; Izza, Y.; Lobo, E.; Viswanathan, V.; and Zick, Y. 2023. Axiomatic Aggregations of Abductive Explanations. Co RR, abs/2310.03131. Biradar, G.; Izza, Y.; Lobo, E.; Viswanathan, V.; and Zick, Y. 2024. Axiomatic Aggregations of Abductive Explanations. In AAAI, 11096 11104. Breiman, L.; Friedman, J.; Olshen, R.; and Stone, C. 1984. Classification and regression trees. Chapman and Hall. ISBN 9781315139470. Brix, C.; Müller, M. N.; Bak, S.; Johnson, T. T.; and Liu, C. 2023. First three years of the international verification of neural networks competition (VNN-COMP). Int. J. Softw. Tools Technol. Transf., 25(3): 329 339. Campbell, T. W.; Roder, H.; Georgantas III, R. W.; and Roder, J. 2022. Exact Shapley values for local and model-true explanations of decision tree ensembles. Machine Learning with Applications, 9: 100345. Chalkiadakis, G.; Elkind, E.; and Wooldridge, M. J. 2012. Computational Aspects of Cooperative Game Theory. Synthesis Lectures on Artificial Intelligence and Machine Learning. Morgan & Claypool Publishers. ISBN 978-3-031-00430-8. Darwiche, A. 2023. Logic for Explainable AI. In LICS, 1 11. Dubey, P. 1975. On the uniqueness of the Shapley value. International Journal of Game Theory, 4: 131 139. Felsenthal, D. S.; and Machover, M. 1998. The measurement of voting power. Edward Elgar Publishing. Felsenthal, D. S.; and Machover, M. 2015. The measurement of a priori voting power. In Heckelman, J. C.; and Miller, N. R., eds., Handbook of Social Choice and Voting, chapter 08, 117 139. Edward Elgar Publishing. Fryer, D. V.; Strümke, I.; and Nguyen, H. D. 2021. Shapley Values for Feature Selection: The Good, the Bad, and the Axioms. IEEE Access, 9: 144352 144360. Goodfellow, I. J.; Shlens, J.; and Szegedy, C. 2015. Explaining and Harnessing Adversarial Examples. In ICLR. Huang, X.; Cooper, M. C.; Morgado, A.; Planes, J.; and Marques-Silva, J. 2023. Feature Necessity & Relevancy in ML Classifier Explanations. In TACAS, 167 186. Huang, X.; and Marques-Silva, J. 2023. The Inadequacy of Shapley Values for Explainability. Co RR, abs/2302.08160. Huang, X.; and Marques-Silva, J. 2024. On the failings of Shapley values for explainability. International Journal of Approximate Reasoning, 109112. Hunter, A.; and Konieczny, S. 2006. Shapley Inconsistency Values. In KR, 249 259. Hunter, A.; and Konieczny, S. 2010. On the measure of conflicts: Shapley Inconsistency Values. Artif. Intell., 174(14): 1007 1026. Ignatiev, A.; Narodytska, N.; and Marques-Silva, J. 2019a. Abduction-Based Explanations for Machine Learning Models. In AAAI, 1511 1519. Ignatiev, A.; Narodytska, N.; and Marques-Silva, J. 2019b. On Relating Explanations and Adversarial Examples. In Neur IPS, 15857 15867. James, G.; Witten, D.; Hastie, T.; and Tibshirani, R. 2017. An introduction to statistical learning. Springer. Janzing, D.; Minorics, L.; and Blöbaum, P. 2020. Feature relevance quantification in explainable AI: A causal problem. In AISTATS, 2907 2916. Kumar, I.; Scheidegger, C.; Venkatasubramanian, S.; and Friedler, S. A. 2021. Shapley Residuals: Quantifying the limits of the Shapley value for explanations. In Neur IPS, 26598 26608. Kumar, I. E.; Venkatasubramanian, S.; Scheidegger, C.; and Friedler, S. A. 2020. Problems with Shapley-value-based explanations as feature importance measures. In ICML, 5491 5500. Letoffe, O.; Huang, X.; and Marques-Silva, J. 2024a. On Correcting SHAP Scores. Co RR, abs/2405.00076. Letoffe, O.; Huang, X.; and Marques-Silva, J. 2024b. SHAP scores fail pervasively even when Lipschitz succeeds. Co RR, abs/2412.13866. Lundberg, S. M.; and Lee, S. 2017. A Unified Approach to Interpreting Model Predictions. In Neur IPS, 4765 4774. Marques-Silva, J. 2022. Logic-Based Explainability in Machine Learning. In Reasoning Web, 24 104. Marques-Silva, J. 2024. Logic-Based Explainability: Past, Present and Future. In ISo LA, 181 204. Marques-Silva, J.; and Huang, X. 2024. Explainability Is Not a Game. Commun. ACM, 67(7): 66 75. Marques-Silva, J.; and Ignatiev, A. 2022. Delivering Trustworthy AI through Formal XAI. In AAAI, 12342 12350. Merrick, L.; and Taly, A. 2020. The Explanation Game: Explaining Machine Learning Models Using Shapley Values. In CDMAKE, 17 38. Mishra, P. 2023. Explainable AI Recipes. Apress. ISBN 978-1-4842-9029-3. Molnar, C. 2023. Interpreting Machine Learning Models With SHAP. Lulu.com. ISBN 979-8857734445. Mothilal, R. K.; Mahajan, D.; Tan, C.; and Sharma, A. 2021. Towards Unifying Feature Attribution and Counterfactual Explanations: Different Means to the Same End. In AIES, 652 663. Roth, A. E. 1988. The Shapley value: essays in honor of Lloyd S. Shapley. Cambridge University Press. Shapley, L. S. 1953. A value for n-person games. Contributions to the Theory of Games, 2(28): 307 317. Shapley, L. S.; and Shubik, M. 1954. A method for evaluating the distribution of power in a committee system. American political science review, 48(3): 787 792. Strumbelj, E.; and Kononenko, I. 2010. An Efficient Explanation of Individual Classifications using Game Theory. J. Mach. Learn. Res., 11: 1 18. Strumbelj, E.; and Kononenko, I. 2014. Explaining prediction models and individual predictions with feature contributions. Knowl. Inf. Syst., 41(3): 647 665. Sundararajan, M.; and Najmi, A. 2020. The Many Shapley Values for Model Explanation. In ICML, 9269 9278. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I. J.; and Fergus, R. 2014. Intriguing properties of neural networks. In ICLR. Valiant, L. G. 1979. The Complexity of Enumeration and Reliability Problems. SIAM J. Comput., 8(3): 410 421. Van den Broeck, G.; Lykov, A.; Schleich, M.; and Suciu, D. 2021. On the Tractability of SHAP Explanations. In AAAI, 6505 6513. Van den Broeck, G.; Lykov, A.; Schleich, M.; and Suciu, D. 2022. On the Tractability of SHAP Explanations. J. Artif. Intell. Res., 74: 851 886. Watson, D. S.; Gultchin, L.; Taly, A.; and Floridi, L. 2021. Local explanations via necessity and sufficiency: unifying theory and practice. In UAI, volume 161, 1382 1392. Wu, M.; Wu, H.; and Barrett, C. W. 2023. Veri X: Towards Verified Explainability of Deep Neural Networks. In Neur IPS. Yan, T.; and Procaccia, A. D. 2021. If You Like Shapley Then You ll Love the Core. In AAAI, 5751 5759. Young, H. P. 1985. Monotonic solutions of cooperative games. International Journal of Game Theory, 14: 65 72. Young, K.; Booth, G.; Simpson, B.; Dutton, R.; and Shrapnel, S. 2019. Deep neural network or dermatologist? Co RR, abs/1908.06612. Yu, J.; Farr, G.; Ignatiev, A.; and Stuckey, P. J. 2023. Anytime Approximate Formal Feature Attribution. Co RR, abs/2312.06973. Yu, J.; Farr, G.; Ignatiev, A.; and Stuckey, P. J. 2024. Anytime Approximate Formal Feature Attribution. In SAT, 30:1 30:23. Yu, J.; Ignatiev, A.; and Stuckey, P. J. 2023. On Formal Feature Attribution and Its Approximation. Co RR, abs/2307.03380.