# testing_semantic_importance_via_betting__9dc36bb0.pdf Testing Semantic Importance via Betting Jacopo Teneggi Johns Hopkins University jtenegg1@jhu.edu Jeremias Sulam Johns Hopkins University jsulam1@jhu.edu Recent works have extended notions of feature importance to semantic concepts that are inherently interpretable to the users interacting with a black-box predictive model. Yet, precise statistical guarantees such as false positive rate and false discovery rate control are needed to communicate findings transparently, and to avoid unintended consequences in real-world scenarios. In this paper, we formalize the global (i.e., over a population) and local (i.e., for a sample) statistical importance of semantic concepts for the predictions of opaque models by means of conditional independence, which allows for rigorous testing. We use recent ideas of sequential kernelized independence testing to induce a rank of importance across concepts, and we showcase the effectiveness and flexibility of our framework on synthetic datasets as well as on image classification using several vision-language models. 1 Introduction Providing guarantees on the decision-making processes of autonomous systems, often based on complex black-box machine learning models, is paramount for their safe deployment. This need motivates efforts towards responsible artificial intelligence, which broadly entails questions of reliability, robustness, fairness, and interpretability. One popular approach to the latter is to use post-hoc explanation methods to identify the features that contribute the most towards the predictions of a model. Several alternatives have been proposed over the past few years, drawing from various definitions of features (e.g., pixels or groups thereof for vision tasks [41], words for language tasks [21], or nodes and edges for graphs [84]) and of importance (e.g, gradients for Grad-CAM [57], Shapley values for SHAP [7, 15, 42, 70], or information-theoretic quantities [43]). While most explanation methods highlight important features in the input space of the predictor, users may care more about their meaning. For example, a radiologist may want to know whether a model considered the size and spiculation of a lung nodule to quantify its malignancy, and not just its raw pixel values. To decouple importance from input features, Kim et al. [34] showed how to learn the vector representation of semantic concepts that are inherently interpretable to users (e.g., stripes , sky , or sand ) and how to study their gradient importance for model predictions. Recent vision-language (VL) models that jointly learn an image and text encoder, such as CLIP [16, 52], have made these representations commonly referred to as concept activation vectors (CAVs) more easily accessible. With these models, obtaining the representation of a concept boils down to a forward pass of the pretrained text encoder, which alleviates the need of a dataset comprising images annotated with their concepts. Several recent works have defined semantic importance both with CAVs and VL models by means of concept bottleneck models (e.g., CBM [36], PCBM [85], La Bo [83]), information-theoretic quantities (e.g., V-IP [37]), sparse coding (e.g., CLIP-IP-OMP [13], Sp Li Ce [6]), network dissection [2] (e.g., CLIP-DISSECT [47], Text Span [26], INVi TE [14]), or causal inference (e.g., Di Con Struct [44], Sani et al. [55]). On the other hand, it is important to communicate findings of important features precisely and transparently in order to avoid unintended consequences in downstream decision tasks. Going back to the example of a radiologist diagnosing lung cancer, how should they interpret two concepts with 38th Conference on Neural Information Processing Systems (Neur IPS 2024). different importance scores? Does their difference in importance carry any statistical meaning? To start addressing similar questions [10, 71] introduced statistical tests for the local (i.e., on a sample) conditional independence structure of a model s predictions. Framing importance by means of conditional independence allows for rigorous testing with false positive rate control. That is, for a user-defined significance level α (0, 1), the probability of wrongly deeming a feature important is no greater than α, which directly conveys the uncertainty in an explanation. Yet these methods consider features as coordinates in the input space, and it is unclear how to extend these ideas to abstract, semantic concepts. In this work, we formalize semantic importance at three distinct levels of statistical independence with null hypotheses of increasing granularity: (i) marginally over a population (i.e., global importance), (ii) conditionally over a population (i.e., global conditional importance), and (iii) for a sample (i.e., local conditional importance).1 Each of these notions will allow us to inquire the extent to which the output of a model depends on specific concepts both over a population and on specific samples and thus deem them important. To test for these notions of semantic importance, instead of classical (or offline [58]) independence testing techniques [5, 10, 11, 28, 29, 69, 86], which are based on p-values and informally follow the rule reject if p α , we propose to use principles of testing by betting (or sequential testing) [59], which are based on e-values [76] and follow the reject when e 1/α rule. As we will expand on, this choice is motivated by the fact that sequential tests are data-efficient and adaptive to the hardness of the problem which naturally induces a rank of importance. We will couple principles of conditional randomization testing (CRT) [11] with recent advances in sequential kernelized independence testing (SKIT) [51, 62], and introduce two novel procedures to test for our definitions of semantic importance: the conditional randomization SKIT (C-SKIT) to study global conditional importance, and following the explanation randomization test (XRT) framework [71] the explanation randomization SKIT (X-SKIT) to study local conditional importance. We will illustrate the validity of our proposed tests on synthetic datasets, and showcase their flexibility on zero-shot image classification on real-world datasets across several and diverse VL models. 1.1 Summary of Contributions and Related Works In this paper, we will rigorously define notions of statistical importance of semantic concepts for the predictions of black-box models via conditional independence both globally over a population and locally for individual samples. For any set of concepts, and for each level of independence, we introduce novel sequential testing procedures that induce a rank of importance. Before presenting the details of our methodology, we briefly expand on a few distinctive features of our work. Explaining nonlinear predictors. Compared to recent methods based on concept bottleneck models [36, 83, 84], our framework does not require training a surrogate linear classifier because we study the semantic importance structure of any given, potentially nonlinear and randomized model. This distinction is not minor training concept bottleneck models results on explanations that pertain to the surrogate (linear) model instead of the original (complex, nonlinear) predictor, and these simpler surrogate models typically reduce performance [48, 85]. In contrast, we provide statistical guarantees directly on the original predictor that would be deployed in the wild. Flexible choice of concepts. Furthermore, our framework does not rely on the presence of a large concept bank (but it can use one if it is available). Instead, we allow users to directly specify which concepts they want to test. This feature is important in settings that involve diverse stakeholders. In medicine, for example, there are physicians, patients, model developers, and members of the regulatory agency tasked with auditing the model each of whom might prefer different semantics for their explanations. Current explanation methods cannot account for these differences off-the-self. Local semantic explanations. Our framework entails explanations for specific (fixed) inputs, whereas prior approaches that rely on the weights of a linear model only inform on global notions of importance. Recently, Shukla et al. [64] and Pham et al. [50] set forth ideas of local semantic importance by combining LIME [54] with T-CAV [34], and by leveraging prototypical part networks [46], respectively. Our work differs in that it does not apply to images only, it considers formal notions of statistical importance rather than heuristics of gradient importance, and it provides guarantees such as Type I error and false discovery rate (FDR) control. 1We adopt the distinction between global and local importance as presented in [18]. Sequential kernelized testing. Motivated by the statistical properties of kernelized independence tests [62], we will employ the maximum mean discrepancy (MMD) [29] as the test statistic in our proposed procedures. The recent related work in Shaer et al. [58] introduces the sequential version of the conditional randomization test (CRT) [11], dubbed e-CRT because of the use of e-values. Unlike our work, Shaer et al. [58] employ residuals of a predictor as test statistic, they do so in the context of global tests only, and unrelated to questions of semantic interpretability. 2 Background In this section, we briefly introduce the necessary notation and general background. Throughout this work, we will denote random variables with capital letters, and their realizations with lowercase. For example, X PX is a random variable sampled from PX, and x indicates an observation of X. Problem setup. We consider k-fold classification problems such that (X, Y ) PXY is a random sample X X with its one-hot label Y {0, 1}k, and (x, y) denotes a particular observation. We assume we are given a fixed predictive model, consisting of an encoder f : X Rd and a classifier g : Rd Rk such that h = f(x) is a d-dimensional representation of x, and ˆyk = g(h)k = g(f(x))k is the prediction of the model for a particular class k (e.g., ˆyk is the output, or score, for class dog ). Naturally, H, ˆY denote the random counterparts of h and ˆy. Although our contributions do not make any assumptions on the performance of the model, f and g can be thought of as good predictors, e.g. those that approximate the conditional expectation of Y given X. Concept bottleneck models (CBMs). Let c = [c1, . . . , cm] Rd m be a dictionary of m concepts such that j [m] := {1, . . . , m}, cj Rd is the representation of the jth concept either obtained via CAVs [34] or a VL model s text encoder. Then, z = c, h is the projection of the embedding h onto the concepts c, and, with appropriate normalization, zj [ 1, 1] is the amount of concept cj in h. Intuitively, CBMs project dense representations onto the subspace of interpretable semantic concepts [36], and their performance strongly depends on the size of the dictionary [48]. For example, it is common for m to be as large as the embedding size (e.g., d = 768 for CLIP:Vi T-L/14). In this work, instead, we let concepts be user-defined, allowing for cases where m d (e.g., m = 20). This is by design as (i) the contributions of this paper apply to any set of concepts, and (ii) it has been shown that humans can only gain valuable information if semantic explanations are succinct [53]. However, we remark that the construction of informative concept banks especially for domainspecific applications is subject of ongoing complementary research [20, 48, 78, 80, 81]. Conditional randomization tests. Recall that two random variables A, B are conditionally independent if, and only if, given a third random variable C, it holds that PA|B,C = PA|C (i.e., A B | C). That is, B does not provide any more information about A with C present. Candes et al. [11] introduced the conditional randomization test (CRT), based on the observation that if A B | C, then the triplets (A, B, C) and (A, e B, C) with e B PB|C, are exchangeable. That is, PABC = PA e BC and one can essentially mask B without changing the joint distribution. Opposite to classical methods, the CRT assumes the conditional distribution of the covariates is known (i.e., PB|C), which lends itself to settings with ample unlabeled data. With this general background, we now present the main technical contributions of this paper. 3 Testing Semantic Importance via Betting Our objective is to test the statistical importance of semantic concepts for the predictions of a fixed, potentially nonlinear model, while inducing a rank of importance. Fig. 1 depicts the problem setup a fixed model, composed of the encoder f and classifier g, is probed via a set of concepts c. This figure also illustrates the key difference with post-hoc concept bottleneck models (PCBMs) [85], in that we do not train a sparse linear layer to approximate E[Y | Z]. Instead, we focus on characterizing the dependence structure between ˆY and Z. Herein, we will drop the ˆyk notation and simply write ˆy, ˆY because we always consider the output of the model for a particular class individually. 3.1 Formalizing Statistical Importance of Semantic Concepts We start by defining global semantic importance as marginal dependence between ˆY and Zj, j [m]. Figure 1: Overview of the problem setup and our contribution. Definition 1 (Global semantic importance). For a concept j [m], HG 0,j : ˆY Zj (1) is the global semantic independence null hypothesis. Rejecting HG 0,j means that we have observed enough evidence to believe the response of the model depends on concept j, i.e. concept j is globally important over the population. Note that both ˆY and Zj are fixed functions of the same random variable H, i.e. ˆY = g(H) and Zj = cj, H . Then, it is reasonable to wonder whether there is any point in testing HG 0,j at all can we obtain two independent random variables from the same one? For example, let g be a linear classifier such that ˆY = w, H , w Rd. Intuition might suggest that if w, cj = 0 then ˆY Zj, i.e. if the classifier is orthogonal to a concept, then the concept cannot be important. We show in the short lemma below (whose proof is in Appendix C.1) that this is false, and that, arguably unsurprisingly, statistical independence is different from orthogonality between vectors, which motivates the need for our testing procedures. Lemma 1. Let ˆY = w, H , w Rd. If d 3, then HG 0,j is true w, cj = 0. The null hypothesis HG 0,j above precisely defines the global importance of a concept, but it ignores the information contained in the rest of them, and concepts may be correlated. For example, predictions for class dog may be independent of stick given tail and paws , although stick is marginally important. To address this, and inspired by [11], we define global conditional semantic importance. Definition 2 (Global conditional semantic importance). For a concept j [m], let j := [m] \ {j} denote all but the jth concept. Then, HGC 0,j : ˆY Zj | Z j (2) is the global conditional semantic independence null hypothesis. Analogous to Definition 1, rejecting HGC 0,j means that we have accumulated enough evidence to believe the response of the model depends on concept j even in the presence of the remaining concepts, i.e. there is information about ˆY in concept j that is missing from the rest. We stress an important distinction between Definition 2 and PCBMs: the latter approximate E[Y | Z] with a sparse linear layer, which is inherently interpretable because the regression coefficients directly inform on the global conditional independence structure of the predictions, i.e. if ˆY = β, Z , β Rm then ˆY Zj | Z j βj = 0. In this work, however, we do not assume any parametrization between the concepts and the labels because we want to provide guarantees on the original, fixed classifier g that acts directly on an embedding h. From this conditional independence perspective, PCBMs can be interpreted as a (parametric) test of true linear independence (i.e., HPCBM 0,j : Y Zj | Z j) between the concepts and the labels (note that HPCBM 0,j has the true label Y and not the prediction ˆY ), whereas we study the semantic structure of the predictions of a complex model, which may have learned spurious, non-linear correlations of these concepts from data. Akin to the CRT [11], we assume we can sample from the conditional distribution of the concepts, i.e. PZj|Z j. Within the scope of this work, m is small (m 20) and we will show how to effectively approximate this distribution with nonparametric methods that do not require prohibitive computational resources. This is an advantage of testing a few semantic concepts compared to input features, especially for imaging tasks where the number of pixels is large ( 105) and learning a conditional generative model (e.g., a diffusion model [67]) may be expensive. Finally, we define the notion of local conditional semantic importance. That is, we are interested in finding the most important concepts for the prediction of the model locally on a particular input x, i.e. ˆy = g(f(x)). Recently, [10, 71] showed how to deploy ideas of conditional randomization testing for local explanations of machine learning models. Briefly, let B, C be random variables and η(B, C) a fixed, possibly randomized, real-valued predictor. For an observation (b, c), the explanation randomization test (XRT) [71] null hypothesis is η(b, c) d= η( e B, c), e B PB|C=c. That is, the observed value of B does not affect the distribution of the response given the observed value of C. We now generalize these ideas. Definition 3 (Local conditional semantic importance). For a fixed z [ 1, 1]m and any C [m], denote ˆYC = g( e HC) with e HC PH|ZC=z C. Then, for a concept j [m] and a subset S [m]\{j}, HLC 0,j,S : ˆYS {j} d= ˆYS (3) is the local conditional semantic independence null hypothesis. In words, rejecting HLC 0,j,S means that, given the observed concepts in S, concept j / S affects the distribution of the response of the model, hence it is important. For this test, we assume we can sample from the conditional distribution of the embeddings given a subset of concepts (i.e., PH|ZC=z C). This is equivalent to solving an inverse problem stochastically, since z = c, h and c is not invertible (c Rd m, m d). Hence, there are several embeddings h that could have generated the observed z C. We will use nonparametric sampling ideas to achieve this, stressing that it suffices to sample the embeddings H and not an entire input image X since the classifier g directly acts on h and the encoder f is deterministic. Finally, we remark that HLC 0,j,S differs from the XRT null hypothesis in that conditioning is performed in the space of semantic concepts instead of the input s. With these precise notions of semantic importance, we now show how to test for each one of them with principles of sequential kernelized independence testing (SKIT) [51]. 3.2 Testing by Betting A classical approach to hypothesis testing consists of formulating a null hypothesis H0, collecting data, and then summarizing evidence by means of a p-value. Under the null, the probability of observing a small p-value is small. Thus, for a significance level α (0, 1), we can reject H0 if p α. In this setting, all data is collected first, and then processed later (i.e., offline). Alternatively, one can instantiate a game between a bettor and nature [59, 60]. At each turn of the game, the bettor places a wager against H0, and then nature reveals the truth. If the bettor wins, they will increase their wealth, otherwise lose some. More formally, and as is commonly done [51, 58, 62], we define a wealth process {Kt}t N0 with K0 = 1 and Kt = Kt 1 (1 + vtκt) where vt, κt [ 1, 1] are a betting fraction and the payoff of the bet, respectively. It is now easy to see that when vtκt 0 (i.e., the bettor wins) the wealth increases, and the opposite otherwise. If the payoff κt guarantees the game is fair, i.e. the bettor cannot accumulate wealth under the null, then we can use the wealth process to reject H0 with Type I error control (details in Appendix A). In particular, for a significance level α (0, 1), we denote τ := min{t 1 : Kt 1/α} the rejection time of H0. The choice of using sequential testing is motivated by two fundamental properties. First, sequential tests are adaptive to the hardness of the problem, sometimes provably [62, Proposition 3]. That is, the harder it is to reject the null, the longer the test will take, and vice versa. This naturally induces a rank of importance across concepts if concept cj rejects faster than cj , then cj is more important (i.e., it is easier to reject the null hypothesis that the predictions do not depend on cj). We stress that this is not always possible by means of p-values because they do not measure effect sizes: consider two concepts that reject their respective nulls at the same significance level; one cannot distinguish which if any is more important. As we will show in our experiments, all tests used in this work are adaptive in practice, but statistical guarantees on their rejection times are currently open questions, and we consider them as future work. Second, sequential tests are sample-efficient because they only analyze the data is needed to reject, which is especially important for conditional randomization tests. In the offline scenario, we would have to resample the entire dataset several times (which is expensive), but the sequential test would terminate in at most the size of the dataset [24]. Algorithm 1 Level-α C-SKIT for concept j Input: Stream ( ˆY (t), Z(t) j , Z(t) j) P ˆY Zj Z j. 1: K0 1 2: Initialize ONS strategy (Algorithm A.1) 3: for t = 1, . . . do 4: Compute ρt as in Eq. (4) 5: D(t) = ( ˆY (t), Z(t) j , Z(t) j) 6: Sample e Z(t) j PZj|Z j=Z(t) j 7: e D(t) ( ˆY (t), e Z(t) j , Z(t) j) 8: κt tanh(ρt(D(t)) ρt( e D(t))) 9: Kt Kt 1 (1 + vtκt) 10: if Kt 1/α then 11: return t 12: end if 13: vt+1 ONS step 14: end for Algorithm 2 Level-α X-SKIT for concept j Input: Observation z, subset S [m] \ {j}. 1: K0 1 2: Initialize ONS strategy (Algorithm A.1) 3: for t = 1, . . . do 4: Compute ρt as in Eq. (5) 5: Sample e H(t) S {j} PH|ZS {j}=z S {j} 6: Sample e H(t) S PH|ZS=z S 7: ˆY (t) S {j} g( e H(t) S {j}), ˆY (t) S g( e H(t) S ) 8: κt tanh(ρt( ˆY (t) S {j}) ρt( ˆY (t) S )) 9: Kt Kt 1 (1 + vtκt) 10: if Kt 1/α then 11: return t 12: end if 13: vt+1 ONS step 14: end for 3.3 Testing Global Semantic Importance with SKIT Podkopaev et al. [51] show how to design sequential kernelized tests of independence (i.e., H0 : A B) by framing them as particular two-sample tests of the form H0 : P = e P, with P = PAB and e P = PA PB. Similarly to [58, 62], they propose to leverage a simple yet powerful observation about the symmetry of the data under H0 [51, Section 4]. We state here the main result we will use in this paper (the proof is included in Appendix A.2). Lemma 2 (See [51, 58, 62]). t 1, let X P and e X e P, and let ρt : X R be any fixed real-valued function on X. Then, κt = tanh(ρt(X) ρt( e X)) provides a fair game for H0 : P = e P. That is, Lemma 2 prescribes how to construct valid payoffs for two-samples tests and, consequently, tests of independence. We note that the choice of tanh provides κt [ 1, 1], but any arbitrary anti-symmetric function can be used (e.g., sign). Furthermore, any fixed function ρt is valid but, in general, this function should have a positive value under the alternative in order for the bettor to increase their wealth and the testing procedure to have good power. Going back to the problem studied in this work, note that the global semantic importance null hypothesis HG 0,j in Definition 1 can be directly rewritten as a two-sample test, i.e. HG 0,j : ˆY Zj is equivalent to HG 0,j : P ˆY Zj = P ˆY PZj. We follow [51] and use the maximum mean discrepancy (MMD) [29] to measure the distance between the joint and the product of marginals. In particular, let R ˆY , RZj be two reproducing kernel Hilbert spaces (RKHSs) on the domains of ˆY and Zj, respectively (recall that ˆY and Zj are univariate). Then, ρSKIT t is the plug-in estimate of the witness function of MMD(P ˆY Zj, P ˆY PZj) at time t.2 We include the SKIT algorithm and technical details on computing ρSKIT t and k SKIT t in Appendix B.1. Computational complexity of SKIT. Analogous to the original presentation in Shekhar and Ramdas [62], the computational complexity of Algorithm B.1 is O(τ 2), where τ is the random rejection time. We now move on to presenting two novel testing procedures: the conditional randomization SKIT (C-SKIT) for HGC 0,j, and the explanation randomization SKIT (X-SKIT) for HLC 0,j,S. 3.4 Testing Global Conditional Semantic Importance with C-SKIT Analogous to the discussion in the previous section, we rephrase the global conditional null hypothesis HGC 0,j in Definition 2 as a two sample test HGC 0,j : P ˆY Zj Z j = P ˆY e Zj Z j, e Zj PZj|Z j. In contrast with other kernel-based notions of distance between conditional distributions [49, 63, 66] and akin to the CRT [11] we assume we can sample from PZj|Z j, which allows us to directly estimate 2Recall that MMD(PAB, PA PB) is the Hilbert-Schmidt Independence Criterion (HSIC) [28]. MMD(P ˆY Zj Z j, P ˆY e Zj Z j) in our testing procedure (we will expand on how to sample from this distribution shortly). Let R ˆY , RZj, RZ j be three RKHSs on the domains of ˆY , Zj, and Z j (i.e., R, R, Rm 1, where m is the number of concepts). Then, at time t, the C-SKIT payoff function is ρC-SKIT t := ˆµ(t 1) ˆY Zj Z j ˆµ(t 1) ˆY e Zj Z j, (4) where ˆµ(t 1) ˆY Zj Z j, ˆµ(t 1) ˆY e Zj Z j are the mean embeddings of their respective distributions in R ˆY RZj RZ j, and is the tensor product (see Appendix B.2 for technical details). Algorithm 1 summarizes the C-SKIT procedure, which provides Type I error control for HGC 0,j, as we briefly state in the following proposition (see Appendix C.2 for the proof). Proposition 1. t 1, let ( ˆY , Zj, Z j) P ˆY Zj Z j and ( ˆY , e Zj, Z j) P ˆY e Zj Z j, e Zj PZj|Z j. Then, κt := tanh(ρC-SKIT t ( ˆY , Zj, Z j) ρC-SKIT t ( ˆY , e Zj, Z j)) provides a fair game for HGC 0,j. Computational complexity of C-SKIT. First note that Z j is an (m 1)-dimensional vector (where m is the number of concepts). So, at each step of the test, the evaluation of the kernel associated with RZ j requires an additional sum over O(m) terms. Furthermore, C-SKIT needs access to samples from PZ|Z j, and we conclude that the computational complexity of Algorithm 1 is O(Tnmτ 2), where Tn represents the cost of the sampler on n samples, and it depends on implementation. For example, in the following, we will use non-parametric samplers with Tn = O(n2). Other choices of samplers, such as variational-autoencoders, may have constant cost (e.g., they are trained once and only used for inference). 3.5 Testing Local Conditional Semantic Importance with X-SKIT Attentive readers will have noticed that the local conditional semantic null hypothesis HLC 0,j,S in Definition 3 is already a two-sample test where the test statistic P is the distribution of the response of the model with the observed amount of concept j (i.e., ˆYS {j} = g( e HS {j})), and the null distribution e P without (i.e., ˆYS = g( e HS)). Herein, we assume we can sample from e HC PH|ZC=z C for any subset C [m], i.e. the conditional distribution of dense embeddings with specific concepts, which we will address via nonparametric methods. Then, for an RKHS R ˆY , the X-SKIT payoff function is ρX-SKIT t := ˆµ(t 1) ˆYS {j} ˆµ(t 1) ˆYS (5) with ˆµ(t 1) ˆYS {j}, ˆµ(t 1) ˆYS mean embeddings of the distributions in R ˆY . That is, ρX-SKIT t is the plug-in estimate of the witness function of MMD( ˆYS {j}, ˆYS) technical details are in Appendix B.3. Then, the X-SKIT testing procedure, which is summarized in Algorithm 2, provides Type I error control for HLC 0,j,S, as the following proposition summarizes (the proof is included in Appendix C.3). Proposition 2. t 1, κt := tanh(ρX-SKIT t ( ˆYS {j}) ρX-SKIT t ( ˆYS)) provides a fair game for HLC 0,j,S. Computational complexity of X-SKIT. Note that Algorithm 2 assumes access to a sampler PH|ZC=z C, so its computational complexity is O(Tnτ 2), where, similarly to above, Tn is the cost of the sampler. We briefly remark that, for the nonparametric samplers used in this work, Tn = n2 (compared to τn2 for C-SKIT) because we only need to estimate one conditional distribution. So far, we have presented our tests for one concept at a time, but we are interested in testing m 1 concepts. In this setting, it is well-known that multiple hypothesis testing requires appropriate corrections to avoid inflated significance levels. We use a result of Wang and Ramdas [79] and devise a greedy post-processor that guarantees false discovery rate control [3] (see Appendix A.4). First, we verify that our tests are valid and that they are adaptive to the hardness of their null hypotheses on two synthetic experiments in Appendix D. Here, we showcase the flexibility and effectiveness of our framework on zero-short image classification across several VL models on three Table 1: Summary of results for each dataset. Metrics are reported as average across all VL models used in the experiments. See main text for details about the models and the metrics used. Method Original model Imagenette Aw A2 CUB Accuracy Rank agreement Accuracy Rank agreement f1 Accuracy Rank agreement f1 SKIT 98.99% 0.51 99.50% 0.50 0.65 89.52% 0.82 0.93 C-SKIT 0.54 0.46 0.57 - - X-SKIT 0.59 - - - - PCBM 95.85% 0.45 95.11% 0.36 0.53 - - - 0.0 0.5 1.0 big (p) black (p) white (p) paws (p) furry (p) vegetation (p) flippers (a) yellow (a) patches (p) hairless (a) old world (p) walks (p) plankton (a) blue (a) bulbous (p) desert (a) hooves (a) horns (a) red (a) flies (a) giant panda 0.0 0.5 1.0 fierce (p) hunter (p) strong (p) stripes (p) muscle (p) bipedal (a) big (p) stalker (p) strain teeth (a) meat (p) small (a) old world (p) meat teeth (p) tusks (a) flippers (a) hops (a) skimmer (a) flies (a) tunnels (a) plankton (a) rejection rate significance level rejection time (a) Global importance with SKIT. 0.0 0.5 1.0 plankton (a) vegetation (p) white (p) blue (a) big (p) furry (p) flies (a) paws (p) hairless (a) old world (p) red (a) hooves (a) yellow (a) bulbous (p) black (p) horns (a) patches (p) flippers (a) walks (p) desert (a) giant panda 0.0 0.5 1.0 fierce (p) stripes (p) hunter (p) strong (p) bipedal (a) meat (p) muscle (p) skimmer (a) meat teeth (p) plankton (a) tunnels (a) flies (a) flippers (a) stalker (p) big (p) small (a) strain teeth (a) old world (p) tusks (a) hops (a) rejection rate significance level rejection time (b) Global conditional importance with C-SKIT. Figure 2: Importance results with CLIP:Vi T-L/14 on 2 classes in the Aw A2 dataset. Concepts are annotated with (p) if they are present in the class, or with (a) otherwise. real-world datasets: Animal with Attributes 2 (Aw A2) [82], CUB-200-2011 (CUB) [77], and the Imagenette subset of Image Net [22].3. We compare performance and transferability of the ranks of importance provided by each method across 8 VL models (see Appendix E for details) and, for all experiments, f is the image encoder of the model and g is the (linear) zero-shot classifier constructed by encoding A photo of a with the text encoder. Herein, we will always use RBF kernels to compute payoffs, and we repeat each test 100 times on independent draws of τ max samples to estimate each concept s expected rejection time and expected rejection rate at a significance level of α = 0.05 with the FDR post-processor described in Appendix A.4. That is, a (normalized) rejection time of 1 means failing to reject in τ max steps. Finally, recall that C-SKIT and X-SKIT need access to PZj|Z j and PH|ZC=z C, and that these distributions are not known in general. Since m is small, we use nonparametric methods to estimate them (see Appendix E.1). Table 1 summarizes the results of all experiments, which we now present and discuss individually. 4.1 Aw A2 Dataset Given the presence of global (i.e., class-level) annotations, we use SKIT and C-SKIT to test the global (and global conditional) semantic importance structure of the predictions for the top-10 best classified animal categories across all models (we describe the dataset, the concepts used, and the hyperparameters of the tests in Appendix E.2). We classify the top-10 concepts reported by each method as important, and we compute the f1 score with the ground-truth annotations. We briefly remark that this choice is informed by the fact that most concepts have rejection rates larger than the significance level of α. When comparing with PCBM since we use different concepts for each class we train 100 independent linear models for each class, and we rank concepts based on their average absolute weights (instead of signed ones) because the null hypotheses presented in this work are two-sided, i.e. a concept is important both if it increases the prediction for a class or if it decreases it. Table 1 shows that both SKIT and C-SKIT outperform PCBM across all three metrics, with SKIT providing the best average rank agreement across different models and importance f1 score (0.50 and 0.65, respectively). The fact that ranks provided by our tests have higher average agreement compared to PCBM suggests that VL models may share a similar semantic independence structure notwithstanding their embedding size or training strategy, i.e. semantic importance may be transferable across models (all individual pairwise agreements are included in Fig. E.4). Finally, Fig. 2 shows ranks of importance with CLIP:Vi T-L/14 on 2 animal categories (see Figs. E.5 and E.6 for all classes). In general, concepts are globally important (rejection rates are greater than α), 3Code to reproduce all experiments is available at https://github.com/Sulam-Group/IBYDMT. 0.0 0.5 1.0 blue forehead (p) blue crown (p) blue under tail (p) grey breast (p) grey underparts (p) multi-colored breast (p) iridescent underparts (a) pink under tail (a) red bill (a) white bill (a) pink nape (a) is green (a) is iridescent (a) small (5 - 9 in) (p) frigatebird 0.0 0.5 1.0 hooked seabird bill (p) forked tail tail (p) solid wing (p) black back (p) is black (p) black upper tail (p) solid tail (p) brown wing (a) green nape (a) orange belly (a) all-purpose bill (a) striped breast (a) pointed tail tail (a) red underparts (a) 0.0 0.5 1.0 spatulate bill (p) duck-like (p) olive throat (p) yellow bill (p) medium (9 - 16 in) (p) bill about the same as head (p) multi-colored wing (p) orange nape (a) red breast (a) masked head (a) buff leg (a) long-legged-like (a) rounded tail tail (a) red underparts (a) cape glossy starling 0.0 0.5 1.0 blue underparts (p) blue upperparts (p) black underparts (p) black breast (p) yellow eye (p) plain head (p) black bill (p) swallow-like (a) iridescent breast (a) brown eye (a) pink eye (a) green leg (a) duck-like (a) pigeon-like (a) rejection rate rejection time Figure 3: Importance results with X-SKIT and CLIP:Vi T-L/14 on 4 images in the CUB dataset. Concepts with (p) are present in the image according to human annotations, and (a) otherwise. and it is harder to reject the global conditional null hypothesis (rejection rates are lower and rejection times larger), naturally reflecting the fact that conditional independence is a stronger condition. 4.2 CUB Dataset This dataset (differently from Aw A2) provides per-image annotations of semantic attributes. So, we use X-SKIT to test the semantic importance structure of VL models locally on particular images and validate its performance against such annotations (we include details about this experiment and extended results in Appendix E.3). The purpose of this experiment is to validate the performance of X-SKIT, hence we use the ground-truth binary semantic annotations as an oracle instead of predicting the presence of concepts. In practical scenarios where ground-truth is not available, one could as done by previous works [12] use LLMs to answer binary questions (e.g., Does this bird have an orange bill? Yes/No ). Furthermore, note that for each concept j [m] there are exponentially many tests with null hypothesis HLC 0,j,S one for each subset S [m] \ {j} which are intractable to compute. Thus, we report average results over 100 tests with random subsets with fixed size s. Fig. 3 depicts prototypical results with CLIP:Vi T-L/14 (Fig. E.9 includes results for all models on the same images). After running X-SKIT, we classify concepts as important by thresholding their rejection rates at level α which is a statistically-valid way of selecting important concepts. Results are included in Table 1, and we conclude that X-SKIT provides ranks of importance that are well-aligned both across models (0.82 rank agreement) and with ground-truth annotations (f1 score of 0.93). We remark that X-SKIT is the first method to provide local semantic explanations, hence why we cannot compare with alternatives. 4.3 Imagenette Dataset Lastly, we use both SKIT, C-SKIT, and X-SKIT on the Imagenette subset of Image Net [22] which does not provide ground-truth semantic annotations to evaluate performance with. So, we use Sp Li Ce [6] to select which concepts to test (see Appendix E.4 for details), but we stress that any user-defined set of concepts would be valid a unique feature of our proposed framework. Figs. 4a and 4b show SKIT and C-SKIT results with CLIP:Vi T-L/14 on 2 classes in the dataset (Fig. E.11 includes all classes). We use Sp Li Ce to encode the entire dataset and test the top-20 concepts. Analogous to the experiment on Aw A2, we can see that rejection rates are lower for C-SKIT (i.e., conditional dependence) compared to SKIT (i.e., marginal dependence). We evaluate rank agreement across all models and compare with PCBM in Table 1. These results confirm that not only are the ranks produced by our tests more transferable across models (rank agreement of 0.51 for SKIT, 0.54 or C-SKIT, 0.59 for X-SKIT, and 0.45 for PCBM), but also they retain the performance of the original classifier (98.99% for our methods vs 95.55% for PCBM). We refer interested readers to Fig. E.12 for all pairwise comparisons. Furthermore, we qualitatively study the stability of our tests as a function of τ max in Fig. E.13. This is important because τ max represents the sample complexity of the tests. Our findings indicate that important concepts tend to exhibit greater stability in their ranks compared to less important ones, with SKIT showing overall more stability than C-SKIT. To conclude, Fig. 4c shows X-SKIT results with CLIP:Vi T-L/14 on three random images from the dataset (see Figs. E.15 and E.16 for all models and more images). We use Sp Li Ce to encode each image and obtain its top-10 concepts, and then add the bottom-4 according to PCBM, for a total of 14 attributes per image. The choice of combining concepts both from Sp Li Ce and PCBM will highlight 0.0 0.5 1.0 spaniel fore selling instrument airshow radio jazz putting battered obsolete fish fishing dispenser bass flew trumpet exterior cathedral band brass English springer rejection rate significance level rejection time 0.0 0.5 1.0 battered fishing fish spaniel bass brass obsolete instrument flew exterior putting cathedral airshow band selling radio fore dispenser trumpet jazz (a) Global importance with SKIT. 0.0 0.5 1.0 fish bass brass fishing band jazz battered instrument trumpet obsolete spaniel dispenser exterior fore cathedral radio flew airshow putting selling 0.0 0.5 1.0 spaniel band brass bass jazz instrument trumpet obsolete fish dispenser fishing fore exterior battered radio flew airshow putting cathedral selling English springer rejection rate significance level rejection time (b) Global conditional importance with C-SKIT. 0.0 0.5 1.0 spaniel fetch cane stick swimming helping fishing (*) help fore (*) rescue launched barge trumpet (*) cathedral (*) English springer rejection rate significance level rejection time 0.0 0.5 1.0 trumpet band instrument conducted obsolete (*) naval cuff chop cathedral (*) sailor bass (*) major fish (*) polishing French horn rejection rate significance level rejection time 0.0 0.5 1.0 launch departs tandem flew instructor slipping descending england fishing (*) uk band (*) cathedral (*) instrument (*) textbook rejection rate significance level rejection time (c) Local importance with X-SKIT. The bottom-4 concepts according to PCBM are annotated with (*). Figure 4: Results with CLIP:Vi T-L/14 on Imagenette. the differences between these methods and our notion of local statistical importance, as we will shortly expand on. Recall that we use nonparametric samplers to approximate PH|ZC=z C, so the cost of using image-specific concepts boils down to projecting the feature embeddings with a different matrix c, which is negligible compared to running the tests. We note that parametric generative models such as variational autoencoders or diffusion models would have required retraining for each set of concepts, which is expensive. Overall, we find that ranks are well-aligned across models (0.71 rank agreement, see Table 1). We can appreciate how the bottom-4 concepts from PCBM, which are annotated with an asterisk, are not always last, i.e. a concept may be locally important even if it is not globally important. For example, concept fishing may not be globally important for class English springer , but it is locally important for an image of a dog in water. Conversely, a concept having a high weight according to Sp Li Ce does not imply it will be statistically important for the predictions of the model, and these distinctions are important in order to communicate findings transparently. 5 Conclusions There exist an increasing interest in explaining modern, unintelligible predictors and, in particular, doing so with inherently interpretable concepts that convey specific meaning to users. This work is the first to formalize precise statistical notions of semantic importance in terms of global (i.e., over a population) and local (i.e., on a sample) conditional hypothesis testing. We propose novel, valid tests for each notion of importance while providing a rank of importance by deploying ideas of sequential testing. Importantly, by approaching importance via conditional independence (and by developing appropriate valid tests), we are able to provide Type I error and FDR control, a feature that is unique to our framework compared to existing alternatives. Furthermore, our tests allow to explain the original potentially nonlinear classifier that would be used in the wild, as opposed to training surrogate linear models as has been the standard so far. Naturally, our work has limitations. First and foremost, the procedures introduced in this work require access to samplers, and there might be settings were learning these models is hard; we used nonparametric estimators in our experiments, but modern generative models could be employed, too. Second, kernel-based tests rely on the assumption that the kernels used are characteristic for the space of distributions considered. Although these assumptions are usually satisfied in Rd for RBF kernels, there may exist data modalities where this is not true (e.g., discrete data, graphs), which would compromise the power of the test. Finally, although we grant full flexibility to users to specify the concepts they care about, there is no guarantee that these are well-represented in the feature space of the model, nor that they are the most informative ones for a specific task. All these points are a matter of ongoing and future work. Acknowledgments We sincerely thank Zhenzhen Wang and the anonymous Neur IPS reviewers for useful conversations that strengthened the presentation of our experimental results. This research was supported by NSF CAREER Award CCF 2239787. [1] Nachman Aronszajn. Theory of reproducing kernels. Transactions of the American mathematical society, 68(3):337 404, 1950. [2] David Bau, Bolei Zhou, Aditya Khosla, Aude Oliva, and Antonio Torralba. Network dissection: Quantifying interpretability of deep visual representations. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 6541 6549, 2017. [3] Yoav Benjamini and Yosef Hochberg. Controlling the false discovery rate: a practical and powerful approach to multiple testing. Journal of the Royal statistical society: series B (Methodological), 57(1):289 300, 1995. [4] Alain Berlinet and Christine Thomas-Agnan. Reproducing kernel Hilbert spaces in probability and statistics. Springer Science & Business Media, 2011. [5] Thomas B Berrett, Yi Wang, Rina Foygel Barber, and Richard J Samworth. The conditional permutation test for independence while controlling for confounders. Journal of the Royal Statistical Society Series B: Statistical Methodology, 82(1):175 197, 2020. [6] Usha Bhalla, Alex Oesterling, Suraj Srinivas, Flavio P Calmon, and Himabindu Lakkaraju. Interpreting clip with sparse linear concept embeddings (splice). ar Xiv preprint ar Xiv:2402.10376, 2024. [7] Beepul Bharti, Paul Yi, and Jeremias Sulam. Sufficient and necessary explanations (and what lies in between). ar Xiv preprint ar Xiv:2409.20427, 2024. [8] Gilles Blanchard and Etienne Roquain. Two simple sufficient conditions for fdr control. Electronic Journal of Statistics, 2:963 992, 2008. [9] Carlo Bonferroni. Teoria statistica delle classi e calcolo delle probabilita. Pubblicazioni del R istituto superiore di scienze economiche e commericiali di firenze, 8:3 62, 1936. [10] Collin Burns, Jesse Thomason, and Wesley Tansey. Interpreting black box models via hypothesis testing. In Proceedings of the 2020 ACM-IMS on foundations of data science conference, pages 47 57, 2020. [11] Emmanuel Candes, Yingying Fan, Lucas Janson, and Jinchi Lv. Panning for gold: modelx knockoffs for high dimensional controlled variable selection. Journal of the Royal Statistical Society Series B: Statistical Methodology, 80(3):551 577, 2018. [12] Aditya Chattopadhyay, Kwan Ho Ryan Chan, and Rene Vidal. Bootstrapping variational information pursuit with large language and vision models for interpretable image classification. In The Twelfth International Conference on Learning Representations, 2024. [13] Aditya Chattopadhyay, Ryan Pilgrim, and Rene Vidal. Information maximization perspective of orthogonal matching pursuit with applications to explainable ai. Advances in Neural Information Processing Systems, 36, 2024. [14] Haozhe Chen, Junfeng Yang, Carl Vondrick, and Chengzhi Mao. Invite: Interpret and control vision-language models with text explanations. In The Twelfth International Conference on Learning Representations, 2024. [15] Jianbo Chen, Le Song, Martin J Wainwright, and Michael I Jordan. L-shapley and c-shapley: Efficient model interpretation for structured data. ar Xiv preprint ar Xiv:1808.02610, 2018. [16] Mehdi Cherti, Romain Beaumont, Ross Wightman, Mitchell Wortsman, Gabriel Ilharco, Cade Gordon, Christoph Schuhmann, Ludwig Schmidt, and Jenia Jitsev. Reproducible scaling laws for contrastive language-image learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2818 2829, 2023. [17] Thomas M Cover. Universal portfolios. Mathematical finance, 1(1):1 29, 1991. [18] Ian Covert, Scott M Lundberg, and Su-In Lee. Understanding global feature contributions with additive importance measures. Advances in Neural Information Processing Systems, 33: 17212 17223, 2020. [19] Ashok Cutkosky and Francesco Orabona. Black-box reductions for parameter-free online learning in banach spaces. In Conference On Learning Theory, pages 1493 1529. PMLR, 2018. [20] Roxana Daneshjou, Mert Yuksekgonul, Zhuo Ran Cai, Roberto Novoa, and James Y Zou. Skincon: A skin disease dataset densely annotated by domain experts for fine-grained debugging and analysis. Advances in Neural Information Processing Systems, 35:18157 18167, 2022. [21] Marina Danilevsky, Kun Qian, Ranit Aharonov, Yannis Katsis, Ban Kawas, and Prithviraj Sen. A survey of the state of explainable ai for natural language processing. ar Xiv preprint ar Xiv:2010.00711, 2020. [22] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A largescale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248 255. Ieee, 2009. [23] Christiane Fellbaum. Word Net: An electronic lexical database. MIT press, 1998. [24] Lasse Fischer and Aaditya Ramdas. Sequential permutation testing by betting. ar Xiv preprint ar Xiv:2401.07365, 2024. [25] Kenji Fukumizu, Arthur Gretton, Xiaohai Sun, and Bernhard Schölkopf. Kernel measures of conditional dependence. Advances in neural information processing systems, 20, 2007. [26] Yossi Gandelsman, Alexei A Efros, and Jacob Steinhardt. Interpreting clip s image representation via text-based decomposition. ar Xiv preprint ar Xiv:2310.05916, 2023. [27] Damien Garreau, Wittawat Jitkrittum, and Motonobu Kanagawa. Large sample analysis of the median heuristic. ar Xiv preprint ar Xiv:1707.07269, 2017. [28] Arthur Gretton, Kenji Fukumizu, Choon Teo, Le Song, Bernhard Schölkopf, and Alex Smola. A kernel statistical test of independence. Advances in neural information processing systems, 20, 2007. [29] Arthur Gretton, Karsten M Borgwardt, Malte J Rasch, Bernhard Schölkopf, and Alexander Smola. A kernel two-sample test. The Journal of Machine Learning Research, 13(1):723 773, 2012. [30] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770 778, 2016. [31] Gabriel Ilharco, Mitchell Wortsman, Ross Wightman, Cade Gordon, Nicholas Carlini, Rohan Taori, Achal Dave, Vaishaal Shankar, Hongseok Namkoong, John Miller, Hannaneh Hajishirzi, Ali Farhadi, and Ludwig Schmidt. Openclip, July 2021. URL https://doi.org/10.5281/ zenodo.5143773. If you use this software, please cite it as below. [32] Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc Le, Yun-Hsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. In International conference on machine learning, pages 4904 4916. PMLR, 2021. [33] Maurice G Kendall. A new measure of rank correlation. Biometrika, 30(1-2):81 93, 1938. [34] Been Kim, Martin Wattenberg, Justin Gilmer, Carrie Cai, James Wexler, Fernanda Viegas, et al. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In International conference on machine learning, pages 2668 2677. PMLR, 2018. [35] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. [36] Pang Wei Koh, Thao Nguyen, Yew Siang Tang, Stephen Mussmann, Emma Pierson, Been Kim, and Percy Liang. Concept bottleneck models. In International conference on machine learning, pages 5338 5348. PMLR, 2020. [37] Stefan Kolek, Aditya Chattopadhyay, Kwan Ho Ryan Chan, Hector Andrade-Loarca, Gitta Kutyniok, and René Vidal. Learning interpretable queries for explainable image classification with information pursuit. ar Xiv preprint ar Xiv:2312.11548, 2023. [38] Yann Le Cun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. [39] Junnan Li, Dongxu Li, Caiming Xiong, and Steven Hoi. Blip: Bootstrapping languageimage pre-training for unified vision-language understanding and generation. In International conference on machine learning, pages 12888 12900. PMLR, 2022. [40] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In Computer Vision ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pages 740 755. Springer, 2014. [41] Pantelis Linardatos, Vasilis Papastefanopoulos, and Sotiris Kotsiantis. Explainable ai: A review of machine learning interpretability methods. Entropy, 23(1):18, 2020. [42] Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions. Advances in neural information processing systems, 30, 2017. [43] Jan Macdonald, Stephan Wäldchen, Sascha Hauch, and Gitta Kutyniok. A rate-distortion framework for explaining neural network decisions. ar Xiv preprint ar Xiv:1905.11092, 2019. [44] Ricardo Moreira, Jacopo Bono, Mário Cardoso, Pedro Saleiro, Mário AT Figueiredo, and Pedro Bizarro. Diconstruct: Causal concept-based explanations through black-box distillation. ar Xiv preprint ar Xiv:2401.08534, 2024. [45] Alfred Müller. Integral probability metrics and their generating classes of functions. Advances in applied probability, 29(2):429 443, 1997. [46] Meike Nauta, Ron Van Bree, and Christin Seifert. Neural prototype trees for interpretable fine-grained image recognition. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 14933 14943, 2021. [47] Tuomas Oikarinen and Tsui-Wei Weng. Clip-dissect: Automatic description of neuron representations in deep vision networks. ar Xiv preprint ar Xiv:2204.10965, 2022. [48] Tuomas Oikarinen, Subhro Das, Lam M Nguyen, and Tsui-Wei Weng. Label-free concept bottleneck models. ar Xiv preprint ar Xiv:2304.06129, 2023. [49] Junhyung Park and Krikamol Muandet. A measure-theoretic approach to kernel conditional mean embeddings. Advances in neural information processing systems, 33:21247 21259, 2020. [50] Thang M Pham, Peijie Chen, Tin Nguyen, Seunghyun Yoon, Trung Bui, and Anh Totti Nguyen. Peeb: Part-based image classifiers with an explainable and editable language bottleneck. ar Xiv preprint ar Xiv:2403.05297, 2024. [51] Aleksandr Podkopaev, Patrick Blöbaum, Shiva Kasiviswanathan, and Aaditya Ramdas. Sequential kernelized independence testing. In International Conference on Machine Learning, pages 27957 27993. PMLR, 2023. [52] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748 8763. PMLR, 2021. [53] Vikram V Ramaswamy, Sunnie SY Kim, Ruth Fong, and Olga Russakovsky. Overlooked factors in concept-based explanations: Dataset choice, concept learnability, and human capability. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10932 10941, 2023. [54] Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. " why should i trust you?" explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pages 1135 1144, 2016. [55] Numair Sani, Daniel Malinsky, and Ilya Shpitser. Explaining the behavior of black-box prediction algorithms with causal learning. ar Xiv preprint ar Xiv:2006.02482, 2020. [56] David W Scott. Multivariate density estimation: theory, practice, and visualization. John Wiley & Sons, 2015. [57] Ramprasaath R Selvaraju, Michael Cogswell, Abhishek Das, Ramakrishna Vedantam, Devi Parikh, and Dhruv Batra. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision, pages 618 626, 2017. [58] Shalev Shaer, Gal Maman, and Yaniv Romano. Model-x sequential testing for conditional independence via testing by betting. In International Conference on Artificial Intelligence and Statistics, pages 2054 2086. PMLR, 2023. [59] Glenn Shafer. Testing by betting: A strategy for statistical and scientific communication. Journal of the Royal Statistical Society Series A: Statistics in Society, 184(2):407 431, 2021. [60] Glenn Shafer and Vladimir Vovk. Game-theoretic foundations for probability and finance, volume 455. John Wiley & Sons, 2019. [61] John Shawe-Taylor and Nello Cristianini. Kernel methods for pattern analysis. Cambridge university press, 2004. [62] Shubhanshu Shekhar and Aaditya Ramdas. Nonparametric two-sample testing by betting. IEEE Transactions on Information Theory, 2023. [63] Tianhong Sheng and Bharath K Sriperumbudur. On distance and kernel measures of conditional dependence. Journal of Machine Learning Research, 24(7):1 16, 2023. [64] Pushkar Shukla, Sushil Bharati, and Matthew Turk. Cavli-using image associations to produce local concept-based explanations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3749 3754, 2023. [65] Amanpreet Singh, Ronghang Hu, Vedanuj Goswami, Guillaume Couairon, Wojciech Galuba, Marcus Rohrbach, and Douwe Kiela. Flava: A foundational language and vision alignment model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15638 15650, 2022. [66] Le Song, Jonathan Huang, Alex Smola, and Kenji Fukumizu. Hilbert space embeddings of conditional distributions with applications to dynamical systems. In Proceedings of the 26th Annual International Conference on Machine Learning, pages 961 968, 2009. [67] Yang Song, Jascha Sohl-Dickstein, Diederik P Kingma, Abhishek Kumar, Stefano Ermon, and Ben Poole. Score-based generative modeling through stochastic differential equations. ar Xiv preprint ar Xiv:2011.13456, 2020. [68] Bharath K Sriperumbudur, Arthur Gretton, Kenji Fukumizu, Bernhard Schölkopf, and Gert RG Lanckriet. Hilbert space embeddings and metrics on probability measures. The Journal of Machine Learning Research, 11:1517 1561, 2010. [69] Wesley Tansey, Victor Veitch, Haoran Zhang, Raul Rabadan, and David M Blei. The holdout randomization test for feature selection in black box models. Journal of Computational and Graphical Statistics, 31(1):151 162, 2022. [70] Jacopo Teneggi, Alexandre Luster, and Jeremias Sulam. Fast hierarchical games for image explanations. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(4):4494 4503, 2022. [71] Jacopo Teneggi, Beepul Bharti, Yaniv Romano, and Jeremias Sulam. Shap-xrt: The shapley value meets conditional independence testing. Transactions on Machine Learning Research, 2023. [72] Isabella Verdinelli and Larry Wasserman. Feature importance: A closer look at shapley values and loco. ar Xiv preprint ar Xiv:2303.05981, 2023. [73] Isabella Verdinelli and Larry Wasserman. Decorrelated variable importance. Journal of Machine Learning Research, 25(7):1 27, 2024. [74] Sebastiano Vigna. A weighted correlation index for rankings with ties. In Proceedings of the 24th international conference on World Wide Web, pages 1166 1176, 2015. [75] Jean Ville. Etude critique de la notion de collectif. Gauthier-Villars Paris, 1939. [76] Vladimir Vovk and Ruodu Wang. E-values: Calibration, combination and applications. The Annals of Statistics, 49(3):1736 1754, 2021. [77] Catherine Wah, Steve Branson, Peter Welinder, Pietro Perona, and Serge Belongie. The caltech-ucsd birds-200-2011 dataset. 2011. [78] Benyou Wang, Qianqian Xie, Jiahuan Pei, Zhihong Chen, Prayag Tiwari, Zhao Li, and Jie Fu. Pre-trained language models in biomedical domain: A systematic survey. ACM Computing Surveys, 56(3):1 52, 2023. [79] Ruodu Wang and Aaditya Ramdas. False discovery rate control with e-values. Journal of the Royal Statistical Society Series B: Statistical Methodology, 84(3):822 852, 2022. [80] Zifeng Wang, Zhenbang Wu, Dinesh Agarwal, and Jimeng Sun. Medclip: Contrastive learning from unpaired medical images and text. ar Xiv preprint ar Xiv:2210.10163, 2022. [81] Shirley Wu, Mert Yuksekgonul, Linjun Zhang, and James Zou. Discover and cure: Conceptaware mitigation of spurious correlation. In International Conference on Machine Learning, pages 37765 37786. PMLR, 2023. [82] Yongqin Xian, Christoph H Lampert, Bernt Schiele, and Zeynep Akata. Zero-shot learning a comprehensive evaluation of the good, the bad and the ugly. IEEE transactions on pattern analysis and machine intelligence, 41(9):2251 2265, 2018. [83] Yue Yang, Artemis Panagopoulou, Shenghao Zhou, Daniel Jin, Chris Callison-Burch, and Mark Yatskar. Language in a bottle: Language model guided concept bottlenecks for interpretable image classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 19187 19197, 2023. [84] Hao Yuan, Haiyang Yu, Shurui Gui, and Shuiwang Ji. Explainability in graph neural networks: A taxonomic survey. IEEE transactions on pattern analysis and machine intelligence, 45(5): 5782 5799, 2022. [85] Mert Yuksekgonul, Maggie Wang, and James Zou. Post-hoc concept bottleneck models. ar Xiv preprint ar Xiv:2205.15480, 2022. [86] Kun Zhang, Jonas Peters, Dominik Janzing, and Bernhard Schölkopf. Kernel-based conditional independence test and application in causal discovery. ar Xiv preprint ar Xiv:1202.3775, 2012. A Testing by Betting In this appendix, we include additional background information on testing by betting that was omitted from the main text for the sake of conciseness of presentation. Recall that the wealth process {Kt}t N0 with N0 := N {0} is defined as K0 = 1 and Kt = Kt 1 (1 + vtκt) (6) where vt [ 1, 1] is the betting fraction and κt [ 1, 1] the payoff of the bet. A.1 Test Martingales We start by introducing the definition of a test martingale (see, for example, Shaer et al. [58]). Definition A.1 (Test martingale). A nonnegative stochastic process {St}t N0 is a test martingale if S0 = 1 and, under a null hypothesis H0, it is a supermartingale, i.e. EH0[St | Ft 1] St 1, (7) where Ft 1 is the filtration (i.e., the smallest σ-algebra) of all previous observations. In the following, we will use Ville s inequality, which we include for the sake of completeness. Lemma A.1 (Ville s inequality [75]). If the stochastic process {St}t N0 is a nonnegative supermartingale, P[ t 0 : St η] E[S0]/η, η > 0. (8) With this, we state a condition under which we can use the wealth process to reject a null hypothesis H0 with Type I error control. Lemma A.2 (See Shaer et al. [58], Shekhar and Ramdas [62]). If EH0[κt | Ft 1] = 0, (9) where Ft 1 denotes the filtration (i.e., the smallest σ-algebra) of all previous observations, then the wealth process {Kt}t N0 describes a fair game and PH0[ t 1 : Kt 1/α] α. (10) Proof. It suffices to show that if EH0[κt | Ft 1] = 0, then the wealth process {Kt}t N0 is a test martingale: 1. K0 = 1 by definition, and 2. It is immediate to see that the wealth process is nonnegative because vtκt [ 1, 1] and the bettor never risks more than their current wealth, i.e. they will never go into debt. Finally, 3. If EH0[κt | Ft 1] = 0, then EH0[Kt | Ft 1] = EH0[Kt 1 (1 + vtκt) | Ft 1] (11) = Kt 1 EH0[1 + vtκt | Ft 1] (Kt 1 | Ft 1 is constant) (12) Kt 1 (1 + EH0[κt | Ft 1]) (vt 1) (13) = Kt 1, (14) and the wealth process is a supermartingale under the null. Then, by Ville s inequality, we conclude that for any significance level α (0, 1) PH0[ t 1 : Kt 1/α] αE[K0] = α (15) which is the statement of the lemma. A.2 Symmetry-based Two-sample Sequential Testing In this section, we show how to leverage symmetry to construct valid sequential tests for a two-sample null hypothesis of the form H0 : P = e P. (16) Lemma A.3 (See [51, 58, 62]). t 1, let X P and e X e P be two random variables sampled from P and e P, respectively. If P = e P, it holds that for any fixed function ρt : X R ρt(X) ρt( e X) d= ρt( e X) ρt(X), (17) that is p0(ρt(X) ρt( e X)) = p0(ρt( e X) ρt(X)), (18) where p0 is the probability density function induced by H0. Proof. The proof is straightforward. If P = e P, then X and e X are exchangeable by assumption. Proof of Lemma 2 Recall the lemma states that for any fixed function ρt : X R, the payoff κt = tanh(ρt(X) ρt( e X)) (19) provides a fair game (i.e., it satisfies Lemma A.2) for a two-sample test with null hypothesis H0 : P = e P. We use Lemma A.3 above to prove a stronger result that implies the desired claim. Lemma A.4 (See [51, 58, 62]). For any t 1, and any fixed anti-symmetric function ξ : R R, it holds that EH0[ξ(ρt(X) ρt( e X)) | Ft 1] = 0. (20) Proof. We can see that EH0[ξ(ρt(X) ρt( e X)) | Ft 1] = EH0[ξ(ρt(X) ρt( e X))] (ρt, ξ are fixed) (21) R ξ(u)p0(u) du (change of variables) (22) R+ (ξ(u) + ξ( u))p0(u) du (by Lemma A.3) (23) R+ (ξ(u) ξ(u))p0(u) du (ξ is anti-symmetric) (24) which concludes the proof. Proof of Lemma 2. Note that tanh is an anti-symmetric function, so Lemma A.4 holds. Then, Lemma A.2 implies that κt = tanh(ρt(X) ρt( e X)) provides a test martingale for H0 : P = e P. A.3 Betting Strategies So far, we have discussed how to construct valid test martingales in terms of the payoff κt. Then, it remains to define a strategy to choose the betting fraction vt. In general, any method that picks vt before data is revealed maintains validity of the test, and we briefly summarize a few alternatives. Constant betting fraction. Naturally, a fixed betting fraction vt = v is valid. However, this strategy may be prone to overshooting, i.e. the wealth may go to zero almost surely under the alternative, and severely impact the power of the test [51, Example 2]. Mixture method [17, 58]. A possible way to overcome the limitations of setting a fixed betting fraction is to average across a distribution, i.e. V K(v) t p(v) dv, (26) where K(v) t is the wealth with constant betting fraction vt = v, and p(v) is a prior over the choice of fractions (e.g., uniform over [ 1, 1]). This choice is valid, and motivated by the intuition that the mixture martingale will be driven by the term that achieves the optimal betting fraction [58, Theorem 1]. Online Newton step (ONS) [19]. Alternatively, one can frame choosing the betting fraction as an online optimization problem that finds the optimal vt in terms of the regret of the strategy. We refer interested readers to [19, 58, 62] for a theoretical analysis of this strategy and simply state here the wealth s growth rate. Algorithm A.1 summarizes this strategy. Lemma A.5 (See Shekhar and Ramdas [62]). For any sequence {vt [ 1, 1] : t 1}, it holds that log t. (27) Algorithm A.1 ONS Betting Strategy Input: Sequence of payoffs {κt}t 1 1: a0 1 2: v1 0 3: for t 1 do 4: zt κt/(1 + vtκt) 5: at at 1 + z2 t 6: vt+1 max(0, min(1, at + 2/(2 log(3)) zt/at)) 7: end for A.4 Controlling False Discovery Rate Finally, we briefly present one way to provide false discovery rate (FDR) control when testing multiple hypotheses with sequential tests. Given m null hypotheses H(1) 0 , . . . , H(m) 0 , denote e(1), . . . , e(m) their respective e-values [60, 76] and let E : [0, ]m 2[m] be an e-testing procedure such that e S = E(e(1), . . . , e(m)) is the set of rejected null hypotheses. Then, FDR is the expected proportion of false discoveries to the number of total findings, i.e. where S0 := {j [m] : H(j) 0 is true} is the set of true null hypotheses (i.e., the ones that should not be rejected). Following [8, 79], we say that E is self-consistent at level α if every rejected e-value satisfies e(j) m/α|e S|, and we now state the lemma we use to construct our FDR post-processor in Algorithm A.2. Lemma A.6 (See Wang and Ramdas [79]). Any self-consistent e-testing procedure at level α controls FDR at level α for arbitrary configurations of e-values. Algorithm A.2 Level-α greedy FDR post-processor. Input: Wealth processes {K(1) t }, . . . , {K(m) t }, t N0. 1: e S 2: for s = 1, . . . , m do 3: j , τ arg min j [m]\e S, τ [0, ] τ s.t. K(j) τ m/αs 4: if τ = then 5: return e S 6: end if 7: e S e S {j } 8: end for 9: return e S Recall that the optional stopping theorem implies that for a test martingale {Kt}t N0, the wealth Kt is an e-value for any t 1. Then, intuitively, Algorithm A.2 transforms an e-testing procedure E into a self-consistent one by greedily rejecting concepts as soon as they cross the adjusted threshold m/α|e S|. Note that we do not know the number of rejections a priori, and that m/α|e S| is a decreasing function of |e S|. Hence, the adjusted threshold for the first concept will be m/α (which matches the common Bonferroni correction [9]), m/2α for the second one, then m/3α, and so on and so forth. The procedure stops when no more concepts reach the threshold, and concepts are sorted by their adjusted rejection times. We remark that Algorithm A.2 runs in O(m) time, and that it does not change the individual testing procedures which is important because concepts are tested concurrently in practice. B Technical Details on Payoff Functions In this appendix, we include technical details on how to compute the payoff functions of all tests presented in this paper. We start with a brief overview of the maximum mean discrepancy (MMD) [29], and we refer interested readers to [1, 4, 61, 68] for rigorous introductions to the theory of reproducing kernel Hilbert spaces (RKHSs) and their applications to probability and statistics. Definition B.1 (Mean embedding (see Gretton et al. [29])). Let P be a probability distribution on X and R an RKHS on the same domain. The mean embedding of P in R is the element µP R with ρ R, EP [ρ(X)] = µP , ρ R. (29) Furthermore, given X(1), . . . , X(n) sampled i.i.d. from P, the plug-in estimate ˆµ(n) P is ˆµ(n) P := 1 i=1 φ(X(i)), (30) where φ is the canonical feature map, i.e. φ(X) = k(X, ), and k is the kernel associated with R. We now define the MMD between two probability distributions P, Q and show that it can be rewritten in terms of their mean embeddings. Definition B.2 (Integral probability metric (see Müller [45])). Let P, Q be two probability distributions over X. Furthermore, denote G = {g : X R} a hypothesis class of real-valued functions over X. Then, DG(P, Q) := sup g G |EP [g(X)] EQ[g(X)]| (31) is the distance between P and Q induced by G, and the function g that achieves the supremum is called witness function. The MMD is defined as DB(R)(P, Q), where B(R) is the unit ball of R, i.e. B(R) := {ρ R : ρ R 1}. (32) Definition B.3 (Maximum mean discrepancy (see Gretton et al. [29])). For P, Q defined as above, let R be an RKHS on their domain. Then, MMD(P, Q) := sup ρ B(R) EP [ρ(X)] EQ[ρ(X)]. (33) We note that we drop the absolute value because if ρ B(R), then ρ B(R) also. From the definition of mean embedding, it follows that MMD(P, Q) = sup ρ B(R) µP , ρ R µQ, ρ R (34) = sup ρ B(R) µP µQ, ρ (35) = µP µQ R, (36) and its witness function satisfies ρ µP µQ. (37) Algorithm B.1 Level-α SKIT for the global importance of concept j Input: Stream ( ˆY (t), Z(t) j ) P ˆY Zj. 1: K0 1 2: Initialize ONS strategy as in Algorithm A.1. 3: for t = 1, . . . do 4: Compute ρt as in Eq. (47) 5: Observe D(2t 1) = ( ˆY (2t 1), Z(2t 1) j ) and D(2t) = ( ˆY (2t), Z(2t) j ) 6: e D(2t 1) ( ˆY (2t 1), Z(2t) j ) and e D(2t) ( ˆY (2t), Z(2t 1) j ) 7: Compute κt as in Eq. (48) 8: Kt Kt 1 (1 + vtκt) 9: if Kt 1/α then 10: return t 11: end if 12: vt+1 ONS step 13: end for B.1 Computing ρSKIT t and κSKIT t Recall that ρSKIT t is the estimate of the witness function of MMD(P ˆY Zj, P ˆY PZj) at time t, i.e. ρSKIT t = ˆµ(2(t 1)) ˆY Zj ˆµ(2(t 1)) ˆY ˆµ(2(t 1)) Zj , (38) ˆµ(2(t 1)) ˆY Zj = 1 2(t 1) t =0 (φ ˆY ( ˆY (t )) φZj(Z(t ) j )), (39) ˆµ(2(t 1)) ˆY = 1 t 1 t =0 φ ˆY ( ˆY (t )), ˆµ(2(t 1)) Zj = 1 t 1 t =0 φZj(Z(t ) j ), (40) and φ ˆY , φZj are the canonical feature maps associated with R ˆY and RZj, respectively. We remark that ρSKIT t is an operator, and, for a sample (ˆy, zj), its value ρSKIT t (ˆy, zj) can be computed as ρSKIT t (ˆy, zj) = (ˆµ(2(t 1)) ˆY Zj ˆµ(2(t 1)) ˆY ˆµ(2(t 1)) Zj )(ˆy, zj) (41) = ˆµ(2(t 1)) ˆY Zj (ˆy, zj) (ˆµ(2(t 1)) ˆY ˆµ(2(t 1)) Zj )(ˆy, zj) (42) ˆµ(2(t 1)) ˆY Zj (ˆy, zj) = 1 2(t 1) t =0 k ˆY ( ˆY (t ), ˆy)k Zj(Z(t ) j , zj) (43) (ˆµ(2(t 1)) ˆY ˆµ(2(t 1)) Zj )(ˆy, zj) = 1 2(t 1) t =0 k ˆY ( ˆY (t ), ˆy) t =0 k Zj(Z(t ) j , zj) where k ˆY , k Zj are the kernels associated with R ˆY and RZj, respectively. Furthermore, note that, in practice, we only have access to samples from the test distribution P ˆY Zj (i.e., the joint) and we swap elements of two consecutive samples to simulate data from the null distribution P ˆY PZj. More formally, let D(2t) = ( ˆY (2t), Z(2t) j ) P ˆY Zj, D(2t 1) = ( ˆY (2t 1), Z(2t 1) j ) P ˆY Zj (45) e D(2t) = ( ˆY (2t), Z(2t 1) j ), e D(2t 1) = ( ˆY (2t 1), Z(2t) j ). (46) Then, ρSKIT t := ˆµ(2(t 1)) ˆY Zj ˆµ(2(t 1)) ˆY ˆµ(2(t 1)) Zj (47) where ˆµ ˆY Zj, ˆµ ˆY , ˆµZj are the mean embeddings of P ˆY Zj, P ˆY , PZj in R ˆY RZj, R ˆY , and RZj, respectively, and is the tensor product as in Gretton et al. [29]. We remark that ρSKIT t is a real-valued operator, i.e. ρSKIT t : R R R, and that we use data up to t 1 to compute ρt in order to maintain validity of the test, i.e. ρt is fixed conditionally on previous observations. Following Lemma 2, we conclude κSKIT t := tanh ρSKIT t (D(2t 1)) + ρSKIT t (D(2t)) ρSKIT t ( e D(2t 1)) ρSKIT t ( e D(2t)) (48) and Algorithm B.1 summarizes the SKIT procedure for the global semantic importance null hypothesis HG 0,j in Definition 1. B.2 Computing ρC-SKIT t and κC-SKIT t Recall that ρC-SKIT t is the estimate of the witness function of MMD(P ˆY Zj Z j, P ˆY e Zj Z j) with e Zj PZj|Z j at time t, i.e. ρC-SKIT t = ˆµ(t 1) ˆY Zj Z j ˆµ(t 1) ˆY e Zj Z j, (49) ˆµ(t 1) ˆY Zj Z j = 1 t 1 φ ˆY ( ˆY (t )) φZj(Z(t ) j ) φZ j(Z(t ) j ) (50) ˆµ(t 1) ˆY e Zj Z j = 1 t 1 φ ˆY ( ˆY (t )) φZj( e Z(t ) j ) φZ j(Z(t ) j ) (51) and φ ˆY , φZj, φZ j are the canonical feature maps associated with their respective RKHSs. We remark that ρC-SKIT t is defined as an operator, and, for a triplet (ˆy, zj, z j) its value can be computed as ρC-SKIT t (ˆy, zj, z j) = (ˆµ(t 1) ˆY Zj Z j ˆµ(t 1) ˆY e Zj Z j)(ˆy, zj, z j) (52) = ˆµ(t 1) ˆY Zj Z j(ˆy, zj, z j) ˆµ(t 1) ˆY e Zj Z j(ˆy, zj, z j) (53) ˆµ(t 1) ˆY Zj Z j(ˆy, zj, z j) = 1 t 1 t =0 k ˆY ( ˆY (t ), y)k Zj(Z(t ) j , zj)k Z j(Z(t ) j , z j) (54) ˆµ(t 1) ˆY e Zj Z j(ˆy, zj, z j) = 1 t 1 t =0 k ˆY ( ˆY (t ), y)k Zj( e Z(t ) j , zj)k Z j(Z(t ) j , z j) (55) where k ˆY , k Zj, k Z j are the kernels associated with R ˆY , RZj, and RZ j, respectively. Following Lemma 2, we conclude κC-SKIT t := tanh(ρC-SKIT t ( ˆY , Zj, Z j) ρC-SKIT t ( ˆY , e Zj, Z j)). (56) B.3 Computing ρX-SKIT t and κX-SKIT t Recall that, for a particular sample z, a concept j [m], and a subset S [m] \ {j} that does not contain j, ρX-SKIT t is the estimate at time t of the witness function of MMD( ˆYS {j}, ˆYS) with ˆYC = g( e HC), e HC PH|ZC=z C, i.e. ρX-SKIT t = ˆµ(t 1) ˆYS {j} ˆµ(t 1) ˆYS , (57) ˆµ(t 1) ˆYS {j} = 1 t 1 t =0 φ ˆY ( ˆY (t ) S {j}), ˆµ(t 1) ˆYS = 1 t 1 t =0 φ ˆY ( ˆY (t ) S ), (58) and φ ˆY is the canonical feature map of R ˆY . Then, for a prediction ˆy, the value of ρX-SKIT t is ρX-SKIT t (ˆy) = (ˆµ(t 1) ˆYS {j} ˆµ(t 1) ˆYS )(ˆy) (59) = ˆµ(t 1) ˆYS {j}(ˆy) ˆµ(t 1) ˆYS (ˆy) (60) t =0 k ˆY ( ˆY (t ) S {j}, ˆy) t =0 k ˆY ( ˆY (t ) S , ˆy) where k ˆY is the kernel associated with R ˆY . To conclude, applying Lemma 2 implies κX-SKIT t := tanh(ρX-SKIT t ( ˆYS {j}) ρX-SKIT t ( ˆYS)). (62) In this appendix, we include the proofs of the results presented in this paper. C.1 Proof of Lemma 1 Recall that c Rd m is a dictionary of m concepts such that cj, j [m] is the vector representation of the jth concept. Then, Z = c, H is the vector where after appropriate normalization Zj [ 1, 1] represents the amount of concept j in h. We want to show that if ˆY = w, H , w Rd, and d 3, then ˆY Zj w, cj = 0. (63) That is, w being orthogonal to cj does not provide any information about the statistical dependence between ˆY and Zj, and vice versa. Proof. Herein, for the sake of simplicity, we will drop the cj notation and consider a single concept c. Furthermore, we will assume that all vectors are normalized, i.e. w = h = c = 1. Note that the Eq. (63) above can directly be rewritten as w, H c, H w, c = 0. (64) ( = ) We show there exist random vectors H such that w, c = 0 but w, H c, H . Let H U(Sd), i.e. H = [H1, . . . , Hd] is sampled uniformly over the sphere in d dimensions. It is easy to see that j [d], Hj = ej, H , where ej is the jth element of the standard basis. Furthermore, it holds that Hj = q j =j H2 j by definition. We conclude that (j, j ), let w = ej and c = ej , then w, c = 0 but w, H c, H . That is, the fact that ej and ej are orthogonal does not mean that their respective projections of H are statistically independent. ( = ) We show how to construct a random vector H such that w, H c, H but w, c = 0. Denote Hη := {h Sd : c, h = η, η = 0} the linear subspace of unit vectors in Rd with the same nonzero inner product with c. Each vector h Hη can be decomposed into a parallel and an orthogonal component to c, i.e. h Hη, h = hc + h = ηc + h , where the last equality follows by definition of Hη. Consider Hη and H η, it follows that for w, h+ Hη and h H η w, h+ = w, h ηc + w , ηc + h+ = ηc + w , ηc + h (65) η2 + w , h+ = η2 + w , h (66) w , h+ h = 2η2 (67) w , = 2η2. ( := h+ h ) (68) Denote S = {(h+, h ) : w , = 2η2} the set of pairs of vector that satisfy Eq. (68), and note that for each pair (h+, h ) there exists a value β such that w, h+ = w, h = β. Then, sampling from S is equivalent to sampling from the set of pairs of vectors in Hη and H η that attain the same correlation with w. Note that when d = 2, h+ , h { w } by construction, hence {0, 2w } and w , {0, 2(1 η2)}. Then, S = because there are no pairs of vectors such that w , = 2η2. For d 3, S is nonempty as long as η p Then, we can construct H as follows: Sample the component parallel to c, i.e. Hc U( ηc), Sample the component orthogonal to c, i.e. (H+ , H ) U(S), and note that by doing so, we have sampled c, H and w, H independently. Finally H = Hc + H+ if Hc = ηc Hc + H if Hc = ηc (69) has w, H c, H by construction, but, since w Hη, w, c = η = 0. This concludes the proof of the lemma. C.2 Proof of Proposition 1 Recall that Proposition 1 states that the payoff function κC-SKIT t = tanh(ρC-SKIT t ( ˆY , Zj, Z j) ρC-SKIT t ( ˆY , e Zj, Z j)) (70) provides a fair game for the global conditional semantic importance null hypothesis HGC 0,j in Definition 2. That is, the wealth process provides Type I error control. Proof. Note that HGC 0,j can be directly rewritten as HGC 0,j : P ˆY Zj Z j = P ˆY e Zj Z j (71) where e Zj PZj|Z j. Then, under the null, the triplets ( ˆY , Zj, Z j) and ( ˆY , e Zj, Z j) are exchangeable by assumption, and the result follows from Lemma 2. C.3 Proof of Proposition 2 Recall that Proposition 2 claims that a wealth process with κX-SKIT t = tanh(ρX-SKIT t ( ˆYS {j}) ρX-SKIT t ( ˆYS)) (72) can be used to reject the local conditional semantic importance HLC 0,j,S in Definition 3 with Type I error control, i.e. the game is fair. Proof. It is easy to see that HLC 0,j,S is already written as a two-sample test. Then, under this null, ˆYS {j} and ˆYS are exchangeable by assumption, and Lemma 2 implies the statement of the proposition. Figure D.1: Pictorial representation of the data-generating process for the synthetic dataset. D Synthetic Experiments In this section, we showcase the main properties of our tests on two synthetic datasets. D.1 Gaussian Data We start by illustrating all sequential tests presented in this work are valid, and that they adapt to the hardness of the problem, i.e. the weaker the dependence structure, the longer their rejection times. We devise a synthetic dataset with a nonlinear response such that all distributions are known and we can sample from the exact conditional distribution. The data-generating process we consider is defined as Z1 N(µ1, σ2 1) µ1 = 1, σ1 = 1 (73) Z2 N(µ2, σ2 2) µ2 = 1, σ2 = 1 (74) Z3 | Z1 N(Z1, σ2 3) σ3 = 1 (75) and Y | Z S(β1Z1 + β2Z2Z3 + β3Z3) + ϵ, (76) where S is the sigmoid function, ϵ N(0, σ2 0), σ0 = 0.01 is independent Gaussian noise, and βi, i = 1, 2, 3 are coefficients that will allow us to test different conditions. Then, it follows that g(z) = E[Y | Z = z] = S(β1z1 + β2z2z3 + β3z3) (77) σ2 1 σ2 1 + σ2 3 Z3 + σ2 3 σ2 1 + σ2 3 µ1, 1 Fig. D.1 depicts the data-generating process. We remark that, for this experiment, we assume there exists a known parametric relation between the response Y and the concepts Z. This is only to verify our tests retrieve the ground-truth structure, and our contributions do not rely on this assumption. With this data-generating process, it holds that: 1. If β2 = 0 then Y Z2, 2. If β1 = 0 then Y Z1 | Z 1, and 3. For an observation Z = z, if z3 = 0 then g( e Z{2,3}) d= g( e Z3) with e ZC PZ|ZC=z C. We test each condition with SKIT, C-SKIT, and X-SKIT, respectively. We use both a linear and RBF kernel with bandwidth set to the median pairwise distance between all previous observations (commonly referred to as the median heuristic [27]). For each test, we estimate the rejection rate (i.e., how often a test rejects), and the expected rejection time (i.e., how many steps of the test it takes to reject) over 100 draws of τ max = 1000 samples, and with a significance level α = 0.05. We remark that a normalized rejection time of 1 means failing to reject in τ max steps. D.1.1 Global Importance with SKIT First, we test that β2 = 0 = Y Z2 (79) 0.00 0.05 0.10 0.15 0.20 Rejection rate (linear) Rejection time (linear) Rejection rate (rbf) Rejection time (rbf) Significance level Figure D.2: Global importance results for H0 : Y Z2 with SKIT. (a) Marginal distributions of Y and Z2 for β2 = 1 and 0, respectively. The red dashed line is the linear regression between the two variables, and, as expected, the slope is 0 for β2 = 0. (b) Mean rejection rate and mean rejection time for SKIT with a linear and RBF kernel, as a function of β2. 0.00 0.05 0.10 0.15 0.20 Rejection rate (linear) Rejection time (linear) Rejection rate (rbf) Rejection time (rbf) Significance level Figure D.3: Global conditional importance results for H0 : Y Z1 | Z 1 with C-SKIT. (a) e Z1 PZ1|Z 1 is independent of Y for Z 1 = [ 1, 3]. As expected, the slope of the linear regression between Y and e Z1 is 0. (b) Mean rejection rate and mean rejection time for C-SKIT with a linear and RBF kernel, as a function of β1. with the symmetry-based SKIT in Algorithm B.1. Fig. D.2a shows samples from the joint distribution PY Z2 for β2 = 1 and β2 = 0. As expected, when β2 = 0, the slope of the linear regression (red dashed line) is 0 because Y and Z2 are independent. Fig. D.2b reports average rejection rate and average rejection time as a function of β2. As β2 increases, the strength of the dependency between Y and Z2 increases, and the rejection time decreases this adaptive behavior is characteristic of sequential tests. We can verify that the rejection rate is below the significance level α = 0.05 when β2 = 0, and that the SKIT procedure provides Type I error control. Finally, we note that both kernels perform similarly for this test, with the linear kernel generally rejecting less than the RBF one, and with longer rejection times. D.1.2 Global Conditional Importance with C-SKIT Then, we test that β1 = 0 = Y Z1 | Z 1 (80) with C-SKIT (Algorithm 1). We remark that we can sample from the exact conditional distribution PZ1|Z 1 = PZ1|Z{2,3} because Z2 is independent of Z1 by construction, and the conditional PZ1|Z3 can be computed analytically as shown in Eq. (78). We verify the conditional distribution behaves as expected in Fig. D.3a. By construction, e Z1 is sampled without looking at Y , hence it is independent, and the slope of the linear regression (red dashed line) is 0. Fig. D.3b shows mean rejection rate and time as a function of β1. First and foremost, we can see that in this case the linear kernel always fails to reject independently of the value of β1. This behavior highlights an important aspect of all kernel-based tests, that is the kernel needs to be characteristic in order for the mean embedding to be an injective function [25, 68]. If this condition is not satisfied, different probability distributions could share the same mean embedding in the RKHS, and it may not be possible to disambiguate them 0.3 0.4 0.5 0.6 0.7 0.8 g(Z) Z2 = 1, Z3 = 0.5 0.3 0.4 0.5 0.6 0.7 0.8 g(Z) Z2 = 1, Z3 = 0 0.10 0.05 0.00 0.05 0.10 z3 g(Z2, 3) d= g(Z3) Rejection rate (linear) Rejection time (linear) Rejection rate (rbf) Rejection time (rbf) Significance level Figure D.4: Local conditional importance results for H0 : g( e Z{2,3}) d= g( e Z3) with X-SKIT. (a) Shows that, as expected, the test and null distributions overlap when z3 = 0. at all. Consequently, the test will not be consistent, and increasing τ max will not increase power. For the RBF kernel which satisfies the characteristic property for probability distributions on Rd the test is valid (i.e., it provides Type I error control for β1 = 0), and it is adaptive to the strength of the conditional dependence structure as β1 increases, the rejection time decreases. D.1.3 Local Conditional Importance with X-SKIT Finally, we test that for a fixed z, z3 = 0 = g( e Z{2,3}) d= g( e Z3), e ZC PZ|ZC=z C, C [m]. (81) with X-SKIT (Algorithm 2). That is because of the multiplicative term z2z3 in g the observed value of z2 does not change the distribution of the response of the model when z3 = 0. Fig. D.4a shows the test (i.e., g( e Z{2,3})) and null (i.e., g( e Z3)) distributions for different values of z3 when z2 = 1. As expected, we can see that when z3 = 0, the two distributions overlap, whereas when z3 = 0.5, the test distribution is slightly shifted to the right. Fig. D.4b shows results of X-SKIT with both a linear and RBF kernel as a function of z3. We use both positive and negative values of z3 to show that X-SKIT has a two-sided alternative, i.e. it rejects both when the test distribution is to the right and to the left of the null. We can see that both the linear and RBF kernel provide Type I error control when z3 = 0, and that their rejection times adapt to the hardness of the problem. Now that we have illustrated all tests in arguably the simplest setting, we move onto a synthetic dataset where the response is learned by means of a neural network. D.2 Counting MNIST Digits In this section, we test the semantic importance structure of a neural network trained to count numbers in synthetic images assembled by placing digits from the MNIST dataset [38] in a 4 4 grid. Fig. D.5 depicts the data-generating process, which satisfies: Blue zeros, orange threes, blue twos, and purple sevens are sampled independently with Zblue zeros U({0, 1, 2}) + ϵ Zorange threes U({0, 1, 2}) + ϵ (82) Zblue twos U({1, 2}) + ϵ Zpurple sevens U({1, 2}) + ϵ (83) Green fives are sampled conditionally on blue zeros with Zgreen fives | Zblue zeros Cat [3/4, 1/8, 1/8] if Nblue zeros = 0 [1/8, 3/4, 1/8] if Nblue zeros = 1 [1/8, 1/8, 3/4] if Nblue zeros = 2 That is, the number of blue zeros changes the probability distribution of green fives over 1,2,3. blue zeros orange threes green fives red threes blue twos purple sevens Figure D.5: Pictorial representation of the data-generating process for the counting dataset. Red threes are sampled conditionally on both orange threes and green fives with Zred threes | Zorange threes, Zgreen fives 2 + Bernoulli(p) + ϵ, (85) p = α if Norange threes Ngreen fives 3 1 α otherwise , α = 0.9. (86) That is, the product of the number of orange threes and green fives changes the distribution of red threes over 1,2. Finally, we remark that N denotes the nearest integer to Z, and ϵ is independent uniform noise (i.e., ϵ U( 0.5, 0.5)) to make the distribution of the concepts continuous. To summarize, in order to generate images, we first sample the concepts Z according to the distribution above, round their values to their respective nearest integers N, and finally randomly place digits from the MNIST dataset in a 4 4 grid according to their number. Note that this data-generating process adds color to the original black and white MNIST digits, and that color matters for the counting task since orange threes and red threes have different distributions. We remark that, with the data generating process above, we can sample from the true conditional distribution of the digits, and, consequently, of images. We omit details on the conditional distribution for the sake of presentation.4 We stress that this setting differs slightly from the general one presented in this paper, where we consider both an encoder f and a classifier g such that ˆy = g(f(x)), and we sample from the conditional distribution of the dense embeddings H given any subset of concepts (i.e., PH|ZC). The scope of this experiment is to showcase the effectiveness of our tests when the response is parametrized by a complex, nonlinear, learned predictor, hence we train a neural network such that ˆy = f(x) and directly sample from the conditional distribution of images given any subset of digits (i.e., PX|ZC). We sample a training dataset of 50,000 images and train a Res Net18 [30] to predict the number of all digits for 6 epochs with batch size of 64 and Adam optimizer [35] with learning rate equal to 10 4, weight decay of 10 5, and a scheduler that halves the learning rate every other epoch (recall that the model needs to learn to disambiguate red and orange threes, so color matters). To evaluate the model, we round predictions to the nearest integer and compute accuracy on a held-out set of 10,000 images from the same distribution (we use the original train and test splits of the MNIST dataset to guarantee no digits showed during training are included in test images), and the model achieves an accuracy greater than 99%. Herein, we study the semantic importance structure of the predicted number of red threes with respect to the predicted number of other digits. Note that the ground-truth distribution satisfies the following conditions: 1. Red threes are independent of blue twos and purple sevens, i.e. Zred threes Zblue twos and Zred threes Zpurple sevens. (87) 4All code necessary to reproduce experiments is available on Git Hub. ground-truth predictions Figure D.6: Distribution of ground-truth data and density estimation of the predictions of the trained model for the validation data in the counting digits experiments. 2. Red threes are independent of blue zeros conditionally on green fives, i.e. Zred threes Zblue zeros | Zgreen fives. (88) 3. If in a specific image there are no orange threes, then red threes are independent of green fives, i.e. Zred threes|(Ngreen fives = ngreen fives, Norange threes = 0) d= Zred threes|Norange threes = 0. (89) D.2.1 Global Importance with SKIT We start by testing whether the predictions of the model satisfy the ground-truth condition Zred threes Zblue twos and Zred threes Zpurple sevens (90) with SKIT (Algorithm B.1). We remark that at inference, we round predictions to the nearest integer and add independent uniform noise ϵ U( 0.5, 0.5) to make the distribution of the response of the model continuous. Fig. D.6 shows the ground-truth distribution of red threes as a function of other digits in the held-out set, and the kernel density estimation of the predictions of the model. As expected, we can see that the ground-truth distribution is marginally dependent on blue zeros, orange threes, and green fives, but it is independent of blue twos and purple sevens. We repeat all tests 100 times with both linear and RBF kernels with bandwidth set to the median of the pairwise distances of previous observations. We perform tests on independent draws of data of size τ max {100, 200, 400, 800, 1600} from the validation set, and study the rank of importance as a function of τ max, i.e. the amount of data available to test. Fig. D.7 includes mean rejection rate and mean rejection time for each digit, and the rank of importance of digits as a function of τ max. We can see that as expected both linear and RBF kernels successfully control Type I error for blue twos and purple sevens , and this confirms that the distribution of the predictions of the model agrees with the underlying ground-truth data-generating process. Furthermore, we can see that the rank of importance is stable across different values of τ max, with purple sevens and blue twos consistently ranked last. D.2.2 Global Conditional Importance with C-SKIT Then, we test whether the predictions of the model satisfy the ground-truth conditional independence condition Zred threes Zblue zeros | Zgreen fives (91) with C-SKIT. Analogous to above, we repeat all tests 100 times with linear and RBF kernels, and Fig. D.8 includes results for τ max {100, 200, 400, 800, 1600}. Here similarly to the synthetic experiment presented in Appendix D.1 we can see that the linear kernel almost always fails to reject, i.e. the mean rejection rates for all digits are close to 0. As discussed earlier, this behavior is due to the fact that the linear kernel is not characteristic for the distributions. On the other hand, the RBF is, and, as expected, it is consistent and it provides Type I error control for the null hypothesis that red threes are independent of blue zeros conditionally on all other digits, which is true. Furthermore, we can see that the rank of importance is less stable compared to the one in Fig. D.2, and in particular, τ max = 100 seems not to be sufficient to retrieve the correct ground-truth structure (i.e., blue twos are ranked before green fives). This highlights how the amount of data available for testing may affect results and findings. 100 200 400 800 1600 max orange threes green fives purple sevens 0.0 0.5 1.0 green fives orange threes purple sevens (a) Linear kernel. 100 200 400 800 1600 max orange threes green fives purple sevens 0.0 0.5 1.0 orange threes green fives purple sevens (b) RBF kernel. Figure D.7: Global semantic importance results for the predicted number of red threes with linear and RBF kernels. In each subfigure, the leftmost panel shows mean rejection rate and mean rejection time over 100 tests with α = 0.05 and τ max = 800. The rightmost panel shows the rank of importance of digits for the prediction of red threes as a function of τ max. 100 200 400 800 1600 max orange threes green fives purple sevens 0.0 0.5 1.0 orange threes green fives purple sevens (a) Linear kernel. 100 200 400 800 1600 max orange threes green fives purple sevens 0.0 0.5 1.0 orange threes green fives purple sevens (b) RBF kernel. Figure D.8: Global conditional semantic importance results for the predicted number of red threes with linear and RBF kernels. In each subfigure, the leftmost panel shows mean rejection rate and mean rejection time over 100 tests with α = 0.05 and τ max = 800. The rightmost panel shows the rank of importance of digits for the prediction of red threes as a function of τ max. D.2.3 Local Conditional Importance with X-SKIT Finally, we test whether the predictions of the model satisfy the ground-truth condition that, for a particular image, if there are no orange threes (i.e., norange threes = 0) then red threes are independent of the observed green fives (i.e., ngreen fives), i.e. Zred threes|(Ngreen fives = ngreen fives, Norange threes = 0) d= Zred threes|Norange threes = 0. (92) We remark that, in the equation above, conditioning is written in terms of the integer norange threes = 0 because of its intuitive meaning, and that this is equivalent to conditioning on zorange threes ( 0.5, 0.5). Similarly, we could replace ngreen fives with zgreen fives, and, in practice, we run tests conditioning on the observed concepts z, and not their integer values n. We use X-SKIT (Algorithm 2) with a linear and RBF kernel with bandwidth set to the median of the pairwise distances of previous observations. We repeat all tests 100 times on individual images with 0, 1, and 2 orange threes, significance level α = 0.05, and τ max = 400. Fig. D.9 shows results grouped by number of orange threes. As expected, we see that when norange threes is grater than 0, the number of green fives is important for the predictions of the model (i.e., rejection rate is close to 1, with short rejection rate), whereas when there are no orange threes in the image, both the linear and RBF kernel control Type I error. We qualitatively compared our findings with pixel-level explanations with Grad-CAM [57], and we can see that they only highlight red threes because that is the digit we are explaining the prediction of. That is, pixel-level explanations cannot convey the full spectrum of semantic importance for the predictions of a model which can be misleading to users. For example, in this case, a user may not understand when the predictions of a model depend on the number of green fives, because they are never highlighted by pixel-level saliency maps. In real-world scenarios, digits may be replaced by sensitive attributes that cannot be inferred by the raw value of pixels. For example, a saliency map highlighting a face does not convey which attributes were used by the model, such as skin color, biological sex, or gender. It is immediate to see how being able to investigate the dependencies of the predictions of a model with respects to these attributes (which our definitions provide) is paramount for their safe deployment. Conditioning concept orange threes = 1 Grad-CAM explanation linear rbf 0.0 Local conditional importance of green fives = 3 Rejection rate Rejection time (a) Results for norange threes = 2. Conditioning concept orange threes = 2 Grad-CAM explanation linear rbf 0.0 Local conditional importance of green fives = 1 Rejection rate Rejection time (b) Results for norange threes = 1. Conditioning concept orange threes = 0 Grad-CAM explanation linear rbf 0.0 Local conditional importance of green fives = 1 Rejection rate Rejection time (c) Results for norange threes = 0. Figure D.9: Local conditional importance of Ngreen fives conditionally on Norange threes. Each row contains the input image, the Grad-CAM explanation for the prediction of the model, and X-SKIT results for 100 repetitions of the test with τ max = 400, with a linear and RBF kernel. Note that X-SKIT finds the observed number of green fives important whenever the number of orange threes is greater than zero, whereas Grad-CAM does not. With these results on synthetic datasets, we now showcase the flexibility of our proposed tests on zero-shot image classification with several and diverse vision-language (VL) models. E Experimental Details In this appendix, we include further details about the real-world experiments that were omitted from the main text for the sake of presentation. All experiments were run on a private server with one 24 GB NVIDIA RTX A5000 GPU and 96 CPU cores with 500 GB of RAM memory. List of VL models used in the experiments. We use 8 different VL models, both CLIPand non-CLIP-based: CLIP:RN50,Vi T-B/32,Vi T-L/14 [52], Open Clip:Vi T-B-32,Vi T-L-14 [31], FLAVA [65], ALIGN [32], and BLIP [39]. Evaluating rank agreement. We use a weighted version of Kendall s tau [33] introduced by Vigna [74] which assigns higher penalties to swaps between elements with higher ranks. This choice reflects the fact that concepts with higher importance should be more stable across different models. We briefly remark that this notion of rank agreement is bounded in [ 1, 1] ( 1 indicates reverse order, and 1 perfect alignment) but not symmetric. Evaluating importance agreement. We threshold rejection rates at level α to classify concepts into important and not important ones. Then, importance agreement is the accuracy between pairs of binarized vectors. E.1 Estimating Conditional Distributions from Data Here, we introduce nonparametric methods to estimate the conditional distributions necessary to run our C-SKIT and X-SKIT tests. Throughout this section, we will assume access to a training set {(h(i), z(i))}n i=1 of n tuples of dense embeddings h Rd with their semantics z [ 1, 1]m. E.1.1 Estimating PZj|Z j for C-SKIT Here, we describe how to sample from the conditional distribution of a concept Zj given the rest, Z j, i.e. e Zj PZj|Z j, which is necessary to run our C-SKIT test. In particular, for a concept Figure E.1: Example marginal and estimated conditional distributions p(zj) and p(zj|z j) for two class-specific concepts on three images from the Imagenette dataset. Distributions are shown as a function of the effective number of points in the weighted KDE (i.e., neff). j [m], and an observation z [ 1, 1]m, we define the unnormalized conditional distribution p(zj | z j) = z(i) j zj νscott by means of weighted kernel density estimation (KDE), where ϕ is the standard normal probability density function, νscott is Scott s factor [56], and , ν > 0 (94) are the weights. That is, the further z(i) j is from z j, the lower its weight in the KDE. As for all kernel-based methods, the bandwidth ν is important for the practical performance of the model. For our experiments, we choose ν adaptively such that the effective number of points in the KDE (i.e. neff = (Pn i=1 wi)2/ Pn i=1 w2 i ) is the same across concepts. This choice is motivated by the fact that different concepts have different distributions, and we want to guarantee the same number of points are used to estimate their conditional distributions. Furthermore, we note that neff controls the strength of the conditioning the larger neff, the slower the decay of the weights, and the weaker the conditioning. That is, in the limit, the weights become uniform, the conditional distribution tends to the marginal p(zj), and all tests presented become of decorrelated semantic importance [72, 73]. With this, sampling e Zj PZj|Z j=z j boils down to first sampling i according to the weights w = [wi, . . . , wn], and then sampling e Zj from the Gaussian distribution centered at z(i ) j . Fig. E.1 shows the marginal (i.e., p(zj)) and estimated conditional distributions (i.e., p(zj|z j)) of two class-specific concepts as a function of effective number of points neff for three images in the Imagenette dataset. We can see how as neff increases, the estimated conditional distribution becomes closer to the marginal, and that the conditional distributions of class-specific concepts tend to be skewed to higher values compared to their marginals. This behavior suggests that images from a specific class have higher values of concepts that are related to the class. We use neff = 2000 for all tests across all real-world experiments. E.1.2 Estimating PH|ZC=z C for X-SKIT Here, we describe how to sample from the conditional distribution of dense embeddings H conditionally on any subset of concepts C [m] of a particular semantic vector z [ 1, 1]m, i.e. e HC PH|ZC=z C, which is necessary to run our X-SKIT test. We show how to achieve this by coupling the nonparametric sampler introduced above with ideas of nearest neighbors. This choice is Figure E.2: Example test (i.e., g( e HS {j})) and null (i.e., g( e HS)) distributions for a class-specific concept and a non-class specific one on three images from the Imagenette dataset as a function of the cardinality of S. motivated by the need to keep samples in-distribution with respects to the downstream classifier g. Intuitively, we propose to: 1. Sample e Z PZ|ZC=z C, and then 2. Retrieve the embedding H(i ) such that its concept representation Z(i ) is the nearest neighbor of e Z. Step 2 makes sure that samples are coming from real images, and it overcomes some of the hurdles of sampling a high-dimensional vector (H Rd, d 102) conditionally on a low-dimensional one (z C [ 1, 1]|C|, C [m], m 20). More precisely, recall that {(h(i), z(i))}n i=1 is a set of n pairs of dense embeddings and their semantics, then e HC = h(i ) s.t. i = arg min i [n] z(i) e Z , e Z PZ|ZC=z C, (95) where PZ|ZC=z C is approximated with p(z C | z C), C := [m] \ C as in Eq. (93). Fig. E.2 shows some example test (i.e., g( e HS {j})) and null distributions (i.e., g( e HS)) for a classspecific concept and a non-class specific one on the same three images from the Imagenette dataset as in Fig. E.1. We remark that S can be any subset of the remaining concepts, so we show results for random subsets of increasing size. We can see that the test distributions of class-specific concepts are skewed to the right, i.e. including the observed class-specific concept increases the output of the predictor. Furthermore, we see the shift decreases the more concepts are included in S, i.e. if S is larger and it contains more information, then the marginal contribution of one additional concept will be smaller. On the other hand, including a non-class-specific concept does not change the distribution of the response of the model, no matter the size of S precisely as our local definition of importance (HLC 0,j,S) demands. E.2 Aw A2 Dataset Here, we include additional information, tables, and figures for the Aw A2 dataset experiment. This dataset comprises 37,322 images (29,841 for training and 7,481 for testing) from 50 animal species along with class-level annotations of 85 attributes (some example figures are included in Fig. E.3). Concept annotations are reported both as frequencies (i.e., how often an attribute appears in images coming from a class) and as binary labels (i.e., 1 means that an attribute is present in a class, and 0 otherwise). Table E.2 shows the zero-shot classification performance of all VL models used in this experiment for the top-10 classes, and, for each class, we test 20 attributes: the 10 most frequent, and a random subset of 10 absent ones (concepts are included in Table E.3). Finally we remark that, for each model, we obtain the dictionary c by encoding concepts with its text encoder. Tests are run with bandwidths of the kernels set to the 90th percentile of the pairwise distances between observations and τ max = 400, 800 for SKIT and C-SKIT, respectively. giant panda tiger giraffe zebra lion squirrel sheep horse elephant dalmatian Figure E.3: Example images from the top-10 best classified classes in the Aw A2 dataset. Table E.2: Zero-shot classification accuracy on the Aw A2 dataset. Model giant panda tiger giraffe zebra lion CLIP:RN50 100.00% 100.00% 100.00% 100.00% 99.51% CLIP:Vi T-B/32 100.00% 100.00% 100.00% 100.00% 99.51% CLIP:Vi T-L/14 100.00% 100.00% 99.59% 100.00% 100.00% Open Clip:Vi T-B-32 100.00% 100.00% 100.00% 100.00% 100.00% Open Clip:Vi T-L-14 100.00% 100.00% 100.00% 100.00% 100.00% FLAVA 100.00% 98.86% 100.00% 100.00% 99.02% ALIGN 100.00% 100.00% 100.00% 99.57% 100.00% BLIP 100.00% 100.00% 99.17% 99.15% 99.51% average 100.00% 0.00 99.86% 0.38 99.84% 0.29 99.84% 0.30 99.69% 0.34 Model squirrel sheep horse elephant dalmatian CLIP:RN50 99.17% 97.54% 98.18% 94.71% 96.36% CLIP:Vi T-B/32 99.58% 98.59% 99.39% 99.04% 98.18% CLIP:Vi T-L/14 100.00% 99.65% 98.78% 100.00% 100.00% Open Clip:Vi T-B-32 99.58% 99.65% 98.78% 100.00% 97.27% Open Clip:Vi T-L-14 100.00% 100.00% 100.00% 100.00% 100.00% FLAVA 99.17% 100.00% 99.39% 99.04% 98.18% ALIGN 99.58% 99.65% 99.70% 100.00% 99.09% BLIP 99.17% 98.94% 99.39% 100.00% 100.00% average 99.53% 0.33 99.25% 0.80 99.20% 0.55 99.10% 1.71 98.64% 1.29 Table E.3: Class-level attributes tested on the Aw A2 dataset. Attributes (20) Class present (10) absent (10) giant panda patches, old world, furry, black, big, white, walks, paws, bulbous, vegetation flies, flippers, hooves, desert, hairless, red, blue, horns, plankton, yellow tiger stripes, stalker, meat, meat teeth, hunter, strong, fierce, old world, muscle, big strain teeth, tunnels, hops, plankton, bipedal, tusks, flippers, flies, small, skimmer giraffe long neck, long legs, big, quadrupedal, spots, vegetation, lean, old world, walks, ground bipedal, hibernate, cave, mountains, ocean, hunter, water, stripes, gray, fish zebra stripes, old world, quadrupedal, black, white, ground, group, grazer, walks, hooves blue, spots, brown, water, coastal, patches, tusks, claws, scavenger, red lion meat, stalker, strong, hunter, meat teeth, big, fierce, paws, furry, old world tunnels, blue, horns, skimmer, long neck, water, flippers, tusks, arctic, spots squirrel tail, furry, forager, small, gray, tree, new world, forest, vegetation, brown horns, blue, yellow, tusks, meat teeth, flippers, scavenger, desert, plankton, strain teeth sheep white, quadrupedal, walks, group, vegetation, grazer, ground, fields, furry, new world arctic, flippers, insects, paws, long neck, red, yellow, swims, plankton, hands horse hooves, fast, grazer, big, long legs tail, quadrupedal, fields, brown, strong arctic, tree, bipedal, plankton, fish, stripes, ocean, strain teeth, scavenger, orange elephant big, old world, gray, tough skin, quadrupedal tusks, hairless, strong, ground, walks claws, flippers, orange, swims, ocean stripes, tunnels, plankton, coastal, strain teeth dalmatian big, old world, gray, tough skin, quadrupedal tusks, hairless, strong, ground, walks claws, flippers, orange, swims, ocean stripes, tunnels, plankton, coastal, strain teeth clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.74 ( 0.09) 0.71 ( 0.13) 0.42 ( 0.20) 0.57 ( 0.12) 0.56 ( 0.11) 0.51 ( 0.16) 0.44 ( 0.19) 0.74 ( 0.10) 1.00 ( 0.00) 0.78 ( 0.10) 0.38 ( 0.25) 0.50 ( 0.20) 0.59 ( 0.16) 0.44 ( 0.26) 0.49 ( 0.19) 0.72 ( 0.10) 0.79 ( 0.10) 1.00 ( 0.00) 0.34 ( 0.23) 0.52 ( 0.19) 0.56 ( 0.16) 0.47 ( 0.25) 0.51 ( 0.14) 0.50 ( 0.14) 0.40 ( 0.24) 0.40 ( 0.15) 1.00 ( 0.00) 0.56 ( 0.09) 0.37 ( 0.15) 0.46 ( 0.14) 0.37 ( 0.15) 0.56 ( 0.17) 0.44 ( 0.33) 0.50 ( 0.23) 0.52 ( 0.17) 1.00 ( 0.00) 0.39 ( 0.17) 0.54 ( 0.13) 0.31 ( 0.24) 0.53 ( 0.10) 0.55 ( 0.21) 0.53 ( 0.17) 0.38 ( 0.13) 0.42 ( 0.15) 1.00 ( 0.00) 0.51 ( 0.08) 0.43 ( 0.12) 0.49 ( 0.16) 0.41 ( 0.29) 0.49 ( 0.23) 0.45 ( 0.13) 0.54 ( 0.14) 0.42 ( 0.13) 1.00 ( 0.00) 0.42 ( 0.20) 0.46 ( 0.15) 0.48 ( 0.17) 0.52 ( 0.11) 0.34 ( 0.12) 0.39 ( 0.17) 0.46 ( 0.14) 0.43 ( 0.22) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.65 ( 0.09) 0.57 ( 0.19) 0.52 ( 0.11) 0.42 ( 0.12) 0.48 ( 0.15) 0.43 ( 0.19) 0.42 ( 0.19) 0.66 ( 0.09) 1.00 ( 0.00) 0.63 ( 0.19) 0.49 ( 0.11) 0.52 ( 0.10) 0.52 ( 0.14) 0.46 ( 0.21) 0.45 ( 0.09) 0.56 ( 0.17) 0.61 ( 0.21) 1.00 ( 0.00) 0.45 ( 0.15) 0.47 ( 0.13) 0.51 ( 0.13) 0.50 ( 0.09) 0.49 ( 0.11) 0.52 ( 0.11) 0.44 ( 0.16) 0.45 ( 0.17) 1.00 ( 0.00) 0.41 ( 0.12) 0.37 ( 0.15) 0.47 ( 0.16) 0.28 ( 0.19) 0.38 ( 0.17) 0.45 ( 0.19) 0.49 ( 0.14) 0.41 ( 0.08) 1.00 ( 0.00) 0.33 ( 0.17) 0.46 ( 0.12) 0.41 ( 0.10) 0.49 ( 0.12) 0.49 ( 0.15) 0.51 ( 0.11) 0.36 ( 0.14) 0.38 ( 0.11) 1.00 ( 0.00) 0.42 ( 0.22) 0.45 ( 0.11) 0.43 ( 0.19) 0.45 ( 0.19) 0.49 ( 0.07) 0.53 ( 0.14) 0.43 ( 0.13) 0.36 ( 0.25) 1.00 ( 0.00) 0.37 ( 0.14) 0.47 ( 0.17) 0.45 ( 0.14) 0.46 ( 0.15) 0.38 ( 0.10) 0.41 ( 0.08) 0.47 ( 0.12) 0.40 ( 0.15) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.53 ( 0.12) 0.55 ( 0.07) 0.38 ( 0.24) 0.31 ( 0.16) 0.40 ( 0.17) 0.41 ( 0.21) 0.28 ( 0.21) 0.47 ( 0.14) 1.00 ( 0.00) 0.54 ( 0.08) 0.44 ( 0.25) 0.42 ( 0.17) 0.38 ( 0.16) 0.41 ( 0.18) 0.25 ( 0.22) 0.44 ( 0.21) 0.56 ( 0.06) 1.00 ( 0.00) 0.39 ( 0.13) 0.37 ( 0.17) 0.37 ( 0.14) 0.40 ( 0.14) 0.17 ( 0.23) 0.34 ( 0.28) 0.41 ( 0.23) 0.36 ( 0.14) 1.00 ( 0.00) 0.37 ( 0.18) 0.27 ( 0.24) 0.45 ( 0.16) 0.26 ( 0.26) 0.32 ( 0.18) 0.42 ( 0.20) 0.41 ( 0.13) 0.38 ( 0.14) 1.00 ( 0.00) 0.30 ( 0.26) 0.44 ( 0.15) 0.25 ( 0.19) 0.41 ( 0.18) 0.37 ( 0.18) 0.40 ( 0.14) 0.33 ( 0.22) 0.34 ( 0.15) 1.00 ( 0.00) 0.30 ( 0.16) 0.17 ( 0.24) 0.44 ( 0.20) 0.37 ( 0.21) 0.38 ( 0.17) 0.42 ( 0.22) 0.44 ( 0.21) 0.34 ( 0.21) 1.00 ( 0.00) 0.22 ( 0.19) 0.26 ( 0.19) 0.29 ( 0.17) 0.18 ( 0.20) 0.23 ( 0.25) 0.26 ( 0.21) 0.15 ( 0.30) 0.21 ( 0.25) 1.00 ( 0.00) Weighted Kendall's tau Figure E.4: Rank agreement comparison between SKIT, C-SKIT, and PCBM on the Aw A2 dataset. Results are reported as mean and standard deviation over the 10 classes considered in this experiment. 0.0 0.5 1.0 big (p) black (p) white (p) paws (p) furry (p) vegetation (p) flippers (a) yellow (a) patches (p) hairless (a) old world (p) walks (p) plankton (a) blue (a) bulbous (p) desert (a) hooves (a) horns (a) red (a) flies (a) giant panda 0.0 0.5 1.0 fierce (p) hunter (p) strong (p) stripes (p) muscle (p) bipedal (a) big (p) stalker (p) strain teeth (a) meat (p) small (a) old world (p) meat teeth (p) tusks (a) flippers (a) hops (a) skimmer (a) flies (a) tunnels (a) plankton (a) 0.0 0.5 1.0 long neck (p) spots (p) vegetation (p) long legs (p) bipedal (a) quadrupedal (p) big (p) old world (p) stripes (a) walks (p) mountains (a) ground (p) hunter (a) lean (p) gray (a) water (a) fish (a) ocean (a) cave (a) hibernate (a) 0.0 0.5 1.0 stripes (p) spots (a) old world (p) quadrupedal (p) tusks (a) grazer (p) hooves (p) walks (p) white (p) black (p) ground (p) water (a) patches (a) brown (a) red (a) blue (a) scavenger (a) coastal (a) group (p) claws (a) 0.0 0.5 1.0 fierce (p) strong (p) hunter (p) furry (p) paws (p) big (p) meat (p) stalker (p) tusks (a) spots (a) meat teeth (p) old world (p) water (a) arctic (a) long neck (a) skimmer (a) flippers (a) blue (a) horns (a) tunnels (a) rejection rate significance level rejection time strain teeth (a) 0.0 0.5 1.0 furry (p) scavenger (a) brown (p) small (p) yellow (a) tree (p) forager (p) tail (p) gray (p) blue (a) forest (p) plankton (a) desert (a) tusks (a) vegetation (p) flippers (a) meat teeth (a) horns (a) new world (p) 0.0 0.5 1.0 fields (p) grazer (p) white (p) walks (p) vegetation (p) ground (p) yellow (a) insects (a) red (a) furry (p) quadrupedal (p) long neck (a) plankton (a) flippers (a) swims (a) paws (a) arctic (a) group (p) hands (a) new world (p) 0.0 0.5 1.0 hooves (p) grazer (p) fields (p) quadrupedal (p) long legs (p) tree (a) big (p) strong (p) brown (p) stripes (a) orange (a) bipedal (a) fast (p) strain teeth (a) ocean (a) plankton (a) tail (p) fish (a) arctic (a) scavenger (a) horse elephant 0.0 0.5 1.0 tusks (p) hairless (p) strong (p) old world (p) big (p) tough skin (p) stripes (a) ocean (a) plankton (a) quadrupedal (p) strain teeth (a) ground (p) walks (p) gray (p) coastal (a) claws (a) orange (a) tunnels (a) swims (a) flippers (a) 0.0 0.5 1.0 spots (p) white (p) lean (p) stripes (a) walks (p) fast (p) black (p) quadrupedal (p) blue (a) grazer (a) vegetation (a) smelly (a) plains (a) mountains (a) paws (p) fish (a) new world (p) hands (a) horns (a) domestic (p) Figure E.5: SKIT importance results with CLIP:Vi T-L/14 on the Aw A2 dataset. 0.0 0.5 1.0 plankton (a) vegetation (p) white (p) blue (a) big (p) furry (p) flies (a) paws (p) hairless (a) old world (p) red (a) hooves (a) yellow (a) bulbous (p) black (p) horns (a) patches (p) flippers (a) walks (p) desert (a) giant panda 0.0 0.5 1.0 fierce (p) stripes (p) hunter (p) strong (p) bipedal (a) meat (p) muscle (p) skimmer (a) meat teeth (p) plankton (a) tunnels (a) flies (a) flippers (a) stalker (p) big (p) small (a) strain teeth (a) old world (p) tusks (a) hops (a) 0.0 0.5 1.0 long neck (p) bipedal (a) fish (a) vegetation (p) spots (p) ocean (a) long legs (p) quadrupedal (p) water (a) gray (a) big (p) hibernate (a) old world (p) stripes (a) ground (p) lean (p) cave (a) hunter (a) mountains (a) walks (p) 0.0 0.5 1.0 stripes (p) spots (a) blue (a) tusks (a) grazer (p) coastal (a) claws (a) old world (p) red (a) ground (p) patches (a) white (p) hooves (p) quadrupedal (p) brown (a) water (a) walks (p) black (p) scavenger (a) group (p) 0.0 0.5 1.0 furry (p) scavenger (a) tree (p) forest (p) small (p) vegetation (p) forager (p) plankton (a) tusks (a) meat teeth (a) blue (a) brown (p) strain teeth (a) gray (p) new world (p) yellow (a) horns (a) desert (a) tail (p) flippers (a) 0.0 0.5 1.0 fields (p) vegetation (p) grazer (p) white (p) long neck (a) swims (a) arctic (a) ground (p) furry (p) flippers (a) walks (p) insects (a) plankton (a) yellow (a) paws (a) red (a) hands (a) new world (p) quadrupedal (p) group (p) 0.0 0.5 1.0 hooves (p) grazer (p) quadrupedal (p) bipedal (a) fields (p) fish (a) strong (p) tree (a) ocean (a) big (p) stripes (a) long legs (p) plankton (a) fast (p) brown (p) orange (a) tail (p) arctic (a) scavenger (a) strain teeth (a) 0.0 0.5 1.0 tusks (p) ocean (a) hairless (p) big (p) plankton (a) swims (a) flippers (a) strong (p) old world (p) tough skin (p) coastal (a) ground (p) tunnels (a) walks (p) claws (a) gray (p) strain teeth (a) quadrupedal (p) orange (a) stripes (a) 0.0 0.5 1.0 spots (p) white (p) vegetation (a) paws (p) grazer (a) new world (p) lean (p) hands (a) fast (p) blue (a) domestic (p) walks (p) fish (a) mountains (a) black (p) smelly (a) plains (a) quadrupedal (p) stripes (a) horns (a) 0.0 0.5 1.0 fierce (p) furry (p) hunter (p) meat (p) water (a) paws (p) strong (p) long neck (a) flippers (a) tusks (a) meat teeth (p) spots (a) arctic (a) old world (p) tunnels (a) stalker (p) big (p) horns (a) skimmer (a) blue (a) rejection rate significance level rejection time importance importance Figure E.6: C-SKIT importance results with CLIP:Vi T-L/14 on the Aw A2 dataset. white pelican brown pelican mallard horned puffin cardinal blue jay cape glossy starling frigatebird vermilion flycatcher northern flicker Figure E.7: Example test images from the CUB dataset with their respective classes. Table E.4: Zero-shot classification accuracy on the CUB dataset. Model white pelican brown pelican mallard horned puffin vermilion flycatcher CLIP:RN50 92.00% 95.00% 95.00% 86.67% 88.33% CLIP:Vi T-B/32 98.00% 96.67% 90.00% 93.33% 93.33% CLIP:Vi T-L/14 98.00% 100.00% 100.00% 95.00% 100.00% Open Clip:Vi T-B-32 100.00% 98.33% 98.33% 88.33% 96.67% Open Clip:Vi T-L-14 100.00% 100.00% 100.00% 96.67% 100.00% FLAVA 100.00% 100.00% 98.33% 95.00% 91.67% ALIGN 96.00% 96.67% 95.00% 98.33% 78.33% BLIP 98.00% 80.00% 96.67% 65.00% 55.00% average 97.75% 2.54% 95.83% 6.24% 96.67% 3.12% 89.79% 10.08% 87.92% 14.09% Model northern flicker cardinal blue jay cape glossy starling frigatebird CLIP:RN50 78.33% 63.16% 65.00% 73.33% 78.33% CLIP:Vi T-B/32 98.33% 71.93% 83.33% 90.00% 80.00% CLIP:Vi T-L/14 98.33% 100.00% 85.00% 100.00% 96.67% Open Clip:Vi T-B-32 96.67% 100.00% 88.33% 96.67% 88.33% Open Clip:Vi T-L-14 100.00% 100.00% 91.67% 95.00% 93.33% FLAVA 76.67% 100.00% 91.67% 88.33% 90.00% ALIGN 86.67% 92.98% 83.33% 80.00% 55.00% BLIP 61.67% 61.40% 90.00% 73.33% 75.00% average 87.08% 12.96% 86.18% 16.42% 84.79% 8.14% 87.08% 9.75% 82.08% 12.47% E.3 CUB Dataset Here, we include additional information and results for the experiment of the CUB dataset [77], which contains 11,788 images of 200 different bird classes, and each image is annotated with the presence of 312 fine-grained concepts that describe the appearance of the bird (e.g., has orange bill , has hook-shaped bill , is small ) with the labelers confidence. Formally, the dataset is a collection {(x(i), y(i), z(i), u(i))}n i=1 of images x with class label y, binary semantic vector z(i) {0, 1}m, and uncertainty values u(i) {1, 2, 3, 4}: not visible (1), guessing (2), probably (3), and definitely (4). We randomly sample 10 images from the 10 classes with highest average accuracy across models: Table E.4 includes accuracies of all VL models used, and Fig. E.7 shows some example test images Table E.5: X-SKIT results on the CUB dataset as a function of conditioning set size s. s Rank agreement Importance agreement Importance f1 score 1 0.81 0.14 0.96% 0.06% 0.93 0.15 2 0.82 0.13 0.97% 0.06% 0.93 0.14 4 0.84 0.12 0.95% 0.08% 0.88 0.15 clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.81 ( 0.16) 0.81 ( 0.14) 0.82 ( 0.15) 0.80 ( 0.15) 0.82 ( 0.13) 0.83 ( 0.13) 0.81 ( 0.15) 0.81 ( 0.16) 1.00 ( 0.00) 0.83 ( 0.13) 0.82 ( 0.17) 0.81 ( 0.14) 0.82 ( 0.15) 0.82 ( 0.13) 0.82 ( 0.16) 0.81 ( 0.15) 0.82 ( 0.14) 1.00 ( 0.00) 0.82 ( 0.16) 0.82 ( 0.13) 0.82 ( 0.11) 0.81 ( 0.14) 0.81 ( 0.14) 0.82 ( 0.15) 0.82 ( 0.16) 0.82 ( 0.15) 1.00 ( 0.00) 0.82 ( 0.14) 0.83 ( 0.14) 0.81 ( 0.14) 0.82 ( 0.13) 0.81 ( 0.15) 0.81 ( 0.14) 0.82 ( 0.14) 0.82 ( 0.16) 1.00 ( 0.00) 0.81 ( 0.14) 0.81 ( 0.13) 0.80 ( 0.15) 0.82 ( 0.14) 0.82 ( 0.15) 0.82 ( 0.13) 0.83 ( 0.13) 0.81 ( 0.15) 1.00 ( 0.00) 0.83 ( 0.13) 0.83 ( 0.13) 0.82 ( 0.14) 0.82 ( 0.14) 0.80 ( 0.15) 0.80 ( 0.17) 0.81 ( 0.13) 0.83 ( 0.14) 1.00 ( 0.00) 0.81 ( 0.12) 0.82 ( 0.14) 0.82 ( 0.15) 0.81 ( 0.13) 0.81 ( 0.17) 0.80 ( 0.15) 0.83 ( 0.13) 0.81 ( 0.13) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.85 ( 0.12) 0.84 ( 0.12) 0.81 ( 0.15) 0.81 ( 0.15) 0.82 ( 0.14) 0.84 ( 0.11) 0.83 ( 0.12) 0.85 ( 0.12) 1.00 ( 0.00) 0.83 ( 0.13) 0.84 ( 0.13) 0.82 ( 0.14) 0.83 ( 0.14) 0.84 ( 0.13) 0.82 ( 0.14) 0.84 ( 0.12) 0.83 ( 0.14) 1.00 ( 0.00) 0.83 ( 0.12) 0.83 ( 0.13) 0.83 ( 0.12) 0.82 ( 0.13) 0.83 ( 0.12) 0.82 ( 0.13) 0.84 ( 0.12) 0.84 ( 0.12) 1.00 ( 0.00) 0.83 ( 0.13) 0.85 ( 0.11) 0.83 ( 0.13) 0.82 ( 0.13) 0.82 ( 0.15) 0.82 ( 0.14) 0.84 ( 0.12) 0.84 ( 0.12) 1.00 ( 0.00) 0.82 ( 0.14) 0.81 ( 0.14) 0.82 ( 0.14) 0.83 ( 0.14) 0.82 ( 0.15) 0.83 ( 0.13) 0.84 ( 0.13) 0.83 ( 0.14) 1.00 ( 0.00) 0.83 ( 0.13) 0.82 ( 0.12) 0.84 ( 0.11) 0.83 ( 0.13) 0.82 ( 0.12) 0.83 ( 0.13) 0.82 ( 0.14) 0.83 ( 0.13) 1.00 ( 0.00) 0.81 ( 0.14) 0.82 ( 0.14) 0.82 ( 0.15) 0.83 ( 0.12) 0.81 ( 0.15) 0.81 ( 0.16) 0.82 ( 0.13) 0.81 ( 0.14) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.85 ( 0.10) 0.84 ( 0.11) 0.87 ( 0.08) 0.84 ( 0.11) 0.84 ( 0.11) 0.84 ( 0.12) 0.85 ( 0.10) 0.85 ( 0.10) 1.00 ( 0.00) 0.84 ( 0.12) 0.85 ( 0.11) 0.85 ( 0.11) 0.83 ( 0.12) 0.84 ( 0.12) 0.85 ( 0.10) 0.85 ( 0.11) 0.84 ( 0.13) 1.00 ( 0.00) 0.85 ( 0.11) 0.85 ( 0.11) 0.84 ( 0.14) 0.84 ( 0.12) 0.84 ( 0.12) 0.86 ( 0.09) 0.85 ( 0.12) 0.85 ( 0.12) 1.00 ( 0.00) 0.85 ( 0.11) 0.85 ( 0.11) 0.85 ( 0.12) 0.84 ( 0.10) 0.84 ( 0.12) 0.84 ( 0.13) 0.85 ( 0.11) 0.84 ( 0.12) 1.00 ( 0.00) 0.83 ( 0.13) 0.84 ( 0.13) 0.83 ( 0.14) 0.84 ( 0.11) 0.83 ( 0.11) 0.84 ( 0.14) 0.85 ( 0.11) 0.84 ( 0.12) 1.00 ( 0.00) 0.84 ( 0.13) 0.84 ( 0.11) 0.85 ( 0.11) 0.85 ( 0.12) 0.84 ( 0.11) 0.86 ( 0.11) 0.85 ( 0.10) 0.84 ( 0.12) 1.00 ( 0.00) 0.84 ( 0.11) 0.84 ( 0.10) 0.85 ( 0.11) 0.84 ( 0.11) 0.84 ( 0.10) 0.83 ( 0.12) 0.83 ( 0.12) 0.84 ( 0.11) 1.00 ( 0.00) Weighted Kendall's tau (a) Rank agreement. clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.97 ( 0.06) 0.97 ( 0.06) 0.97 ( 0.05) 0.97 ( 0.05) 0.97 ( 0.05) 0.97 ( 0.06) 0.98 ( 0.05) 0.97 ( 0.06) 1.00 ( 0.00) 0.95 ( 0.08) 0.96 ( 0.08) 0.96 ( 0.06) 0.96 ( 0.06) 0.97 ( 0.05) 0.96 ( 0.07) 0.97 ( 0.06) 0.95 ( 0.08) 1.00 ( 0.00) 0.97 ( 0.06) 0.97 ( 0.05) 0.96 ( 0.08) 0.95 ( 0.08) 0.96 ( 0.06) 0.97 ( 0.05) 0.96 ( 0.08) 0.97 ( 0.06) 1.00 ( 0.00) 0.97 ( 0.06) 0.96 ( 0.07) 0.96 ( 0.08) 0.97 ( 0.05) 0.97 ( 0.05) 0.96 ( 0.06) 0.97 ( 0.05) 0.97 ( 0.06) 1.00 ( 0.00) 0.96 ( 0.07) 0.97 ( 0.06) 0.97 ( 0.06) 0.97 ( 0.05) 0.96 ( 0.06) 0.96 ( 0.08) 0.96 ( 0.07) 0.96 ( 0.07) 1.00 ( 0.00) 0.97 ( 0.06) 0.97 ( 0.05) 0.97 ( 0.06) 0.97 ( 0.05) 0.95 ( 0.08) 0.96 ( 0.08) 0.97 ( 0.06) 0.97 ( 0.06) 1.00 ( 0.00) 0.97 ( 0.06) 0.98 ( 0.05) 0.96 ( 0.07) 0.96 ( 0.06) 0.97 ( 0.05) 0.97 ( 0.06) 0.97 ( 0.05) 0.97 ( 0.06) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.96 ( 0.07) 0.97 ( 0.05) 0.97 ( 0.04) 0.97 ( 0.05) 0.97 ( 0.05) 0.97 ( 0.05) 0.97 ( 0.05) 0.96 ( 0.07) 1.00 ( 0.00) 0.96 ( 0.08) 0.96 ( 0.07) 0.96 ( 0.07) 0.97 ( 0.05) 0.97 ( 0.04) 0.96 ( 0.07) 0.97 ( 0.05) 0.96 ( 0.08) 1.00 ( 0.00) 0.98 ( 0.04) 0.97 ( 0.04) 0.97 ( 0.06) 0.96 ( 0.07) 0.97 ( 0.05) 0.97 ( 0.04) 0.96 ( 0.07) 0.98 ( 0.04) 1.00 ( 0.00) 0.98 ( 0.04) 0.97 ( 0.06) 0.96 ( 0.06) 0.97 ( 0.05) 0.97 ( 0.05) 0.96 ( 0.07) 0.97 ( 0.04) 0.98 ( 0.04) 1.00 ( 0.00) 0.97 ( 0.06) 0.96 ( 0.07) 0.96 ( 0.06) 0.97 ( 0.05) 0.97 ( 0.05) 0.97 ( 0.06) 0.97 ( 0.06) 0.97 ( 0.06) 1.00 ( 0.00) 0.97 ( 0.05) 0.97 ( 0.06) 0.97 ( 0.05) 0.97 ( 0.04) 0.96 ( 0.07) 0.96 ( 0.06) 0.96 ( 0.07) 0.97 ( 0.05) 1.00 ( 0.00) 0.96 ( 0.07) 0.97 ( 0.05) 0.96 ( 0.07) 0.97 ( 0.05) 0.97 ( 0.05) 0.96 ( 0.06) 0.97 ( 0.06) 0.96 ( 0.07) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.96 ( 0.07) 0.95 ( 0.07) 0.95 ( 0.08) 0.96 ( 0.08) 0.96 ( 0.05) 0.95 ( 0.08) 0.95 ( 0.08) 0.96 ( 0.07) 1.00 ( 0.00) 0.96 ( 0.06) 0.95 ( 0.08) 0.96 ( 0.08) 0.96 ( 0.08) 0.96 ( 0.07) 0.95 ( 0.08) 0.95 ( 0.07) 0.96 ( 0.06) 1.00 ( 0.00) 0.95 ( 0.07) 0.95 ( 0.07) 0.95 ( 0.07) 0.94 ( 0.07) 0.94 ( 0.09) 0.95 ( 0.08) 0.95 ( 0.08) 0.95 ( 0.07) 1.00 ( 0.00) 0.95 ( 0.09) 0.95 ( 0.07) 0.95 ( 0.08) 0.94 ( 0.09) 0.96 ( 0.08) 0.96 ( 0.08) 0.95 ( 0.07) 0.95 ( 0.09) 1.00 ( 0.00) 0.96 ( 0.08) 0.95 ( 0.08) 0.96 ( 0.07) 0.96 ( 0.05) 0.96 ( 0.08) 0.95 ( 0.07) 0.95 ( 0.07) 0.96 ( 0.08) 1.00 ( 0.00) 0.95 ( 0.07) 0.95 ( 0.08) 0.95 ( 0.08) 0.96 ( 0.07) 0.94 ( 0.07) 0.95 ( 0.08) 0.95 ( 0.08) 0.95 ( 0.07) 1.00 ( 0.00) 0.94 ( 0.09) 0.95 ( 0.08) 0.95 ( 0.08) 0.94 ( 0.09) 0.94 ( 0.09) 0.96 ( 0.07) 0.95 ( 0.08) 0.94 ( 0.09) 1.00 ( 0.00) Importance agreement (b) Importance agreement. Figure E.8: X-SKIT agreement results on the CUB dataset as a function of conditioning set size, s. Results are reported as means and standard deviations over the random 100 images used in the experiment. for each class. For each image, we select 14 concepts to test: we first restrict ourselves to annotations with good confidence (i.e., probably or definitely ), and then, for each concept j, we estimate its marginal (i.e., pj := P[Zj = 1]) and class-conditional (i.e., pj|y := P[Zj = 1 | Y = y]) rates over the dataset. Finally, for each test image x with label y, we score concepts by the difference between their class-conditional and marginal rates, i.e. sj(y) = pj|y pj. Intuitively, a large value of sj(y) indicates that concept j has a higher occurrence in class y compared to the population, and we say that it is discriminative for class y. We test the 7 most discriminative concepts that are present in the observed image x, and a random subset of 7 concepts that are absent according to the ground-truth annotations. We remark that, since concepts are binary, we do not use the KDE-based methods presented in Appendix E.1, and instead we approximate PH|ZC=z C by sampling uniformly from the entries in the dataset that match the conditioning vector z C. Finally, we use RBF kernels with bandwidths set to the median of the pairwise distances of observations, and τ max = 200. After running X-SKIT, we classify concepts as important by thresholding their rejection rates at level α which is a statistically-valid way of selecting important concepts. Table E.5 summarizes agreement and detection results as a function of conditioning set size s (i.e., the number of concepts in S [m] \ {j}), and Fig. E.8 includes all pairwise agreement values. We note that the f1 score for s = 4 (i.e., when conditioning on 4 concepts) is lower compared to s = 1, 2. This is expected, as the more concepts one conditions on, the smaller the effect of including one additional concept. Finally, Fig. E.9 shows ranks of importance for all models on the 4 example images used in the main body of the paper. 0.0 0.5 1.0 blue under tail (p) blue crown (p) blue forehead (p) grey underparts (p) grey breast (p) multi-colored breast (p) is iridescent (a) red bill (a) white bill (a) pink nape (a) is green (a) small (5 - 9 in) (p) pink under tail (a) iridescent underparts (a) 0.0 0.5 1.0 blue under tail (p) blue forehead (p) blue crown (p) grey breast (p) grey underparts (p) multi-colored breast (p) small (5 - 9 in) (p) red bill (a) white bill (a) pink nape (a) is green (a) is iridescent (a) pink under tail (a) iridescent underparts (a) CLIP:Vi T-B/32 0.0 0.5 1.0 blue forehead (p) blue crown (p) blue under tail (p) grey breast (p) grey underparts (p) multi-colored breast (p) iridescent underparts (a) pink under tail (a) red bill (a) white bill (a) pink nape (a) is green (a) is iridescent (a) small (5 - 9 in) (p) CLIP:Vi T-L/14 0.0 0.5 1.0 blue under tail (p) blue forehead (p) blue crown (p) grey underparts (p) grey breast (p) multi-colored breast (p) red bill (a) white bill (a) pink nape (a) is green (a) is iridescent (a) small (5 - 9 in) (p) pink under tail (a) iridescent underparts (a) Open Clip:Vi T-B-32 0.0 0.5 1.0 blue crown (p) blue forehead (p) blue under tail (p) grey breast (p) grey underparts (p) multi-colored breast (p) red bill (a) white bill (a) pink nape (a) is green (a) is iridescent (a) small (5 - 9 in) (p) pink under tail (a) iridescent underparts (a) Open Clip:Vi T-L-14 0.0 0.5 1.0 blue under tail (p) blue forehead (p) blue crown (p) grey underparts (p) grey breast (p) multi-colored breast (p) red bill (a) white bill (a) pink nape (a) is green (a) is iridescent (a) small (5 - 9 in) (p) pink under tail (a) iridescent underparts (a) 0.0 0.5 1.0 blue under tail (p) blue crown (p) blue forehead (p) grey breast (p) grey underparts (p) multi-colored breast (p) small (5 - 9 in) (p) red bill (a) white bill (a) pink nape (a) is green (a) is iridescent (a) pink under tail (a) iridescent underparts (a) 0.0 0.5 1.0 blue crown (p) blue forehead (p) blue under tail (p) grey breast (p) grey underparts (p) multi-colored breast (p) small (5 - 9 in) (p) red bill (a) white bill (a) pink nape (a) is green (a) is iridescent (a) pink under tail (a) iridescent underparts (a) rejection rate rejection time 0.0 0.5 1.0 hooked seabird bill (p) forked tail tail (p) solid wing (p) is black (p) black back (p) black upper tail (p) solid tail (p) all-purpose bill (a) red underparts (a) brown wing (a) striped breast (a) orange belly (a) green nape (a) pointed tail tail (a) 0.0 0.5 1.0 hooked seabird bill (p) forked tail tail (p) solid wing (p) black back (p) is black (p) black upper tail (p) solid tail (p) striped breast (a) red underparts (a) all-purpose bill (a) brown wing (a) orange belly (a) green nape (a) pointed tail tail (a) CLIP:VIT-B/32 0.0 0.5 1.0 hooked seabird bill (p) forked tail tail (p) solid wing (p) black back (p) is black (p) black upper tail (p) solid tail (p) brown wing (a) green nape (a) orange belly (a) all-purpose bill (a) striped breast (a) pointed tail tail (a) red underparts (a) CLIP:VIT-L/14 0.0 0.5 1.0 hooked seabird bill (p) is black (p) forked tail tail (p) solid wing (p) black back (p) black upper tail (p) solid tail (p) striped breast (a) brown wing (a) all-purpose bill (a) orange belly (a) green nape (a) pointed tail tail (a) red underparts (a) Open Clip:Vi T-B-32 0.0 0.5 1.0 hooked seabird bill (p) forked tail tail (p) is black (p) black upper tail (p) black back (p) solid wing (p) solid tail (p) all-purpose bill (a) brown wing (a) pointed tail tail (a) orange belly (a) striped breast (a) green nape (a) red underparts (a) Open Clip:Vi T-L-14 0.0 0.5 1.0 hooked seabird bill (p) forked tail tail (p) solid wing (p) is black (p) black back (p) solid tail (p) black upper tail (p) orange belly (a) brown wing (a) all-purpose bill (a) striped breast (a) green nape (a) pointed tail tail (a) red underparts (a) FLAVA ALIGN 0.0 0.5 1.0 hooked seabird bill (p) forked tail tail (p) is black (p) solid wing (p) black back (p) black upper tail (p) solid tail (p) brown wing (a) pointed tail tail (a) all-purpose bill (a) orange belly (a) striped breast (a) green nape (a) red underparts (a) 0.0 0.5 1.0 hooked seabird bill (p) forked tail tail (p) black back (p) solid wing (p) is black (p) black upper tail (p) solid tail (p) all-purpose bill (a) brown wing (a) orange belly (a) striped breast (a) green nape (a) pointed tail tail (a) red underparts (a) 0.0 0.5 1.0 spatulate bill (p) duck-like (p) olive throat (p) yellow bill (p) medium (9 - 16 in) (p) bill about the same as head (p) multi-colored wing (p) red underparts (a) orange nape (a) buff leg (a) red breast (a) masked head (a) long-legged-like (a) rounded tail tail (a) 0.0 0.5 1.0 spatulate bill (p) duck-like (p) medium (9 - 16 in) (p) olive throat (p) yellow bill (p) bill about the same as head (p) multi-colored wing (p) buff leg (a) masked head (a) rounded tail tail (a) red breast (a) orange nape (a) long-legged-like (a) red underparts (a) CLIP:Vi T-B/32 0.0 0.5 1.0 spatulate bill (p) duck-like (p) olive throat (p) yellow bill (p) medium (9 - 16 in) (p) bill about the same as head (p) multi-colored wing (p) orange nape (a) red breast (a) masked head (a) buff leg (a) long-legged-like (a) rounded tail tail (a) red underparts (a) CLIP:Vi T-L/14 0.0 0.5 1.0 duck-like (p) spatulate bill (p) medium (9 - 16 in) (p) olive throat (p) yellow bill (p) multi-colored wing (p) bill about the same as head (p) long-legged-like (a) masked head (a) red breast (a) buff leg (a) orange nape (a) rounded tail tail (a) red underparts (a) Open Clip:Vi T-B-32 0.0 0.5 1.0 spatulate bill (p) duck-like (p) olive throat (p) medium (9 - 16 in) (p) yellow bill (p) bill about the same as head (p) multi-colored wing (p) red underparts (a) red breast (a) masked head (a) buff leg (a) orange nape (a) long-legged-like (a) rounded tail tail (a) Open Clip:Vi T-L-14 0.0 0.5 1.0 spatulate bill (p) duck-like (p) yellow bill (p) medium (9 - 16 in) (p) olive throat (p) bill about the same as head (p) multi-colored wing (p) buff leg (a) red breast (a) masked head (a) orange nape (a) long-legged-like (a) rounded tail tail (a) red underparts (a) 0.0 0.5 1.0 duck-like (p) spatulate bill (p) olive throat (p) medium (9 - 16 in) (p) yellow bill (p) multi-colored wing (p) bill about the same as head (p) red breast (a) masked head (a) buff leg (a) orange nape (a) long-legged-like (a) rounded tail tail (a) red underparts (a) 0.0 0.5 1.0 spatulate bill (p) duck-like (p) olive throat (p) yellow bill (p) medium (9 - 16 in) (p) multi-colored wing (p) bill about the same as head (p) red breast (a) masked head (a) buff leg (a) orange nape (a) long-legged-like (a) rounded tail tail (a) red underparts (a) 0.0 0.5 1.0 blue underparts (p) blue upperparts (p) black underparts (p) black breast (p) yellow eye (p) plain head (p) black bill (p) iridescent breast (a) brown eye (a) pink eye (a) green leg (a) duck-like (a) pigeon-like (a) swallow-like (a) 0.0 0.5 1.0 blue underparts (p) blue upperparts (p) black underparts (p) black breast (p) yellow eye (p) plain head (p) black bill (p) swallow-like (a) iridescent breast (a) brown eye (a) pigeon-like (a) pink eye (a) green leg (a) duck-like (a) CLIP:Vi T-B/32 0.0 0.5 1.0 blue underparts (p) blue upperparts (p) black underparts (p) black breast (p) yellow eye (p) plain head (p) black bill (p) swallow-like (a) iridescent breast (a) brown eye (a) pink eye (a) green leg (a) duck-like (a) pigeon-like (a) CLIP:Vi T-L/14 0.0 0.5 1.0 blue underparts (p) blue upperparts (p) black underparts (p) black breast (p) yellow eye (p) plain head (p) green leg (a) black bill (p) iridescent breast (a) brown eye (a) pink eye (a) duck-like (a) pigeon-like (a) swallow-like (a) Open Clip:Vi T-B-32 0.0 0.5 1.0 blue underparts (p) blue upperparts (p) black underparts (p) black breast (p) yellow eye (p) plain head (p) black bill (p) brown eye (a) iridescent breast (a) pink eye (a) green leg (a) duck-like (a) pigeon-like (a) swallow-like (a) Open Clip:Vi T-L-14 0.0 0.5 1.0 blue underparts (p) blue upperparts (p) black underparts (p) black breast (p) yellow eye (p) plain head (p) black bill (p) iridescent breast (a) brown eye (a) pink eye (a) green leg (a) duck-like (a) pigeon-like (a) swallow-like (a) 0.0 0.5 1.0 blue underparts (p) blue upperparts (p) black underparts (p) yellow eye (p) black breast (p) plain head (p) black bill (p) duck-like (a) iridescent breast (a) brown eye (a) pink eye (a) green leg (a) pigeon-like (a) swallow-like (a) 0.0 0.5 1.0 blue underparts (p) black underparts (p) black breast (p) blue upperparts (p) plain head (p) yellow eye (p) black bill (p) swallow-like (a) iridescent breast (a) brown eye (a) pink eye (a) green leg (a) duck-like (a) pigeon-like (a) Figure E.9: X-SKIT importance results (s = 1) across all models for four example images in the CUB dataset. church French horn cassette player English springer tench parachute golf ball gas pump garbage truck chainsaw Figure E.10: Example images from the Imagenette dataset with their respective labels. Table E.6: Zero-shot classification accuracy on the Imagenette dataset. Model tench English springer cassette player French horn church CLIP:RN50 99.48% 99.49% 95.80% 99.49% 100.00% CLIP:Vi T-B/32 99.74% 99.24% 96.08% 98.73% 100.00% CLIP:Vi T-L/14 99.22% 99.75% 99.72% 99.75% 100.00% Open Clip:Vi T-B-32 100.00% 100.00% 98.32% 99.24% 99.76% Open Clip:Vi T-L-14 100.00% 100.00% 99.72% 99.75% 100.00% FLAVA 99.74% 99.24% 97.76% 98.22% 100.00% ALIGN 99.74% 100.00% 99.44% 99.75% 100.00% BLIP 100.00% 98.48% 93.56% 99.75% 100.00% average 99.74% 0.26% 99.53% 0.50% 97.55% 2.09% 99.33% 0.54% 99.97% 0.08% Model parachute golf ball gas pump garbage truck chainsaw CLIP:RN50 99.74% 97.74% 91.65% 98.71% 96.63% CLIP:Vi T-B/32 98.97% 99.25% 97.61% 99.49% 99.22% CLIP:Vi T-L/14 99.74% 99.75% 100.00% 99.74% 99.74% Open Clip:Vi T-B-32 99.49% 99.25% 97.61% 99.23% 97.67% Open Clip:Vi T-L-14 100.00% 99.50% 99.28% 99.49% 98.96% FLAVA 98.72% 99.00% 97.61% 99.23% 96.37% ALIGN 99.49% 99.75% 99.28% 99.49% 98.96% BLIP 99.49% 99.50% 98.09% 99.23% 98.19% average 99.46% 0.39% 99.22% 0.61% 97.64% 2.43% 99.33% 0.29% 98.22% 1.15% E.4 Imagenette Dataset Here, we present additional information and results on the Imagenette dataset presented in the main body of this manuscript.5 This dataset contains 13,394 images (9,469 for training and 3,925 for testing) from ten easily separable classes in Image Net [22]. Fig. E.10 includes example images from the classes in the dataset, and Table E.6 reports the classification accuracy across all vision-language models (98.99% 0.01% average accuracy). Recall that the Image Net dataset does not provide ground-truth semantic annotations, hence we use Sp Li Ce [6] to find which concepts to test. In particular, we use the 10,000 most frequent words in the vocabulary from the MSCOCO dataset [40], and we set the ℓ1 regularization term in Sp Li Ce to 0.20. Following previous work [85], we filter the selected concepts such that they are different from the classes in the dataset. We use Word Net [23] to lemmatize both concepts and class names (e.g., churches becomes church ), and we check that concepts are not contained in class names and vice versa. For example, the concept churches would be skipped because church is already the name of a class, and gasoline would be skipped because it contains part of the class gas pump . 5The Imagenette dataset is available at https://github.com/fastai/imagenette. 0.0 0.5 1.0 battered fishing fish spaniel bass brass obsolete instrument flew exterior putting cathedral airshow band selling radio fore dispenser trumpet jazz 0.0 0.5 1.0 spaniel fore selling instrument airshow radio jazz putting battered obsolete fish fishing dispenser bass flew trumpet exterior cathedral band brass English springer 0.0 0.5 1.0 obsolete instrument radio dispenser band selling jazz trumpet bass flew airshow spaniel brass putting exterior fishing fore fish cathedral battered cassette player 0.0 0.5 1.0 trumpet instrument band jazz brass bass fishing obsolete cathedral radio dispenser exterior flew airshow selling fish battered fore putting spaniel French horn 0.0 0.5 1.0 cathedral exterior obsolete band trumpet instrument radio bass jazz brass airshow dispenser fish selling spaniel flew battered putting fishing fore rejection rate significance level rejection time 0.0 0.5 1.0 flew airshow exterior cathedral trumpet fore putting brass fish dispenser battered jazz radio band obsolete fishing spaniel bass instrument selling 0.0 0.5 1.0 fore putting flew battered spaniel exterior cathedral fishing jazz airshow radio fish band brass selling dispenser trumpet bass obsolete instrument 0.0 0.5 1.0 dispenser obsolete exterior instrument trumpet battered fore airshow radio flew selling putting band brass fishing cathedral bass fish jazz spaniel 0.0 0.5 1.0 obsolete dispenser instrument band exterior trumpet fishing spaniel jazz airshow radio bass selling fish flew fore battered cathedral brass putting garbage truck 0.0 0.5 1.0 obsolete instrument dispenser jazz radio trumpet band airshow battered exterior fore brass fishing cathedral flew fish spaniel bass selling putting (a) Global importance with SKIT. 0.0 0.5 1.0 fish bass brass fishing band jazz battered instrument trumpet obsolete spaniel dispenser exterior fore cathedral radio flew airshow putting selling 0.0 0.5 1.0 spaniel band brass bass jazz instrument trumpet obsolete fish dispenser fishing fore exterior battered radio flew airshow putting cathedral selling English springer 0.0 0.5 1.0 obsolete radio instrument bass band brass trumpet jazz dispenser fish fishing spaniel exterior selling flew cathedral fore airshow putting battered cassette player 0.0 0.5 1.0 trumpet band brass bass jazz instrument obsolete fishing fish dispenser exterior spaniel radio cathedral fore flew airshow battered putting selling French horn 0.0 0.5 1.0 cathedral exterior bass band brass obsolete instrument trumpet jazz fish fishing dispenser spaniel radio fore battered airshow flew selling putting 0.0 0.5 1.0 flew airshow brass bass trumpet band exterior obsolete jazz instrument fish fishing cathedral dispenser spaniel fore radio putting battered selling 0.0 0.5 1.0 fore putting bass obsolete brass band jazz instrument trumpet dispenser fish spaniel fishing exterior flew battered radio cathedral airshow selling 0.0 0.5 1.0 dispenser bass obsolete trumpet band brass instrument jazz fish exterior fishing spaniel radio cathedral battered fore airshow putting flew selling 0.0 0.5 1.0 obsolete band bass brass trumpet instrument jazz dispenser exterior fishing spaniel fish cathedral airshow fore flew radio battered putting selling garbage truck 0.0 0.5 1.0 instrument bass jazz brass band obsolete trumpet dispenser fishing fish battered spaniel exterior fore flew cathedral radio airshow putting selling rejection rate significance level rejection time (b) Global conditional importance with C-SKIT. Figure E.11: Global importance results with CLIP:Vi T-L/14 on the Imagenette dataset. clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.72 ( 0.10) 0.61 ( 0.11) 0.62 ( 0.14) 0.62 ( 0.12) 0.55 ( 0.20) 0.35 ( 0.26) 0.54 ( 0.19) 0.72 ( 0.11) 1.00 ( 0.00) 0.61 ( 0.17) 0.60 ( 0.12) 0.58 ( 0.10) 0.51 ( 0.22) 0.39 ( 0.20) 0.51 ( 0.16) 0.62 ( 0.12) 0.63 ( 0.17) 1.00 ( 0.00) 0.45 ( 0.21) 0.56 ( 0.15) 0.47 ( 0.26) 0.39 ( 0.24) 0.45 ( 0.14) 0.61 ( 0.14) 0.61 ( 0.11) 0.45 ( 0.19) 1.00 ( 0.00) 0.60 ( 0.11) 0.55 ( 0.13) 0.44 ( 0.18) 0.50 ( 0.20) 0.61 ( 0.15) 0.58 ( 0.11) 0.57 ( 0.14) 0.57 ( 0.15) 1.00 ( 0.00) 0.57 ( 0.25) 0.42 ( 0.19) 0.50 ( 0.20) 0.57 ( 0.21) 0.55 ( 0.20) 0.49 ( 0.19) 0.56 ( 0.17) 0.59 ( 0.22) 1.00 ( 0.00) 0.34 ( 0.23) 0.65 ( 0.09) 0.39 ( 0.22) 0.41 ( 0.20) 0.35 ( 0.30) 0.42 ( 0.21) 0.46 ( 0.16) 0.29 ( 0.32) 1.00 ( 0.00) 0.27 ( 0.24) 0.51 ( 0.21) 0.49 ( 0.19) 0.42 ( 0.15) 0.47 ( 0.21) 0.47 ( 0.24) 0.65 ( 0.11) 0.30 ( 0.19) 1.00 ( 0.00) SKIT c-SKIT PCBM clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.87 ( 0.04) 0.74 ( 0.09) 0.56 ( 0.08) 0.66 ( 0.09) 0.57 ( 0.14) 0.57 ( 0.07) 0.44 ( 0.20) 0.87 ( 0.04) 1.00 ( 0.00) 0.74 ( 0.07) 0.60 ( 0.10) 0.68 ( 0.09) 0.57 ( 0.13) 0.64 ( 0.05) 0.40 ( 0.24) 0.73 ( 0.13) 0.72 ( 0.11) 1.00 ( 0.00) 0.45 ( 0.16) 0.56 ( 0.09) 0.56 ( 0.13) 0.61 ( 0.06) 0.42 ( 0.16) 0.52 ( 0.09) 0.57 ( 0.11) 0.43 ( 0.19) 1.00 ( 0.00) 0.62 ( 0.10) 0.51 ( 0.15) 0.40 ( 0.14) 0.45 ( 0.14) 0.63 ( 0.08) 0.66 ( 0.09) 0.60 ( 0.08) 0.63 ( 0.08) 1.00 ( 0.00) 0.59 ( 0.11) 0.41 ( 0.14) 0.44 ( 0.15) 0.59 ( 0.15) 0.59 ( 0.14) 0.58 ( 0.11) 0.52 ( 0.09) 0.55 ( 0.12) 1.00 ( 0.00) 0.45 ( 0.18) 0.60 ( 0.15) 0.56 ( 0.09) 0.60 ( 0.08) 0.52 ( 0.15) 0.36 ( 0.18) 0.41 ( 0.10) 0.33 ( 0.25) 1.00 ( 0.00) 0.31 ( 0.23) 0.41 ( 0.22) 0.41 ( 0.23) 0.42 ( 0.16) 0.42 ( 0.18) 0.40 ( 0.19) 0.61 ( 0.16) 0.32 ( 0.23) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.68 ( 0.11) 0.61 ( 0.17) 0.49 ( 0.15) 0.50 ( 0.16) 0.40 ( 0.13) 0.47 ( 0.15) 0.45 ( 0.16) 0.69 ( 0.10) 1.00 ( 0.00) 0.65 ( 0.09) 0.56 ( 0.09) 0.49 ( 0.18) 0.46 ( 0.16) 0.50 ( 0.19) 0.44 ( 0.15) 0.60 ( 0.19) 0.63 ( 0.12) 1.00 ( 0.00) 0.48 ( 0.20) 0.54 ( 0.18) 0.32 ( 0.24) 0.50 ( 0.13) 0.37 ( 0.21) 0.50 ( 0.13) 0.51 ( 0.13) 0.42 ( 0.21) 1.00 ( 0.00) 0.42 ( 0.22) 0.37 ( 0.14) 0.40 ( 0.25) 0.42 ( 0.13) 0.49 ( 0.17) 0.46 ( 0.17) 0.53 ( 0.15) 0.44 ( 0.25) 1.00 ( 0.00) 0.36 ( 0.17) 0.48 ( 0.12) 0.28 ( 0.25) 0.35 ( 0.21) 0.47 ( 0.16) 0.34 ( 0.26) 0.29 ( 0.25) 0.35 ( 0.22) 1.00 ( 0.00) 0.39 ( 0.17) 0.30 ( 0.25) 0.47 ( 0.19) 0.55 ( 0.13) 0.48 ( 0.11) 0.37 ( 0.34) 0.51 ( 0.13) 0.37 ( 0.17) 1.00 ( 0.00) 0.35 ( 0.26) 0.43 ( 0.18) 0.38 ( 0.21) 0.38 ( 0.21) 0.42 ( 0.15) 0.31 ( 0.22) 0.33 ( 0.24) 0.38 ( 0.28) 1.00 ( 0.00) Weighted Kendall's tau Figure E.12: Rank agreement comparison between SKIT, C-SKIT, and PCBM on Imagenette. Result are reported as mean and standard deviation over the 10 classes considered in the experiment. 20 19 18 17 16 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 battered fishing fish spaniel bass obsolete brass instrument flew exterior putting cathedral dispenser airshow band trumpet radio selling fore jazz spaniel fore selling instrument radio putting airshow battered obsolete jazz fishing bass fish flew dispenser cathedral trumpet band exterior brass English springer obsolete instrument radio dispenser band selling jazz trumpet bass flew airshow spaniel fishing brass exterior putting fore fish cathedral battered cassette player trumpet band jazz instrument brass bass fishing obsolete cathedral radio exterior dispenser airshow flew selling fish battered fore putting spaniel French horn cathedral exterior obsolete instrument band trumpet radio jazz bass airshow brass fish dispenser spaniel selling flew battered putting fishing fore 100 200 400 800 1600 max 20 19 18 17 16 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 flew airshow exterior cathedral fore trumpet putting brass fish dispenser battered radio jazz band obsolete spaniel fishing bass instrument selling 100 200 400 800 1600 max fore putting flew battered spaniel cathedral fishing airshow exterior jazz fish radio brass dispenser band obsolete selling trumpet instrument bass 100 200 400 800 1600 max dispenser obsolete exterior instrument trumpet battered fore airshow radio selling putting flew band brass fishing cathedral jazz fish bass spaniel 100 200 400 800 1600 max obsolete dispenser instrument exterior trumpet band fishing spaniel jazz airshow bass radio fish selling fore flew battered cathedral brass putting garbage truck 100 200 400 800 1600 max obsolete instrument dispenser jazz radio airshow band trumpet exterior battered fore cathedral brass fishing spaniel fish flew bass selling putting (a) Global importance with SKIT. 20 19 18 17 16 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 bass fish brass fishing band trumpet battered jazz instrument obsolete spaniel dispenser exterior fore cathedral radio flew airshow putting selling spaniel brass band bass instrument trumpet jazz obsolete fish fishing exterior fore dispenser flew radio battered cathedral airshow putting selling English springer obsolete radio instrument bass brass band jazz trumpet fishing dispenser fish exterior spaniel selling fore flew cathedral putting battered airshow cassette player trumpet band instrument brass jazz bass obsolete fishing fish dispenser exterior spaniel radio flew cathedral fore battered airshow putting selling French horn cathedral exterior bass band brass obsolete instrument trumpet jazz fish fishing dispenser spaniel radio fore flew airshow battered putting selling 100 200 400 800 1600 max 20 19 18 17 16 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 flew airshow bass brass band trumpet jazz exterior instrument obsolete fish cathedral fore fishing spaniel dispenser radio putting battered selling 100 200 400 800 1600 max fore putting bass brass band obsolete trumpet jazz instrument dispenser fish fishing exterior flew spaniel cathedral radio battered airshow selling 100 200 400 800 1600 max dispenser obsolete bass band brass trumpet instrument jazz exterior fish fishing radio spaniel cathedral battered fore putting flew airshow selling 100 200 400 800 1600 max obsolete bass band trumpet brass instrument jazz dispenser exterior fishing fish spaniel cathedral airshow radio flew fore battered putting selling garbage truck 100 200 400 800 1600 max instrument bass brass band jazz obsolete trumpet fish fishing dispenser fore exterior battered spaniel radio cathedral flew airshow putting selling (b) Global conditional importance with C-SKIT. Figure E.13: Importance results with CLIP:Vi T-L/14 on Imagenette as a function of τ max. We run SKIT and C-SKIT with RBF kernels with bandwidths equal to the 90th percentile of the pairwise distances of previous observations and τ max = 400, 800, respectively. We encode the entire dataset using Sp Li Ce and keep the top-20 concepts to test. Fig. E.11 shows global and global conditional importance results with CLIP:Vi T-L/14 for all classes in the dataset. Furthermore, Fig. E.12 shows all pairwise rank agreement comparisons for SKIT, C-SKIT, and PCBM across all 8 VL models used in the experiment. Lastly, Fig. E.13 qualitatively shows ranks of importance as a function of τ max with CLIP:Vi T-L/14. We use X-SKIT on 2 random images from all classes in the dataset (20 images total). Recall that we use Sp Li Ce to encode each image and keep the top-10 concepts, and finally add the bottom-4 concepts according to PCBM, for a total of 14 concepts per image. We set the bandwidth of the RBF kernel used in the test to the median of the distance of the observations, and τ max = 200. As in the CUB dataset experiment, we classify concepts as important by thresholding their rejection rates at Table E.7: X-SKIT results on the Imagenette dataset as a function of conditioning set size s. s Rank agreement Importance agreement 1 0.59 0.21 0.71 0.14 2 0.56 0.21 0.67 0.13 4 0.53 0.23 0.68 0.14 clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.75 ( 0.09) 0.65 ( 0.18) 0.67 ( 0.15) 0.59 ( 0.24) 0.59 ( 0.12) 0.62 ( 0.17) 0.53 ( 0.25) 0.75 ( 0.09) 1.00 ( 0.00) 0.68 ( 0.17) 0.68 ( 0.13) 0.62 ( 0.17) 0.59 ( 0.13) 0.61 ( 0.18) 0.56 ( 0.25) 0.65 ( 0.15) 0.68 ( 0.15) 1.00 ( 0.00) 0.57 ( 0.24) 0.58 ( 0.25) 0.56 ( 0.19) 0.59 ( 0.20) 0.53 ( 0.24) 0.63 ( 0.20) 0.65 ( 0.18) 0.56 ( 0.30) 1.00 ( 0.00) 0.65 ( 0.24) 0.51 ( 0.20) 0.57 ( 0.24) 0.51 ( 0.26) 0.60 ( 0.17) 0.60 ( 0.16) 0.57 ( 0.24) 0.61 ( 0.26) 1.00 ( 0.00) 0.56 ( 0.21) 0.62 ( 0.16) 0.58 ( 0.19) 0.59 ( 0.11) 0.59 ( 0.12) 0.55 ( 0.22) 0.54 ( 0.17) 0.54 ( 0.21) 1.00 ( 0.00) 0.56 ( 0.17) 0.50 ( 0.23) 0.61 ( 0.20) 0.58 ( 0.21) 0.59 ( 0.17) 0.55 ( 0.28) 0.62 ( 0.17) 0.55 ( 0.23) 1.00 ( 0.00) 0.57 ( 0.21) 0.51 ( 0.27) 0.54 ( 0.26) 0.50 ( 0.30) 0.51 ( 0.29) 0.57 ( 0.22) 0.50 ( 0.25) 0.59 ( 0.16) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.71 ( 0.11) 0.60 ( 0.17) 0.63 ( 0.15) 0.54 ( 0.26) 0.55 ( 0.13) 0.58 ( 0.16) 0.54 ( 0.23) 0.72 ( 0.11) 1.00 ( 0.00) 0.65 ( 0.15) 0.67 ( 0.12) 0.58 ( 0.20) 0.55 ( 0.15) 0.60 ( 0.14) 0.55 ( 0.31) 0.62 ( 0.13) 0.65 ( 0.14) 1.00 ( 0.00) 0.57 ( 0.20) 0.53 ( 0.23) 0.53 ( 0.18) 0.58 ( 0.19) 0.50 ( 0.24) 0.62 ( 0.16) 0.65 ( 0.15) 0.53 ( 0.24) 1.00 ( 0.00) 0.61 ( 0.26) 0.49 ( 0.18) 0.56 ( 0.22) 0.50 ( 0.24) 0.56 ( 0.18) 0.56 ( 0.19) 0.53 ( 0.25) 0.60 ( 0.22) 1.00 ( 0.00) 0.50 ( 0.22) 0.61 ( 0.14) 0.55 ( 0.21) 0.56 ( 0.12) 0.58 ( 0.12) 0.53 ( 0.23) 0.52 ( 0.17) 0.48 ( 0.27) 1.00 ( 0.00) 0.52 ( 0.17) 0.49 ( 0.21) 0.58 ( 0.16) 0.60 ( 0.13) 0.56 ( 0.20) 0.57 ( 0.22) 0.63 ( 0.13) 0.51 ( 0.19) 1.00 ( 0.00) 0.60 ( 0.16) 0.51 ( 0.28) 0.54 ( 0.31) 0.46 ( 0.31) 0.49 ( 0.31) 0.56 ( 0.24) 0.44 ( 0.26) 0.59 ( 0.18) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.64 ( 0.12) 0.58 ( 0.16) 0.63 ( 0.15) 0.52 ( 0.25) 0.57 ( 0.13) 0.55 ( 0.17) 0.48 ( 0.24) 0.64 ( 0.14) 1.00 ( 0.00) 0.57 ( 0.16) 0.64 ( 0.13) 0.56 ( 0.25) 0.51 ( 0.14) 0.58 ( 0.17) 0.45 ( 0.28) 0.57 ( 0.16) 0.56 ( 0.16) 1.00 ( 0.00) 0.54 ( 0.25) 0.51 ( 0.28) 0.49 ( 0.22) 0.54 ( 0.25) 0.45 ( 0.27) 0.61 ( 0.18) 0.62 ( 0.15) 0.52 ( 0.27) 1.00 ( 0.00) 0.62 ( 0.25) 0.52 ( 0.17) 0.54 ( 0.19) 0.48 ( 0.26) 0.50 ( 0.28) 0.54 ( 0.24) 0.52 ( 0.26) 0.58 ( 0.29) 1.00 ( 0.00) 0.48 ( 0.29) 0.59 ( 0.22) 0.53 ( 0.23) 0.53 ( 0.17) 0.49 ( 0.16) 0.48 ( 0.22) 0.51 ( 0.20) 0.47 ( 0.28) 1.00 ( 0.00) 0.53 ( 0.16) 0.49 ( 0.18) 0.53 ( 0.22) 0.57 ( 0.16) 0.53 ( 0.23) 0.55 ( 0.19) 0.58 ( 0.24) 0.54 ( 0.16) 1.00 ( 0.00) 0.56 ( 0.19) 0.43 ( 0.32) 0.44 ( 0.31) 0.45 ( 0.29) 0.45 ( 0.31) 0.53 ( 0.23) 0.44 ( 0.29) 0.58 ( 0.17) 1.00 ( 0.00) Weighted Kendall's tau (a) Rank agreement. clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.80 ( 0.09) 0.75 ( 0.10) 0.74 ( 0.14) 0.74 ( 0.14) 0.69 ( 0.10) 0.69 ( 0.15) 0.69 ( 0.12) 0.80 ( 0.09) 1.00 ( 0.00) 0.73 ( 0.13) 0.77 ( 0.14) 0.77 ( 0.10) 0.73 ( 0.11) 0.70 ( 0.14) 0.72 ( 0.13) 0.75 ( 0.10) 0.73 ( 0.13) 1.00 ( 0.00) 0.69 ( 0.12) 0.69 ( 0.13) 0.69 ( 0.10) 0.66 ( 0.13) 0.67 ( 0.13) 0.74 ( 0.14) 0.77 ( 0.14) 0.69 ( 0.12) 1.00 ( 0.00) 0.76 ( 0.13) 0.66 ( 0.12) 0.64 ( 0.17) 0.70 ( 0.17) 0.74 ( 0.14) 0.77 ( 0.10) 0.69 ( 0.13) 0.76 ( 0.13) 1.00 ( 0.00) 0.71 ( 0.15) 0.71 ( 0.14) 0.72 ( 0.14) 0.69 ( 0.10) 0.73 ( 0.11) 0.69 ( 0.10) 0.66 ( 0.12) 0.71 ( 0.15) 1.00 ( 0.00) 0.70 ( 0.15) 0.68 ( 0.15) 0.69 ( 0.15) 0.70 ( 0.14) 0.66 ( 0.13) 0.64 ( 0.17) 0.71 ( 0.14) 0.70 ( 0.15) 1.00 ( 0.00) 0.68 ( 0.13) 0.69 ( 0.12) 0.72 ( 0.13) 0.67 ( 0.13) 0.70 ( 0.17) 0.72 ( 0.14) 0.68 ( 0.15) 0.68 ( 0.13) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.77 ( 0.14) 0.70 ( 0.14) 0.71 ( 0.17) 0.69 ( 0.15) 0.66 ( 0.10) 0.66 ( 0.11) 0.64 ( 0.18) 0.77 ( 0.14) 1.00 ( 0.00) 0.69 ( 0.14) 0.73 ( 0.11) 0.70 ( 0.09) 0.67 ( 0.11) 0.64 ( 0.14) 0.66 ( 0.16) 0.70 ( 0.14) 0.69 ( 0.14) 1.00 ( 0.00) 0.66 ( 0.14) 0.64 ( 0.12) 0.70 ( 0.13) 0.66 ( 0.12) 0.65 ( 0.14) 0.71 ( 0.17) 0.73 ( 0.11) 0.66 ( 0.14) 1.00 ( 0.00) 0.74 ( 0.14) 0.65 ( 0.10) 0.63 ( 0.14) 0.66 ( 0.10) 0.69 ( 0.15) 0.70 ( 0.09) 0.64 ( 0.12) 0.74 ( 0.14) 1.00 ( 0.00) 0.64 ( 0.09) 0.67 ( 0.12) 0.65 ( 0.10) 0.66 ( 0.10) 0.67 ( 0.11) 0.70 ( 0.13) 0.65 ( 0.10) 0.64 ( 0.09) 1.00 ( 0.00) 0.67 ( 0.10) 0.63 ( 0.11) 0.66 ( 0.11) 0.64 ( 0.14) 0.66 ( 0.12) 0.63 ( 0.14) 0.67 ( 0.12) 0.67 ( 0.10) 1.00 ( 0.00) 0.66 ( 0.14) 0.64 ( 0.18) 0.66 ( 0.16) 0.65 ( 0.14) 0.66 ( 0.10) 0.65 ( 0.10) 0.63 ( 0.11) 0.66 ( 0.14) 1.00 ( 0.00) clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 1.00 ( 0.00) 0.72 ( 0.12) 0.73 ( 0.15) 0.73 ( 0.13) 0.71 ( 0.17) 0.69 ( 0.11) 0.73 ( 0.11) 0.69 ( 0.14) 0.72 ( 0.12) 1.00 ( 0.00) 0.62 ( 0.13) 0.72 ( 0.13) 0.65 ( 0.12) 0.61 ( 0.14) 0.67 ( 0.14) 0.62 ( 0.11) 0.73 ( 0.15) 0.62 ( 0.13) 1.00 ( 0.00) 0.68 ( 0.15) 0.66 ( 0.15) 0.69 ( 0.12) 0.65 ( 0.15) 0.68 ( 0.12) 0.73 ( 0.13) 0.72 ( 0.13) 0.68 ( 0.15) 1.00 ( 0.00) 0.76 ( 0.09) 0.68 ( 0.15) 0.71 ( 0.11) 0.68 ( 0.12) 0.71 ( 0.17) 0.65 ( 0.12) 0.66 ( 0.15) 0.76 ( 0.09) 1.00 ( 0.00) 0.67 ( 0.13) 0.71 ( 0.10) 0.66 ( 0.16) 0.69 ( 0.11) 0.61 ( 0.14) 0.69 ( 0.12) 0.68 ( 0.15) 0.67 ( 0.13) 1.00 ( 0.00) 0.68 ( 0.12) 0.65 ( 0.11) 0.73 ( 0.11) 0.67 ( 0.14) 0.65 ( 0.15) 0.71 ( 0.11) 0.71 ( 0.10) 0.68 ( 0.12) 1.00 ( 0.00) 0.68 ( 0.16) 0.69 ( 0.14) 0.62 ( 0.11) 0.68 ( 0.12) 0.68 ( 0.12) 0.66 ( 0.16) 0.65 ( 0.11) 0.68 ( 0.16) 1.00 ( 0.00) Importance agreement (b) Importance agreement. Figure E.14: X-SKIT agreement results on the Imagenette dataset across 8 different vision-language models as a function of conditioning set size, s. Results are reported as means and standard deviations over the random 20 images used in the experiment. level α. Table E.7 summarizes rank and importance agreement as a function of conditioning set size s (i.e., the number of concepts in S), and Fig. E.14 includes all pairwise agreement values. We can see that ranks are generally well-aligned across models, and that agreement slightly decreases as the number of conditioning concepts increases. Finally, Figs. E.15 and E.16 include results with X-SKIT for 2 images from three classes in the dataset across all models used in the experiment, and Fig. E.17 summarizes ranks across models on the same images. 0.0 0.5 1.0 spaniel whisky fluffy royalty sibling youngster shamrock patch twin marble fore (*) fishing (*) trumpet (*) cathedral (*) 0.0 0.5 1.0 spaniel whisky fluffy sibling twin shamrock marble royalty patch youngster trumpet (*) fore (*) fishing (*) cathedral (*) CLIP:Vi T-B/32 0.0 0.5 1.0 spaniel sibling fluffy marble royalty twin shamrock whisky youngster trumpet (*) fishing (*) cathedral (*) patch fore (*) CLIP:Vi T-L/14 0.0 0.5 1.0 spaniel fluffy sibling royalty marble twin fore (*) cathedral (*) patch trumpet (*) fishing (*) shamrock whisky youngster Open Clip:Vi T-B-32 0.0 0.5 1.0 spaniel fluffy sibling youngster marble patch cathedral (*) fore (*) twin whisky trumpet (*) fishing (*) royalty shamrock Open Clip:Vi T-L-14 0.0 0.5 1.0 spaniel patch whisky youngster sibling cathedral (*) fluffy royalty trumpet (*) shamrock fishing (*) marble fore (*) twin 0.0 0.5 1.0 spaniel whisky royalty sibling fluffy patch marble twin youngster shamrock fishing (*) fore (*) cathedral (*) trumpet (*) 0.0 0.5 1.0 spaniel trumpet (*) sibling royalty whisky youngster patch fishing (*) fluffy twin shamrock fore (*) cathedral (*) marble 0.0 0.5 1.0 spaniel whisky fluffy royalty sibling youngster shamrock patch twin marble fore (*) fishing (*) trumpet (*) cathedral (*) 0.0 0.5 1.0 spaniel whisky fluffy sibling twin shamrock marble royalty patch youngster trumpet (*) fore (*) fishing (*) cathedral (*) CLIP:Vi T-B/32 0.0 0.5 1.0 spaniel fetch cane stick swimming fishing (*) rescue barge help helping launched trumpet (*) fore (*) cathedral (*) Open Clip:Vi T-B-32 0.0 0.5 1.0 spaniel fetch cane stick helping swimming fishing (*) fore (*) barge launched rescue cathedral (*) help trumpet (*) CLIP:Vi T-L/14 0.0 0.5 1.0 spaniel fetch cane stick swimming helping fishing (*) help fore (*) rescue launched barge trumpet (*) cathedral (*) Open Clip:Vi T-L-14 0.0 0.5 1.0 spaniel stick cane fetch swimming fishing (*) trumpet (*) launched helping fore (*) barge cathedral (*) rescue help 0.0 0.5 1.0 spaniel fetch cane stick rescue fishing (*) fore (*) swimming trumpet (*) helping barge help cathedral (*) launched 0.0 0.5 1.0 spaniel help stick helping launched fishing (*) barge cathedral (*) fore (*) swimming fetch cane rescue trumpet (*) 0.0 0.5 1.0 band quartet bass (*) prom formal ensemble fish (*) dress riverfront graduate bulldog squad obsolete (*) cathedral (*) 0.0 0.5 1.0 band quartet ensemble bass (*) dress squad riverfront formal fish (*) prom obsolete (*) graduate cathedral (*) bulldog CLIP:Vi T-B/32 0.0 0.5 1.0 band ensemble quartet obsolete (*) fish (*) bass (*) cathedral (*) bulldog dress prom graduate squad riverfront formal CLIP:Vi T-L/14 0.0 0.5 1.0 band quartet formal ensemble prom graduate riverfront bass (*) bulldog dress cathedral (*) squad fish (*) obsolete (*) Open Clip:Vi T-B-32 0.0 0.5 1.0 band quartet ensemble prom squad bass (*) formal dress riverfront obsolete (*) fish (*) cathedral (*) bulldog graduate Open Clip:Vi T-L-14 0.0 0.5 1.0 band bass (*) obsolete (*) bulldog ensemble fish (*) prom quartet squad formal cathedral (*) riverfront graduate dress 0.0 0.5 1.0 band quartet ensemble prom formal bass (*) riverfront cathedral (*) squad fish (*) bulldog graduate obsolete (*) dress 0.0 0.5 1.0 band ensemble quartet bass (*) formal prom squad dress cathedral (*) obsolete (*) graduate bulldog fish (*) riverfront 0.0 0.5 1.0 trumpet band conducted instrument sailor chop bass (*) naval cathedral (*) fish (*) major cuff obsolete (*) polishing 0.0 0.5 1.0 trumpet band conducted cathedral (*) instrument sailor naval cuff bass (*) obsolete (*) polishing chop major fish (*) CLIP:Vi T-B/32 0.0 0.5 1.0 trumpet band conducted instrument chop sailor fish (*) cathedral (*) bass (*) obsolete (*) major polishing naval cuff CLIP:Vi T-L/14 0.0 0.5 1.0 trumpet band conducted instrument cathedral (*) cuff sailor naval bass (*) obsolete (*) polishing chop fish (*) major Open Clip:Vi T-B-32 0.0 0.5 1.0 trumpet band instrument conducted obsolete (*) naval cuff chop cathedral (*) sailor bass (*) major fish (*) polishing Open Clip:Vi T-L-14 0.0 0.5 1.0 trumpet band instrument obsolete (*) bass (*) fish (*) chop major cuff naval conducted sailor cathedral (*) polishing 0.0 0.5 1.0 trumpet band conducted instrument bass (*) sailor major chop cathedral (*) obsolete (*) naval fish (*) cuff polishing 0.0 0.5 1.0 trumpet instrument band conducted bass (*) naval major sailor obsolete (*) polishing cuff fish (*) cathedral (*) chop Figure E.15: Local importance results with SKIT and CLIP:Vi T-L/14 for 2 images from three classes in the Imagenette dataset (part I of II. 0.0 0.5 1.0 launch flew descending textbook instructor tandem slipping fishing (*) instrument (*) uk england band (*) cathedral (*) departs 0.0 0.5 1.0 flew descending launch departs tandem instructor textbook slipping uk england band (*) instrument (*) fishing (*) cathedral (*) CLIP:Vi T-B/32 0.0 0.5 1.0 flew descending tandem launch instructor departs slipping england cathedral (*) instrument (*) band (*) fishing (*) uk textbook CLIP:Vi T-L/14 0.0 0.5 1.0 launch flew descending instructor departs tandem slipping cathedral (*) fishing (*) band (*) england textbook instrument (*) uk Open Clip:Vi T-B-32 0.0 0.5 1.0 launch departs tandem flew instructor slipping descending england fishing (*) uk band (*) cathedral (*) instrument (*) textbook Open Clip:Vi T-L-14 0.0 0.5 1.0 launch descending flew tandem instructor instrument (*) cathedral (*) slipping uk band (*) fishing (*) textbook departs england 0.0 0.5 1.0 tandem launch descending instructor departs flew slipping instrument (*) uk england fishing (*) cathedral (*) band (*) textbook 0.0 0.5 1.0 tandem slipping departs england flew instrument (*) descending instructor cathedral (*) fishing (*) textbook uk launch band (*) 0.0 0.5 1.0 skiing violin band (*) rose lollipop instrument (*) fishing (*) iv surgery string earbuds cathedral (*) lead science 0.0 0.5 1.0 skiing lollipop string violin fishing (*) rose iv instrument (*) earbuds surgery band (*) science cathedral (*) lead CLIP:Vi T-B/32 0.0 0.5 1.0 iv string lollipop violin skiing band (*) cathedral (*) instrument (*) earbuds rose fishing (*) lead science surgery CLIP:Vi T-L/14 0.0 0.5 1.0 lollipop skiing surgery fishing (*) violin string earbuds cathedral (*) science rose lead instrument (*) band (*) iv Open Clip:Vi T-B-32 0.0 0.5 1.0 lollipop skiing surgery violin cathedral (*) fishing (*) science earbuds instrument (*) rose iv band (*) string lead Open Clip:Vi T-L-14 0.0 0.5 1.0 skiing lollipop violin surgery science earbuds fishing (*) string cathedral (*) lead instrument (*) band (*) iv rose 0.0 0.5 1.0 skiing surgery fishing (*) instrument (*) science iv violin string cathedral (*) earbuds lead lollipop band (*) rose 0.0 0.5 1.0 lollipop skiing rose surgery string science lead fishing (*) violin instrument (*) earbuds cathedral (*) band (*) iv Figure E.16: Local importance results with SKIT and CLIP:Vi T-L/14 for 2 images from three classes in the Imagenette dataset (part II of II). clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 14 13 12 11 10 9 8 7 6 5 4 3 2 1 spaniel help stick fishing launched helping cathedral fore rescue barge fetch trumpet cane swimming clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 14 13 12 11 10 9 8 7 6 5 4 3 2 1 spaniel trumpet sibling fishing youngster whisky fluffy twin royalty patch cathedral shamrock fore marble clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 14 13 12 11 10 9 8 7 6 5 4 3 2 1 band ensemble quartet formal bass prom squad dress cathedral obsolete bulldog fish graduate riverfront clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 14 13 12 11 10 9 8 7 6 5 4 3 2 1 trumpet instrument band bass conducted naval cathedral obsolete major sailor fish polishing cuff chop clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 14 13 12 11 10 9 8 7 6 5 4 3 2 1 tandem slipping departs flew fishing textbook descending instrument england cathedral uk instructor band launch clip:Vi T-B/32 clip:Vi T-L/14 open_clip:Vi T-B-32 open_clip:Vi T-L-14 14 13 12 11 10 9 8 7 6 5 4 3 2 1 skiing lollipop string fishing instrument surgery cathedral science band violin rose lead earbuds iv Figure E.17: Summary of X-SKIT ranks of importance across all models for 2 examples images from 3 classes in the Imagenette dataset. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: In the abstract, we state that we introduce precise definitions of semantic importance for the predictions of black-box models, and novel procedures to test for each of them. These are the contributions of the paper. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: Throughout the paper, we clearly state what guarantees our proposed methods provide, and what they do not. We comment on possible future directions. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: We provide complete proofs and background information for each claim. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: We provide pseudocode for all algorithms used in the paper, detailed descriptions of the data generating processes for synthetic datasets, links to the publicly available datasets that were used, description of hyperparameters used for training and testing models, and link to a Git Hub repository to reproduce all experiments presented in the paper. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: We provide all code necessary to reproduce the experiments presented in this paper. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: We dedicate an appendix to the full experimental details. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: We report p-values and uncertainty estimates wherever they are appropriate. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: We specify all computational resources used in this work. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: The work presented in this paper did not involve human subjects, and all data used was from current publicly available datasets. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [Yes] Justification: The objective of this paper is to develop statistically-valid methods to study the semantic importance of human-interpretable concepts for the predictions of black-box models. We envision this research will support better practices towards a responsible use of artificial intelligence in society. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: The presented paper does not pose such risks. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: We stated all original references of papers and code repositories used for this paper, inclusive of their respective licenses. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: The submitted paper does not release new assets. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: The submitted paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: The submitted paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.