# leaveoneout_distinguishability_in_machine_learning__82dbc6bd.pdf Published as a conference paper at ICLR 2024 LEAVE-ONE-OUT DISTINGUISHABILITY IN MACHINE LEARNING Jiayuan Ye , Anastasia Borovykh , Soufiane Hayou , and Reza Shokri National University of Singapore, Imperial College London We introduce an analytical framework to quantify the changes in a machine learning algorithm s output distribution following the inclusion of a few data points in its training set, a notion we define as leave-one-out distinguishability (LOOD). This is key to measuring data memorization and information leakage as well as the influence of training data points in machine learning. We illustrate how our method broadens and refines existing empirical measures of memorization and privacy risks associated with training data. We use Gaussian processes to model the randomness of machine learning algorithms, and validate LOOD with extensive empirical analysis of leakage using membership inference attacks. Our analytical framework enables us to investigate the causes of leakage and where the leakage is high. For example, we analyze the influence of activation functions, on data memorization. Additionally, our method allows us to identify queries that disclose the most information about the training data in the leave-one-out setting. We illustrate how optimal queries can be used for accurate reconstruction of training data.1 1 INTRODUCTION A key question in interpreting a model involves identifying which members of the training set have influenced the model s predictions for a particular query (Koh & Liang, 2017; Koh et al., 2019; Pruthi et al., 2020). Our goal is to understand the impact of the presence of a specific data point in the training set of a model on its predictions. This becomes important, for example, when model s prediction on a data point is primarily due to its own presence (or that of points similar to it) in the training set - a phenomenon referred to as memorization (Feldman, 2020) and can cause leakage of sensitive information about the training data. This enables an adversary to infer which data points were included in the training set (using membership inference attacks), by observing the model s predictions (Shokri et al., 2017; Zarifzadeh et al., 2023). One way to measure such influence, memorization, and information leakage, and thus addressing related questions, is by (re)training models with various combinations of data points. This empirical approach, however, is computationally expensive (Feldman & Zhang, 2020), and does not enable efficient analysis of, for example, what data points are influential, which queries they influence the most, and what properties of model (architectures) can impact their influence. These limitations demand an efficient and constructive modeling approach to influence and leakage analysis. This is the focus of the current paper. We generalize the above-mentioned closely-related concepts (Section 2), and unify them as the statistical divergence between the output distribution of models, trained using (stochastic) machine learning algorithms, when one or more data points are added to (or removed from) the training set. We refer to this measure as leave-one-out distinguishability (LOOD). A larger LOOD of a training data S in relation to query data Q suggests a larger potential for information leakage about S, when an adversary observes the prediction of models trained on S and queried at Q. As a special variant of LOOD, if we measure the gap between the mean of prediction distributions over leave-one-out models on a query, the resultant mean distance LOOD recovers existing definitions for influence (function) (Steinhardt et al., 2017; Koh & Liang, 2017; Koh et al., 2019) and memorization (selfinfluence) (Feldman & Zhang, 2020; Feldman, 2020) (i.e., a greater mean distance LOOD suggests a larger influence of training data S on query data Q). Authors AB, SH and RS are ordered alphabetically. 1The code is available through this link. Published as a conference paper at ICLR 2024 We propose an analytical method that estimates LOOD accurately and efficiently (Section 3), by using Gaussian Processes (GPs) to model the output distribution of neural networks in response to querying on any data point(s).2 We empirically validate that LOOD under NNGP models correlates well with the performance of membership inference attacks (Ye et al., 2022; Carlini et al., 2023) against neural networks in the leave-one-out setting for benchmark datasets (Figure 1b). We also experimentally show that mean distance LOOD under NNGP agrees with the prediction differences under leave-one-out retraining (Figure 1c)of deep neural networks. This demonstrates that our approach allows an accurate assessment of information leakage and memorization (self-influence) for deep neural networks, while significantly reducing the required computation time (by more than two orders of magnitude - Footnote 6) as we compute LOOD analytically. Our method offers theoretical explanations as well as efficient algorithms (Section 4) to examine which query points are most influenced by any training data S (i.e., to quantify the leakage of models in black-box settings). This further allows one to gauge the severity of this leakage by examining to what extent the most-influenced query can help reconstruct the training data. We prove that, for GPs under weak regularity conditions (Theorem 4.1), the differing data itself is a stationary point of LOOD. Experiments on image datasets further show that, LOOD under NNGP model is maximized when the query equals the differing datapoint (4b). These conclusions (for NNGP models) experimentally generalize to deep neural networks trained via SGD, where the empirical leakage is maximized when query data is within a small perturbation of the differing data (Figure 4c). Additionally, LOOD under NNGP remains almost unchanged even as the number of queries increases, even when the queries are optimized to have high LOOD (Appendix E.1). In other words, for NNGP models, all the information related to the differing point in the leave-one-out setting is contained within the model s prediction distribution solely on that differing point. We show how to exploit this to reconstruct training data in the leave-one-out or leave-one group-out settings. We present samples of reconstructed images in Figure 10 and 15-19, which strikingly resemble the training data, even when our optimization lands on sub-optimal queries. Finally, the analytical structure of LOOD guides us in theoretically understanding how the model architecture can affect information leakage. In Section 6, we investigate how the analytical LOOD for Neural Network Gaussian Processes (NNGPs) is affected by activation functions. We prove that a low-rank kernel matrix implies low LOOD (Proposition 6.1) while existing literature (Proposition G.1) prove that the NNGP kernel for fully connected networks is closer to a low-rank matrix under Re LU activation (thus lead to less leakage) compared to smooth activation functions such as Ge LU. Thus, smooth activation functions induce higher LOOD (information leakage) than nonsmooth activation functions in deep neural networks. We validate that this conclusion aligns with empirical experiments for both NNGPs and for deep neural network trained via SGD (Figure 5). 2 LOOD DEFINITION AND ITS CONNECTIONS TO RELATED CONCEPTS Definition 2.1 (LOOD). Let D and D be two training datasets that only differ in the set S.3 Let f D(Q) be the distribution of model predictions given training dataset D and query dataset Q, where the probability is taken over the randomness of the training algorithm. We define leave-one-out distinguishability (LOOD) as the statistical distance between f D(Q) and f D (Q): LOOD(Q; D, D ) := dist(f D(Q), f D (Q)) (1) In the paper, we mainly measure LOOD using the KL divergence Kullback & Leibler (1951) to capture information leakage. However, as we explain below, different choices of dist() and f() allow us to recover established definitions of memorization, information leakage, and influence. Information Leakage as LOOD. Information Leakage refers to how much a model s predictions can reveal information about specific data records in its training set, which would not be extractable if those entries were removed from the training set. In the black-box setting, where adversary observes model s prediction (distribution) at a query set Q, this is exactly LOOD. This is closely related 2NNGPs are widely used in the literature as standalone models (Neal, 2012), as modeling tools for different types of neural networks (Lee et al., 2017; Novak et al., 2018; Hron et al., 2020; Yang, 2019), and as building blocks for model architecture search (Park et al., 2020) or data distillation (Loo et al., 2022). 3S = D \ D and D D. Published as a conference paper at ICLR 2024 to Differential privacy (DP) (Dwork et al., 2006) LOOD measures privacy loss in a black-box setting, for specific datasets, and DP determines the maximum (white-box) privacy loss across all possible datasets. 4 In this paper, we measure LOOD using KL divergence, which is different from Hocky-stick divergence used for equivalent definitions of (ε, δ)-DP (Balle et al., 2018). The closest DP definition to LOOD is Rényi DP (Mironov, 2017) (consider Rényi divergence with order α 1). LOOD gives useful information about leakage in DP ML algorithms. DP guarantees that the prediction distributions cannot change beyond a certain bound under the change of any training data. However, such a guarantee reflects the worst-case scenario and the same bound holds for any possible training data and query. DP bound does not help measure the privacy risk of specific data points in certain training sets and queries. LOOD, however, helps compute this risk and explore how information leakage varies with different predictions, which prediction leads to the most leakage, and how leakage evolves with more predictions (See Figure 4, Section 4, and Appendix E.1). LOOD controls the leakage quantified by inference attacks. One established method for empirically assessing information leakage is through the performance of a membership inference attack (MIA) (Shokri et al., 2017), where an attacker attempts to guess whether a specific data point S was part of the model s training dataset. The power of any MIA (true positive rate given a fixed false positive rate) is controlled by the likelihood ratio test of observing the prediction given D versus D as the training set (Neyman & Pearson, 1933). Thus, the leakage as quantified by MIA is upper bounded by statistical distance relates to likelihood ratio, e.g., KL divergence (Dong et al., 2019). 5 See Figure 1a for the high correlation between LOOD and MIA performance, and see Appendix C for why KL divergence is better suited than mean distance for measuring information leakage. Influence (Function) as (Derivatives of) Mean Distance LOOD. The problem of influence estimation arises in the context of explaining machine learning predictions (Koh & Liang, 2017) and analyzing the robustness of a training algorithm against adversarial data (Steinhardt et al., 2017), where the influence of an (adversarial) training data z D on the trained model s loss at test data ztest is defined as Iloss(z, ztest) = L(ˆθ z; ztest) L(ˆθ; ztest), where ˆθ = arg min L(θ; D) on dataset D and ˆθ z = arg minθ L(θ; D \{z}). This quantity is thus the change of a trained model s loss on the test point ztest due to the removal of one point z from the training dataset (when z = ztest, it reflects self-influence of z). This is exactly LOOD given by mean distance (Definition C.1) between loss distributions, for an (idealized) training algorithm that outputs the optimal model given the training dataset. Estimating influence incurs computational difficulties, as it requires leave-one-out retraining and exact minimizers. Koh & Liang (2017) approximately compute this influence without retraining for small perturbation of training data, via a first-order Taylor approximation. They compute Ipert,loss(z, ztest) = δL(ˆθzδ, z; ztest) where ˆθzδ, z = arg minθ L(θ; D) L(θ; z)+L(θ; zδ) via non-exact estimations of the Hessian (as exact Hessian computation is computationally expensive for large model). For simple models such as linear model and shallow CNN, the influence function approximation correlate well with leave-one-out retraining. However, for deep neural networks, the influence function approximation worsens significantly ((Koh & Liang, 2017, Figure 2) and Figure 1c). By contrast, we propose an analytical method for computing mean distance LOOD that shows high correlation to leave-one-out retraining even for deep neural networks (Figure 1c). Memorization as LOOD. Memorization, as defined in (Feldman, 2020, page 5), is given by mem(A, D, S) = Prh A(D)(h(x S) = y S) Prh A(D\S)(h(x S) = y S); this measures the change of prediction probability on training data S after removing it from the training dataset D, given training algorithm A. To avoid training many models on D and D \ S for each pair of training data and dataset, one approach is to resort to boosting-like techniques (Feldman & Zhang, 2020; Zhang et al., 2021) to reduce the number of retrained models (at a drop of estimation quality), which still requires training hundreds or thousands of models. The memorization measure is exactly mean distance LOOD Definition C.1 when query equals the differing data (i.e., self-influence) and output function f is the prediction accuracy. Thus, we can analytically compute and reason about memorization in ML using LOOD. 4DP guarantee certainly applies to black-box setting too, but its primary analysis is done for the stronger white-box threat model, where adversary observes the model parameters (or parameter updates during training). 5KL divergence (and thus LOOD) can also be used for semantically measuring the advantage of reconstruction attacks and attribute inference attacks. See (Guo et al., 2022, Section 5.2) for a concrete example. Published as a conference paper at ICLR 2024 (a) High correlation between the LOOD and MIA performance. The leave-one-out dataset D is a classbalanced subset of CIFAR-10 with size 1000, and D = D S for a differing record S. We evaluate over 200 randomly chosen S (from CIFAR-10 or uniform distribution over the pixel domain). This indicate that LOOD faithfully captures information leakage under MIA. See Appendix C.3 for more discussions on this correlation curve. 0.4 0.5 0.6 0.7 0.8 0.9 1 0.4 AUC score of MIA (under NNGP) AUC score of MIA (under NN models) (b) MIA performance on NNGP and NN models matches, for 90 random training data records, each represented by a dot. For MIA on NNs, we evaluate over 50 models on D and 50 models on D . For MIA on NNGP, we compute the prediction mean and variance of NNGP (on the differing data) given D and D , and evaluate MIA over 5000 samples from the two Gaussian distributions. Models are trained on car and plane images from CIFAR10. (c) Mean distance LOOD matches the prediction difference after leave-oneout retraining of NN models, for 90 randomly chosen differing data (each represented by a dot). Dataset and model setups are the same as Figure 1b. As comparison, we plot the influence function estimates (Koh and Liang [10]) and Trac In scores (Pruthi et al.[18]). We normalize all estimates to [0, 1]. We report Pearson s correlation and p-value in the legends. Figure 1: Validation of our analytical framework (based on LOOD and mean distance LOOD for NNGP), according to information leakage measure (the performance of membership inference attacks) and influence definition (prediction difference under leave-one-out retraining). 3 OUR ANALYTICAL FRAMEWORK FOR ESTIMATING LOOD Prior approaches for estimating leakage and memorization are heavily based on experiments, thus suffering from (a) approximation error (e.g., first-order Taylor expansion and non-exact Hessian approximation); (b) modelling error (e.g., suboptimal MIAs for quantifying leakage); (c) high computation cost (e.g., training many models for accurate estimation of memorization); and (d) experimental instability. To address this, we propose an analytical method that estimates LOOD accurately and efficiently, by modelling the prediction distribution of neural networks as Gaussian process. Gaussian process. To analytically capture the randomness of machine learning on model outputs we use Gaussian Process (GP). A GP f GP(µ, K) is specified by its mean function µ(x) = Ef[f(x)] and covariance kernel function K(x, x ) = Ef[(f(x) µ(x))(f(x ) µ(x ))] for any inputs x and x (Williams & Rasmussen, 2006). The mean function is typically set to zero, under which the kernel function K completely captures the structure of a GP. Some commonly used kernel functions are isotropic kernels and correlation kernels (Appendix D.2). Given training data D with noisy labels y D = f(D) + N(0, σ2I), the posterior prediction distribution of GP on a query data Q follows multivariate Gaussian f D,σ2(Q) N(µD,σ2(Q), ΣD,σ2(Q)) with analytical mean µD,σ2(Q) = KQD(KDD + σ2I) 1y D and covariance matrix ΣD,σ2(Q) = KQQ KQD(KDD + σ2I) 1KDQ (by GP regression). Here for brevity, we denote KXZ = (K(xi, zj))xi D,zj Z as the kernel matrix between X and Z. Neural Network Gaussian process and connections to NN training. Prior works (Lee et al., 2017; Schoenholz et al., 2017; 2016) have shown that the output of neural networks on a query converges to a Gaussian process, at initialization as the width of the NN tends to infinity. The kernel function of the neural network Gaussian process (NNGP) is given by K(x, x ) = Eθ prior [ fθ(x), fθ(x ) ], where θ refers to the neural network weights (sampled from the prior distribution), and fθ refers to the prediction function of the model with weights θ. In this case, instead of the usual gradient-based training, we can analytically compute the Gaussian Process regression posterior distribution of the network prediction given the training data and query inputs. See Appendix B for more discussions on the connections between NNGP and the training of its corresponding neural networks. Published as a conference paper at ICLR 2024 LOOD for NNGP effectively captures leakage in practical NN training. Under Gaussian process (GP) modelling, we can analytically compute and analyze LOOD on the GP models associated with a given neural network (i.e., NNGP). For the underlying NNGP models, we first investigate whether the LOOD accurately reflects the level of information leakage in the idealized leave-oneout membership inference attack (Ye et al., 2022, Section 4.5). Specifically, the leave-one-out MIA distinguish between two prediction distributions and of models trained on leave-one-out datasets that only differs in one record S. In Figure 1a, we indeed observe a strong correlation between LOOD and MIA performance. This concludes that LOOD effectively reflects information leakage in NNGP. We also empirically validate whether the leakage under NNGP models accurately reflects the leakage in neural networks trained using SGD. In Figure 1b, we observe a strong correlation between the MIA performance on NNGP models and the MIA performance on neural network models. This suggests that NNGP captures the leakage in the corresponding neural networks. Mean distance LOOD for NNGP agrees with memorization in practical NN training. We further investigates whether our analytical framework allows accurate estimates for memorization (selfinfluence). In Figure 1c, we show the estimates given by different methods (y-axis), and compare it with the actual prediction difference (x-axis) between 100 NN models trained on each of the leave-one-out datasets, over 50 randomly chosen differing data. That is, the x-axis is exactly the memorization or self-influence as defined by (Feldman, 2020, Page 5) and (Feldman & Zhang, 2020, Equation 1), and serves as a ground-truth baseline that should be matched by good influence estimations. In the y-axis, we show the normalized mean distance LOOD (under NNGP), influence function (Koh & Liang, 2017) and Trac In estimates (Pruthi et al., 2020). We observe that mean distance LOOD has the highest correlation to the actual prediction difference among all methods. Analytical LOOD allows improved computational efficiency. One significant advantage of our method is the efficiency of computing analytic expressions. In Figure 1b1c, we observed over 140 speed up for estimating leakage and influence using our framework versus empirically measuring it over retrained models as it is performed in the literature (Ye et al., 2022; Feldman & Zhang, 2020).6 This allows us to efficiently answer many advanced questions (that are previously computationally expensive or infeasible to answer via experimental approaches), such as: Where, i.e., at which query Q, does the model exhibit the most information leakage or influence of its training data? How do different model architectures impact influence? We discuss these questions in Section 4-6. 4 OPTIMIZING LOOD TO IDENTIFY QUERY WITH MAXIMUM LEAKAGE LOOD (6) measures the information leakage of a model about its training data S when queried on a query data point Q. For assessing the maximum possible leakage, we need to understand which query Q should the attacker use to extract the highest possible amount of information about S. This question is closely related to the open problem of constructing a worst-case query data for tight privacy auditing via MIAs (Steinke et al., 2024; Jagielski et al., 2020; Nasr et al., 2021; Wen et al., 2022; Nasr et al., 2023). However, evaluating leakage via MIAs can be computationally expensive due to the associated costs of retraining many models on a fixed pair of leave-one-out datasets. To address this, we will show that estimating LOOD under GPs offers an efficient alternative for answering this question. In this section, we consider single query (|Q| = 1) and single differing point (|S| = 1) setting and analyze the most influenced point by maximizing the LOOD: arg max Q LOOD(Q; D, D ) := dist(f D,σ2(Q), f D ,σ2(Q)), (2) We theoretically analyze and empirical solve the LOOD optimization problem (2). Our analysis also extends to the multi-query (|Q| > 1) and a group of differing data (|S| > 1) setting (Appendix E). Analysis: querying the differing data itself incurs stationary LOOD (information leakage). We first prove that for a wide class of kernel functions, the differing point is a stationary solution for (2), i.e., satisfies the first-order optimality condition. In the following theorem, we prove that if the kernel function satisfies certain regularity conditions, the gradient of LOOD at the differint point is zero, i.e., querying the differing incurs stationary LOOD over its local neighborhood. 6Evaluating the MIA and prediction difference under leave-one-out retraining in Figure 1b took 100 GPU hours for training 50 90 FC networks. For the same setting, estimating LOOD only took 40 GPU minutes. Published as a conference paper at ICLR 2024 Theorem 4.1. Let the kernel function be K. Assume that (i) K(x, x) = 1 for all training data x, i.e., all training data are normalized under kernel function K; (ii) K(x,x ) x |x =x = 0 for all training data x, i.e., the kernel distance between two data records x and x is stable when they are equal. Then, the differing data is a stationary point for LOOD as a function of query data, as follows QLOOD(f D,σ2(Q) f D ,σ2(Q)) |Q=S= 0, where by definition we have that QLOOD(f D,σ2(Q) f D ,σ2(Q)) = 1 Σ (Q) Σ(Q) Q Σ(Q) Σ (Q) + Q µ(Q) µ (Q) 2 2 Σ (Q) 1 + µ(Q) µ (Q) 2 2 Q(Σ (Q) 1) The full proof is deferred in Appendix D.1. Theorem 4.1 proves that as long as the kernel function K satisfies weak regularity conditions (which hold for commonly used RBF kernel and NNGP kernel on a sphere as shown in Appendix D.2), the first-order stationary condition of LOOD is satisfied when the query equals differing data. This stationarity is a necessary condition for maximal leakage, and suggests that the differing point itself could be the query that incurs maximal leakage. (a) Exact reconstruction (b) Color-flipped reconstruction 0.000 0.025 0.050 0.075 0 (c) Prediction similarity in KL (Balle et al., 2022, Section V.B.) between differ data & opt query. Figure 2: Examples of optimized query after LOOD optimization, given all car and plane images from CIFAR10 as training dataset, under NNGP for 10-layer FC network with Re LU activation. We show in (a) the optimized query gets close to the differing point in LOOD; and in (b) the optimized query converges to a query with significantly lower LOOD than the differing data, yet they visually resemble each other. In (c) we further statistically check the prediction similarity between query and differing data in KL divergence (Balle et al., 2022, Section V.B.) across 500 runs with randomly chosen differing data (see Figures 14, 15, 16, 17, 18 and 19 for all the images for these runs). Experimentally Optimizing LOOD enables a Data Reconstruction Attack. To further understand what query Q induce maximal leakage about training data S, we run Adam on the non-convex objective (2). Interestingly, we observe that the optimized query by (2) generally recovers the differing data with high visual precision across random experiments: see Figure 2a, Appendix D.3 and Figure 14-19 for examples under RBF and NNGP kernels. 7 The optimized query also tends to have very similar LOOD to the differing data itself (see Figure 2c), even when there are more than one differing data, i.e., leave-one-group-out setting (see Figure 3 only in its right-most column does the optimized single query have slightly higher LOOD than the differing record that it recovers). We remark that, possibly due to the nonconvex structure of LOOD, cases exist where the optimized query has significantly lower LOOD than the differing data (Figure 2b). Noteworthy is that even in this case the query still resembles the shape of the differing data. This shows the significant amount of information leaked through the model s prediction on the differing data itself. Such results show that LOOD can be leveraged for a data reconstruction attack in a setting where the adversary has access to the prediction distribution of models trained on leave-one-out datasets. This setting aligns with prior works (Balle et al., 2022; Guo et al., 2022) that focus on reconstruction attacks under the assumption that the adversary possesses knowledge of all data points except for one. Our reconstruction via optimizing LOOD methodology extends to the group setting. Figure 3 demonstrates that even when the attacker does not know a group of data points, they can successfully 7We tried several image similarity metrics in the reconstruction literature, e.g., per-pixel distance (Carlini et al., 2023, Section 5.1), SSIM (Haim et al., 2022, Figure 1) and LPIPS (Balle et al., 2022, Section V.B) but they do not reflect the visual similarity in our setting. This is possibly because LOOD optimization only recovers the highest amount of information about the differing data which need not be the exact RGB values. It is an open problem to design better image similarity metrics for such reconstructions beyond RGB values. Published as a conference paper at ICLR 2024 Figure 3: Query optimization (data reconstruction) by leave-one-group-out distinguishability, when there are three differing records (the three leftmost images). Above each differing record, we show whether it is predicted correctly by the NNGP model trained without it. The three middle images are the optimal queries in the multi-query setting, and the rightmost image is the optimal single query; each row presents the results for a different random query initialization. Observe that differing images with highest LOOD (and for certain initialization even all differing images) can be recovered. reconstruct the whole group with high visual similarity. Additionally, by examining the frequency of LOOD optimization converging to a query close to each data point in the differing group (across optimized queries in multiple runs), we can assess which S s are most vulnerable to reconstruction (Appendix E.2). Reconstruction under the leave-one-group-out setting mimics a practical real world adversary that aims to reconstruct more than one training data. In the limit of large group size that equals the size of the training dataset, such leave-one-group-out setting recovers one of the difficult reconstruction attack scenario studied in (Haim et al., 2022), where the adversary does not have exact knowledge about any training data. Therefore, extending our results within the leave-one-group-out setting to practical reconstruction attacks presents a promising open problem for future exploration. Experimental maximality of leakage when query equals differing data for NNGPs and NNs. To complement Theorem 4.1 and explain the success of our reconstruction attack, we experimentally investigate whether querying on differing point itself incurs maximal LOOD. We observe many empirical evidences that support this hypothesis, e.g., for a one-dimensional toy dataset in Figure 4a. In Figure 4b and its caption, we further observe that querying on differing data S itself consistently incurs larger LOOD than querying on random perturbations of S for CIFAR10 under NNGP. Appendix E also support this empirical maximality of LOOD under multiple queries and a group of differing data (Appendix E). Finally, Figure 4c and its caption show that MIA performance on NN models is generally the highest when queries are near the differing data. On the other hand, we note that in practice data-dependent normalization is often used prior to training. In such a setting there exists optimized query with higher LOOD than the differing data itself see Appendix D.4. 5 OPTIMIZING LOOD TO IDENTIFY THE MOST INFLUENCED POINT Mean distance LOOD (Definition C.1) measures the influence of training data S on the prediction at query Q. Thus we can use mean distance LOOD to efficiently analyze where (at which Q) the model is most influenced by the change of one train data S. This question shed light on the fundamental question of which prediction is influenced the most by (adversarial) training data (Koh & Liang, 2017) (and from there connections to robustness against data poisoning Steinhardt et al. (2017)). We now use mean distance LOOD for a more fine-grained analysis of influence and memorization within our framework. In Proposition 5.1, we analyze the different behaviors of mean distance LOOD, based on where the differing point is located with respect to the rest of the training set. Proposition 5.1 (First-order optimality condition for influence). Consider a dataset D, a differing point S, and a kernel function K satisfying the same conditions of Theorem 4.1. Then, we have that QM(Q) |Q=S= α 2(1 KSDM 1 D KDS)(y S KSDM 1 D y D)2 KSDM 1 D KDS, Published as a conference paper at ICLR 2024 5 0 5 Qeury data Q Training dataset D differ S = -2.5 differ S = 0 differ S = 2.5 (a) LOOD under RBF (b) LOOD under NNGP (c) MIA performance on NNs Figure 4: Empirical maximality of information leakage when query equals differing data. Figure (a) uses a toy dataset D = (x, y) with x [ 5, 5], y = sin(x), and computes LOOD for all queries in the data domain. Figure (b), (c) use car and plane images in CIFAR-10, and evaluate leakage on perturbation queries {Q = S + x R : x [ 50, 50]} of the differing data S along a random direction R L unit ball. For five randomly chosen differing data, we show in (b) and (c) that LOOD and MIA performance are visually maximal around x = 0. For statistical pattern, we repeat the experiment for 70 randomly chosen differing data, and observe the scale of perturbation with maximal information leakage is x = 0.000 0.000 for NNGPs, and x = 2.375 5.978 for NNs. where M(Q) is the mean distance LOOD in C.1, KSD, KDS, KDD are kernel matrices as defined in Section 2, KSD = QKQD |Q=S, MD = KDD + σ2I, and α = 1 KSDM 1 D KDS + σ2. Proposition 5.1 proves that when S is far from the dataset D (i.e., KSD 0) or close to the dataset D (i.e., y S KSDM 1 D y D, that is S is perfectly predicted by models trained on D), the mean distance LOOD gradient QM(Q) |Q=S is close to 0 (i.e., the differing data S has stationary influence on itself). When S is far from the remaining training dataset D and the data labels are bounded (which is the case in a classification task), we further analyze the Hessian of mean distance LOOD and prove that differing data S has locally maximal influence on itself (Lemma F.4). Between the two cases, when the differing data is neither too close nor too far from the remaining dataset, Proposition 5.1 proves that QM(Q) |Q=S = 0, i.e., the differing data S does not incur maximal influence on itself. To give an intuition, we can think of the Mean LOOD as the result of an interaction of two forces, a force due to the training points, and another force due to the differing point. When the differing point is neither too far nor too close to the training data, the synergy of the two forces creates a region that does not include the differing point where the LOOD is maximal. See Figures 11, 12 and 13 for more experimental illustrations of the behaviors of mean distance LOOD. 6 EXPLAINING THE EFFECT OF ACTIVATION FUNCTIONS ON LOOD LOOD under NNGP enjoys analytical dependence on model architectures, thus enabling us to efficiently investigate questions such as: how does the choice of activation function affect the magnitute of information leakage? Answers to such questions would provide guidance for how to train models in a more privacy-preserving manner. To theoretically analyze how the choice of activation function affect the magnitude of LOOD, we will leverage Proposition G.1 in the Appendix; it shows that for the same depth, smooth activations such that Ge LU are associated with kernels that are farther away from a low rank all-constant matrix (more expressive) than kernel obtained with non-smooth activations, e.g. Re LU. Therefore to understand the effect of activation, it suffices to analyze how the low-rank property of kernel matrix affects the magnitude of LOOD. Thus, we establish the following proposition, which shows that LOOD is small if the NNGP kernel matrix has low rank property. Proposition 6.1 (Informal: Low rank kernel matrix implies low LOOD). Let D and D = D S be an arbitrarily fixed pair of leave-one-out datasets, where D contains n records. Define the function h(α) = supx,x D |K(x, x ) α|. Let αmin = argminα R+h(α), and thus h(αmin) quantifies how close the kernel matrix is to a low rank all-constant matrix. Assume that αmin > 0. Then, there exist constants An, B > 0 such that max Q LOOD(f D,σ2(Q) f D ,σ2(Q)) Anh(αmin) + B n 1. Thus the smaller h(αmin) is, the smaller LOOD could be. See G for formal statement and proof. Published as a conference paper at ICLR 2024 Proposition 6.1 proves that the LOOD (leakage) over worst-case query is upper bounded by a term h(αmin) that is small if the kernel matrix is close to a low-rank all-constant matrix. Intuitively, this happens because the network output forgets most of the information about the input dataset under low rank kernel matrix. By combining Proposition G.1 and Proposition 6.1, we conclude that smooth activations induce higher leakage than non-smooth activations. This is consistent with the intuition that smooth activations allow deeper information propagation, thus inducing more expressive kernels and more memorization of the training data. Interestingly, recent work (Dosovitskiy et al., 2021) also showed that Ge LU enables better model performance than Re LU even in modern architectures such as Vi T. Our results thus suggest there is a privacy-accuracy tradeoffs by activation choices. Figure 5: Empirical validation of our analytical results in of Section 6 that per-record information leakage is higher under GELU activation than under Re LU, for both NNGPs and NN models. We evaluate perrecord leakage with membership inference attack performance on models trained on leave-one-out datasets. Dataset contains car and plane images from CIFAR-10. The NN model is fully connected network with depth 10 and width 1024. Experimental results on NNGP validates that Ge LU activation induces higher LOOD than Re LU activation. We now empirically investigate how the activation function affects the LOOD under NNGP. We evaluate NNGP for fully connected neural networks with varying depths {2, 4, , 10}, trained on plane and car images from CIFAR10. We compute LOOD over 500 randomly chosen differing data, and observe under fixed depth, the LOOD under Re LU is more than 1.1 higher than LOOD under Ge LU in 95% cases. We visualize randomly chosen 20 differing data in Figure 5. For each fixed differing data, we also observe that the gap between LOOD under Re LU and LOOD under Ge LU grows with the depth. This is consistent with the pattern predicted by our analysis in Proposition G.1, i.e., Ge LU gives rise to higher leakage in LOOD than Re LU activation, and their gap increases with depth. It is worth noting that our experiments for the LOOD optimization are done without constraining the query data norm. Therefore, our experiments shows that Proposition G.1 could be valid in more general settings. Experimental results on NNs validate the effect of activation choice on leakage. We further investigate whether leakage (measured by MIA performance on leave-one-out retrained models) for deep neural networks are similarly affected by activation choice. Figure 5 shows that DNNs with Ge LU activation incurs higher perrecord leakage (MIA performance in AUC score) compared to DNNs with Re LU activation, for 20 randomly chosen differing data. This is consistent with our analytical results (Proposition G.1 and 6.1), and aligns with prior empirical observations Wei et al. (2020); Haim et al. (2022); Shamsabadi et al. (2023) that data reconstruction is harder under Re LU activation (than its smooth counterparts). 7 CONCLUSIONS We show that the estimations of information leakage, influence, and memorization are intrinsically the same problem of estimating leave-one-out distinguishability (LOOD). We propose one analytical framework for accurately and efficiently computing LOOD in machine learning via Gaussian process modeling. This framework facilitates more efficient and interpretable answers to existing questions and enables us to explore new questions about how a data point influences model predictions. Notably, the analytical nature of LOOD enables performing optimization to identify predictions that leak the most information about each training data record (Section 4), which enables exact reconstruction of each training data. Future Works. An interesting direction is to use LOOD to analyze how other factors affect information leakage, such as the training data (distribution) and architecture choices. It is also interesting to extend our framework for estimating other concepts closely related to leave-one-out distinguishability, such as Shapley value (Shapley et al., 1953) in the literature of data valuation. 8 8Shapley value is a linear combination of LOOD on all subsets of the training dataset. Thus, in certain scenarios e.g., submodular utility, LOOD regarding the complete dataset is a lower bound for Shapley value. Published as a conference paper at ICLR 2024 ACKNOWLEDGEMENTS The authors would like to thank Martin Strobel for the help in reproducing influence function, and to thank Yao Tong, Tuan-Duy H. Nguyen and Yaxi Hu for helpful feedback on earlier drafts of the paper. The work of Reza Shokri was supported by the Asian Young Scientist Fellowship 2023, and the Ministry of Education, Singapore, Academic Research Fund (Ac RF) Tier 1. Published as a conference paper at ICLR 2024 Borja Balle, Gilles Barthe, and Marco Gaboardi. Privacy amplification by subsampling: Tight analyses via couplings and divergences. Advances in neural information processing systems, 31, 2018. Borja Balle, Giovanni Cherubin, and Jamie Hayes. Reconstructing training data with informed adversaries. In 2022 IEEE Symposium on Security and Privacy (SP), pp. 1138 1156. IEEE, 2022. Nicholas Carlini, Jamie Hayes, Milad Nasr, Matthew Jagielski, Vikash Sehwag, Florian Tramer, Borja Balle, Daphne Ippolito, and Eric Wallace. Extracting training data from diffusion models. ar Xiv preprint ar Xiv:2301.13188, 2023. Jinshuo Dong, Aaron Roth, and Weijie J Su. Gaussian differential privacy. ar Xiv preprint ar Xiv:1905.02383, 2019. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations, 2021. Cynthia Dwork, Frank Mc Sherry, Kobbi Nissim, and Adam Smith. Calibrating noise to sensitivity in private data analysis. In Theory of Cryptography: Third Theory of Cryptography Conference, TCC 2006, New York, NY, USA, March 4-7, 2006. Proceedings 3, pp. 265 284. Springer, 2006. Cynthia Dwork, Aaron Roth, et al. The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci., 9(3-4):211 407, 2014. Vitaly Feldman. Does learning require memorization? a short tale about a long tail. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, pp. 954 959, 2020. Vitaly Feldman and Chiyuan Zhang. What neural networks memorize and why: Discovering the long tail via influence estimation. Advances in Neural Information Processing Systems, 33:2881 2891, 2020. Chuan Guo, Brian Karrer, Kamalika Chaudhuri, and Laurens van der Maaten. Bounding training data reconstruction in private (deep) learning. In International Conference on Machine Learning, pp. 8056 8071. PMLR, 2022. Niv Haim, Gal Vardi, Gilad Yehudai, Ohad Shamir, and Michal Irani. Reconstructing training data from trained neural networks. ar Xiv preprint ar Xiv:2206.07758, 2022. Soufiane Hayou, Arnaud Doucet, and Judith Rousseau. On the impact of the activation function on deep neural networks training. In International Conference on Machine Learning, 2019. Soufiane Hayou, Eugenio Clerico, Bobby He, George Deligiannidis, Arnaud Doucet, and Judith Rousseau. Stable resnet. In Arindam Banerjee and Kenji Fukumizu (eds.), Proceedings of The 24th International Conference on Artificial Intelligence and Statistics, volume 130 of Proceedings of Machine Learning Research, pp. 1324 1332. PMLR, 13 15 Apr 2021. Jiri Hron, Yasaman Bahri, Jascha Sohl-Dickstein, and Roman Novak. Infinite attention: Nngp and ntk for deep attention networks. In International Conference on Machine Learning, pp. 4376 4386. PMLR, 2020. Matthew Jagielski, Jonathan Ullman, and Alina Oprea. Auditing differentially private machine learning: How private is private sgd? Advances in Neural Information Processing Systems, 33: 22205 22216, 2020. Pang Wei Koh and Percy Liang. Understanding black-box predictions via influence functions. In International conference on machine learning, pp. 1885 1894. PMLR, 2017. Pang Wei W Koh, Kai-Siang Ang, Hubert Teo, and Percy S Liang. On the accuracy of influence functions for measuring group effects. Advances in neural information processing systems, 32, 2019. Published as a conference paper at ICLR 2024 Solomon Kullback and Richard A Leibler. On information and sufficiency. The annals of mathematical statistics, 22(1):79 86, 1951. Jaehoon Lee, Yasaman Bahri, Roman Novak, Samuel S Schoenholz, Jeffrey Pennington, and Jascha Sohl-Dickstein. Deep neural networks as gaussian processes. ar Xiv preprint ar Xiv:1711.00165, 2017. Noel Loo, Ramin Hasani, Alexander Amini, and Daniela Rus. Efficient dataset distillation using random feature approximation. ar Xiv preprint ar Xiv:2210.12067, 2022. Yizhang Lou, Chris E Mingard, and Soufiane Hayou. Feature learning and signal propagation in deep neural networks. In International Conference on Machine Learning, pp. 14248 14282. PMLR, 2022. Ilya Mironov. Rényi differential privacy. In 2017 IEEE 30th computer security foundations symposium (CSF), pp. 263 275. IEEE, 2017. Milad Nasr, Shuang Songi, Abhradeep Thakurta, Nicolas Papernot, and Nicholas Carlin. Adversary instantiation: Lower bounds for differentially private machine learning. In 2021 IEEE Symposium on Security and Privacy (SP), pp. 866 882. IEEE, 2021. Milad Nasr, Jamie Hayes, Thomas Steinke, Borja Balle, Florian Tramèr, Matthew Jagielski, Nicholas Carlini, and Andreas Terzis. Tight auditing of differentially private machine learning. ar Xiv preprint ar Xiv:2302.07956, 2023. Radford M Neal. Bayesian learning for neural networks, volume 118. Springer Science & Business Media, 2012. Jerzy Neyman and Egon Sharpe Pearson. Ix. on the problem of the most efficient tests of statistical hypotheses. Philosophical Transactions of the Royal Society of London. Series A, Containing Papers of a Mathematical or Physical Character, 231(694-706):289 337, 1933. Roman Novak, Lechao Xiao, Jaehoon Lee, Yasaman Bahri, Greg Yang, Jiri Hron, Daniel A Abolafia, Jeffrey Pennington, and Jascha Sohl-Dickstein. Bayesian deep convolutional networks with many channels are gaussian processes. ar Xiv preprint ar Xiv:1810.05148, 2018. Daniel S. Park, Jaehoon Lee, Daiyi Peng, Yuan Cao, and Jascha Sohl-Dickstein. Towards nngpguided neural architecture search, 2020. Garima Pruthi, Frederick Liu, Satyen Kale, and Mukund Sundararajan. Estimating training data influence by tracing gradient descent. Advances in Neural Information Processing Systems, 33: 19920 19930, 2020. Samuel S Schoenholz, Justin Gilmer, Surya Ganguli, and Jascha Sohl-Dickstein. Deep information propagation. ar Xiv preprint ar Xiv:1611.01232, 2016. S.S. Schoenholz, J. Gilmer, S. Ganguli, and J. Sohl-Dickstein. Deep information propagation. In International Conference on Learning Representations, 2017. Ali Shahin Shamsabadi, Jamie Hayes, Borja Balle, and Adrian Weller. Mnemonist: locating model parameters that memorize training examples. In Uncertainty in Artificial Intelligence, pp. 1879 1888. PMLR, 2023. Lloyd S Shapley et al. A value for n-person games. 1953. Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. Membership inference attacks against machine learning models. In 2017 IEEE symposium on security and privacy (SP), pp. 3 18. IEEE, 2017. Jacob Steinhardt, Pang Wei W Koh, and Percy S Liang. Certified defenses for data poisoning attacks. Advances in neural information processing systems, 30, 2017. Thomas Steinke, Milad Nasr, and Matthew Jagielski. Privacy auditing with one (1) training run. Advances in Neural Information Processing Systems, 36, 2024. Published as a conference paper at ICLR 2024 Wenqi Wei, Ling Liu, Margaret Loper, Ka-Ho Chow, Mehmet Emre Gursoy, Stacey Truex, and Yanzhao Wu. A framework for evaluating gradient leakage attacks in federated learning. ar Xiv preprint ar Xiv:2004.10397, 2020. Yuxin Wen, Arpit Bansal, Hamid Kazemi, Eitan Borgnia, Micah Goldblum, Jonas Geiping, and Tom Goldstein. Canary in a coalmine: Better membership inference with ensembled adversarial queries. ar Xiv preprint ar Xiv:2210.10750, 2022. Christopher KI Williams and Carl Edward Rasmussen. Gaussian processes for machine learning. MIT press Cambridge, MA, 2006. Greg Yang. Wide feedforward or recurrent neural networks of any architecture are gaussian processes. Advances in Neural Information Processing Systems, 32, 2019. Greg Yang and Etai Littwin. Tensor programs ivb: Adaptive optimization in the infinite-width limit. ar Xiv preprint ar Xiv:2308.01814, 2023. Jiayuan Ye, Aadyaa Maddi, Sasi Kumar Murakonda, Vincent Bindschaedler, and Reza Shokri. Enhanced membership inference attacks against machine learning models. In Proceedings of the 2022 ACM SIGSAC Conference on Computer and Communications Security, pp. 3093 3106, 2022. Sajjad Zarifzadeh, Philippe Liu, and Reza Shokri. Low-cost high-power membership inference attacks. ar Xiv e-prints, pp. ar Xiv 2312, 2023. Chiyuan Zhang, Daphne Ippolito, Katherine Lee, Matthew Jagielski, Florian Tramèr, and Nicholas Carlini. Counterfactual memorization in neural language models. ar Xiv preprint ar Xiv:2112.12938, 2021. Published as a conference paper at ICLR 2024 1 Introduction 1 2 LOOD Definition and its Connections to Related Concepts 2 3 Our Analytical Framework for Estimating LOOD 4 4 Optimizing LOOD to Identify Query with maximum leakage 5 5 Optimizing LOOD to Identify the Most Influenced Point 7 6 Explaining the Effect of Activation Functions on LOOD 8 7 Conclusions 9 A Table of Notations 15 B Additional discussion on the connections between NNGP and NN training 15 C Deferred details for Section 2 15 C.1 Mean Distance LOOD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 C.2 KL Divergence LOOD . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 C.3 Comparison between Mean Distance LOOD and KL Divergence LOOD . . . . . . 17 D Deferred proofs and experiments for Section 4 19 D.1 Deferred proof for Theorem 4.1 . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 D.2 Proofs for regularity conditions of commonly used kernel functions . . . . . . . . 21 D.3 More examples for query optimization under NNGP kernel functions . . . . . . . . 23 D.4 When the differing point is not the optimal query . . . . . . . . . . . . . . . . . . 23 E Analyzing LOOD for Multiple Queries and Leave-one-group-out Setting 24 E.1 LOOD under multiple queries . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 E.2 Leave-one-group out distinguishability . . . . . . . . . . . . . . . . . . . . . . . . 25 F Optimal query under mean distance and deferred proofs for Section 5 25 F.1 Differing point is in general not optimal for Mean Distance LOOD . . . . . . . . . 25 F.2 Differing point is optimal in Mean Distance LOOD when it is far from the dataset . 27 F.3 The optimal query for Mean Distance LOOD on image datasets . . . . . . . . . . 29 G Deferred proofs for Section 6 29 Published as a conference paper at ICLR 2024 A TABLE OF NOTATIONS Table 1: Symbol reference Symbol Meaning D Leave-one-out dataset D Leave-one-out dataset combined with the differing data S Differing data record(s) s number of differing data records Q Queried data record (s) q number of queries d dimension of the data K(x, x ) Kernel function K(X, Z) Rp k, for two finite collections of vectors X = (x1, . . . , xp) and Z = (z1, . . . , zk) (K(xi, zj))1 i p 1 j k KXZ Rp k Abbreviation for K(X, Z) MD Abbreviation for KDD + σ2I MD Abbreviation for KD D + σ2I M(Q) Abbreviation for mean distance LOOD under query Q and leave-one-out datasets D and D MIA Abbreviation for membership inference attack B ADDITIONAL DISCUSSION ON THE CONNECTIONS BETWEEN NNGP AND NN TRAINING Note that in general, there is a performance gap between a trained NN and its equivalent NNGP model. However, they are fundamentally connected even when there is a performance gap: the NNGP describes the (geometric) information flow in a randomly initialized NN and therefore captures the early training stages of SGD-trained NNs. In a highly non-convex problem such as NN training, the initialization is crucial and its impact on different quantities (including performance) is well-documented Schoenholz et al. (2016). Moreover, it is also well-known that the bulk of feature learning occurs during the first few epochs of training (see (Lou et al., 2022, Figure 2)), which makes the initial training stages even more crucial. Hence, it should be expected that properties of NNGP transfer (at least partially) to trained NNs. More advanced tools such as Tensor Programs Yang & Littwin (2023) try to capture the covariance kernel during training, but such dynamics are generally intractable in closed-form. C DEFERRED DETAILS FOR SECTION 2 C.1 MEAN DISTANCE LOOD The simplest way of quantifying the LOOD is by using the distance between the means of the two distributions formed by coupled Gaussian processes: Definition C.1 (Mean distance LOOD). By computing the distance between the mean of coupled Gaussian processes on query set Q, we obtain the following mean distance LOOD. M(Q; D, D ) =1 2 µD,σ2(Q) µD ,σ2(Q) 2 2 (3) 2||KQD(KDD + σ2I) 1y D KQD (KD D + σ2) 1y D ||2 2. (4) where Q is a query data record, and µD,σ2(Q) and µD ,σ2(Q) are the mean of the prediction function for GP trained on D and D respectively, as defined in Section 2. When the context is clear, we abbreviate D and D in the notation and use M(Q) under differing data S to denote M(Q; D, D ) where D = D S. Published as a conference paper at ICLR 2024 Figure 6: Correlation between the LOOD and MIA success. The leave-one-out dataset D is a classbalanced subset of CIFAR-10 with size 1000, and D = D S for a randomly chosen differing record S. We evaluate over 200 random choices of the differing data record (sampled from the CIFAR-10 dataset or uniform distribution over pixel domain [255].) We also evaluate under different LOOD metrics (mean distance and KL divergence) and different kernel functions (RBF kernel or NNGP kernel for fully connected network with depth 1). We observe that the correlation between KL and AUC (right figures) is consistently stronger and more consistent than the correlation between mean distance LOOD and auc (left figures). Example C.2 (Mean distance LOOD is not well-correlated with MIA success). We show the correlations between Mean distance LOOD and the success of leave-one-out membership inference attack (MIA) 9 in Figure 6 (left). We observe that low mean distance and high AUC score occur at the same time, i.e., low Mean distance LOOD does not imply low attack success. This occurs especially when the differing point is sampled from a uniform distribution over the pixel domain [0, 255]. Another downside of the mean distance LOOD metric is that solely first-order information is incorporated; this results in it being additive in the queries: M(Q) = Pm i=1 M(Qi). That is, Mean distance LOOD does not allow to incorporate the impact of correlations between multiple queries. Moreover, the mean distance LOOD has a less clear interpretation under certain kernel functions. For example, under the NNGP kernel, a query with large ℓ2 norm will also have high Mean Distance LOOD (despite normalized query being the same), due to the homogeneity of NNGP kernel. By contrast, as we show below, KL divergence (relative entropy) LOOD mitigates these limitations. C.2 KL DIVERGENCE LOOD To mitigate the downsides of Mean distance LOOD and incorporate higher-order information, we use KL divergence (relative entropy) LOOD. Definition C.3 (KL Divergence LOOD). Let Q Rd q be the set of queried data points, where d is the data feature dimension and q is the number of queries. Let D, D be a pair of leave-one-out datasets such that D = D S where S is the differing data point. Let f D,σ2 and f D ,σ2 as coupled Gaussian processes given D and D . Then the Leave-one-out KL distinguishability between f D,σ2 9This is measured by the AUC of the attacker s false positive rate (FPR) versus his true positive rate (TPR). Published as a conference paper at ICLR 2024 and f D ,σ2 on queries Q is as follows, KL(f D,σ2(Q) f D ,σ2(Q)) =1 log |Σ (Q)| |Σ(Q)| q + Tr(Σ (Q) 1Σ(Q)) (5) + (µ(Q) µ (Q)) Σ (Q) 1(µ(Q) µ (Q)) (6) where for brevity, we denote µ(Q) = µD,σ2(Q), µ (Q) = µD ,σ2(Q), Σ(Q) = ΣD,σ2(Q), and Σ (Q) = ΣD ,σ2(Q) as defined in Section 2. Example C.4 (KL divergence is better correlated with MIA success). We show correlations between KL Divergence LOOD and AUC score of the MIA in Figure 6 (right). We observe that KL Divergence LOOD and AUC score are well-correlated with each other. On the assymmetry of KL divergence. In this paper we restrict our discussions to the case when the base distribution in KL divergence is specified by the predictions on the larger dataset D , which is a reasonable upper bound on the KL divergence with base distribution specified by predictions on the smaller dataset. This is due to the following observations: 1. In numerical experiments (e.g., the middle figure in Figure 7) that Σ(Q)/Σ (Q) 1 across varying single query Q. This is aligned with the intuition that the added information in the larger dataset D reduce its prediction uncertainty (variance). 2. If Σ(Q)/Σ (Q) 1, then we have that KL(f D,σ2(Q) f D ,σ2(Q)) KL(f D ,σ2(Q) f D,σ2(Q)) Proof. Denote r = Σ(Q)/Σ (Q) 1, then by Equation (6), for single query Q, we have that KL(f D,σ2(Q) f D ,σ2(Q)) = 1 2 log r 1 + r + 2M(Q)Σ (Q) 1 (7) where M(Q) is the abbreviation for mean distance LOOD as defined in Definition C.1. Similarly, when the base distribution of KL divergence changes, we have that KL(f D ,σ2(Q) f D,σ2(Q)) = 1 log r 1 + 1 r + 2M(Q)Σ(Q) 1 (8) Therefore, we have that KL(f D,σ2(Q) f D ,σ2(Q)) KL(f D ,σ2(Q) f D,σ2(Q)) r 2 log r + 2M(Q)Σ(Q) 1(r 1) For r 1, we have that r 1 r 2 log r 0 and M(Q)Σ(Q) 1(r 1) 1. Therefore, we have that KL(f D,σ2(Q) f D ,σ2(Q)) KL(f D ,σ2(Q) f D,σ2(Q)) 0 as long as r = Σ(Q)/Σ (Q) 1. Therefore, the KL divergence with base prediction distribution specified by the larger dataset is an upper bound for the worst-case KL divergence between prediction distributions on leave-one-out datasets. C.3 COMPARISON BETWEEN MEAN DISTANCE LOOD AND KL DIVERGENCE LOOD KL Divergence LOOD is better correlated with attack performance. As observed from Figure 6, KL divergence LOOD exhibits strong correlation to attack performances in terms of AUC score of the corresponding leave-one-out MIA. By contrast, Mean distance LOOD is not well-correlated with MIA AUC score. For example, there exist points that incur small mean distance LOOD while exhibiting high privacy risk in terms of AUC score under leave-one-out MIA. Moreover, from Figure 6 (right), empirically there exists a general threshold of KL Divergence LOOD that implies small AUC (that persists under different choices of kernel function, and dataset distribution). This is reasonable Published as a conference paper at ICLR 2024 M(Q) with S * S * = argmax S|| QM(Q) Q = S|| Q * = argmax QM(Q) variance (ratio) log( (Q)/ (Q)) KL(Q) with S * Figure 7: An example on a one-dimensional toy sine training dataset generated as Appendix F.1, where each training data is shown as a red dot. We denote S as a crafted differing data record that incurs maximum gradient for the mean distance LOOD objective, which is constructed following the instructions in Algorithm 1. In the left plot, we observe that the optimal query Q for mean distance LOOD does not equal the differing point S . On the contrary, the optimal query for KL LOOD (which incurs maximal information leakage) is the differing point S (right plot). This shows the discrepancy between mean distance LOOD and KL LOOD, and shows the inadequacy of using mean distance LOOD to capture information leakage. since KL divergence is the mean of negative log-likelihood ratio, and is thus closely related to the power of log-likelihood ratio membership inference attack. Why the mean is (not) informative for leakage. To understand when and why there are discrepancies between the mean distance LOOD and KL LOOD, we numerically investigate the optimal query Q that incurs maximal mean distance LOOD. In Figure 7, we observe that although Q indeed incurs the largest mean distance among all possible query data, the ratio log(Σ(Q)/Σ (Q)) between prediction variance of GP trained on D and that of GP trained on D , is also significantly smaller at query Q when compared to the differing point S. Consequently, after incorporating the second-order variance information, the KL divergence at query Q is significantly lower than the KL divergence at the differing point S , despite of the fact that Q incurs maximal mean distance LOOD. A closer look at Figure 7 middle indicates that the variance ratio peaks at the differing point S , and drops rapidly as the query becomes farther away from the differing data and the training dataset. Intuitively, this is because Σ (Q), i.e., the variance at the larger dataset, is extremely small at the differing data point S. As the query gets further away from the larger training dataset, Σ (Q) then increases rapidly. By contrast, the variance on the smaller dataset Σ(Q) is large at the differing data point S, but does not change a lot around its neighborhood, thus contributing to low variance ratio. This shows that there exist settings where the second order information contributes significantly to the information leakage, which is not captured by mean distance LOOD (as it only incorporates first-order information). As another example case of when the variance of prediction distribution is non-negligible for faithfully capturing information leakage, we investigate under kernel functions that are homogeneous with regard to their input data. Specifically, for a homogeneous kernel function K that satisfifes K(λx, x ) = K(x, λx ) = λαK(x, x ) for any real number λ R, we have that ΣD,σ2(λQ) = λ2αΣD,σ2(Q) µD,σ2(λQ) = λαµD,σ2(Q), where σ2 > 0 denotes the variance of the label noise. Therefore, for any λ R, we have that M(λQ; D, D ) = λαM(Q; D, D ) (9) KL(f D,σ2(λQ) f D ,σ2(λQ)) = KL(f D,σ2(Q) f D ,σ2(Q)) (10) Consequently, the query that maximizes Mean distance LOOD explodes to infinity for data with unbounded domain. On the contrary, the KL Divergence LOOD is scale-invariant and remains stable as the query scales linearly (because it takes the variance into account, which also grows with the query scale). In this case, the mean distance is not indicative of the actual information leakage as the variance is not a constant and is non-negligible. In summary, KL LOOD is a metric that is better-suited for the purpose of measuring information leakage, compared to mean distance LOOD. This is because KL LOOD captures the correlation between multiple queries, and incorporates the second-order information of prediction distribution. Published as a conference paper at ICLR 2024 D DEFERRED PROOFS AND EXPERIMENTS FOR SECTION 4 D.1 DEFERRED PROOF FOR THEOREM 4.1 In this section, we prove that the differing data record is the stationary point for optimizing the LOOD objective. We will need the following lemma. Lemma D.1. For A Rn n, b Rn and c R, we have that A b b c 1 = A 1 + α 1A 1bb A 1 α 1A 1b α 1b A 1 α 1 where α = c b A 1b. We now proceed to prove Theorem 4.1 Theorem 4.1. Let the kernel function be K. Assume that (i) K(x, x) = 1 for all training data x, i.e., all training data are normalized under kernel function K; (ii) K(x,x ) x |x =x = 0 for all training data x, i.e., the kernel distance between two data records x and x is stable when they are equal. Then, the differing data is a stationary point for LOOD as a function of query data, as follows QLOOD(f D,σ2(Q) f D ,σ2(Q)) |Q=S= 0. Proof. In this statement, LOOD refers to KL LOOD Definition C.3, therefore by analytical differentiation of Definition C.3, we have that QLOOD(f D,σ2(Q) f D ,σ2(Q)) = QKL(f D,σ2(Q) f D ,σ2(Q)) (12) + µ(Q) µ (Q) 2 2 Q Σ (Q) 1 + µ(Q) µ (Q) 2 2 Σ (Q) 1 Note that here Σ(Q) and Σ (Q) are both scalars, as we consider single query. We denote the three terms by F1(Q), F2(Q), and F3(Q). We will show that F1(S) = 0 and F2(S) = F3(S) which is sufficient to conclude. Let us start with the first term. 1. F1(S) = 0: Recall from the Gaussian process definition in Section 2 that Σ(Q) = KQQ KQDM 1 D KDQ. where MD = KDD + σ2I. Therefore, QΣ(Q) = 2 KQD Q M 1 D KDQ, Q Rd n (d is the dimension of Q and n is the number of datapoints in the training set D). A similar formula holds for QΣ (Q). From Lemma D.1, and by using the kernel assumption that K(S, S) = 1, we have that M 1 D = KDD + σ2I KDS KSD 1 + σ2 = M 1 D + α 1M 1 D KDSKSDM 1 D α 1M 1 D KDS α 1KSDM 1 D α 1 where α = 1 + σ2 KSDM 1 D KDS = σ2 + Σ(S). Published as a conference paper at ICLR 2024 Using the fact that MD en+1 = KD S + σ2en+1 where en+1 = (0, . . . , 0, 1) Rn+1, simple calculations yield Σ (S) = 1 KSD M 1 D KD S = σ2(1 σ2e n+1M 1 D en+1) = σ2(1 σ2α 1). We now have all the ingredients for the first term. To alleviate the notation, we denote by KSD = QKQD |Q=S. |Q=S= 2 KSDM 1 D KDS Σ (S) + 2Σ(S) KSD M 1 D KD S Σ (S)2 . A key observation here is the fact that the last entry of KSD is 0 by assumption. Using the formula above for M 1 D , we obtain KSD M 1 D KD S = σ2 KSD M 1 D en+1 = σ2 KSD( α 1M 1 D KDS) = α 1σ2 KSDM 1 D KDS). Now observe that α 1σ2Σ(S)Σ (S) 1 = α 1σ2(α σ2)σ 2(1 σ2α 1) 1 = 1, which concludes the proof for the first term. 2. F2(S) + F3(S) = 0: Let us start by simplifying F2(S) = µ(Q) µ (Q) 2 2 Q |Q=S Σ (S) 1. The derivative here is that of M(Q), and we have µ(Q) µ (Q) 2 2 Q |Q=S= 2(µ(S) µ (S))( KSDM 1 D y D KSD M 1 D y D ). Using the formula of M 1 D and observe that the last entry of KSD is zero, we obtain µ(Q) µ (Q) 2 2 Q |Q=S= 2α 1(µ(S) µ (S)) y S KSDM 1 D y D KSDM 1 D KDS. Moreover, we have that µ(S) µ (S) =KSDM 1 D y D KSD M 1 D y D (16) =KSDM 1 D y D e n+1(MD σ2I)M 1 D y D (17) =KSDM 1 D y D y S + σ2e n+1M 1 D y D (18) =KSDM 1 D y D y S + σ2( α 1KSDM 1 D y D + α 1y S) (19) =(σ2α 1 1)(y S KSDM 1 D y D). (20) F2(S) = 2α 1(σ2α 1 1)Σ (S) 1 y S KSDM 1 D y D 2 KSDM 1 D KDS (21) Let us now deal with F3(S) = (µ(S) µ (S))2 Σ (Q) 1 Q |Q=S. We have that Published as a conference paper at ICLR 2024 Q |Q=S = 2Σ (S) 2 KSD M 1 D KD S = 2σ2Σ (S) 2 KSD M 1 D en+1 = 2σ2Σ (S) 2 KSD( α 1M 1 D KDS) = 2α 1σ2Σ (S) 2 KSDM 1 D KDS. (22) By plugging Equation (22) and 20 into F3(S) = (µ(S) µ (S))2 Σ (Q) 1 Q |Q=S, we prove that F3(S) = 2α 1σ2(σ2α 1 1)2Σ (S) 2 y S KSDM 1 D y D 2 KSD M 1 D KDS. We conclude by observing that σ2(σ2α 1 1)Σ (S) 1 = σ2(σ2α 1 1)σ 2(1 σ2α 1) 1 = 1. D.2 PROOFS FOR REGULARITY CONDITIONS OF COMMONLY USED KERNEL FUNCTIONS We first show that both the isotropic kernel and the correlation kernel satisfies the condition of Theorem 4.1. We then show that the RBF kernel, resp. the NNGP kernel on a sphere, is an isotropic kernel, resp. a correlation kernel. Proposition D.2 (Isotropic kernels satisfy conditions in Theorem 4.1). Assume that the kernel function K is isotropic, i.e. there exists a continuously differentiable function g such that K(x, y) = g( x y ) for all x, y. Assume that g satisfies z 1g (z) is bounded10 and g(0) = 1. Then, K satisfies the conditions of Theorem 4.1. Proof. It is straightforward that K(x, x) = 1 for all x. Simple calculations yield x = g ( x y ) By assumption, the term g ( x y ) x y is bounded for all x = y and also in the limit x y. Therefore, it is easy to see that K(x, y) More rigorously, the partial derivative above exists by continuation of the function z g (z) z z near 0+. Proposition D.3 (Correlation kernels satisfy conditions in Theorem 4.1). Assume that the kernel function K is a correlation kernel, i.e. there exists a function g such that K(x, y) = g x,y x y for all x, y = 0. Assume that g(1) = 1. Then, K satisfies the conditions of Theorem 4.1. Proof. The first condition is satisfied by assumption on g. For the second condition, observe that x 2 x g x, y which yields K(x,y) We have the following result for the NNGP kernel. 10Actually, only the boundedness near zero is needed. Published as a conference paper at ICLR 2024 Proposition D.4 (NNGP kernel on the sphere satisfies conditions in Theorem Theorem 4.1). The NNGP kernel function is similar to the correlation kernel. For Re LU activation function, it has the form K(x, y) = x y g x, y This kernel function does not satisfy the conditions of Theorem 4.1 and therefore we cannot conclude that the differing point is a stationary point. Besides, in practice, NNGP is used with normalized data in order to avoid any numerical issues; the datapoints are normalized to have x = r for some fixed r > 0. With this formulation, we can actually show that the kernel K restricted to the sphere satisfies the conditions of Theorem 4.1. However, as one might guess, it is not straightforward to compute a "derivative on the sphere" as this requires a generalized definition of the derivative on manifolds (the coordinates are not free). To avoid dealing with unnecessary complications, we can bypass this problem by considering the spherical coordinates instead. For x Rd, there exists a unique set of parameters (r, φ1, φ2, . . . , φd 1) Rd s.t. x1 = r cos(φ1) x2 = r sin(φ1) cos(φ2) x3 = r sin(φ1) sin(φ2) cos(φ3) ... xd 1 = r sin(φ1) sin(φ2) . . . sin(φd 2) cos(φd 1) xd = r sin(φ1) sin(φ2) . . . sin(φd 2) sin(φd 1). This is a generalization of the polar coordinates in 2D. Without loss of generality, let us assume that r = 1 and let φ = (φ1, φ2, . . . , φd 1). In this case, the NNGP kernel, evaluated at (φ, φ ) is given by K(φ, φ ) = g j=1 sin(φj) sin(φ j) cos(φi) cos(φ i) + j=1 sin(φj) sin(φ j) with the convention Q0 i=1 = 1. Consider the spherical coordinates introduced above, and assume that r = 1. Then, the NNGP kernel given by K(φ, φ ) satisfies the conditions of Theorem 4.1. Proof. Computing the derivative with respect to φ is equivalent to computing the derivative on the unit sphere. For l {1, . . . , d 1}, we have φl = J(φ, φ ) g j=1 sin(φj) sin(φ j) cos(φi) cos(φ i) + j=1 sin(φj) sin(φ j) j=1 sin(φ j) cos(φi) cos(φ i) j=1 sin(φ j) j=1 sin(φj) j=1 sin(φ j) sin(φl) cos(φ l). When φ = φ , we obtain Published as a conference paper at ICLR 2024 (a) A run of LOOD optimization under NNGP kernel for fully connected network with depth 5 (b) A run of LOOD optimization under NNGP kernel for CNN with depth 5 Figure 8: Additional LOOD optimization results under NNGP kernels with different architectures. We evaluate on class-balanced subset of CIFAR-10 dataset of size 1000. The optimized query consistently has similar LOOD and MIA AUC score to the differing data record. J(φ, φ) = cos(φl) j=1 sin(φj)2 cos(φi)2 + cos(φl) j=1 sin(φj)2 j=1 sin(φj)2 By observing that j=1 sin(φj)2 j=1 sin(φj)2, we conclude that K(φ,φ ) φl |φ=φ = 0 for all l. D.3 MORE EXAMPLES FOR QUERY OPTIMIZATION UNDER NNGP KERNEL FUNCTIONS We show the optimized query for NNGP kernel under different model architectures, including fully connected network and convolutional neural network with various depth, in Figure 8. To further understand the query optimization process in details, we visualize an example LOOD optimization process below in Figure 9. D.4 WHEN THE DIFFERING POINT IS NOT THE OPTIMAL QUERY In practice data-dependent normalization is used prior to training, such as batch and layer normalization. Note that this normalization of the pair of leave-one-out datasets results in them differing in more than one record 11 Then the optimized query data record obtains a higher LOOD than the LOOD of the differing points; we show this in Figure 10; thus in such a setting there is value for the attacker to optimise the query; this is consistent with results from (Wen et al., 2022) that exploit the training data augmentation. 11During inference, the same test record will again go through a normalization step that depends on the training dataset before being passed as a query the corresponding Gaussian process. Published as a conference paper at ICLR 2024 (a) Query after 0 iterations of optimization (b) Query after 10 iterations of optimization (c) Query after 20 iterations of optimization (d) LOOD of query versus number of iterations Figure 9: Illustration of an example LOOD optimization process at different number of optimization iterations. We evaluate on a class-balanced subset of CIFAR-10 dataset of size 1000 and NNGP kernel for fully connected network with depth 5. The query is initialized by uniform sampling from the pixel domain, and starts recovering the differing data record as the optimization proceeds. We plot the optimized queries after 0, 10 and 20 iterations and show the convergence of LOOD for query with regard to the increasing number of iterations for LOOD optimization. (a) Differing data and optimized query in run one (b) Differing data and optimized query in run two Figure 10: Example runs of LOOD optimization under data-dependent training data normalization (that uses per-channel mean and standard deviation of training dataset). The optimized query has a slightly higher LOOD than differing point, and visually has darker color. Our setup uses 500 training data points of the CIFAR10 dataset (50 per class) and NNGP specified by CNN with depth 5. E ANALYZING LOOD FOR MULTIPLE QUERIES AND LEAVE-ONE-GROUP-OUT SETTING E.1 LOOD UNDER MULTIPLE QUERIES We now turn to a more general question of quantifying the LOOD when having a model answer multiple queries {Q1, ..., Qq} rather than just one query Q (this corresponds to the practical setting of e.g. having API access to a model that one can query multiple times). In this setting, a natural question is: as the number of queries grows, does the distinguishability between leave-one-out models predictions get stronger (when compared to the single query setting)? Our empirical observations indicate that the significant information regarding the differing data point resides within its prediction on that data record. Notably, the optimal single query Q = S exhibits a higher or equal degree of leakage compared to any other query set. Specifically, to show that the optimal single query Q = S exhibits a higher or equal degree of leakage compared to any other query set, we performed multi-query optimization max Q1, ,Qq LOOD(Q1, , Qq; D, D ), which uses gradient descent to optimize the LOOD across q queries, for varying numbers of queries q = 1, 2, 4, 8. We evaluated this setting under the following conditions: NNGP using a fully connected network with depths of 2, 4, 6, 8, 10; using 500 random selections of leave-one-out datasets (where the larger dataset is of size 10000) sampled from the car and plane classes of the CIFAR10 dataset. Our observations indicate that, in all Published as a conference paper at ICLR 2024 evaluation trials, the LOOD values obtained from the optimized m queries are lower than the LOOD value from the single differing data record. E.2 LEAVE-ONE-GROUP OUT DISTINGUISHABILITY Here we extend our analysis to the leave-one group-out setting, which involves two datasets, D and D , that differ in a group of s data points {S1, ..., Ss}. There are no specific restrictions on how these points are selected. In this setup, we evaluate the model using queries {Q1, ..., Qq}, where q 1. The leave-one group-out approach can provide insights into privacy concerns when adding a group to the training set, such as in datasets containing records from members of the same organization or family Dwork et al. (2014). Additionally, this setting can shed light on the success of data poisoning attacks, where an adversary adds a set of points to the training set to manipulate the model or induce wrong predictions. Here, we are interested in constructing and studying queries that result in the highest information leakage about the group, i.e., where the model s predictions are most affected under the leave-one group-out setting. Similar to our previous approach, we run maximization of LOOD to identify the queries that optimally distinguishes the leave-one group-out models. Figure 3 presents the results for single-query (q = 1) setting and multi-query (q = 3) setting, with a differing group size |S| = 3. We have conducted in total 10 rounds of experiments, where in each round we randomly select a fixed group of differing datapoints and then perform 10 independent optimization runs (starting from different initializations). Among all the runs, 80% of them converged, in which we observe that the optimal queries are visually similar to the members of the differing group, but the chance of recovering them varies based on how much each individual data point is exposed through a single query (by measuring LOOD when querying the point). We observe that among the points in set S, the differing data point with the highest single-query LOOD was recovered 61% of the times, whereas the one with the lowest single-query LOOD was recovered only 8% of the times. This can be leveraged for data reconstruction attacks where the adversary has access to part of the train dataset of the model. F OPTIMAL QUERY UNDER MEAN DISTANCE AND DEFERRED PROOFS FOR SECTION 5 We discuss here the computation of the optimal query under mean distance. This is equivalent to the question of where the function is most influenced by a change in the dataset. F.1 DIFFERING POINT IS IN GENERAL NOT OPTIMAL FOR MEAN DISTANCE LOOD Numerically identifying cases where the differing point is not optimal. We define an algorithm that finds non-optimal differing points S such that when optimizing the query Q, the resulting query is different from S. When the differing point S is not the optimal query, the gradient of M(Q) evaluated at S is non-zero. Hence, intuitively, finding points S such that QM(Q) |Q=S is maximal should yield points S for which the optimal query is not the differing point (as there is room for optimising the query due to its positive gradient). We describe this algorithm in Algorithm 1. Algorithm 1 Identifying non-optimal differing points S Given D and σ, we solve the optimization problem S = argmax S QM(Q) |Q=S to identify a non-optimal differing point. (We use gradient descent to solve this problem.) A toy example of where the differing point is not optimal. To assess the efficacy of Algorithm 1, we consider a simple example in the one-dimensional case. The setup is as follows: (1) Data D = {(xi, yi), i = 1, . . . , 10} is generated using the following distribution: xi N(0, 1), yi = sin(xi). (2) Noise parameter is given by σ2 = 0.01. (3) Kernel is given by K(x, x ) = exp( (x x )2/2) (RBF kernel). (4) In Algorithm 1, only Sx is optimized, the label Sy is fixed to Sy = sin(Sx). We show in Figure 11a the gradient of M(Q) evaluated at S for varying S, and highlight the point S (the result of Algorithm 1) that has the largest non-stationarity for mean distance LOOD objective. Additionally, if we look at the query with maximal LOOD under differing data S = S , the optimal query Q that we found is indeed different from the differing point S , as shown in Figure 11b. Published as a conference paper at ICLR 2024 8 6 4 2 0 2 4 6 8 S QM(Q) Q = S S * = argmax S|| QM(Q) Q = S|| = 3.899 Training data (a) The gradient of M(Q) evaluated at S. 8 6 4 2 0 2 4 6 8 Q M(Q) with S * = 3.899 S * = argmax S|| QM(Q) Q = S|| Q * = argmax QM(Q) (b) The mean distance M(Q) under S . Figure 11: We plot the stationary condition indicator function QM(Q) |Q=S of LOOD when query Q equals the differing data S, over different values of differing data S with Sx [ 5, 5]. This allows us to run Algorithm 1 to identify a differing point S = max S QM(Q) |Q=S that has the highest objective (maximal non-stationarity), as shown in Algorithm 1 (where S = 3.899). In Figure 11b, we validate that under the optimized differing data S , the query Q that maximizes LOOD (over leave-one-out datasets that differ in S ) is indeed not the differing point S . Let us now look at what happens when we choose other differing points S = S . Figure 12 shows M = argmax QM(Q) Training data Figure 12: The plot of M(Q) for varying Q and S. We also show the optimal query Q and the differing point S. the optimal query for three different choices of S. When S is far from the training data D (Sx { 5, 5}), the optimal query approximately coincides with the differing point S. When S is close to the training data (Sx = 0), the behaviour of the optimal query is non-trivial, but it should be noted that the value of the MSE is very small in this case. Explaining when the differing point is suboptimal for Mean Distance LOOD. We now give a theoretical explanation regarding why the differing point is suboptimal for mean distance LOOD when it is close (but not so close) to the training dataset (as observed in Figure 12 middle plot). We first give a proposition that shows that the gradient of M(Q) is large when the differing point S is close (but not so close) to the training dataset D. Proposition 5.1. Consider a dataset D, a differing point S, and a kernel function K satisfying the same conditions of Theorem 4.1. Then, we have that QM(Q) |Q=S= α 2(1 KSDM 1 D KDS)(y S KSDM 1 D y D)2 KSDM 1 D KDS, (23) where M(Q) denotes the mean distance LOOD as defined in Definition C.1, KSD R1 |D| is the kernel matrix between data point S and dataset D, KDS = K SD, KSD = QKQD |Q=S, MD = KDD + σ2I, and α = 1 KSDM 1 D KDS + σ2. Proof. The left-hand-side of Equation (23) is QM(Q) = Q µ(Q) µ (Q) 2 2 by definition of the mean distance LOOD M(Q) in Definition C.1. This term occurs in the term F2(Q) in Equation (13), in the proof for Theorem 4.1. Specifically, we have QM(Q) = F2(S) Σ (S). The detailed expression for F2(S) is computed in line Equation (21) during the proof for Theorem 4.1. Therefore, by multiplying F2(S) in Equation (21) with Σ (S), we obtain the expression for QM(Q) in the statement Equation (23). Published as a conference paper at ICLR 2024 As a result, it should be generally expected that QM(Q) |Q=S = 0. We can further see that when S is far from the data in the sense that KSD 0, the gradient is also close to 0. In the opposite case where S is close to D, e.g. satisfying y S KSDM 1 D y D, then the gradient is also close to zero. Between the two cases, there exists a region where the gradient norm is maximal. F.2 DIFFERING POINT IS OPTIMAL IN MEAN DISTANCE LOOD WHEN IT IS FAR FROM THE DATASET Explaining when the differing point is optimal for Mean Distance LOOD. We now prove that the differing point is (nearly) optimal when it is far from remaining dataset. We assume the inputs belong to some compact set C Rd. Definition F.1 (Isotropic kernels). We say that a kernel function K : C C R is isotropic is there exists a function g : R R such that K(x, x ) = g( x x ) for all x, x C. Examples includes RBF kernel, Laplace kernel, etc. For this type of kernel, we have the following result. Lemma F.2. Let K be an isotropic kernel, and assume that its corresponding function g is differentiable on C (or just the interior of C). Then, for all Q C, we have that QM(Q) ζ (d (Q, S) + d (Q, D)) max(d(Q, S), d(Q, D)), D = {D1, . . . , Dn}, D = D {S} d (Q, D) = maxi [n] |g ( Q Di )| and d (Q, S) = g ( Q S ) d(Q, D) = maxi [n] |g( Q Di )| and d(Q, S) = g( Q S ) ζ > 0 is a constant that depends only on D. Proof. Let Q C. For the sake of simplification, we use the following notation: u = (KDD + σ2I) 1y D, and u = (KD D + σ2I) 1y D . Simple calculations yield QM(Q) = (KQD u KQDu) i=1 (u i ui) QKQDi + u n+1 QKQ,S The first term is upperbounded by u [n] u d(Q, D) + |u n+1|d(Q, S), where u [n] is the vector containing the first n coordinates of u , and ui refers to the ith coordinate of the vector u. Now observe that u is bounded for all S, provided the data labels y are bounded. For this, it suffices to see that (KD D + σ2I) 1 2 σ 2 because of the positive semi-definiteness of KD D . Moreover, we have QKQDi = g ( Q Di ) Q Di Q Di , which implies that QKQDi = |g ( Q Di )|. Therefore, using the triangular inequality, we obtain i=1 (u i ui) QKQDi + u n+1 QKQ,S u [n] u 1d (Q, D) + |u n+1||g ( Q S )|. The existence of ζ is trivial. From Lemma F.2, we have the following corollary. Corollary F.3. Assume that supz R g(z) < , lim|z| g (z) = 0 and g (0) = 0. Then, provided that C is large enough, for any D = {D1, . . . , Dn} and ϵ > 0, there exists a point S, such that if we define D = D {S}, then QM(S) ϵ. That is, if we choose S far enough from the data D, then the point S is close to a stationary point. Published as a conference paper at ICLR 2024 Although this shows that gradient is small at S (for S far enough from the data D), it does not say whether this is close to a local minimum or maximum. T determine the nature of the stationary points in this region, an analysis of the second-order geometry is needed. More precisely, we would like to understand the behaviour of the Hessian matrix in a neighbourhood of S. In the next result, we show that the Hessian matrix is strictly negative around S (for S sufficiently far from the data D), which confirms that any stationary point in the neighbourhood of S should be a local maximum. Lemma F.4 (Second-order analysis). Assume that g(0) > 0, limz 0 z 1g (z) = l , where l is a negative constant that depends on the kernel function, and that limz z 1g (z) = 0. Then, provided that C is large enough, for any D = {D1, . . . , Dn} and ϵ > 0, there exists a point S, such that if we define D = D {S}, then there exists η > 0 satisfying: for all Q B(S, η), 2 QM(Q) 0. Lemma F.4. Assume that g(0) > 0, limz 0 z 1g (z) = l , where l < 0 is a negative constant that depends on the kernel function, and that limz z 1g (z) = 0. Then, provided that C is large enough, for any D = {D1, . . . , Dn} and ϵ > 0, there exists a point S, such that if we define D = D {S}, then there exists η > 0 satisfying: for all Q B(S, η), 2 QM(Q) 0 is a negatively definite matrix. Proof. Let Q C. Let v(Q) = KQD u KQT u. The Hessian is given by 2 QM(Q) = v(Q) 2 Qv(Q) + Qv(Q) Qv(Q) where 2 Qv(Q) = Pn i=1(u i ui) 2 QKQDi + u n+1 2 QKQ,S and Qv(Q) = Pn i=1(u i ui) QKQDi + u n+1 QKQ,S. By Lemma F.2, we have that for S far enough from the dataset D, we have that Qv(S) ϵ. Therefore, the Hessian at S simplilfies to the following. 2 Q|Q=SM(Q) v(Q) i=1 (u i ui) 2 QKQDi + u n+1 2 QKQ,S For i [n], simple calculations yield 2 QKQDi = α(Q, Di)(Q Di)(Q Di) Q Di 2 + g ( Q Di ) where α(Q, Di) = g ( Q Di ) g ( Q Di ) Notice that u n+1 y S KSDu. Without loss of generality, assume that y S > 0. For S far enough from the dataset D, we then have u n+1 > 0 and v(S) y S > 0. Let us now see what happens to the Hessian. For such S, and for any ϵ > 0, there exists a neighbourhood B(S, η), η > 0, such that for all Q B(S, η), we have |g ( Q Di )| Q Di < ϵ, and |α(Q, Di)| < ϵ. This ensures that 2 QKQDi F 2ϵ where . F denotes the Frobenius norm of a matrix. On the other hand, we have 2 QKQ,S = limz 0 g (z) z I = l I. It remains to deal with v(Q). Observe that for S far enough from the data D, and Q close enough to S, v(Q) is close to v(S) y S > 0. Thus, taking ϵ small enough, there exists η satisfying the requirements, which concludes the proof. Lemma F.4 shows that whenever the point S is isolated from the dataset D, the Hessian is negative in a neighbourhood of S, suggesting that any local stationary point can only be a local maximum. This is true for isotropic kernels such as the Radial Basis Function kernel (RBF) and the Laplace kernel. As future work, Lemma F.2 and Lemma F.4 have important computational implications for the search for the worst poisonous point S. Finding the worst poisonous point S is computationally expensive for two reasons: we cannot use gradient descent to optimize over S since the target function is implicit in S, and for each point S, a matrix inversion is required to find max Q M(Q). Our analysis suggests that if S is far enough from the data D, we can use M(S) as a lower-bound estimate of max Q M(Q). Published as a conference paper at ICLR 2024 F.3 THE OPTIMAL QUERY FOR MEAN DISTANCE LOOD ON IMAGE DATASETS As we concluded in the analysis in Section 4, only when the differing datapoint is far away (in some notion of distance) from the rest of the dataset is the optimal query (in mean distance LOOD) the differing datapoint. In Figure 11b, we successfully identified settings on a toy dataset where the differing point is not optimal for mean distance LOOD. We now investigate whether this also happens on an image dataset. (a) Differing data and optimized query in run one (b) Differing data and optimized query in run two (c) AUC gap on Q, S in 50 trials Figure 13: Results for running single query optimization for Mean distance LOOD on class-balanced subset of CIFAR-10 dataset of size 1000 and RBF kernel. The optimized query has higher mean distance LOOD than the differing data record, despite having lower MIA AUC score. Across the 50 trials that we evaluated, more than 50% of trials result in an optimized query that has lower AUC score (by more than 0.025) than the differing data. In Figure 13, we observe that the optimized record is indeed similar (but not identical) to the differing record for a majority fraction of cases. Therefore, also in image data the removal of a certain datapoint may influence the learned function most in a location that is not the removed datapoint itself. However, the optimized query for Mean distance LOOD still has much smaller MIA AUC score than the differing data. This again indicates that MSE optimization finds a suboptimal query for information leakage, despite achieving higher influence (mean distance). G DEFERRED PROOFS FOR SECTION 6 Our starting point is the following observation in prior work (Schoenholz et al., 2017; Hayou et al., 2019; 2021), which states that the NNGP kernel is sensitive to the choice of the activation function. Proposition G.1. Assume that the inputs are normalized, x = 1 and let KL NNGP denote the NNGP kernel function of a an MLP of depth L 1. Then, the for any two inputs, we have the following [Non-smooth activations] With Re LU: there exists a constant α > 0 such that |k L NNGP (x, x ) α| = Θ(L 2). [Smooth activations] With Ge LU, ELU, Tanh: there exists a constant α > 0 such that |k L NNGP (x, x ) α| = Θ(L 1). As a result, for the same depth L, one should expect that NNGP kernel matrix with Re LU activation is closer to a low rank constant matrix than with smooth activations such as Ge LU, ELU, Tanh.12 For proving Proposition 6.1, we will need the following lemma. Lemma G.2. Let A and B be two positive definite symmetric matrices. Then, the following holds A 1 B 1 2 max(λmin(A) 1, λmin(B) 1) A B 2, where λmin(A) is the smallest eigenvalue of A, and . 2 is the usual 2-norm for matrices. Proof. Define the function f(t) = (t A + (1 t)B) 1 for t [0, 1]. Standard results from linear algebra yield f (t) = f(t)(A B)f(t) 12The different activations are given by Ge LU(x) = xΦ(x) where Φ is the c.d.f of a standard Gaussian variable, ELU(x) = 1{x 1}x + 1{x<1}(ex 1), and Tanh(x) = (ex e x)(ex + e x) 1. Published as a conference paper at ICLR 2024 Therefore, we have f(1) f(0) 2 = Z 1 0 f (t)dt 2 0 f (t) 2dt 0 f(t) 2 2 A B 2dt 0 (λmin(t A + (1 t)B)) 2 A B 2dt. Using Weyl s inequality on the eigenvalues of the sum of two matrices, we have the following λmin(t A + (1 t)B) λmin(t A) + λmin((1 t)B) = tλmin(A) + (1 t)λmin(B) min(λmin(A), λmin(B)). Combining the two inequalities concludes the proof. We are now ready to prove Proposition 6.1, as follows. Proposition 6.1. Let D and D = D S be an arbitrary pair of leave-one-out datasets, where D contains n records and D contains n+1 records. Define the function h(α) = supx,x D |K(x, x ) α|. Let αmin = argminα R+h(α), and thus h(αmin) quantifies how close the kernel matrix is to a low rank all-constant matrix. Assume that αmin > 0. Then, it holds that max Q M(Q) Cn,αmin,ζ n5/2h(αmin) + Cα,ζ n 1, Cn,αmin,ζ = max (αmin + h(αmin))ζ min(αminσ2, |αminσ2 nh(αmin)|), (1 + n 1)5/2(αmin + h(αmin))ζ min(αminσ2, |αminσ2 (n + 1)h(αmin)|) and Cαmin,ζ = ζ(1 + α 1 min). As a result, if n5/2h(αmin) 1 and n 1, then the maximum mean distance LOOD (i.e., maximum influence) satisfies max Q M(Q) 1. Similarly, there exist constants An, B > 0 such that max Q KL(f D,σ2(Q) f D ,σ2(Q)) Anh(αmin) + B n 1. Proof. Let Q be a query point and α R. We have the following 2 KQDM 1 D y D KQD M 1 D y D , where MD = (KDD + σ2I) and the same holds for MD . Let U = αUn + σ2I and U = αUn+1 + σ2I , where Un is the n n matrix having ones everywhere. With this notation, we have the following bound 2 M(Q) KQD(M 1 D U 1)y D + KQD (M 1 D (U ) 1)y D + KQD (U ) 1y D KQDU 1y D , Let us deal with first term. We have that KQD(M 1 D U 1)y D KQD y D M 1 D U 1 n1/2(α+h(α)) y D 2 M 1 D U 1 . Published as a conference paper at ICLR 2024 Using Lemma G.2, we have the following M 1 D U 1 max(λmin(A) 1, λmin(B) 1) MD U , which implies that KQD(M 1 D U 1)y D n3/2(α + h(α)) y D 2 min(ασ2, |ασ2 nh(α)|) h(α), where we have used the inequality |λmin(A) λmin(B)| A B 2 for any two symmetric positive matrices A, B. Similarly, we have the following KQD (M 1 D (U ) 1)y D (n + 1)3/2(α + h(α)) y D 2 min(ασ2, |ασ2 (n + 1)h(α)|) h(α).s For the last term, we have KQD (U ) 1y D KQDU 1y D (KQD αe n+1)(U ) 1y D + (KQD αe n )U 1y D + αe n+1(U ) 1y D αe n U 1y D . Moreover, we have (KQD αe n )U 1y D σ 2 y D nh(α), and, (KQD αe n+1)(U ) 1y D σ 2 y D For the last term, we have αe n+1(U ) 1y D αe n U 1y D α |αn + σ2||α(n + 1) + σ2| + α |α(n + 1) + σ2||y S| (1 + α 1)ζn 1. Combining all these inequalities yield the desired bound. For the second upperbound, observe that KL(f D,σ2(Q) f D ,σ2(Q)) = log Σ (Q) Σ(Q) + Σ(Q) 2Σ (Q) + (µ(Q) µ(Q ))2 2 by the definition of KL LOOD in Definition C.3. Meanwhile, for sufficiently small h(αmin), we have that Σ(Q) = αmin α2 minn(σ2 + nαmin) 1 + O(h(αmin)) = αminσ2(σ2 + nαmin) 1 + O(h(αmin)), and a similar formula holds for Σ (Q) with (n + 1) instead of n. Therefore, we obtain Σ(Q) Σ (Q) = 1 + αmin σ2 + αminn. Combining this result with the upperbound on the mean distance, the existence of An and B is straightforward. Published as a conference paper at ICLR 2024 differ image LOOD:0.11267 differ image LOOD:0.08864 differ image LOOD:0.08516 differ image LOOD:0.07705 differ image LOOD:0.07506 differ image LOOD:0.07432 differ image LOOD:0.07211 differ image LOOD:0.06895 differ image LOOD:0.06663 differ image LOOD:0.06372 differ image LOOD:0.06230 differ image LOOD:0.05962 opt query LOOD:0.11267 opt query LOOD:0.08864 opt query LOOD:0.08516 opt query LOOD:0.07705 opt query LOOD:0.07506 opt query LOOD:0.07432 opt query LOOD:0.07211 opt query LOOD:0.06895 opt query LOOD:0.06663 opt query LOOD:0.06372 opt query LOOD:0.06230 opt query LOOD:0.05962 differ image LOOD:0.05820 differ image LOOD:0.05808 differ image LOOD:0.05660 differ image LOOD:0.05623 differ image LOOD:0.05472 differ image LOOD:0.05416 differ image LOOD:0.05218 differ image LOOD:0.05067 differ image LOOD:0.04894 differ image LOOD:0.04579 differ image LOOD:0.04475 differ image LOOD:0.04438 opt query LOOD:0.05820 opt query LOOD:0.05808 opt query LOOD:0.05660 opt query LOOD:0.05623 opt query LOOD:0.05472 opt query LOOD:0.05416 opt query LOOD:0.05218 opt query LOOD:0.05067 opt query LOOD:0.04894 opt query LOOD:0.04579 opt query LOOD:0.04475 opt query LOOD:0.04438 differ image LOOD:0.04369 differ image LOOD:0.04327 differ image LOOD:0.04316 differ image LOOD:0.04199 differ image LOOD:0.04198 differ image LOOD:0.04183 differ image LOOD:0.04144 differ image LOOD:0.04109 differ image LOOD:0.04094 differ image LOOD:0.04037 differ image LOOD:0.04006 differ image LOOD:0.03800 opt query LOOD:0.04369 opt query LOOD:0.04327 opt query LOOD:0.04316 opt query LOOD:0.04199 opt query LOOD:0.04198 opt query LOOD:0.04183 opt query LOOD:0.04144 opt query LOOD:0.04109 opt query LOOD:0.04094 opt query LOOD:0.04037 opt query LOOD:0.04006 opt query LOOD:0.03800 differ image LOOD:0.03678 differ image LOOD:0.03734 differ image LOOD:0.03663 differ image LOOD:0.03644 differ image LOOD:0.03642 differ image LOOD:0.03631 differ image LOOD:0.03600 differ image LOOD:0.03504 differ image LOOD:0.03473 differ image LOOD:0.03451 differ image LOOD:0.03403 differ image LOOD:0.03401 opt query LOOD:0.03678 opt query LOOD:0.03677 opt query LOOD:0.03663 opt query LOOD:0.03644 opt query LOOD:0.03642 opt query LOOD:0.03631 opt query LOOD:0.03600 opt query LOOD:0.03504 opt query LOOD:0.03473 opt query LOOD:0.03451 opt query LOOD:0.03403 opt query LOOD:0.03401 differ image LOOD:0.03363 differ image LOOD:0.03357 differ image LOOD:0.03341 differ image LOOD:0.03324 differ image LOOD:0.03323 differ image LOOD:0.03300 differ image LOOD:0.03251 differ image LOOD:0.03212 differ image LOOD:0.03169 differ image LOOD:0.03089 differ image LOOD:0.03089 differ image LOOD:0.03078 opt query LOOD:0.03363 opt query LOOD:0.03357 opt query LOOD:0.03341 opt query LOOD:0.03324 opt query LOOD:0.03323 opt query LOOD:0.03300 opt query LOOD:0.03238 opt query LOOD:0.03212 opt query LOOD:0.03169 opt query LOOD:0.03089 opt query LOOD:0.03089 opt query LOOD:0.03078 differ image LOOD:0.03059 differ image LOOD:0.02947 differ image LOOD:0.02879 differ image LOOD:0.02858 differ image LOOD:0.02829 differ image LOOD:0.02823 differ image LOOD:0.02802 differ image LOOD:0.02781 differ image LOOD:0.02774 differ image LOOD:0.02730 differ image LOOD:0.02723 differ image LOOD:0.02672 opt query LOOD:0.03059 opt query LOOD:0.02947 opt query LOOD:0.02879 opt query LOOD:0.02858 opt query LOOD:0.02829 opt query LOOD:0.02823 opt query LOOD:0.02802 opt query LOOD:0.02781 opt query LOOD:0.02774 opt query LOOD:0.02730 opt query LOOD:0.02723 opt query LOOD:0.02672 differ image LOOD:0.02656 differ image LOOD:0.02639 differ image LOOD:0.02600 differ image LOOD:0.02598 differ image LOOD:0.02576 differ image LOOD:0.02570 differ image LOOD:0.02520 differ image LOOD:0.02514 differ image LOOD:0.02503 differ image LOOD:0.02490 differ image LOOD:0.02481 differ image LOOD:0.02466 opt query LOOD:0.02656 opt query LOOD:0.02639 opt query LOOD:0.02600 opt query LOOD:0.02598 opt query LOOD:0.02576 opt query LOOD:0.02570 opt query LOOD:0.02520 opt query LOOD:0.02514 opt query LOOD:0.02503 opt query LOOD:0.02490 opt query LOOD:0.02481 opt query LOOD:0.02466 differ image LOOD:0.02421 differ image LOOD:0.02408 differ image LOOD:0.02400 differ image LOOD:0.02399 differ image LOOD:0.02392 differ image LOOD:0.02384 differ image LOOD:0.02382 differ image LOOD:0.02376 differ image LOOD:0.02373 differ image LOOD:0.02336 differ image LOOD:0.02331 differ image LOOD:0.02318 opt query LOOD:0.02421 opt query LOOD:0.02408 opt query LOOD:0.02400 opt query LOOD:0.02399 opt query LOOD:0.02392 opt query LOOD:0.02384 opt query LOOD:0.02382 opt query LOOD:0.02376 opt query LOOD:0.02373 opt query LOOD:0.02336 opt query LOOD:0.02331 opt query LOOD:0.02318 Figure 14: Visualization of 500 randomly chosen differing data and query optimized by LOOD (Part 1). We show differing data above the optimized query, and use their LOOD gap to sort the images. Published as a conference paper at ICLR 2024 differ image LOOD:0.02303 differ image LOOD:0.02302 differ image LOOD:0.02289 differ image LOOD:0.02274 differ image LOOD:0.02260 differ image LOOD:0.02229 differ image LOOD:0.02221 differ image LOOD:0.02193 differ image LOOD:0.02186 differ image LOOD:0.02174 differ image LOOD:0.02174 differ image LOOD:0.02156 opt query LOOD:0.02303 opt query LOOD:0.02302 opt query LOOD:0.02289 opt query LOOD:0.02274 opt query LOOD:0.02260 opt query LOOD:0.02229 opt query LOOD:0.02221 opt query LOOD:0.02193 opt query LOOD:0.02186 opt query LOOD:0.02174 opt query LOOD:0.02174 opt query LOOD:0.02156 differ image LOOD:0.02154 differ image LOOD:0.02152 differ image LOOD:0.02143 differ image LOOD:0.02128 differ image LOOD:0.02103 differ image LOOD:0.02099 differ image LOOD:0.02080 differ image LOOD:0.02078 differ image LOOD:0.02068 differ image LOOD:0.02067 differ image LOOD:0.02065 differ image LOOD:0.02059 opt query LOOD:0.02154 opt query LOOD:0.02152 opt query LOOD:0.02143 opt query LOOD:0.02128 opt query LOOD:0.02103 opt query LOOD:0.02099 opt query LOOD:0.02080 opt query LOOD:0.02077 opt query LOOD:0.02068 opt query LOOD:0.02067 opt query LOOD:0.02065 opt query LOOD:0.02059 differ image LOOD:0.02059 differ image LOOD:0.02021 differ image LOOD:0.02015 differ image LOOD:0.02011 differ image LOOD:0.02011 differ image LOOD:0.02009 differ image LOOD:0.02009 differ image LOOD:0.02002 differ image LOOD:0.02001 differ image LOOD:0.01996 differ image LOOD:0.01988 differ image LOOD:0.01986 opt query LOOD:0.02059 opt query LOOD:0.02021 opt query LOOD:0.02015 opt query LOOD:0.02011 opt query LOOD:0.02011 opt query LOOD:0.02009 opt query LOOD:0.02009 opt query LOOD:0.02002 opt query LOOD:0.02001 opt query LOOD:0.01996 opt query LOOD:0.01988 opt query LOOD:0.01986 differ image LOOD:0.01976 differ image LOOD:0.01973 differ image LOOD:0.01972 differ image LOOD:0.01946 differ image LOOD:0.01917 differ image LOOD:0.01917 differ image LOOD:0.01917 differ image LOOD:0.01897 differ image LOOD:0.01893 differ image LOOD:0.01889 differ image LOOD:0.01847 differ image LOOD:0.01835 opt query LOOD:0.01976 opt query LOOD:0.01973 opt query LOOD:0.01972 opt query LOOD:0.01946 opt query LOOD:0.01917 opt query LOOD:0.01917 opt query LOOD:0.01917 opt query LOOD:0.01897 opt query LOOD:0.01893 opt query LOOD:0.01889 opt query LOOD:0.01847 opt query LOOD:0.01835 differ image LOOD:0.01825 differ image LOOD:0.01822 differ image LOOD:0.01815 differ image LOOD:0.01801 differ image LOOD:0.01783 differ image LOOD:0.01776 differ image LOOD:0.01771 differ image LOOD:0.01762 differ image LOOD:0.01723 differ image LOOD:0.01694 differ image LOOD:0.01651 differ image LOOD:0.01642 opt query LOOD:0.01825 opt query LOOD:0.01822 opt query LOOD:0.01815 opt query LOOD:0.01801 opt query LOOD:0.01783 opt query LOOD:0.01776 opt query LOOD:0.01771 opt query LOOD:0.01762 opt query LOOD:0.01723 opt query LOOD:0.01694 opt query LOOD:0.01651 opt query LOOD:0.01642 differ image LOOD:0.01642 differ image LOOD:0.01625 differ image LOOD:0.01584 differ image LOOD:0.01568 differ image LOOD:0.01547 differ image LOOD:0.01527 differ image LOOD:0.01514 differ image LOOD:0.01514 differ image LOOD:0.01511 differ image LOOD:0.01494 differ image LOOD:0.01485 differ image LOOD:0.01460 opt query LOOD:0.01642 opt query LOOD:0.01625 opt query LOOD:0.01584 opt query LOOD:0.01568 opt query LOOD:0.01547 opt query LOOD:0.01527 opt query LOOD:0.01514 opt query LOOD:0.01514 opt query LOOD:0.01511 opt query LOOD:0.01494 opt query LOOD:0.01485 opt query LOOD:0.01460 differ image LOOD:0.01450 differ image LOOD:0.01425 differ image LOOD:0.01419 differ image LOOD:0.01419 differ image LOOD:0.01404 differ image LOOD:0.01365 differ image LOOD:0.01335 differ image LOOD:0.01315 differ image LOOD:0.01313 differ image LOOD:0.01303 differ image LOOD:0.01238 differ image LOOD:0.01216 opt query LOOD:0.01450 opt query LOOD:0.01425 opt query LOOD:0.01419 opt query LOOD:0.01419 opt query LOOD:0.01404 opt query LOOD:0.01365 opt query LOOD:0.01335 opt query LOOD:0.01315 opt query LOOD:0.01313 opt query LOOD:0.01303 opt query LOOD:0.01238 opt query LOOD:0.01216 differ image LOOD:0.01197 differ image LOOD:0.01190 differ image LOOD:0.01174 differ image LOOD:0.01173 differ image LOOD:0.03048 differ image LOOD:0.01160 differ image LOOD:0.01160 differ image LOOD:0.01128 differ image LOOD:0.01079 differ image LOOD:0.01045 differ image LOOD:0.01037 differ image LOOD:0.01002 opt query LOOD:0.01197 opt query LOOD:0.01190 opt query LOOD:0.01174 opt query LOOD:0.01172 opt query LOOD:0.01166 opt query LOOD:0.01160 opt query LOOD:0.01160 opt query LOOD:0.01128 opt query LOOD:0.01079 opt query LOOD:0.01045 opt query LOOD:0.01037 opt query LOOD:0.01002 Figure 15: Visualization of 500 randomly chosen differing data and query optimized by LOOD (Part 2). We show differing data above the optimized query, and use their LOOD gap to sort the images. Published as a conference paper at ICLR 2024 differ image LOOD:0.00943 differ image LOOD:0.00942 differ image LOOD:0.00903 differ image LOOD:0.00881 differ image LOOD:0.00833 differ image LOOD:0.00832 differ image LOOD:0.00822 differ image LOOD:0.00812 differ image LOOD:0.00759 differ image LOOD:0.00750 differ image LOOD:0.01980 differ image LOOD:0.00741 opt query LOOD:0.00943 opt query LOOD:0.00942 opt query LOOD:0.00903 opt query LOOD:0.00881 opt query LOOD:0.00833 opt query LOOD:0.00832 opt query LOOD:0.00822 opt query LOOD:0.00812 opt query LOOD:0.00758 opt query LOOD:0.00750 opt query LOOD:0.00750 opt query LOOD:0.00741 differ image LOOD:0.00725 differ image LOOD:0.00715 differ image LOOD:0.00595 differ image LOOD:0.00556 differ image LOOD:0.00465 differ image LOOD:0.00464 differ image LOOD:0.05335 differ image LOOD:0.08214 differ image LOOD:0.01642 differ image LOOD:0.00367 differ image LOOD:0.00347 differ image LOOD:0.04034 opt query LOOD:0.00725 opt query LOOD:0.00715 opt query LOOD:0.00595 opt query LOOD:0.00556 opt query LOOD:0.00465 opt query LOOD:0.00464 opt query LOOD:0.00460 opt query LOOD:0.00389 opt query LOOD:0.00383 opt query LOOD:0.00367 opt query LOOD:0.00347 opt query LOOD:0.00344 differ image LOOD:0.03648 differ image LOOD:0.00307 differ image LOOD:0.03577 differ image LOOD:0.03706 differ image LOOD:0.05191 differ image LOOD:0.04711 differ image LOOD:0.08392 differ image LOOD:0.02822 differ image LOOD:0.05252 differ image LOOD:0.02144 differ image LOOD:0.03490 differ image LOOD:0.04911 opt query LOOD:0.00338 opt query LOOD:0.00307 opt query LOOD:0.00282 opt query LOOD:0.00271 opt query LOOD:0.00247 opt query LOOD:0.00206 opt query LOOD:0.00199 opt query LOOD:0.00198 opt query LOOD:0.00193 opt query LOOD:0.00192 opt query LOOD:0.00162 opt query LOOD:0.00160 differ image LOOD:0.05124 differ image LOOD:0.02470 differ image LOOD:0.05657 differ image LOOD:0.02189 differ image LOOD:0.06925 differ image LOOD:0.03109 differ image LOOD:0.02962 differ image LOOD:0.04066 differ image LOOD:0.07845 differ image LOOD:0.04868 differ image LOOD:0.07299 differ image LOOD:0.02364 opt query LOOD:0.00154 opt query LOOD:0.00151 opt query LOOD:0.00148 opt query LOOD:0.00143 opt query LOOD:0.00143 opt query LOOD:0.00142 opt query LOOD:0.00137 opt query LOOD:0.00136 opt query LOOD:0.00134 opt query LOOD:0.00128 opt query LOOD:0.00122 opt query LOOD:0.00122 differ image LOOD:0.07113 differ image LOOD:0.04005 differ image LOOD:0.03044 differ image LOOD:0.03634 differ image LOOD:0.01994 differ image LOOD:0.02035 differ image LOOD:0.03902 differ image LOOD:0.02847 differ image LOOD:0.02229 differ image LOOD:0.04828 differ image LOOD:0.03866 differ image LOOD:0.01505 opt query LOOD:0.00115 opt query LOOD:0.00113 opt query LOOD:0.00113 opt query LOOD:0.00112 opt query LOOD:0.00111 opt query LOOD:0.00109 opt query LOOD:0.00109 opt query LOOD:0.00106 opt query LOOD:0.00104 opt query LOOD:0.00104 opt query LOOD:0.00103 opt query LOOD:0.00102 differ image LOOD:0.02578 differ image LOOD:0.02855 differ image LOOD:0.00437 differ image LOOD:0.01112 differ image LOOD:0.06310 differ image LOOD:0.01424 differ image LOOD:0.02354 differ image LOOD:0.02091 differ image LOOD:0.03824 differ image LOOD:0.03619 differ image LOOD:0.06928 differ image LOOD:0.01827 opt query LOOD:0.00101 opt query LOOD:0.00101 opt query LOOD:0.00099 opt query LOOD:0.00096 opt query LOOD:0.00095 opt query LOOD:0.00091 opt query LOOD:0.00090 opt query LOOD:0.00089 opt query LOOD:0.00089 opt query LOOD:0.00087 opt query LOOD:0.00087 opt query LOOD:0.00087 differ image LOOD:0.05793 differ image LOOD:0.04239 differ image LOOD:0.03110 differ image LOOD:0.03124 differ image LOOD:0.02729 differ image LOOD:0.03789 differ image LOOD:0.03684 differ image LOOD:0.04845 differ image LOOD:0.01951 differ image LOOD:0.03032 differ image LOOD:0.02307 differ image LOOD:0.02515 opt query LOOD:0.00087 opt query LOOD:0.00084 opt query LOOD:0.00083 opt query LOOD:0.00080 opt query LOOD:0.00080 opt query LOOD:0.00080 opt query LOOD:0.00079 opt query LOOD:0.00079 opt query LOOD:0.00078 opt query LOOD:0.00078 opt query LOOD:0.00077 opt query LOOD:0.00075 differ image LOOD:0.02202 differ image LOOD:0.02633 differ image LOOD:0.01066 differ image LOOD:0.03576 differ image LOOD:0.01265 differ image LOOD:0.03237 differ image LOOD:0.03547 differ image LOOD:0.06689 differ image LOOD:0.03437 differ image LOOD:0.02633 differ image LOOD:0.01637 differ image LOOD:0.01051 opt query LOOD:0.00075 opt query LOOD:0.00071 opt query LOOD:0.00071 opt query LOOD:0.00071 opt query LOOD:0.00067 opt query LOOD:0.00064 opt query LOOD:0.00062 opt query LOOD:0.00061 opt query LOOD:0.00061 opt query LOOD:0.00061 opt query LOOD:0.00059 opt query LOOD:0.00059 Figure 16: Visualization of 500 randomly chosen differing data and query optimized by LOOD (Part 3). We show differing data above the optimized query, and use their LOOD gap to sort the images. Published as a conference paper at ICLR 2024 differ image LOOD:0.01583 differ image LOOD:0.01854 differ image LOOD:0.01069 differ image LOOD:0.01774 differ image LOOD:0.00778 differ image LOOD:0.01150 differ image LOOD:0.01709 differ image LOOD:0.02363 differ image LOOD:0.03169 differ image LOOD:0.03482 differ image LOOD:0.01416 differ image LOOD:0.01411 opt query LOOD:0.00058 opt query LOOD:0.00057 opt query LOOD:0.00056 opt query LOOD:0.00056 opt query LOOD:0.00053 opt query LOOD:0.00053 opt query LOOD:0.00051 opt query LOOD:0.00051 opt query LOOD:0.00050 opt query LOOD:0.00048 opt query LOOD:0.00048 opt query LOOD:0.00048 differ image LOOD:0.04709 differ image LOOD:0.02629 differ image LOOD:0.02209 differ image LOOD:0.01680 differ image LOOD:0.02008 differ image LOOD:0.02257 differ image LOOD:0.04513 differ image LOOD:0.01833 differ image LOOD:0.01362 differ image LOOD:0.02947 differ image LOOD:0.02422 differ image LOOD:0.03767 opt query LOOD:0.00047 opt query LOOD:0.00047 opt query LOOD:0.00046 opt query LOOD:0.00046 opt query LOOD:0.00046 opt query LOOD:0.00046 opt query LOOD:0.00045 opt query LOOD:0.00045 opt query LOOD:0.00044 opt query LOOD:0.00044 opt query LOOD:0.00044 opt query LOOD:0.00043 differ image LOOD:0.02585 differ image LOOD:0.02182 differ image LOOD:0.00717 differ image LOOD:0.01367 differ image LOOD:0.01369 differ image LOOD:0.01359 differ image LOOD:0.00866 differ image LOOD:0.01945 differ image LOOD:0.02525 differ image LOOD:0.02015 differ image LOOD:0.02675 differ image LOOD:0.00433 opt query LOOD:0.00042 opt query LOOD:0.00042 opt query LOOD:0.00041 opt query LOOD:0.00039 opt query LOOD:0.00037 opt query LOOD:0.00036 opt query LOOD:0.00036 opt query LOOD:0.00035 opt query LOOD:0.00034 opt query LOOD:0.00033 opt query LOOD:0.00032 opt query LOOD:0.00031 differ image LOOD:0.00779 differ image LOOD:0.02493 differ image LOOD:0.02392 differ image LOOD:0.00882 differ image LOOD:0.02291 differ image LOOD:0.02842 differ image LOOD:0.03416 differ image LOOD:0.02719 differ image LOOD:0.03477 differ image LOOD:0.02603 differ image LOOD:0.02004 differ image LOOD:0.01747 opt query LOOD:0.00031 opt query LOOD:0.00031 opt query LOOD:0.00031 opt query LOOD:0.00031 opt query LOOD:0.00030 opt query LOOD:0.00029 opt query LOOD:0.00029 opt query LOOD:0.00027 opt query LOOD:0.00026 opt query LOOD:0.00026 opt query LOOD:0.00026 opt query LOOD:0.00025 differ image LOOD:0.01911 differ image LOOD:0.00400 differ image LOOD:0.02012 differ image LOOD:0.00856 differ image LOOD:0.01670 differ image LOOD:0.00872 differ image LOOD:0.02935 differ image LOOD:0.01848 differ image LOOD:0.03761 differ image LOOD:0.01219 differ image LOOD:0.01360 differ image LOOD:0.02007 opt query LOOD:0.00025 opt query LOOD:0.00023 opt query LOOD:0.00023 opt query LOOD:0.00023 opt query LOOD:0.00022 opt query LOOD:0.00022 opt query LOOD:0.00022 opt query LOOD:0.00022 opt query LOOD:0.00021 opt query LOOD:0.00021 opt query LOOD:0.00019 opt query LOOD:0.00018 differ image LOOD:0.00368 differ image LOOD:0.00975 differ image LOOD:0.00647 differ image LOOD:0.01842 differ image LOOD:0.01831 differ image LOOD:0.02481 differ image LOOD:0.03865 differ image LOOD:0.01153 differ image LOOD:0.01375 differ image LOOD:0.02420 differ image LOOD:0.02273 differ image LOOD:0.02417 opt query LOOD:0.00018 opt query LOOD:0.00018 opt query LOOD:0.00017 opt query LOOD:0.00017 opt query LOOD:0.00016 opt query LOOD:0.00016 opt query LOOD:0.00016 opt query LOOD:0.00015 opt query LOOD:0.00014 opt query LOOD:0.00014 opt query LOOD:0.00014 opt query LOOD:0.00013 differ image LOOD:0.02124 differ image LOOD:0.01064 differ image LOOD:0.01941 differ image LOOD:0.02026 differ image LOOD:0.01117 differ image LOOD:0.01717 differ image LOOD:0.02648 differ image LOOD:0.01433 differ image LOOD:0.01058 differ image LOOD:0.01447 differ image LOOD:0.01330 differ image LOOD:0.02837 opt query LOOD:0.00013 opt query LOOD:0.00013 opt query LOOD:0.00012 opt query LOOD:0.00012 opt query LOOD:0.00012 opt query LOOD:0.00011 opt query LOOD:0.00011 opt query LOOD:0.00010 opt query LOOD:0.00010 opt query LOOD:0.00010 opt query LOOD:0.00009 opt query LOOD:0.00009 differ image LOOD:0.04528 differ image LOOD:0.00291 differ image LOOD:0.00686 differ image LOOD:0.02288 differ image LOOD:0.00579 differ image LOOD:0.01565 differ image LOOD:0.02777 differ image LOOD:0.02554 differ image LOOD:0.00456 differ image LOOD:0.00594 differ image LOOD:0.00925 differ image LOOD:0.03961 opt query LOOD:0.00009 opt query LOOD:0.00009 opt query LOOD:0.00008 opt query LOOD:0.00007 opt query LOOD:0.00007 opt query LOOD:0.00007 opt query LOOD:0.00006 opt query LOOD:0.00006 opt query LOOD:0.00006 opt query LOOD:0.00006 opt query LOOD:0.00006 opt query LOOD:0.00005 Figure 17: Visualization of 500 randomly chosen differing data and query optimized by LOOD (Part 4). We show differing data above the optimized query, and use their LOOD gap to sort the images. Published as a conference paper at ICLR 2024 differ image LOOD:0.01168 differ image LOOD:0.01904 differ image LOOD:0.02628 differ image LOOD:0.01840 differ image LOOD:0.01897 differ image LOOD:0.00175 differ image LOOD:0.01233 differ image LOOD:0.00614 differ image LOOD:0.03230 differ image LOOD:0.00726 differ image LOOD:0.00734 differ image LOOD:0.00226 opt query LOOD:0.00005 opt query LOOD:0.00005 opt query LOOD:0.00005 opt query LOOD:0.00005 opt query LOOD:0.00005 opt query LOOD:0.00005 opt query LOOD:0.00004 opt query LOOD:0.00004 opt query LOOD:0.00004 opt query LOOD:0.00003 opt query LOOD:0.00003 opt query LOOD:0.00003 differ image LOOD:0.01252 differ image LOOD:0.01731 differ image LOOD:0.01999 differ image LOOD:0.00641 differ image LOOD:0.00956 differ image LOOD:0.01840 differ image LOOD:0.01836 differ image LOOD:0.01037 differ image LOOD:0.02315 differ image LOOD:0.01998 differ image LOOD:0.00975 differ image LOOD:0.02696 opt query LOOD:0.00003 opt query LOOD:0.00003 opt query LOOD:0.00003 opt query LOOD:0.00002 opt query LOOD:0.00002 opt query LOOD:0.00002 opt query LOOD:0.00001 opt query LOOD:0.00001 opt query LOOD:0.00001 opt query LOOD:0.00001 opt query LOOD:0.00001 opt query LOOD:0.00000 differ image LOOD:0.01009 differ image LOOD:0.02939 differ image LOOD:0.03047 differ image LOOD:0.01789 differ image LOOD:0.01005 differ image LOOD:0.01233 differ image LOOD:0.01546 differ image LOOD:0.00195 differ image LOOD:0.01883 differ image LOOD:0.03404 differ image LOOD:0.00579 differ image LOOD:0.05763 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 differ image LOOD:0.03037 differ image LOOD:0.02439 differ image LOOD:0.00886 differ image LOOD:0.01063 differ image LOOD:0.03072 differ image LOOD:0.03627 differ image LOOD:0.02162 differ image LOOD:0.01157 differ image LOOD:0.01389 differ image LOOD:0.00880 differ image LOOD:0.01145 differ image LOOD:0.02190 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 differ image LOOD:0.02863 differ image LOOD:0.02917 differ image LOOD:0.02089 differ image LOOD:0.03417 differ image LOOD:0.02626 differ image LOOD:0.00657 differ image LOOD:0.00738 differ image LOOD:0.01073 differ image LOOD:0.00952 differ image LOOD:0.00174 differ image LOOD:0.00458 differ image LOOD:0.01520 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 differ image LOOD:0.00992 differ image LOOD:0.01582 differ image LOOD:0.01820 differ image LOOD:0.00635 differ image LOOD:0.00505 differ image LOOD:0.00255 differ image LOOD:0.00570 differ image LOOD:0.00645 differ image LOOD:0.00446 differ image LOOD:0.03138 differ image LOOD:0.00661 differ image LOOD:0.01677 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 differ image LOOD:0.00846 differ image LOOD:0.01485 differ image LOOD:0.02242 differ image LOOD:0.01209 differ image LOOD:0.02104 differ image LOOD:0.00935 differ image LOOD:0.00601 differ image LOOD:0.01314 differ image LOOD:0.04021 differ image LOOD:0.00694 differ image LOOD:0.00436 differ image LOOD:0.03243 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 differ image LOOD:0.02508 differ image LOOD:0.01027 differ image LOOD:0.00296 differ image LOOD:0.02684 differ image LOOD:0.00767 differ image LOOD:0.00742 differ image LOOD:0.01070 differ image LOOD:0.02532 differ image LOOD:0.02327 differ image LOOD:0.00152 differ image LOOD:0.01116 differ image LOOD:0.01594 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 Figure 18: Visualization of 500 randomly chosen differing data and query optimized by LOOD (Part 5). We show differing data above the optimized query, and use their LOOD gap to sort the images. Published as a conference paper at ICLR 2024 differ image LOOD:0.00994 differ image LOOD:0.00327 differ image LOOD:0.01066 differ image LOOD:0.00181 differ image LOOD:0.03200 differ image LOOD:0.01688 differ image LOOD:0.02805 differ image LOOD:0.00352 differ image LOOD:0.00961 differ image LOOD:0.00846 differ image LOOD:0.01028 differ image LOOD:0.01484 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 differ image LOOD:0.01617 differ image LOOD:0.00921 differ image LOOD:0.03189 differ image LOOD:0.02536 differ image LOOD:0.00318 differ image LOOD:0.01481 differ image LOOD:0.00467 differ image LOOD:0.00643 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 opt query LOOD:0.00000 Figure 19: Visualization of 500 randomly chosen differing data and query optimized by LOOD (Part 6). We show differing data above the optimized query, and use their LOOD gap to sort the images.