# quantification_of_uncertainty_with_adversarial_models__17b21084.pdf Quantification of Uncertainty with Adversarial Models Kajetan Schweighofer Lukas Aichberger Mykyta Ielanskyi Günter Klambauer Sepp Hochreiter ELLIS Unit Linz and LIT AI Lab, Institute for Machine Learning, Johannes Kepler University Linz, Austria Joint first authors Quantifying uncertainty is important for actionable predictions in real-world applications. A crucial part of predictive uncertainty quantification is the estimation of epistemic uncertainty, which is defined as an integral of the product between a divergence function and the posterior. Current methods such as Deep Ensembles or MC dropout underperform at estimating the epistemic uncertainty, since they primarily consider the posterior when sampling models. We suggest Quantification of Uncertainty with Adversarial Models (QUAM) to better estimate the epistemic uncertainty. QUAM identifies regions where the whole product under the integral is large, not just the posterior. Consequently, QUAM has lower approximation error of the epistemic uncertainty compared to previous methods. Models for which the product is large correspond to adversarial models (not adversarial examples!). Adversarial models have both a high posterior as well as a high divergence between their predictions and that of a reference model. Our experiments show that QUAM excels in capturing epistemic uncertainty for deep learning models and outperforms previous methods on challenging tasks in the vision domain. 1 Introduction Actionable predictions typically require risk assessment based on predictive uncertainty quantification [Apostolakis, 1991]. This is of utmost importance in high stake applications, such as medical diagnosis or drug discovery, where human lives or extensive investments are at risk. In such settings, even a single prediction has far-reaching real-world impact, thus necessitating the most precise quantification of the associated uncertainties. Furthermore, foundation models or specialized models that are obtained externally are becoming increasingly prevalent, also in high stake applications. It is crucial to assess the robustness and reliability of those unknown models before applying them. Therefore, the predictive uncertainty of given, pre-selected models at specific test points should be quantified, which we address in this work. We consider predictive uncertainty quantification (see Fig. 1) for deep neural networks [Gal, 2016, Hüllermeier and Waegeman, 2021]. According to Vesely and Rasmuson [1984], Apostolakis [1991], Helton [1993], Mc Kone [1994], Helton [1997], predictive uncertainty can be categorized into two types. First, aleatoric (Type A, variability, stochastic, true, irreducible) uncertainty refers to the variability when drawing samples or when repeating the same experiment. Second, epistemic (Type B, lack of knowledge, subjective, reducible) uncertainty refers to the lack of knowledge about the true model. Epistemic uncertainty can result from imprecision in parameter estimates, incompleteness in modeling, or indefiniteness in the applicability of the model. While aleatoric uncertainty cannot be reduced, epistemic uncertainty can be reduced by more data, better models, or more knowledge about the problem. We follow Helton [1997] and consider epistemic uncertainty as the imprecision or variability of parameters that determine the predictive distribution. Vesely and Rasmuson [1984] 37th Conference on Neural Information Processing Systems (Neur IPS 2023). Figure 1: Adversarial models. For the red test point, the predictive uncertainty is high as it is far from the training data. High uncertainties are detected by different adversarial models that assign the red test point to different classes, although all of them explain the training data equally well. As a result, the true class of the test point remains ambiguous. calls this epistemic uncertainty parameter uncertainty , which results from an imperfect learning algorithm or from insufficiently many training samples. Consequently, we consider predictive uncertainty quantification as characterizing a probabilistic model of the world. In this context, aleatoric uncertainty refers to the inherent stochasticity of sampling outcomes from the predictive distribution of the model and epistemic uncertainty refers to the uncertainty about model parameters. Current uncertainty quantification methods such as Deep Ensembles [Lakshminarayanan et al., 2017] or Monte-Carlo (MC) dropout [Gal and Ghahramani, 2016] underperform at estimating the epistemic uncertainty [Wilson and Izmailov, 2020, Parker-Holder et al., 2020, Angelo and Fortuin, 2021], since they primarily consider the posterior when sampling models. Thus they are prone to miss important posterior modes, where the whole integrand of the integral defining the epistemic uncertainty is large. We introduce Quantification of Uncertainty with Adversarial Models (QUAM) to identify those posterior modes. QUAM searches for those posterior modes via adversarial models and uses them to reduce the approximation error when estimating the integral that defines the epistemic uncertainty. Adversarial models are characterized by a large value of the integrand of the integral defining the epistemic uncertainty. Thus, they considerably differ to the reference model s prediction at a test point while having a similarly high posterior probability. Consequently, they are counterexamples of the reference model that predict differently for a new input, but explain the training data equally well. Fig. 1 shows examples of adversarial models which assign different classes to a test point, but agree on the training data. A formal definition is given by Def. 1. It is essential to note that adversarial models are a new concept that is to be distinguished from other concepts that include the term adversarial in their naming, such as adversarial examples [Szegedy et al., 2013, Biggio et al., 2013], adversarial training [Goodfellow et al., 2015], generative adversarial networks [Goodfellow et al., 2014] or adversarial model-based RL [Rigter et al., 2022]. Our main contributions are: We introduce QUAM as a framework for uncertainty quantification. QUAM approximates the integral that defines the epistemic uncertainty substantially better than previous methods, since it reduces the approximation error of the integral estimator. We introduce the concept of adversarial models for estimating posterior integrals with non-negative integrands. For a given test point, adversarial models have considerably different predictions than a reference model while having similarly high posterior probability. We introduce a new setting for uncertainty quantification, where the uncertainty of a given, preselected model is quantified. 2 Current Methods to Estimate the Epistemic Uncertainty Definition of Predictive Uncertainty. Predictive uncertainty quantification is about describing a probabilistic model of the world, where aleatoric uncertainty refers to the inherent stochasticity of sampling outcomes from the predictive distribution of the model and epistemic uncertainty refers to the uncertainty about model parameters. We consider two distinct settings of predictive uncertainty quantification. Setting (a) concerns with the predictive uncertainty at a new test point expected under all plausible models given the training dataset [Gal, 2016, Hüllermeier and Waegeman, 2021]. This definition of uncertainty comprises how differently possible models predict (epistemic) and how confident each model is about its prediction (aleatoric). Setting (b) concerns with the predictive uncertainty at a new test point for a given, pre-selected model. This definition of uncertainty comprises how likely this model is the true model that generated the training dataset (epistemic) [Apostolakis, 1991, Helton, 1997] and how confident this model is about its prediction (aleatoric). As an example, assume we have initial data from an epidemic, but we do not know the exact infection rate, which is a parameter of a prediction model. The goal is to predict the number of infected persons at a specific time in the future, where each point in time is a test point. In setting (a), we are interested in the uncertainty of test point predictions of all models using infection rates that explain the initial data. If all likely models agree for a given new test point, the prediction of any of those models can be trusted, otherwise we can not trust the prediction regardless of which model is selected in the end. In setting (b), we have selected a specific infection rate from the initial data as parameter for our model to make predictions. We refer to this model as the given, pre-selected model. However, we do not know the true infection rate of the epidemic. All models with infection rates that are consistent with the initial data are likely to be the true model. If all likely models agree with the given, pre-selected model for a given new test point, the prediction of the model can be trusted. 2.1 Measuring Predictive Uncertainty We consider the predictive distribution of a single model p(y | x, w), which is a probabilistic model of the world. Depending on the task, the predictive distribution of this probabilistic model can be a categorical distribution for classification or a Gaussian distribution for regression. The Bayesian framework offers a principled way to treat the uncertainty about the parameters through the posterior p(w | D) p(D | w)p(w) for a given dataset D. The Bayesian model average (BMA) predictive distribution is given by p(y | x, D) = R W p(y | x, w)p( w | D)d w. Following Gal [2016], Depeweg et al. [2018], Smith and Gal [2018], Hüllermeier and Waegeman [2021], the uncertainty of the BMA predictive distribution is commonly measured by the entropy H[p(y | x, D)]. It refers to the total uncertainty, which can be decomposed into an aleatoric and an epistemic part. The BMA predictive entropy is equal to the posterior expectation of the cross-entropy CE[ , ] between the predictive distribution of candidate models and the BMA, which corresponds to setting (a). In setting (b), the cross-entropy is between the predictive distribution of the given, pre-selected model and candidate models. Details about the entropy and cross-entropy as measures of uncertainty are given in Sec. B.1.1 in the appendix. In the following, we formalize how to measure the notions of uncertainty in setting (a) and (b) using the expected cross-entropy over the posterior. Setting (a): Expected uncertainty when selecting a model. We estimate the predictive uncertainty at a test point x when selecting a model w given a training dataset D. The total uncertainty is the expected cross-entropy between the predictive distribution of candidate models p(y | x, w) and the BMA predictive distribution p(y | x, D), where the expectation is with respect to the posterior: Z W CE[p(y | x, w) , p(y | x, D)] p( w | D) d w = H[p(y | x, D)] (1) W H[p(y | x, w)] p( w | D) d w + I[Y ; W | x, D] W H[p(y | x, w)] p( w | D) d w | {z } aleatoric W DKL(p(y | x, w) p(y | x, D)) p( w | D) d w | {z } epistemic The aleatoric uncertainty characterizes the uncertainty due to the expected stochasticity of sampling outcomes from the predictive distribution of candidate models p(y | x, w). The epistemic uncertainty characterizes the uncertainty due to the mismatch between the predictive distribution of candidate models and the BMA predictive distribution. It is measured by the mutual information I[ ; ], between the prediction Y and the model parameters W for a given test point and dataset, which is equivalent to the posterior expectation of the KL-divergence DKL( ) between the predictive distributions of candidate models and the BMA predictive distribution. Derivations are given in appendix Sec. B.1. Setting (b): Uncertainty of a given, pre-selected model. We estimate the predictive uncertainty of a given, pre-selected model w at a test point x. We assume that the dataset D is produced according to the true distribution p(y | x, w ) parameterized by w . The posterior p( w | D) is an estimate of how likely w match w . For epistemic uncertainty, we should measure the difference between the predictive distributions under w and w , but w is unknown. Therefore, we measure the expected difference between the predictive distributions under w and w. In accordance with Apostolakis [1991] and Helton [1997], the total uncertainty is therefore the expected cross-entropy between the predictive distributions of a given, pre-selected model w and candidate models w, any of which could be the true model w according to the posterior: Z W CE[p(y | x, w) , p(y | x, w)] p( w | D) d w (2) = H[p(y | x, w)] | {z } aleatoric W DKL(p(y | x, w) p(y | x, w)) p( w | D) d w | {z } epistemic The aleatoric uncertainty characterizes the uncertainty due to the stochasticity of sampling outcomes from the predictive distribution of the given, pre-selected model p(y | x, w). The epistemic uncertainty characterizes the uncertainty due to the mismatch between the predictive distribution of the given, pre-selected model and the predictive distribution of candidate models that could be the true model. Derivations and further details are given in appendix Sec. B.1. 2.2 Estimating the Integral for Epistemic Uncertainty Current methods for predictive uncertainty quantification suffer from underestimating the epistemic uncertainty [Wilson and Izmailov, 2020, Parker-Holder et al., 2020, Angelo and Fortuin, 2021]. The epistemic uncertainty is given by the respective terms in Eq. (1) for setting (a) and Eq. (2) for our new setting (b). To estimate these integrals, almost all methods use gradient descent on the training data. Thus, posterior modes that are hidden from the gradient flow remain undiscovered and the epistemic uncertainty is underestimated [Shah et al., 2020, Angelo and Fortuin, 2021]. An illustrative example is depicted in Fig. 2. Posterior expectations as in Eq. (1) and Eq. (2) that define the epistemic uncertainty are generally approximated using Monte Carlo integration. A good approximation of posterior integrals through Monte Carlo integration requires to capture all large values of the nonnegative integrand [Wilson and Izmailov, 2020], which is not only large values of the posterior, but also large values of the KL-divergence. Variational inference [Graves, 2011, Blundell et al., 2015, Gal and Ghahramani, 2016] and ensemble methods [Lakshminarayanan et al., 2017] estimate the posterior integral based on models with high posterior. Posterior modes may be hidden from gradient descent based techniques as they only discover mechanistically similar models. Two models are mechanistically similar if they rely on the same input attributes for making their predictions, that is, they are invariant to the same input attributes [Lubana et al., 2022]. However, gradient descent will always start by extracting input attributes that are highly correlated to the target as they determine the steepest descent in the error (a) Deep Ensembles (b) MC dropout (c) Training data + new test point Figure 2: Model prediction analysis. Softmax outputs (black) of individual models of Deep Ensembles (a) and MC dropout (b), as well as their average output (red) on a probability simplex. Models were selected on the training data, and evaluated on the new test point (red) depicted in (c). The background color denotes the maximum likelihood of the training data that is achievable by a model having a predictive distribution (softmax values) equal to the respective location on the simplex. Deep Ensembles and MC dropout fail to find models predicting the orange class, although there would be likely models that do so. Details on the experimental setup are given in the appendix, Sec. C.2. landscape. These input attributes create a large basin in the error landscape into which the parameter vector is drawn via gradient descent. Consequently, other modes further away from such basins are almost never found [Shah et al., 2020, Angelo and Fortuin, 2021]. Thus, the epistemic uncertainty is underestimated. Another reason that posterior modes may be hidden from gradient descent is the presence of different labeling hypotheses. If there is more than one way to explain the training data, gradient descent will use all of them as they give the steepest error descent [Scimeca et al., 2022]. Other work focuses on MCMC sampling according to the posterior distribution, which is approximated by stochastic gradient variants [Welling and Teh, 2011, Chen et al., 2014] for large datasets and models. Those are known to face issues to efficiently explore the highly complex and multimodal parameter space and escape local posterior modes. There are attempts to alleviate the problem [Li et al., 2016, Zhang et al., 2020]. However, those methods do not explicitly look for important posterior modes, where the predictive distributions of sampled models contribute strongly to the approximation of the posterior integral, and thus have large values for the KL-divergence. 3 Adversarial Models to Estimate the Epistemic Uncertainty Intuition. The epistemic uncertainty in Eq. (1) for setting (a) compares possible models with the BMA. Thus, the BMA is used as reference model. The epistemic uncertainty in Eq. (2) for our new setting (b) compares models that are candidates for the true model with the given, pre-selected model. Thus, the given, pre-selected model is used as reference model. If the reference model makes some prediction at the test point, and if other models (the adversaries) make different predictions while explaining the training data equally well, then one should be uncertain about the prediction. Adversarial models are plausible outcomes of model selection, while having a different prediction at the test data point than the reference model. In court, the same principle is used: if the prosecutor presents a scenario but the advocate presents alternative equally plausible scenarios, the judges become uncertain about what happened and rule in favor of the defendant. We use adversarial models to identify locations where the integrand of the integral defining the epistemic uncertainty in Eq. (1) or Eq. (2) is large. These locations are used to construct a mixture distribution that is used for mixture importance sampling to estimate the desired integrals. Using the mixture distribution for sampling, we aim to considerably reduce the approximation error of the estimator of the epistemic uncertainty. Mixture Importance Sampling. We estimate the integrals of epistemic uncertainty in Eq. (1) and in Eq. (2). In the following, we focus on setting (b) with Eq. (2), but all results hold for setting (a) with Eq. (1) as well. Most methods sample from a distribution q( w) to approximate the integral: W DKL(p(y | x, w) p(y | x, w)) p( w | D) d w = Z q( w) q( w) d w , (3) where u(x, w, w) = DKL(p(y | x, w) p(y | x, w))p( w | D). As with Deep Ensembles or MC dropout, posterior sampling is often approximated by a sampling distribution q( w) that is close to p( w | D). Monte Carlo (MC) integration estimates v by u(x, w, wn) q( wn) , wn q( w) . (4) If the posterior has different modes, the estimate under a unimodal approximate distribution has high variance and converges very slowly [Steele et al., 2006]. Thus, we use mixture importance sampling (MIS) [Hesterberg, 1995]. MIS utilizes a mixture distribution instead of the unimodal distribution in standard importance sampling [Owen and Zhou, 2000]. Furthermore, many MIS methods iteratively enhance the sampling distribution by incorporating new modes [Raftery and Bao, 2010]. In contrast to the usually applied iterative enrichment methods which find new modes by chance, we have a much more favorable situation. We can explicitly search for posterior modes where the KL divergence is large, as we can cast it as a supervised learning problem. Each of these modes determines the location of a mixture component of the mixture distribution. Theorem 1. The expected mean squared error of importance sampling with q( w) can be bounded by Eq( w) h (ˆv v)2i Eq( w) " u(x, w, w) 2# 4 N . (5) Proof. The inequality Eq. (5) follows from Theorem 1 in Akyildiz and Míguez [2021], when considering 0 u(x, w, w) as an unnormalized distribution and setting φ = 1. Approximating only the posterior p( w | D) as done by Deep Ensembles or MC dropout is insufficient to guarantee a low expected mean squared error, since the sampling variance cannot be bounded (see appendix Sec. B.2). Corollary 1. With constant c, Eq( w) h (ˆv v)2i 4c2/N holds if u(x, w, w) c q( w). Consequently, q( w) must have modes where u(x, w, w) has modes even if the q-modes are a factor c smaller. The modes of u(x, w, w) are models w with both high posterior and high KL-divergence. We are searching for these modes to determine the locations wk of the components of a mixture distribution q( w): k=1 αk P( w ; wk, θ) , (6) with αk = 1/K for K such models wk that determine a mode. Adversarial model search finds the locations wk of the mixture components, where wk is an adversarial model. The reference model does not define a mixture component, as it has zero KL-divergence to itself. We then sample from a distribution P at the local posterior mode with mean wk and a set of shape parameters θ. The simplest choice for P is a Dirac delta distribution, but one could use e.g. a local Laplace approximation of the posterior [Mac Kay, 1992], or a Gaussian distribution in some weight-subspace [Maddox et al., 2019]. Furthermore, one could use wk as starting point for SG-MCMC chains [Welling and Teh, 2011, Chen et al., 2014, Zhang et al., 2020, 2022]. More details regarding MIS are given in the appendix in Sec. B.2. In the following, we propose an algorithm to find those models with both high posterior and high KL-divergence to the predictive distribution of the reference model. Adversarial Model Search. Adversarial model search is the concept of searching for a model that has a large distance / divergence to the reference predictive distribution and at the same time a high posterior. We call such models adversarial models as they act as adversaries to the reference model by contradicting its prediction. A formal definition of an adversarial model is given by Def. 1: Definition 1. Given are a new test data point x, a reference conditional probability model p(y | x, w) from a model class parameterized by w, a divergence or distance measure D( , ) for probability distributions, γ > 0, Λ > 0, and a dataset D. Then a model with parameters w that satisfies the inequalities | log p(w | D) log p( w | D)| γ and D(p(y | x, w), p(y | x, w)) Λ is called an (γ, Λ) adversarial model. Adversarial model search corresponds to the following optimization problem: max δ D(p(y | x, w) , p(y | x, w + δ)) s.t. log p(w | D) log p(w + δ | D) γ . (7) We are searching for a weight perturbation δ that maximizes the distance D( , ) to the reference distribution without decreasing the log posterior more than γ. The search for adversarial models is restricted to δ , for example by only optimizing the last layer of the reference model or by bounding the norm of δ. This optimization problem can be rewritten as: max δ D(p(y | x, w) , p(y | x, w + δ)) + c (log p(w + δ | D) log p(w | D) + γ) . (8) where c is a hyperparameter. According to the Karush-Kuhn-Tucker (KKT) theorem [Karush, 1939, Kuhn and Tucker, 1950, May, 2020, Luenberger and Ye, 2016]: If δ is the solution to the problem Eq. (7), then there exists a c 0 with δL(δ , c ) = 0 (L is the Lagrangian) and c (log p(w | D) log p(w + δ | D) γ) = 0. This is a necessary condition for an optimal point according to Theorem on Page 326 of Luenberger and Ye [2016]. We solve this optimization problem by the penalty method, which relies on the KKT theorem [Zangwill, 1967]. A penalty algorithm solves a series of unconstrained problems, solutions of which converge to the solution of the original constrained problem (see e.g. Fiacco and Mc Cormick [1990]). The unconstrained problems are constructed by adding a weighted penalty function measuring the constraint violation to the objective function. At every step, the weight of the penalty is increased, thus the constraints are less violated. If exists, the solution to the constraint optimization problem is an adversarial model that is located within a posterior mode but has a different predictive distribution compared to the reference model. We summarize the adversarial model search in Alg. 1. Algorithm 1 Adversarial Model Search (used in QUAM) Supplies: Adversarial model w with maximum Ladv and Lpen 0 Requires: Test point x, training dataset D = {(xk, yk)}K k=1, reference model w, loss function l, loss of reference model on the training dataset Lref = 1 K PK k=1 l(p(y | xk, w), yk), minimization procedure MINIMIZE, number of penalty iterations M, initial penalty parameter c0, penalty parameter increase scheduler η, slack parameter γ, distance / divergence measure D( , ). 1: w w; w w; c c0 2: for m 1 to M do 3: Lpen 1 K PK k=1 l(p(y | xk, w), yk) (Lref + γ) 4: Ladv D(p(y | x, w) , p(y | x, w)) 5: L Ladv + c Lpen 6: w MINIMIZE(L( w)) 7: if Ladv larger than all previous and Lpen 0 then 8: w w 9: c η(c) 10: return w Practical Implementation. Empirically, we found that directly executing the optimization procedure defined in Alg. 1 tends to result in adversarial models with similar predictive distribution for a given input across multiple searches. The vanilla implementation of Alg. 1 corresponds to an untargeted attack, known from the literature on adversarial attacks [Szegedy et al., 2013, Biggio et al., 2013]. To prevent the searches from converging to a single solution, we optimize the cross-entropy loss for one specific class during each search, which corresponds to a targeted attack. Each resulting adversarial model represents a local optimum of Eq. (7). We execute as many adversarial model searches as there are classes, dedicating one search to each class, unless otherwise specified. To compute Eq. (4), we use the predictive distributions p(y | x, w) of all models w encountered during each penalty iteration of all searches, weighted by their posterior probability. The posterior probability is approximated with the negative exponential training loss, the likelihood, of models w. This approximate posterior probability is scaled with a temperature parameter, set as a hyperparameter. Further details are given in the appendix Sec. C.1. 4 Experiments In this section, we compare previous uncertainty quantification methods and our method QUAM in a set of experiments. First, we assess the considered methods on a synthetic benchmark, on which it is feasible to compute a ground truth epistemic uncertainty. Then, we conduct challenging out-ofdistribution (OOD) detection, adversarial example detection, misclassification detection and selective prediction experiments in the vision domain. We compare (1) QUAM, (2) cyclical Stochastic Gradient Hamiltonian Monte Carlo (c SG-HMC) [Zhang et al., 2020], (3) an efficient Laplace approximation (Laplace) [Daxberger et al., 2021], (4) MC dropout (MCD) [Gal and Ghahramani, 2016] and (5) Deep Ensembles (DE) [Lakshminarayanan et al., 2017] on their ability to estimate the epistemic uncertainty. Those baseline methods, especially Deep Ensembles, are persistently among the best performing uncertainty quantification methods across various benchmark tasks [Filos et al., 2019, Ovadia et al., 2019, Caldeira and Nord, 2020, Band et al., 2022] 4.1 Epistemic Uncertainty on Synthetic Dataset We evaluated all considered methods on the two-moons dataset, created using the implementation of Pedregosa et al. [2011]. To obtain the ground truth uncertainty, we utilized Hamiltonian Monte Carlo (HMC) [Neal, 1996]. HMC is regarded as the most precise algorithm to approximate posterior expectations [Izmailov et al., 2021], but necessitates extreme computational expenses to be applied to models and datasets of practical scale. The results are depicted in Fig. 3. QUAM most closely matches the uncertainty estimate of the ground truth epistemic uncertainty obtained by HMC and excels especially on the regions further away from the decision boundary such as in the top left and bottom right of the plots. All other methods fail to capture the epistemic uncertainty in those regions as gradient descent on the training set fails to capture posterior modes with alternative predictive distributions in those parts and misses the important integral components. Experimental details and results for the epistemic uncertainty as in Eq. (2) are given in the appendix Sec. C.3. (a) Ground Truth - HMC (b) c SG-HMC (c) Laplace (d) MC dropout (e) Deep Ensembles (f) Our Method - QUAM Figure 3: Epistemic uncertainty as in Eq. (1) for two-moons. Yellow denotes high epistemic uncertainty. Purple denotes low epistemic uncertainty. HMC is considered as ground truth [Izmailov et al., 2021] and is most closely matched by QUAM. Artifacts for QUAM arise because it is applied to each test point individually, whereas other methods use the same sampled models for all test points. 4.2 Epistemic Uncertainty on Vision Datasets We benchmark the ability of different methods to estimate the epistemic uncertainty of a given, pre-selected model (setting (b) as in Eq. (2)) in the context of (i) out-of-distribution (OOD) detection, (ii) adversarial example detection, (iii) misclassification detection and (iv) selective prediction. In all experiments, we assume to have access to a pre-trained model on the in-distribution (ID) training dataset, which we refer to as reference model. The epistemic uncertainty is expected to be higher for OOD samples, as they can be assigned to multiple ID classes, depending on the utilized features. Adversarial examples indicate that the model is misspecified on those inputs, thus we expect a higher epistemic uncertainty, the uncertainty about the model parameters. Furthermore, we expect higher epistemic uncertainty for misclassified samples than for correctly classified samples. Similarly, we expect the classifier to perform better on a subset of more certain samples. This is tested by evaluating the accuracy of the classifier on retained subsets of a certain fraction of samples with the lowest epistemic uncertainty [Filos et al., 2019, Band et al., 2022]. We report the AUROC for classifying the ID vs. OOD samples (i), the ID vs. the adversarial examples (ii), or the correctly classified vs. the misclassified samples (iii), using the epistemic uncertainty as score to distinguish the two classes respectively. For the selective prediction experiment (iv), we report the AUC of the accuracy vs. fraction of retained samples, using the epistemic uncertainty to determine the retained subsets. MNIST. We perform OOD detection on the FMNIST [Xiao et al., 2017], KMNIST [Clanuwat et al., 2018], EMNIST [Cohen et al., 2017] and OMNIGLOT [Lake et al., 2015] test datasets as OOD datasets, using the Le Net [Le Cun et al., 1998] architecture. The test dataset of MNIST [Le Cun et al., 1998] is used as ID dataset. We utilize the aleatoric uncertainty of the reference model (as in Eq. (2)) as a baseline to assess the added value of estimating the epistemic uncertainty of the reference model. The results are listed in Tab. 1. QUAM outperforms all other methods on this task, with Deep Ensembles being the runner up method on all dataset pairs. Furthermore, we observed, that only the epistemic uncertainties obtained by Deep Ensembles and QUAM are able to surpass the performance of using the aleatoric uncertainty of the reference model. Image Net-1K. We conduct OOD detection, adversarial example detection, misclassification detection and selective prediction experiments on Image Net-1K [Deng et al., 2009]. As OOD dataset, we use Image Net-O [Hendrycks et al., 2021], which is a challenging OOD dataset that was explicitly created to be classified as an ID dataset with high confidence by conventional Image Net-1K classifiers. Similarly, Image Net-A [Hendrycks et al., 2021] is a dataset consisting of natural adversarial exam- Table 1: MNIST results: AUROC using the epistemic uncertainty of a given, pre-selected model (as in Eq. (2)) as a score to distinguish between ID (MNIST) and OOD samples. We also report the AUROC when using the aleatoric uncertainty of the reference model (Reference). Dood Reference c SG-HMC Laplace MCD DE QUAM FMNIST .986 .005 .977 .004 .978 .004 .978 .005 .988 .001 .994 .001 KMNIST .966 .005 .957 .005 .959 .006 .956 .006 .990 .001 .994 .001 EMNIST .888 .007 .869 .012 .877 .011 .876 .008 .924 .003 .937 .008 OMNIGLOT .973 .003 .963 .004 .963 .003 .965 .003 .983 .001 .992 .001 Table 2: Image Net-1K results: AUROC using the epistemic uncertainty of a given, pre-selected model (as in Eq. (2)) to distinguish between ID (Image Net-1K) and OOD samples. Furthermore, we report the AUROC when using the epistemic uncertainty for misclassification detection and the AUC of accuracy over fraction of retained predictions on the Image Net-1K validation dataset. We also report results for all experiments, using the aleatoric uncertainty of the reference model (Reference). Dood // Task Reference c SG-HMC MCD DE (LL) DE (all) QUAM Image Net-O .626 .004 .677 .005 .680 .003 .562 .004 .709 .005 .753 .011 Image Net-A .792 .002 .799 .001 .827 .002 .686 .001 .874 .004 .872 .003 Misclassification .867 .007 .772 .011 .796 .014 .657 .009 .780 .009 .904 .008 Selective prediction .958 .003 .931 .003 .935 .006 .911 .004 .950 .002 .969 .002 ples, which belong to the ID classes of Image Net-1K, but are misclassified with high confidence by conventional Image Net-1K classifiers. Furthermore, we evaluated the utility of the uncertainty score for misclassification detection of predictions of the reference model on the Image Net-1K validation dataset. On the same dataset, we evaluated the accuracy of the reference model when only predicting on fractions of samples with the lowest epistemic uncertainty. All Image Net experiments were performed on variations of the Efficient Net architecture [Tan and Le, 2019]. Recent work by Kirichenko et al. [2022] showed that typical Image Net-1K classifiers learn desired features of the data even if they rely on simple, spurious features for their prediction. Furthermore, they found, that last layer retraining on a dataset without the spurious correlation is sufficient to re-weight the importance that the classifier places on different features. This allows the classifier to ignore the spurious features and utilize the desired features for its prediction. Similarly, we apply QUAM on the last layer of the reference model. We compare against c SG-HMC applied to the last layer, MC dropout and Deep Ensembles. MC dropout was applied to the last layer as well, since the Efficient Net architectures utilize dropout only before the last layer. Two versions of Deep Ensembles were considered. First, Deep Ensembles aggregated from pre-trained Efficient Nets of different network sizes (DE (all)). Second, Deep Ensembles of retrained last layers on the same encoder network (DE (LL)). We further utilize the aleatoric uncertainty of the reference model (as in Eq. (2)) as a baseline to assess the additional benefit of estimating the epistemic uncertainty of the reference model. The Laplace approximation was not feasible to compute on our hardware, even only for the last layer. The results are listed in Tab. 2. Plots showing the respective curves of each experiment are depicted in Fig. C.8 in the appendix. We observe that using the epistemic uncertainty provided by DE (LL) has the worst performance throughout all experiments. While DE (all) performed second best on most tasks, MC dropout outperforms it on OOD detection on the Image Net-O dataset. QUAM outperforms all other methods on all tasks we evaluated, except for Image Net-A, where it performed on par with DE (all). Details about all experiments and additional results are given in the appendix Sec. C.4. Compute Efficiency. As an ablation study, we investigate the performance of QUAM under a restricted computational budget. Therefore, the searches for adversarial models were performed on only a subset of classes instead of each eligible class, specifically the top N most probable classes according to the predictive distribution of the given, pre-selected model. The computational budget between QUAM and MC dropout was matched by accounting for the number of forward pass equivalents required by each method. In this context, we assume that the backward pass corresponds to the computational cost of two forward passes. The results depicted in Fig. 4 show that QUAM Figure 4: Inference speed vs. performance. MCD and QUAM evaluated on equal computational budget in terms of forward pass equivalents on Image Net-O (left) and Image Net-A (right) tasks. outperforms MC dropout even under a very limited computational budget. Furthermore, training a single additional ensemble member for Deep Ensembles requires more compute than evaluating the entire Image Net-O and Image Net-A datasets with QUAM when performed on all 1000 classes. 5 Related Work Quantifying predictive uncertainty, especially for deep learning models, is an active area of research. Classical uncertainty quantification methods such as Bayesian Neural Networks (BNNs) [Mac Kay, 1992, Neal, 1996] are challenging for deep learning, since (i) the Hessian or maximum-a-posterior (MAP) is difficult to estimate and (ii) regularization and normalization techniques cannot be treated [Antoráan et al., 2022]. Epistemic neural networks [Osband et al., 2021] add a variance term (the epinet) to the output only. Bayes By Backprop [Blundell et al., 2015] and variational neural networks [Oleksiienko et al., 2022] work only for small models as they require considerably more parameters. MC dropout [Gal and Ghahramani, 2016] casts applying dropout during inference as sampling from an approximate distribution. MC dropout was generalized to MC dropconnect [Mobiny et al., 2021]. Deep Ensembles [Lakshminarayanan et al., 2017] are often the best-performing uncertainty quantification method [Ovadia et al., 2019, Wursthorn et al., 2022]. Masksembles or Dropout Ensembles combine ensembling with MC dropout [Durasov et al., 2021]. Stochastic Weight Averaging approximates the posterior over the weights [Maddox et al., 2019]. Single forward pass methods are fast and they aim to capture a different notion of epistemic uncertainty through the distribution or distances of latent representations [Bradshaw et al., 2017, Liu et al., 2020, Mukhoti et al., 2021, van Amersfoort et al., 2021, Postels et al., 2021] rather than through posterior integrals. For further methods and a general overview of uncertainty estimation see e.g. Hüllermeier and Waegeman [2021], Abdar et al. [2021] and Gawlikowski et al. [2021]. 6 Conclusion We have introduced QUAM, a novel method that quantifies predictive uncertainty using adversarial models. Adversarial models identify important posterior modes that are missed by previous uncertainty quantification methods. We conducted various experiments on deep neural networks, for which epistemic uncertainty is challenging to estimate. On a synthetic dataset, we highlighted the strength of our method to capture epistemic uncertainty. Furthermore, we conducted experiments on large-scale benchmarks in the vision domain, where QUAM outperformed all previous methods. Searching for adversarial models is computationally expensive and has to be done for each new test point. However, more efficient versions can be utilized. One can search for adversarial models while restricting the search to a subset of the parameters, e.g. to the last layer as was done for the Image Net experiments, to the normalization parameters, or to the bias weights. Furthermore, there have been several advances for efficient fine-tuning of large models [Houlsby et al., 2019, Hu et al., 2021]. Utilizing those for more efficient versions of our algorithm is an interesting direction for future work. Nevertheless, high stake applications justify this effort to obtain the best estimate of predictive uncertainty for each new test point. Furthermore, QUAM is applicable to quantify the predictive uncertainty of any single given model, regardless of whether uncertainty estimation was considered during the modeling process. This allows to assess the predictive uncertainty of foundation models or specialized models that are obtained externally. Acknowledgements We would like to thank Angela Bitto-Nemling for providing relevant literature, organizing meetings, and giving feedback on this research project. Furthermore, we would like to thank Angela Bitto Nemling, Daniel Klotz, and Sebastian Lehner for insightful discussions and provoking questions. The ELLIS Unit Linz, the LIT AI Lab, the Institute for Machine Learning, are supported by the Federal State Upper Austria. We thank the projects AI-MOTION (LIT-2018-6-YOU-212), Deep Flood (LIT2019-8-YOU-213), Medical Cognitive Computing Center (MC3), INCONTROL-RL (FFG-881064), PRIMAL (FFG-873979), S3AI (FFG-872172), DL for Granular Flow (FFG-871302), EPILEPSIA (FFG-892171), AIRI FG 9-N (FWF-36284, FWF-36235), AI4Green Heating Grids(FFG899943), INTEGRATE (FFG-892418), ELISE (H2020-ICT-2019-3 ID: 951847), Stars4Waters (HORIZONCL6-2021-CLIMATE-01-01). We thank Audi.JKU Deep Learning Center, TGW LOGISTICS GROUP GMBH, Silicon Austria Labs (SAL), FILL Gesellschaft mb H, Anyline Gmb H, Google, ZF Friedrichshafen AG, Robert Bosch Gmb H, UCB Biopharma SRL, Merck Healthcare KGa A, Verbund AG, GLS (Univ. Waterloo) Software Competence Center Hagenberg Gmb H, TÜV Austria, Frauscher Sensonic, TRUMPF and the NVIDIA Corporation. M. Abdar, F. Pourpanah, S. Hussain, D. Rezazadegan, L. Liu, M. Ghavamzadeh, P. Fieguth, X. Cao, A. Khosravi, U. R. Acharya, V. Makarenkov, and S. Nahavandi. A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Information Fusion, 76:243 297, 2021. A. Adler, R. Youmaran, and W. R. B. Lionheart. A measure of the information content of EIT data. Physiological Measurement, 29(6):S101 S109, 2008. Ö. D. Akyildiz and J. Míguez. Convergence rates for optimised adaptive importance samplers. Statistics and Computing, 31(12), 2021. F. D'Angelo and V. Fortuin. Repulsive deep ensembles are bayesian. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan, editors, Advances in Neural Information Processing Systems, volume 34, pages 3451 3465. Curran Associates, Inc., 2021. J. Antoráan, D. Janz, J. U. Allingham, E. Daxberger, R. Barbano, E. Nalisnick, and J. M. Hernández Lobato. Adapting the linearised Laplace model evidence for modern deep learning. Ar Xiv, 2206.08900, 2022. G. Apostolakis. The concept of probability if safety assessments of technological systems. Science, 250(4986):1359 1364, 1991. N. Band, T. G. J. Rudner, Q. Feng, A. Filos, Z. Nado, M. W. Dusenberry, G. Jerfel, D. Tran, and Y. Gal. Benchmarking bayesian deep learning on diabetic retinopathy detection tasks. Ar Xiv, 2211.12717, 2022. B. Biggio, I. Corona, D. Maiorca, B. Nelson, N. Šrndi c, P. Laskov, G. Giacinto, and F. Roli. Evasion attacks against machine learning at test time. In H. Blockeel, K. Kersting, S. Nijssen, and F. Železný, editors, Machine Learning and Knowledge Discovery in Databases, pages 387 402, Berlin, Heidelberg, 2013. Springer Berlin Heidelberg. C. Blundell, J. Cornebise, K. Kavukcuoglu, and D. Wierstra. Weight uncertainty in neural network. In F. Bach and D. Blei, editors, Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pages 1613 1622, 2015. J. Bradshaw, A. G. de G. Matthews, and Z. Ghahramani. Adversarial examples, uncertainty, and transfer testing robustness in gaussian process hybrid deep networks. Ar Xiv, 1707.02476, 2017. J. Caldeira and B. Nord. Deeply uncertain: comparing methods of uncertainty quantification in deep learning algorithms. Machine Learning: Science and Technology, 2(1):015002, 2020. O. Cappé, A. Guillin, J.-M. Marin, and C. P. Robert. Population Monte Carlo. Journal of Computational and Graphical Statistics, 13(4):907 929, 2004. T. Chen, E. Fox, and C. Guestrin. Stochastic gradient Hamiltonian Monte Carlo. In International Conference on Machine Learning, pages 1683 1691. Proceedings of Machine Learning Research, 2014. T. Clanuwat, M. Bober-Irizar, A. Kitamoto, A. Lamb, K. Yamamoto, and D. Ha. Deep learning for classical japanese literature. Ar Xiv, 1812.01718, 2018. A. D. Cobb and B. Jalaian. Scaling hamiltonian monte carlo inference for bayesian neural networks with symmetric splitting. Uncertainty in Artificial Intelligence, 2021. G. Cohen, S. Afshar, J. Tapson, and A. Van Schaik. Emnist: Extending mnist to handwritten letters. In 2017 international joint conference on neural networks. IEEE, 2017. T. M. Cover and J. A. Thomas. Elements of Information Theory. Wiley Series in Telecommunications and Signal Processing. Wiley-Interscience, 2nd edition, 2006. ISBN 0471241954. E. Daxberger, A. Kristiadi, A. Immer, R. Eschenhagen, M. Bauer, and P. Hennig. Laplace reduxeffortless bayesian deep learning. Advances in Neural Information Processing Systems, 34: 20089 20103, 2021. J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition. IEEE, 2009. S. Depeweg, J. M. Hernández-Lobato, F. Doshi-Velez, and S. Udluft. Decomposition of uncertainty in Bayesian deep learning for efficient and risk-sensitive learning. In Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pages 1192 1201, 2018. N. Durasov, T. Bagautdinov, P. Baque, and P. Fua. Masksembles for uncertainty estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13539 13548, 2021. V. Elvira, L. Martino, D. Luengo, and M. F. Bugallo. Efficient multiple importance sampling estimators. IEEE Signal Processing Letters, 22(10):1757 1761, 2015. V. Elvira, L. Martino, D. Luengo, and M. F. Bugallo. Generalized multiple importance sampling. Statistical Science, 34(1), 2019. R. Eschenhagen, E. Daxberger, P. Hennig, and A. Kristiadi. Mixtures of laplace approximations for improved post-hoc uncertainty in deep learning. ar Xiv, 2111.03577, 2021. A. V. Fiacco and G. P. Mc Cormick. Nonlinear programming: sequential unconstrained minimization techniques. Society for Industrial and Applied Mathematics, 1990. A. Filos, S. Farquhar, A. N. Gomez, T. G. J. Rudner, Z. Kenton, L. Smith, M. Alizadeh, A. De Kroon, and Y. Gal. A systematic comparison of bayesian deep learning robustness in diabetic retinopathy tasks. Ar Xiv, 1912.10481, 2019. S. Fort, H. Hu, and B. Lakshminarayanan. Deep ensembles: A loss landscape perspective. Ar Xiv, 1912.02757, 2019. Y. Gal. Uncertainty in Deep Learning. Ph D thesis, Department of Engineering, University of Cambridge, 2016. Y. Gal and Z. Ghahramani. Dropout as a Bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the 33nd International Conference on Machine Learning, 2016. J. Gawlikowski, C. R. N. Tassi, M. Ali, J. Lee, M. Humt, J. Feng, A. Kruspe, R. Triebel, P. Jung, R. Roscher, M. Shahzad, W. Yang, R. Bamler, and X. X. Zhu. A survey of uncertainty in deep neural networks. Ar Xiv, 2107.03342, 2021. I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. In Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 27. Curran Associates, Inc., 2014. I. J. Goodfellow, J. Shlens, and C. Szegedy. Explaining and harnessing adversarial examples. ar Xiv, 1412.6572, 2015. A. Graves. Practical variational inference for neural networks. In J. Shawe-Taylor, R. Zemel, P. Bartlett, F. Pereira, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 24. Curran Associates, Inc., 2011. C. Guo, G. Pleiss, Y. Sun, and K. Q. Weinberger. On calibration of modern neural networks. In D. Precup and Y. W. Teh, editors, Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 1321 1330, 2017. F. K. Gustafsson, M. Danelljan, and T. B. Schön. Evaluating scalable bayesian deep learning methods for robust computer vision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2020. J. Hale. A probabilistic earley parser as a psycholinguistic model. In Proceedings of the Annual Conference of the North American Chapter of the Association for Computational Linguistics, pages 1 8. Association for Computational Linguistics, 2001. J. C. Helton. Risk, uncertainty in risk, and the EPA release limits for radioactive waste disposal. Nuclear Technology, 101(1):18 39, 1993. J. C. Helton. Uncertainty and sensitivity analysis in the presence of stochastic and subjective uncertainty. Journal of Statistical Computation and Simulation, 57(1-4):3 76, 1997. D. Hendrycks, K. Zhao, S. Basart, J. Steinhardt, and D. Song. Natural adversarial examples. IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021. T. Hesterberg. Weighted average importance sampling and defensive mixture distributions. Technometrics, 37(2):185 194, 1995. T. Hesterberg. Estimates and confidence intervals for importance sampling sensitivity analysis. Mathematical and Computer Modelling, 23(8):79 85, 1996. N. Houlsby, F. Huszar, Z. Ghahramani, and M. Lengyel. Bayesian active learning for classification and preference learning. Ar Xiv, 1112.5745, 2011. N. Houlsby, A. Giurgiu, S. Jastrzebski, B. Morrone, Q. De Laroussilhe, A. Gesmundo, M. Attariyan, and S. Gelly. Parameter-efficient transfer learning for nlp. In International Conference on Machine Learning. Proceedings of Machine Learning Research, 2019. E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, L. Wang, and W. Chen. Lora: Low-rank adaptation of large language models. Ar Xiv, 2106.09685, 2021. G. Huang, Y. Li, G. Pleiss, Z. Liu, J. E. Hopcroft, and K. Q. Weinberger. Snapshot ensembles: Train 1, get m for free. Ar Xiv, 1704.00109, 2017. E. Hüllermeier and W. Waegeman. Aleatoric and epistemic uncertainty in machine learning: an introduction to concepts and methods. Machine Learning, 3(110):457 506, 2021. P. Izmailov, S. Vikram, M. D. Hoffman, and A. G. Wilson. What are Bayesian neural network posteriors really like? In Proceedings of the 38th International Conference on Machine Learning, pages 4629 4640, 2021. E. T. Jaynes. Information theory and statistical mechanics. Phys. Rev., 106:620 630, 1957. S. Kapoor. torch-sgld: Sgld as pytorch optimizer. https://pypi.org/project/torch-sgld/, 2023. Accessed: 12-05-2023. W. Karush. Minima of functions of several variables with inequalities as side conditions. Master s thesis, University of Chicago, 1939. A. Kendall and Y. Gal. What uncertainties do we need in Bayesian deep learning for computer vision? In Advances in Neural Information Processing Systems, volume 30, 2017. D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. Ar Xiv, 1412.6980, 2014. D. P. Kingma and M. Welling. Auto-encoding variational Bayes. In 2nd International Conference on Learning Representations, 2014. P. Kirichenko, P. Izmailov, and A. G. Wilson. Last layer re-training is sufficient for robustness to spurious correlations. Ar Xiv, 2204.02937, 2022. H. W. Kuhn and A. W. Tucker. Nonlinear programming. In J. Neyman, editor, Second Berkeley Symposium on Mathematical Statistics and Probability, pages 481 492, Berkeley, 1950. University of California Press. O. Kviman, H. Melin, H. Koptagel, V. Elvira, and J. Lagergren. Multiple importance sampling ELBO and deep ensembles of variational approximations. In Proceedings of the 25th International Conference on Artificial Intelligence and Statistics, volume 151 of Proceedings of Machine Learning Research, pages 10687 10702, 2022. B. M. Lake, R. Salakhutdinov, and J. B. Tenenbaum. Human-level concept learning through probabilistic program induction. Science, 2015. B. Lakshminarayanan, A. Pritzel, and C. Blundell. Simple and scalable predictive uncertainty estimation using deep ensembles. In Proceedings of the 31st International Conference on Neural Information Processing Systems, page 6405 6416. Curran Associates Inc., 2017. Y. Le Cun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. C. Li, C. Chen, D. Carlson, and L. Carin. Preconditioned stochastic gradient Langevin dynamics for deep neural networks. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI), page 1788 1794. AAAI Press, 2016. J. Liu, Z. Lin, S. Padhy, D. Tran, T. Bedrax Weiss, and B. Lakshminarayanan. Simple and principled uncertainty estimation with deterministic deep learning via distance awareness. Advances in Neural Information Processing Systems, 33:7498 7512, 2020. E. S. Lubana, E. J. Bigelow, R. P. Dick, D. Krueger, and H. Tanaka. Mechanistic mode connectivity. Ar Xiv, 2211.08422, 2022. D. G. Luenberger and Y. Ye. Linear and nonlinear programming. International Series in Operations Research and Management Science. Springer, 2016. D. J. C. Mac Kay. A practical Bayesian framework for backprop networks. Neural Computation, 4: 448 472, 1992. W. J. Maddox, P. Izmailov, T. Garipov, D. P. Vetrov, and A. G. Wilson. A simple baseline for Bayesian uncertainty in deep learning. In Advances in Neural Information Processing Systems, 2019. A. Malinin and M. Gales. Predictive uncertainty estimation via prior networks. Advances in neural information processing systems, 31, 2018. R. May. A simple proof of the Karush-Kuhn-Tucker theorem with finite number of equality and inequality constraints. Ar Xiv, 2007.12483, 2020. T. E. Mc Kone. Uncertainty and variability in human exposures to soil contaminants through homegrown food: A monte carlo assessment. Risk Analysis, 14, 1994. A. Mobiny, P. Yuan, S. K. Moulik, N. Garg, C. C. Wu, and H. Van Nguyen. Drop Connect is effective in modeling uncertainty of Bayesian deep networks. Scientific Reports, 11:5458, 2021. J. Mukhoti, A. Kirsch, J. van Amersfoort, P. H. S. Torr, and Y. Gal. Deep deterministic uncertainty: A simple baseline. Ar Xiv, 2102.11582, 2021. R. Neal. Bayesian Learning for Neural Networks. Springer Verlag, New York, 1996. I. Oleksiienko, D. T. Tran, and A. Iosifidis. Variational neural networks. Ar Xiv, 2207.01524, 2022. I. Osband, Z. Wen, S. M. Asghari, V. Dwaracherla, M. Ibrahimi, X. Lu, and B. Van Roy. Weight uncertainty in neural networks. Ar Xiv, 2107.08924, 2021. Y. Ovadia, E. Fertig, J. Ren, Z. Nado, D. Sculley, S. Nowozin, J. Dillon, B. Lakshminarayanan, and J. Snoek. Can you trust your model s uncertainty? evaluating predictive uncertainty under dataset shift. In Advances in Neural Information Processing Systems, volume 32, 2019. A. Owen and Y. Zhou. Safe and effective importance sampling. Journal of the American Statistical Association, 95(449):135 143, 2000. J. Parker-Holder, L. Metz, C. Resnick, H. Hu, A. Lerer, A. Letcher, A. Peysakhovich, A. Pacchiano, and J. Foerster. Ridge rider: Finding diverse solutions by following eigenvectors of the hessian. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 753 765. Curran Associates, Inc., 2020. A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems, 32, 2019. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, et al. Scikit-learn: Machine learning in python. The Journal of Machine Learning Research, 12:2825 2830, 2011. J. Postels, M. Segu, T. Sun, L. Van Gool, F. Yu, and F. Tombari. On the practicality of deterministic epistemic uncertainty. Ar Xiv, 2107.00649, 2021. A. E. Raftery and L. Bao. Estimating and projecting trends in HIV/AIDS generalized epidemics using incremental mixture importance sampling. Biometrics, 66, 2010. M. Rigter, B. Lacerda, and N. Hawes. Rambo-rl: Robust adversarial model-based offline reinforcement learning. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, editors, Advances in Neural Information Processing Systems, volume 35, pages 16082 16097. Curran Associates, Inc., 2022. L. Scimeca, S. J. Oh, S. Chun, M. Poli, and S. Yun. Which shortcut cues will dnns choose? a study from the parameter-space perspective. ar Xiv, 2110.03095, 2022. T. Seidenfeld. Entropy and uncertainty. Philosophy of Science, 53(4):467 491, 1986. H. Shah, K. Tamuly, A. Raghunathan, P. Jain, and P. Netrapalli. The pitfalls of simplicity bias in neural networks. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors, Advances in Neural Information Processing Systems, volume 33, pages 9573 9585. Curran Associates, Inc., 2020. C. E. Shannon and C. Elwood. A mathematical theory of communication. The Bell System Technical Journal, 27:379 423, 1948. L. Smith and Y. Gal. Understanding measures of uncertainty for adversarial example detection. In A. Globerson and R. Silva, editors, Proceedings of the Thirty-Fourth Conference on Uncertainty in Artificial Intelligence, pages 560 569. AUAI Press, 2018. R. J. Steele, A. E. Raftery, and M. J. Emond. Computing normalizing constants for finite mixture models via incremental mixture importance sampling (IMIS). Journal of Computational and Graphical Statistics, 15, 2006. M. Sundararajan, A. Taly, and Q. Yan. Axiomatic attribution for deep networks. In International Conference on Machine Learning, 2017. C. Szegedy, W. Zaremba, I. Sutskever, J. Bruna, D. Erhan, I. J. Goodfellow, and R. Fergus. Intriguing properties of neural networks. Ar Xiv, 1312.6199, 2013. M. Tan and Q. Le. Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning, pages 6105 6114. Proceedings of Machine Learning Research, 2019. M. Tribus. Thermostatics and Thermodynamics: An Introduction to Energy, Information and States of Matter, with Engineering Applications. University series in basic engineering. Van Nostrand, 1961. J. van Amersfoort, L. Smith, Y. W. Teh, and Y. Gal. Uncertainty estimation using a single deep deterministic neural network. In International Conference on Machine Learning, pages 9690 9700. Proceedings of Machine Learning Research, 2020. J. van Amersfoort, L. Smith, A. Jesson, O. Key, and Y. Gal. On feature collapse and deep kernel learning for single forward pass uncertainty. Ar Xiv, 2102.11409, 2021. E. Veach and L. J. Guibas. Optimally combining sampling techniques for Monte Carlo rendering. In Proceedings of the 22nd Annual Conference on Computer Graphics and Interactive Techniques, pages 419 428. Association for Computing Machinery, 1995. W. E. Vesely and D. M. Rasmuson. Uncertainties in nuclear probabilistic risk analyses. Risk Analysis, 4, 1984. S. Weinzierl. Introduction to Monte Carlo methods. Ar Xiv, hep-ph/0006269, 2000. M. Welling and Y. W. Teh. Bayesian learning via stochastic gradient Langevin dynamics. In Proceedings of the 28th International Conference on Machine Learning, page 681 688, Madison, WI, USA, 2011. Omnipress. A. G. Wilson and P. Izmailov. Bayesian deep learning and a probabilistic perspective of generalization. Advances in Neural Information Processing Systems, 33:4697 4708, 2020. K. Wursthorn, M. Hillemann, and M. M. Ulrich. Comparison of uncertainty quantification methods for CNN-based regression. The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, XLIII-B2-2022:721 728, 2022. H. Xiao, K. Rasul, and R. Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. Ar Xiv, 1708.07747, 2017. W. I. Zangwill. Non-linear programming via penalty functions. Management Science, 13(5):344 358, 1967. R. Zhang, C. Li, J. Zhang, C. Chen, and A. G. Wilson. Cyclical stochastic gradient MCMC for Bayesian deep learning. In International Conference on Learning Representations, 2020. R. Zhang, A. G. Wilson, and C. De Sa. Low-precision stochastic gradient Langevin dynamics. Ar Xiv, 2206.09909, 2022. J. V. Zidek and C. van Eeden. Uncertainty, entropy, variance and the effect of partial information. Lecture Notes-Monograph Series, 42:155 167, 2003. This is the appendix of the paper Quantification of Uncertainty with Adversarial Models . It consists of three sections. In view of the increasing influence of contemporary machine learning research on the broader public, section A gives a societal impact statement. Following to this, section B gives details of our theoretical results, foremost about the measure of uncertainty used throughout our work. Furthermore, Mixture Importance Sampling for variance reduction is discussed. Finally, section C gives details about the experiments presented in the main paper, as well as further experiments. Contents of the Appendix A Societal Impact Statement 18 B Theoretical Results 19 B.1 Measuring Predictive Uncertainty . . . . . . . . . . . . . . . . . . . . . . . . . . 19 B.1.1 Entropy and Cross-Entropy as Measures of Predictive Uncertainty . . . . . 19 B.1.2 Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 B.1.3 Regression . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 B.2 Mixture Importance Sampling for Variance Reduction . . . . . . . . . . . . . . . . 23 C Experimental Details and Further Experiments 27 C.1 Details on the Adversarial Model Search . . . . . . . . . . . . . . . . . . . . . . . 27 C.2 Simplex Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 C.3 Epistemic Uncertainty on Synthetic Dataset . . . . . . . . . . . . . . . . . . . . . 29 C.4 Epistemic Uncertainty on Vision Datasets . . . . . . . . . . . . . . . . . . . . . . 29 C.4.1 MNIST . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 C.4.2 Image Net . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 C.5 Comparing Mechanistic Similarity of Deep Ensembles vs. Adversarial Models . . 38 C.6 Prediction Space Similarity of Deep Ensembles and Adversarial Models . . . . . . 38 C.7 Computational Expenses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 List of Figures B.1 Asymptotic variance for multimodal target and unimodal sampling distribution . . 26 C.1 Illustrative example of QUAM . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 C.2 HMC and Adversarial Model Search on simplex . . . . . . . . . . . . . . . . . . . 28 C.3 Epistemic uncertainty (setting (b)) on synthetic classification dataset . . . . . . . . 30 C.4 Model variance of different methods on toy regression dataset . . . . . . . . . . . 30 C.5 Histograms MNIST . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 C.6 Calibration on Image Net . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 C.7 Histograms Image Net . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 C.8 Detailed results of Image Net experiments . . . . . . . . . . . . . . . . . . . . . . 37 C.9 Comparing mechanistic similarity of Deep Ensembles vs. Adversarial Models . . . 39 C.10 Prediction Space Similarity of Deep Ensembles vs. Adversarial Models . . . . . . 39 List of Tables C.1 Results for additional baseline (Mo LA) . . . . . . . . . . . . . . . . . . . . . . . 31 C.2 Detailed results of MNIST OOD detection experiments . . . . . . . . . . . . . . . 32 C.3 Results for calibration on Image Net . . . . . . . . . . . . . . . . . . . . . . . . . 34 C.4 Detailed results of Image Net experiments . . . . . . . . . . . . . . . . . . . . . . 35 A Societal Impact Statement In this work, we have focused on improving the predictive uncertainty estimation for machine learning models, specifically deep learning models. Our primary goal is to enhance the robustness and reliability of these predictions, which we believe have several positive societal impacts. 1. Improved decision-making: By providing more accurate predictive uncertainty estimates, we enable a broad range of stakeholders to make more informed decisions. This could have implications across various sectors, including healthcare, finance, and autonomous vehicles, where decision-making based on machine learning predictions can directly affect human lives and economic stability. 2. Increased trust in machine learning systems: By enhancing the reliability of machine learning models, our work may also contribute to increased public trust in these systems. This could foster greater acceptance and integration of machine learning technologies in everyday life, driving societal advancement. 3. Promotion of responsible machine learning: Accurate uncertainty estimation is crucial for the responsible deployment of machine learning systems. By advancing this area, our work promotes the use of those methods in an ethical, transparent, and accountable manner. While we anticipate predominantly positive impacts, it is important to acknowledge potential negative impacts or challenges. 1. Misinterpretation of uncertainty: Even with improved uncertainty estimates, there is a risk that these might be misinterpreted or misused, potentially leading to incorrect decisions or unintended consequences. It is vital to couple advancements in this field with improved education and awareness around the interpretation of uncertainty in AI systems. 2. Increased reliance on machine learning systems: While increased trust in machine learning systems is beneficial, there is a risk it could lead to over-reliance on these systems, potentially resulting in reduced human oversight or critical thinking. It s important that robustness and reliability improvements don t result in blind trust. 3. Inequitable distribution of benefits: As with any technological advancement, there is a risk that the benefits might not be evenly distributed, potentially exacerbating existing societal inequalities. We urge policymakers and practitioners to consider this when implementing our findings. In conclusion, while our work aims to make significant positive contributions to society, we believe it is essential to consider these potential negative impacts and take steps to mitigate them proactively. B Theoretical Results B.1 Measuring Predictive Uncertainty In this section, we first discuss the usage of the entropy and the cross-entropy as measures of predictive uncertainty. Following this, we introduce the two settings (a) and (b) (see Sec. 2) in detail for the predictive distributions of probabilistic models in classification and regression. Finally, we discuss Mixture Importance Sampling for variance reduction of the uncertainty estimator. B.1.1 Entropy and Cross-Entropy as Measures of Predictive Uncertainty Shannon and Elwood [1948] defines the entropy H[p] = PN i=1 pi log pi as a measure of the amount of uncertainty of a discrete probability distribution p = (p1, . . . , p N) and states that it measures how much choice is involved in the selection of a class i. See also Jaynes [1957], Cover and Thomas [2006] for an elaboration on this topic. The value log pi has been called "surprisal" [Tribus, 1961] (page 64, Subsection 2.9.1) and has been used in computational linguistics [Hale, 2001]. Hence, the entropy is the expected or mean surprisal. Instead of surprisal also the terms information content , self-information , or Shannon information are used. The cross-entropy CE[p , q] = PN i=1 pi log qi between two discrete probability distributions p = (p1, . . . , p N) and q = (q1, . . . , q N) measures the expectation of the surprisal of q under distribution p. Like the entropy, the cross-entropy is a mean of surprisals, therefore can be considered as a measure to quantify uncertainty. The higher surprisals are on average, the higher the uncertainty. The cross-entropy has increased uncertainty compared to the entropy since more surprising events are expected when selecting events via p instead of q. Only if those distributions coincide, there is no additional surprisal and the cross-entropy is equal to the entropy of the distributions. The cross-entropy depends on the uncertainty of the two distributions and how different they are. In particular, high surprisal of qi and low surprisal of pi strongly increase the cross-entropy since unexpected events are more frequent, that is, we are more often surprised. Thus, the cross-entropy does not only measure the uncertainty under distribution p, but also the difference of the distributions. The average surprisal via the cross-entropy depends on the uncertainty of p and the difference between p and q: CE[p , q] = i=1 pi log qi (9) i=1 pi log pi + i=1 pi log pi = H[p] + DKL(p q) , where the Kullback-Leibler divergence DKL( ) is i=1 pi log pi The Kullback-Leibler divergence measures the difference in the distributions via their average difference of surprisals. Furthermore, it measures the decrease in uncertainty when shifting from the estimate p to the true q [Seidenfeld, 1986, Adler et al., 2008]. Therefore, the cross-entropy can serve to measure the total uncertainty, where the entropy is used as aleatoric uncertainty and the difference of distributions is used as the epistemic uncertainty. We assume that q is the true distribution that is estimated by the distribution p. We quantify the total uncertainty of p as the sum of the entropy of p (aleatoric uncertainty) and the Kullback-Leibler divergence to q (epistemic uncertainty). In accordance with Apostolakis [1991] and Helton [1997], the aleatoric uncertainty measures the stochasticity of sampling from p, while the epistemic uncertainty measures the deviation of the parameters p from the true parameters q. In the context of quantifying uncertainty through probability distributions, other measures such as the variance have been proposed [Zidek and van Eeden, 2003]. For uncertainty estimation in the context of deep learning systems, e.g. Gal [2016], Kendall and Gal [2017], Depeweg et al. [2018] proposed to use the variance of the BMA predictive distribution as a measure of uncertainty. Entropy and variance capture different notions of uncertainty and investigating measures based on the variance of the predictive distribution is an interesting avenue for future work. B.1.2 Classification Setting (a): Expected uncertainty when selecting a model. We assume to have training data D and an input x. We want to know the uncertainty in predicting a class y from x when we first choose a model w based on the posterior p( w | D) an then use the chosen model w to choose a class for input x according to the predictive distribution p(y | x, w). The uncertainty in predicting the class arises from choosing a model (epistemic) and from choosing a class using this probabilistic model (aleatoric). Through Bayesian model averaging, we obtain the following probability of selecting a class: p(y | x, D) = Z W p(y | x, w) p( w | D) d w . (11) The total uncertainty is commonly measured as the entropy of this probability distribution [Houlsby et al., 2011, Gal, 2016, Depeweg et al., 2018, Hüllermeier and Waegeman, 2021]: H[p(y | x, D)] . (12) We can reformulate the total uncertainty as the expected cross-entropy: H[p(y | x, D)] = X y Y p(y | x, D) log p(y | x, D) (13) y Y log p(y | x, D) Z W p(y | x, w) p( w | D) d w y Y p(y | x, w) log p(y | x, D) p( w | D) d w W CE[p(y | x, w) , p(y | x, D)] p( w | D) d w . We can split the total uncertainty into the aleatoric and epistemic uncertainty [Houlsby et al., 2011, Gal, 2016, Smith and Gal, 2018]: Z W CE[p(y | x, w) , p(y | x, D)] p( w | D) d w (14) W (H[p(y | x, w)] + DKL(p(y | x, w) p(y | x, D))) p( w | D) d w W H[p(y | x, w)] p( w | D) d w + Z W DKL(p(y | x, w) p(y | x, D)) p( w | D) d w W H[p(y | x, w)] p( w | D) d w + I[Y ; W | x, D] . We verify the last equality in Eq. (14), i.e. that the Mutual Information is equal to the expected Kullback-Leibler divergence: I[Y ; W | x, D] = Z y Y p(y, w | x, D) log p(y, w | x, D) p(y | x, D) p( w | D) d w (15) y Y p(y | x, w) p( w | D) log p(y | x, w) p( w | D) p(y | x, D) p( w | D) d w y Y p(y | x, w) log p(y | x, w) p(y | x, D) p( w | D) d w W DKL(p(y | x, w) p(y | x, D)) p( w | D) d w . This is possible because the label is dependent on the selected model. First, a model is selected, then a label is chosen with the selected model. To summarize, the predictive uncertainty is measured by: H[p(y | x, D)] = Z W H[p(y | x, w)] p( w | D) d w + I[Y ; W | x, D] (16) W H[p(y | x, w)] p( w | D) d w W DKL(p(y | x, w) p(y | x, D)) p( w | D) d w W CE[p(y | x, w) , p(y | x, D)] p( w | D) d w . The total uncertainty is given by the entropy of the Bayesian model average predictive distribution, which we showed is equal to the expected cross-entropy between the predictive distributions of candidate models w selected according to the posterior and the Bayesian model average predictive distribution. The aleatoric uncertainty is the expected entropy of candidate models drawn from the posterior, which can also be interpreted as the entropy we expect when selecting a model according to the posterior. Therefore, if all models likely under the posterior have low surprisal, the aleatoric uncertainty in this setting is low. The epistemic uncertainty is the expected KL divergence between the the predictive distributions of candidate models and the Bayesian model average predictive distribution. Therefore, if all models likely under the posterior have low divergence of their predictive distribution to the Bayesian model average predictive distribution, the epistemic uncertainty in this setting is low. Setting (b): Uncertainty of a given, pre-selected model. We assume to have training data D, an input x, and a given, pre-selected model with parameters w and predictive distribution p(y | x, w). Using the predictive distribution of the model, a class y is selected based on x, therefore there is uncertainty about which y is selected. Furthermore, we assume that the true model with predictive distribution p(y | x, w ) and parameters w has generated the training data D and will also generate the observed (real world) y from x that we want to predict. The true model is only revealed later, e.g. via more samples or by receiving knowledge about w . Hence, there is uncertainty about the parameters of the true model. Revealing the true model is viewed as drawing a true model from all possible true models according to their agreement with D. Note, to reveal the true model is not necessary in our framework but helpful for the intuition of drawing a true model. We neither consider uncertainty about the model class nor the modeling nor about the training data. In summary, there is uncertainty about drawing a class from the predictive distribution of the given, pre-selected model and uncertainty about drawing the true parameters of the model distribution. According to Apostolakis [1991] and Helton [1997], the aleatoric uncertainty is the variability of selecting a class y via p(y | x, w). Using the entropy, the aleatoric uncertainty is H[p(y | x, w)] . (17) Also according to Apostolakis [1991] and Helton [1997], the epistemic uncertainty is the uncertainty about the parameters w of the distribution, that is, a difference measure between w and the true parameters w . We use as a measure for the epistemic uncertainty the Kullback-Leibler divergence: DKL(p(y | x, w) p(y | x, w )) . (18) The total uncertainty is the aleatoric uncertainty plus the epistemic uncertainty, which is the crossentropy between p(y | x, w) and p(y | x, w ): CE[p(y | x, w) , p(y | x, w )] = H[p(y | x, w)] + DKL(p(y | x, w) p(y | x, w )) . (19) However, we do not know the true parameters w . The posterior p( w | D) gives us the likelihood of w being the true parameters w . We assume that the true model is revealed later. Therefore we use the expected Kullback-Leibler divergence for the epistemic uncertainty: Z W DKL(p(y | x, w) p(y | x, w)) p( w | D) d w . (20) Consequently, the total uncertainty is H[p(y | x, w)] + Z W DKL(p(y | x, w) p(y | x, w)) p( w | D) d w . (21) The total uncertainty can therefore be expressed by the expected cross-entropy as it was in setting (a) (see Eq. (16)), but between p(y | x, w) and p(y | x, w): W CE[p(y | x, w) , p(y | x, w)] p( w | D) d w (22) W (H[p(y | x, w)] + DKL(p(y | x, w) p(y | x, w))) p( w | D) d w = H[p(y | x, w)] + Z W DKL(p(y | x, w) p(y | x, w)) p( w | D) d w . B.1.3 Regression We follow Depeweg et al. [2018] and measure the predictive uncertainty in a regression setting using the differential entropy H[p(y | x, w)] = R Y p(y | x, w) log p(y | x, w)dy of the predictive distribution p(y | x, w) of a probabilistic model. In the following, we assume that we are modeling a Gaussian distribution, but other continuous probability distributions e.g. a Laplace lead to similar results. The model thus has to provide estimators for the mean µ(x, w) and variance σ2(x, w) of the Gaussian. The predictive distribution is given by p(y | x, w) = (2π σ2(x, w)) 1 2 exp (y µ(x, w))2 The differential entropy of a Gaussian distribution is given by H[p(y | x, w)] = Z Y p(y | x, w) log p(y | x, w) dy (24) 2 log(σ2(x, w)) + log(2π) + 1 The KL divergence between two Gaussian distributions is given by DKL(p(y | x, w) p(y | x, w)) (25) Y p(y | x, w) log p(y | x, w) p(y | x, w) 2 log σ2(x, w) + σ2(x, w) + (µ(x, w) µ(x, w))2 2 σ2(x, w) 1 Setting (a): Expected uncertainty when selecting a model. Depeweg et al. [2018] consider the differential entropy of the Bayesian model average p(y | x, D) = R W p(y | x, w)p( w | D)d w, which is equal to the expected cross-entropy and can be decomposed into the expected differential entropy and Kullback-Leibler divergence. Therefore, the expected uncertainty when selecting a model is given by W CE[p(y | x, w) , p(y | x, D)] p( w | D) d w = H[p(y | x, D)] (26) W H[p(y | x, w)] p( w | D) d w + Z W DKL(p(y | x, w) p(y | x, D)) p( w | D) d w 1 2 log(σ2(x, w)) p( w | D) d w + log(2π) W DKL(p(y | x, w) p(y | x, D)) p( w | D) d w . Setting (b): Uncertainty of a given, pre-selected model. Synonymous to the classification setting, the uncertainty of a given, pre-selected model w is given by Z W CE[p(y | x, w) , p(y | x, w)] p( w | D) d w (27) = H[p(y | x, w)] + Z W DKL(p(y | x, w) p(y | x, w)) p( w | D) d w 2 log(σ2(x, w)) + log(2π) 1 2 log σ2(x, w) + σ2(x, w) + (µ(x, w) µ(x, w))2 2 σ2(x, w) p( w | D) d w . Homoscedastic, Model Invariant Noise. We assume, that noise is homoscedastic for all inputs x X, thus σ2(x, w) = σ2(w). Furthermore, most models in regression do not explicitly model the variance in their training objective. For such a model w, we can estimate the variance on a validation dataset Dval = {(xn, yn)}|N n=1 as n=1 (yn µ(xn, w))2 . (28) If we assume that all reasonable models under the posterior will have similar variances (ˆσ2(w) σ2( w) for w p( w | D)), the uncertainty of a prediction using the given, pre-selected model w is given by Z W CE[p(y | x, w) , p(y | x, w)] p( w | D) d w (29) 2 log(ˆσ2(w)) + log(2π) 1 2 log ˆσ2(w) + ˆσ2(w) + (µ(x, w) µ(x, w))2 2 ˆσ2(w) p( w | D) d w 2 log(ˆσ2(w)) + 1 ˆσ2(w) W (µ(x, w) µ(x, w))2 p( w | D) d w + 1 2 + log(2π) . B.2 Mixture Importance Sampling for Variance Reduction The epistemic uncertainties in Eq. (1) and Eq. (2) are expectations of KL divergences over the posterior. We have to approximate these integrals. If the posterior has different modes, a concentrated importance sampling function has a high variance of estimates, therefore converges very slowly [Steele et al., 2006]. Thus, we use mixture importance sampling (MIS) [Hesterberg, 1995]. MIS uses a mixture model for sampling, instead of a unimodal model of standard importance sampling [Owen and Zhou, 2000]. Multiple importance sampling Veach and Guibas [1995] is similar to MIS and equal to it for balanced heuristics [Owen and Zhou, 2000]. More details on these and similar methods can be found in Owen and Zhou [2000], Cappé et al. [2004], Elvira et al. [2015, 2019], Steele et al. [2006], Raftery and Bao [2010]. MIS has been very successfully applied to estimate multimodal densities. For example, the evidence lower bound (ELBO) [Kingma and Welling, 2014] has been improved by multiple importance sampling ELBO [Kviman et al., 2022]. Using a mixture model should ensure that at least one of its components will locally match the shape of the integrand. Often, MIS iteratively enrich the sampling distribution by new modes [Raftery and Bao, 2010]. In contrast to iterative enrichment, which finds modes by chance, we are able to explicitly search for posterior modes, where the integrand of the definition of epistemic uncertainty is large. For each of these modes, we define a component of the mixture from which we then sample. We have the huge advantage to have explicit expressions for the integrand. The integrand of the epistemic uncertainty in Eq. (1) and Eq. (2) has the form D(p(y | x, w) , p(y | x, w)) p( w | D) , (30) where D( , ) is a distance or divergence of distributions which is computed using the parameters that determine those distributions. The distance/divergence D( , ) eliminates the aleatoric uncertainty, which is present in p(y | x, w) and p(y | x, w). Essentially, D( , ) reduces distributions to functions of their parameters. Importance sampling is applied to estimate integrals of the form X f(x) p(x) dx = Z q(x) q(x) dx , (31) with integrand f(x) and probability distributions p(x) and q(x), when it is easier to sample according to q(x) than p(x). The estimator of Eq. (31) when drawing xn according to q(x) is given by f(xn) p(xn) q(xn) . (32) The asymptotic variance σ2 s of importance sampling is given by (see e.g. Owen and Zhou [2000]): q(x) s 2 q(x) dx (33) 2 q(x) dx s2 , and its estimator when drawing xn from q(x) is given by f(xn) p(xn) q(xn) s 2 (34) f(xn) p(xn) We observe, that the variance is determined by the term f(x)p(x) q(x) , thus we want q(x) to be proportional to f(x)p(x). Most importantly, q(x) should not be close to zero for large f(x)p(x). To give an intuition about the severity of unmatched modes, we depict an educational example in Fig. B.1. Now we plug in the form of the integrand given by Eq. (30) into Eq. (31), to calculate the expected divergence D( , ) under the model posterior p( w | D). This results in D(p(y | x, w) , p(y | x, w)) p( w | D) q( w) q( w) d w , (35) with estimate D(p(y | x, w) , p(y | x, wn)) p( wn | D) q( wn) . (36) The variance is given by D(p(y | x, w) , p(y | x, w)) p( w | D) q( w) v 2 q( w) d w (37) D(p(y | x, w) , p(y | x, w)) p( w | D) 2 q( w) d w v2 . The estimate for the variance is given by D(p(y | x, w) , p(y | x, wn)) p( wn | D) q( wn) v 2 (38) D(p(y | x, w) , p(y | x, wn)) p( wn | D) where wn is drawn according to q( w). The asymptotic (N ) confidence intervals are given by lim N Pr a σv N ˆv v b σv a exp( 1/2 t2) dt . (39) Thus, ˆv converges with σv N to v. The asymptotic confidence interval is proofed in Weinzierl [2000] and Hesterberg [1996] using the Lindeberg Lévy central limit theorem which ensures the asymptotic normality of the estimate ˆv. The q( w) that minimizes the variance is q( w) = D(p(y | x, w) , p(y | x, w)) p( w | D) Thus we want to find a density q( w) that is proportional to D(p(y | x, w) , p(y | x, w)) p( w | D). Only approximating the posterior p( w | D) as Deep Ensembles or MC dropout is insufficient to guarantee a low expected error, since the sampling variance cannot be bounded, as σ2 v could get arbitrarily big if the distance is large but the probability under the sampling distribution is very small. For q( w) p( w | D) and non-negative, unbounded, but continuous D( , ), the variance σ2 v given by Eq. (37) cannot be bounded. For example, if D( , ) is the KL-divergence and both p(y | x, w) and p(y | x, w) are Gaussians where the means µ(x, w), µ(x, w) and variances σ2(x, w), σ2(x, w) are estimates provided by the models, the KL is unbounded. The KL divergence between two Gaussian distributions is given by DKL(p(y | x, w) p(y | x, w)) (41) Y p(y | x, w) log p(y | x, w) p(y | x, w) 2 log σ2(x, w) + σ2(x, w) + (µ(x, w) µ(x, w))2 2 σ2(x, w) 1 For σ2(x, w) going towards zero and a non-zero difference of the mean values, the KL-divergence can be arbitrarily large. Therefore, methods that only consider the posterior p( w | D) cannot bound the variance σ2 v if D( , ) is unbounded and the parameters w allow distributions which can make D( , ) arbitrary large. 4 2 0 2 4 6 8 x 0.2 0.4 0.6 0.8 1.0 1.2 2 p, 2 Asymptotic Variance of IS Estimate Asymptotic Variance of IS Estimate 1 2 3 4 5 2 q Asymptotic Variance of IS Estimate Figure B.1: Analysis of asymptotic variance of importance sampling for multimodal target distribution p(x) and a unimodal sampling distribution q(x). The target distribution is a mixture of two Gaussian distributions with means µp,1, µp,2 and variances σ2 p,1, σ2 p,2. The sampling distribution is a single Gaussian with mean µq and variance σ2 q. q(x) matches one of the modes of p(x), but misses the other. Both distributions are visualized for their standard parameters µp,1 = µq = 0, µp,2 = 3 and σ2 p,1 = σ2 p,2 = σ2 q = 1, where both mixture components of p(x) are equally weighted. We calculate the asymptotic variance (Eq. (33) with f(x) = 1) for different values of σ2 p,2, µp,2 and σ2 q and show the results in the top right, bottom left and bottom right plot respectively. The standard value for the varied parameter is indicated by the black dashed line. We observe, that slightly increasing the variance of the second mixture component of p(x), which is not matched by the mode of q(x), rapidly increases the asymptotic variance. Similarly, increasing the distance between the center of the unmatched mixture component of p(x) and q(x) strongly increases the asymptotic variance. On the contrary, increasing the variance of the sampling distribution q(x) does not lead to a strong increase, as the worse approximation of the matched mode of p(x) is counterbalanced by putting probability mass where the second mode of p(x) is located. Note, that this issue is even more exacerbated if f(x) is non-constant. Then, q(x) has to match the modes of f(x) as well. C Experimental Details and Further Experiments Our code is publicly available at https://github.com/ml-jku/quam. C.1 Details on the Adversarial Model Search During the adversarial model search, we seek to maximize the KL divergence between the prediction of the reference model and adversarial models. For an example, see Fig. C.1. We found that directly maximizing the KL divergence always leads to similar solutions to the optimization problem. Therefore, we maximized the likelihood of a new test point to be in each possible class. The optimization problem is very similar, considering the predictive distribution p(y | x, w) of a reference model and the predictive distribution p(y | x, w) of an adversarial model, the model that is updated. The KL divergence between those two is given by DKL(p(y | x, w) p(y | x, w)) (42) = X p(y | x, w) log p(y | x, w) p(y | x, w) = X p(y | x, w) log (p(y | x, w)) X p(y | x, w) log (p(y | x, w)) = H[p(y | x, w)] + CE[p(y | x, w) , p(y | x, w)] . Only the cross-entropy between the predictive distributions of the reference model parameterized by w and the adversarial model parameterized by w plays a role in the optimization, since the entropy of pw stays constant during the adversarial model search. Thus, the optimization target is equivalent to the cross-entropy loss, except that pw is generally not one-hot encoded but an arbitrary categorical distribution. This also relates to targeted / untargeted adversarial attacks on the input. Targeted attacks try to maximize the output probability of a specific class. Untargeted attacks try to minimize the probability of the originally predicted class, by maximizing all other classes. We found that attacking individual classes works better empirically, while directly maximizing the KL divergence always leads to similar solutions for different searches. The result often is a further increase of the probability associated with the most likely class. Therefore, we conducted as many adversarial model searches for a new test point, as there are classes in the classification task. Thereby, we optimize the cross-entropy loss for one specific class in each search. Figure C.1: Illustrative example of QUAM. We illustrate quantifying the predictive uncertainty of a given, pre-selected model (blue), a classifier for images of cats and dogs. For each of the input images, we search for adversarial models (orange) that make different predictions than the given, pre-selected model while explaining the training data equally well (having a high likelihood). The adversarial models found for an image of a dog or a cat still make similar predictions (low epistemic uncertainty), while the adversarial model found for an image of a lion makes a highly different prediction (high epistemic uncertainty), as features present in images of both cats and dogs can be utilized to classify the image of a lion. For regression, we add a small perturbation to the bias of the output linear layer. This is necessary to ensure a gradient in the first update step, as the model to optimize is initialized with the reference model. For regression, we perform the adversarial model search two times, as the output of an adversarial model could be higher or lower than the reference model if we assume a scalar output. We force, that the two adversarial model searches get higher or lower outputs than the reference model respectively. While the loss of the reference model on the training dataset Lref is calculated on the full training dataset (as it has to be done only once), we approximate Lpen by randomly drawn mini-batches for each update step. Therefore, the boundary condition might not be satisfied on the full training set, even if the boundary condition is satisfied for the mini-batch estimate. As described in the main paper, the resulting model of each adversarial model search is used to define the location of a mixture component of a sampling distribution q( w) (Eq. (6)). The epistemic uncertainty is estimated by Eq. (4), using models sampled from this mixture distribution. The simplest choice of distributions for each mixture distribution is a delta distribution at the location of the adversarial model wk. While this performs well empirically, we discard a lot of information by not utilizing predictions of models obtained throughout the adversarial model search. The intermediate solutions of the adversarial model search allow to assess how easily models with highly divergent predictive distributions to the reference model can be found. Furthermore, the expected mean squared error (Eq. (5)) decreases with 1 N with the number of samples N and the expected variance of the estimator (Eq. (38)) decreases with 1 N . Therefore, using more samples is beneficial empirically, even though we potentially introduce a bias to the estimator. Consequently, we utilize all sampled models during the adversarial model search as an empirical sampling distribution for our experiments. This is the same as how members of an ensemble can be seen as an empirical sampling distribution [Gustafsson et al., 2020] and conceptually similar to Snapshot ensembling [Huang et al., 2017]. To compute Eq. (4), we use the negative exponential training loss of each model to approximate its posterior probability p( w | D). Note that the training loss is the negative log-likelihood, which in turn is proportional to the posterior probability. Note we temperature-scale the approximate posterior probability by p( w | D) 1 T , with the temperature parameter T set as a hyperparameter. C.2 Simplex Example We sample the training dataset D = {(xk, yk)}K k=1 from three Gaussian distributions (21 datapoints from each Gaussian) at locations µ1 = ( 4, 2)T , µ2 = (4, 2)T , µ3 = (0, 2 2)T and the same two-dimensional covariance with σ2 = 1.5 on both entries of the diagonal and zero on the off-diagonals. The labels yk are one-hot encoded vectors, signifying which Gaussian the input xk was sampled from. The new test point x we evaluate for is located at ( 6, 2). To attain the likelihood (b) Adversarial Model Search (c) Training data + new test point Figure C.2: Softmax outputs (black) of individual models of HMC (a) as well as their average output (red) on a probability simplex. Softmax outputs of models found throughout the adversarial model search (b), colored by the attacked class. Left, right and top corners denote 100% probability mass at the blue, orange and green class in (c) respectively. Models were selected on the training data, and evaluated on the new test point (red) depicted in (c). The background color denotes the maximum likelihood of the training data that is achievable by a model having equal softmax output as the respective location on the simplex. for each position on the probability simplex, we train a two-layer fully connected neural network (with parameters w) with hidden size of 10 on this dataset. We minimize the combined loss k=1 l(p(y | xk, w), yk) + l(p(y | x, w), y) , (43) where l is the cross-entropy loss function and y is the desired categorical distribution for the output of the network. We report the likelihood on the training dataset upon convergence of the training procedure for y on the probability simplex. To average over different initializations of w and alleviate the influence of potentially bad local minima, we use the median over 20 independent runs to calculate the maximum. For all methods, we utilize the same two-layer fully connected neural network with hidden size of 10; for MC dropout we additionally added dropout with dropout probability 0.2 after every intermediate layer. We trained 50 networks for the Deep Ensemble results. For MC dropout we sampled predictive distributions using 1000 forward passes. Fig. C.2 (a) shows models sampled using HMC, which is widely regarded as the best approximation to the ground truth for predictive uncertainty estimation. Furthermore, Fig. C.2 (b) shows models obtained by executing the adversarial model search for the given training dataset and test point depicted in Fig. C.2 (c). HMC also provides models that put more probability mass on the orange class. Those are missed by Deep Ensembles and MC dropout (see Fig. 2 (a) and (b)). The adversarial model search used by QUAM helps to identify those regions. C.3 Epistemic Uncertainty on Synthetic Dataset We create the two-moons dataset using the implementation of Pedregosa et al. [2011]. All experiments were performed on a three-layer fully connected neural network with hidden size 100 and Re LU activations. For MC dropout, dropout with dropout probability of 0.2 was applied after the intermediate layers. We assume to have a trained reference model w of this architecture. Results of the same runs as in the main paper, but calculated for the epistemic uncertainty in setting (b) (see Eq. (2)) are depicted in Fig. C.3. Again, QUAM matches the ground truth best. Furthermore, we conducted experiments on a synthetic regression dataset, where the input feature x is drawn randomly between [ π, π] and the target is y = sin(x) + ϵ, with ϵ N(0, 0.1). The results are depicted in Fig. C.4. As for the classification results, the estimate of QUAM is closest to the ground truth provided by HMC. The HMC implementation of Cobb and Jalaian [2021] was used to obtain the ground truth epistemic uncertainties. For the Laplace approximation, we used the implementation of Daxberger et al. [2021]. For SG-MCMC we used the python package of Kapoor [2023]. C.4 Epistemic Uncertainty on Vision Datasets Several vision datasets and their corresponding OOD datasets are commonly used for benchmarking predictive uncertainty quantification in the literature, e.g. in Blundell et al. [2015], Gal and Ghahramani [2016], Malinin and Gales [2018], Ovadia et al. [2019], van Amersfoort et al. [2020], Mukhoti et al. [2021], Postels et al. [2021], Band et al. [2022]. Our experiments focused on two of those: MNIST [Le Cun et al., 1998] and its OOD derivatives as the most basic benchmark and Image Net1K [Deng et al., 2009] to demonstrate our method s ability to perform on a larger scale. Four types of experiments were performed: (i) OOD detection (ii) adversarial example detection, (iii) misclassification detection and (iv) selective prediction. Our experiments on adversarial example detection did not utilize a specific adversarial attack on the input images, but natural adversarial examples [Hendrycks et al., 2021], which are images from the ID classes, but wrongly classified by standard Image Net classifiers. Misclassification detection and selective prediction was only performed for Imagenet1K, since MNIST classifiers easily reach accuracies of 99% on the test set, thus hardly misclassifying any samples. In all cases except selective prediciton, we measured AUROC, FPR at TPR of 95% and AUPR of classifying ID vs. OOD, non-adversarial vs. adversarial and correctly classified vs. misclassified samples (on ID test set), using the epistemic uncertainty estimate provided by the different methods. For selective prediction, we utilized the epistemic uncertainty estimate to select a subset of samples on the ID test set. (a) Ground Truth - HMC (b) c SG-HMC (c) Laplace (d) MC dropout (e) Deep Ensembles (f) Our Method - QUAM Figure C.3: Epistemic uncertainty as in Eq. (2). Yellow denotes high epistemic uncertainty. Purple denotes low epistemic uncertainty. The black lines show the decision boundary of the reference model w. HMC is considered to be the ground truth epistemic uncertainty. The estimate of QUAM is closest to the ground truth. All other methods underestimate the epistemic uncertainty in the top left and bottom right corner, as all models sampled by those predict the same class with high confidence for those regions. 6 4 2 0 2 4 6 (a) Ground Truth - HMC 6 4 2 0 2 4 6 (b) c SG-HMC 6 4 2 0 2 4 6 (c) Laplace 6 4 2 0 2 4 6 (d) MC dropout 6 4 2 0 2 4 6 (e) Deep Ensembles 6 4 2 0 2 4 6 (f) Our Method - QUAM Figure C.4: Variance between different models found by different methods on synthetic sine dataset. Orange line denotes the empirical mean of the averaged models, shades denote one, two and three standard deviations respectively. HMC is considered to be the ground truth epistemic uncertainty. The estimate of QUAM is closest to the ground truth. All other methods fail to capture the variance between points as well as the variance left outside the region ([ π, π]) datapoints are sampled from. Table C.1: Additional baseline Mo LA: AUROC using the epistemic uncertainty of a given, preselected model as a score to distinguish between ID (MNIST) and OOD samples. Results for additional baseline method Mo LA, comparing to Laplace approximation, Deep Ensembles (DE) and QUAM. Results are averaged over three independent runs. Dood Laplace Mo LA DE QUAM FMNIST .978 .004 .986 .002 .988 .001 .994 .001 KMNIST .959 .006 .984 .000 .990 .001 .994 .001 EMNIST .877 .011 .920 .002 .924 .003 .937 .008 OMNIGLOT .963 .003 .979 .000 .983 .001 .992 .001 C.4.1 MNIST OOD detection experiments were performed on MNIST with Fashion MNIST (FMNIST) [Xiao et al., 2017], EMNIST [Cohen et al., 2017], KMNIST [Clanuwat et al., 2018] and OMNIGLOT [Lake et al., 2015] as OOD datasets. In case of EMNIST, we only used the letters subset, thus excluding classes overlapping with MNIST (digits). We used the MNIST (test set) vs FMNIST (train set) OOD detection task to tune hyperparameters for all methods. The evaluation was performed using the complete test sets of the above-mentioned datasets (n = 10000). For each seed, a separate set of Deep Ensembles was trained. Ensembles with the size of 10 were found to perform best. MC dropout was used with a number of samples set to 2048. This hyperparameter setting was found to perform well. A higher sampling size would increase the performance marginally while increasing the computational load. Noteworthy is the fact, that with these settings the computational requirements of MC dropout surpassed those of QUAM. Laplace approximation was performed only for the last layer, due to the computational demand making it infeasible on the full network with our computational capacities. Mixture of Laplace approximations Eschenhagen et al. [2021] was evaluated as well using the parameters provided in the original work. Notably, the results from the original work suggesting improved performance compared to the Deep Ensembles on these tasks could not be reproduced. Comparison is provided in Table C.1. SG-HMC was performed on the full network using the Python package from Kapoor [2023]. Parameters were set in accordance with those of the original authors [Zhang et al., 2020]. For QUAM, the initial penalty parameter found by tuning was c0 = 6, which was exponentially increased (ct+1 = ηct) with η = 2 every 14 gradient steps for a total of two epochs through the training dataset. Gradient steps were performed using Adam [Kingma and Ba, 2014] with a learning rate of 5.e-3 and weight decay of 1.e-3, chosen equivalent to the original training parameters of the model. A temperature of 1.e-3 was used for scaling the cross-entropy loss, an approximation for the posterior probabilities when calculating Eq. (4). Detailed results and additional metrics and replicates of the experiments can be found in Tab. C.2. Experiments were performed three times with seeds: {42, 142, 242} to provide confidence intervals. Histograms of the scores on the ID dataset and the OOD datasets for different methods are depicted in Fig. C.5. C.4.2 Image Net For Image Net1K [Deng et al., 2009], OOD detection experiments were performed with Image Net-O [Hendrycks et al., 2021], adversarial example detection experiments with Image Net-A [Hendrycks et al., 2021], and misclassification detection as well as selective prediction experiments on the official validation set of Image Net1K. For each experiment, we utilized a pre-trained Efficient Net [Tan and Le, 2019] architecture with 21.5 million trainable weights available through Py Torch [Paszke et al., 2019], achieving a top-1 accuracy of 84.2% as well as a top-5 accuracy of 96.9%. c SG-HMC was performed on the last layer using the best hyperparameters that resulted from a hyperparameter search around the ones suggested by the original authors [Zhang et al., 2020]. The Laplace approximation with the implementation of [Daxberger et al., 2021] was not feasible to compute for this problem on our hardware, even only for the last layer. Similarly to the experiments in section C.4.1, we compare against a Deep Ensemble consisting of 10 pre-trained Efficient Net architectures ranging from 5.3 million to 66.3 million trainable weights (DE (all)). Also, we retrained the last layer of 10 ensemble members (DE (LL)) given the same base network. We also compare Table C.2: Detailed results of MNIST OOD detection experiments, reporting AUROC, AUPR and FPR@TPR=95% for individual seeds. OOD dataset Method Seed AUPR AUROC FPR@TPR=95% c SG-HMC 42 0.8859 0.8823 0.5449 142 0.8714 0.8568 0.8543 242 0.8797 0.8673 0.7293 Laplace 42 0.8901 0.8861 0.5273 142 0.8762 0.8642 0.7062 242 0.8903 0.8794 0.6812 Deep Ensembles 42 0.9344 0.9239 0.4604 142 0.9325 0.9236 0.4581 242 0.9354 0.9267 0.4239 MC dropout 42 0.8854 0.8787 0.5636 142 0.8769 0.8630 0.6718 242 0.8881 0.8751 0.6855 QUAM 42 0.9519 0.9454 0.3405 142 0.9449 0.9327 0.4538 242 0.9437 0.9317 0.4325 c SG-HMC 42 0.9532 0.9759 0.0654 142 0.9610 0.9731 0.0893 242 0.9635 0.9827 0.0463 Laplace 42 0.9524 0.9754 0.0679 142 0.9565 0.9739 0.0788 242 0.9613 0.9824 0.0410 Deep Ensembles 42 0.9846 0.9894 0.0319 142 0.9776 0.9865 0.0325 242 0.9815 0.9881 0.0338 MC dropout 42 0.9595 0.9776 0.0644 142 0.9641 0.9748 0.0809 242 0.9696 0.9848 0.0393 QUAM 42 0.9896 0.9932 0.0188 142 0.9909 0.9937 0.0210 242 0.9925 0.9952 0.0132 c SG-HMC 42 0.9412 0.9501 0.2092 142 0.9489 0.9591 0.1551 242 0.9505 0.9613 0.1390 Laplace 42 0.9420 0.9520 0.1915 142 0.9485 0.9617 0.1378 242 0.9526 0.9640 0.1165 Deep Ensembles 42 0.9885 0.9899 0.0417 142 0.9875 0.9891 0.0458 242 0.9884 0.9896 0.0473 MC dropout 42 0.9424 0.9506 0.2109 142 0.9531 0.9618 0.1494 242 0.9565 0.9651 0.1293 QUAM 42 0.9928 0.9932 0.0250 142 0.9945 0.9952 0.0194 242 0.9925 0.9932 0.0260 c SG-HMC 42 0.9499 0.9658 0.1242 142 0.9459 0.9591 0.1498 242 0.9511 0.9637 0.1222 Laplace 42 0.9485 0.9647 0.1238 142 0.9451 0.9597 0.1345 242 0.9526 0.9656 0.1077 Deep Ensembles 42 0.9771 0.9822 0.0621 142 0.9765 0.9821 0.0659 242 0.9797 0.9840 0.0581 MC dropout 42 0.9534 0.9663 0.1248 142 0.9520 0.9619 0.1322 242 0.9574 0.9677 0.1063 QUAM 42 0.9920 0.9930 0.0274 142 0.9900 0.9909 0.0348 242 0.9906 0.9915 0.0306 Figure C.5: MNIST: Histograms of uncertainty scores calculated for test set samples of the specified datasets. Table C.3: Calibration: expected calibration error (ECE) based on the weighted average predictive distribution. Reference refers to the predictive distribution of the given, pre-selected model. Experiment was performed on three distinct splits, each containing 7000 Image Net-1K validation samples. Reference c SG-HMC MCD DE QUAM .159 .004 .364 .001 .166 .004 .194 .004 .096 .006 (a) Reference (b) c SG-HMC Figure C.6: Calibration: confidence vs. accuracy based on (weighted) average predictive distribution of different uncertainty quantification methods. Point size indicates number of samples in the bin. against MC dropout used with 2048 samples with a dropout probability of 20%. The Efficient Net architectures utilize dropout only before the last layer. The adversarial model search for QUAM was performed on the last layer of the Efficient Net, which has 1.3 million trainable parameters. To enhance the computational efficiency, the output of the second-to-last layer was computed once for all samples, and this output was subsequently used as input for the final layer when performing the adversarial model search. We fixed c0 to 1 and exponentially updated it at every of the 256 update steps. Also, weight decay was fixed to 1.e-4 for the Adam optimizer [Kingma and Ba, 2014]. Two hyperparameters have jointly been optimized on Image Net-O and Image Net-A using a small grid search, with learning rate α {5.e-3, 1.e-3, 5.e-4, 1.e-4} and the exponential schedule update constant η {1.15, 1.01, 1.005, 1.001}. The hyperparameters α = 1.e-3 and η = 1.01 resulted in the overall highest performance and have thus jointly been used for each of the three experiments. This implies that c0 increases by 1% after each update step. We additionally searched for the best temperature and the best number of update steps for each experiment separately. The best temperature for scaling the cross-entropy loss when calculating Eq. (4) was identified as 0.05, 0.005, and 0.0005, while the best number of update steps was identified as 50, 100, and 100 for Image Net-O OOD detection, Image Net-A adversarial example detection, and Image Net1K misclassification detection, respectively. Selective prediction was performed using the same hyperparameters as misclassification detection. We observed that the adversarial model search is relatively stable with respect to these hyperparameters. The detailed results on various metrics and replicates of the experiments can be found in C.4. Histograms of the scores on the ID dataset and the OOD dataset, the adversarial example dataset and the correctly and incorrectly classified samples are depicted in Fig. C.7 for all methods. ROC curves, as well as accuracy over retained sample curves, are depicted in Fig. C.8. To provide confidence intervals, we performed all experiments on three distinct dataset splits of the ID datasets, matching the number of OOD samples. Therefore we used three times 2000 ID samples for Imagenet-O and three times 7000 ID samples for Imagenet-A and misclassification detection as well as selective prediction. Calibration. Additionally, we analyze the calibration of QUAM compared to other baseline methods. Therefore, we compute the expected calibration error (ECE) [Guo et al., 2017] on the Image Net-1K validation dataset using the expected predictive distribution. Regarding QUAM, the predictive distribution was optained using the same hyperparameters as for misclassification detection reported above. We find that QUAM improves upon the other considered baseline methods, although it was not directly designed to improve the calibration of the predictive distribution. Tab. C.3 states the ECE of considered uncertainty quantification methods and in Fig. C.6 the accuracy and number of samples (depicted by the size) for specific confidence bins is depicted. Table C.4: Detailed results of Image Net OOD detection, adversarial example detection and misclassification experiments, reporting AUROC, AUPR and FPR@TPR=95% for individual splits. OOD dataset / task Method Split AUPR AUROC FPR@TPR=95% Image Net-O Reference I 0.615 0.629 0.952 II 0.600 0.622 0.953 III 0.613 0.628 0.954 c SG-HMC I 0.671 0.682 0.855 II 0.661 0.671 0.876 III 0.674 0.679 0.872 MC dropout I 0.684 0.681 0.975 II 0.675 0.677 0.974 III 0.689 0.681 0.972 Deep Ensembles (LL) I 0.573 0.557 0.920 II 0.566 0.562 0.916 III 0.573 0.566 0.928 Deep Ensembles (all) I 0.679 0.713 0.779 II 0.667 0.703 0.787 III 0.674 0.710 0.786 QUAM I 0.729 0.758 0.766 II 0.713 0.740 0.786 III 0.734 0.761 0.764 Image Net-A Reference I 0.779 0.795 0.837 II 0.774 0.791 0.838 III 0.771 0.790 0.844 c SG-HMC I 0.800 0.800 0.785 II 0.803 0.800 0.785 III 0.799 0.798 0.783 MC dropout I 0.835 0.828 0.748 II 0.832 0.828 0.740 III 0.826 0.825 0.740 Deep Ensembles (LL) I 0.724 0.687 0.844 II 0.723 0.685 0.840 IIII 0.721 0.686 0.838 Deep Ensembles (all) I 0.824 0.870 0.385 II 0.837 0.877 0.374 III 0.832 0.875 0.375 QUAM I 0.859 0.875 0.470 II 0.856 0.872 0.466 III 0.850 0.870 0.461 Misclassification Reference I 0.623 0.863 0.590 II 0.627 0.875 0.554 III 0.628 0.864 0.595 c SG-HMC I 0.478 0.779 0.755 II 0.483 0.779 0.752 III 0.458 0.759 0.780 MC dropout I 0.514 0.788 0.719 II 0.500 0.812 0.704 III 0.491 0.788 0.703 Deep Ensembles (LL) I 0.452 0.665 0.824 II 0.421 0.657 0.816 III 0.425 0.647 0.815 Deep Ensembles (all) I 0.282 0.770 0.663 II 0.308 0.784 0.650 III 0.310 0.786 0.617 QUAM I 0.644 0.901 0.451 II 0.668 0.914 0.305 III 0.639 0.898 0.399 Figure C.7: Image Net: Histograms of uncertainty scores calculated for test set samples of the specified datasets. (a) OOD detection (b) Adversarial example detection (c) Misclassification (d) Selective prediction Figure C.8: Image Net-1K OOD detection results on Image Net-O, adversarial example detection results on Image Net-A, misclassification detection and selective prediction results on the validation dataset. ROC curves using the epistemic uncertainty of a given, pre-selected model (as in Eq. (2)) to distinguish between (a) the Image Net-1K validation dataset and Image Net-O, (b) the Image Net-1K validation dataset and Image Net-A and (c) the reference model s correctly and incorrectly classified samples. (d) Accuracy of reference model on subset composed of samples that exhibit lowest epistemic uncertainty. C.5 Comparing Mechanistic Similarity of Deep Ensembles vs. Adversarial Models The experiments were performed on MNIST, EMNIST, and KMNIST test datasets, using 512 images of each using Deep Ensembles, and the reference model w, trained on MNIST. Results are depicted in Fig. C.9. For each image and each ensemble member, gradients were integrated over 64 steps from 64 different random normal sampled baselines for extra robustness [Sundararajan et al., 2017]. Since the procedure was also performed on the OOD sets as well as our general focus on uncertainty estimation, no true labels were used for the gradient computation. Instead, predictions of ensemble members for which the attributions were computed were used as targets. Principal Component Analysis (PCA) was performed for the attributions of each image separately, where for each pixel the attributions from different ensemble members were treated as features. The ratios of explained variance, which are normalized to sum up to one, are collected from each component. If all ensemble members would utilize mutually exclusive features for their prediction, all components would be weighted equally, leading to a straight line in the plots in the top row in Fig. C.9. Comparatively high values of the first principal component to the other components in the top row plots in Fig. C.9 indicate low diversity in features used by Deep Ensembles. The procedure was performed similarly for an ensemble of adversarial models. The main difference was that for each image an ensemble produced as a result of an adversarial model search on that specific image was used. We observe, that ensembles of adversarial models utilize more dissimilar features, indicated by the decreased variance contribution of the first principal component. This is especially strong for ID data, but also noticeable for OOD data. C.6 Prediction Space Similarity of Deep Ensembles and Adversarial Models In the following, ensembles members and adversarial models are analyzed in prediction space. We used the same Deep Ensembles as the one trained on MNIST for the OOD detection task described in Sec. C.4.1. Also, 10 adversarial models were retrieved from the reference model w and a single OOD sample (KMNIST), following the same procedure as described in Sec. C.4.1. For the analysis, PCA was applied to the flattened softmax output vectors of each of the 20 models applied to ID validation data. The resulting points represent the variance of the model s predictions across different principal components [Fort et al., 2019]. The results in Fig. C.10 show, that the convex hull of blue points representing adversarial models, in general, is much bigger than the convex hull of orange points representing ensemble members across the first four principal components, which explain 99.99% of the variance in prediction space. This implies that even though adversarial models achieve similar accuracy as Deep Ensembles on the validation set, they are capable of capturing more diversity in prediction space. C.7 Computational Expenses Experiments on Synthetic Datasets The example in Sec. C.2 was computed within half an hour on a GTX 1080 Ti. Experiments on synthetic datasets shown in Sec. C.3 were also performed on a single GTX 1080 Ti. Note that the HMC baseline took approximately 14 hours on 36 CPU cores for the classification task. All other methods except QUAM finish within minutes. QUAM scales with the number of test samples. Under the utilized parameters and 6400 test samples, QUAM computation took approximately 6 hours on a single GPU and under one hour for the regression task, where the number of test points is much smaller. Experiments on Vision Datasets Computational Requirements for the vision domain experiments depend a lot on the exact utilization of the baseline methods. While Deep Ensembles can take a long time to train, depending on the ensemble size, we utilized either pre-trained networks for ensembling or only trained last layers, which significantly reduces the runtime. Noteworthy, MC-dropout can result in extremely high runtimes depending on the number of forward passes and depending on the realizable batch size for inputs. The same holds for SG-HMC. Executing the QUAM experiments on MNIST (Sec. C.4.1) took a grand total of around 120 GPU-hours on a variety of mostly older generation and low-power GPUs (P40, Titan V, T4), corresponding to roughly 4 GPU-seconds per sample. Executing the experiments on Image Net (Sec. C.4.2) took about 100 GPU-hours on a mix of A100 and A40 GPUs, corresponding to around 45 GPU-seconds per sample. The experiments presented in Sec.C.5 and C.6 took around 2 hours each on 4 GTX 1080 Ti. Figure C.9: The differences between significant component distribution are marginal on OOD data but pronounced on the ID data. The ID data would be subject to optimization by gradient descent during training, therefore the features are learned greedily and models are similar to each other mechanistically. We observe, that the members of Deep Ensembles show higher mechanistic similarity than the members of ensembles obtained from adversarial model search. Figure C.10: Convex hull of the down-projected softmax output from 10 Ensemble Members (orange) as well as 10 adversarial models (blue). PCA is used for down-projection, all combinations of the first four principal components (99.99% variance explained) are plotted against each other. Softmax outputs are obtained on a batch of 10 random samples from the ID validation dataset. The black cross marks the given, pre-selected model w.