# credal_bayesian_deep_learning__c69d21b0.pdf Published in Transactions on Machine Learning Research (10/2024) Credal Bayesian Deep Learning Michele Caprio michele.caprio@manchester.ac.uk Department of Computer Science, University of Manchester Souradeep Dutta souradeep@ece.ubc.ca Department of Computer Science, University of British Columbia Kuk Jin Jang jangkj@seas.upenn.edu Department of Computer and Information Science, University of Pennsylvania Vivian Lin vilin@seas.upenn.edu Department of Computer and Information Science, University of Pennsylvania Radoslav Ivanov ivanor@rpi.edu Department of Computer Science, Rensselaer Polytechnic Institute Oleg Sokolsky sokolsky@seas.upenn.edu Department of Computer and Information Science, University of Pennsylvania Insup Lee lee@seas.upenn.edu Department of Computer and Information Science, University of Pennsylvania Reviewed on Open Review: https: // openreview. net/ forum? id= 4NHF9AC5ui Uncertainty quantification and robustness to distribution shifts are important goals in machine learning and artificial intelligence. Although Bayesian Neural Networks (BNNs) allow for uncertainty in the predictions to be assessed, different sources of predictive uncertainty cannot be distinguished properly. We present Credal Bayesian Deep Learning (CBDL). Heuristically, CBDL allows to train an (uncountably) infinite ensemble of BNNs, using only finitely many elements. This is possible thanks to prior and likelihood finitely generated credal sets (FGCSs), a concept from the imprecise probability literature. Intuitively, convex combinations of a finite collection of prior-likelihood pairs are able to represent infinitely many such pairs. After training, CBDL outputs a set of posteriors on the parameters of the neural network. At inference time, such posterior set is used to derive a set of predictive distributions that is in turn utilized to distinguish between (predictive) aleatoric and epistemic uncertainties, and to quantify them. The predictive set also produces either (i) a collection of outputs enjoying desirable probabilistic guarantees, or (ii) the single output that is deemed the best, that is, the one having the highest predictive lower probability another imprecise-probabilistic concept. CBDL is more robust than single BNNs to prior and likelihood misspecification, and to distribution shift. We show that CBDL is better at quantifying and disentangling different types of (predictive) uncertainties than single BNNs and ensemble of BNNs. In addition, we apply CBDL to two case studies to demonstrate its downstream tasks capabilities: one, for motion prediction in autonomous driving scenarios, Published in Transactions on Machine Learning Research (10/2024) and two, to model blood glucose and insulin dynamics for artificial pancreas control. We show that CBDL performs better when compared to an ensemble of BNNs baseline. 1 Introduction One of the greatest virtues an individual can have is arguably being aware of their own ignorance, and acting cautiously as a consequence. Similarly, an autonomous system using neural networks (NNs) would greatly benefit from understanding the probabilistic properties of the NN s output (for example, its robustness to distribution shift), in order to incorporate them into any further decision-making. This paper collocates on the path of giving a machine such a desirable quality. In the last few years, there has been a proliferation of work on calibrating (classification) NNs, in order to estimate the confidence in their outputs (Guo et al., 2017) or to produce conformal sets that are guaranteed to contain the true label, in a probably approximately correct (PAC) sense (Park et al., 2020). While such methods are a promising first step, they require a calibration set (in addition to the original training set) and cannot be directly used on Out-Of-Distribution data without further examples. Bayesian Neural Networks (BNNs) offer one approach to overcome the above limitations. The Bayesian paradigm provides a rigorous framework to analyze and train uncertainty-aware neural networks, and more generally to support the development of learning algorithms (Jospin et al., 2022). In addition, it overcomes some of the drawbacks of deep learning models, namely that they are prone to overfitting, which adversely affects their generalization capabilities, and that they tend to be overconfident about their predictions when they provide a confidence interval. BNNs, though, are trained using a single prior, which may still suffer from miscalibration and robustness issues (Lenk & Orme, 2009). In this work we introduce Credal Bayesian Deep Learning (CBDL), a procedure that draws on concepts from the imprecise probability (IP) literature (Augustin et al., 2014; Troffaes & de Cooman, 2014; Walley, 1991). Unlike other techniques in the fields of artificial intelligence (AI) and machine learning (ML) involving imprecise probabilities that typically only focus on classification problems CBDL can be used for both classification and regression. It captures the ambiguity the designer faces when selecting which prior to choose for the parameters of a neural network, and which likelihood distribution to choose for the training data at hand. CBDL can be thought of as a NN trained using prior and likelihood finitely generated credal sets (FGCSs), Pprior and Plik, respectively.1 They are convex sets of probability measures having finitely many extreme elements (the elements that cannot be written as a convex combination of one another), ex Pprior and ex Plik, respectively (see also Remark 2). A very simple example of a prior FGCS Pprior is the collection of all the convex combinations of two one-dimensional Normal distributions N(µ1, σ2 1) and N(µ2, σ2 2), i.e., Pprior = {P : P = βN(µ1, σ2 1) + (1 β)N(µ2, σ2 2), for all β [0, 1]}. Consequently, in this toy example we have that ex Pprior = {N(µ1, σ2 1), N(µ2, σ2 2)}. FGCSs are further examined in section 2.2. Given the use of finitely generated credal sets, CBDL can also be seen as a non-condensed (uncountably) infinite ensemble of BNNs each BNN corresponding to a pair (P, L) of prior P from the prior FGCS Pprior and likelihood L from the likelihood FGCS Plik. Such infinite ensemble, though, is carried out using only finitely many elements, that is, only pairs (P ex, Lex) of components of the sets of extreme elements ex Pprior and ex Plik. This because every element P of Pprior can be obtained from a convex combinations of the elements of ex Pprior, and similarly for the likelihood FGCS. After training, CBDL produces a posterior FGCS Ppost on the parameters of the neural network. At inference time, Ppost is used to derive a predictive FGCS Ppred, that is, given a new input, a set of plausible distributions over the space of outputs.2 In turn, 1Intuitively, the larger these credal sets, the higher prior and likelihood ambiguity the user faces. 2As we shall see in section 3.1, since computing the posteriors and the predictive distributions is oftentimes an intractable problem, we approximate the elements of Ppost and of Ppred using Variational Inference (VI). As a consequence, we denote the VI-approximated posterior and predictive credal sets as Ppost and ˆPpred, respectively. Published in Transactions on Machine Learning Research (10/2024) Ppred generates a set of outputs or a single output, depending on the user s needs that enjoys desirable probabilistic guarantees, even when the elements of credal set Ppred are approximated, e.g. using variational inference. CBDL also gives a way of quantifying and disentangling different types of uncertainties within Ppred. We use a credal set approach to overcome some of the drawbacks of single BNNs. In particular, CBDL allows to counter the criticism to the practice in (standard) Bayesian statistics of (i) using a single, arbitrary prior to represent the initial state of ignorance of the agent, (ii) using non-informative priors to model ignorance, and (iii) using a single, arbitrary likelihood to represent the agent s knowledge about the sampling model.3 As a consequence, credal sets make the analysis more robust to prior and likelihood misspecification. In addition, they make it possible to quantify and distinguish between epistemic and aleatoric uncertainties (EU and AU, respectively). This is desirable in light of several areas of recent ML research, such as Bayesian deep learning (Depeweg et al., 2018; Kendall & Gal, 2017), adversarial example detection (Smith & Gal, 2018), and data augmentation in Bayesian classification (Kapoor et al., 2022). AU refers to the uncertainty that is inherent to the data generating process; as such, it is irreducible. Think, for example, of a coin toss. No matter how many times the coin is tossed, the stochastic variability of the experiment cannot be eliminated. EU, instead, refers to the lack of knowledge about the data generating process; as such, it is reducible. It can be lessened on the basis of additional data. For example, after only a few tosses, we are unable to gauge whether a coin is biased or not, but if we repeat the experiment long enough, this type of uncertainty vanishes. We note in passing that predictive EU cannot be adequately captured using a single BNN (Hüllermeier & Waegeman, 2021; Fellaji & Pennerath, 2024). One reason for this is that selecting a unique prior and a unique likelihood implicitly assumes perfect knowledge around the true prior and the true data generating process. Another, more subtle one, is explored in the last paragraph of section 2.1. Methods that disentangle between the two types of (predictive) uncertainties that are based on a single distribution are ad-hoc, but are not theoretically well-justified. EU can be typically reduced by retraining the model using an augmented training set (Lin et al., 2024) (e.g. via semantic preserving transformations (Kaur et al., 2023), Puzzle Mix (Kim et al., 2020), etc.). On the other hand, since AU is irreducible, there is an increasing need for ML techniques that are able to detect and flag its excess, so that the user can proceed with caution . Remark 1. EU should not be confused with the concept of epistemic probability (de Finetti, 1974; 1975; Walley, 1991). In the subjective probability literature, epistemic probability can be captured by a single distribution. Its best definition can be found in Walley (1991, Sections 1.3.2 and 2.11.2). There, the author specifies how epistemic probabilities model logical or psychological degrees of partial belief of the agent. We remark, though, how de Finetti and Walley work with finitely additive probabilities, while in this paper we use countably additive probabilities. The motivation for working with credal sets is threefold: (i) it allows to be robust against prior and likelihood misspecification; (ii) unlike when using a single probability distribution, it permits to represent ignorance in the sense of lack of knowledge; (iii) it allows to quantify and disentangle between EU and AU. A more in-depth discussion can be found in Appendix A. In addition, we point out that despite a hierarchical Bayesian model (HBM) approach seems to be a viable alternative to one based on credal sets, these latter are better justified philosophically, and do not suffer from the same theoretical shortcomings of HBM procedures (Bernardo, 1979; Hüllermeier & Waegeman, 2021; Jeffreys, 1946; Walley, 1991). A more detailed explanation can be found in Appendix B. Let us mention that we only compare CBDL against Bayesian techniques because a comparison against non-Bayesian ones is either difficult to justify, or unfair. Indeed, the initial knowledge component, captured by the prior distribution (or the set of priors, in our case), has no direct comparisons with other approaches. Take conformal prediction (Vovk et al., 2022; Shafer & Vovk, 2008; Gibbs et al., 2023; Barber et al., 2023) for instance. It is a model-free method, which means that no prior knowledge around the experiment at hand is required. The only choice the user makes is which non-conformity score to use. How would we compare this approach with the choice of different priors, which conveys the idea that the user knows that a true distribution over the space of neural network parameters exists, but it is not fully known? In addition, despite the ubiquitous claim that conformal prediction (CP) is an uncertainty quantification tool, it is actually not. CP is an uncertainty representation tool. Indeed, CP represents uncertainty via 3Criticisms (i) and (iii) are also pointed out in Manchingal & Cuzzolin (2022, Section 2.2). Published in Transactions on Machine Learning Research (10/2024) the conformal prediction region (on the other hand, in this paper we represent uncertainty via credal sets). It does not quantify it: there is no real value attached to any kind of predictive uncertainty (aleatoric or epistemic). Some claim that the diameter of the conformal prediction region quantifies the uncertainty, but even in that case, it is unable to distinguish between AU and EU. Indeed, the diameter is a positive function of both: it increases as both increase, and hence it cannot be used to distinguish between the two. We summarize our contributions next: (1) We present CBDL, and develop the theoretical tools and the algorithm required to use it in practice. (2) We show that a CBDL approach is more robust than single BNNs to prior and likelihood misspecification, and to distribution shifts. We also explain how, during inference, the credal set of posteriors Ppost on the network parameters obtained during training is used to derive a credal set of predictive distributions Ppred, and in turn a set of outcomes or a single outcome that enjoys probabilistic guarantees.4 (3) We show how CBDL is better at quantifying and disentangling predictive AU and EU than single BNNs and ensemble of BNNs.5 We also apply CBDL to two safety critical systems to demonstrate its downstream tasks capabilities. One, motion prediction for autonomous driving, and two, the human insulin and blood glucose dynamics for artificial pancreas control. We demonstrate improvements in both these settings with respect to ensemble of BNNs methods, the reason being that better uncertainty quantification leads to a positive impact on the decisions made using them. Before moving on, let us point out that while CBDL pays a computational price coming from the use of credal sets, it is able to quantify both (predictive) EU and AU, and to do so in a principled manner, unlike single and ensembles of BNNs. In addition, it requires less stringent assumptions on the nature of the prior and likelihood ambiguity faced by the agent than other imprecise-probabilities-based techniques, as explained in Appendices B and L. We also stress that if the user prioritizes computational efficiency, they should use non-Bayesian methods, as they are faster to implement than Bayesian techniques. In safety-critical situations instead where using uncertainty-informed methods is crucial for quantifying the types of uncertainties, but also for the outcome of the analysis at hand CBDL is a natural choice. Structure of the Paper. Section 2 presents the needed preliminary concepts, followed by section 3 that introduces and discusses the CBDL algorithm, together with its theoretical properties. We present our experimental results in section 4, and we examine the related work in section 5. Section 6 concludes our work. In the appendices, we give further theoretical and philosophical arguments and we prove our claims. Notation Summary. In this paper, we make extensive use of different notation and acronyms. We summarize them in Table 1 to help the reader navigate through them. 2 Background and Preliminaries In this section, we present the background notions that are needed to understand our main results. In section 2.1 we introduce Bayesian Neural Networks. Section 2.2 discusses (finitely generated) credal sets, upper and lower probabilities, and imprecise highest density regions. Section 2.3 introduces the concepts of upper and lower entropy, which are used to quantify and disentangle AU and EU. The reader familiar with these concepts can skip to section 3. 2.1 Bayesian Neural Networks In line with the recent survey on BNNs by Jospin et al. (2022), Bayes theorem can be stated as P(H | D) = [P(D | H)P(H)]/P(D) = P(D, H)/ R P(D, H )d H , where H is a hypothesis about which the agent holds some prior beliefs, and D is the data the agent uses to update their initial opinion. Probability distribution P(D | H) represents how likely it is to observe data D if hypothesis H were to be true, and is called likelihood, while probability distribution P(H) represents the agent s initial opinion around the plausibility 4Once again, as we shall see in section 3.1, we approximate the elements of the posterior and the predictive credal sets using Variational Inference (VI). 5We note in passing how recently Mucsányi et al. (2024) show that (finite) deep ensembles are the current state-of-the-art methods for quantifying and disentangling different types of uncertainties on Image Net. Our experiments show that CBDL improves on (finite) ensembles of BNNs, both in uncertainty quantification and disentanglement, and in downstream tasks performances. Together with the findings in Mucsányi et al. (2024), this is an additional argument in favor of the effectiveness of our methodology. Published in Transactions on Machine Learning Research (10/2024) Notation and Acronyms Meaning FGCS Finitely Generated Credal Set CBDL Credal Bayesian Deep Learning Upper case latin letters, such as P and L Probability measures, e.g. prior or likelihood Lower case latin letters, such as p and ℓ The pdf/pmf associated to probability measures Π A finite collection of probability measures Conv( ) The convex hull operator Π The convex hull of Π, i.e. Π = Conv(Π) Pprior A prior FGCS Plik A likelihood FGCS Ppost A posterior FGCS VI Variational Inference Ppost A posterior FGCS whose extreme elements are the VI approximations of the correct ones Ppred A predictive FGCS ˆPpred A predictive FGCS whose extreme elements are the VI approximations of the correct ones ex P The extreme elements of an FGCS P BNN Bayesian Neural Network D = Dx Dy Training set; Dx is the set of training inputs, and Dy is the set of training outputs EBNN Ensemble of Bayesian Neural Networks x 7 Φ(x) Function representing the neural network architecture P, P Lower and upper probabilities, respectively HDR Highest Density Region IHDR Imprecise Highest Density Region IRα(P) Imprecise Highest Density Region of level α associated with FGCS P H(P) Entropy of probability measure P H(P), H(P) Upper and lower entropy associated with a credal set EU(P) Epistemic Uncertainty encoded in an FGCS P AU(P) Aleatoric Uncertainty encoded in an FGCS P TU(P) Total Uncertainty encoded in an FGCS P #A Cardinality of a generic set A post, pred The act of computing a posterior and a posterior predictive distributions, respectively BMA Bayesian Model Averaging Table 1: Notation and acronyms that we use throughout the paper, together with their meaning. Published in Transactions on Machine Learning Research (10/2024) of hypothesis H, and is called prior. The evidence available is encoded in P(D) = R P(D, H )d H , while posterior probability P(H | D) represents the agent s updated opinion. Using Bayes theorem to train a predictor can be understood as learning from data D: the Bayesian paradigm offers an established way of quantifying uncertainty in deep learning models. BNNs are stochastic artificial neural networks (ANNs) trained using a Bayesian approach (Caprio & Mukherjee, 2023b; Goan & Fookes, 2020; Jospin et al., 2022; Lampinen & Vehtari, 2001; Titterington, 2004; Wang & Yeung, 2021). The goal of ANNs is to represent an arbitrary function y = Φ(x). Let θ represent the parameters of the network, and call Θ the space θ belongs to. Stochastic neural networks are a type of ANN built by introducing stochastic components to the network. This is achieved by giving the network either a stochastic activation or stochastic weights to simulate multiple possible models with their associated probability distribution. This can be summarized as θ p(θ), y = Φθ(x) + ε, where Φ depends on θ to highlight the stochastic nature of the neural network, p is the density of a probability measure P on Θ,6 and ε represents random noise to account for the fact that function Φθ is just an approximation. The connection between the BNN notation and the general one of the Bayes theorem is explained in the next two paragraphs. To design a BNN, the first step is to choose a deep neural network architecture, that is, functional model Φθ. Then, the agent specifies the stochastic model, that is, a prior distribution over the possible model parametrization p(θ), and a prior confidence in the predictive power of the model p(y | x, θ). Given the usual assumption that multiple data points from the training set are independent, the product Q (x,y) D p(y | x, θ) represents the likelihood of outputs y Dy given inputs x Dx and parameter θ, where (a) D = Dx Dy is the training set; (b) Dx = {xi}n i=1 is the collection of training inputs, which is a subset of the space X of inputs; (c) Dy = {yi}n i=1 is the collection of training outputs, which is a subset of the space Y of outputs. The model parametrization can be considered to be hypothesis H. Following Jospin et al. (2022), we assume independence between model parameters θ and training inputs Dx, in formulas Dx θ. Hence, Bayes theorem can be rewritten as p(θ | D) = p(Dy | Dx, θ)p(θ) R Θ p(Dy | Dx, θ )p(θ )dθ p(Dy | Dx, θ)p(θ). Notice that the equality comes from having assumed Dx θ. Posterior density p(θ | D) is typically high dimensional and highly nonconvex (Izmailov et al., 2021c; Jospin et al., 2022), so computing it and sampling from it is a difficult task. The first issue is tackled using Variational Inference (VI) procedures, while Markov Chain Monte Carlo (MCMC) methods address the second challenge. Both are reviewed in the context of machine learning in Jospin et al. (2022, Section V), where the authors also inspect their limitations. BNNs can be used for both regression and classification (Jospin et al., 2022, Section II); besides having a solid theoretical justification, there are practical benefits from using BNNs, as presented in Jospin et al. (2022, Section III). Suppose now we collect a new input x of interest, and that we want to predict its correct output y. Then, the posterior distribution p(θ | D) over the network parameters comes to the rescue, as it allows us to compute the so-called posterior predictive distribution, p( y | x, D) = Z Θ p( y | x, θ)p(θ | D)dθ = Eθ p(θ|D)[p( y | x, θ)], where p( y | x, θ) is the model distribution we specified before. Posterior predictive p( y | x, D) tells us how likely output y is to be the correct one for input x, given the knowledge encapsulated in the data D we collected, which enters the computation via the posterior probability p(θ | D). Let us add a remark here: oftentimes, scholars claim that, in a Bayesian setting, the distribution on the parameters θ captures the epistemic uncertainty (EU) faced by the agent. This is a somehow agreeable premise, akin to a second-order distribution reasoning. If we accept this assertion, though, we see how EU at the predictive level is not 6We can write p as the Radon-Nikodym derivative of P with respect to some σ-finite dominating measure µ, that is, p = d P/dµ. Published in Transactions on Machine Learning Research (10/2024) quantifiable any more, since it gets washed away by taking the expectation Eθ p(θ|D)[ ]. As we shall see later, CBDL allows to overcome this conceptual shortcoming. 2.2 Imprecise Probabilities As CBDL is rooted in the theory of imprecise probabilities (IPs), in this section we give a gentle introduction to the IP concepts we will use throughout the paper. CBDL is based on the Bayesian sensitivity analysis (BSA) approach to IPs, that in turn is grounded in the dogma of ideal precision (DIP) (Berger, 1984), (Walley, 1991, Section 5.9). The DIP posits that in any problem there is an ideal probability model which is precise, but which may not be precisely known. We call this condition ambiguity (Ellsberg, 1961; Gilboa & Marinacci, 2013). Facing ambiguity can be represented mathematically by a set Pprior of priors and a set Plik of likelihoods that seem plausible or fit to express the agent s beliefs on the parameters of interest and their knowledge of the data generating process (DGP). Generally speaking, the farther apart the boundary elements of the sets (i.e., their infimum and supremum), the higher the agent s ambiguity. Of course, if Pprior and Plik are singletons we go back to the usual Bayesian paradigm. A procedure based on sets Pprior and Plik yields results that are more robust to prior and likelihood misspecification than a regular Bayesian method. In the presence of prior ignorance and indecisiveness about the sampling model, it is better to give answers in the form of intervals or sets, rather than arbitrarily select a prior and a likelihood, and then update. Sets Pprior and Plik allow to represent indecision, thus leading to less informative but more robust and valid conclusions. Remark 2. Throughout the paper, we denote by Π = {P1, . . . , Pk}, k N, a finite set of probabilities on a generic space Ω, such that for all j {1, . . . , k}, Pj cannot be written as a convex combination of the other k 1 components of Π. We denote by Π its convex hull Π Conv(Π), i.e., the set of probabilities Q on Ωthat can be written as Q(A) = Pk j=1 βj Pj(A), for all A Ω, where the βj s are elements of [0, 1] that sum up to 1. In the literature, it is referred to as a Finitely Generated Credal Set (FGCS, Levi (1980); Cozman (2000b)). Notice then that the extreme elements of Π correspond to the elements of Π, in formulas exΠ = Π. Simple graphical representations of finitely generated credal sets are given in Figures 1 and 2. p(ω1) p(ω2) P({ω1}) P({ω2}) Figure 1: Suppose we are in a 3-class classification setting, so Ω= {ω1, ω2, ω3}. Then, any probability measure P on Ωcan be seen as a probability vector. For example, suppose P({ω1}) = 0.6, P({ω2}) = 0.3, and P({ω3}) = 0.1. We have that P (0.6, 0.3, 0.1) . Since its elements are positive and sum up to 1, probability vector P belongs to the unit simplex, the purple triangle in the figure. Then, we can specify Π = {P1, . . . , P5}, and obtain as a consequence that Π = Conv(Π) is the orange pentagon. It is a convex polygon with finitely many extreme elements, and it is the geometric representation of a finitely generated credal set. Let us now introduce the concepts of lower and upper probabilities. The lower probability P associated with Π is given by P(A) = inf P Π P(A), for all A Ω. The upper probability P associated with Π is defined as the conjugate to P, that is, P(A) := 1 P(Ac) = sup P Π P(A),7 for all A Ω. These definitions hold even if Π is not finite. Then, we have the following important result. 7Here, Ac := Ω\ A. Published in Transactions on Machine Learning Research (10/2024) Proposition 3. P is the upper probability for Π if and only if it is also the upper probability for Π . That is, P(A) = sup P Π P(A) = sup P Π P (A), for all A Ω. The same holds for the lower probability. A version of Proposition 3 was proven in Dantzig (1963), while a variant for finitely additive probability measures can be found in Walley (1991, Section 3.6). A simple graphical representation of upper and lower probabilities for a set A is given in Figure 2. We now use lower probability P to define the α-level Imprecise Highest Density Region (IHDR), for some α [0, 1]. Definition 4. (Coolen, 1992, Section 2) Let α be any value in [0, 1]. Then, set IRα(Π ) Ωis called a (1 α)-Imprecise Highest Density Region (IHDR) if 1. P[{ω IRα(Π )}] 1 α; IRα(Π ) dω is a minimum. If Ωis at most countable, we replace R IRα(Π ) dω with #IRα(Π ), where # denotes the cardinality operator. Figure 2: In this figure, a replica of Flint et al. (2017, Figure 1), Π = {P1, P2}, where P1 and P2 are two Normal distributions whose probability density functions (pdf s) p1 and p2 are given by the dashed blue and brown curves, respectively. Their convex hull is Π = Conv(Π) = {Q : Q = βP1 + (1 β)P2, for all β [0, 1]}. The pdf q of an element Q of Π is depicted by a solid black curve. In addition, let A = [ 0.8, 0.4]. Then, P(A) = R 0.4 0.8 p2(ω)dω 0, while P(A) is given by the red shaded area under p1, that is, P(A) = R 0.4 0.8 p1(ω)dω. Definition 4 holds also if Π = exΠ is not finite. Notice that condition 2 is needed so that IRα(Π ) is the subset of Ωhaving the lowest possible cardinality, which still satisfies condition 1. By the definition of lower probability, Definition 4 implies that P [{ω IRα(Π )}] 1 α, for all P Π . Here lies the appeal of the IHDR concept. Let us give a simple example, borrowed from Caprio et al. (2024a). Suppose Ω= {ω1, . . . , ω5}, Π = {P1, P2, P3}, and α = 0.1. The numerical values for Ps({ωj}) are given in Table 2, for all s {1, 2, 3} and all j {1, . . . , 5}. Then, from Proposition 3 and Definition 4, we have that IRα(Π ) = {ω1, ω2, ω3}. ω1 ω2 ω3 ω4 ω5 P1 0.7 0.25 0.03 0.01 0.01 P2 0.6 0.2 0.1 0.05 0.05 P3 0.5 0.3 0.15 0.025 0.025 Table 2: Numerical values for our example. It is easy to see that the smallest subset of Ωthat is assigned a probability of at least 0.9 by all the elements of Π is {ω1, ω2, ω3}. An operative way of building the IHDR is to consider the union of the (precise) Highest Density Regions (HDRs) of the elements of Π = exΠ .8 Let us define them formally. 8Here, by operative we mean that this procedure to compute the IHDR is easy to carry out in practice. Published in Transactions on Machine Learning Research (10/2024) Definition 5. (Coolen, 1992, Section 1) Pick any Pj Π, j {1, . . . , k}. Let α be any value in [0, 1]. Then, set Rα(Pj) Ωis called a (1 α)-Highest Density Region (HDR) for Pj if Pj[{ω Rα(Pj)}] 1 α and Z Rα(Pj) dω is a minimum. If Ωis at most countable, we replace R Rα(Pj) dω with #Rα(Pj). Equivalently (Hyndman, 1996), Rα(Pj) = {ω Ω: pj(ω) pα j } Ω, where pj is the pdf or the probability mass function (pmf) of Pj, and pα j is a constant value. In particular, it is the largest constant such that Pj[{ω Rα(Pj)}] 1 α. In dimension 1, Rα(Pj) can be interpreted as the smallest collection of elements of Ω(interval or union of intervals) to which distribution Pj assigns probability of at least 1 α. As we can see, HDRs are a Bayesian counterpart of confidence intervals.9 We give a simple visual example in Figure 3. Figure 3: The 0.25-HDR from a Normal Mixture density. This picture is a replica of Hyndman (1996, Figure 1). The geometric representation of 75% probability according to Pj is the area between the pdf curve pj(ω) and the horizontal bar corresponding to p0.25 j . A higher probability coverage (according to Pj) would correspond to a lower constant, so pα j < p0.25 j , for all α < 0.25. In the limit, we recover 100% coverage at p0 j = 0. As we mentioned earlier, an operative way of obtaining IHDR IRα(Π ) is by putting IRα(Π ) = k j=1Rα(Pj). Thanks to Proposition 3, by taking the union of the HDRs, we ensure that all the probability measures in the credal set Π = Conv(Π) assign probability of at least 1 α to the event {ω IRα(Π )}. In turn, this implies that P[{ω IRα(Π )}] = minj {1,...,k} Pj[{ω IRα(Π )}] 1 α. We also have that the difference between the upper and lower probabilities of {ω IRα(Π )} is bounded by α. To see this, notice that P[{ω IRα(Π )}] 1, so P[{ω IRα(Π )}] P[{ω IRα(Π )}] α. 2.3 Quantifying and Disentangling Aleatoric and Epistemic Uncertainties Recall that, given a probability measure P on a generic space Ω, the (Shannon) entropy of P is defined as H(P) := E[ log p] = R Ωlog[p(ω)]P(dω) if Ωis uncountable, where p denotes the pdf of P. If Ωis at most countable, we have that H(P) = P ω ΩP({ω}) log[P({ω})]. As pointed out by Dubois & Hüllermeier (2007); Hüllermeier & Waegeman (2021), the entropy primarily captures the shape of the distribution, namely its peakedness or non-uniformity, and hence informs about the predictability of the outcome of a random experiment: the higher its value, the lower the predictability. Then, we can define the imprecise versions of the Shannon entropy as proposed by Abellán et al. (2006); Hüllermeier & Waegeman (2021), H(P) := sup P Π H(P) and H(P) := inf P Π H(P), called the upper and lower Shannon entropy, respectively.10 Notice that these definitions hold for all sets of probabilities, not just for (finitely generated) credal sets. The upper entropy is a measure of total uncertainty since it represents the minimum level of predictability associated with the elements of Π . In Abellán et al. (2006); Hüllermeier & Waegeman (2021), the authors 9Hence standard choices for the value of α are 0.1, 0.05, 0.01. 10In Appendix E, we provide bounds to the values of upper and lower entropy. Published in Transactions on Machine Learning Research (10/2024) posit that it can be decomposed as a sum of aleatoric and epistemic uncertainties, and that this latter can be specified as the difference between upper and lower entropy, thus obtaining H(P ) | {z } TU(Π ) = H(P ) | {z } AU(Π ) H(P ) H(P ) | {z } EU(Π ) where TU(Π ) denotes the total uncertainty associated with set Π , AU(Π ) is the AU associated with Π , and EU(Π ) represents the EU associated with Π . As we can see, if Π is a singleton, then TU(Π ) = AU(Π ), and EU(Π ) = 0. This captures the idea that, in general, a single distribution can only gauge aleatoric uncertainty.11 We have the following proposition. Proposition 6. Let Π, Π be sets of probability measures as the ones considered in Remark 2. Then, sup P Π H(P) = H(P) H(P ) = sup P Π H(P ) and inf P Π H(P) = H(P) = H(P ) = inf P Π H(P ). Proposition 6 tells us that the upper entropy of the extreme elements in Π = exΠ is a lower bound for the upper entropy of the whole credal set Π , and that the lower entropy of the extreme elements in Π is equivalent to the lower entropy of the whole credal set Π . These facts imply that AU(Π ) = AU(Π), and that EU(Π ) EU(Π). In addition, as a consequence of (Smieja & Tabor, 2012, Theorem III.1), we have that EU(Π ) EU(Π) + log(#Π). In turn, we have that EU(Π ) [EU(Π), EU(Π) + log(#Π)]. We briefly mention that other uncertainty measures based on credal sets are also available (see Bronevich & Rozenberg (2021); Hofman et al. (2024); Hüllermeier & Waegeman (2021) or Appendix E for a few examples) and they can be used in place of upper and lower entropy12 to quantify EU and AU within our credal region P, as long as the measure chosen for the total uncertainty is bounded. Let us also add a small remark. Although EU can be reduced with an increasing amount of data, it is very seldom the case that it goes to zero when a finite amount of data is collected. When that happens, it means that the initial uncertainty was very low in the first place. Typically, EU goes to zero only asymptotically, as a finite amount of data is almost never enough to overcome initial uncertainty (Walley, 1991; Wimmer et al., 2023). 3 Our Procedure and Its Properties This is the main portion of the paper. In section 3.1, we present and discuss the CBDL algorithm. Its theoretical properties are derived in section 3.2. 3.1 CBDL algorithm Recall that D = Dx Dy denotes the training set, where Dx = {xi}n i=1 X is the collection of training inputs, Dy = {yi}n i=1 Y is the collection of training outputs, and X and Y denote the input and output spaces, respectively. We then denote by P a generic prior on the parameters θ Θ of a BNN having pdf p, and by L Lx,θ a generic likelihood on the space Y of outputs having pdf ℓ ℓx,θ. The act of computing the posterior from prior P and likelihood L using a BNN is designated by post[P, L], and the act of deriving the posterior predictive from posterior P( | D) and likelihood L is designated by pred[P( | D), L]. The CBDL procedure is presented in Algorithm 1, and it is discussed in the following paragraphs. During training, in Step 1 the user specifies K priors on the parameters of the neural network, that constitute the extrema ex Pprior of the prior FGCS Pprior. Similarly, in Step 2 they elicit S likelihoods capturing the possible architectures of the neural network, that correspond to the extrema ex Plik of the likelihood FGCS Plik. Let us give an example. Outlined in Jospin et al. (2022, Sections IV-B and IV-C1), for classification, the standard process for BNNs involves 11It can gauge EU, though, if it is a second-order distribution, or if it is the result of an ensemble of probabilities. 12Which, although easy to compute, can sometimes suffer from shortcomings coming from lacking the monotonicity property. That is, there are credal sets that, although nested into each other, have the same upper and/or lower entropy (Hüllermeier & Waegeman, 2021). Published in Transactions on Machine Learning Research (10/2024) Algorithm 1 Credal Bayesian Deep Learning (CBDL) Training and Inference During Training Step 1 Specify K priors ex Pprior = {P ex k }K k=1 Step 2 Specify S likelihoods ex Plik = {Lex s }S s=1 Step 3 Compute Pk,s( | D) = post[P ex k , Lex s ], for all k and all s % Approximated via VI by Pk,s% During Inference Inputs: New input x X Parameters: Confidence parameter α [0, 1] Outputs: Predictive Aleatoric and Epistemic Uncertainties, α-level IHDR Step 4 Compute P pred k,s = pred[ Pk,s, Lex s ], for all k and all s % Approximated via VI by ˆP pred k,s % Step 5 Compute and return AU( ˆPpred) and the bounds for EU( ˆPpred) Step 6 Compute and return the (1 α)-IHDR IRα( ˆPpred) A Normal prior with zero mean and diagonal covariance σ2I on the coefficients of the network, that is, p(θ) = N(0, σ2I). In the context of CBDL, we could specify, e.g. ex Pprior = {P : p(θ) = N(µ, σ2I), µ {µ , 0, µ+}, σ2 {3, 7}}. That is, the extreme elements of the prior credal set are five independent Normals having different levels of fatness of the tails, and centered at a vector µ+ having positive entries, a vector µ having negative entries, and a vector 0 having entries equal to 0. They capture the ideas of positive bias, negative bias, and no bias of the coefficients, respectively. This is done to hedge against possible prior misspecification. A Categorical likelihood, p(y | x, θ) = Cat(Φθ(x)), whose parameter is given by the output of a functional model Φθ. In the context of CBDL, we could specify the set of extreme elements of the likelihood credal set as ex Plik = {L : ℓx,θ(y) = Cat(Φs,θ(x)), s {1, . . . , S}}. Specifying set ex Plik, then, corresponds to eliciting a finite number S of possible (parametrized) architectures for the neural network Φs,θ, s {1, . . . , S}, and obtain, as a consequence, S categorical distributions {Cat(Φs,θ(x))}S s=1. This captures the ambiguity around the true data generating process faced by the agent, and allows them to hedge against likelihood misspecification. More in general, we can use the priors and likelihoods that better fit the type of analysis we are performing.13 For example, for the choice of the priors we refer to Fortuin et al. (2021), where the authors study the problem of selecting the right type of prior for BNNs. Step 3 performs an element-wise application of Bayes rule for all the elements of ex Pprior and ex Plik. Each posterior is approximated using the Variational Inference (VI) method (Jospin et al., 2022, Section V). That is, we project every posterior Pk,s( | D) onto a set S of well-behaved distributions (e.g. Normals) using the KL divergence. In formulas, Pk,s = arg min Q S KL[Q Pk,s( | D)]. By well-behaved , we mean that they have to satisfy the conditions in Zhang & Gao (2020, Sections 2 and 3).14 This ensures that as the sample size goes to infinity, the approximated posteriors converge to the true data generating process. We also point out how despite using a VI approximation for the exact posteriors, as we shall see in the next section, the credal set of approximated posteriors is closer to the oracle posterior P o( | D) than any of its elements taken individually. As a consequence, working with credal sets leads to VI posterior approximations that are better than the ones resulting from a single BNN, or an ensemble of BNNs, where several BNNs are combined into one. Remark 7. Although highly unlikely in practice, it is theoretically possible that the VI approximation of the (finite) set {Pk,s( | D)}k,s of posteriors is a singleton, see Figure 4. While the conditions in Zhang & Gao (2020) guarantee that asymptotically the approximated posteriors coincide with the true data generating process, this typically does not happen with finite-dimensional datasets. As a consequence, obtaining a singleton when projecting {Pk,s( | D)}k,s onto S may result in an underestimation of the uncertainties faced by the user. In that case, we either consider a different set whose elements still satisfy the conditions in Zhang 13Choosing 2 to 5 priors and likelihoods is usually enough to safely hedge against prior and likelihood misspecification. 14We assume that the conditions on the priors and the likelihoods given in Zhang & Gao (2020) are satisfied. Published in Transactions on Machine Learning Research (10/2024) & Gao (2020, Sections 2 and 3) on which to project {Pk,s( | D)}k,s according to the KL divergence, or we use a different projection operator , that is, a divergence different from the KL. For example, Rényi and χ2 divergences, or Hellinger and total variation metrics are suggested by Zhang & Gao (2020). Alternatively, we can consider a different approximation strategy altogether, for instance the Laplace approximation (Ritter et al., 2018). K L K L K L P2( D) P3( D) Figure 4: Let Θ denote the space of probability measures on Θ. Suppose that in the analysis at hand we specified three priors and only one likelihood, so S = 1 and we can drop the s index. Let {Pk( | D)}3 k=1 be the collection of exact posteriors, so that the black segment represents the exact posterior FGCS. Then, if we project the elements of {Pk( | D)}3 k=1 onto S1 via the KL divergence, we obtain the same distribution P. This is detrimental to the analysis because such an approximation underestimates the (posterior) epistemic (and possibly also aleatoric) uncertainty faced by the agent. Then, the user could specify a different set S2 of well-behaved distributions onto which project the elements of {Pk( | D)}3 k=1. In the figure, we see that they are projected onto S2 via the KL divergence to obtain P1, P2, and P3. The convex hull of these latter, captured by the red shaded triangle, represents the variational approximation of the exact posterior FGCS. After Step 3, we obtain a finite set { Pk,s}k,s of VI approximation of the posteriors on the network parameters, whose cardinality is K S. Its convex hull constitutes Ppost, that is, the VI-approximated posterior FGCS. We assume that ex Ppost = { Pk,s}k,s. This is an assumption because it may well be that due to the approximation procedure some of the elements of { Pk,s}k,s are not independent of one another. We defer to future work the design of a procedure that finds the elements of { Pk,s}k,s that cannot be written as a convex combination of one another. Being a combinatorial task, Step 3 is a computational bottleneck of Algorithm 1. We have to calculate K S VI approximations to as many posteriors, but this allows us to forego any additional assumptions on the nature of the lower and upper probabilities that are oftentimes required by other imprecise-probabilitiesbased techniques.15 Clearly, CBDL is simplified if either Pprior or Plik are singletons. Notice that in the case that Pprior and Plik are both not singletons, for all A Θ the interval [ P(A), P(A)] is wider than the case when one or the other is a singleton. In the limiting case where both are singletons, we retrieve the usual Bayesian updating, so the interval shrinks down to a point. Before moving to inference time, let us remark a difference between Bayesian Model Averaging (BMA) and CBDL. In BMA, the user specifies a distribution on the models. Translated in the notation we use in this work, this means having a discrete distribution Q over the K S prior-likelihood combinations, that is, over the elements of ex Ppost. Such Q is then used to select an element P from the (VI-approximated) posterior FGCS Ppost as Ppost P = X k,s Q({ Pk,s}) Pk,s, (1) 15If we are willing to make such assumptions, Theorem 10 in Appendix D shows how to compute the upper posterior using only upper prior and upper likelihood. Published in Transactions on Machine Learning Research (10/2024) where Q({ Pk,s}) [0, 1] for all k and all s, and P k,s Q({ Pk,s}) = 1 (Caprio & Mukherjee, 2023a). For example, if Q is the discrete uniform distribution on {1, . . . , K S}, then P is the so-called center of gravity of the credal set Ppost (Miranda & Montes, 2023, Section 3.2). Instead, CBDL does not select a unique distribution from Ppost. Rather, distributions are kept separate so to be able to derive a predictive FGCS in Step 4 of Algorithm 1, that is in turn used to quantify and disentangle predictive uncertainties, and to compute the predictive IHDR, that is, a collection of outputs having a high probability of being the correct ones for a new input x. During inference, a new input x is provided. In Step 4, every element of ex Ppost is used to derive a predictive distribution P pred k,s = pred[ Pk,s, Lex s ] on the output space Y. In particular, for every k and every s, the pdf ppred k,s of P pred k,s is obtained as ppred k,s ( y | x, x1, y1, . . . , xn, yn) = Z Θ ℓex s ( y | x, θ) pk,s(θ | x1, y1, . . . , xn, yn)dθ Θ ℓex s ( y | x, θ) pk,s(θ)dθ, where ℓex s is the pdf of likelihood Lex s , pk,s is the pdf of the true posterior Pk,s( | D), pk,s is the pdf of the VI-approximated posterior Pk,s, and y is the output associated to the new input x (see Appendix F for more details). Each predictive distribution P pred k,s is approximated to ˆP pred k,s (e.g. using Normals) via Variational Inference. The convex hull of the collection { ˆP pred k,s }k,s having cardinality K S constitutes the VI-approximated predictive FGCS ˆPpred. Similarly to what we did for ex Ppost, we assume that ex ˆPpred = { ˆP pred k,s }k,s. In Step 5, building on the results in section 2.3, and in particular on Proposition 6, we compute and return AU( ˆPpred) and the bounds for EU( ˆPpred). In particular, we have AU(Ppred) AU( ˆPpred) = H( ˆP pred) (2) and EU(Ppred) EU( ˆPpred) [H( ˆP pred) H( ˆP pred), H( ˆP pred) H( ˆP pred) + log(K S)], (3) where (i) H( ˆP pred) = mink,s H( ˆP pred k,s ); (ii) H( ˆP pred) = maxk,s H( ˆP pred k,s ); and (iii) Ppred is the true predictive FGCS, that is, the one we would have obtained had we been able to compute the exact posteriors in Step 3 and the exact predictive distributions in Step 4. Let us point out a salient feature of CBDL. The AU and the (bounds for) the EU associated with the VI-approximated predictive FGCS ˆPpred embed uncertainty comparable to an uncountably infinite ensemble of BNNs, i.e., an ensemble of BNNs of cardinality ℵ1, despite the simple and intuitive mathematics over the finite set ex ˆPpred. They are not merely pessimistic results on the uncertainties associated with a finite ensemble of BNNs. We observe that we compute the AU and EU associated with the (VI-approximated) predictive credal set ˆPpred, rather than those related to the (VI-approximated) posterior credal set Ppost. We do so because we are ultimately interested in reporting the uncertainty around the predicted outputs given a new input in the problem at hand, more than the uncertainty on the parameters of the NN. Before commenting on the next step, let us pause here and add a remark. While a bad choice of priors and likelihoods may lead to large values of upper and lower entropy H( ˆP pred) and H( ˆP pred), respectively this is not a risk confined to our procedure. Poor modeling choices are an unavoidable risk in model-based techniques. This gave rise to the famous adage by George Box essentially, all models are wrong but some are useful (Box, 1976).16 We maintain that our method is indeed useful, since it overcomes some of the shortcomings of traditional Bayesian techniques as explained in Appendices A and B. As for regular Bayesian methods, though, for our approach too the designer will need to make plausible choices for priors 16Curiously, a similar motivation was brought forward by Pseudo-Dionysius the Areopagite in favor of the use of sacred images in the Christian tradition (Migne, 1857). While they do not capture the essence of God, these defective approximations help the believer s thought to elevate. Published in Transactions on Machine Learning Research (10/2024) and likelihoods. A further positive aspect of working with credal sets is that they are able to self-regulate . That is, in the case of prior-likelihood conflict which happens e.g. if the prior set is ill-specified (Reimherr et al., 2021) the posterior credal set, and in turn the predictive credal set too, will be wider. This will be reflected in the uncertainty measures which will register an excess of posterior, and in turn predictive, epistemic uncertainties. Finally, in Step 6, we compute and return the α-level Imprecise Highest Density Region IRα( ˆPpred) for the VI-approximated predictive FGCS ˆPpred, which approximates the IHDR for the true predictive FGCS Ppred. It is the smallest subset of Y such that ˆP pred[{ y IRα( ˆPpred)}] 1 α, for all ˆP pred ˆPpred, and for some α [0, 1]. It can be interpreted as the smallest collection of outputs y that have a high probability of being the correct ones for the new input x, according to all the distributions in ˆPpred. Or equivalently, we can say that the correct output for the new input x belongs to IHDR IRα( ˆPpred) with lower probability of at least 1 α, a probabilistic guarantee for the set of outputs generated by our procedure. Notice also that the size of IRα( ˆPpred) is an increasing function of both (predictive) AU and EU. As a consequence, it is related, but it is not equal, to the (predictive) AU the agent faces. If we want to avoid to perform the procedure only to discover that IRα( ˆPpred) is too large , then we can add an AU check after Step 5. This, together with computing IRα( ˆPpred) in a classification setting, is explored in Appendices G and H. As we have seen in section 2.2, we find IRα( ˆPpred) by taking the union of the K S HDRs Rα( ˆP pred k,s ) of the extreme elements ex ˆPpred of ˆPpred. The HDR of a well-known distribution (for instance, a Normal) can be routinely obtained in R, e.g. using package HDInterval (Juat et al., 2022). We conclude this section with two remarks. First, we point out how CBDL does not depend on the method used to approximate the posterior and the predictive distributions: Variational Inference can be substituted by other approaches. CBDL can also be easily adapted to other TU, AU, and EU measures, as long as the measure chosen for the total uncertainty is bounded. Second, Step 6 can be effortlessly modified so that CBDL produces the collection of outputs having the highest lower density. This is in line with the Naive Credal Classifier theory (Cozman, 2000a; Zaffalon, 2002). In this case, we forego control on the accuracy level 1 α of the output region produced by CBDL. It is implemented as follows. In the modified version of Step 6, we compute arg max y Y min k,s ˆppred k,s ( y), where ˆppred k,s is the pdf of the VI-approximated predictive distribution ˆP pred k,s . If such arg max is not a singleton, and the user is set on CBDL outputting a unique value y for the new input x, they can select one element uniformly at random from the arg max. 3.2 Theoretical Properties of CBDL Working with credal sets makes CBDL more robust to distribution misspecification and shifts than single BNNs. To see this, we present the following general result, and then we apply it to our case. Let Π be an FGCS as in Remark 2, and consider a probability measure Ψ such that Ψ Π . Proposition 8. Call d any metric and div any divergence on the space of probability measures of interest. Let d(Π , Ψ) := inf P Π d(P , Ψ) and div(Π Ψ) := inf P Π div(P Ψ). Then, for all P Π , d(Π , Ψ) d(P , Ψ) and div(Π , Ψ) div(P Ψ). Proposition 8 holds if Π is any set of probabilities, not just an FGCS.17 In Appendix I, we show that the above result still holds if the elements of Π and Ψ are defined on Euclidean spaces having different dimensions (Cai & Lim, 2022; Caprio, 2022). Let us now apply Proposition 8 to CBDL. Suppose that, when designing a single BNN, an agent chooses likelihood L, while when implementing CBDL, they specify in Step 2 a finite set of likelihoods ex Plik = {Lex s }S s=1, S 2, and then let the induced credal set Plik = Conv(ex Plik) represent their uncertainty around 17More in general, Proposition 8 holds for any type of function f, not just metrics or divergences, because Proposition 8 essentially only applies the definition of the infimum. Published in Transactions on Machine Learning Research (10/2024) the sampling model. Assume that L ex Plik. This means that when designing the single BNN, the agent chooses arbitrarily which of the elements of ex Plik to use. Suppose also that the oracle data generating process Lo is different from L, Lo = L, so that we are actually in the presence of likelihood misspecification. Then, we have two cases. (1) If the true sampling model Lo belongs to Plik, then the distance measured via a metric or a divergence between Plik and Lo is 0, while that between L and Lo is positive. (2) If Lo Plik, then the distance between Plik and Lo is no larger than the distance between L and Lo, no matter (i) which metric or distance we use (Proposition 8), (ii) whether or not Lo and the elements of Plik are defined on the same Euclidean space (Appendix I, Lemma 17). A visual representation is given in Figure 5. A similar argument holds if part of the training dataset D is generated by one distribution Lo 1, and the remaining part is generated by Lo 2, a situation that we can categorize as distribution shift. Even if in the case of a single BNN the user were able to elicit exactly L = Lo 1, the likelihood would still be misspecified for the data that come from Lo 2. This shortcoming is overcome by a credal set approach, in which either both oracle distributions Lo 1 and Lo 2 belong to the credal set Plik, or the distance between Plik and either of them is no larger than that between the single chosen one L, and Lo 1 and Lo 2. Lo d(𝒫lik, Lo) Figure 5: CBDL is more robust to distribution shifts than single BNNs. Here Plik is the convex hull of five plausible likelihoods, and d denotes a generic metric on the space Y of probabilities on Y. We see how d(Plik, Lo) < d(L, Lo); if we replace metric d by a generic divergence div, the inequality would still hold. Let us add a remark here. Assume that the oracle prior P o is in the prior credal set Pprior and that the oracle likelihood Lo is in the likelihood credal set Plik. Then, it is immediate to see that the oracle posterior P o( | D) belongs to the posterior credal set Ppost. Naturally, this does not imply that the posterior credal set collapses to P o( | D). In general, it is unlikely that a finite amount of data is able to completely annihilate all the epistemic uncertainty faced by the agent. What may happen is that if the training set is large enough, Ppost may be inscribed in a ball of small radius around P o( | D). This does not mean that we suffer from under-confidence due to larger-than-necessary epistemic uncertainty. Rather, the relative epistemic uncertainty (measured by the difference between prior and posterior uncertainty, divided by the prior uncertainty) drops significantly. In addition, working with sets of prior and likelihoods allows us to hedge against prior and likelihood misspecification, a consequence of Proposition 8. 4 Experiments From the previous sections, it is clear that CBDL improves on the uncertainty quantification capabilities of single BNNs.18 CBDL allows a better quantification of predictive AU because of its robustness to misspecification stemming from Proposition 8. In a sense, the predictive AU quantified by a single BNN is a function of the choices of prior and likelihood made by the user. In addition, as we pointed out before (e.g. in the last paragraph of section 2.1), predictive EU cannot be obtained in a theoretically principled manner from a single BNN (Hüllermeier & Waegeman, 2021; Fellaji & Pennerath, 2024). Here, we show that these theoretical arguments are also backed by empirical evidence. In particular, in section 4.1 we compare CBDL to the method proposed in Krishnan & Tickoo (2020), in which a single Bayesian Neural Network s output is utilized to estimate both predictive epistemic and aleatoric uncertainties. We also show that CBDL improves on the ensemble of BNNs (EBNN) proposed in Cobb et al. (2019), that we 18This includes empirical Bayes methodologies (Krishnan et al., 2020). Published in Transactions on Machine Learning Research (10/2024) treat as an ablation that is a naive extension of (Krishnan & Tickoo, 2020) to multiple BNNs. In section 4.2, we analyze the downstream task performance of CBDL. We show that it performs better than an ensemble of BNNs (EBNN) (Cobb et al., 2019). To demonstrate the utility of our method, we study the behavior of certain safety-critical settings under distribution shifts and its ramifications. One, for motion prediction in autonomous driving scenarios, and two, to model blood glucose and insulin dynamics for artificial pancreas control. 4.1 (Predictive) Uncertainty Quantification Distribution shifts can introduce uncertainties in a system, which in turn can render the predictions meaningless. This can be due to naturally occurring corruptions, as introduced in Hendrycks & Dietterich (2019) for image datasets. The authors introduced 18 different noise types, which can be varied across 5 different severity levels, ranging from low (severity = 1), to medium (severity = 2, 3), and high (severity = 4, 5). The intuition is that in the current context, increasing the noise severity should generally result in higher uncertainty. We evaluate our CBDL method on 4 standard image datasets, CIFAR-10 (Krizhevsky et al., 2009), SVHN (Netzer et al., 2011), Fashion-MNIST (Xiao et al., 2017), and MNIST (Lecun et al., 1998). We use a slightly different set of perturbations than those introduced in Mu & Gilmer (2019) for gray-scale images like MNIST and Fashion-MNIST. Additionally, we perform cross domain testing for each dataset, where we expect the predictive uncertainties to be higher. We implement and train a Resnet-20 Bayesian Neural Network model inside the library Bayesian-torch (Krishnan et al., 2019). For each dataset, we train 4 different networks initialized with different seeds on the prior and with the same architecture. This corresponds to eliciting a prior FGCS Pprior such that ex Pprior = {P ex 1 , . . . , P ex 4 }, so that K in Step 1 of Algorithm 1 is equal to 4, and a likelihood FGCS that is a singleton, Plik = ex Plik = {L}, so that S in Step 2 of Algorithm 1 is equal to 1. We use a learning-rate of 0.001, batch-size of 128, and train the networks using Mean-Field Variational Inference for 200 epochs. The inference is carried out by performing multiple forward passes through parameters drawn from the posterior distribution. We used 20 Monte-Carlo samples in the experiments. Baselines. In Krishnan & Tickoo (2020), for a single BNN output, the predictive distribution is obtained through multiple stochastic forward passes on the network while sampling from the weight posteriors using Monte Carlo estimators. In their work, they define the overall entropy and the entropy of the predictive distribution as the predictive entropy (Krishnan & Tickoo, 2020, (C.2)). This quantity captures a combination of predictive aleatoric and epistemic uncertainties. Additionally, they define the mutual information between weight posterior and predictive distribution as the predictive epistemic uncertainty. Finally, the predictive aleatoric uncertainty is defined as the expected entropy. The predictive epistemic and aleatoric uncertainties sum to the predictive entropy. In our experiments, we use these definitions to compute the respective quantities. In Cobb et al. (2019), the authors consider different BNNs, but instead of keeping them separate and use them to build a predictive credal set, they average them out. Similar to theirs, we elicit the following procedure, that we call ensemble of BNNs (EBNN). Consider R N 2 different BNNs, and compute the posterior distribution on the parameters. They induce R predictive distributions on the output space Y, each having mean µr and variance σ2 r, r {1, . . . , R}. We call EBNN distribution Pens a Normal having mean µens = 1/R PR r=1 µr and covariance matrix σ2 ens I, where σ2 ens = 1/R PR r=1 σ2 r + 1/(R 1) PR r=1(µr µens)2. In section 4.2, we use the α-level HDR Rα(Pens) associated with Pens as a baseline for the IHDR IRα( ˆPpred) computed at Step 6 of Algorithm 1. Following Cobb et al. (2019), for EBNN we posit that 1/R PR r=1 σ2 r captures the (predictive) aleatoric uncertainty associated with Pens, and 1/(R 1) PR r=1(µr µens)2 captures the (predictive) epistemic uncertainty associated with Pens.19 For CBDL, we look at the value of AU( ˆPpred) = H( ˆP pred) from equation 2, and at the lower bound H( ˆP pred) H( ˆP pred) for EU( ˆPpred) from equation 3. It is enough to focus on the lower 19Notice that in this case we retain the assumption that the EBNN distribution Pens on Y has mean µens and covariance matrix σ2 ens I, but we do not require that it is a Normal. That is because in the four image datasets that we consider, the output space Y is finite. Published in Transactions on Machine Learning Research (10/2024) bound because the upper bound is given by H( ˆP pred) H( ˆP pred) + log(4), and log(4) 0.6 is a fixed value. Hence, the trend of both lower and upper bounds for EU( ˆPpred) as the severity of corruption changes is the same. Summary of our Results. Two scenarios are evaluated: an In-Distribution (ID) scenario with corruption by noises of increasing severity, and an Out-Of-Distribution (OOD) assessment. The following main points summarize our findings: 1. (ID Evaluation) For both our proposed method and that of Krishnan & Tickoo (2020), for increasing severities of noise corruption, predictive epistemic and aleatoric uncertainties increase. In comparison, the baseline EBNN method demonstrates a counterintuitive result, where the predictive aleatoric uncertainty decreases as the severity of corruption increases. The table of values can be found in Appendix N. 2. (ID Evaluation) To assess the utility of our proposed method in downstream tasks, we analyze the accuracy versus rejection rate for all methods. Table 3 summarizes the results. Across MNIST-C and Fashion-MNIST-C, CBDL exhibits a higher average accuracy across the range of rejection rates. For the remaining datasets, it demonstrates results comparable to the baseline (see section 4.1.1 for evaluation details) 3. (OOD Evaluation) To evaluate OOD detection, models trained on a particular dataset are tested on the other three datasets with the goal of analyzing the behavior of the predictive uncertainties. Tables 4 and 5 summarize the results. Across all the evaluated datasets, CBDL had the most consistent behavior of exhibiting a large increase in predictive AU, relative to the increase in predictive EU when tested on an OOD dataset. In comparison, EBNN shows a decrease in predictive AU, and the single BNN shows inconsistent behavior depending on the dataset. CIFAR-10C MNIST-C Fashion MNIST-C SVHN Severity BNN Ensemble CBDL BNN Ensemble CBDL BNN Ensemble CBDL BNN Ensemble CBDL 1 0.754 0.983 0.982 0.645 0.919 0.929 0.525 0.730 0.748 0.698 0.913 0.909 2 0.761 0.947 0.946 0.680 0.879 0.896 0.537 0.655 0.641 0.703 0.843 0.835 3 0.749 0.906 0.907 0.617 0.804 0.841 0.442 0.498 0.529 0.669 0.764 0.758 4 0.733 0.838 0.836 0.639 0.713 0.737 0.332 0.421 0.489 0.608 0.586 0.588 5 0.661 0.734 0.734 0.404 0.495 0.539 0.300 0.351 0.391 0.455 0.378 0.385 Mean 0.732 0.882 0.881 0.597 0.762 0.788 0.433 0.531 0.560 0.627 0.697 0.695 Table 3: In-Distribution evaluation on noise corruptions of increasing severity. Average area under accuracy vs rejection rate curve. Higher numbers indicate overall better accuracy over a range of rejection rates. 4.1.1 In-distribution Evaluation Overall, CBDL demonstrates the most consistent behavior in terms of increasing levels of predictive epistemic and aleatoric uncertainties as the severity of noise corruption increases. The full table of results can be found in Appendix N. In our experiments, for CIFAR-10 the single BNN also exhibits similarly consistent behavior. In MNIST and Fashion-MNIST, the BNN seems to be more consistent than CBDL, whereas for SVHN it seems to show a more inconsistent behavior. These results allow us to conclude that CBDL is comparable if not better than BNN in terms of In-Distribution behavior. Accuracy vs Rejection Rate. As the quantities pertaining to different approaches are not comparable, we perform an additional evaluation comparing the accuracy vs the rejection rate of a given method. In this task, for a given uncertainty threshold, if the uncertainty quantity exceeds such a threshold, then the prediction is rejected. Of those not rejected, the accuracy is computed. A range of uncertainty thresholds is tested for each method and plotted. Figure 6 depicts an example of the plot. There, for lower severities (e.g. 1-2) of Gaussian blur noise, CBDL exhibits better performance and converges to 100% accuracy at a lower rejection rate. Intuitively, for the same range of rejection rates, a higher curve signifies better performance, as less samples are rejected while achieving a higher accuracy. However, for higher severities (above 3), the single BNN seems to show better performance. Even so, the overall rejection rate is high to achieve such accuracy. The entirety of the results can be found in Appendix P. To quantify the differences, the areas Published in Transactions on Machine Learning Research (10/2024) CBDL Ensemble BNN Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric In Dist-Clean 0.102 0.031 0.008 0.076 0.151 0.179 MNIST 0.184 (1.801) 0.142 (4.626) 0.021 (2.723) 0.044 (0.572) 0.445 (2.95) 0.772 (4.312) Fashion MNIST 0.183 (1.791) 0.147 (4.808) 0.022 (2.822) 0.042 (0.557) 0.431 (2.861) 0.811 (4.53) SVHN 0.183 (1.789) 0.141 (4.606) 0.019 (2.421) 0.046 (0.608) 0.394 (2.612) 0.647 (3.616) Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric In Dist-Clean 0.188 0.057 0.012 0.063 0.198 0.071 CIFAR 0.185 (0.986) 0.167 (2.94) 0.023 (1.948) 0.042 (0.67) 0.821 (4.155) 0.372 (5.124) Fashion MNIST 0.162 (0.866) 0.187 (3.287) 0.019 (1.618) 0.035 (0.564) 0.919 (4.654) 0.461 (6.496) SVHN 0.190 (1.015) 0.168 (2.963) 0.022 (1.847) 0.043 (0.69) 0.923 (4.674) 0.297 (4.183) Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric Fashion MNIST In Dist-Clean 0.121 0.029 0.007 0.075 0.197 0.351 CIFAR 0.205 (1.705) 0.104 (3.561) 0.022 (3.178) 0.050 (0.667) 0.408 (2.073) 0.221 (0.629) MNIST 0.165 (1.367) 0.182 (6.192) 0.021 (3.069) 0.033 (0.443) 0.909 (4.618) 0.328 (0.932) SVHN 0.218 (1.81) 0.109 (3.742) 0.026 (3.827) 0.046 (0.618) 0.598 (3.036) 0.305 (0.867) Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric In Dist-Clean 0.061 0.01 0.004 0.083 0.097 0.107 CIFAR 0.181 (2.992) 0.121 (11.778) 0.026 (6.029) 0.040 (0.479) 0.655 (6.765) 0.716(6.69) MNIST 0.198 (3.272) 0.063 (6.154) 0.024 (5.495) 0.056 (0.675) 0.329 (3.393) 0.336 (3.139) Fashion MNIST 0.199 (3.288) 0.113 (10.984) 0.026 (6.002) 0.041 (0.491) 0.103 (6.375) 0.125 (6.178) Table 4: Out-Of-Distribution evaluation. Across all datasets, when tested on datasets outside of the training dataset, CBDL demonstrated the most consistent behavior in terms of predictive uncertainty quantification. The quantity in the parenthesis is the magnitude of increase relative to the quantity when tested on the clean dataset. CBDL Ensemble BNN Epi. AUROC Alea. AUROC Epi. AUROC Alea. AUROC Epi. AUROC Alea. AUROC MNIST 0.905 0.826 0.805 0.104 0.860 0.929 Fashion MNIST 0.920 0.836 0.816 0.093 0.808 0.937 SVHN 0.851 0.823 0.789 0.117 0.826 0.909 CBDL Ensemble BNN Epi. AUROC Alea. AUROC Epi. AUROC Alea. AUROC Epi. AUROC Alea. AUROC CIFAR 0.930 0.912 0.827 0.040 0.835 0.882 Fashion MNIST 0.956 0.935 0.804 0.030 0.929 0.960 SVHN 0.971 0.926 0.808 0.033 0.874 0.898 CBDL Ensemble BNN Epi. AUROC Alea. AUROC Epi. AUROC Alea. AUROC Epi. AUROC Alea. AUROC Fashion MNIST CIFAR 0.933 0.724 0.862 0.121 0.703 0.770 MNIST 0.925 0.881 0.770 0.085 0.962 0.933 SVHN 0.972 0.830 0.865 0.066 0.855 0.904 CBDL Ensemble BNN Epi. AUROC Alea. AUROC Epi. AUROC Alea. AUROC Epi. AUROC Alea. AUROC CIFAR 0.979 0.941 0.890 0.014 0.964 0.952 MNIST 0.781 0.679 0.777 0.205 0.728 0.709 Fashion MNIST 0.967 0.881 0.899 0.030 0.921 0.917 Table 5: Out-Of-Distribution evaluation. AUROC for OOD detection is computed using both aleatoric and epistemic predictive uncertainty measures of the different approaches. Each model is trained on the training partition of the dataset and tested on the respective testing set of the OOD dataset. When using predictive AU, a single BNN performs relatively better than CBDL for OOD detection, though CBDL is still comparable and outperforms the single BNN when trained on Fashion-MNIST. When using predictive EU, CBDL outperforms the other approaches on most datasets. under the curve are computed and averaged for a given noise severity level. Finally, this quantity is averaged across all severity levels. The results are shown in Table 3. As we can see, when EBNN outperforms CBDL in specific instances, the difference is minor, and when the opposite happens, the difference is significant. Over all datasets, CBDL outperforms a single BNN. Published in Transactions on Machine Learning Research (10/2024) Figure 6: Accuracy vs Rejection Rate - CIFAR10C Gaussian blur. For lower severities (e.g. 1-2) of Gaussian blur noise, CBDL exhibits better performance and converges to 100% accuracy at a lower rejection rate. For higher severities (above 3), the single BNN seems to show better performance. Even so, the overall rejection rate is high to achieve such accuracy. The ensemble method (ENS) exhibits comparable performance in this specific instance. 4.1.2 Out-of-distribution Evaluation In this final evaluation, we analyze the behavior of the three approaches (CBDL, EBNN, and BNN) when tested on an OOD dataset. For a given dataset, for example CIFAR, a model is trained on the dataset, and evaluated on the remaining datasets, i.e. MNIST, SVHN, and Fashion-MNIST. The relative increase in the average predictive EU and AU is computed compared to when an approach is tested on the In-Distribution testing set. The In-Distribution uncertainty quantities for each dataset are shown in Table 4. The relative magnitude of the increase in each quantity is shown in the parenthesis for each respective OOD dataset. The baseline EBNN exhibits a counterintuitive behavior, viz. a decrease in predictive AU. Although this behavior is consistent across the different dataset combinations, this would be an undesirable quality for OOD detection scenarios. The single BNN exhibits inconsistent behavior in terms of the relative increase or decrease in magnitude. For CIFAR-10, the predictive AU increases by a magnitude larger than the predictive EU, while for SVHN the increase is smaller. Moreover, for Fashion-MNIST the predictive AU decreases, and for MNIST the behavior is inconsistent. Only our proposed CBDL method demonstrates a stable, consistent increase in predictive AU, which is greater in magnitude relative to the increase in predictive EU. These results demonstrate a clear improvement in performance by our proposed approach. We also give results pertaining AUROCs for the OOD detection performance, presented in Table 5. Each group of results shows which dataset the models are trained on, and which they are tested on. We report AUROC when using both the predictive EU and AU of each approach. In general, the baseline EBNN method performs the worst in all cases. When using predictive AU, a single BNN performs relatively better than CBDL for OOD detection, though CBDL is still comparable and outperforms it when trained on Fashion MNIST. When using predictive EU, CBDL clearly outperforms the other approaches on most datasets. This is in line with other works where predictive EU is considered important for OOD detection (Kendall & Gal, 2017). We conjecture that this may be due to the fact that a single BNN is not able to gauge predictive EU properly. Hence, the single BNN flags an instance as OOD when it comes from the tail of the distribution; this is well-captured by (predictive) aleatoric uncertainty. On the contrary, CBDL is able to gauge predictive EU properly, and hence it is able to capture when OOD happens by looking at the disagreement between the elements of the predictive credal set, captured by the difference between upper and lower entropy. Published in Transactions on Machine Learning Research (10/2024) 4.2 Downstream Tasks Performance As we have shown in the previous section, CBDL is better than single BNNs and ensemble of BNNs at quantifying and disentangling predictive AU and EU. In this section, we show with two applications motion prediction in autonomous driving scenarios, and blood glucose and insulin dynamics for artificial pancreas control that CBDL has better downstream tasks capability than EBNN. We do not compare CBDL against belief tracking techniques because these latter require extra assumptions that CBDL do not, as we further expand on in Appendix K. 4.2.1 Motion Prediction for Autonomous Racing In this case study, we demonstrate the utility of CBDL for motion prediction in autonomous driving scenarios. An important challenge in autonomous driving is understanding the intent of other agents and predicting their future trajectories to allow for safety-aware planning. In autonomous racing, where control is pushed to the dynamical limits, accurate and robust predictions are even more essential for outperforming opponent agents while assuring safety. CBDL provides a straightforward method for quantifying uncertainty and deriving robust prediction regions for anticipating an agent s behavior. We use the problem settings in Tumu et al. (2023) to define the problem of obtaining prediction sets for future positions of an autonomous racing agent. Our results show that the prediction regions have improved coverage when compared to EBNN. These results hold in both In-Distribution and Out-Of-Distribution settings, which are described below. Problem. Let Oi(t, l) Oi = {πi t l, . . . , πi t} denote the i-th trajectory instance of an agent at time t, consisting of the observed positions from time t l up to time t. Let then Ci be a time-invariant context variable. Let also F i(t, h) F i = {πi t+1, . . . , πi t+h} be the collection of the next h future positions. We wish to obtain a model M that predicts region Rα with probabilistic guarantees. In particular, for EBNN Rα is the α-level HDR Rα(Pens) of Pens, so that Pens[F i Rα(Pens)] 1 α, while for CBDL Rα = IRα( ˆPpred), so pred[F i IRα( ˆPpred)] 1 α, or equivalently, ˆP pred[F i IRα( ˆPpred)] 1 α, for all ˆP pred ˆPpred. The dataset consists of instances of (Oi, F i) divided into a training set Dtrain and a testing set Dtest. We train an uncertainty-aware model on Dtrain that computes the triplet (F i l , F i m, F i u) = M(Oi, Ci) where F i l , F i u, F i m are the lower, upper, and mean predictions of the future positions, respectively. The dataset Dall is created by collecting simulated trajectories of autonomous race cars in the F1Tenth-Gym (O Kelly et al., 2020); for details, see Tumu et al. (2023). As shown in Figure 7, different racing lines were utilized including the center, right, left, and optimal racing line for the Spielberg track. Figure 7: Motion Prediction for F1Tenth-Gym Environment (O Kelly et al., 2020). Data is collected by simulating various racing lines on the Spielberg Track. Published in Transactions on Machine Learning Research (10/2024) In-Distribution Results Ensemble CBDL 1 α 0.9 0.95 0.99 0.9 0.95 0.99 One-step 0.962 0.980 0.992 0.992 0.995 0.997 Multi-step 0.638 0.826 0.937 0.914 0.948 0.979 Out-of-Distribution Results Ensemble CBDL 1 α 0.9 0.95 0.99 0.9 0.95 0.99 One-step 0.919 0.950 0.980 0.979 0.988 0.995 Multi-step 0.532 0.703 0.860 0.825 0.884 0.943 Table 6: F1Tenth coverage results. We report one-step coverage and multi-step coverage across 3 different values of α. CBDL exceed coverage of EBNNs in all settings. We denote these by Dcenter, Dright, Dleft, and Drace, respectively. Position π is a vector π = (a, b, ϑ, v) , where a and b are coordinates in a 2-dimensional Euclidean space, and ϑ and v are the heading and speed, respectively. In total, the Dall consists of 34686 train instances, 4336 validation instances, and 4336 test instances. In-Distribution vs Out-Of-Distribution. We consider the prediction task to be In-Distribution when Dtrain, Dtest Dall. It is Out-Of-Distribution (OOD) when Dtrain Dcenter Dright Dleft and Dtest Drace. Metrics. We train the ensemble of BNNs and the CBDL models, Mens and MCBDL respectively, using the same architecture and different seeds. As for section 4.1, for CBDL this corresponds to having a nonsingleton prior FGCS, and a singleton likelihood FGCS. We compare the performance with respect to the test set by computing the single-step coverage, where each prediction time-step is treated independently, and the multi-step coverage, which considers the entire h-step prediction. Figure 8.(a) depicts a sample of the In-Distribution evaluation for each of the models. For a given trajectory, the red boxes indicate when the prediction region did not cover the actual trajectory at that time-step. Qualitatively, MCBDL has less missed time steps when compared to Mens. Table 6 shows that CBDL performs better in terms of both one-step and multi-step coverage. Similar results can be observed for the OOD scenario. There, all models were trained on racing lines which are predominantly parallel to the track curvature. As a consequence, when the test set consists of instances with higher curvatures, the overall coverage of all models degrades. This can be seen in Figure 8.(b), where the prediction of the models (orange) tends to be straight while the actual trajectory is more curved (green). Despite this, the figure and the coverage metrics in Table 6 show how CBDL exhibits a more robust behavior. 4.2.2 Artificial Pancreas Control Overall Setup. In this next case study we consider the problem of data-driven control of human blood glucose-insulin dynamics, using an artificial pancreas system, see Figure 9. External insulin delivery is accomplished by using an insulin pump controlled by the artificial pancreas software, which attempts to regulate the blood-glucose (BG) level of the patient within the euglycemic range of [70, 180] mg/dl (Kushner et al., 2018). Levels below 70 mg/dl lead to hypoglycemia, which can lead to loss of consciousness, coma or even death. On the other hand, levels above 300 mg/dl lead to a condition called ketoacidosis, where the body can break down fat due to lack of insulin, and lead to build up of ketones. In order to treat this situation, patients receive external insulin delivery through insulin pumps. Artificial Pancreas (AP) systems can remedy this situation by measuring the blood glucose level, Published in Transactions on Machine Learning Research (10/2024) Figure 8: In both pictures, the red boxes indicate when the prediction region did not cover the actual trajectory at that time-step. Left: F1Tenth In-Distribution results. Given an input of past observations, CBDL exhibits better coverage of the future target trajectory. Predictions which do not cover the target within the desired 1 α level are indicated in red. Right: F1Tenth Out-Of-Distribution (OOD) results. Robust performance is exhibited by CBDL when compared to EBNN in OOD settings. Figure 9: The Bayesian Neural Networks predict a future blood glucose value. These individual predictions are combined to get a robust estimate of the true value as an interval. This is used by the Model Predictive Control (MPC) algorithm to recommend insulin dosage for the patient. The patient block in our experiment is simulated using the virtual patient models from the UVa-Padova simulator. and automatically injecting insulin into the blood stream. Thus, we define the unsafe regions of the space as G(t) ( , 70) (300, ), where G(t) is the BG value at time t. This is the shaded region in Figure 10. Figure 10: Starting from an initial glucose value, the task of the artificial pancreas controller is to maintain the blood glucose value within safe operating limits using insulin as a mode of control. Neural Network Models and Controller. Deep Neural Networks are effective in capturing the BGinsulin dynamics for personalized medical devices (Kushner et al., 2018). This allows for improved device performance. Even though standard Feedforward Neural Networks can be used, Bayesian Neural Networks Published in Transactions on Machine Learning Research (10/2024) (BNNs), and especially a collection of multiple BNNs, offer a better alternative towards uncertainty aware predictions. Here, we test the ramifications of these prediction sets, when used inside an online receding horizon control scheme for insulin delivery. We use the standard MPC control scheme for this purpose, well-known in the literature (Dutta et al., 2018). More formally, let G(t) and I(t) It be the bloodglucose and insulin values at time t, respectively. We denote the finite length trajectory of length H as G H(t) := [G(t H + 1), . . . , G(t)], and I H(t) := [I(t H + 1), . . . , I(t)]. An uncertainty aware model M computes the triplet (Gl(t + l), Gm(t + l), Gu(t + l)) = M( G H(t), I H(t)), where Gm is the mean prediction output, and Gl, Gu are the lower and upper predictions of the glucose value, respectively. By design, it is true that Gl Gm Gu. A model predictive control algorithm whose cost function we denote by J solves arg min I0,I1,...,Ik 1 Pk 1 i=0 J(M( G H(t + i), I H(t + i))). After every time step, the control algorithm picks the first insulin input I0 as the insulin bolus for the patient, and discards the rest. Cost function J takes into account three factors, (i) Distance of the mean prediction level Gm at each time step from a target value of 120 mg/dl, (ii) Distance of upper and lower predictions (Gu and Gl) from the unsafe regions of the state space G(t) > 300 and G(t) < 70, and (iii) Total insulin injected Pk 1 t=0 It. Starting with some initial glucose value G(0), we measure the performance of the artificial pancreas controller as the fraction of time it spends in the unsafe regions, tunsafe = 1 t=1 1 {G(t) ( , 70) (300, )} , where 1{ } denotes the indicator function. A lower value is more desirable. We compare EBNN and CBDL as different realizations of the model M. Distribution Shift using Meals. A well known problem with learned models is distribution shift. Bayesian Neural Networks can address this issue by apprising the end user of the increased uncertainty. For regression models of the type described above, this appears as larger prediction intervals [Gl, Gu]. The artificial pancreas controller can run into this situation in the following way: the insulin-glucose time series data collected for training the data-driven model M can be without meals, while at test time the patient can have meals. This creates a distribution shift between the training and test time data. Fortunately, the UVa-Padova simulator (Dalla Man et al., 2013) allows us to create datasets with and without meal inputs. In this case study, the training data was obtained by randomly initializing the BG value in the range [120, 190], and simulating the patient for 720 minutes. The controller was executed at 5 minutes intervals. At test time the patient was supplied meals at specific time intervals (for details, see Appendix J). This creates a significant distribution shift since meals are effectively an unknown variable which can affect the system state. However, from the controller s perspective this is practical, since patients can have unannounced meals. Results and Discussion. To capture the difference in performance between EBNN and CBDL, we compute Perfdiff := (t EBNN unsafe t CBDL unsafe)/t EBNN unsafe. Both t EBNN unsafe and t CBDL unsafe depend on interval [Gl, Gu]; for EBNN, this latter corresponds to the α-level HDR Rα(Pens) associated with EBNN distribution Pens, while for CBDL it corresponds to the IHDR IRα( ˆPpred). We consider one case in which CBDL is trained using a credal prior set and only one likelihood (we choose different seeds which initialize the prior distributions but we keep the same architecture for the BNNs), and another case in which we do the opposite (we use the same seed and different architectures). We report Perfdiff, across different choices in Table 7. We observe that for lower values of α (or equivalently, for larger values of 1 α, e.g. 0.95 and 0.99) the gains of CBDL are more pronounced. This means that when larger significance levels need to be ensured, CBDL is to be preferred to ensemble of BNNs. As discussed before, the CBDL procedure considers all the infinitely many possible priors that can be expressed as a convex combination of the priors that the user specifies at the beginning of the analysis. The same holds for the likelihoods. While this results in a more conservative estimate as compared to the EBNN framework, CBDL produces controllers which respect the safety limits better. To see that CBDL is more conservative than EBNN, notice that when combining predictive distributions from multiple BNNs, CBDL combines the predictions via a finitely generated credal set (FGCS) whose extrema are the individual distributions. On Published in Transactions on Machine Learning Research (10/2024) the contrary, an EBNN takes an average of the individual distributions to compute the ensemble distribution Pens. The union of the HDRs of the predictive distributions is more conservative (i.e., broader) than the HDR of the single ensemble distribution. While the approach by EBNN seems like a reasonable choice on the surface, it falls short in capturing the uncertainty necessary for the downstream task. For more details on this case study, see Appendix J. 1 α 0.9 0.95 0.99 Varying Seeds -5.8% 5.5% 0.6% Varying Architectures -6.6% 0.9% 0.3% Table 7: We report the performance improvements when using IBNNs as compared to EBNNs across 3 different values of α. Row 1 corresponds to the case where the individual BNNs are trained with different seeds for the prior distribution; and Row 2 is the case when the BNNs have different architectures. 5 Related Work In Corani et al. (2012), the authors introduce credal classifiers (CCs) as a generalization of classifiers based on Bayesian networks. Unlike CCs, CBDL does not require independence assumptions between non-descendant, non-parent variables. In addition, CBDL avoids NP-hard complexity issues of searching for optimal structure in the space of Bayesian networks (Chickering et al., 2004). In Manchingal & Cuzzolin (2022), an epistemic convolutional neural network (ECNN) is developed that explicitly models the epistemic uncertainty induced by training data of limited size and quality. A clear distinction is that ECNNs measure uncertainty in targetlevel representations whereas CBDL identifies the uncertainty measure on the output space Y. Despite the merit of their work, we believe CBDL achieves greater generality, since it is able to quantify both aleatoric and epistemic predictive uncertainties, and is applicable to problems beyond classification. For a review of the state of the art concerning the distinction between EU and AU we refer to Hüllermeier & Waegeman (2021) and to Manchingal & Cuzzolin (2022). We also point out how CBDL has been recently used to solve prior-likelihood conflicts in Bayesian statistics (Marquardt et al., 2023). Further references can be found in Appendix L. We also point out how there exist other efficient methods which perform approximate Variational Inference via dropouts in deep neural networks (Kendall & Gal, 2017; Gal & Ghahramani, 2016). As mentioned at the end of section 3.1, we can easily adapt CBDL to use such dropout approximations in Steps 3-4 of Algorithm 1. Since our contribution is centered around how different predictions can be combined via an FGCS, we used the de-facto standard for performing inference on BNNs, which is based on off-the-shelf VI techniques. In the future, we plan to study the effect on computational complexity and uncertainty quantification capability of a CBDL procedure that approximates posterior and predictive distributions via dropout. We do not consider Bayesian Model Averaging (BMA) as a baseline for CBDL for two main reasons. First, BMA applied to deep learning needs to implement full batch Hamiltonian Monte Carlo in order to get to the true posterior (Izmailov et al., 2021b). Given the number of parameters in modern deep learning architectures in the order of millions this is realistically possible at an experimental level only to labs with access to industry scale computational resources. In order to be practically relevant, we limit our experiments to the more well-understood realm of Variational Inference on BNNs. In addition, Izmailov et al. (2021a) show the pitfalls of BMA in the context of Bayesian Neural Networks, a further reason not to use a Highest Density Region resulting from BMA as a baseline for CBDL. Such pitfalls are related to the fact that BMA can be seen as a model featuring second-order distributions, i.e. distributions over distributions. In particular, the distribution Q in equation 1 is a second order distribution. These types of models have been recently shown to suffer from major pitfalls when used to quantify predictive EU due to their sensitivity to regularization parameters, and to underestimate predictive AU (Bengs et al., 2022; Pandey & Yu, 2023; Juergens et al., 2024). Published in Transactions on Machine Learning Research (10/2024) 6 Conclusion We presented CBDL, a procedure that can be seen as a non-condensed, uncountably infinite ensemble of BNNs, carried out using only finitely many elements. It allows to distinguish between predictive AU and EU, and to quantify them. We showed how it can be used to specify a set of outputs the IHDR that enjoys probabilistic guarantees. We showed empirically that it improves on the Bayesian state of the art at gauging predictive AU and EU, and we also demonstrated its downstream tasks capabilities. We point out how a region that improves on IRα( ˆPpred), meaning that it would be tighter, is IR α( ˆPpred) = {y Y : ˆppred(y) ˆpα}, where ˆppred := mink,s ˆppred k,s , and ˆpα is the largest constant such that ˆP pred[y IR α( ˆPpred)] 1 α. The problem with IR α is that, while the highest density regions Rα( ˆP pred k,s ) associated with the predictive distributions in ex ˆPpred can be computed using off-the-shelf tools, calculating ˆppred and ˆpα would have been much more computationally expensive. In addition, it would have required to come up with a new technique to find ˆppred and ˆpα. We defer studying this to future work. We also plan to apply CBDL to continual learning to overcome the curse of dimensionality and to capture an agent s preference over the tasks to perform, similarly to Lu et al. (2023), and to active learning, to be able to sample from the regions of the state space exhibiting the highest epistemic uncertainty, similarly to Dutta et al. (2023). Furthermore, we intend to relate CBDL to Bayesian Model Selection (BMS) (Ghosh et al., 2019). This latter suffers from the same problem as regular Bayesian inference. That is, while it tries to come up with a sophisticate prior that induces shrinkage, it still relies on the correctness of that prior, i.e. on correctly specifying the prior s parameters. In the future, an interesting way of combining CBDL with BMS will be to use a finite number of regularized horseshoe priors, as suggested by Ghosh et al. (2019, Section 3.2), as extreme elements of the prior credal set. We also call attention to the fact that CBDL is a model-based approach. The relationship with model-free approaches such as conformal prediction (Shafer & Vovk, 2008) will be the object of future studies. In particular, we are interested in finding in which cases IHDRs are narrower than conformal regions, and vice versa, and which credal sets give rise to IHDRs enjoying the same probabilistic guarantees as conformal regions. Finally, we point out how one possible way of easing the burden of the combinatorial task in Step 3 of Algorithm 1 is to specify a prior credal set whose size strikes the perfect balance between being vague enough so that we do not underestimate the EU, and being small enough so that CBDL is actually implementable. We suspect conjugacy of the priors may play a key role in this endeavor. Because of its centrality, we defer the study of optimal prior credal sets to future work. We also point out how an evidential approach (Amini et al., 2020; Charpentier et al., 2020; Denœux, 2022; 2023; Sensoy et al., 2018) at least in classification problems could allows us to bypass the bottleneck in Algorithm 1 by merging an imprecise probabilistic approach, with some tricks resulting from smartly choosing our priors. A CBDL-adjacent research question of great interest pertaining ensemble learning, then, is how do single components of an ensemble contribute to the quantification of the global predictive EU faced by the agent. When the uncertainty captured by an ensemble is distilled into only one distribution, there could indeed be uncertainty spills , like when one pours water in a glass too hastily, and some finishes on the table instead of in the glass. This remains an unexplored venure in the uncertainty quantification community. Published in Transactions on Machine Learning Research (10/2024) Joaquín Abellán and Serafín Moral. A non-specificity measure for convex sets of probability distributions. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems, 8(3):357 367, 2000. Joaquín Abellán, George Jiří Klir, and Serafín Moral. Disaggregated total uncertainty measure for credal sets. International Journal of General Systems, 1(35):29 44, 2006. Alexander Amini, Wilko Schwarting, Ava Soleimany, and Daniela Rus. Deep evidential regression. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 14927 14937. Curran Associates, Inc., 2020. Thomas Augustin, Frank P.A. Coolen, Gert De Cooman, and Matthias C.M. Troffaes (eds.). Introduction to imprecise probabilities. Wiley Series in Probability and Statistics. John Wiley and Sons, 2014. David Avis and Komei Fukuda. Reverse search for enumeration. Discrete Applied Mathematics, 65:21 46, 1996. Rina Foygel Barber, Emmanuel J. Candès, Aaditya Ramdas, and Ryan J. Tibshirani. Conformal prediction beyond exchangeability. The Annals of Statistics, 51(2):816 845, 2023. doi: 10.1214/23-AOS2276. URL https://doi.org/10.1214/23-AOS2276. Viktor Bengs, Eyke Hüllermeier, and Willem Waegeman. Pitfalls of epistemic uncertainty quantification through loss minimisation. In Alice H. Oh, Alekh Agarwal, Danielle Belgrave, and Kyunghyun Cho (eds.), Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id= epjx T_ARZW5. James O. Berger. The robust Bayesian viewpoint. In Joseph B. Kadane (ed.), Robustness of Bayesian Analyses. Amsterdam : North-Holland, 1984. Jose M. Bernardo. Reference posterior distributions for Bayesian inference. Journal of the Royal Statistical Society: Series B, 41(2):113 128, 1979. George E. P. Box. Science and statistics. Journal of the American Statistical Association, 71(356):791 799, 1976. Andrey G. Bronevich and Igor N. Rozenberg. Generalized Hartley measures on credal sets. In Andrés Cano, Jasper De Bock, Enrique Miranda, and Serafín Moral (eds.), Proceedings of the Twelveth International Symposium on Imprecise Probability: Theories and Applications, volume 147 of Proceedings of Machine Learning Research, pp. 32 41. PMLR, 2021. Yuhang Cai and Lek-Heng Lim. Distances between probability distributions of different dimensions. IEEE Transactions on Information Theory, 68(6):4020 4031, 2022. Michele Caprio. Refined Pinsker s and reverse Pinsker s inequalities for probability distributions of different dimensions. IEEE Access, 10:116425 116431, 2022. Michele Caprio. Imprecise markov semigroups and their ergodicity, 2024. URL https://arxiv.org/abs/ 2405.00081. Michele Caprio and Ruobin Gong. Dynamic precise and imprecise probability kinematics. In Enrique Miranda, Ignacio Montes, Erik Quaeghebeur, and Barbara Vantaggi (eds.), Proceedings of the Thirteenth International Symposium on Imprecise Probability: Theories and Applications, volume 215 of Proceedings of Machine Learning Research, pp. 72 83. PMLR, 11 14 Jul 2023. Michele Caprio and Sayan Mukherjee. Finite admixture models: a bridge with stochastic geometry and choquet theory. Available at ar Xiv:2002.08409, 2023a. Michele Caprio and Sayan Mukherjee. Concentration inequalities and optimal number of layers for stochastic deep neural networks. IEEE Access, 11:38458 38470, 2023b. doi: 10.1109/ACCESS.2023.3268034. Published in Transactions on Machine Learning Research (10/2024) Michele Caprio and Sayan Mukherjee. Ergodic theorems for dynamic imprecise probability kinematics. International Journal of Approximate Reasoning, 152:325 343, 2023c. Michele Caprio and Sayan Mukherjee. Extended probabilities and their application to statistical inference, chapter 2. World Scientific, 2023d. Michele Caprio and Teddy Seidenfeld. Constriction for sets of probabilities. In Enrique Miranda, Ignacio Montes, Erik Quaeghebeur, and Barbara Vantaggi (eds.), Proceedings of the Thirteenth International Symposium on Imprecise Probability: Theories and Applications, volume 215 of Proceedings of Machine Learning Research, pp. 84 95. PMLR, 11 14 Jul 2023. URL https://proceedings.mlr.press/v215/ caprio23b.html. Michele Caprio, Kuk Jin Jang, Souradeep Dutta, Shireen Manchingal, Fabio Cuzzolin, Oleg Sokolsky, and insup Lee. Credal and interval deep evidential classification. Technical Reports of the PRECISE Center, 2024a. Michele Caprio, Yusuf Sale, Eyke Hüllermeier, and Insup Lee. A Novel Bayes Theorem for Upper Probabilities. In Fabio Cuzzolin and Maryam Sultana (eds.), Epistemic Uncertainty in Artificial Intelligence, pp. 1 12, Cham, 2024b. Springer Nature Switzerland. Michele Caprio, Maryam Sultana, Eleni Elia, and Fabio Cuzzolin. Credal learning theory, 2024c. URL https://arxiv.org/abs/2402.00957. Simone Cerreia-Vioglio, Fabio Maccheroni, and Massimo Marinacci. Ergodic theorems for lower probabilities. Proceedings of the American Mathematical Society, 144:3381 3396, 2015. Bertrand Charpentier, Daniel Zügner, and Stephan Günnemann. Posterior network: Uncertainty estimation without OOD samples via density-based pseudo-counts. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 1356 1367. Curran Associates, Inc., 2020. David M. Chickering, David Heckerman, and Christopher Meek. Large-sample learning of Bayesian networks is NP-hard. The Journal of Machine Learning Research, 5:1287 1330, 2004. Adam D. Cobb, Michael D. Himes, Frank Soboczenski, Simone Zorzan, Molly D. O Beirne, Atılım Güneş Baydin, Yarin Gal, Shawn D. Domagal-Goldman, Giada N. Arney, Daniel Angerhausen, and 2018 NASA FDL Astrobiology Team II. An ensemble of Bayesian neural networks for exoplanetary atmospheric retrieval. The Astronomical Journal, 158(1):33, 2019. Frank P. A. Coolen. Imprecise highest density regions related to intervals of measures. Memorandum COSOR, 9254, 1992. Giorgio Corani, Alessandro Antonucci, and Marco Zaffalon. Bayesian Networks with Imprecise Probabilities: Theory and Application to Classification, chapter 4 of Data Mining: Foundations and Intelligent Paradigms: Volume 1: Clustering, Association and Classification, pp. 49 93. Berlin, Germany : Springer, 2012. Fabio G. Cozman. Credal networks. Artificial Intelligence, 120(2):199 233, 2000a. ISSN 0004-3702. doi: https://doi.org/10.1016/S0004-3702(00)00029-1. URL https://www.sciencedirect.com/science/ article/pii/S0004370200000291. Fabio Gagliardi Cozman. Credal networks. Artificial Intelligence, 120:199 233, 2000b. Chiara Dalla Man, Francesco Micheletto, Dayu Lv, Marc D. Breton, Boris Kovatchev, and Claudio Cobelli. The UVA/PADOVA type 1 diabetes simulator: New features. Journal of Diabetes Science and Technology, 8(1):26 34, 2013. George B. Dantzig. Linear Programming and Extensions. Princeton University Press, 1963. URL http: //www.jstor.org/stable/j.ctt1cx3tvg. Published in Transactions on Machine Learning Research (10/2024) Bruno de Finetti. Theory of Probability, volume 1. New York : Wiley, 1974. Bruno de Finetti. Theory of Probability, volume 2. New York : Wiley, 1975. Thierry Denœux. An evidential neural network model for regression based on random fuzzy numbers. In Sylvie Le Hégarat-Mascle, Isabelle Bloch, and Emanuel Aldea (eds.), Belief Functions: Theory and Applications, pp. 57 66, Cham, 2022. Springer International Publishing. Thierry Denœux. Quantifying prediction uncertainty in regression using random fuzzy sets: the ENNreg model. IEEE Transactions on Fuzzy Systems, pp. 1 10, 2023. Stefan Depeweg, Jose-Miguel Hernandez-Lobato, Finale Doshi-Velez, and Steffen Udluft. Decomposition of uncertainty in Bayesian deep learning for efficient and risk-sensitive learning. In International Conference on Machine Learning, pp. 1184 1193. PMLR, 2018. Didier Dubois and Eyke Hüllermeier. Comparing probability measures using possibility theory: A notion of relative peakedness. International Journal of Approximate Reasoning, 45(2):364 385, 2007. Eighth European Conference on Symbolic and Quantitative Approaches to Reasoning with Uncertainty (ECSQARU 2005). Souradeep Dutta, Taisa Kushner, and Sriram Sankaranarayanan. Robust data-driven control of artificial pancreas systems using neural networks. In Milan Češka and David Šafránek (eds.), Computational Methods in Systems Biology, pp. 183 202, Cham, 2018. Springer International Publishing. ISBN 978-3-319-99429-1. Souradeep Dutta, Michele Caprio, Vivian Lin, Matthew Cleaveland, Kuk Jin Jang, Ivan Ruchkin, Oleg Sokolsky, and Insup Lee. Distributionally robust statistical verification with imprecise neural networks. ar Xiv preprint ar Xiv:2308.14815, 2023. Daniel Ellsberg. Risk, ambiguity, and the Savage axioms. The Quarterly Journal of Economics, 75(4): 643 669, 1961. Mohammed Fellaji and Frédéric Pennerath. The epistemic uncertainty hole: an issue of bayesian neural networks. Available at ar Xiv:2407.01985, 2024. Emlyn Flint, Florence Chikurunhe, and Anthony Seymour. Regime-based tactical allocation for equity factors and balanced portfolios. SSRN Electronic Journal, 01 2017. Benoît Fortin, Samir Hachour, and François Delmotte. Multi-target PHD tracking and classification using imprecise likelihoods. International Journal of Approximate Reasoning, 90:17 36, 2017. Vincent Fortuin, Adrià Garriga-Alonso, Mark van der Wilk, and Laurence Aitchison. BNNpriors: A library for bayesian neural network inference with different prior distributions. Software Impacts, 9:100079, 2021. Yarin Gal and Zoubin Ghahramani. Dropout as a bayesian approximation: Representing model uncertainty in deep learning, 2016. Soumya Ghosh, Jiayu Yao, and Finale Doshi-Velez. Model selection in Bayesian neural networks via horseshoe priors. Journal of Machine Learning Research, 20(182):1 46, 2019. URL http://jmlr.org/papers/v20/ 19-236.html. Isaac Gibbs, John J. Cherian, and Emmanuel J. Candès. Conformal prediction with conditional guarantees. Available at ar Xiv:2305.12616, 2023. Itzhak Gilboa and Massimo Marinacci. Ambiguity and the Bayesian paradigm. In Daron Acemoglu, Manuel Arellano, and Eddie Dekel (eds.), Advances in Economics and Econometrics, Tenth World Congress, volume 1. Cambridge : Cambridge University Press, 2013. Ethan Goan and Clinton Fookes. Bayesian Neural Networks: An Introduction and Survey, pp. 45 87. Cham, Switzerland : Springer International Publishing, 2020. Published in Transactions on Machine Learning Research (10/2024) Ruobin Gong and Xiao-Li Meng. Judicious judgment meets unsettling updating: dilation, sure loss, and Simpson s paradox. Statistical Science, 36(2):169 190, 2021. Chuan Guo, Geoff Pleiss, Yu Sun, and Kilian Q Weinberger. On calibration of modern neural networks. In International conference on machine learning, pp. 1321 1330. PMLR, 2017. Dan Hendrycks and Thomas G. Dietterich. Benchmarking neural network robustness to common corruptions and perturbations. Co RR, abs/1903.12261, 2019. Paul Hofman, Yusuf Sale, and Eyke Hüllermeier. Quantifying aleatoric and epistemic uncertainty: A credal approach. In ICML 2024 Workshop on Structured Probabilistic Inference {\&} Generative Modeling, 2024. Peter J. Huber and Elvezio M. Ronchetti. Robust statistics. Wiley Series in Probability and Statistics. Hoboken, New Jersey : Wiley, 2nd edition, 2009. Rob J. Hyndman. Computing and graphing highest density regions. The American Statistician, 50(2): 120 126, 1996. Eyke Hüllermeier and Willem Waegeman. Aleatoric and epistemic uncertainty in machine learning: an introduction to concepts and methods. Machine Learning, 3(110):457 506, 2021. Pavel Izmailov, Patrick Nicholson, Sanae Lotfi, and Andrew G Wilson. Dangers of bayesian model averaging under covariate shift. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 3309 3322. Curran Associates, Inc., 2021a. Pavel Izmailov, Sharad Vikram, Matthew D. Hoffman, and Andrew Gordon Wilson. What are bayesian neural network posteriors really like? Co RR, abs/2104.14421, 2021b. URL https://arxiv.org/abs/ 2104.14421. Pavel Izmailov, Sharad Vikram, Matthew D Hoffman, and Andrew Gordon Gordon Wilson. What are Bayesian neural network posteriors really like? In Marina Meila and Tong Zhang (eds.), Proceedings of the 38th International Conference on Machine Learning, volume 139, pp. 4629 4640. PMLR, 2021c. Harold Jeffreys. An invariant form for the prior probability in estimation problems. Proceedings of the Royal Society A, 186(1007):453 461, 1946. Laurent Valentin Jospin, Hamid Laga, Farid Boussaid, Wray Buntine, and Mohammed Bennamoun. Handson Bayesian neural networks A tutorial for deep learning users. IEEE Computational Intelligence Magazine, 17(2):29 48, 2022. Ngumbang Juat, Mike Meredith, and John Kruschke. Package hdinterval , 2022. URL https://cran. r-project.org/web/packages/HDInterval/HDInterval.pdf. Accessed on May 9, 2023. Mira Juergens, Nis Meinert, Viktor Bengs, Eyke Hüllermeier, and Willem Waegeman. Is epistemic uncertainty faithfully represented by evidential deep learning methods? In Forty-first International Conference on Machine Learning, 2024. URL https://openreview.net/forum?id=mxj B0LIgp T. Sanyam Kapoor, Wesley J Maddox, Pavel Izmailov, and Andrew Gordon Wilson. On uncertainty, tempering, and data augmentation in bayesian classification. ar Xiv preprint ar Xiv:2203.16481, 2022. Ramneet Kaur, Xiayan Ji, Souradeep Dutta, Michele Caprio, Yahan Yang, Elena Bernardis, Oleg Sokolsky, and Insup Lee. Using semantic information for defining and detecting OOD inputs. Available at ar Xiv:2302.11019, 2023. Alex Kendall and Yarin Gal. What uncertainties do we need in bayesian deep learning for computer vision? Advances in neural information processing systems, 30, 2017. Published in Transactions on Machine Learning Research (10/2024) Jang-Hyun Kim, Wonho Choo, and Hyun Oh Song. Puzzle mix: Exploiting saliency and local statistics for optimal mixup. In Hal Daumé III and Aarti Singh (eds.), Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pp. 5275 5285. PMLR, 13 18 Jul 2020. John Klein, Christèle Lecomte, and Pierre Miché. Hierarchical and conditional combination of belief functions induced by visual tracking. International Journal of Approximate Reasoning, 51(4):410 428, 2010. Ranganath Krishnan and Omesh Tickoo. Improving model calibration with accuracy versus uncertainty optimization. In Advances in Neural Information Processing Systems, volume 33, pp. 18237 18248. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/hash/ d3d9446802a44259755d38e6d163e820-Abstract.html. Ranganath Krishnan, Mahesh Subedar, and Omesh Tickoo. MOPED: efficient priors for scalable variational inference in bayesian deep neural networks. Co RR, abs/1906.05323, 2019. Ranganath Krishnan, Mahesh Subedar, and Omesh Tickoo. Specifying weight priors in Bayesian deep neural networks with empirical Bayes. Proceedings of the AAAI Conference on Artificial Intelligence, 34 (04):4477 4484, April 2020. doi: 10.1609/aaai.v34i04.5875. URL https://ojs.aaai.org/index.php/ AAAI/article/view/5875. Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. CIFAR-10 (Canadian Institute for Advanced Research). Available at http://www.cs.toronto.edu/ kriz/cifar.html, 2009. Meelis Kull and Peter A. Flach. Reliability Maps: A Tool to Enhance Probability Estimates and Improve Classification Accuracy. In Toon Calders, Floriana Esposito, Eyke Hüllermeier, and Rosa Meo (eds.), Machine Learning and Knowledge Discovery in Databases, Lecture Notes in Computer Science, pp. 18 33, Berlin, Heidelberg, 2014. Springer. ISBN 978-3-662-44851-9. Taisa Kushner, David Bortz, David M. Maahs, and Sriram Sankaranarayanan. A data-driven approach to artificial pancreas verification and synthesis. In Proceedings of the 9th ACM/IEEE International Conference on Cyber-Physical Systems, ICCPS 18, pp. 242 252. IEEE Press, 2018. ISBN 9781538653012. Jouko Lampinen and Aki Vehtari. Bayesian approach for neural networks - review and case studies. Neural Networks, 4:257 274, 2001. Daniel Lassiter. Representing credal imprecision: from sets of measures to hierarchical Bayesian models. Philosophical Studies, 177(6):1463 1485, 2020. Yann Lecun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. Peter Lenk and Bryan Orme. The value of informative priors in Bayesian inference with sparse data. Journal of Marketing Research, 46(6):832 845, 2009. ISSN 00222437. Isaac Levi. The Enterprise of Knowledge. London, UK : MIT Press, 1980. Vivian Lin, Kuk Jin Jang, Souradeep Dutta, Michele Caprio, Oleg Sokolsky, and Insup Lee. DC4L: Distribution shift recovery via data-driven control for deep learning models. In Alessandro Abate, Mark Cannon, Kostas Margellos, and Antonis Papachristodoulou (eds.), Proceedings of the 6th Annual Learning for Dynamics and Control Conference, volume 242 of Proceedings of Machine Learning Research, pp. 1526 1538. PMLR, 15 17 Jul 2024. URL https://proceedings.mlr.press/v242/lin24b.html. Pengyuan Lu, Michele Caprio, Eric Eaton, and Insup Lee. Zero-shot task preference addressing enabled by imprecise Bayesian continual learning. Available at ar Xiv:2305.14782, 2023. Shireen Kudukkil Manchingal and Fabio Cuzzolin. Epistemic deep learning. Available at arxiv:2206.07609, 2022. Published in Transactions on Machine Learning Research (10/2024) Massimo Marinacci and Luigi Montrucchio. Introduction to the mathematics of ambiguity. In Itzhak Gilboa (ed.), Uncertainty in economic theory: a collection of essays in honor of David Schmeidler s 65th birthday. London : Routledge, 2004. Alexander Marquardt, Julian Rodemann, and Thomas Augustin. An empirical study of prior-data conflicts in Bayesian neural networks. Poster presented at 13th International Symposium on Imprecise Probabilities: Theories and Applications (ISIPTA) 2023, 2023. Jacques Paul Migne. De coelesti hierarchia. In Patrologia Graeca, volume 3. Paris : Imprimerie Catholique, 1857. Enrique Miranda and Ignacio Montes. Centroids of the core of exact capacities: a comparative study. Annals of Operations Research, 321:409 449, 2023. Norman Mu and Justin Gilmer. MNIST-C: A robustness benchmark for computer vision. Co RR, abs/1906.02337, 2019. Bálint Mucsányi, Michael Kirchhof, and Seong Joon Oh. Benchmarking uncertainty disentanglement: Specialized uncertainties for specialized tasks. Available at ar Xiv:2402.19460, 2024. Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y. Ng. Reading digits in natural images with unsupervised feature learning. In NIPS Workshop on Deep Learning and Unsupervised Feature Learning 2011, 2011. Matthew O Kelly, Hongrui Zheng, Dhruv Karthik, and Rahul Mangharam. F1TENTH: An Open-source Evaluation Environment for Continuous Control and Reinforcement Learning. In Proceedings of the Neur IPS 2019 Competition and Demonstration Track, pp. 77 89. PMLR, 2020. Deep Shankar Pandey and Qi Yu. Learn to accumulate evidence from all training samples: Theory and practice. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp. 26963 26989. PMLR, 23 29 Jul 2023. URL https://proceedings.mlr.press/v202/pandey23a.html. Sangdon Park, Osbert Bastani, Nikolai Matni, and Insup Lee. PAC confidence sets for deep neural networks via calibrated prediction. In 8th International Conference on Learning Representations, 2020. Luis Raul Pericchi. Sets of prior probabilites and bayesian robustness. Documentation Section on the website of the Society for Imprecise Probability Theory and Applications (SIPTA), 1998. Matthew Reimherr, Xiao-Li Meng, and Dan L. Nicolae. Prior sample size extensions for assessing prior impact and prior-likelihood discordance. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 83(3):413 437, 2021. doi: https://doi.org/10.1111/rssb.12414. URL https://rss.onlinelibrary. wiley.com/doi/abs/10.1111/rssb.12414. Hippolyt Ritter, Aleksandar Botev, and David Barber. A scalable laplace approximation for neural networks. In International Conference on Learning Representations, 2018. Yusuf Sale, Michele Caprio, and Eyke Hüllermeier. Is the volume of a credal set a good measure for epistemic uncertainty? In Robin J. Evans and Ilya Shpitser (eds.), Proceedings of the Thirty-Ninth Conference on Uncertainty in Artificial Intelligence, volume 216 of Proceedings of Machine Learning Research, pp. 1795 1804. PMLR, 31 Jul 04 Aug 2023. URL https://proceedings.mlr.press/v216/sale23a.html. Yusuf Sale, Viktor Bengs, Michele Caprio, and Eyke Hüllermeier. Second-order uncertainty quantification: A distance-based approach. In Ruslan Salakhutdinov, Zico Kolter, Katherine Heller, Adrian Weller, Nuria Oliver, Jonathan Scarlett, and Felix Berkenkamp (eds.), Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 43060 43076. PMLR, 21 27 Jul 2024. URL https://proceedings.mlr.press/v235/sale24a.html. Published in Transactions on Machine Learning Research (10/2024) Robin Senge, Stefan Bösner, Krzysztof Dembczyński, Jörg Haasenritter, Oliver Hirsch, Norbert Donner Banzhoff, and Eyke Hüllermeier. Reliable classification: Learning classifiers that distinguish aleatoric and epistemic uncertainty. Information Sciences, 255:16 29, January 2014. ISSN 0020-0255. Murat Sensoy, Lance Kaplan, and Melih Kandemir. Evidential deep learning to quantify classification uncertainty. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. Glenn Shafer and Vladimir Vovk. A tutorial on conformal prediction. Journal of Machine Learning Research, 9:371 421, 2008. Marek Smieja and Jacek Tabor. Entropy of the mixture of sources and entropy dimension. IEEE Transactions on Information Theory, 58(5):2719 2728, 2012. Lewis Smith and Yarin Gal. Understanding measures of uncertainty for adversarial example detection. ar Xiv preprint ar Xiv:1803.08533, 2018. Bjørnar Tessem. Interval probability propagation. International Journal of Approximate Reasoning, 7(3): 95 120, 1992. D. Michael Titterington. Bayesian methods for neural networks and related models. Statistical Science, 19 (1):128 139, 2004. Krasymyr Tretiak, Georg Schollmeyer, and Scott Ferson. Neural network model for imprecise regression with interval dependent variables. Available at ar Xiv:2206.02467, 2022. Matthias C.M. Troffaes and Gert de Cooman. Lower Previsions. Chichester, United Kingdom : John Wiley and Sons, 2014. Renukanandan Tumu, Lars Lindemann, Truong Nghiem, and Rahul Mangharam. Physics Constrained Motion Prediction with Uncertainty Quantification, 2023. ar Xiv:2302.01060. Vladimir Vovk, Alexander Gammerman, and Glenn Shafer. Algorithmic Learning in a Random World. Cham : Springer, second edition, 2022. Peter Walley. Coherent lower (and upper) probabilities. Technical report, University of Warwick, Coventry, 1981. Peter Walley. Statistical Reasoning with Imprecise Probabilities, volume 42 of Monographs on Statistics and Applied Probability. London : Chapman and Hall, 1991. Hao Wang and Dit-Yan Yeung. A survey on Bayesian deep learning. ACM Computing Surveys, 53(5):1 37, 2021. Larry A. Wasserman and Joseph B. Kadane. Bayes theorem for Choquet capacities. The Annals of Statistics, 18(3):1328 1339, 1990. Lisa Wimmer, Yusuf Sale, Paul Hofman, Bernd Bischl, and Eyke Hüllermeier. Quantifying aleatoric and epistemic uncertainty in machine learning: Are conditional entropy and mutual information appropriate measures? In Robin J. Evans and Ilya Shpitser (eds.), Proceedings of the Thirty-Ninth Conference on Uncertainty in Artificial Intelligence, volume 216 of Proceedings of Machine Learning Research, pp. 2282 2292. PMLR, 31 Jul 04 Aug 2023. URL https://proceedings.mlr.press/v216/wimmer23a.html. Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-MNIST: a novel image dataset for benchmarking machine learning algorithms. Co RR, abs/1708.07747, 2017. Marco Zaffalon. The naive credal classifier. Journal of Statistical Planning and Inference, 105(1): 5 21, 2002. ISSN 0378-3758. doi: https://doi.org/10.1016/S0378-3758(01)00201-4. URL https: //www.sciencedirect.com/science/article/pii/S0378375801002014. Imprecise Probability Models and their Applications. Fengshuo Zhang and Chao Gao. Convergence rates of variational posterior distributions. The Annals of Statistics, 48(4):2180 2207, 2020. Published in Transactions on Machine Learning Research (10/2024) A Why do we need IPs? The main motivations for working with credal sets are two. Let (Ω, F) be the measurable space of interest. (i) A single probability distribution does not suffice to represent ignorance in the sense of lack of knowledge; this is well documented in the literature, see e.g. Hüllermeier & Waegeman (2021) and references therein. Consider the example of complete ignorance (CI) in the case of a finite state space Ω(Hüllermeier & Waegeman, 2021, Section 3.3). In standard Bayesian analysis, CI is modeled in terms of the uniform distribution Unif(Ω); this is justified by Laplace s principle of indifference . Then, however, it is not possible to distinguish between precise probabilistic knowledge about a random event called prior indifference; think of the tossing of a fair coin and a complete lack of knowledge due to an incomplete description of the experiment called prior ignorance. Another problem is given by the additive nature of probability distributions. Consider again the example of a uniform distribution. First, let us observe that it is not invariant under reparametrization. In addition, if we model the ignorance about the length x of the side of a cube in R3 via a uniform measure on the interval [l, u] R, then this does not yield a uniform distribution of x3 on [l3, u3], which suggests some degree of informedness about the cube s volume. Finally, as pointed out in Walley (1991), if we ask a subject even an expert about their opinion regarding some events, it is much more likely that they will report interval of probabilities rather than single values. (ii) Working with credal sets allows to achieve prior and likelihood robustness: realistically large sets Pprior of priors and Plik of likelihoods are elicited. Using credal sets, the agent recognizes that prior beliefs and knowledge about the sampling model are limited and imprecise. Combining each pair of functions in Pprior and Plik using Bayes rule, a class of posterior distributions reflecting the updated state of uncertainty is formed. If the available information is not sufficient to identify a unique posterior distribution, or a set of posteriors whose diameter is small, credal sets allow to represent indecision, thus leading to a less informative but more robust conclusions.20 B On the use of credal sets Let us address a critique raised against the use of credal sets. Lassiter (2020) argues against the use of sets of probabilities to model an agent s prior beliefs and their knowledge of the sampling model, while debating in favor of using hierarchical Bayesian models. As reported in Hüllermeier & Waegeman (2021, Secton 4.6.2), the argument against credal sets that is more cogent for the machine learning literature is that modeling a lack of knowledge in a set-based manner may hamper the possibility of inductive inference, up to a point where learning from empirical data is not possible any more. With this, we mean the following. As Pericchi (1998) points out, the natural candidate for a class of priors to represent complete ignorance is the class Pall of all distributions. When this class leads to non-vacuous and useful conclusions, these are quite compelling and uncontroversial. It turns out that the posterior probabilities obtained from this class are vacuous, that is, their lower and upper bounds are 0 and 1: no finite sample is enough to annihilate a sufficiently extreme prior belief. There is then a compromise to be made, and this is the compromise of near-ignorance. The nearignorance class should be vacuous a priori in some respects, typically the ones that are the most important for the analysis at hand. This way of proceeding is labeled as arbitrary by Lassiter (2020), who instead advocates for the use of hierarchical Bayesian procedures. We find this critique not compelling, as during the analysis the job of the agent is to model reality: as pointed out in Hüllermeier & Waegeman (2021, Secton 5), statistical inference is not possible without underlying assumptions, and conclusions drawn from data are always conditional on those assumptions. If we were to work every time with the maximum level of generality, we would hardly be able to reach any conclusions. For example, in a statistical analysis we never consider the state Ωof apparently possible states (Walley, 1991, section 2.1.2), that is, the one that contains all the states ω that are logically consistent with the available information. If we consider a coin toss, we let the state space be Ω= {heads, tails}, certainly not Ω= {heads, tails, coin landing on its edge, coin braking into pieces on landing, coin disappearing down a crack in the floor}. The same holds for sets 20Here diameter has to be understood as the distance between upper and lower probability of event A, for all A F. Published in Transactions on Machine Learning Research (10/2024) of probabilities: it makes much more sense to work with near-ignorance credal sets than to work with Pall. A final reason to rebut the point in Lassiter (2020) is that the problems indicated in (i) in section A that make the use of the uniform prior distribution often interpreted as representing epistemic uncertainty in standard Bayesian inference at least debatable are inherited by hierarchical Bayesian modeling, as specified in Bernardo (1979); Hüllermeier & Waegeman (2021); Jeffreys (1946). Furthermore, a single distribution that is the result of a hierarchical Bayesian procedure is unable to gauge epistemic uncertainty (Hüllermeier & Waegeman, 2021). B.1 Further notes on credal sets As pointed out in Corani et al. (2012, Section 3.3), there is a way of obtaining credal sets starting from sets of probability intervals; in addition, standard algorithms can compute the extreme elements of a credal set for which a probability interval has been provided (Avis & Fukuda, 1996). However, the resulting number of extrema is exponential in the size of the possibility space (Tessem, 1992).21 For this reason we prefer to specify prior and likelihood finitely generated credal sets instead. The way credal sets behave after conditioning on new available evidence has been recently studied in Caprio & Seidenfeld (2023), and a way of using credal sets in the open-world scenario (that is, when the support of the elements of the credal set can become larger as more data become available) is explored in Caprio & Mukherjee (2023d). Furthermore, the use of credal sets in statistical learning theory has been studied in Caprio et al. (2024c), and in computer vision in Caprio (2024). C A further IP concept: the core Let again (Ω, F) be the measurable space of interest. Because of the conjugacy property of upper and lower probabilities, let us focus on upper probabilities only. We say that upper probability P is concave if P(A B) P(A)+P(B) P(A B), for all A, B F. Recall that (Ω, F) denotes the set of all probability measures on (Ω, F). Upper probability P is compatible with the convex set (Gong & Meng, 2021) core(P) : = {P (Ω, F) : P(A) P(A), A F} = {P (Ω, F) : P(A) P(A) P(A), A F} where the second equality is a characterization (Cerreia-Vioglio et al., 2015, Page 3389). Notice that the core is convex (Marinacci & Montrucchio, 2004, Section 2.2). We assume it is nonempty. Then, it is weak -compact as a result of Marinacci & Montrucchio (2004, Proposition 3).22 Since the core is convex, the set ex[core(P)] of extreme points of the core is well defined. It contains all the elements of the core that cannot be written as a convex combination of one another. The following important result is a consequence of Walley (1991, Theorem 3.6.2). Theorem 9. Suppose core(P) is nonempty. Then, the following holds. (a) ex[core(P)] = . (b) core(P) is the closure in the weak topology of the convex hull of ex[core(P)]. (c) If P(A) = sup P core(P ) P(A), for all A F, then P(A) = sup P ex[core(P )] P(A), for all A F. So in order to define an upper probability P that setwise dominates the elements of core(P) it is enough to specify the extreme points of the core. 21Recall that the possibility space of a random variable is the space of the values it can take on. 22Recall that in the weak topology, a net (Pα)α I converges to P if and only if Pα(A) P(A), for all A F. Published in Transactions on Machine Learning Research (10/2024) D A new Bayes theorem for IPs We present Theorem 10, a result that although appealing does not lend itself well to be applied to the CBDL procedure. An extension of Theorem 10 is given in Caprio et al. (2024b). Call Θ the parameter space of interest and assume it is Polish, that is, the topology for Θ is complete, separable, and metrizable. This ensures that the set (Θ, B) of probability measures on Θ is Polish as well, where B denotes the Borel σ-algebra for Θ. Let X be the set of all bounded, non-negative, B-measurable functionals on Θ. Call D = X Y the sample space endowed with the product σ-algebra A = Ax Ay, where Ax is the σ-algebra endowed to X and Ay is the σ-algebra endowed to Y. Let the agent elicit Lθ := {Pθ (D, A) : θ Θ}. Assume that each Pθ Lθ has density L(θ) = p(D | θ) with respect to some σ-finite dominating measure ν on (D, A); this represents the likelihood function for θ having observed data D D. We assume for now that L X , for all D D. Let the agent specify a set P of probabilities on (Θ, B). Then, compute P, and consider Pco := core(P); it represents the agent s initial beliefs.23 We assume that every P Pco has density p with respect to some σ-finite dominating measure µ on (Θ, B), that is, p = d P dµ . We require the agent s beliefs to be represented by the core for two main reasons. The first, mathematical, one is to ensure that the belief set is compatible with the upper probability. The second, philosophical, one is the following (Caprio & Gong, 2023; Caprio & Mukherjee, 2023c). A criticism brought forward by Walley (1991, Section 2.10.4.(c)) is that, given an upper probability P, there is no cogent reason for which the agent should choose a specific PT that is dominated by P, or for that matter a collection of plausible probabilities. Because the core considers all (countably additive) probability measures that are dominated by P, it is the perfect instrument to reconcile Walley s behavioral and sensitivity analysis interpretations. Let the agent compute P θ, and consider Lco θ := core(P θ); it represents the set of plausible likelihoods. Let L := L = d Pθ dν , Pθ Lco θ and denote by L(θ) := sup L L L(θ) and by L(θ) := inf L L L(θ), for all θ Θ. Call PD (Θ, B) : d PD dµ = p(θ | D) = L(θ)p(θ) R Θ L(θ)p(θ)dθ, dµ , P Pco, L = d Pθ dν , Pθ Lco θ the class of posterior probabilities when the prior is in Pco and the likelihood is in Lco θ , and let P D(A) = sup PD Pco D PD(A), for all A B. Then, the following is a generalization of Bayes theorem in Wasserman & Kadane (1990). Theorem 10. Suppose Pco, Lco θ are nonempty. Then for all A B, P D(A) sup P Pco R Θ L(θ)1A(θ)P(dθ) provided that the ratio is well defined. Here, c := sup P Pco R Θ L(θ)1A(θ)P(dθ) + inf P Pco R Θ L(θ)1Ac(θ)P(dθ), and 1A denotes the indicator function for A B. In addition, if P is concave, then the inequality in equation 5 is an equality for all A B. This result is particularly appealing because, given some assumptions, it allows to perform a (generalized) Bayesian update of a prior upper probability (PUP) by carrying out only one operation, even when the likelihood is ill specified so that a set of likelihoods is needed. We also have the following. Lemma 11. Suppose Pco, Lco θ are nonempty. Then, if P is concave, we have that P D is concave as well. 23Superscript co stands for convex and core. Published in Transactions on Machine Learning Research (10/2024) This lemma is important because it tells us that the generalized Bayesian update of Theorem 10 preserves concavity, and so it can be applied to successive iterations. If at time t the PUP is concave, then the PUP at time t + 1 that is, the posterior upper probability at time t will be concave too. Necessary and sufficient conditions for a generic upper probability to be concave are given in Marinacci & Montrucchio (2004, Section 5). In the future, these results can be generalized to the case in which the elements of X are unbounded using techniques in Troffaes & de Cooman (2014), and to the case in which the elements of X are Rd-valued, for some d N, since we never used specific properties of R in our proofs. Despite being attractive, the generalized Bayesian update of Theorem 10 hinges upon three assumptions, namely that Pco and Lco θ are both cores of an upper probability, that they are nonempty, and that the prior upper probability P is concave. As the proverb goes, there is no free lunch. Having to check these assumptions, together with computing a supremum, an infimum, and the integrals in equation 5, makes Theorem 10 inadequate to be applied in the context of CBDL. E Bounds on upper and lower entropy In this section we find an upper bound for H(P) and a lower bound for H(P) that are extremely interesting. We first need to introduce three new concepts. Definition 12. Consider a set P of probabilities on a generic measurable space (Ω, F). We say that lower probability P is convex if P(A B) P(A) + P(B) P(A B), for all A, B F. Then, let P be either an upper or a lower probability, and consider a generic bounded measurable function f on (Ω, F), that is, f B(Ω). We define the Choquet integral of f with respect to P as follows Ω f(ω)P(dω) := Z 0 P ({ω Ω: f(ω) t}) dt + Z 0 [P ({ω Ω: f(ω) t}) P(Ω)] dt, where the right hand side integrals are (improper) Riemann integrals. If P is additive, then the Choquet integral reduces to the standard additive integral. Finally, if Ωis uncountable, define for all ω Ω π(ω) := inf P P d P dµ (ω) and π(ω) := sup P P We call them lower and upper densities, respectively. The following theorem gives the desired bounds. Theorem 13. Consider a set P of probabilities on a generic measurable space (Ω, F). If Ωis uncountable, assume that every P P is dominated by a σ-finite measure µ and that the Radon-Nikodym derivatives d P dµ are continuous and bounded, for all P P. Define Ωlog [π(ω)] P(dω) if Ωis uncountable P ω ΩP({ω}) log[P({ω})] if Ωis at most countable and similarly Ωlog [π(ω)] P(dω) if Ωis uncountable P ω ΩP({ω}) log[P({ω})] if Ωis at most countable , Then, H(P) H(P) and H(P) H(P). In addition, if P is concave, the first bound is tighter, and if P is convex, the second bound is tighter. Remark 14. In Hüllermeier & Waegeman (2021, Section 4.6.1), the authors point out that Abellán & Moral (2000) presents a generalization of the Hartley measure, called generalized Hartley measure GH(P), that can be used to disaggregate the total uncertainty captured by H(P) into aleatoric and epistemic uncertainties. Published in Transactions on Machine Learning Research (10/2024) We prefer not to introduce it in the present work because GH(P) is defined based on the mass function of a belief function (Gong & Meng, 2021, Definition 2.4).24 This entails that the authors assume that lower probability P associated with the set of probabilities of interest is a belief function, and so that for every collection {A, A1, . . . , Ak} such that A Ai, the following holds =I {1,...,k} ( 1)#I 1P( i IAi), for all k N. As it is immediate to see, this is assumption is not needed in the context of CBDL. Other ways of disentangling and quantifying AU and EU can be found e.g. in Sale et al. (2023; 2024) F How to derive a predictive distribution Suppose we performed a Bayesian updating procedure so to obtain posterior pdf p(θ | x1, y1, . . . , xn, yn). Recall that {(xi, yi)}n i=1 (X Y)n denotes the training set. We obtain the predictive distribution p( y | x, x1, y1, . . . , xn, yn) on Y as follows p( y | x, x1, y1, . . . , xn, yn) = Z Θ p( y, θ | x, x1, y1, . . . , xn, yn)dθ Θ p( y | θ, x, x1, y1, . . . , xn, yn) p(θ | x, x1, y1, . . . , xn, yn)dθ Θ p( y | x, θ) p(θ | x1, y1, . . . , xn, yn)dθ, where p( y | x, θ) is the likelihood used to derive the posterior. Notice that the last equality comes from output y only depending on input x and parameter θ, and from having assumed Dx θ (see section 2.1). From an applied point of view, a sample from p( y | x, x1, y1, . . . , xn, yn) is obtained as follows: 1. specify input x; 2. sample a parameter θ from the posterior, θ p(θ | x1, y1, . . . , xn, yn); 3. plug θ in the likelihood and sample y p(y | θ, x). G Aleatoric uncertainty check for α-level IHDR The diameter of IRα( ˆPpred) is a function of the predictive aleatoric uncertainty (AU) faced by the agent.25 If we want to avoid to perform the computation in Step 5 of Algorithm 1 only to discover that IRα( ˆPpred) is too large , then we can add a predictive AU check . At the beginning of the analysis, compute the lower entropy H( ˆP pred) = mink,s H( ˆP pred k,s ) associated with the set ex ˆPpred of extreme elements of the VI-approximated predictive credal set ˆPpred. By equation 2, it is equal to the predictive aleatoric uncertainty encoded in ˆPpred. We then verify whether the lower entropy H( ˆP pred) is too high . That is, if H( ˆP pred) > φ, for some φ > 0, we want our procedure to abstain. This means that if the predictive aleatoric uncertainty in set ˆPpred is too high, then our procedure does not return any output set for input x. The value of φ can be set equal to the entropy of the probability measures that are typically used in the context the agent works in. For example, in medical applications the agent may consider the entropy of a Normal distribution, while in financial applications the entropy of a distribution with fatter tails, such as a t-distribution or a Cauchy. We call these reference φ values. If we add this predictive AU check , at inference time (that is, before Step 4 of Algorithm 1), the agent needs to specify the pair of parameters (φ, α). 24A belief function is a mathematical concept that should not be confused with the term belief we used throughout the paper to address the agent s knowledge. 25Since the diameter is a metric concept, we assume that we can find a well-defined metric dy on Y. If that is not the case, we substitute the diameter with the notion of cardinality. Published in Transactions on Machine Learning Research (10/2024) H α-level IHDR in a Classification setting In classification problems, BNNs compute the probability vector θ Θ Φθ|D(x), where we write Φθ|D to highlight the fact that θ is sampled from posterior p(θ | D), and then select the most likely class ˆy := arg maxj ϖj, where the ϖj s are the elements of ϖ. When applied to a classification setting, the general procedure introduced in Algorithm 1 becomes the following. Recall that we denote by K the cardinality of ex Pprior, and by S the cardinality of ex Plik. Assume that Y = {y1, . . . , y J}, that is, there are J N 2 possible labels. Then, a VI-approximated predictive distribution ˆP pred k,s in ex ˆPpred can be seen as J-dimensional probability vector ˆppred k,s = (ˆppred k,s,1, . . . , ˆppred k,s,J) , where ˆppred k,s,j = ˆP pred k,s ({yj}), for all k {1, . . . , K}, s {1, . . . , S}, and j {1, . . . , J}. Now fix any k and any s, and define the partial order k,s on Y as yl k,s yi ˆppred k,s,l ˆppred k,s,i and yl k,s yi ˆppred k,s,l > ˆppred k,s,i, where i, l {1, . . . , J}, i = l. This means that we can order the labels according to the probability that ˆP pred k,s assigns to them: the first label will be the one having highest probability according to ˆP pred k,s , the second label will have the second-highest probability according to ˆP pred k,s , and so on. Now order the label space Y according to k,s so to obtain Yk,s := {yk,s 1 , . . . , yk,s J }. This means that yk,s 1 k,s yk,s j , for all j {2, . . . , J}, yk,s 2 k,s yk,s j , for all j {3, . . . , J} (but yk,s 1 k,s yk,s 2 ), and so on. That is, we order the labels from the most to the least likely according to ˆP pred k,s . Then, we call α-level credible set according to ˆP pred k,s , α [0, 1], the set CSα( ˆP pred k,s ) := yk,s 1 , . . . , yk,s j : i=1 ˆP pred k,s ({yk,s i }) [1 α, 1 α + ε], j J, and j < j : i=1 ˆP pred k,s ({yk,s i }) [1 α, 1 α + ε] , for some ε > 0. It corresponds to the α-level HDR Rα( ˆP pred k,s ). Notice that we require Pj i=1 ˆP pred k,s ({yk,s i }) [1 α, 1 α+ε] because we may need to go slightly above level 1 α. Just as a toy example, we may have 7 labels, 3 of which would give a 0.945 coverage, while 4 would give a coverage of 0.953. If we are interested in the α = 0.05-level credible set, we ought to include the fourth label, thus yielding a coverage slightly higher than 1 α = 0.95. The interpretation to CSα( ˆP pred k,s ) is the following: it consists of the smallest collection of labels to which ˆP pred k,s assigns probability of at least 1 α (that is, those having the highest probability of being the correct one for the new input x). Finally, we call α-level imprecise credible set, α [0, 1], the set ICSα( ˆPpred) := [ k,s CSα( ˆP pred k,s ). In turn, we have that ˆP pred[{ y ICSα( ˆPpred)] 1 α. Remark 15. Notice that if a credible set of level α is enough, then we can replace the left endpoint of the interval in equation 6 with 1 (α + εk,s), for some εk,s > 0. Strictly speaking, in this case we obtain an (α +εk,s)-level credible set, which we denote by CSαk,s( ˆP pred k,s ), where αk,s := α +εk,s. Going back to our toy example, we will have a credible set with 3 labels that yields a coverage of 1 (α+εk,s) = 0.945 0.95 = 1 α, so εk,s = 0.005. In turn, this implies that the imprecise credible set will have a coverage of 1 (α+maxk,s εk,s), that is, it will have level α + maxk,s εk,s. We denote it by ICS α( ˆPpred), where α := α + maxk,s εk,s. Published in Transactions on Machine Learning Research (10/2024) I Distance between a set of distributions P and a single distribution P having different dimensions Many concepts in this section are derived from Cai & Lim (2022). Let m, n N such that m n, and let p [1, ]. Call M p(Rj) the set of probability measures on Rj having finite p-th moment, and Md(Rj) the set of probability measures on Rj having density with respect to some σ-finite dominating measure µ, j {m, n}. Let O(m, n) := {V Rm n : V V = Im}, where Im denotes the m-dimensional identity matrix, and for any V O(m, n) and any b Rm, define the following function φV,b : Rn Rm, x 7 φV,b(x) := V x + b. Let B(Rn) be the Borel σ-algebra on Rn, and for any Q (Rn, B(Rn)), define φV,b(Q) := Q φ 1 V,b, the pushforward measure. Consider then two generic probability measures Q, S such that Q M p(Rm) and S M p(Rn), and call Φ+ p (Q, n) := {α M p(Rn) : φV,b(α) = Q, for some V O(m, n), b Rm}, Φ+ d (Q, n) := {α Md(Rn) : φV,b(α) = Q, for some V O(m, n), b Rm}, Φ (S, m) := {β M(Rm) : φV,b(S) = β, for some V O(m, n), b Rm}. Recall now the definition of p-Wasserstein metric between two generic distributions defined on the same Euclidean space. Let P1, P2 M p(Rn), for some n N and some p [1, ]. Then, the p-Wasserstein distance between them is defined as Wp(P1, P2) := inf γ Γ(P1,P2) R2n x y p 2γ(d(x, y)) 1/p , where 2 denotes the Euclidean distance, p = is interpreted as the essential supremum, and Γ(P1, P2) := {γ (R2n, B(R2n)) : projn 1(γ) = P2, projn 2(γ) = P1} is the set of couplings between P1 and P2, where projn 1 is the projection onto the first n coordinates, and projn 2 is the projection to the last n coordinates. Recall then the definition of f-divergence between two generic distributions defined on the same Euclidean space. Let P1, P2 (Rn, B(Rn)), for some n N, and assume P1 P2. Then, for any convex functional f on R such that f(1) = 0, the f-divergence between P1 and P2 is defined as divf(P1 P2) := Z d P2 (x) P2(dx). Aside from the Rényi divergence, the f-divergence includes just about every known divergences as special case (Cai & Lim, 2022). The following are the main results of this section. Lemma 16. Let m, n N such that m n, and let p [1, ] and f be any convex functional on R such that f(0) = 1. Consider a generic P M p(Rm) and P M p(Rn). Let Φ+ p (P, n) := P PΦ+ p (P, n) and Φ+ d (P, n) := P PΦ+ d (P, n). Define W + p (P, P ) := infα Φ+ p (P,n) Wp(α, P ), for all P P; div+ f (P P ) := infα Φ+ d (P,n) divf(P P ), for all P P; W + p (P, P ) := infα Φ+ p (P,n) Wp(α, P ); div+ f (P P ) := infα Φ+ d (P,n) divf(P P ); Then, for all P P the following holds W + p (P, P ) W + p (P, P ) and div+ f (P P ) div+ f (P P ). Published in Transactions on Machine Learning Research (10/2024) Lemma 17. Let m, n N such that m n, and let p [1, ] and f be any convex functional on R such that f(0) = 1. Consider a generic P M p(Rn) and P M p(Rm). Let Φ (P, m) := P PΦ (P, m). Define W p (P, P ) := infα Φ (P,m) Wp(α, P ), for all P P; div f (P P ) := infα Φ (P,m) divf(P P ), for all P P; W p (P, P ) := infα Φ (P,m) Wp(α, P ); div f (P P ) := infα Φ (P,m) divf(P P ); Then, for all P P the following holds W p (P, P ) W p (P, P ) and div f (P P ) div f (P P ). A visual representation of the application of Lemma 17 in the context of CBDL is given in Figure 11. Φ (𝒫lik, m) Φ (L, m) Lo p (𝒫lik, Lo) Figure 11: We assume that n > m and that the oracle distribution Lo belongs to M p(Rm), while likelihood FGCS Plik is a subset of M p(Rn), for some finite p 1. We also assume that L is one of the extreme elements of Plik. We see how W p (Plik, Lo) < W p (L, Lo); if we replace metric W p by a generic f-divergence divf, the inequality would still hold thanks to Lemma 17. J Details on Artificial Pancreas Example Artificial Pancreas Model. An important factor when designing the controller for an artificial pancreas is to adapt the insulin delivery algorithm to the particular details of the patient. This is because patients display a wide range of variability in their response to insulin, depending on age, Body Mass Index (BMI), and other physiological parameters. The Bayesian Neural Network models have 2 hidden layers, with 10 Published in Transactions on Machine Learning Research (10/2024) neurons each, for the case study with 4 different seeds. This choice was informed by the experiments in Dutta et al. (2018). For the case study with different architectures, we trained BNNs with 4 different widths: 10, 20, 30, and 40. The horizon length is H = 10 time steps, and the prediction horizon is 5 steps into the future. The neural networks were trained for 200 time steps, with a learning rate of 0.001, and batch size of 128 using Mean-Field Variational Inference. The training dataset consisted of 28400 training samples, recorded without meals. Controller. We implemented a simple model predictive controller, using an off-the-shelf implementation of the covariance matrix adaptation evolution strategy (CMA-ES). The model predictive control planning horizon was k = 5, with a fixed seed for the randomized solver. K Other possible baselines for CBDL Our experiments may seem like a type of belief tracking (Fortin et al., 2017; Klein et al., 2010). Two comments are in order. First, this line of literature does not use deep learning techniques. Second, in these works the authors rely on Dempster-Shafer theory, a field in imprecise probability theory where lower probabilities are assumed to be belief functions, see Remark 14. We do not rely on this assumption in our work. L Further related work Modeling uncertainty has been a longstanding goal of ML/AI research and a variety of approaches have been developed for doing so (Guo et al., 2017; Park et al., 2020; Jospin et al., 2022). Recently, emphasis has been placed on discerning between aleatoric and epistemic uncertainties (Senge et al., 2014; Kull & Flach, 2014; Kendall & Gal, 2017). In Tretiak et al. (2022), the authors present an IP-based neural network which uses a regression technique based on probability intervals. Contrary to CBDL, their NN is rooted in the frequentist approach to imprecise probabilities (Huber & Ronchetti, 2009). In Manchingal & Cuzzolin (2022, Sections 2.1, 2.3), the authors focus on belief-functions-based classification methods. CBDL cannot be directly compared with these methodologies because (i) they do not require that the user expresses their knowledge via a belief function, but rather through a credal set; (ii) they can be used for regression and classification; (iii) they are rooted in Bayesian theory, as opposed to Dempster Shafer theory. Other works in ML using a belief function approach are those from the field of evidential machine learning (EML), see e.g. (Amini et al., 2020; Charpentier et al., 2020; Denœux, 2022; 2023; Sensoy et al., 2018) and references therein. Existing models mainly address clustering, classification, and regression problems. The reasons why CBDL cannot be directly compared with methods from the EML literature are (i) and (iii) above, the fact that EML methods are not (derived from) Bayesian ones, and that their definitions of AU and EU are slightly different from the canonical ones in ensemble deep learning. Proof of Proposition 3. If P(A) = sup P Π P (A), for all A F, then it is immediate to see that P(A) = sup P Π P(A) = sup P exΠ P (A) = sup P Π P (A), for all A F, since Π Π . Suppose now that P(A) = sup P Π P(A), for all A F. Then, we have that for all P Π and all A F, P(A) P(A). Pick now any P Π . We can write it as P = Pk j=1 βj Pj, where βj [0, 1], for all j {1, . . . , k}, Pk j=1 βj = 1, and {Pj}k j=1 = Π. Pick then any A F; we have j=1 βj Pj(A) j=1 βj P(A) = P(A). So P(A) P (A). Because this holds for all P Π and all A F, the claim is proven. Published in Transactions on Machine Learning Research (10/2024) Proof of Proposition 6. The lower bound for the upper entropy of Π comes immediately from Π being a superset of Π. Let us now prove the lower entropy equality. Let Π = {P1, . . . , Pk} and Π = ConvΠ. Pick any P Π . By the definition of Π , there exists a collection of non-negative reals {βj}k j=1 such that Pk j=1 βj = 1 and Pk j=1 βj Pj = P . By the concavity of the entropy, we have that j=1 βj H(Pj) j=1 βj H(P) = H(P) := inf P Π H(P). Since P was chosen arbitrarily and Π is finite, this implies that inf P Π H(P ) =: H(P ) H(P). In addition, we have that, since Π Π , H(P ) H(P). Combining this with the above result, we obtain H(P ) = H(P). Proof of Proposition 8. Pick any metric d on the space Ω (Ω, F) of probabilities on (Ω, F), and any P Π . Because P belongs to Π , inf P Π d(P , Ψ) can only be either equal to or smaller than d(P , Ψ). By the definition of d(Π , Ψ), if inf P Π d(P , Ψ) = d(P , Ψ), then d(Π , Ψ) = d(P , Ψ). If instead inf P Π d(P , Ψ) < d(P , Ψ), then d(Π , Ψ) < d(P , Ψ). The proof is similar for a generic divergence div on (Ω, F). Proof of Theorem 10. Assume Pco, Lco θ are nonempty. Then, by Marinacci & Montrucchio (2004, Proposition 3), they are weak -compact. Pick any A B. Recall that we can rewrite the usual Bayes updating rule as Θ L(θ)1A(θ)P(dθ) R Θ L(θ)1A(θ)P(dθ) + R Θ L(θ)1Ac(θ)P(dθ) Θ L(θ)1Ac(θ)P (dθ) R Θ L(θ)1A(θ)P (dθ) which is maximized when R Θ L(θ)1Ac(θ)P(dθ) R Θ L(θ)1A(θ)P(dθ) is minimized. But R Θ L(θ)1Ac(θ)P(dθ) R Θ L(θ)1A(θ)P(dθ) inf P Pco R Θ L(θ)1Ac(θ)P(dθ) sup P Pco R Θ L(θ)1A(θ)P(dθ) , which proves the inequality in equation 5. Assume now that P is concave. By Wasserman & Kadane (1990, Lemma 1), we have that there exists P Pco such that Θ L(θ)1A(θ)P(dθ) = Z Θ L(θ)1A(θ)P(dθ), (7) for all L L . In addition, by Wasserman & Kadane (1990, Lemma 4), we have that for all X X and all ϵ > 0, there exists a non-negative, upper semi-continuous function h X such that sup P Pco Θ X(θ)P(dθ) ϵ < sup P Pco Θ h(θ)P(dθ) Θ X(θ)P(dθ). (8) Published in Transactions on Machine Learning Research (10/2024) Let now X = L1A. Notice that since Lco θ is weak -compact, by equation 4 so is L . This implies that L, L L , since a compact set always contains its boundary, so L X as well, and in turn L1A X . Fix then any L L and put h = L1A. It is immediate to see that h is non-negative and upper semi-continuous. Then, by equation 8, we have that for all ϵ > 0 sup P Pco L(θ)1A(θ)P(dθ) ϵ < Θ L(θ)1A(θ)P(dθ) sup P Pco L(θ)1A(θ)P(dθ). (9) Combining equation 7 andequation 9, we obtain sup P Pco L(θ)1A(θ)P(dθ) ϵ Θ L(θ)1A(θ)P(dθ) sup P Pco L(θ)1A(θ)P(dθ), (10) for all L L . Pick now any ϵ > 0 and put k := sup P Pco L(θ)1A(θ)P(dθ) + inf P Pco Θ L(θ)1Ac(θ)P(dθ) > 0. Choose any L L and δ (0, ϵk). By equation 10 we have that [sup P Pco R Θ L(θ)1A(θ)P(dθ)] δ < R Θ L(θ)1A(θ)P(dθ) and that [inf P Pco R Θ L(θ)1Ac(θ)P(dθ)] + δ > R Θ L(θ)1Ac(θ)P(dθ). Recall that c := sup P Pco R Θ L(θ)1A(θ)P(dθ) + inf P Pco R Θ L(θ)1Ac(θ)P(dθ), and define d := R Θ L(θ)1A(θ)P(dθ) + R Θ L(θ)1Ac(θ)P(dθ). Then, Θ L(θ)1A(θ)P(dθ) sup P Pco R Θ L(θ)1A(θ)P(dθ) δ c + δ δ = sup P Pco R Θ L(θ)1A(θ)P(dθ) > sup P Pco R Θ L(θ)1A(θ)P(dθ) Since this holds for all ϵ > 0, we have that sup PD Pco D PD(A) = sup P Pco R Θ L(θ)1A(θ)P(dθ) concluding the proof. Proof of Lemma 11. Walley (1981); Wasserman & Kadane (1990) show that concave upper probabilities are closed with respect to the generalized Bayes rule. In particular, this means that, if we let b := sup P Pco R Θ L(θ)1A(θ)P(dθ) + inf P Pco R Θ L(θ)1Ac(θ)P(dθ), for any fixed A B, if P is concave, then for all L L P D(A) = sup P Pco R Θ L(θ)1A(θ)P(dθ) is concave. But since Lco θ is weak -compact (by our assumption and Marinacci & Montrucchio (2004, Proposition 3)), by equation 4 so is L . This implies that L, L L , since a compact set always contains its Published in Transactions on Machine Learning Research (10/2024) boundary. Call then L = L1A + L1Ac. It is immediate to see that L L . Then, by equation 11 we have that if we call b := sup P Pco R Θ L (θ)1A(θ)P(dθ) + inf P Pco R Θ L (θ)1Ac(θ)P(dθ), it follows that P D(A) = sup P Pco R Θ L (θ)1A(θ)P(dθ) = sup P Pco R Θ L(θ)1A(θ)P(dθ) is concave, concluding the proof. Proof of Theorem 13. Suppose Ωis uncountable. First notice that, since we assumed d P dµ to be continuous and bounded for all P P, then so is log d P dµ , for all P P, since composing a continuous function with a continuous and bounded one gives us a continuous and bounded function. This entails that the Choquet integrals of log d P dµ with respect to P and P are both well defined. In addition, being continuous and bounded, both d P dµ and log d P dµ attain their infima and suprema thanks to Weierstrass extreme value theorem, for all P P. Hence, all the Choquet integrals used in this proof are well defined. Then, we have the following H(P) : = sup P P H(P) = sup P P dµ (ω) P(dω) Ω ( log) d P dµ (ω) P(dω) dµ (ω) P(dω) (12) Ω log inf P P d P Ω log [π(ω)] P(dω) = H(P). The inequality in equation 12 is true because for all ω Ω, dµ (ω) ( log) d P The inequality in equation 13 is a property of Choquet integrals taken with respect to upper probabilities (Marinacci & Montrucchio, 2004). The equality in equation 14 is true because for a generic function f, we have that sup f = inf f. Finally, the equality in equation 15 is true because the logarithm is a strictly increasing function. By Marinacci & Montrucchio (2004, Theorem 38), if P is concave, then inequality equation 13 holds with an equality, and so the bound is tighter. The proof for H(P) H(P) is similar; we use the facts that inf P P n ( log) d P dµ (ω) o ( log) d P dµ (ω) , for all ω Ω; by Marinacci & Montrucchio (2004), dµ (ω) P(dω) dµ (ω) P(dω); (16) Published in Transactions on Machine Learning Research (10/2024) for a generic function f, inf f = sup f; by Marinacci & Montrucchio (2004, Theorem 38), if P is convex, then equation 16 holds with an equality. Suppose now Ωis at most countable; in this case, we do not need any assumptions to make the Choquet integrals well defined, since we will not deal with density functions. The following holds H(P) : = sup P P H(P) = sup P P ω Ω P({ω}) log [P({ω})] ω Ω P({ω})( log) [P({ω})] ω Ω sup P P {P({ω})( log) [P({ω})]} (17) P({ω}) sup P P ( log) [P({ω})] (18) P({ω}) inf P P log [P({ω})] (19) P({ω}) log inf P P P({ω}) (20) P({ω}) log [P({ω})] = H(P). The inequality in equation 17 comes from the well known fact that the sum of the suprema is at least equal to the supremum of the sum. The inequality in equation 18 comes from the fact that for differentiable functions, the product of the suprema is at least equal to the supremum of the product. The equality in equation 19 is true because for a generic function f, we have that sup f = inf f. Finally, the equality in equation 20 is true because the logarithm is a strictly increasing function. By Marinacci & Montrucchio (2004, Theorem 38), if P is concave, then inequality equation 17 holds with an equality, and so the bound is tighter. The proof for H(P) H(P) is similar; we use the facts that the sum of the infima is at most equal to the infimum of the sum; for differentiable functions, the product of the infima is at most equal to the infimum of the product; for a generic function f, inf f = sup f; by Marinacci & Montrucchio (2004, Theorem 38), if P is convex, then inf P P P ω ΩP({ω})( log) [P({ω})] = P ω Ωinf P P {P({ω})( log) [P({ω})]}. Proof of Lemma 16. Fix any p [1, ] and pick any P P. Because Φ+ p (P, n) Φ+ p (P, n), then infα Φ+ p (P,n) Wp(α, P ) can only be either equal or smaller than infα Φ+ p (P,n) Wp(α, P ). Now, if infα Φ+ p (P,n) Wp(α, P ) = infα Φ+ p (P,n) Wp(α, P ), then W + p (P, P ) = W + p (P, P ). If instead infα Φ+ p (P,n) Wp(α, P ) < infα Φ+ p (P,n) Wp(α, P ), then W + p (P, P ) < W + p (P, P ). This concludes the first part of the proof. Fix then any convex functional f on R such that f(0) = 1; the proof is similar for f-divergences. Proof of Lemma 17. The proof is very similar to that of Lemma 16. Published in Transactions on Machine Learning Research (10/2024) N All Results: In-distribution and Out-of-distribution Evaluation N.1 CIFAR-10 Results CBDL Ensemble BNN Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric Low Med High Low Med High Low Med High Low Med High Low Med High Low Med High gaussian 0.145 0.169 0.176 0.066 0.099 0.102 0.012 0.016 0.016 0.065 0.056 0.055 0.268 0.359 0.38 0.292 0.388 0.412 shot noise 0.129 0.158 0.174 0.053 0.084 0.104 0.01 0.014 0.016 0.069 0.061 0.055 0.224 0.315 0.378 0.249 0.342 0.409 speckle noise 0.128 0.158 0.174 0.053 0.081 0.102 0.01 0.014 0.016 0.069 0.061 0.055 0.224 0.31 0.375 0.247 0.336 0.407 impulse 0.134 0.161 0.174 0.052 0.085 0.118 0.011 0.015 0.017 0.069 0.059 0.051 0.231 0.338 0.423 0.257 0.364 0.454 defocus blur 0.105 0.129 0.176 0.032 0.048 0.095 0.008 0.01 0.016 0.076 0.071 0.057 0.154 0.203 0.325 0.185 0.25 0.434 gaussian blur 0.105 0.154 0.186 0.032 0.069 0.11 0.008 0.013 0.017 0.076 0.064 0.053 0.154 0.258 0.348 0.185 0.329 0.497 motion blur 0.137 0.166 0.175 0.055 0.085 0.1 0.011 0.015 0.016 0.068 0.06 0.056 0.227 0.322 0.36 0.268 0.385 0.446 zoom blur 0.148 0.161 0.176 0.063 0.077 0.1 0.012 0.014 0.016 0.066 0.062 0.056 0.253 0.287 0.338 0.318 0.365 0.457 snow 0.131 0.157 0.163 0.05 0.077 0.086 0.01 0.014 0.015 0.07 0.062 0.059 0.221 0.31 0.337 0.242 0.326 0.355 fog 0.104 0.122 0.154 0.033 0.045 0.09 0.008 0.009 0.013 0.076 0.072 0.06 0.155 0.199 0.321 0.187 0.246 0.397 brightness 0.103 0.108 0.124 0.031 0.034 0.043 0.008 0.008 0.01 0.076 0.075 0.072 0.153 0.164 0.195 0.179 0.188 0.225 contrast 0.107 0.145 0.165 0.035 0.066 0.148 0.008 0.012 0.017 0.075 0.065 0.046 0.162 0.265 0.419 0.197 0.342 0.69 elastic 0.134 0.144 0.168 0.053 0.059 0.089 0.011 0.012 0.015 0.069 0.067 0.058 0.219 0.238 0.33 0.273 0.297 0.38 pixelate 0.116 0.135 0.162 0.042 0.065 0.096 0.009 0.011 0.015 0.073 0.066 0.058 0.191 0.259 0.342 0.215 0.275 0.38 jpeg 0.13 0.147 0.156 0.049 0.064 0.075 0.01 0.012 0.013 0.07 0.066 0.063 0.217 0.262 0.296 0.246 0.293 0.331 spatter 0.117 0.145 0.147 0.041 0.065 0.062 0.009 0.012 0.012 0.073 0.065 0.066 0.191 0.279 0.266 0.216 0.299 0.291 saturate 0.112 0.113 0.131 0.041 0.039 0.046 0.008 0.009 0.011 0.073 0.074 0.07 0.183 0.183 0.219 0.212 0.212 0.257 frost 0.124 0.153 0.166 0.047 0.075 0.099 0.01 0.013 0.015 0.071 0.063 0.056 0.208 0.298 0.361 0.23 0.326 0.403 Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric In Dist-Clean 0.102 0.031 0.008 0.076 0.151 0.179 MNIST 0.184 (1.801) 0.142 (4.626) 0.021 (2.723) 0.044 (0.572) 0.445 (2.95) 0.772 (4.312) Fashion MNIST 0.183 (1.791) 0.147 (4.808) 0.022 (2.822) 0.042 (0.557) 0.431 (2.861) 0.811 (4.53) SVHN 0.183 (1.789) 0.141 (4.606) 0.019 (2.421) 0.046 (0.608) 0.394 (2.612) 0.647 (3.616) Table 8: CIFAR-10 Results. The 4 BNNs trained have the following accuracies : 90, 89, 90, 89 in percentage terms and rounded to the nearest whole number. For different categories of corruptions, increasing severity leads to higher levels of aleatoric uncertainty for CBDL. When exposed to completely unseen datasets, this reaches its peak. In contrast, the Ensemble has a reverse trend. For the single BNN, the network with the highest accuracy was selected. Published in Transactions on Machine Learning Research (10/2024) N.2 MNIST Results CBDL Ensemble BNN Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric Low Med High Low Med High Low Med High Low Med High Low Med High Low Med High brightness 0.21 0.246 0.195 0.128 0.087 0.158 0.023 0.026 0.024 0.045 0.05 0.045 0.323 0.599 0.767 0.11 0.229 0.301 canny edges 0.162 0.158 0.162 0.174 0.176 0.174 0.016 0.015 0.016 0.042 0.042 0.042 0.641 0.658 0.651 0.296 0.289 0.296 dotted line 0.18 0.173 0.176 0.097 0.102 0.102 0.014 0.013 0.014 0.055 0.055 0.055 0.375 0.377 0.376 0.153 0.152 0.15 fog 0.185 0.187 0.19 0.163 0.167 0.166 0.023 0.023 0.023 0.043 0.042 0.043 0.749 0.827 0.826 0.326 0.359 0.361 glass blur 0.161 0.145 0.155 0.181 0.206 0.2 0.016 0.016 0.02 0.038 0.033 0.033 0.634 0.706 0.751 0.323 0.389 0.386 impulse noise 0.186 0.172 0.181 0.083 0.132 0.158 0.013 0.015 0.021 0.057 0.049 0.042 0.297 0.469 0.766 0.122 0.219 0.379 motion blur 0.184 0.161 0.157 0.123 0.185 0.195 0.016 0.018 0.019 0.049 0.035 0.029 0.404 0.765 0.918 0.188 0.449 0.542 rotate 0.189 0.167 0.142 0.072 0.134 0.196 0.013 0.014 0.015 0.059 0.048 0.033 v0.247 0.451 0.749 0.096 0.211 0.379 scale 0.196 0.169 0.121 0.08 0.152 0.232 0.014 0.015 0.013 0.057 0.044 0.025 0.269 0.478 0.859 0.103 0.217 0.541 shear 0.188 0.177 0.153 0.065 0.102 0.184 0.012 0.014 0.017 0.061 0.055 0.036 0.219 0.346 0.736 0.082 0.147 0.366 shot noise 0.188 0.179 0.18 0.061 0.08 0.113 0.012 0.012 0.014 0.062 0.059 0.052 0.213 0.247 0.376 0.078 0.091 0.162 spatter 0.186 0.174 0.176 0.074 0.132 0.127 0.013 0.016 0.016 0.059 0.047 0.049 0.247 0.455 0.421 0.098 0.226 0.195 stripe 0.18 0.182 0.184 0.165 0.163 0.161 0.021 0.02 0.021 0.04 0.04 0.04 0.745 0.76 0.76 0.225 0.22 0.229 translate 0.191 0.192 0.192 0.061 0.086 0.128 0.013 0.015 0.019 0.061 0.056 0.046 0.214 0.292 0.474 0.078 0.112 0.204 zigzag 0.18 0.176 0.179 0.119 0.123 0.122 0.016 0.016 0.016 0.051 0.05 0.05 0.476 0.477 0.476 0.201 0.202 0.201 Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric In Dist-Clean 0.188 0.057 0.012 0.063 0.198 0.071 CIFAR 0.185 (0.986) 0.167 (2.94) 0.023 (1.948) 0.042 (0.67) 0.821 (4.155) 0.372 (5.124) Fashion MNIST 0.162 (0.866) 0.187 (3.287) 0.019 (1.618) 0.035 (0.564) 0.919 (4.654) 0.461 (6.496) SVHN 0.190 (1.015) 0.168 (2.963) 0.022 (1.847) 0.043 (0.69) 0.923 (4.674) 0.297 (4.183) Table 9: The 4 BNNs trained have the following accuracy: 99, 99, 99, 98 in percentage terms and rounded to the nearest whole number. These are the probabilities of the most likely label to be the correct one according to the 4 different BNNs. For the single BNN, the network with the highest accuracy was selected. For different categories of corruptions, increasing severity leads to higher levels of aleatoric uncertainty for IBNNs. When exposed to completely unseen datasets, this gets close to the highest aleatoric uncertainty. The same is not true for EBNN. The epistemic uncertainty for IBNNs shows a less consistent trend in this case. Published in Transactions on Machine Learning Research (10/2024) N.3 SVHN Results CBDL Ensemble BNN Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric Low Med High Low Med High Low Med High Low Med High Low Med High Low Med High brightness 0.062 0.074 0.102 0.011 0.019 0.053 0.005 0.006 0.008 0.083 0.08 0.07 0.113 0.161 0.293 0.136 0.198 0.405 contrast 0.077 0.099 0.143 0.022 0.048 0.174 0.006 0.008 0.018 0.08 0.072 0.04 0.17 0.261 0.37 0.247 0.487 1.212 defocus blur 0.109 0.154 0.138 0.044 0.147 0.218 0.009 0.017 0.02 0.072 0.045 0.029 0.272 0.486 0.431 0.353 0.877 1.334 elastic 0.182 0.177 0.192 0.14 0.122 0.102 0.023 0.021 0.022 0.04 0.046 0.05 0.562 0.495 0.489 0.84 0.779 0.639 fog 0.157 0.171 0.181 0.095 0.124 0.135 0.016 0.02 0.023 0.056 0.047 0.042 0.448 0.534 0.599 0.602 0.709 0.704 frost 0.089 0.118 0.131 0.026 0.052 0.063 0.007 0.01 0.012 0.078 0.069 0.066 0.2 0.319 0.372 0.238 0.381 0.442 gaussian blur 0.068 0.131 0.14 0.017 0.087 0.208 0.005 0.012 0.019 0.081 0.06 0.032 0.143 0.38 0.546 0.176 0.603 1.151 gaussian noise 0.116 0.163 0.182 0.031 0.061 0.103 0.01 0.018 0.025 0.074 0.061 0.044 0.251 0.429 0.655 0.251 0.41 0.586 impulse noise 0.13 0.169 0.185 0.035 0.06 0.102 0.012 0.018 0.026 0.072 0.061 0.045 0.3 0.449 0.653 0.296 0.43 0.589 jpeg 0.07 0.083 0.129 0.014 0.018 0.041 0.005 0.006 0.012 0.082 0.08 0.071 0.124 0.153 0.287 0.134 0.166 0.306 motion blur 0.1 0.157 0.167 0.027 0.078 0.149 0.008 0.015 0.02 0.077 0.06 0.041 0.233 0.438 0.605 0.272 0.568 0.889 pixelate 0.061 0.123 0.18 0.011 0.036 0.095 0.004 0.011 0.02 0.083 0.073 0.052 0.102 0.222 0.516 0.114 0.254 0.591 saturate 0.063 0.065 0.141 0.011 0.012 0.075 0.005 0.005 0.014 0.083 0.083 0.062 0.105 0.108 0.376 0.116 0.12 0.415 shot noise 0.119 0.17 0.183 0.031 0.067 0.106 0.011 0.019 0.026 0.074 0.059 0.044 0.255 0.462 0.659 0.256 0.441 0.584 snow 0.131 0.16 0.16 0.043 0.076 0.096 0.012 0.017 0.017 0.07 0.059 0.055 0.3 0.435 0.492 0.31 0.443 0.519 spatter 0.088 0.127 0.163 0.021 0.037 0.06 0.007 0.012 0.017 0.079 0.072 0.063 0.169 0.277 0.404 0.191 0.301 0.403 speckle noise 0.1 0.143 0.181 0.023 0.046 0.085 0.008 0.015 0.023 0.077 0.067 0.052 0.201 0.345 0.655 0.207 0.343 0.516 zoom blur 0.061 0.061 0.064 0.011 0.012 0.013 0.004 0.004 0.005 0.083 0.083 0.083 0.105 0.108 0.121 0.117 0.123 0.136 Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric In Dist-Clean 0.061 0.01 0.004 0.083 0.097 0.107 CIFAR 0.181 (2.992) 0.121 (11.778) 0.026 (6.029) 0.040 (0.479) 0.655 (6.765) 0.716(6.69) MNIST 0.198 (3.272) 0.063 (6.154) 0.024 (5.495) 0.056 (0.675) 0.329 (3.393) 0.336 (3.139) Fashion MNIST 0.199 (3.288) 0.113 (10.984) 0.026 (6.002) 0.041 (0.491) 0.103 (6.375) 0.125 (6.178) Table 10: The 4 BNNs trained have the following accuracy : 95, 95, 96, 95 in percentage terms and rounded to the nearest whole number. These are the probabilities of the most likely label to be the correct one according to the 4 different BNNs. For the single BNN, the network with the highest accuracy was selected. For different categories of corruptions, increasing severity leads to higher levels of aleatoric uncertainty for IBNNs. When exposed to completely unseen datasets, this reaches its peak. The same is not true for EBNN. For the epistemic uncertainty as well, there is a clear trend with increasing corruption severity. Published in Transactions on Machine Learning Research (10/2024) N.4 Fashion MNIST CBDL Baseline BNN Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric Low Med High Low Med High Low Med High Low Med High Low Med High Low Med High Fashion MNIST-C brightness 0.199 0.212 0.205 0.108 0.106 0.094 0.021 0.024 0.022 0.051 0.048 0.051 0.568 0.519 0.475 0.261 0.294 0.262 canny edges 0.18 0.175 0.177 0.157 0.164 0.16 0.021 0.02 0.02 0.043 0.042 0.043 0.729 0.734 0.753 0.216 0.215 0.217 dotted line 0.167 0.168 0.168 0.051 0.051 0.051 0.012 0.013 0.013 0.067 0.066 0.066 0.336 0.349 0.36 0.127 0.131 0.129 fog 0.19 0.185 0.187 0.128 0.132 0.129 0.021 0.02 0.021 0.046 0.045 0.046 0.489 0.463 0.448 0.264 0.25 0.237 glass blur 0.189 0.189 0.193 0.114 0.127 0.133 0.017 0.019 0.022 0.052 0.048 0.045 0.517 0.641 0.777 0.209 0.274 0.354 impulse noise 0.169 0.196 0.202 0.052 0.09 0.111 0.013 0.019 0.02 0.066 0.055 0.051 0.304 0.498 0.675 0.118 0.203 0.276 motion blur 0.182 0.171 0.163 0.106 0.157 0.174 0.017 0.02 0.019 0.053 0.041 0.038 0.505 0.616 0.641 0.216 0.296 0.285 rotate 0.185 0.173 0.162 0.074 0.164 0.184 0.016 0.02 0.02 0.06 0.037 0.035 0.392 0.754 0.797 0.162 0.338 0.341 scale 0.154 0.172 0.127 0.054 0.123 0.223 0.01 0.015 0.014 0.067 0.05 0.028 0.302 0.657 0.917 0.11 0.209 0.398 shear 0.172 0.177 0.182 0.063 0.13 0.144 0.014 0.019 0.02 0.063 0.045 0.039 0.333 0.609 0.719 0.134 0.257 0.279 shot noise 0.16 0.178 0.197 0.046 0.059 0.088 0.01 0.013 0.017 0.068 0.064 0.056 0.268 0.344 0.49 0.097 0.121 0.17 spatter 0.15 0.189 0.191 0.044 0.082 0.074 0.01 0.018 0.017 0.07 0.057 0.059 0.279 0.479 0.459 0.112 0.21 0.19 stripe 0.21 0.208 0.208 0.125 0.126 0.125 0.025 0.025 0.025 0.041 0.041 0.041 0.686 0.698 0.697 0.349 0.352 0.341 translate 0.153 0.189 0.17 0.046 0.094 0.158 0.01 0.018 0.019 0.069 0.054 0.041 0.27 0.499 0.721 0.102 0.21 0.317 zigzag 0.187 0.19 0.189 0.065 0.065 0.066 0.016 0.016 0.016 0.061 0.061 0.061 0.39 0.404 0.411 0.151 0.154 0.152 Epistemic Aleatoric Epistemic Aleatoric Epistemic Aleatoric In Dist-Clean 0.121 0.029 0.007 0.075 0.197 0.351 CIFAR 0.205 (1.705) 0.104 (3.561) 0.022 (3.178) 0.050 (0.667) 0.408 (2.073) 0.221 (0.629) MNIST 0.165 (1.367) 0.182 (6.192) 0.021 (3.069) 0.033 (0.443) 0.909 (4.618) 0.328 (0.932) SVHN 0.218 (1.81) 0.109 (3.742) 0.026 (3.827) 0.046 (0.618) 0.598 (3.036) 0.305 (0.867) Table 11: The 4 BNNs trained have the following accuracies : 93, 92, 92, 92 in percentage terms and rounded to the nearest whole number. These are the probabilities of the most likely label to be the correct one according to the 4 different BNNs. For the single BNN, the network with the highest accuracy was selected. For different categories of corruptions, increasing severity leads to higher levels of aleatoric uncertainty for IBNNs. Which is high when exposed to completely unseen datasets as well. The same is not true for EBNN. Published in Transactions on Machine Learning Research (10/2024) O Evaluation Summary by Noise O.1 CIFAR-10 Results BNN Ensemble CBDL gaussian noise 0.613 0.116 0.703 0.116 0.715 0.168 shot noise 0.669 0.077 0.794 0.077 0.793 0.144 impulse noise 0.66 0.132 0.776 0.132 0.791 0.191 speckle noise 0.683 0.057 0.829 0.057 0.839 0.115 gaussian blur 0.741 0.073 0.836 0.073 0.825 0.191 defocus blur 0.756 0.044 0.919 0.044 0.912 0.105 motion blur 0.813 0.036 0.902 0.036 0.889 0.067 zoom blur 0.78 0.029 0.899 0.029 0.888 0.066 fog 0.753 0.034 0.946 0.034 0.94 0.07 frost 0.745 0.025 0.884 0.025 0.884 0.062 snow 0.763 0.039 0.927 0.039 0.929 0.044 spatter 0.773 0.036 0.955 0.036 0.956 0.019 contrast 0.665 0.155 0.785 0.155 0.782 0.234 brightness 0.737 0.025 0.991 0.025 0.99 0.004 saturate 0.751 0.019 0.984 0.019 0.983 0.009 jpeg 0.775 0.011 0.94 0.011 0.937 0.023 pixelate 0.685 0.097 0.848 0.097 0.855 0.132 elastic 0.804 0.021 0.952 0.021 0.95 0.024 Table 12: Results for CIFAR-10C, organized by noise. For different categories of corruptions, increasing severity leads to higher levels of aleatoric uncertainty for CBDL. When exposed to completely unseen datasets, this reaches its peak. In contrast, the baseline has a reverse trend. O.2 MNIST Results BNN Ensemble CBDL shot noise 0.597 0.07 0.999 0.07 0.999 0.001 impulse noise 0.56 0.296 0.716 0.296 0.7 0.396 glass blur 0.365 0.217 0.406 0.217 0.41 0.271 motion blur 0.498 0.317 0.65 0.317 0.654 0.394 fog 0.343 0.269 0.364 0.269 0.579 0.282 spatter 0.714 0.089 0.991 0.089 0.991 0.007 brightness 0.63 0.323 0.803 0.323 0.904 0.113 shear 0.663 0.121 0.916 0.121 0.915 0.168 rotate 0.64 0.066 0.990 0.066 0.999 0.0 scale 0.618 0.32 0.755 0.32 0.749 0.382 translate 0.694 0.142 0.992 0.142 0.987 0.023 dotted line 0.728 0.019 0.995 0.019 0.994 0.001 zigzag 0.777 0.023 0.931 0.023 0.927 0.014 stripe 0.288 0.279 0.06 0.279 0.106 0.08 canny edges 0.84 0.043 0.849 0.043 0.911 0.036 Table 13: Results for MNIST-C, organized by noise. For different categories of corruptions, increasing severity leads to higher levels of aleatoric uncertainty for CBDL. When exposed to completely unseen datasets, this reaches its peak. In contrast, the baseline has a reverse trend. Published in Transactions on Machine Learning Research (10/2024) O.3 Fashion MNIST Results BNN Ensemble CBDL Fashion MNIST shot noise 0.538 0.134 0.636 0.134 0.624 0.161 impulse noise 0.411 0.251 0.511 0.251 0.519 0.344 glass blur 0.407 0.164 0.374 0.164 0.385 0.227 motion blur 0.66 0.061 0.928 0.061 0.933 0.034 fog 0.045 0.031 0.122 0.031 0.312 0.094 spatter 0.665 0.043 0.937 0.043 0.936 0.019 brightness 0.095 0.117 0.273 0.117 0.371 0.201 shear 0.593 0.143 0.691 0.143 0.719 0.165 rotate 0.45 0.277 0.488 0.277 0.451 0.347 scale 0.437 0.259 0.38 0.259 0.396 0.363 translate 0.586 0.147 0.716 0.147 0.752 0.178 dotted line 0.661 0.024 0.945 0.024 0.947 0.01 zigzag 0.669 0.019 0.915 0.019 0.919 0.015 stripe 0.141 0.046 0.016 0.046 0.071 0.049 canny edges 0.143 0.086 0.035 0.086 0.062 0.023 Table 14: Results for Fashion MNIST-C, organized by noise. For different categories of corruptions, increasing severity leads to higher levels of aleatoric uncertainty for CBDL. When exposed to completely unseen datasets, this reaches its peak. In contrast, the baseline has a reverse trend. O.4 SVHN Results BNN Ensemble CBDL Gaussian Noise 0.475 0.304 0.544 0.304 0.549 0.339 Shot Noise 0.447 0.298 0.502 0.298 0.504 0.332 Impulse Noise 0.478 0.304 0.522 0.304 0.532 0.315 Speckle Noise 0.586 0.211 0.69 0.211 0.695 0.28 Gaussian Blur 0.634 0.325 0.627 0.325 0.62 0.39 Defocus Blur 0.374 0.361 0.363 0.361 0.351 0.405 Motion Blur 0.765 0.094 0.803 0.094 0.795 0.133 Zoom Blur 0.675 0.02 0.998 0.02 0.998 0.002 Fog 0.432 0.264 0.363 0.264 0.34 0.246 Frost 0.79 0.032 0.952 0.032 0.945 0.035 Snow 0.738 0.034 0.779 0.034 0.775 0.08 Spatter 0.714 0.093 0.807 0.093 0.798 0.188 Contrast 0.852 0.115 0.781 0.115 0.803 0.284 Brightness 0.715 0.067 0.992 0.067 0.989 0.016 Saturate 0.707 0.078 0.93 0.078 0.916 0.14 JPEG 0.744 0.051 0.951 0.051 0.949 0.073 Pixelate 0.577 0.275 0.721 0.275 0.729 0.35 Elastic 0.274 0.23 0.213 0.23 0.227 0.208 Table 15: Results for SVHN-C, organized by noise. For different categories of corruptions, increasing severity leads to higher levels of aleatoric uncertainty for CBDL. When exposed to completely unseen datasets, this reaches its peak. In contrast, the baseline has a reverse trend. Published in Transactions on Machine Learning Research (10/2024) P Accuracy vs Rejection Rate - All Results P.1 CIFAR10-C Figure 12: Accuracy vs Rejection Rate - CIFAR10C gaussian noise. Figure 13: Accuracy vs Rejection Rate - CIFAR10C shot noise. Figure 14: Accuracy vs Rejection Rate - CIFAR10C impulse noise. Published in Transactions on Machine Learning Research (10/2024) Figure 15: Accuracy vs Rejection Rate - CIFAR10C speckle noise. Figure 16: Accuracy vs Rejection Rate - CIFAR10C gaussian blur. Figure 17: Accuracy vs Rejection Rate - CIFAR10C defocus blur. Figure 18: Accuracy vs Rejection Rate - CIFAR10C motion blur. Published in Transactions on Machine Learning Research (10/2024) Figure 19: Accuracy vs Rejection Rate - CIFAR10C zoom blur. Figure 20: Accuracy vs Rejection Rate - CIFAR10C fog. Figure 21: Accuracy vs Rejection Rate - CIFAR10C frost. Figure 22: Accuracy vs Rejection Rate - CIFAR10C snow. Published in Transactions on Machine Learning Research (10/2024) Figure 23: Accuracy vs Rejection Rate - CIFAR10C spatter. Figure 24: Accuracy vs Rejection Rate - CIFAR10C contrast. Figure 25: Accuracy vs Rejection Rate - CIFAR10C brightness. Figure 26: Accuracy vs Rejection Rate - CIFAR10C saturate. Published in Transactions on Machine Learning Research (10/2024) Figure 27: Accuracy vs Rejection Rate - CIFAR10C jpeg compression. Figure 28: Accuracy vs Rejection Rate - CIFAR10C pixelate. Figure 29: Accuracy vs Rejection Rate - CIFAR10C elastic transform. Published in Transactions on Machine Learning Research (10/2024) P.2 MNIST-C Figure 30: Accuracy vs Rejection Rate - MNISTC shot noise. Figure 31: Accuracy vs Rejection Rate - MNISTC impulse noise. Figure 32: Accuracy vs Rejection Rate - MNISTC glass blur. Published in Transactions on Machine Learning Research (10/2024) Figure 33: Accuracy vs Rejection Rate - MNISTC motion blur. Figure 34: Accuracy vs Rejection Rate - MNISTC fog. Figure 35: Accuracy vs Rejection Rate - MNISTC spatter. Published in Transactions on Machine Learning Research (10/2024) Figure 36: Accuracy vs Rejection Rate - MNISTC brightness. Figure 37: Accuracy vs Rejection Rate - MNISTC shear. Figure 38: Accuracy vs Rejection Rate - MNISTC rotate. Published in Transactions on Machine Learning Research (10/2024) Figure 39: Accuracy vs Rejection Rate - MNISTC scale. Figure 40: Accuracy vs Rejection Rate - MNISTC translate. Figure 41: Accuracy vs Rejection Rate - MNISTC dotted line. Published in Transactions on Machine Learning Research (10/2024) Figure 42: Accuracy vs Rejection Rate - MNISTC zigzag. Figure 43: Accuracy vs Rejection Rate - MNISTC stripe. Figure 44: Accuracy vs Rejection Rate - MNISTC canny edges. Published in Transactions on Machine Learning Research (10/2024) P.3 Fashion MNIST-C Figure 45: Accuracy vs Rejection Rate - Fashion MNISTC shot noise. Figure 46: Accuracy vs Rejection Rate - Fashion MNISTC impulse noise. Figure 47: Accuracy vs Rejection Rate - Fashion MNISTC glass blur. Published in Transactions on Machine Learning Research (10/2024) Figure 48: Accuracy vs Rejection Rate - Fashion MNISTC motion blur. Figure 49: Accuracy vs Rejection Rate - Fashion MNISTC fog. Figure 50: Accuracy vs Rejection Rate - Fashion MNISTC spatter. Published in Transactions on Machine Learning Research (10/2024) Figure 51: Accuracy vs Rejection Rate - Fashion MNISTC brightness. Figure 52: Accuracy vs Rejection Rate - Fashion MNISTC shear. Figure 53: Accuracy vs Rejection Rate - Fashion MNISTC rotate. Published in Transactions on Machine Learning Research (10/2024) Figure 54: Accuracy vs Rejection Rate - Fashion MNISTC scale. Figure 55: Accuracy vs Rejection Rate - Fashion MNISTC translate. Figure 56: Accuracy vs Rejection Rate - Fashion MNISTC dotted line. Published in Transactions on Machine Learning Research (10/2024) Figure 57: Accuracy vs Rejection Rate - Fashion MNISTC zigzag. Figure 58: Accuracy vs Rejection Rate - Fashion MNISTC stripe. Figure 59: Accuracy vs Rejection Rate - Fashion MNISTC canny edges. Published in Transactions on Machine Learning Research (10/2024) Figure 60: Accuracy vs Rejection Rate - SVHNC Gaussian Noise. Figure 61: Accuracy vs Rejection Rate - SVHNC Shot Noise. Figure 62: Accuracy vs Rejection Rate - SVHNC Impulse Noise. Published in Transactions on Machine Learning Research (10/2024) Figure 63: Accuracy vs Rejection Rate - SVHNC Speckle Noise. Figure 64: Accuracy vs Rejection Rate - SVHNC Gaussian Blur. Figure 65: Accuracy vs Rejection Rate - SVHNC Defocus Blur. Figure 66: Accuracy vs Rejection Rate - SVHNC Motion Blur. Published in Transactions on Machine Learning Research (10/2024) Figure 67: Accuracy vs Rejection Rate - SVHNC Zoom Blur. Figure 68: Accuracy vs Rejection Rate - SVHNC Fog. Figure 69: Accuracy vs Rejection Rate - SVHNC Frost. Figure 70: Accuracy vs Rejection Rate - SVHNC Snow. Published in Transactions on Machine Learning Research (10/2024) Figure 71: Accuracy vs Rejection Rate - SVHNC Spatter. Figure 72: Accuracy vs Rejection Rate - SVHNC Contrast. Figure 73: Accuracy vs Rejection Rate - SVHNC Brightness. Figure 74: Accuracy vs Rejection Rate - SVHNC Saturate. Published in Transactions on Machine Learning Research (10/2024) Figure 75: Accuracy vs Rejection Rate - SVHNC JPEG. Figure 76: Accuracy vs Rejection Rate - SVHNC Pixelate. Figure 77: Accuracy vs Rejection Rate - SVHNC Elastic. Published in Transactions on Machine Learning Research (10/2024) Q Code and Dataset Availability Code and datasets for implementation and replication of results will be available at https://github.com/ PRECISE/credal-bayesian-deep-learning.git.