# transitional_uncertainty_with_layered_intermediate_predictions__8219cc3e.pdf Transitional Uncertainty with Layered Intermediate Predictions Ryan Benkert 1 Mohit Prabhushankar 1 Ghassan Al Regib 1 Abstract In this paper, we discuss feature engineering for single-pass uncertainty estimation. For accurate uncertainty estimates, neural networks must extract differences in the feature space that quantify uncertainty. This could be achieved by current single-pass approaches that maintain feature distances between data points as they traverse the network. While initial results are promising, maintaining feature distances within the network representations frequently inhibits information compression and opposes the learning objective. We study this effect theoretically and empirically to arrive at a simple conclusion: preserving feature distances in the output is beneficial when the preserved features contribute to learning the label distribution and act in opposition otherwise. We then propose Transitional Uncertainty with Layered Intermediate Predictions (TULIP) as a simple approach to address the shortcomings of current single-pass estimators. Specifically, we implement feature preservation by extracting features from intermediate representations before information is collapsed by subsequent layers. We refer to the underlying preservation mechanism as transitional feature preservation. We show that TULIP matches or outperforms current single-pass methods on standard benchmarks and in practical settings where these methods are less reliable (imbalances, complex architectures, medical modalities). 1. Introduction Effective single-pass uncertainty estimation in deep learning is governed by two design principals. The first is defining an output score that reflects uncertainty. For instance, we can measure uncertainty through distance from the training data 1School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, USA. Correspondence to: Ryan Benkert . Proceedings of the 41 st International Conference on Machine Learning, Vienna, Austria. PMLR 235, 2024. Copyright 2024 by the author(s). (Liu et al., 2020), or the softmax confidence of the output (Mukhoti et al., 2023). While score design plays a crucial role in measuring uncertainty, the choice is dictated by application and uncertainty characteristics (Kendall & Gal, 2017). The second principle concerns information availability, namely whether the network can preserve features that reflect uncertainty information and does not collapse uncertain data points to certain representations (Van Amersfoort et al., 2020). We refer to the latter principle as feature preservation. Despite their critical importance, preserving features is not trivial in neural networks. In particular, information compression is a desirable property of neural networks and a central component of the learning problem (Tishby et al., 2000). In spite of this discrepancy, current single-pass uncertainty methods preserve features by maintaining distances between data points in the output and risk inhibiting compression of application irrelevant information (Liu et al., 2020; Van Amersfoort et al., 2020; van Amersfoort et al., 2021; Mukhoti et al., 2023; Kwon et al., 2020; Prabhushankar & Al Regib). As a result, several recent studies have shown practical limitations of current single-pass methods such as their susceptibility to distributional shift (Postels et al., 2022). We provide a simple illustration of this effect in Figure 1a. The left plot shows the 2D neural network output features of two clusters when trained without explicit feature preservation. The network collapses the class clusters to single points creating a challenging setting for uncertainty estimation. The center plot shows the same 2D classification problem, but depicts a network trained with feature preservation constraints on the output. These disadvantages motivate the search for alternative feature preservation approaches. To this end, a more conservative strategy preserves features in the collection of multiple representations of the same sample (Malinin & Gales, 2018). From a preservation perspective, the approach is preferred as feature distances are encoded in differences (e.g. spread) of the individual sample representations and does not require constraining the compression property. We term this approach as transitional feature preservation and illustrate a toy example in the right plot of Figure 1a. We show three different representations of the same two clusters collected from different sources hw1,2,3. While the sources collapse points individu- Transitional Uncertainty with Layered Intermediate Predictions Figure 1. a) overview of different feature preservation paradigms. We show 2D representations of neural networks where h1 and h2 denote the output dimensions in the feature space. Left: conventional neural network. Samples are collapsed to two tight clusters with little uncertainty information. Center: feature preservation in the output. Feature differences are maintained resulting in higher uncertainty related content but also in a cluster overlap. Right: transitional feature preservation. uncertainty is measured from differences between several representations of the same sample (denoted as hw1,2,3). b) uncertainty comparison of our intuition to ensembles. Transitions in between network layers provide an accurate signal for uncertainty estimation in comparison to ensembles. ally, the inter-source difference for a given sample reflects uncertainty (signified by h). A classic candidate that could implement this property are deep ensembles (Lakshminarayanan et al., 2017), where each source representation is collected from a separate network. However, ensembles face two central challenges for uncertainty estimation: 1) the evaluation requires several forward passes limiting their practicality; and 2) there is no guarantee that ensemble transitions preserve features because all ensemble members can collapse features to the same point collectively. We show an example of this phenomenon with a synthetic 2D spiral dataset (Figure 1b left plot). Uncertainties collected from ensemble transitions fail to capture uncertainty far away from the training data even though multiple networks are used to preserve features. Our method addresses both of these limitations. We provide experimental details and further discussion of our synthetic example in Appendix B.6 and C.2 respectively. In this paper, we address shortcomings of current singlepass uncertainty estimators and ensembles by leveraging intermediate representations. Figure 1b (right plot) shows the intuition behind our approach. Features are extracted from intermediate layers before information is collapsed allowing the uncertainty calculation within a single forward pass. From the uncertainty plot, we can see the advantage of our approach in comparison to ensembles. Uncertainty increases systematically further away from the training data showing while ensembles are uncertain only around decision boundaries. In addition to empirical evidence, we provide theoretical guarantees for intermediate representations in Section 4. In particular, we show that the linear combination of representations is feature preserving when the first layer is collapse resistant. Finally, we combine intermediate representations with a single-pass uncertainty estimation layer, approximate Gaussian Processes, and arrive at a new single-pass model: Transitional Uncertainty with Layered Intermediate Predictions or TULIP in short. Our estimator requires less labeled training data and outperforms current single-pass estimators both on standard benchmarks and other data modalities (CT scans) with little additional space overhead. Further, we show that TULIP is preferable under complex architectures, class imbalance, and several other challenging settings for current single-pass estimators. 2. Background 2.1. Neural Networks in the Information Plane Consider the input space X with a corresponding probabilistic random variable X. Further, let Y denote a lower dimensional target space characterized by variable Y . The learning problem for neural networks is equivalent to finding the minimally sufficient statistical mapping h (X) with respect to the mutual information I(X; Y ) (Shwartz-Ziv & Tishby, 2017). Transitional Uncertainty with Layered Intermediate Predictions h (X) = argmin hw:I(hw(X);Y )=I(X;Y ) I(hw(X); X) (1) Equation 1 is intuitive. During training, we optimize the network hw to fit the lower dimensional distribution Y - i.e. maximize the mutual information I(hw(X); Y ) between the representation distribution hw(X) and the target distribution Y . At the same time, the neural network must compress information irrelevant to the lower dimensional target variable Y . In Equation 1, we minimize the mutual information I(hw(X); X) between the representation hw(X) and the input distribution X. In practice, we derive hw from training data D = {yi, xi}N i=1 often collected from a subset of the full input space XID X. As a result, the network optimizes with respect to the in-distribution variable XID and produces arbitrarily bad results when exposed to out-ofdistribution (OOD) data XOOD X : XOOD XID = . For this reason, accurate uncertainty estimation is contingent on modeling information related to the full input distribution without over-fitting to XID (Liu et al., 2020). 2.2. Distance-Based Feature Preservation in the Output An intuitive approach to uncertainty estimation involves modeling distributional information in the output of the neural network feature extractor (Van Amersfoort et al., 2020; van Amersfoort et al., 2021; Mukhoti et al., 2023; Liu et al., 2020). The approach is practical as we can compute the network output in a single forward pass and measure uncertainty from the logits directly. To ensure accurate uncertainty estimates, current single-pass methods model input information by maintaining the distances between data points as they traverse the network. By preserving meaningful distances, we can estimate uncertainty by measuring the distance to the training domain XID (Liu et al., 2020). More formally, given an input space X equipped with a meaningful distance d X, we learn a neural network hw : X H that allows a distance d H within the feature manifold that reflects the true distance d X (Liu et al., 2020): d H(hw(x1), hw(x2)) = d X(x1, x2) (2) Unfortunately, neural networks do not naturally implement distance preservation and collapse data points to the same output. For this purpose, current single-pass methods enforce distance preservation artificially through constraints on the network representation. Popular examples of constraints include the two-sided gradient penalty (Gulrajani et al., 2017) and spectral normalization in combination with residual connections (Miyato et al., 2018). 3. Theoretical Analysis 3.1. Pitfalls of Feature Preservation in the Output While distance preservation in the output is a desirable property for uncertainty estimation, enforcing Equation 2 frequently results in performance degradation in neural networks: directly preserving distances in the network output can inhibit compression of information; a learning objective according to Equation 1. In this section, we provide theoretical justification that enforcing distance preservation on the network can act in direct opposition to the learning problem. We further arrive at an intuitive conclusion: distance preservation in the output is beneficial only when the preserved distances contain information related to the label distribution Y - i.e. when they are relevant to the application. We start our discussion by connecting the learning problem in Equation 1 to distances in the feature space. In particular, the minimally sufficient statistic shares the following dependency to feature distances for networks preserving distances in the output: h (X) = argmin hw:{I(f k H(hw(M k));Y k)=I(M k;Y k), k [1,Np]} X k I(f k H(hw(M k)); f k X(M k)). (3) Here, M k is the corresponding random variable of a subset of the input space Mk X, where each point in Mk has a unique distance to a fixed anchor point xk X. Together, all Np partition subsets form the entire input space X = S k [1,Np] Mk : T k [1,Np] Mk = . Further, f k L(.) = d L(xk; .) is a distance function with respect to the anchor point. We provide a formal definition of unique distance sets Mk, as well as proof for Equation 3 in Appendix A.1. Equation 3 is conceptually important as it provides a direct dependency between the learning problem of neural networks and distances in the feature plane. In particular, we note that the compression objective involves minimizing I(d H(hw(xk); hw(M k)); d X(xk; M k)) which is in direct opposition to preservation constraints that aim to maximize the similarities between the input and feature distances. We further note the maximization objective between the feature distances and the label distribution I(f k H(hw(M k)); Y k). Here, preservation constraints can be beneficial: when additional distances are preserved that contain information related to the label distribution Y k, the term is increased. We arrive at the following observation: Observation 1. Preserving distances in the output is beneficial if the preserved distances contribute to the application objective (i.e. contain information of label distribution), and oppose the learning problem otherwise. Transitional Uncertainty with Layered Intermediate Predictions Figure 2. Classification accuracy with and without distance preservation in the output: a) uniformely removed training and test data (left); b) class imbalance at different severity levels (right). In both graphs, we show the classification accuracy on the y-axis. The x-axis on the left graph represents the percentage of uniformely removed data, on the right the axis represents the fraction of imbalanced classes. The zero point on the x-axis is equivalent for both scenarios and represents the standard CIFAR100 benchmark without imbalance or data removal. In the following, we analyze the practical scenario of class imbalance to showcase Observation 1. 3.2. Distance Preservation under Class Imbalance In practice, relevancy information of features is not available and popular options preserve differences between features blindly without regard of application (Miyato et al., 2018; Gulrajani et al., 2017). A common example where this characteristic is problematic is class imbalance. Here, information is either overor under-represented in the training set, resulting in an increase of application-irrelevant data. In this subsection, we investigate the generalization performance when preserving features in the output under different severities of class imbalance. We find that distance constraints result in performance decline under high imbalance severities. In addition to generalization performance, we further investigate the uncertainty estimates under class imbalance in Appendix C.1. Experimental Setup To illustrate class imbalance, we artificially imbalance the CIFAR100 benchmark by removing either training or test samples of a previously balanced class. The portion of classes we artificially imbalance determines the severity of imbalance. For our experiments, we enforce distance preservation through spectral normalization in combination with residual connections (Miyato et al., 2018). We choose spectral normalization due to its simplicity and often stronger performance than the double sided gradient penalty (Gulrajani et al., 2017). We compare other distance-based methods in our benchmark experiments in Section 6. Full details on both imbalance method and experimental setup are provided in Appendix B.1. Accuracy Curves We compare the classification accuracy with and without spectral normalization in Figure 2. In addition to class imbalance, we consider settings where we randomly remove training samples (left graph). We show this setting to determine that the accuracy gain/loss from output feature preservation is dependent on the available information, not the number of samples. Our experiments highlight both advantages and disadvantages of preserving features in the output representation: if the target distribution is sufficiently similar to the input distribution, additional preserved features correlate with the generalization objective and results in a performance increase. This can be seen from the accuracy improvement with spectral normalization under low imbalance severities or when samples are removed randomly. The opposite can be observed where, in contrast to random sample removal (left graph), we explicitly remove training and test samples to imbalance classes (right graph). Here, the training set contains more significant amounts of irrelevant information and spectral normalization significantly decreases the generalization performance. 4. Our Method: TULIP 4.1. Transitional Feature Preservation Within the previous sections we found that feature preservation in the output can oppose the learning objective; an undesirable property for neural networks. To this end, a more prudent strategy involves preserving distance information in a collection of representations instead of a single output. Classic models that implement this property are ensembles (Lakshminarayanan et al., 2017). Uncertainty is encoded in the collection of ensemble models without explicit preservation constraints and the feature distance is then preserved in the difference or spread of the individual representations (Malinin et al., 2019). We formalize the concept within the context of distance preservation: given a set of neural network representations {hw1(x), ..., hw N (x)} and a transitional function h : H1 H2 ... HN V, we seek to learn feature mappings that allow a distance d V within the transitional space that reflects the true distance d X: d V ( h(x1), h(x2)) = d X(x1, x2) (4) We refer to methods along Equation 4 as transitional feature preservation or TFP in short. While powerful iterative methods such as ensembles implement TFP, their evaluation requires several forward passes and is frequently infeasible due to time or space limitations. In this paper, we implement TFP in single-pass uncertainty estimation by considering a linear combination of intermediate layer representations within the neural network. In particular, we find that the linear combination of intermediate distances PL l=0 rld Hl(hwl(x1), hwl(x2)) is distance preserving when the first layer is collapse resistant (i.e. d H0(hw0(x1), hw0(x2)) = 0) for d X(x1, x2) = 0). This requirement is different from Equation 2 as it allows distance contraction or expansion of d H0 with respect to d X, not full preservation. Proposition 4.1 makes the concept Transitional Uncertainty with Layered Intermediate Predictions more precise. Proposition 4.1 (Transitional Feature Preservation in Intermediate Representations). Consider the neural network mapping hw : X H with the layered architecture hw = hw0 hw1... hw L, where the first layer hw0 is collapse resistant with respect to the input space, d H0(hw0(x1), hw0(x2)) = 0 for d X(x1, x2) = 0. Then there exists a C R such that l=0 rld Hl(hwl(x1), hwl(x2)) = C d X(x1, x2), where C = 1 under an appropriate choice of rl. In other words, there exists a linear combination of intermediate representations that is feature preserving in transition - i.e. satisfies Equation 4. We provide proof of Proposition 4.1 in Appendix A.2. Note that Proposition 4.1 assumes collapse resistance in the first layer. In practice, this can be achieved by enforcing preservation constraints such as spectral normalization on the first layer only, which theoretically comes with the same risks outlined in Section 3 (to a significantly lesser degree). Empirically, we found that omitting the constraint entirely does not compromise the quality of the uncertainty estimates suggesting that the first layer rarely collapses data points in practice. 4.2. Algorithm Our method consists of three components: a principal feedforward network, a constellation of shallow-deep network exits with individual internal classifiers, and a combination head (Figure 3). During training, the shallow-deep network exits are trained jointly with the feed-forward component, while the combination head is fitted after optimization on a validation set extracted from the training data XID. We emphasize that our method does not require out-of-distribution validation samples as the combination head is fitted with data from the training set. During inference, the input sample traverses both the feed-forward network, as well as the shallow exits. The final output prediction is derived from the main network, while the uncertainty score is derived from a combination of the intermediate output logits. In addition to our description in the main paper, we provide implementation details and algorithm pseudo code in Appendix D. Shallow-Deep Network Exits Shallow-Deep networks (SDNs) (Kaya et al., 2019) were originally introduced in the context of computation reduction. A SDN is a modified version of a conventional DNN where additional internal classifiers are placed on intermediate representations to produce preliminary predictions. In our context, we utilize internal classifiers to produce intermediate logits for uncertainty estimation. Formally, given the intermediate layer l of the principal feed-forward network with the latent representation hl wl(x) RMl, an internal classifier is a shallow network f l wl : RMl RK that produces prediction logits for K classes. Regarding positioning, we place internal classifiers uniformly across the feed forward network (Kaya et al., 2019). Training Procedure We train internal classifiers jointly with the feed-forward network, similar to (Kaya et al., 2019). For this purpose, we propose a weighted loss function for each internal classifier, as well as the principal network output. Given the final logits of the principle network hw(x), an appropriate proper scoring rule (or loss) Ls, the SDN Loss is defined as LSDN(x, y) = p0Ls(hw(x), y) + i=1 pi Ls(f i wl(x), y). where pi represent individual loss weights, and NIC is the total number of internal classifiers. In our experiments, we found that equal weighting (i.e. pk = 1 NIC+1)) produces sufficient results but emphasize that other weighting schemes may be more desirable depending on the application. Combination Head While heavily utilizing early layer representations is intuitive, we found that relying on shallow outputs exclusively results in poor uncertainty scores. Specifically, in addition to increased domain awareness, early intermediate outputs contain significant amounts of application-irrelevant information that results in unreliable uncertainty estimates. For this purpose, we propose a fusion scheme by first calculating individual uncertainty scores and subsequently combining them by a weighted scaled sum. ufinal(x) = 1 PNIC i=1 ri i=1 rius(f i wi(x)). (6) In Equation 6, us represents the individual uncertainty score, and ri the score weights. While principally any uncertainty score can be used, we replace the output layer of the internal classifiers with approximate gaussian processes and calculate uncertainty similar to (Liu et al., 2020). Our choice is based on simplicity and the strong performance of approximate Gaussian processes. We call our resulting method Transitional Uncertainty with Layered Intermediate Predictions or TULIP in short. Fitting the Combination Head Fitting the weights ri in Equation 6 is more involved. While several methods exist to fit uncertainty parameters (Lee et al., 2023; 2018; Liang et al., 2017), they require either a) access to a small set of XOOD, and/or b) access to all labels in XID. In this Transitional Uncertainty with Layered Intermediate Predictions Figure 3. Workflow of our method during inference. The architecture consists of a main network, internal classifiers (IC), as well as a combination head. During inference, the input traverses the main network, as well as the internal classifiers. The prediction is obtained from the main network output, while the uncertainty score is obtained from a combination of internal classifier outputs. Table 1. OOD detection and classification accuracy on the CIFAR10 dataset. OOD AUROC Architecture Algorithms Runtime Space Complexity Preservation Constraint Accuracy CIFAR10-C CIFAR100-C SVHN Ensemble of 5 (Lakshminarayanan et al., 2017) 5x 5x No Constraint 96.327 0.045 0.725 0.002 0.914 0.003 0.943 0.008 DNN 1x 1x No Constraint 95.533 0.080 0.715 0.007 0.903 0.000 0.927 0.015 Energy-Based (Liu et al., 2021) 0.688 0.016 0.879 0.003 0.881 0.006 DUN(Antor an et al., 2020) 1x 1.5x Div. Loss 94.675 0.045 0.713 0.003 0.881 0.006 0.880 0.008 Early Exit(Qendro et al., 2021) 1x 1.5x Div. Loss 93.510 0.100 0.714 0.005 0.881 0.007 0.870 0.009 SNGP(Liu et al., 2020) 1x 1x SN 95.033 0.076 0.721 0.007 0.928 0.005 0.976 0.003 DUQ(Van Amersfoort et al., 2020) 1x 1x GP 88.867 0.211 0.618 0.003 0.824 0.008 0.829 0.016 TULIP 1x 1.5x No Constraint 94.880 0.324 0.738 0.006 0.936 0.003 0.946 0.012 Res Net-101 Ensemble of 5 (Lakshminarayanan et al., 2017) 5x 5x No Constraint 96.237 0.057 0.706 0.001 0.908 0.000 0.939 0.004 DNN 1x 1x No Constraint 95.837 0.103 0.690 0.006 0.894 0.001 0.922 0.012 Energy-Based (Liu et al., 2021) 0.715 0.006 0.849 0.009 0.890 0.034 DUN(Antor an et al., 2020) 1x 1.5x Div. Loss 95.645 0.075 0.711 0.003 0.889 0.005 0.917 0.014 Early Exit(Qendro et al., 2021) 1x 1.5x Div. Loss 94.255 0.085 0.706 0.001 0.875 0.003 0.867 0.045 SNGP(Liu et al., 2020) 1x 1x SN 91.907 0.183 0.636 0.009 0.912 0.015 0.906 0.018 DUQ(Van Amersfoort et al., 2020) 1x 1x GP 89.427 0.315 0.620 0.003 0.833 0.004 0.830 0.004 TULIP 1x 1.5x No Constraint 94.257 0.349 0.722 0.006 0.937 0.003 0.938 0.004 Res Net-152 Ensemble of 5 (Lakshminarayanan et al., 2017) 5x 5x No Constraint 96.330 0.160 0.706 0.001 0.910 0.000 0.945 0.000 DNN 1x 1x No Constraint 95.877 0.097 0.690 0.006 0.891 0.000 0.928 0.005 Energy-Based (Liu et al., 2021) 0.665 0.008 0.849 0.004 0.905 0.018 DUN(Antor an et al., 2020) 1x 1.5x Div. Loss 95.850 0.190 0.693 0.006 0.884 0.003 0.883 0.019 Early Exit(Qendro et al., 2021) 1x 1.5x Div. Loss 94.340 0.085 0.679 0.000 0.836 0.001 0.796 0.000 SNGP(Liu et al., 2020) 1x 1x SN 90.510 0.814 0.636 0.009 0.899 0.016 0.847 0.020 DUQ(Van Amersfoort et al., 2020) 1x 1x GP 91.263 0.185 0.623 0.002 0.731 0.059 0.842 0.012 TULIP 1x 1.5x No Constraint 94.213 0.777 0.722 0.006 0.927 0.004 0.945 0.004 paper, we assume neither. Our fitting algorithm requires two steps. First, we derive proxy labels from the small validation set. Second, we derive the weight parameters by formulating a binary classification problem, using the individual uncertainty scores us(f i wi(x)), and proxy labels s(x) derived from the disagreement in between different SDN exits. In Appendix C.2, we study the proxy labels in detail and compare disagreement in SDN exits with ensembles. Note, that our algorithm assumes that the validation set represents a small amount of unlabeled samples originating from XID and does not require out-of-distribution samples or additional data of any kind. We define our proxy labels through disagreement in the form of prediction switches between internal classifiers. Given a validation sample xval, we define a prediction switch as s(xval) = bool( i=2 1f iwi!=f i 1 wi 1 < Ns) (7) where f i wi are abbreviations for the internal classifier predictions pred(f i wi(xval), and 1f iwi!=f i 1 wi 1 is a binary variable reducing to one if two subsequent classifier predictions differ or zero otherwise. Ns represents the number of switches to determine a positive sample and is an integer in the range [1, NIC]. Our choice regarding the disagreement label is based on simplicity. By evaluating prediction switches, we reduce the tuning process to a binary classification problem allowing a partition of the validation set into coarse high-, and one low-uncertainty subgroups. Specifically, we classify the sample xval as high-uncertainty if s(xval) amounts to one and as low-uncertainty otherwise. Subsequently, Transitional Uncertainty with Layered Intermediate Predictions Table 2. OOD detection and classification accuracy on the CIFAR100 dataset. OOD AUROC Architecture Algorithms Preservation Constraint Accuracy CIFAR10-C CIFAR100-C SVHN Ensemble of 5 (Lakshminarayanan et al., 2017) No Constraint 79.475 0.265 0.828 0.003 0.680 0.002 0.857 0.009 DNN No Constraint 77.623 0.276 0.832 0.003 0.680 0.004 0.859 0.013 Energy-Based (Liu et al., 2021) 0.829 0.008 0.682 0.006 0.815 0.050 DUN(Antor an et al., 2020) Div. Loss 75.760 0.440 0.774 0.002 0.648 0.000 0.789 0.009 Early Exit(Qendro et al., 2021) Div. Loss 70.715 0.065 0.823 0.004 0.883 0.009 0.815 0.014 SNGP(Liu et al., 2020) SN 75.083 0.889 0.821 0.010 0.707 0.005 0.900 0.008 DUQ(Van Amersfoort et al., 2020) GP - - - - TULIP No Constraint 78.437 0.223 0.868 0.002 0.738 0.002 0.955 0.008 Res Net-101 Ensemble of 5 (Lakshminarayanan et al., 2017) No Constraint 79.545 0.015 0.829 0.002 0.680 0.002 0.849 0.006 DNN No Constraint 77.257 0.285 0.834 0.003 0.671 0.005 0.851 0.024 Energy-Based (Liu et al., 2021) 0.836 0.004 0.674 0.007 0.856 0.042 DUN(Antor an et al., 2020) Div. Loss 77.680 0.013 0.780 0.001 0.647 0.000 0.783 0.002 Early Exit(Qendro et al., 2021) Div. Loss 71.215 0.045 0.810 0.001 0.670 0.002 0.828 0.005 SNGP(Liu et al., 2020) SN 74.380 1.978 0.816 0.028 0.674 0.031 0.906 0.027 DUQ(Van Amersfoort et al., 2020) GP - - - - TULIP No Constraint 78.550 0.213 0.863 0.004 0.730 0.002 0.959 0.001 Res Net-152 Ensemble of 5 (Lakshminarayanan et al., 2017) No Constraint 78.487 0.133 0.725 0.002 0.680 0.002 0.822 0.002 DNN No Constraint 78.160 0.242 0.830 0.002 0.674 0.002 0.851 0.010 Energy-Based (Liu et al., 2021) 0.828 0.002 0.674 0.002 0.852 0.009 DUN(Antor an et al., 2020) Div. Loss 78.845 0.025 0.721 0.004 0.637 0.003 0.796 0.009 Early Exit(Qendro et al., 2021) Div. Loss 72.490 0.005 0.816 0.007 0.675 0.006 0.828 0.005 SNGP(Liu et al., 2020) SN 74.077 2.631 0.822 0.028 0.659 0.039 0.889 0.015 DUQ(Van Amersfoort et al., 2020) GP - - - - TULIP No Constraint 78.877 0.311 0.857 0.010 0.715 0.010 0.926 0.031 we derive the weighting parameters through logistic regression, where we map the individual uncertainty scores to the us(f i wi(xval)) to the corresponding subgroup s(xval): r1, ..., r NIC = LR({s(xi), vi}Nval i=1 ) (8) In our notation, LR is an abbreviation for logistic regression, and vi are the individual uncertainty scores [us(f 1 w1(xi)), ..., us(f NIC w NIC (xi))] bundled into a single vector. 5. Related Work Our work most closely relates with estimating uncertainty in a single forward pass. Several initial studies in singlepass uncertainty estimation were considered within the context of online settings or regression tasks. Here, notable methods include quantile regression (Koenker & Bassett, 1978), conformal prediction (Shafer & Vovk, 2008), or direct variance prediction (Nix & Weigend, 1994). The initial concepts were followed by a large body of uncertainty estimation methods including but not limited to replacing the loss function (Malinin & Gales, 2018; Hein et al., 2019; Sensoy et al., 2018), the output layer (Liu et al., 2021; Padhy et al., 2020; Bendale & Boult, 2016; Macˆedo & Ludermir, 2022; Tagasovska & Lopez-Paz, 2019; Gast & Roth, 2018), alternative gradient representations (Kwon et al., 2020; Prabhushankar & Al Regib; Lee et al., 2023), or interval arithmetic (Oala et al., 2020). Notably, (Gast & Roth, 2018) modify the network to produce probabilistic outputs instead of point estimates and (Oala et al., 2020) propagate error intervals directly through the neural network. While several methods are promising, they do not explicitly consider information preservation within the representations of the network and are susceptible to collapsing features to single entities. As a result, recent methods consider feature preservation within the output as a vital component for reliable uncertainty scores (Liu et al., 2020; Van Amersfoort et al., 2020; van Amersfoort et al., 2021; Mukhoti et al., 2023). Our work complements these approaches by considering feature preservation with intermediate representations without explicit constraints. Further, additional approaches exist for uncertainty estimation (Lee et al., 2023; 2018; Guo et al., 2017; Liang et al., 2017) where several explore intermediate representations. However, they assume access to out-ofdistribution validation samples and/or a fully labeled training set. Our work is complementary by investigating feature preservation without access to a out-of-distribution validation set and does not assume that all samples are labeled in training. Additionally, there exists several recent studies to bayesian models (Kandemir et al., 2021). While theoretically appealing, the framework relies on a bayesian model which requires several iterative forward passes for evaluation. Finally, our work closely relates to several approaches utilizing early-exit neural networks with the application of uncertainty estimation (Qendro et al., 2021; Antor an et al., 2020). While initial results are promising, these approaches utilize a diversity loss term which can be interpreted as a feature preservation constraint along Equation 2. Our work is complementary by utilizing early-exit neural networks without feature preservation constraints. We compare against two other early-exit approaches in our experiments. In Appendix E, we provide further related work on iterative uncertainty estimation, as well as distance-preserving neural networks. Transitional Uncertainty with Layered Intermediate Predictions Table 3. AUROC and classification accuracy on on the organ{C, A, S} datasets. AUROC Dataset Algorithms Preservation Constraint Accuracy organ A organ C organ S DNN No Constraint 94.602 0.388 - 0.906 0.004 0.850 0.007 Energy-Based (Liu et al., 2021) - 0.887 0.005 0.841 0.008 DUN(Antor an et al., 2020) Div. Loss 93.784 0.106 - 0.850 0.008 0.747 0.006 Early Exit(Qendro et al., 2021) Div. Loss 92.009 0.213 - 0.894 0.004 0.732 0.002 SNGP(Liu et al., 2020) SN 93.906 0.297 - 0.907 0.010 0.857 0.006 TULIP No Constraint 95.254 0.191 - 0.915 0.003 0.869 0.003 DNN No Constraint 92.106 0.176 0.884 0.001 - 0.780 0.005 Energy-Based (Liu et al., 2021) 0.857 0.001 - 0.751 0.006 DUN(Antor an et al., 2020) Div. Loss 91.231 0.000 0.874 0.002 - 0.801 0.001 Early Exit(Qendro et al., 2021) Div. Loss 90.826 0.236 0.867 0.000 - 0.791 0.001 SNGP(Liu et al., 2020) SN 90.941 0.530 0.849 0.008 - 0.765 0.007 TULIP No Constraint 92.122 0.227 0.894 0.003 - 0.794 0.003 DNN No Constraint 80.258 0.299 0.754 0.010 0.815 0.003 - Energy-Based (Liu et al., 2021) 0.733 0.012 0.775 0.005 - DUN(Antor an et al., 2020) Div. Loss 81.657 0.198 0.751 0.003 0.824 0.005 - Early Exit(Qendro et al., 2021) Div. Loss 78.888 0.000 0.747 0.001 0.815 0.003 - SNGP(Liu et al., 2020) SN 79.918 0.280 0.707 0.009 0.790 0.010 - TULIP No Constraint 80.002 0.126 0.778 0.001 0.822 0.003 - 6. Benchmark Experiments 6.1. CIFAR10 and CIFAR100 We start with standardized benchmarks in out-of-distribution (OOD) detection. The following combinations are evaluated: CIFAR10 vs. CIFAR10-C/CIFAR100-C/SVHN and CIFAR100 vs. CIFAR10-C/CIFAR100-C/SVHN (Krizhevsky et al., 2009; Netzer et al., 2011; Hendrycks & Dietterich, 2019). In addition to a standard DNN, we compare against five single-pass uncertainty baselines that do not require additional OOD data: the energy-based model (Liu et al., 2021), DUN (Antor an et al., 2020), early-exit ensembles (Qendro et al., 2021), DUQ (Van Amersfoort et al., 2020), and SNGP (Liu et al., 2020). We choose these three methods because 1) their strong empirical performance and 2) they utilize four popular methods for feature preservation in the network output. DUQ preserves features with a double sided gradient penalty (GP) (Gulrajani et al., 2017), SNGP implements spectral normalization (SN) (Miyato et al., 2018), DUN and early-exit ensembles utilize a an additional diversity loss term (Div. loss) (Qendro et al., 2021; Antor an et al., 2020), and the energy based model relies on the softmax density without regularization (No constraint). To investigate robustness with respect to network complexity, we consider three architectures with residual connections and varying depth: Res Net architectures (He et al., 2016) with 50, 101, and 152 layers. We restrict our experiments to these architectures as SNGP requires residual connections for its functionality. For fine-tuning the uncertainty weights, we partition 10% of training the samples and remove the labels to perform the unsupervised fitting algorithm. In Table 1 and Table 2, we report the AUROC scores for training on CIFAR10 and CIFAR100 respectively. When evaluating the corruption datasets, CIFAR10-C and CIFAR100-C, we average all corruption types and intensities. Further details on implementation and feature preservation can be found in Appendix B.4 and B.3. In addition, we investigate calibration, runtime, and imbalanced settings in Appendix C.4, Our method outperforms the other single-pass methods despite having access to less training annotations. This holds true over varying architectures and training datasets. In particular, methods trained with spectral normalization achieve lower accuracy with increasing network depth. For instance, the accuracy of SNGP reduces nearly three percent when extending Res Net-50 to Res Net-101. We relate this observation to the scaling of the Lipschitz bounds. Specifically, the lower and upper bounds scale exponentially with the number of layers and are tighter for shallow models while looser for deeper ones. Hence, the preservation constraint is weaker for deeper models and fails to maintain the distance between data-points. The remaining methods deploy a different feature preservation strategy and are thus agnostic to this effect. Further, DUQ did not converge on CIFAR100 due to training instabilities. These arise when the class centroids get noisy from increasing class and data complexity. Finally, we acknowledge the runtime and space complexity. While our method is favorable in terms of performance adding internal classifiers does increase the space complexity of the model. However, the effect does not scale with the network depth and is significantly less than ensembles which is 2x minimum. 6.2. Medical Modalities Further limitations of current single-pass methods include their strong modification the baseline algorithm. This renders them difficult to scale to different data modalities. For this purpose, we consider a medical setting where the training data contains different information as the test set. We benchmark TULIP on three CT scan datasets from (Yang et al., 2023). All three datasets contain CT scans of the same eleven body organs and are named after the three planes (axial, coronal, and saggittal) in which the data was collected. In our experiments, we train on one plane and perform misclassification detection on the combined test Transitional Uncertainty with Layered Intermediate Predictions Table 4. OOD detection and classification accuracy on the Image Net dataset. Algorithms Accuracy ECE Image Net-C DNN 72.617 0.019 1.138 0.005 0.867 0.006 Energy-Based(Liu et al., 2021) 0.876 0.004 MC-Dropout(Gal & Ghahramani, 2016) 0.837 0.001 0.877 0.000 TULIP 72.617 0.019 0.678 0.001 0.885 0.000 set of the original plane and an additional plane. We report AUROC and accuracy on the in-domain test set in Table 3. Each row shows a different training set, while each column refers to the test set that is combined with the in-distribution test set. All experiments are performed with a Res Net-50 architecture and we use the same hyperparameters as in our previous experiments. TULIP matches or outperforms competing methods. The observation is important as it shows that our modifications scale to complex data modalities which is not the case for several other single-pass methods. In particular, algorithms with strong architectural modifications such as SNGP does not perform well in the medical modality. The result is expected due to imbalance in the dataset. In particular, SNGP is challenged when the information within the training set does not correlate well with the test set and supports our usage of TFP in TULIP. These results are complementary and support conclusions from existing studies (Postels et al., 2022). 6.3. Image Net In this section, we investigate how TULIP adapts when internal classifiers are not trained together with the baseline architecture. A further limitation of several single-pass uncertainty estimators is that their implementation requires a significant change to the existing architecture. In several applications the modification is non-trivial, especially in large-scale benchmarks. For this purpose, we consider settings where we do not modify the baseline architecture, allowing the same generalization performance to a standard DNN. In our experiments, we place individual internal classifiers on top of a Res Net-101 architecture fully trained on imagenet and compare against other methods that require no change to the existing architecture. In our experiments, this is the energy-based model (Liu et al., 2021) as a single-pass example and MC-Dropout (Gal & Ghahramani, 2016) as an iterative example. For MC-Dropout, we perform inference in a single forward pass without any dropout layers rendering the same behavior as the baseline DNN. To derive uncertainty estimates, we perform several forward passes with dropout enabled making it an iterative method. We show our results in Table 4 and provide experimental details in Appendix B.5. From our results, we see that TULIP outperforms several other methods on large scale benchmarks. We interpret these results as a further testament to the feature extraction capabilities of intermediate represen- tations. In particular, our results show that the intermediate representations do not need to be explicitly trained to extract additional information but can be used post-hoc. We further note that TULIP can scale to large scale benchmarks which is not the case for other single-pass methods that need to be trained from scratch. 7. Discussion and Limitations A central observation we made in this work is that enforcing feature preservation by constraining model representations can be harmful to the model performance. We highlight application relevance as a key requirement for effective representation constraints. In practice, the characteristic can be undesirable as training distributions can severely differ from deployment. We propose single-pass transitional feature preservation through intermediate representations to address these disadvantages. While our approach is effective, we do not claim that our improvements solve feature preservation in uncertainty estimation entirely. In particular, we propose one instance of TFP through SDNs that comes with its own set of limitations: similar to iterative methods, the success depends on the amount of source representations in h to preserve features. For this purpose, SDNs are less effective on shallow architectures with fewer intermediate options to extract from. Further, we acknowledge that implementing internal classifiers comes with a large set of hyperparameters increasing space complexity. In our implementation, we use standard settings that show promising performance without significantly impacting the space complexity. However, other applications may require more elaborate hyperparameter explorations and additional internal classifiers. For this reason, we see reducing the space complexity of internal classifiers or the development of efficient hyperparameter tuners as promising research directions. Finally, we chose the combination of SDNs with Gaussian Processes out of simplicity and the strong empirical performance. However, key novelties of this paper (SDNs and singe-pass TFP) are not limited to one uncertainty score (Gaussian Processes) and are building blocks for single-pass uncertainty methods. We encourage researchers to implement different combinations of single-pass TFP and uncertainty scores to advance the field of single-pass uncertainty estimation. Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal Transitional Uncertainty with Layered Intermediate Predictions consequences of our work, none which we feel must be specifically highlighted here. Abraham, I., Bartal, Y., and Neimany, O. Advances in metric embedding theory. In Proceedings of the thirtyeighth annual ACM symposium on Theory of computing, pp. 271 286, 2006. Antor an, J., Allingham, J., and Hern andez-Lobato, J. M. Depth uncertainty in neural networks. Advances in neural information processing systems, 33:10620 10634, 2020. Arjovsky, M., Chintala, S., and Bottou, L. Wasserstein gan, 2017. Behrmann, J., Grathwohl, W., Chen, R. T., Duvenaud, D., and Jacobsen, J.-H. Invertible residual networks. In International Conference on Machine Learning, pp. 573 582. PMLR, 2019. Bendale, A. and Boult, T. E. Towards open set deep networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1563 1572, 2016. Chennuru Vankadara, L. and von Luxburg, U. Measures of distortion for machine learning. Advances in Neural Information Processing Systems, 31, 2018. Cover, T. M. and Thomas, J. A. Elements of Information Theory (Wiley Series in Telecommunications and Signal Processing). Wiley-Interscience, USA, 2006. ISBN 0471241954. Dinh, L., Krueger, D., and Bengio, Y. Nice: Non-linear independent components estimation. ar Xiv preprint ar Xiv:1410.8516, 2014. Dinh, L., Sohl-Dickstein, J., and Bengio, S. Density estimation using real nvp. ar Xiv preprint ar Xiv:1605.08803, 2016. Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et al. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv:2010.11929, 2020. Dusenberry, M., Jerfel, G., Wen, Y., Ma, Y., Snoek, J., Heller, K., Lakshminarayanan, B., and Tran, D. Efficient and scalable bayesian neural nets with rank-1 factors. In International conference on machine learning, pp. 2782 2792. PMLR, 2020. Gal, Y. and Ghahramani, Z. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In international conference on machine learning, pp. 1050 1059. PMLR, 2016. Gast, J. and Roth, S. Lightweight probabilistic deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3369 3378, 2018. Gouk, H., Frank, E., Pfahringer, B., and Cree, M. J. Regularisation of neural networks by enforcing lipschitz continuity. Machine Learning, 110:393 416, 2021. Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., and Courville, A. C. Improved training of wasserstein gans. Advances in neural information processing systems, 30, 2017. Guo, C., Pleiss, G., Sun, Y., and Weinberger, K. Q. On calibration of modern neural networks. In International conference on machine learning, pp. 1321 1330. PMLR, 2017. He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Hein, M., Andriushchenko, M., and Bitterwolf, J. Why relu networks yield high-confidence predictions far away from the training data and how to mitigate the problem. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 41 50, 2019. Hendrycks, D. and Dietterich, T. Benchmarking neural network robustness to common corruptions and perturbations. ar Xiv preprint ar Xiv:1903.12261, 2019. Jacobsen, J.-H., Smeulders, A., and Oyallon, E. irevnet: Deep invertible networks. ar Xiv preprint ar Xiv:1802.07088, 2018. Kandemir, M., Akg ul, A., Haußmann, M., and Unal, G. Evidential turing processes. Co RR, abs/2106.01216, 2021. URL https://arxiv.org/abs/2106.01216. Kaya, Y., Hong, S., and Dumitras, T. Shallow-deep networks: Understanding and mitigating network overthinking. In International conference on machine learning, pp. 3301 3310. PMLR, 2019. Kendall, A. and Gal, Y. What uncertainties do we need in bayesian deep learning for computer vision?, 2017. Koenker, R. and Bassett, G. Regression quantiles. Econometrica, 46(1):33 50, 1978. ISSN 00129682, 14680262. URL http://www.jstor. org/stable/1913643. Transitional Uncertainty with Layered Intermediate Predictions Krizhevsky, A., Hinton, G., et al. Learning multiple layers of features from tiny images. 2009. Kwon, G., Prabhushankar, M., Temel, D., and Al Regib, G. Backpropagated gradient representations for anomaly detection. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XXI 16, pp. 206 226. Springer, 2020. Lakshminarayanan, B., Pritzel, A., and Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles. Advances in neural information processing systems, 30, 2017. Lawrence, N. D. and Qui nonero Candela, J. Local distance preservation in the gp-lvm through back constraints. In Proceedings of the 23rd International Conference on Machine Learning, ICML 06, pp. 513 520, New York, NY, USA, 2006. Association for Computing Machinery. doi: 10.1145/1143844.1143909. Lee, J., Lehman, C., Prabhushankar, M., and Al Regib, G. Probing the purview of neural networks via gradient analysis. IEEE Access, 11:32716 32732, 2023. Lee, K., Lee, K., Lee, H., and Shin, J. A simple unified framework for detecting out-of-distribution samples and adversarial attacks. Advances in neural information processing systems, 31, 2018. Liang, S., Li, Y., and Srikant, R. Enhancing the reliability of out-of-distribution image detection in neural networks. ar Xiv preprint ar Xiv:1706.02690, 2017. Liu, J., Lin, Z., Padhy, S., Tran, D., Bedrax Weiss, T., and Lakshminarayanan, B. Simple and principled uncertainty estimation with deterministic deep learning via distance awareness. Advances in Neural Information Processing Systems, 33:7498 7512, 2020. Liu, W., Wang, X., Owens, J. D., and Li, Y. Energy-based out-of-distribution detection, 2021. Macˆedo, D. and Ludermir, T. Enhanced isotropy maximization loss: Seamless and high-performance out-ofdistribution detection simply replacing the softmax loss, 2022. Malinin, A. and Gales, M. Predictive uncertainty estimation via prior networks, 2018. Malinin, A., Mlodozeniec, B., and Gales, M. Ensemble distribution distillation. ar Xiv preprint ar Xiv:1905.00076, 2019. Miyato, T., Kataoka, T., Koyama, M., and Yoshida, Y. Spectral normalization for generative adversarial networks. ar Xiv preprint ar Xiv:1802.05957, 2018. Mukhoti, J., Kirsch, A., van Amersfoort, J., Torr, P. H., and Gal, Y. Deep deterministic uncertainty: A new simple baseline. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 24384 24394, 2023. Naeini, M. P., Cooper, G., and Hauskrecht, M. Obtaining well calibrated probabilities using bayesian binning. In Proceedings of the AAAI conference on artificial intelligence, volume 29, 2015. Netzer, Y., Wang, T., Coates, A., Bissacco, A., Wu, B., and Ng, A. Y. Reading digits in natural images with unsupervised feature learning. 2011. Nix, D. A. and Weigend, A. S. Learning local error bars for nonlinear regression. In Proceedings of the 7th International Conference on Neural Information Processing Systems, NIPS 94, pp. 489 496, Cambridge, MA, USA, 1994. MIT Press. Oala, L., Heiß, C., Macdonald, J., M arz, M., Samek, W., and Kutyniok, G. Interval neural networks: Uncertainty scores. ar Xiv preprint ar Xiv:2003.11566, 2020. Osawa, K., Swaroop, S., Khan, M. E. E., Jain, A., Eschenhagen, R., Turner, R. E., and Yokota, R. Practical deep learning with bayesian principles. Advances in neural information processing systems, 32, 2019. Padhy, S., Nado, Z., Ren, J., Liu, J., Snoek, J., and Lakshminarayanan, B. Revisiting one-vs-all classifiers for predictive uncertainty and out-of-distribution detection in neural networks, 2020. Perrault-Joncas, D. and Meila, M. Metric learning and manifolds: Preserving the intrinsic geometry. Preprint Department of Statistics, University of Washington, 2012. Postels, J., Seg u, M., Sun, T., Sieber, L. D., Van Gool, L., Yu, F., and Tombari, F. On the practicality of deterministic epistemic uncertainty. In Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., and Sabato, S. (eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp. 17870 17909. PMLR, 17 23 Jul 2022. URL https://proceedings.mlr.press/ v162/postels22a.html. Prabhushankar, M. and Al Regib, G. Introspective learning: A two-stage approach for inference in neural networks. In Advances in Neural Information Processing Systems. Qendro, L., Campbell, A., Lio, P., and Mascolo, C. Early exit ensembles for uncertainty quantification. In Machine Learning for Health, pp. 181 195. PMLR, 2021. Transitional Uncertainty with Layered Intermediate Predictions Sensoy, M., Kaplan, L., and Kandemir, M. Evidential deep learning to quantify classification uncertainty. Advances in neural information processing systems, 31, 2018. Shafer, G. and Vovk, V. A tutorial on conformal prediction. Journal of Machine Learning Research, 9(3), 2008. Shwartz-Ziv, R. and Tishby, N. Opening the black box of deep neural networks via information. ar Xiv preprint ar Xiv:1703.00810, 2017. Tagasovska, N. and Lopez-Paz, D. Single-model uncertainties for deep learning. Advances in Neural Information Processing Systems, 32, 2019. Thiagarajan, J., Anirudh, R., Narayanaswamy, V. S., and Bremer, T. Single model uncertainty estimation via stochastic data centering. Advances in Neural Information Processing Systems, 35:8662 8674, 2022. Tishby, N., Pereira, F. C., and Bialek, W. The information bottleneck method. ar Xiv preprint physics/0004057, 2000. Trockman, A. and Kolter, J. Z. Patches are all you need? ar Xiv preprint ar Xiv:2201.09792, 2022. Van Amersfoort, J., Smith, L., Teh, Y. W., and Gal, Y. Uncertainty estimation using a single deep deterministic neural network. In International conference on machine learning, pp. 9690 9700. PMLR, 2020. van Amersfoort, J., Smith, L., Jesson, A., Key, O., and Gal, Y. On feature collapse and deep kernel learning for single forward pass uncertainty. ar Xiv preprint ar Xiv:2102.11409, 2021. Wen, Y., Tran, D., and Ba, J. Batchensemble: an alternative approach to efficient ensemble and lifelong learning. ar Xiv preprint ar Xiv:2002.06715, 2020. Wenzel, F., Roth, K., Veeling, B. S., Swiatkowski, J., Tran, L., Mandt, S., Snoek, J., Salimans, T., Jenatton, R., and Nowozin, S. How good is the bayes posterior in deep neural networks really? ar Xiv preprint ar Xiv:2002.02405, 2020. Yang, J., Shi, R., Wei, D., Liu, Z., Zhao, L., Ke, B., Pfister, H., and Ni, B. Medmnist v2-a large-scale lightweight benchmark for 2d and 3d biomedical image classification. Scientific Data, 10(1):41, 2023. Transitional Uncertainty with Layered Intermediate Predictions A.1. Mutual Information of Distances In this section, we discuss connecting the learning problem in Equation 1 with distances in the feature plane. To establish a dependency between Equation 1 and feature distances, we first define sets where the distance d X is bijective. Mathematically, this is equivalent to restricting sets to unique distance values; sets we define as unique distance sets. Definition A.1 formalizes the concept. Definition A.1 (Unique Distance Set and Partition). Consider the metric space (X, d X) with a corresponding random variable X p X describing the input distribution with density p X. We define a unique distance set M X as a set in X possessing unique distances with respect to d X and an arbitrary but fixed anchor point xa X. M = {xi, xj, xa X : d X(xa, xj) = d X(xa, xi), i = j = a} Further, we define a partition X = S k [1,Np] Mk : T k [1,Np] Mk = over unique distance sets Mk as a unique partition set. Equivalently, we define M p M and M k p M k as the corresponding random variables with their respective probability densities. With the help of unique distance sets, we can formulate the proof for Equation 3. Proof. We first formulate Equation 1 as an optimization objective over unique distance sets. Since the optimization over each individual set can be viewed as a separate learning task, we can rewrite the objective as a summary of the mutual information over individual distance sets: h (X) = argmin hw:{I(hw(M k);Y k)=I(M k;Y k), k [1,Np]} k I(hw(M k); M k) (9) Given the anchor point xk for Mk, f X(x) = d X(xk; x) represents an injection within the individual subset Mk. The characteristic is relevant, as we utilize the transformation invariance property of the mutual information. Assuming hw preserves the unique distance property according to Equation 2, we rewrite Equation 9 as h (X) = argmin hw:{I(hw(M k);Y k)=I(M k;Y k), k [1,N]} k I(hw(M k); M k) = argmin hw:{I(d H(hw(xk);hw(M k));Y k)=I(M k;Y k), k [1,N]} k I(d H(hw(xk); hw(M k)); d X(xk; M k)) = argmin hw:{I(f k H(hw(M k));Y k)=I(M k;Y k), k [1,Np]} k I(f k H(hw(M k)); f k X(M k)). A.2. Transitional Feature Preservation of Intermediate Representations In this section, we discuss the proof of Proposition 4.1. Proof. To prove Proposition 4.1, we utilize concepts of metric distortion from metric embedding theory (Abraham et al., 2006; Chennuru Vankadara & von Luxburg, 2018). Specifically, neural network layers can be characterized by the distortion they introduce to the input space. We define the network distortion coefficient of a given layer l as ρl(x1, x2) = ( d Hl(hwl(x1),hwl(x2)) d Hl 1(hwl 1(x1),hwl 1(x2)), if d Hl 1(hwl 1(x1), hwl 1(x2)) = 0 1, otherwise . (11) Transitional Uncertainty with Layered Intermediate Predictions The interpretation of Equation 11 is simple. If the previous layer does not collapse the input distances (d Hl 1(hwl 1(x1), hwl 1(x2)) = 0), the ratio between both distances characterizes the distance distortion of layer l. Specifically, we have distance contraction if ρl(x1, x2) is less than one. Here, zero represents the corner case when the layer collapses the input to a single point (d Hl(hwl(x1), hwl(x2)) = 0). Further, we have distance expansion when ρl(x1, x2) is greater than one, i.e. when the layer increases the distance between the points. Finally, we have perfect distance preservation in the case of ρl(x1, x2) = 1. When the previous layer collapses the input points, the layer l receives the same input for x1 and x2, resulting a perfect distance preservation as the distance between the same point is trivially zero. The following relationship between distortion and network distances is important. d Hl(hwl(x1), hwl(x2)) = ρl(x1, x2) d Hl 1(hwl 1(x1), hwl 1(x2)) (12) Equation 12 follows directly from the the fact that the input of a given layer l is the output of the previous layer hw = hw0 hw1... hw L. When the previous layer does not collapse the input Equation 12 directly follows from the definition of ρl. In the case of feature collapse within the previous layer (d Hl 1(hwl 1(x1), hwl 1(x2)) = 0), the input to the next layer is the collapsed point and d Hl(hwl(x1), hwl(x2)) = 0 satisfying Equation 12. Using the network distortion coefficient we can rewrite the linear combination of distances as a function of the input distances d X: d SDN( h(x1), h(x2)) = l=0 rld Hl(hwl(x1), hwl(x2)), l=0 rl d X(x1, x2) i=0 ρi(x1, x2), = d X(x1, x2) i=0 ρi(x1, x2), = d X(x1, x2) C. The first derivation follows from a recursive application of Equation 12. The second, from the independence of d X(x1, x2) from both i and l. We note that Equation 13 satisfies Equation 4 when an appropriate weight choice rl results in C = 1. A solution for rl only exists when the first layer is collapse resistant, i.e. when ρ0(x1, x2) = 0 for d X(x1, x2) = 0; a requirement for Proposition 4.1. A.3. Mutual Information of Intermediate Representations In this section, we discuss how intermediate representations aid in increasing the information of the full input distribution within the uncertainty source representation. As discussed in Section 2.1, effective uncertainty estimation is contingent on modelling information of the full input space X (not just XID) to differentiate the training distribution from the test distribution. Within the context of the mutual information in neural networks (Tishby et al., 2000), achieving this is equivalent to maintaining the mutual information between the uncertainty source representation Z (i.e. the representation used to compute the uncertainty u(Z)), and the input X. I(Z; X). (14) A differentiator for uncertainty estimators is therefore their uncertainty source Z. We show that combining intermediate layers has favorable uncertainty properties in comparison to a conventional neural network output - i.e. our method maintains the mutual information I(Z; X) more effectively. Transitional Uncertainty with Layered Intermediate Predictions In our algorithm, we measure uncertainty from a combination of intermediate layers hl wl instead of the final output. Within the context of mutual information, the joint representation Z = h1 w1, ..., h L w L maintains the following relationship for a layered neural network: I(h1 w1, ..., h L w L; X) I(hw(X); X) (15) The interpretation of Equation 15 is simple. Our method preserves information by extracting features before they are collapsed by subsequent network components. Hence, the mutual information with respect to the input is larger when intermediate representations are utilized in comparison to the final output exclusively. We further provide proof for Equation 15: Proof. For our discussion, we utilize data processing inequality (Cover & Thomas, 2006) within the context of neural networks. Specifically, given an intermediate layer hl wl the following relationship holds for any subsequent layers I(hl wl(X); X) I(hl+1 wl+1(X); X). (16) The mutual information of the joint variable Z = h1 w1, ..., h L w L and input X can be expressed with the chain rule of mutual information I(Z; X) = I(h1 w1(X); X) I(h2 w2(X), ..., h L w L(X); X|h1 w1(X)) = I(h1 w1(X); X) H[X] + H[X|h2 w2(X), ..., h L w L(X)] = I(h1 w1(X); X) I(hw(X); X) The first derivation comes from the chain rule of mutual information, the seconde from the definition of mutual information, and the third from the fact that layered neural networks form a Markov chain with X h1 w1(X) ... h L w L(X) (Tishby et al., 2000). The final inequality is a direct manifestation of Equation 16. B. Impelementation Details In this appendix, we provide details of the different experimental setups and comparison methods used in this paper. All experiments are implemented with pytorch. When a implementation was publicly available, we heavily relied on it in our own code. This is the case for DUQ (https://github.com/y0ast/deterministic-uncertaintyquantification), and SNGP (https://github.com/google/uncertainty-baselines/blob/master/baselines/imagenet/sngp.py, as well as https://github.com/y0ast/DUE). B.1. Surface Plots and Class Distribution Experiments Hyperparameter and Architecture Details In all experiments, we train a resnet-18 architecture (He et al., 2016) over 200 epochs and optimize with stochastic gradient descent with a learning rate of 0.01. We further decrease the learning rate by a factor of 0.2 in epochs 100, 125, 150, and 175 respectively, and use the data augmentations random crop, random horizontal flip, and cutout to increase the generalization performance. For our experiments, we deploy direct spectral normalization of the convolutional, and batch normalization layers to implement representational feature preservation. On the full CIFAR100 dataset, we achieve an overall classification accuracy of 77.41 % and 75.93 % for the model with and without spectral normalization respectively. We average our results over three random seeds. Transitional Uncertainty with Layered Intermediate Predictions Imbalanced CIFAR100 We imbalance the dataset as follows: for a certain subset of classes A {1, ..., K}, we reduce the the number of training samples by 80% and do not change the number of test samples. For a second subset B {1, ..., K}, A B = , we reduce the number of test samples by 90% and do not change the number of training samples. As a result, the first subset of classes contains few training samples and a large amount of test samples, while the other set suffers from the opposite problem. The imbalance severity can be adjusted by the number of classes in both imbalanced subsets A B {1, ..., K}. For simplicity, we keep the same amount of classes in both subsets A and B. In Figure 4, we show an overview of class distributions at different severity levels. Here, the top and bottom row contain the training and test distribution respectively. We perform our entire analysis on the CIFAR100 dataset as it represents a challenging benchmark with a large class variety. In addition, the dataset is fully balanced and contains the same class distribution for training and testing. Figure 4. Toy example of class distributions at different imbalance severity levels. Each column represents a different severity level, and each column the training and test set distribution respectively. B.2. Ablation Study We further provide an ablation study on utilizing different constellations of TULIP. We investigate equal weighting of the individual intermediate scores, as well as removing the GP layer. We show our results in Table 5. As shown, the final version of TULIP outperforms the other constellations further supporting the algorithm. Table 5. Weight Ablation Study. Algorithms CIFAR10-C CIFAR100-C SVHN TULIP - Equal Weighting 0.716 0.006 0.883 0.009 0.840 0.023 TULIP - No GP, Switch Weighting 0.722 0.005 0.885 0.008 0.811 0.031 TULIP 0.738 0.006 0.936 0.003 0.946 0.012 B.3. Comparison Method Details In the following, we provide details on each feature preservation method. Energy-Based Model (Liu et al., 2021): for the energy based model, uncertainty estimates are derived by replacing the final softmax layer with the unnormalized softmax density. No additional feature preservation constraint is used during training. In our experiments, we only compare against the version that does not require additional out-of-distribution data. Transitional Uncertainty with Layered Intermediate Predictions DUQ (Van Amersfoort et al., 2020): in DUQ, uncertainty is inferred by the closest kernel distance. To ensure the preservation of features, DUQ implements the double sided gradient penalty, penalizing the squared distance of the gradient from a fixed value at every input point. In contrast to spectral normalization, it implements feature preservation with a local constraint, without explicit guarantees for data points outside of XID. SNGP (Liu et al., 2020): SNGP infers uncertainty through distance awareness within the output. In the original paper, this is achieved by replacing the final output layer with a gaussian process approximation, and implementing spectral normalization in combination with residual layers. Direct spectral normalization on the weights provides a upper lipschitz bound while the combination with residual connections further ensures a lower lipschitz bound on the distance between two input points. Both bounds jointly result in feature preservation, as distances in between input points are approximately preserved when traversing the network. In comparison to the gradient penalty, spectral normalization enforces a global constraint. B.4. Out-of-Distribution Experiments on CIFAR10/100 In our out-of-distribution experiments, we use the same backbone residual architectures (Res Net-50, -101, and -152) with a batch size of 128. For all setups, we use the standard data augmentations random horizontal flip, random crop, and cutout. In the following, we describe the details for each method. All Results are averaged over three random seeds. Softmax DNN and Energy-Based Model We train both models with the SGD optimizer and an initial learning rate of 0.01. We optimize the model for 200 epochs and reduce the learning rate by a factor of 0.2 in epochs 100 and 150. For the energy-based model, we use the unnormalized softmax density, similar to other implementations (Mukhoti et al., 2023). DUQ Our DUQ models are trained with the SGD optimizer and a learning rate of 0.05. We train for 600 epochs and reduce the learning rate by factor 0.2 in epochs [300, 375, 450, 525]. For the gradient penalty weight, we perform the experiment with hyperparameters from the original paper (Van Amersfoort et al., 2020), as well as a newer implementation from (Postels et al., 2022), and report the constellation with the highest accuracy value. SNGP We trained SNGP with the SGD optimizer and an initial learning rate of 0.01. We reduce the learning rate by a factor of 0.2 in epochs 100 and 150, and train for 200 epochs. We further use a spectral normalization coefficient of three. Our Method We train each SDN model with the SGD optimizer using an initial learning rate of 0.01. Further, we optimize the architecture for 400 epochs and reduce the learning rate by a factor of 0.2 in epoch 200, and 300. The architecture of the internal classifiers is similar to (Kaya et al., 2019), with a single linear layer combined with a mixture of average-/max-pooling where necessary. The output layer is then fed into a GP layer, which has the same architecture as SNGP. We distribute the internal classifiers equally distanced across the network by placing a internal classifier on top of every third residual block. For Res Net-50 this is equivalent to every sixth layer, and every ninth layer for the remaining larger models. Our selection is geared towards simplicity and performance may be further improved with other uncertainty scores such as entropy or energy functions. We train each model with a equally weighted SDN loss. We set the switch parameter Ns to the average number of switches occurring in the validation set as a general rule, but set it to 1 for the small scale datasets where few switches occur and resulting in a negligible switch mean. B.5. Image Net For our experiments on Image Net, we train the baseline DNN with standard training settings similar to baseline recipe in https://pytorch.org/blog/how-to-train-state-of-the-art-models-using-torchvision-latest-primitives/. For our MC-Dropout implementation, we place one dropout layer in between every residual bottleneck block. We set the dropout probability to 0.01 when deriving uncertainty estimates. For our SDN implementation, we use the same number of internal classifiers as in our previous experiments and set Ns to the mean number of switches in the validation set. We choose this setting due to simplicity and it allows us to choose hyperparameters in an automated fashion. We train our model with an equally weighted SDN loss and train for 100 expochs with an adam optimizer and a learning rate of 0.0001. B.6. Artificial Dataset For our experiments in Section C.2, we train our models on an artificial spiral dataset with three different classes. Here, each spiral arm represents a class that starts in the center, and spirals for one full loop of 360 . In our ensemble experiments, we Transitional Uncertainty with Layered Intermediate Predictions use a three-layer MLP architecture in an ensemble of ten. During training, we use a SGD optimizer with a learning rate of 0.008 and train each ensemble element for 400 epochs. To measure disagreement in between layers, we adjust the MLP architecture into a SDN, by placing an internal classifier on the first, and second layer respectively. We train the model with an SDN loss as described in Equation 5, optimize with the adam variant of SGD, and select a learning rate of 0.001. Due to the higher loss complexity, we train the SDN model for 800 epochs. To further measure disagreement, we utilize the same measure as previous works (Mukhoti et al., 2023; Malinin et al., 2019). disagreement(x) = H[ 1 NIC/E i=1 p(y|x, wi)] 1 NIC/E i=1 H[p(y|x, wi)]. (18) Here, NIC/E denotes the number of internal classifiers or ensembles respectively, and p(y|x, wi) the target posterior distribution of the individual model elements. C. Additional Experiments C.1. Surface Plots for Distance Preservation under Class Imbalance In addition to class accuracy, we further wish to analyze uncertainty estimates under class imbalance. For this purpose, we plot both accuracy and the number of samples (sample concentration) with respect to imbalance severity and uncertainty scores (Figure 5). The accuracy plots provide information of the calibration capabilities in relation to class imbalance. Ideally, the uncertainty fully informs of the accuracy of a sample and the dependency is linear on the y-/z-plane (Guo et al., 2017). We note that a conventional neural network is not calibrated and overconfident in its prediction - the dependency is not linear. Spectral normalization significantly improves along this characteristic (top right plot), and improves linearity regardless of the imbalance severity. The bottom row complements our accuracy curves. For low imbalance severities the majority of samples concentrate low uncertainty/high accuracy regions on the x-/y-plane. However, the dependency inverts with increasing imbalance. Samples concentrate in high-uncertainty/low-accuracy regions complementing the accuracy decline in Figure 2. C.2. Disagreement Analysis of Proxy Labels In this section, we provide a detailed analysis on the uncertainty proxy scores used in TULIP. In our analysis, we compare the properties of intermediate representation with ensembles uncertainty scores and analyze the disagreement among both methods. Within the context of ensembles, disagreement is used to derive the difference or spread of ensembles and is frequently used directly as an uncertainty score in several contexts (Malinin et al., 2019). To showcase disagreement within intermediate representations, we compare against ensembles on an artificial spiral dataset (Figure 6). The left image shows the total uncertainty of an ensemble of ten and is among the most common usages of ensembles. The center shows disagreement among the different ensemble models as derived by previous work (Malinin et al., 2019), and approximates uncertainty occurring due to data scarcity (Kendall & Gal, 2017). Ideally, the measure is low where sufficient data samples are available (center of the spiral), and increases where little or no data is available. The right shows the same measure of disagreement, with the exception of measuring in between individual SDN outputs instead of ensemble models. We note, that ensembles exhibit high disagreement near the decision boundaries exclusively, while the SDN model comprehensively approximates data scarcity in between layers. Our observations can be explained along the intuition of feature preservation. Ensembles measure disagreement among the output of the entire network architectures, and collapse important information from the input distribution that can be leveraged for uncertainty estimation. In contrast, intermediate representations contain more information of the input distribution and provide a coarse measure of data scarcity. Overall, we gather that disagreement in between internal classifiers is a reasonable approach to determine the hyperparameters ri. Experimental details to produce Figure 6 can be found in Appendix B.6. C.3. Additional Architectures We further compare TULIP with additional architectures. Within this context, we explore both the vision transformer (Dosovitskiy et al., 2020), as well as the convmixer (Trockman & Kolter, 2022). For the vision transformer, we use a similar setup to our imagenet experiments in Section 6.3. For the convmixer, we investigate medical modalities similar to Transitional Uncertainty with Layered Intermediate Predictions Figure 5. Accuracy and number of samples (sample concentration) with respect to dataset imbalance and average uncertainty score. In the the top row we show accuracy, in the bottom row we show sample concentration. The left column represents a conventional DNN while the right shows feature distance preservation through spectral normalization. Table 6. OOD detection and classification accuracy on the Image Net dataset. Algorithms Accuracy Image Net-C DNN 78.546 0.000 0.822 0.000 Energy-Based(Liu et al., 2021) 0.819 0.000 MC-Dropout(Gal & Ghahramani, 2016) 0.839 0.001 TULIP 78.546 0.000 0.851 0.004 Section 6.2. We show our results in Table 6 and Table 7. Similar to our previous experiments, TULIP overwhelmingly matches or outperforms the comparison methods supporting our approach with transitional feature preservation. C.4. Calibration and Runtime We further investigate calibration as characteristic for uncertainty score quality. With calibration, we refer to the capability of the output score to be reflective of the actual generalization performance (Guo et al., 2017). For this purpose, we consider the expected calibration error (ECE) (Naeini et al., 2015) to measure miscalibration and show our results in Figure 8. Our setup is equivalent to our experiments on out-of-distribution detection and we further show the runtime normalized by the latency of a conventional DNN. Specifically, we measure the latency of a single batch for each model and divide by the latency of a conventional DNN. For all of our experiments we use a single NVIDIA Ge Force GTX 1080 Ti. We note, that all algorithms are equivalent in terms of runtime and that our method matches or outperforms comparable single-pass methods in the majority of benchmarked constellations. Our results show both the runtime benefits and a high uncertainty estimation quality for our method. C.5. Imbalanced Out-of-Distribution Experiments In addition to standard out-of-distribution detection, we consider less informative training sets in the form of class imbalance. For this purpose, we unbalance the full 100 classes of the CIFAR100 dataset, as described in our previous analysis in Transitional Uncertainty with Layered Intermediate Predictions Figure 6. Comparison of disagreement in ensembles and intermediate representations trained on an artificial spiral dataset. The left image represents the total uncertainty of an ensemble of ten. In the center, we display model disagreement within the ensembles. The right depicts the disagreement within the intermediate representations of a single deterministic neural network. Dark red, or black represents low uncertainty while lighter shades of red or orange depict the opposite. Table 7. AUROC and classification accuracy on on the organ{C, A, S} datasets. AUROC Dataset Algorithms Preservation Constraint Accuracy organ A organ C organ S DNN No Constraint 93.717 0.197 - 0.904 0.003 0.904 0.003 Energy-Based (Liu et al., 2021) - 0.894 0.004 0.847 0.002 SNGP(Liu et al., 2020) SN 89.810 0.000 - 0.834 0.015 0.826 0.014 TULIP No Constraint 94.880 0.324 - 0.897 0.013 0.855 0.006 DNN No Constraint 90.221 0.018 0.836 0.007 - 0.762 0.006 Energy-Based (Liu et al., 2021) 0.803 0.015 - 0.738 0.010 SNGP(Liu et al., 2020) SN 87.530 0.000 0.755 0.000 - 0.729 0.000 TULIP No Constraint 90.941 0.109 0.842 0.001 - 0.772 0.002 DNN No Constraint 80.672 0.085 0.688 0.028 0.733 0.010 - Energy-Based (Liu et al., 2021) 0.720 0.028 0.777 0.014 - SNGP(Liu et al., 2020) SN 70.671 0.980 0.640 0.019 0.713 0.007 - TULIP No Constraint 78.984 0.017 0.757 0.037 0.794 0.017 - Table 8. Expected-Calibration-Error and Runtime on CIFAR10 and CIFAR100. ECE Architecture Algorithms Runtime Space Complexity CIFAR10 CIFAR100 DNN 1x 1x 0.065 0.001 1.110 0.002 Energy-Based (Liu et al., 2021) SNGP(Liu et al., 2020) 1x 1x 0.010 0.002 0.720 0.002 DUQ(Van Amersfoort et al., 2020) 1x 1x 0.973 0.003 - TULIP 1x 1.5x 0.030 0.005 0.658 0.004 Res Net-101 DNN 1x 1x 0.056 0.001 1.108 0.011 Energy-Based (Liu et al., 2021) SNGP(Liu et al., 2020) 1x 1x 0.039 0.013 0.671 0.021 DUQ(Van Amersfoort et al., 2020) 1x 1x 0.973 0.001 - TULIP 1x 1.5x 0.025 0.021 0.672 0.031 Res Net-152 DNN 1x 1x 0.055 0.001 1.126 0.004 Energy-Based (Liu et al., 2021) SNGP(Liu et al., 2020) 1x 1x 0.050 0.008 0.698 0.005 DUQ(Van Amersfoort et al., 2020) 1x 1x 0.623 0.002 - TULIP 1x 1.5x 0.026 0.011 0.662 0.055 Transitional Uncertainty with Layered Intermediate Predictions Table 9. OOD detection and classification accuracy on the imbalanced CIFAR100 dataset vs. SVHN. Architecture Algorithms Accuracy AUROC DNN 50.630 0.155 0.747 0.019 Energy-Based (Liu et al., 2021) 50.630 0.155 0.772 0.042 SNGP(Liu et al., 2020) 45.752 3.375 0.783 0.041 DUQ(Van Amersfoort et al., 2020) - - TULIP 52.388 0.568 0.840 0.019 Res Net-101 DNN 49.782 0.161 0.751 0.006 Energy-Based 49.782 0.161 0.793 0.003 SNGP(Liu et al., 2020) 44.073 2.186 0.748 0.018 DUQ(Van Amersfoort et al., 2020) - - TULIP 52.648 1.027 0.807 0.028 Section 3. For the first 50 classes we reduce the training samples by 80% (400 samples) and maintain the same test set. For the remaining 50 classes, we reduce the test set by 90% (90 samples) and maintain all training samples. We use the same implementations as our previous out-of-distribution experiments and consider the three residual backbones Res Net-50, -101, and -152. We show the AUROC scores in Table 9. Complementary to our previous results, our method outperforms other single-pass uncertainty estimators despite having access to only 90% of the training labels. These results illustrate the importance of using feature preservation methods that do not oppose the training objective and show that intermediate representations are attractive options for maintaining information of the input distribution. D. Method Details In our implementation, we calculate the individual uncertainty scores through distance awareness, similar to SNGP (Liu et al., 2020). Applied top our algorithm, the final layer of both the internal classifier, as well as the prediction output are Laplace-approximated Gaussian processes, and we calculate uncertainty with the Dempster-Shafer metric: K + PK k=1 exp(gk(x)) (19) Here, gk is the k-th logit of the output g (either model prediction or internal classifier), and K represents the number of classes. Our choice is based on simplicity and the past success of Gaussian process layers in single-pass uncertainty estimation (van Amersfoort et al., 2021; Liu et al., 2020). While our design produces sufficient results, we emphasize that other implementations of both us and hw may further improve the performance. Transitional Uncertainty with Layered Intermediate Predictions Algorithm 1 Training 1: Input 2: Labeled Training Set 3: {xi XID, yi}N i=1 4: Unlabeled Validation Set 5: {xi XID}Nval i=1 6: SDN Training 7: for epoch [1, epochs] do 8: hw SGD Update LSDN 9: end for 10: Fit Combination Head 11: Derive prediction switches and scores 12: s {s(xi)}Nval i=1 13: v {[us(f 1 w1(xi)), .., us(f NIC w NIC (xi))]}Nval i=1 14: Fit logistic regression weights 15: r1, ..., r NIC = LR(s, v) 16: Return hw, r1, ..., r NIC Algorithm 2 Prediction 1: Input 2: Test Sample xte 3: SDN Prediction 4: Internal uncertainty scores 5: u [us(f 1 w1(xte)), ..., us(f NIC w NIC (xte))] 6: Combine scores 7: ufinal 1 PNIC i=1 ri PNIC i=1 rius(f i wi(x)) 8: Prediction 9: yte hw(xte) 10: Return yte, ufinal E. Additional Related Work Distance Preserving Neural Networks The goal of learning a distance-preserving mapping has been an important objective in a wide range of fields such as generative modeling (Lawrence & Qui nonero Candela, 2006; Dinh et al., 2014; 2016) and dimensionality reduction (Abraham et al., 2006; Perrault-Joncas & Meila, 2012). Recently, the concept has been expanded to uncertainty estimation for neural networks and is used to enable single-pass uncertainty estimators (Liu et al., 2020). Several methods exist to control distance preservation in neural networks and each comes with its own set of trade-offs: the two-sided gradient penalty (Gulrajani et al., 2017) was originally introduced in the context of GANs as an alternative to weight clipping (Arjovsky et al., 2017). The penalty regularizes the network by penalizing the squared distance of the gradient from a fixed value for every input point. The approach is popular due to its simple implementation, but represents only a soft constraint. Spectral normalization (Gouk et al., 2021; Miyato et al., 2018) combines spectral normalization with residual connections to implement distance preservation. The method represents a global constraint as it regularizes by normalizing the weights and suffers less from training instabilities as the gradient penalty. Finally, there exists work on reversible networks that force distance preservation through reversible layers and avoiding down-scaling operations (Jacobsen et al., 2018; Behrmann et al., 2019). In practice, reversible models are difficult to train and consume considerably more memory in practice (van Amersfoort et al., 2021). For this purpose, recent single-pass approaches utilize either spectral normalization, or the gradient penalty. Iterative Uncertainty Estimation In addition to single-pass uncertainty, significant related work exists in iterative uncertainty estimation. With iterative uncertainty estimation, we refer to methods requiring several forward passes for computation. Here, the state-of-the-art are deep ensembles (Lakshminarayanan et al., 2017), as well as several parameterefficient counterparts (Wen et al., 2020; Dusenberry et al., 2020; Thiagarajan et al., 2022). Further examples include Bayesian Neural Networks (Wenzel et al., 2020; Osawa et al., 2019) and MC-Dropout (Gal & Ghahramani, 2016). In practice, these methods tend to render powerful uncertainty estimates but require several forward passes to compute, limiting their applicability in practice. In Appendix E, we provide additional related work on distance preserving neural networks.