# detecting_extrapolation_with_local_ensembles__9259b83b.pdf Published as a conference paper at ICLR 2020 DETECTING EXTRAPOLATION WITH LOCAL ENSEMBLES David Madras University of Toronto Vector Institute madras@cs.toronto.edu James Atwood Google Brain atwoodj@google.com Alex D Amour Google Brain alexdamour@google.com We present local ensembles, a method for detecting extrapolation at test time in a pre-trained model. We focus on underdetermination as a key component of extrapolation: we aim to detect when many possible predictions are consistent with the training data and model class. Our method uses local second-order information to approximate the variance of predictions across an ensemble of models from the same class. We compute this approximation by estimating the norm of the component of a test point s gradient that aligns with the low-curvature directions of the Hessian, and provide a tractable method for estimating this quantity. Experimentally, we show that our method is capable of detecting when a pretrained model is extrapolating on test data, with applications to out-of-distribution detection, detecting spurious correlates, and active learning. 1 INTRODUCTION As machine learning is deployed in increasingly vital areas, there is increasing demand for metrics that draw attention to potentially unreliable predictions. One important source of unreliability is extrapolation. Extrapolation can be formalized in a number of ways: it can refer to making predictions on inputs outside the support of the training data, making predictions with high Bayesian or Frequentist uncertainty, or making predictions that depend strongly on arbitrary choices outside of the learning problem specification (e.g., a random seed). In this paper, we develop a method for detecting this last form of extrapolation. Specifically, we say that a trained model is extrapolating on a test input if the prediction at this input is underdetermined meaning that many different predictions are all equally consistent with the constraints posed by the training data and the learning problem specification (i.e., the model architecture and the loss function). Underdetermination is just one form of extrapolation, but it is particularly relevant in the context of overparameterized model classes (e.g. deep neural networks). Recently, simple (but computationally expensive) ensembling methods (Lakshminarayanan et al., 2017), which train many models on the same data from different random seeds, have proven highly effective at uncertainty quantification tasks (Ovadia et al., 2019). This suggests that underdetermination is a key threat to reliability in deep learning, and motivates flexible methods that can detect underdetermined predictions cheaply. With this motivation, we present local ensembles, a post-hoc method for measuring the extent to which a pre-trained model s prediction is underdetermined for a particular test input. Given a trained model, our method returns an extrapolation score that measures the variability of test predictions across a local ensemble, i.e. a set of local perturbations of the trained model parameters that fit the training data equally well. Local ensembles are a computationally cheap, post-hoc alternative to fully trained ensembles, and do not require special training procedures of approximate ensembling methods that measure related, but distinct, notions of uncertainty (Gal & Ghahramani, 2015; Blundell et al., 2015). Local ensembles also address a gap in approximate methods for estimating prediction uncertainty. Specifically, whereas exact Bayesian or Frequentist uncertainty includes underdetermination as one component, approximate methods such as Laplace approximations (Mac Kay, 1992) or influence function-based methods (Schulam & Saria, 2019) break down when underdetermination is present. Some of this work was done while this author was an intern at Google Brain. Published as a conference paper at ICLR 2020 In contrast, our method leverages the pathology that makes these methods struggle (an ill-conditioned Hessian). Our contributions in this paper are as follows: We present local ensembles, a test-time method for detecting underdetermination-based extrapolation in overparameterized models. We demonstrate theoretically that our method approximates the variance of a trained ensemble with local second-order information. We give a practical method for tractably approximating this quantity, which is simpler and cheaper than alternative second-order reliability methods. Through experiments aimed at testing underdetermination, we show our method approximates the behavior of trained ensembles, and can detect extrapolation in a range of scenarios. 2 EXTRAPOLATION SCORE AND LOCAL ENSEMBLES Let z = (x, y) be an example input-output pair, where x is a vector of features and y is a label. We define a model in terms of a loss function L with parameters θ as a sum over training examples (zi)n i=1, i.e., L(θ) = Pn i ℓ(zi, θ), where ℓis an example-wise loss (e.g., mean-squared error or cross entropy). Let θ be the parameters of the trained model, obtained by, e.g., minimizing the loss over this dataset, i.e., θ = arg minθ L(θ). We write the prediction function given by parameters θ at an input x as ˆy(x, θ). We consider the problem of auditing a trained model, where unlabeled test points x arrive one at a time in a stream, and we wish to assess extrapolation on a point-by-point basis. Figure 1: In this quadratic bowl, arrows denote the small eigendirection, where predictions are slow to change. We argue this direction is key to extrapolation. In this section, we introduce our local ensemble extrapolation score Em(x ) for an unlabeled test point x (significance of m explained below). The score is designed to measure the variability that would be induced by randomly choosing predictions from an ensemble of models with similar training loss. Our score has a simple form: it is the norm of the prediction gradient gθ (x ) := θˆy(x , θ ) multiplied by a matrix of Hessian eigenvectors spanning a subspace of low curvature Um (defined in more detail below). Em(x ) = U mgθ (x ) 2 (1) Here, we show that this score is proportional to the standard deviation of predictions across a local ensemble of models with near-identical training loss, and demonstrate that this approximation holds in practice. 2.2 DERIVATION Our derivation proceeds in two steps. First, we define a local ensemble of models with similar training loss, then we state the relationship between our extrapolation score and the variability of predictions within this ensemble. The spectral decomposition of the Hessian Hθ plays a key role in our derivation. Let Hθ = UΛU , (2) where U is a square matrix whose columns are the orthonormal eigenvectors of Hθ , written (ξ(1), , ξ(p)), and Λ is a square, diagonal matrix with the eigenvalues of Hθ , written (λ(1), , λ(p)), along its diagonal. As a convention, we index the eigenvectors and eigenvalues in decreasing order of the eigenvalue magnitude. To construct a local ensemble of loss-preserving models, we exploit the fact that eigenvectors with large corresponding eigenvalues represent directions of high curvature, whereas eigenvectors with small corresponding eigenvalues represent directions of low curvature. In particular, under the Published as a conference paper at ICLR 2020 assumption that the model has been trained to a local minimum or saddle point, parameter perturbations in flat directions (those corresponding to small eigenvalues λ(j)) do not change the training loss substantially (see Fig. 1). 1 We characterize this subspace by the span of eigenvectors with corresponding small eigenvalues. Formally, let m be the eigenvalue index such that the eigenvalues {λ(j) : j > m}, are sufficiently small to be considered flat . 2 We call the subspace spanned by {ξ(j) : j σ} the ensemble subspace. Parameter perturbations in the ensemble subspace generate an ensemble of models with near-identical training loss. 3 Our score Em(x ) characterizes the variance of test predictions with respect to random parameter perturbations θ that occur in the ensemble subspace. We now show that our extrapolation score Em(x ) is, to the first order, proportional to the standard deviation of predictions at a point x . Proposition 1. Let θ be a random parameter perturbation with mean zero and covariance proportional to the identity ϵ I, projected into the ensemble subspace spanned by {ξ(j) : j > m}. Let P (x ) be the linearized change in a test prediction at x induced by perturbing the learned parameters θ by θ: P (x ) := gθ (x ) θ ˆy(x , θ + θ) ˆy(x , θ ). Then Em(x ) = ϵ 1/2 SD(P (x )). Proof. First, we characterize the distribution of the loss-preserving parameter perturbation θ. Let Um be the matrix whose columns {ξ(j) : j > m} span the ensemble subspace. Then Um U m is a projection matrix that projects vectors into the ensemble subspace, so the projected perturbation θ has covariance ϵ Um U m. From this, we derive the variance of the linearized change in prediction V ar(P ) = ϵ gθ (x ) Um U mgθ (x ) = ϵ U mgθ (x ) 2 2 = ϵ Em(x )2. 2.5 2.0 1.5 1.0 0.5 0.0 Ensemble Standard Deviation Extrapolation Score Figure 2: Mean extrapolation score vs. ensemble prediction standard deviation on Wine Quality dataset. Shows line of best fit and Pearson coefficient R. Both axes log-scaled. We test this hypothesized relationship on several tabular datasets. We train an ensemble of twenty neural networks using the same architecture and training set, varying only the random seed. Then, for each model in the ensemble, we calculate our local ensemble extrapolation score Em(x) for each input x in the test set (see Sec. 3 for details). For each x , we compare, across the ensemble, the mean value of Em(x ) to the standard deviation ˆy(x ). In Fig. 2, we plot these two quantities against each other for one of the datasets, finding a nearly linear relationship. On each dataset, we found a similar, significantly linear relationship (see Table 1 and Appendix A). We note the relationship is weaker with the Diabetes dataset; the standard deviations of these ensembles are an order of magnitude higher than the other datasets, indicating much noisier data. Finally, we note that we can obtain similar results for the standard deviation of the loss at test points if we redefine g as the loss gradient rather than the prediction gradient. 3 COMPUTING EXTRAPOLATION SCORES 1To see this, consider a second-order Taylor approximation to the training loss when the parameters are perturbed by a vector θ: L(θ + θ) L(θ ) 1 2 θ Hθ θ = Pp j=1 λ(j) ξ (j) θ 2. Note that the first-order term drops out because we assume θ lies at a local minimum or saddle point. Perturbations θ that lie in the subspace spanned by eigenvectors ξ(j) whose corresponding eigenvalues λ(j) are close to zero contribute negligibly to this sum. 2For now, we take m to be given, and discuss tradeoffs for choosing m in practice in Section 3. 3We note that not all loss-preserving ensembles will necessarily be local; however, the existence of some such local ensemble is sufficient to enable underdetermination. Published as a conference paper at ICLR 2020 Dataset Pearson Boston 0.76 Diabetes 0.50 Abalone 0.76 Wine 0.87 Table 1: Correlation of extrapolation scores and ensemble std. deviations on 4 datasets. We now discuss a practical method for computing our extrapolation scores. The key operation is constructing the set of eigenvectors that span a suitably loss-preserving ensemble subspace (i.e. have sufficiently small corresponding eigenvalues). Because eigenvectors corresponding to large eigenvalues are easier to estimate, we construct this subspace by finding the top m eigenvectors and defining the ensemble subspace as their orthogonal complement. Our method can be implemented with any algorithm that returns the top m eigenvectors. See below for discussion on some tradeoffs in the choice of m, as well as the algorithm we choose (the Lanczos iteration (Lanczos, 1950)). Our method proceeds as follows. For a given choice of m, we calculate the m eigenvectors of H with the largest eigenvalues. These eigenvectors define an m-dimensional subspace which is the orthogonal complement to the ensemble subspace. We use these eigenvectors to construct a matrix that projects gradients into the ensemble subspace. Specifically, let Um be the matrix whose columns are these large-eigenvalue eigenvectors {ξ(j) : j m}. Then Um U m is the projection matrix that projects vectors into the large eigenvalue subspace, and I Um U m projects into its complement: the ensemble subspace. Now, for any test input x , we take the norm of this projected gradient to compute our score Em(x ) = I Um U m gθ (x ) 2 . The success of this approach depends on the choice of m. Specifically, the extrapolation score Em(x ) is the most sensitive to underdetermination in the region of the trained parameters if we set m to be the smallest index for which the training loss is relatively flat in the implied ensemble subspace. If m is set too low, the ensemble subspace will include well-constrained directions, and Em(x ) will over-estimate the prediction s sensitivity to loss-preserving perturbations. If m is set too high, the ensemble subspace will omit some under-constrained directions, and Em(x ) will be less sensitive. For models where all parameters are well-constrained by the training data, a suitable m may not exist. This will usually not be the case for deep neural network models, which are known to have very ill-conditioned Hessians (see, e.g., Sagun et al., 2017). Whether a direction is well-constrained is ultimately a judgment call for the user. One potential heuristic is to consider a small parameter perturbation of a fixed norm in the ensemble subspace, and to set a tolerance for how much that perturbation can change the training loss; for small perturbations, the change in loss is a linear function of the curvature in the direction of the perturbation, which is upper bounded by λ(m). We use the Lanczos iteration to estimate the top m eigenvectors, which presents a number of practical advantages for usage in our scenario. Firstly, it performs well under early stopping, returning good estimates of the top m eigenvectors after m iterations. Secondly, we can cache intermediate steps, meaning that computing the m + 1-th eigenvector is fast once we have computed the first m. Thirdly, it requires only implicit access to the Hessian through a function which applies matrix multiplication, meaning we can take advantage of efficient Hessian-vector product methods (Pearlmutter, 1994). Finally, the Lanczos iteration is simple it can be implemented in less than 20 lines of Python code (see Appendix B.1). It contains only one hyperparameter, the stopping value m. Fortunately, tuning this parameter is efficient given a maximum value M, we can try many values m < M at once, by estimating M eigenvectors and then calculating Em by using the first m eigenvectors. The main constraint of our method is space rather than time while estimating the first m eigenvectors enables easy caching for later use, it may be difficult to work with these eigenvectors in memory as m and model size p increase. This tradeoff informed our choice of m in this paper; we note in some cases that increasing m further could have improved performance (see Appendix E). This suggests that further work on techniques for mitigating this tradeoff, e.g. online learning of sparse representations (Wang & Lu, 2016; Wang et al., 2012), could improve the performance of our method. See Appendix B for more details on the Lanczos iteration. Published as a conference paper at ICLR 2020 4 RELATED WORK 4.1 RELATION TO BAYESIAN AND FREQUENTIST SECOND-ORDER METHODS It is instructive to compare our extrapolation score to two other approximate reliability quantification methods that are aimed at Bayesian and Frequentist notions of extrapolation, respectively. Like our extrapolation score, both of these methods make use of local information in the Hessian to make an inference about the variance of a prediction. First, consider the Laplace approximation of the posterior predictive variance. This metric is derived by interpreting the loss function as being equivalent to a Bayesian log-posterior distribution over the model parameters θ, and approximating it with a Gaussian. Specifically, (see, e.g., Mac Kay, 1992) V ar(y | x ) gθ (x ) H 1 θ gθ (x ) = j=1 λ 1 (j) ξ (j)gθ (x ) 2 . (3) Second, consider scores such as RUE (Resampling Under Uncertainty) designed to approximate the variability of predictions by resampling the training data (Schulam & Saria, 2019). These methods approximate the change in trained parameter values induced by perturbing the training data via influence functions (Koh & Liang, 2017; Cook & Weisberg, 1982). Specifically, the gradient of the parameters with respect to the weight of a given training example zi is given by I(zi) = H 1 θ θℓ(zi, θ ) = j=1 λ 2 (j) ξ (j) θℓ(zi, θ ) 2 . (4) Schulam & Saria (2019) combine this influence function with a specific random distribution of weights to approximate the variance of predictions under bootstrap resampling; other similar formulations are possible, sometimes with theoretical guarantees (Giordano et al., 2019). Importantly, both of these methods work well when model parameters are well-constrained by the training data, but they struggle when predictions are (close to) underdetermined. This is because, in the presence of underdetermination, the Hessian becomes ill-conditioned. Practical advice for dealing with this ill-conditioning is available (Koh & Liang, 2017), but we note that this not merely a numerical pathology; by our argument above, a poorly conditioned Hessian is a clear signal of extrapolation. In contrast to these methods, our method focuses specifically on prediction variability induced by underconstrained parameters. Our extrapolation score incorporates only those terms with small eigenvalues, and removes the inverse eigenvalue weights that make inverse-Hessian methods break down. This is clear from its summation representation: Em(x ) = P j>m ξ (j)gθ (x ) 2 . Our method also has computational advantages over approaches that rely on inverting the Hessian. Firstly, implicitly inverting the Hessian is a complex and costly process by finding only the important components of the projection explicitly, our method is simpler and more efficient. Furthermore, we only need to find these components once; we can reuse them for future test inputs. This type of caching is not possible with methods which require us to calculate the inverse Hessian-vector product for each new test input (e.g. conjugate gradient descent). 4.2 RELATED WORK IN DETECTING EXTRAPOLATION Some recent works explore the relationship between test points, the learned model, and the training set. Several papers examine reliability criteria that are based on distance in some space: within/betweengroup distances (Jiang et al., 2018), a pre-specified kernel in a learned embedding space (Card et al., 2019), or the activation space of a neural network (Papernot & Mc Daniel, 2018). We implement some nearest-neighbor baselines inspired by this work in Sec. 5. Additionally, a range of methods exist for related tasks, such as OOD detection (Choi & Jang, 2018; Liang et al., 2017; Gal & Ghahramani, 2015; Sch olkopf et al., 2001) and calibration (Naeini et al., 2015; Guo et al., 2017). Some work using generative models for OOD detection uses related second-order analysis (Nalisnick et al., 2018). A line of work explores the benefits of training ensemble methods explicitly, discussed in detail in Dietterich (2000). These methods have been discussed for usage in some of the applications we present in Sec. 5, including uncertainty detection (Lakshminarayanan et al., 2017), active learning (Melville & Mooney, 2004) and OOD detection (Choi & Jang, 2018). A number of approximate Published as a conference paper at ICLR 2020 ensembling methods have also been proposed. For example, MC-Dropout (Gal & Ghahramani, 2015) and Bayes by Backprop (Blundell et al., 2015) are approximate Bayesian model averaging methods that leverage special training procedures to represent an ensemble of prediction functions. These target a distinct notion of uncertainty from our loss-preserving ensembles (see Appendix F). Finally, a line of work on Rashomon sets explores loss-preserving ensembles more formally in simpler model classes (Semenova & Rudin, 2019; Fisher et al., 2018). 5 EXPERIMENTS In this section, we give evidence that local ensembles can detect extrapolation due to underdetermination in trained models. In order to explicitly evaluate underdetermination, we present a range of experiments where a pre-trained model has a blind spot , and evaluate its ability to detect when an input is in that blind spot. We probe our method s ability to detect a range of extrapolation, exploring cases where the blind spot is: 1. easily visualized, 2. well-defined by the feature distribution, 3. well-defined by a latent distribution, and 4. unknown, but where we can evaluate our model s detection performance through an auxiliary task. See Appendix D for experimental details. Code for running the local ensembles method can be found at https://github.com/dmadras/local-ensembles. 5.1 VISUALIZING EXTRAPOLATION DETECTION 4 2 0 2 4 6 x Train Test OOD (a) Data and trained model 3 2 1 0 1 2 3 4 x 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 Standard Deviation (b) Extrapolation scores 0 5 10 15 20 0.3 Cosine Similarity (c) Eigenvector estimates & AUC Figure 3: We train a neural network ensemble (Fig. 3a). We compute extrapolation scores (solid line), which correlate with the standard deviation of the ensemble (dotted line) (Fig. 3b). Our OOD performance achieves high AUC (solid line) even though some of our eigenvector estimates have low cosine similarity to ground truth (dotted line) (Fig. 3c). 0 5 10 15 20 0 40 Lanczos True Figure 4: True and estimated eigenspectrums for toy model Hessian. We note that the first few eigenvalues account for most of the variation, and that our estimates are accurate. We begin with an easily visualized toy experiment. In Fig. 3a, we show our data (y = sin 4x + N(0, 1 4)). We generate in-distribution training data from x [ 1, 0] [1, 2], but at test time, we consider all x [ 3, 4]. We train 20 neural networks with the same architecture (two hidden layers of three units each). As shown in Fig. 3a, the ensemble disagrees most on x < 1, x > 2. This means that we should most mistrust predictions from this model class on these extreme values, since there are many models within the class that perform equally well on the training data, but differ greatly on those inputs. We should also mistrust predictions from x [0, 1], although detecting this extrapolation may be harder since the ensemble agrees more strongly on these points. For each model in the ensemble, we test our method on an OOD task: can we flag test points which fall outside the training distribution? Since OOD examples may be uncommon in practice, we use AUC to measure our performance. We show that the extrapolation score is empirically related to the standard deviation of the ensemble s predictions at the input, which in turn is related to whether the input is OOD (Fig. 3b). Examining one model from this ensemble, we observe that by estimating only m = 2 eigenvectors, we achieve > 90% AUC (Fig. 3c). It turns out that m = 10 performs best on this task/model. As we complete more iterations (m > 10) we start finding smaller eigenvalues, which are more important to the ensemble subspace and whose eigenvector we do not wish to project out. Published as a conference paper at ICLR 2020 1 2 3 4 5 Number of Features Simulated NN (Inputs) NN (Final Layer) NN (Reprs) Local Ensembles 1 2 3 4 5 Number of Features Simulated 1 2 3 4 5 Number of Features Simulated 0.60 0.65 0.70 0.75 0.80 0.85 0.90 0.95 1 2 3 4 5 Number of Features Simulated Figure 5: AUC achieved on Wine Quality simulated features task (y-axis) compared to the number of extra features simulated (x-axis). Noise parameter σ increased from left to right σ {0, 0.1, 0.2, 0.5}. Solid line is our method. Results averaged across 5 random seeds, standard deviations shown. Results for the other three datasets are qualitatively similar and shown in Appendix C. We note our AUC improves even with some eigenvector estimates having low cosine similarity to the ground truth eigenvectors (the Lanczos iteration has some stochasticity due to minibatch estimation see Appendix B for details). We hypothesize this robustness is because the ensemble subspace of this model class is relatively low-dimensional. Even if an estimated vector is noisy, the non-overlapping parts of the projection will likely be mostly perpendicular to the ensemble subspace, due to the properties of high-dimensional space. 5.2 SIMULATED FEATURES In this experiment, we create a blind spot in a given dataset by extending and manipulating the data s feature distribution. We induce a collinearity in feature space by generating new features which are a linear combination of two other randomly selected features in the training data. This means there is potential for underdetermination: multiple, equally good learnable relationships exist between those features and the target. However, at test time, we will sometimes sample these simulated features from their marginal distribution instead. This breaks the linear dependence, requiring extrapolation (the model is by definition underconstrained), without making the new data trivially out-of-distribution. We can make this extrapolation detection task easier by generating several features this way, or make it harder by adding some noise N(0, σ2) to these features. We run this experiment on four tabular datasets, training a two hidden-layer neural network for each task. We compare to three nearest-neighbour baselines, where the metric is the distance (in some space) of the test point to its nearest neighbour by Euclidean distance in an in-distribution validation set. NN (Inputs) uses input space; NN (Reprs) uses hidden representation space, which is formed by concatenating all the activations of the network together (inspired by Papernot & Mc Daniel (2018), who propose a similar method for adversarial robustness); and NN (Final Layer), uses just the final hidden layer of representations. We note that since our method is post-hoc and can be applied to any twice-differentiable pre-trained model, we do not compare to training-based methods e.g. those producing Bayesian predictive distributions (Gal & Ghahramani, 2016; Blundell et al., 2015). Our metric is AUC: we aim to assign higher scores to inputs which break the collinearity (where the feature is drawn from the marginal), than those which do not. In Figure 5, we show that local ensembles (LE) outperform the baselines for each number of extra simulated features, and that this performance is fairly robust to added noise. 5.3 CORRELATED LATENT FACTORS Here, we extend the experiment from Section 5.2 by inducing a blind spot in latent space. The rationale is similar: if two latent factors are strongly correlated at training time, the model may grow reliant on that correlation, and at test-time may be underdetermined if that correlation is broken. We use the Celeb A dataset (Liu et al., 2015) of celebrity faces, which annotates every image with 40 latent binary attributes describing the image (e.g. brown hair ). To induce the blind spot, we choose two attributes: a label L and a spurious correlate C. We then create a training set where L and C are perfectly correlated: a point is only included in the training set if L = C. We train a convolutional neural network (CNN), with two convolutional layers and a fully-connected layer at the end, as a binary classifier to predict L. Then, we create a test set of held-out data where P(L = C) = P(L = C). The test data where L = C is in our model s blind spot; these are the inputs for which we want to output high extrapolation scores. We show in Appendix E that the models Published as a conference paper at ICLR 2020 dramatically fail to classify these inputs (L = C). We compare to four baseline extrapolation scores: the three nearest-neighbour methods described in Sec. 5.2, as well as Max Prob, where we use 1 the maximum outputted probability of the softmax. We test two values of L (Male and Attractive) and two values of C (Eyeglasses and Wearing Hat). We chose these specific values of L because they are difficult to predict and holistic i.e.and not localized to particular areas of image space. In Table 2, we present results for each of the four L, C settings, showing both the loss gradient and the prediction gradient variant of local ensembles. Note that the loss gradient cannot be calculated at test time since we do not have labels available instead, we calculate a separate extrapolation score using the gradient for the loss with respect to each possible label, and take the minimum. Our method achieves the best performance on most settings, and is competitive with the best baseline on each. However, the variation between the tasks is quite noteworthy. We note two patterns in particular. Firstly, we note the performance of Max Prob and the loss gradient variant of our method are quite correlated, which we hypothesize relates to ˆY ℓ. Additionally, the effect of increasing m is inconsistent between experiments: we discuss possible relationships to the eigenspectrum of the trained models. See Appendix E for a discussion on these patterns. 5.4 ACTIVE LEARNING 0 20 40 60 80 100 Rounds of Active Learning Classification Error Random Local Ensembles 0 20 40 60 80 100 Rounds of Active Learning Classification Error Random Local Ensembles (b) Fashion MNIST Figure 6: Active learning results. Average of 5 random seeds shown with std. deviation error bars. Method M/E M/H A/E A/H Max Prob 0.738 0.677 0.461 0.433 NN (Pixels) 0.561 0.550 0.521 0.547 NN (Reprs) 0.584 0.578 0.503 0.533 NN (Final Layer) 0.589 0.517 0.480 0.497 LE (Loss) 0.770 0.684 0.454 0.456 LE (Predictions) 0.364 0.544 0.519 0.582 Table 2: AUC for Latent Factors OOD detection task. Column heading denotes in-distribution definitions: labels are M (Male) and A (Attractive); spurious correlates are E (Eyeglasses) and H (Wearing Hat). Image is in-distribution iff label = spurious correlate. LE stands for local ensembles. Each Lanczos iteration uses 3000 eigenvectors. 500 examples from each test set are used. 95% CI is bolded. Finally, we consider the case where we know our model has blind spots, but do not know where they are. We use active learning to probe this situation. If underdetermination is an important factor in extrapolation, then local ensembles should be useful for selecting the most useful points to add to our training set. We use MNIST (Le Cun et al., 2010) and Fashion MNIST (Xiao et al., 2017) for our active learning experiments. We begin the first round with a training set of twenty: two labelled data points from each of the ten classes. In each round, we train our model (a small CNN with two convolutional layers and a fully connected layer at the end) to a minimum validation loss using the current training set. After each round, we select ten new points from a randomly selected pool of 500 unlabelled points, and add those ten points and their labels to our training set. We compare local ensembles (selecting the points with the highest extrapolation scores using the loss-gradient variant) to a random baseline selection mechanism. In Fig. 6, we show that our method outperforms the baseline on both datasets, and this improvement increases in later rounds of active learning. We only used 10 eigenvectors in our Lanczos approximation, which we found to be a surprisingly effective approximation; we did not observe improvement with more eigenvectors. This Published as a conference paper at ICLR 2020 experiment serves to emphasize the flexibility of our method: by detecting an underlying property of the model, we can use the method for a range of tasks (active learning as well as OOD detection). 6 CONCLUSION We present local ensembles, a post-hoc method for detecting extrapolation due to underdetermination in a trained model. Our method uses local second-order information to approximate the variance of an ensemble. We give a tractable implementation using the Lanczos iteration to estimate the largest eigenvectors of the Hessian, and demonstrate its practical flexibility and utility. Although this method is not a full replacement for ensemble methods, which can characterize more complexity (e.g. multiple modes), we believe it fills an important role in isolating one component of prediction unreliability. In future work, we hope to scale these methods to larger models and to further explore the properties of different stopping points m. We also hope to explore applications in fairness and interpretability, where understanding model and training bias is of paramount importance. 7 ACKNOWLEDGEMENTS Thank you to Jamie Smith, Yaniv Ovadia, Yoni Halpern, Pang Wei Koh, Jackson Wang, James Lucas, Marc-Etienne Brunet, Kamyar Ghasemipour and Elliot Creager for helpful comments and discussions. Charles Blundell, Julien Cornebise, Koray Kavukcuoglu, and Daan Wierstra. Weight uncertainty in neural networks. ar Xiv preprint ar Xiv:1505.05424, 2015. Dallas Card, Michael Zhang, and Noah A Smith. Deep weighted averaging classifiers. In Proceedings of the Conference on Fairness, Accountability, and Transparency, pp. 369 378. ACM, 2019. Hyunsun Choi and Eric Jang. Generative ensembles for robust anomaly detection. ar Xiv preprint ar Xiv:1810.01392, 2018. R Dennis Cook and Sanford Weisberg. Residuals and influence in regression. New York: Chapman and Hall, 1982. Paulo Cortez, Ant onio Cerdeira, Fernando Almeida, Telmo Matos, and Jos e Reis. Modeling wine preferences by data mining from physicochemical properties. Decision Support Systems, 47(4): 547 553, 2009. Jan JM Cuppen. A divide and conquer method for the symmetric tridiagonal eigenproblem. Numerische Mathematik, 36(2):177 195, 1980. Inderjit S Dhillon. A new o (n2) algorithm for the symmetric tridiagonal eigenvalue/eigenvector problem. Technical report, CALIFORNIA UNIV BERKELEY GRADUATE DIV, 1997. Thomas G Dietterich. Ensemble methods in machine learning. In International workshop on multiple classifier systems, pp. 1 15. Springer, 2000. Dheeru Dua and Casey Graff. UCI machine learning repository, 2017. URL http://archive. ics.uci.edu/ml. Bradley Efron, Trevor Hastie, Iain Johnstone, Robert Tibshirani, et al. Least angle regression. The Annals of statistics, 32(2):407 499, 2004. Aaron Fisher, Cynthia Rudin, and Francesca Dominici. All models are wrong but many are useful: Variable importance for black-box, proprietary, or misspecified prediction models, using model class reliance. ar Xiv preprint ar Xiv:1801.01489, 2018. Y Gal and Z Ghahramani. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. arxiv, 2015. Published as a conference paper at ICLR 2020 Yarin Gal and Zoubin Ghahramani. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In international conference on machine learning, pp. 1050 1059, 2016. Ryan Giordano, William Stephenson, Runjing Liu, Michael Jordan, and Tamara Broderick. A swiss army infinitesimal jackknife. In The 22nd International Conference on Artificial Intelligence and Statistics, pp. 1139 1147, 2019. Luc Giraud, Julien Langou, Miroslav Rozloˇzn ık, and Jasper van den Eshof. Rounding error analysis of the classical gram-schmidt orthogonalization process. Numerische Mathematik, 101(1):87 100, 2005. Chuan Guo, Geoff Pleiss, Yu Sun, and Kilian Q Weinberger. On calibration of modern neural networks. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pp. 1321 1330. JMLR. org, 2017. David Harrison Jr and Daniel L Rubinfeld. Hedonic housing prices and the demand for clean air. Journal of environmental economics and management, 5(1):81 102, 1978. Heinrich Jiang, Been Kim, Melody Guan, and Maya Gupta. To trust or not to trust a classifier. In Advances in Neural Information Processing Systems, pp. 5541 5552, 2018. Pang Wei Koh and Percy Liang. Understanding black-box predictions via influence functions. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pp. 1885 1894. JMLR. org, 2017. Balaji Lakshminarayanan, Alexander Pritzel, and Charles Blundell. Simple and scalable predictive uncertainty estimation using deep ensembles. In Advances in Neural Information Processing Systems, pp. 6402 6413, 2017. Cornelius Lanczos. An iteration method for the solution of the eigenvalue problem of linear differential and integral operators. United States Governm. Press Office Los Angeles, CA, 1950. Yann Le Cun, Corinna Cortes, and CJ Burges. Mnist handwritten digit database. AT&T Labs [Online]. Available: http://yann. lecun. com/exdb/mnist, 2:18, 2010. Shiyu Liang, Yixuan Li, and R Srikant. Enhancing the reliability of out-of-distribution image detection in neural networks. ar Xiv preprint ar Xiv:1706.02690, 2017. Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang. Deep learning face attributes in the wild. In Proceedings of the IEEE international conference on computer vision, pp. 3730 3738, 2015. David JC Mac Kay. Information-based objective functions for active data selection. Neural computation, 4(4):590 604, 1992. Prem Melville and Raymond J Mooney. Diverse ensembles for active learning. In Proceedings of the twenty-first international conference on Machine learning, pp. 74. ACM, 2004. Mahdi Pakdaman Naeini, Gregory Cooper, and Milos Hauskrecht. Obtaining well calibrated probabilities using bayesian binning. In Twenty-Ninth AAAI Conference on Artificial Intelligence, 2015. Eric Nalisnick, Akihiro Matsukawa, Yee Whye Teh, Dilan Gorur, and Balaji Lakshminarayanan. Do deep generative models know what they don t know? ar Xiv preprint ar Xiv:1810.09136, 2018. Warwick J Nash, Tracy L Sellers, Simon R Talbot, Andrew J Cawthorn, and Wes B Ford. The population biology of abalone (haliotis species) in tasmania. i. blacklip abalone (h. rubra) from the north coast and islands of bass strait. Sea Fisheries Division, Technical Report, 48, 1994. Travis E Oliphant. A guide to Num Py, volume 1. Trelgol Publishing USA, 2006. Yaniv Ovadia, Emily Fertig, Jie Ren, Zachary Nado, D Sculley, Sebastian Nowozin, Joshua V Dillon, Balaji Lakshminarayanan, and Jasper Snoek. Can you trust your model s uncertainty? evaluating predictive uncertainty under dataset shift. ar Xiv preprint ar Xiv:1906.02530, 2019. Published as a conference paper at ICLR 2020 Nicolas Papernot and Patrick Mc Daniel. Deep k-nearest neighbors: Towards confident, interpretable and robust deep learning. ar Xiv preprint ar Xiv:1803.04765, 2018. Barak A Pearlmutter. Fast exact multiplication by the hessian. Neural computation, 6(1):147 160, 1994. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825 2830, 2011. Levent Sagun, Utku Evci, V Ugur Guney, Yann Dauphin, and Leon Bottou. Empirical analysis of the hessian of over-parametrized neural networks. ar Xiv preprint ar Xiv:1706.04454, 2017. Bernhard Sch olkopf, John C Platt, John Shawe-Taylor, Alex J Smola, and Robert C Williamson. Estimating the support of a high-dimensional distribution. Neural computation, 13(7):1443 1471, 2001. Peter Schulam and Suchi Saria. Auditing pointwise reliability subsequent to training. In Proceedings of the 22nd International Conference on Ar-tificial Intelligence and Statistics (AISTATS), 2019. Lesia Semenova and Cynthia Rudin. A study in rashomon curves and volumes: A new perspective on generalization and model simplicity in machine learning. ar Xiv preprint ar Xiv:1908.01755, 2019. Chuang Wang and Yue M Lu. Online learning for sparse pca in high dimensions: Exact dynamics and phase transitions. In 2016 IEEE Information Theory Workshop (ITW), pp. 186 190. IEEE, 2016. Dong Wang, Huchuan Lu, and Ming-Hsuan Yang. Online object tracking with sparse prototypes. IEEE transactions on image processing, 22(1):314 325, 2012. Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms, 2017. Published as a conference paper at ICLR 2020 0.25 0.50 0.75 1.00 1.25 1.50 1.75 2.00 2.25 Ensemble Standard Deviation Extrapolation Score 2 4 6 8 10 12 14 16 Ensemble Standard Deviation Extrapolation Score 0.2 0.4 0.6 0.8 Ensemble Standard Deviation Extrapolation Score Figure 7: Datasets are (left to right) Boston, Diabetes, Abalone. Dotted line represents linear fit of data. R is Pearson correlation coefficient. 1.5 1.0 0.5 0.0 0.5 Ensemble Standard Deviation Extrapolation Score 0.5 1.0 1.5 2.0 2.5 Ensemble Standard Deviation Extrapolation Score 2.5 2.0 1.5 1.0 0.5 0.0 Ensemble Standard Deviation Extrapolation Score Figure 8: Datasets are (left to right) Boston, Diabetes, Abalone. Each axis is log-scaled. Dotted line represents linear fit of data. R is Pearson correlation coefficient. A ENSEMBLES: OTHER DATASETS We show in Figures 7 and 8 the same plots as Fig. 2 for other tabular datasets, demonstrating the strongly linear relationship between extrapolation score and ensemble standard deviation. B THE LANCZOS ITERATION: FUTHER DETAILS The Lanczos iteration (Lanczos, 1950) is a method for tridiagonalizing a Hermitian matrix. It can be thought of as a variant of power iteration, iteratively building a larger basis through repeatedly multiplying an initial vector by M, ensuring orthogonality at each step. Once M has been tridiagonalized, computing the final eigendecomposition is relatively simple a number of specialized algorithms exist which are relatively fast (O(p2)) (Dhillon, 1997; Cuppen, 1980). The Lanczos iteration is simple to implement, but presents some challenges. The first challenge is numerical instability when computed in floating point arithmetic, the algorithm is no longer quarnteed to return a good approximation to the true eigenbasis, or even an orthogonal one. As such, the standard implementation of the Lanczos iteration is unstable, and can be inaccurate even on simple problems. Fortunately, solutions exist: a procedure known as two-step classical Gram-Schmidt orthogonalization which involves ensuring twice that each new vector is linearly independent of the previous ones is guaranteed to produce an orthogonal basis, with errors on the order of machine roundoff (Giraud et al., 2005). A second potential challenge is presented by the stochasticity of minibatch computation. Since we access our Hessian H only through Hessian-vector products, we must use only minibatch computation at each stage. This means that each iteration will be stochastic, which will decrease the accuracy (but not the orthogonality) of the eigenvector estimates provided. However, in practice, we found that even fairly noisy estimates were nonetheless useful see Sec. 5.1 for more discussion. B.1 LANCZOS ALGORITHM CODE SNIPPET The Lanczos algorithm is quite simple to implement. Figure 9 shows a a short implementation using Python/Numpy (Oliphant, 2006). Published as a conference paper at ICLR 2020 import numpy as np def lanczos(matmul_fn, dim, num_iters, eps=1e-8): Given implicit access to a matrix M of size (dim x dim) through mamtul_fn, tridiagonalize M into diagonal elements Alpha and off-diagonal elements Beta. Also return the associated orthonormal basis Q. # Initialize orthonormal basis. Q = [np.zeros((dim, 1))] # Initialize off-diagonal elements. Beta = [] # Initialize diagonal elements. Alpha = [] # Begin with random initial vector. q = np.random.uniform(size=(dim, 1)) Q.append(q / np.linalg.norm(q)) Q_k_range = Q[0] for k in range(1, num_iters + 1): # Compute next step of power iteration. z = matmul_fn(Q[k]) Alpha.append(np.matmul(Q[k].T, z)) Q_k_range = np.concatenate([Q_k_range, Q[k]], axis=1) # Reorthogonalize using two-step classical Gram-Schmidt. for _ in range(2): z_orth = np.sum(np.matmul(z.T, Q_k_range) * Q_k_range, axis=1, keepdims=True) z = z - z_orth Beta.append(np.linalg.norm(z)) Q.append(z / Beta[k]) # Check for convergence. if np.linalg.norm(z) < eps: break return Q, Beta, Alpha Figure 9: Example Python implementation of Lanczos algorithm for tridiagonalizing an implicit matrix M. 1 2 3 4 5 Number of Features Simulated NN (Inputs) NN (Final Layer) NN (Reprs) Local Ensembles 1 2 3 4 5 Number of Features Simulated 1 2 3 4 5 Number of Features Simulated 0.50 0.55 0.60 0.65 0.70 0.75 0.80 0.85 0.90 1 2 3 4 5 Number of Features Simulated Figure 10: Boston dataset C SIMULATED FEATURES - OTHER DATASETS In Section 5.2, we present results for an experiment where we aim to detect broken collinearities in the feature space. In Figures 10, 11, and 12, we show results on three more tabular datasets. See the main text for more experimental details. Published as a conference paper at ICLR 2020 1 2 3 4 5 Number of Features Simulated NN (Inputs) NN (Final Layer) NN (Reprs) Local Ensembles 1 2 3 4 5 Number of Features Simulated 1 2 3 4 5 Number of Features Simulated 1 2 3 4 5 Number of Features Simulated 0.45 0.50 0.55 0.60 0.65 0.70 0.75 0.80 Figure 11: Diabetes dataset 1 2 3 4 5 Number of Features Simulated 0.800 0.825 0.850 0.875 0.900 0.925 0.950 0.975 1.000 NN (Inputs) NN (Final Layer) NN (Reprs) Local Ensembles 1 2 3 4 5 Number of Features Simulated 1 2 3 4 5 Number of Features Simulated 1 2 3 4 5 Number of Features Simulated Figure 12: Abalone dataset D EXPERIMENTAL DETAILS Here we give more experimental details on the datasets and models used. D.1 DATASETS D.1.1 TABULAR DATASETS We subtracted the mean and divided by the standard deviation for each feature (as calculated from the training set). Boston (Harrison Jr & Rubinfeld, 1978) and Diabetes (Efron et al., 2004). These datasets were loaded from Scikit-Learn (Pedregosa et al., 2011). Abalone (Nash et al., 1994). This dataset was downloaded from the UCI repository (Dua & Graff, 2017) at http://archive.ics.uci.edu/ml/datasets/Abalone. We converted sex to a three-dimensional one-hot vector (M, F, I). Wine Quality (Cortez et al., 2009). This dataset was downloaded from the UCI repository (Dua & Graff, 2017) at http://archive.ics.uci.edu/ml/datasets/Wine+Quality. We used both red and white wines. D.2 MNIST, FASHIONMNIST AND CELEBA. These datasets were loaded using Tensorflow Datasets https://github.com/tensorflow/ datasets. We divided the pixel values by 255 and, for MNIST and Fashion MNIST, binarized by thresholding at 0.7. D.3 EXPERIMENTAL DETAILS D.3.1 TOY DATA EXPERIMENTS For the first experiment, we train a two-layer neural network with 3 hidden units in each layer and tanh units. We train for 400 optimization steps using minibatch size 32. Our data is generated from y = sin(4x) + N(0, 1 4). We generate 200 training points, 100 test points, and 200 OOD points. We aggregate Y over a grid of 10 points from -1 to 1, with aggregation function min. We run the Lanczos algorithm until convergence. For the second experiment, we train a two-layer neural network with 5 hidden units in each layer and Re LU units. Our data is generated from y = βx2 + N(0, 1). Our training set consists of x drawn Published as a conference paper at ICLR 2020 Test Set M/E M/H A/E A/H In-Distribution 0.03 0.06 0.01 0.02 Out-of-Distribution 0.98 0.96 0.90 0.93 Table 3: Error rate for in and out of distribution test set with correlated latent factors setup. Column heading denotes in-distribution definitions: labels are M (Male) and A (Attractive); spurious correlate are E (Eyeglasses) and H (Wearing Hat). Image is in-distribution iff label == spurious correlate. uniformly from [ 0.5, 0.5] and [2.5, 3.5]. However, at test time, we will consider x [ 3, 6]. We generate 200 training points, 100 test points, and 200 OOD points. We aggregate Y over a grid of 5 points from -6 to 9, with aggregation function min. D.3.2 SIMULATED FEATURES For each dataset we use the same setup. We use a two-layer MLP with Re LU activations and hidden layer sizes of 20 and 100. We trained all models with mean squared error loss. We use batch size 64, patience 100 and a 100-step running average window for estimating current performance. For the Lanczos iteration, we run up to 2000 iterations. We always report numbers from the final iteration run. For estimating the Hessian in the HVPs in the Lanczos iteration, we use batch size 32 and sample 5 minibatches. To pre-process the data, we first split randomly out 30% of the dataset as OOD. We choose 2 random features i, j and a number β U(0, 1), and generate the new feature x = βx[i] + (1 β)x[j]. We also normalize this feature by its mean and standard deviation as calculated on the training set. We add random noise to the features after splitting into in-distribution and OOD meaning we are not redrawing from the same marginal noise distribution. We use 1000 examples from in-distribution and OOD for testing. D.3.3 CORRELATED LATENT FACTORS We use a CNN with Re LU hidden activations. We use two convolutional layers with 50 and 20 filters each and stride size 2 for a total of 1.37 million parameters. We trained all models with cross entropy loss. We use an extra dense layer on top with 30 units. We use batch size 32, patience 100 steps, and a 100-step running average window for estimating current performance. We sample the validation set randomly as 20% of the training set. For the Lanczos iteration, we run 3000 iterations. We always report numbers from the final iteration run. We use 500 examples from in-distribution and OOD for testing. For estimating the Hessian in the HVPs in the Lanczos iteration, we use batch size 16 and sample 5 minibatches. D.3.4 ACTIVE LEARNING We use a CNN with Re LU hidden activations. We use two convolutional layers with 16 and 32 layers, stride size 5, and a dense layer on top with 64 units. We trained all models with mean squared error loss. We use batch size 32, patient 100 steps, and a 100-step running average window for estimating current performance. E CORRELATED LATENT FACTORS E.1 PERFORMANCE OF BINARY CLASSIFIERS In Section 5.3, we discuss an experiment where we correlated a latent label L and confounder C attribute. Table 3 shows the in-distribution and out-of-distribution test error. These are drastically different, meaning that learning to detect this type of extrapolation is critical to maintain model performance. Published as a conference paper at ICLR 2020 0 500 1000 1500 2000 2500 3000 0 500 10001500200025003000 0 500 1000 1500 2000 2500 3000 0 500 1000 1500 2000 2500 3000 Figure 13: Latent Factors OOD detection task, using loss gradient. Y-axis shows AUC, X-axis shows the number of eigenvectors estimated by the Lanczos algorithm, data sampled every 500 eigenvectors. Tasks from left to right are M/E, M/H, A/E, A/H. 0 500 1000 1500 2000 2500 3000 0 500 1000 1500 2000 2500 3000 0.48 0.49 0.50 0.51 0.52 0.53 0.54 0 500 10001500200025003000 0.5180 0.5185 0.5190 0.5195 0.5200 0.5205 0.5210 0.5215 0.5220 0 500 1000 1500 2000 2500 3000 0.52 0.53 0.54 0.55 0.56 0.57 0.58 Figure 14: Latent Factors OOD detection task, using loss gradient. Y-axis shows AUC, X-axis shows the number of eigenvectors estimated by the Lanczos algorithm, data sampled every 500 eigenvectors. Tasks from left to right are M/E, M/H, A/E, A/H. E.2 BEHAVIOUR OF AUC WITH MORE ESTIMATED EIGENVECTORS In Fig. 13 and 14, we show that the tasks present differing behaviours as more eigenvectors are estimated. We observe that for the Male/Eyeglasses and Attractive/Wearing Hat tasks, we get improved performance with more eigenvectors, but for the others we do not necessarily see the same improvements. Interestingly, this upward trend occurs both times that our method achieves a statistically significant improvement over baselines. It is unclear why this occurs for some settings of the task and not others, but we hypothesize that this is a sign that the method is working more correctly in these settings. E.3 RELATIONSHIP BETWEEN LOSS GRADIENT AND Max Prob METHOD As discussed in Sec. 5.3, we have the relationship between the loss ℓ, prediction ˆY , and parameters θ: θℓ= ˆY ℓ θ ˆY . Using min as an aggregation function, we find that min Y {0,1} | ˆY ℓ(Y, ˆY )| has an inverted V-shape (Fig. 15). This is a similar shape to 1 max( ˆY , 1 ˆY ), which is the metric implicitly measured by Max Prob. E.4 ESTIMATED EIGENSPECTRUM OF DIFFERENT CORRELATED LATENT FACTOR TASKS 0.0 0.2 0.4 0.6 0.8 1.0 Figure 15: X-axis is ˆY , Y-axis is min Y {0,1} ˆY ℓ(Y, ˆY ). In Fig. 16, we examine the estimated eigenspectrums of the four tasks we present in the correlated latent factors experiment, to see if we can detect a reason why performance might differ on these tasks. In Fig. 16a, we show that the two tasks with the label Attractive, the eigenvalues are larger. These are also the tasks where the lossgradient-based variant of local ensembles failed, indicating that that variant of the method may be worse at handling larger eigenvalues. In Fig. 16b, we show that the two tasks where the local ensembles methods performed best (M/E, A/H, achieving statistically significant improvements over the baselines at a 95% confidence level, and also showing improvement as more eigenvectors were estimated), the most prominent negative eigenvalue is relatively smaller magnitude compared to the most prominent positive eigenvalue. This could mean that the local ensembles method was less successful in the other tasks (M/H, A/E) Published as a conference paper at ICLR 2020 0 500 1000 1500 2000 2500 3000 15.0 log(|Eigenvalue|) Male_Eyeglasses Male_Wearing_Hat Attractive_Eyeglasses Attractive_Wearing_Hat (a) Log absolute eigenvalues 0 500 1000 1500 2000 2500 3000 Normalized Eigenvalue Male_Eyeglasses Male_Wearing_Hat Attractive_Eyeglasses Attractive_Wearing_Hat (b) Normalized eigenvalues Figure 16: We show the estimated eigenspectrum of the four CNNs we train on the correlated latent factors task. In Fig. 16a, we show the absolute estimated eigenvalues sorted by absolute value, on a log scale. In Fig. 16b, we show the estimated eigenvalues divided by the maximum estimated eigenvalues. We only show 3000 estimated eigenvalues because we ran the Lanczos iteration for only 3000 iterations, meaning we did not estimate the rest of the eigenspectrum. simply because those models were not trained close enough to a convex minimum and still had fairly significant eigenvalues. F EMPIRICAL COMPARISON TO MC-DROPOUT One of the strengths of our method is that it can be applied post-hoc, making it usable in a wide range of situations. Therefore, we choose to compare the local ensembles method to only baselines which can also be applied post-hoc. However, one might still be interested in how some of these methods compare against ours. One such method is the MC-Dropout method (Gal & Ghahramani, 2016), which constructs an ensemble at test time from a model trained with dropout, by averaging over several random dropout inferences. This method is not considered post-hoc, since it is only applicable to methods which have been trained with dropout (and works best for models with dropout on each layer, rather than just a single layer). Figure 17: Performance of MC-dropout on the simulated features task from Sec. 5.2. Y-axis is AUC, X-axis is number of features simulated. This is the 0 noise setting. Both MC-dropout and local ensembles estimate the variance of an ensemble achieved through perturbations of a trianed model. In MC-dropout, those perturbations take the form of stochastically setting some weights of the model to 0; on local ensembles, they are Gaussian perturbation projected into the ensemble subspace. However, despite these similarities, we provide empirical evidence here that our method is categorically different than MCDropout, by showing that our perturbations find a flatter (more loss-preserving) ensemble. This means that our method is more directly measuring underdetermination, since we are comparing nearby models which are equally justified by the training data. We use the four tabular datasets discussed in Sec 5.2. For each dataset, we train a neural network with dropout on each weight (as in MC-dropout), to minimize a mean squared error objective. We use a dropout parameter p = 0.1, as suggested in Gal & Ghahramani (2016). We then create an MC-dropout ensemble of size 50: that is, we randomly dropout each weight with probability 0.1, and repeat 50 times. We calculate the training loss for each model in the MC-dropout ensemble. To estimate how flat this ensemble is (i.e. how loss-preserving on the training set), we simply subtract the training loss of the full model (no weights dropped out) from each model in the dropout- Published as a conference paper at ICLR 2020 0 500 1000 1500 2000 2500 3000 3500 Increase in Training Loss MC-Dropout Local Ensembles 0 500 1000 1500 2000 2500 3000 3500 Increase in Training Loss MC-Dropout Local Ensembles (b) Diabetes 0 500 1000 1500 2000 2500 3000 3500 Increase in Training Loss MC-Dropout Local Ensembles (c) Abalone 0 500 1000 1500 2000 2500 3000 3500 Increase in Training Loss MC-Dropout Local Ensembles (d) Wine Quality Figure 18: Comparison of loss-preservingness of ensembles found by MC-Dropout vs Local Ensembles methods. X-axis shows different values for m (number of dimensions projected out), with p m dimensions in the ensemble subspace. Y-axis shows expected increase in training loss across the ensemble, as compared to the non-perturbed model (no weights dropped out/perturbation = 0). Gaussian perturbation is scaled to have the same expected size as dropout perturbations for that task and m-value. Standard deviations shown across 20 random initializations. For each initialization, 50 perturbations were performed. Dotted red horizontal line shows average training loss increase for an MC-dropout ensemble (20 models, 50 perturbations each). ensemble, and average these differences. We repeat this process 20 times, changing only the random seeds. We can now estimate a similar quantity (average increase in training loss across the ensemble) for our method. For each dataset, we train a neural network with the same architecture as in the dropout experiment. We then choose a standard deviation σ equal to the average magnitude of the perturbations created through MC-dropout on the corresponding task. This is to ensure a fair comparison for each method, we are using perturbations on the same scale. We run the Lanczos iteration to convergence. We then compute the ensemble subspace for each Lanczos iteration m (recall that p m is the dimensionality of the ensemble subspace, and m is the number of dimensions projected out). For each ensemble subspace of dimensionality p m, we sample Gaussian noise from N(0, σ p m), project it into the ensemble subspace, and add it to the model s trained parameters. This ensures that when projected into the ensemble subspace, the Gaussian perturbation has the same expected magnitude as the dropout perturbations for the corresponding task. Using these projected Gaussian perturbations, we create an ensemble of size 50 for each m, and calculate the training loss for each model in the local ensemble. We subtract the training loss of the original model (perturbation 0) from each model in the local ensemble, and average across the ensemble. We repeat this process 20 times, changing only the random seeds. In Figure 18, we show the results of this comparison for each of the four datasets. The decreasing curve shows that as we project out more eigenvectors (increasing m), the resulting ensemble subspace becomes flatter, i.e. perturbations in this subspace increase training loss by a smaller amount. The Published as a conference paper at ICLR 2020 horizontal dotted line is the average training loss increase for model in an an MC-dropout ensemble. We see that this horizontal line is well above the descending curve representing local ensembles. Recall that we scaled the perturbations to be the same size for both MC-Dropout and all value of m for local ensembles. This means that, for perturbations of the same magnitude, for most values of m, the ensembles found by our method are much flatter than those found by MC-dropout. While MC-dropout ensembles may be useful for other tasks, this shows that our method is more directly measuring underdetermination in the model. Additionally, we run the MC dropout method on the simulated features task from Sec. 5.2. We find (see Fig. 17) that it is unable to match the performance of our method, reinforcing the point that it is concerned mostly with other types of uncertainty.