# locally_adaptive_label_smoothing_improves_predictive_churn__b6c4de7d.pdf Locally Adaptive Label Smoothing for Predictive Churn Dara Bahri 1 Heinrich Jiang 1 Training modern neural networks is an inherently noisy process that can lead to high prediction churn disagreements between re-trainings of the same model due to factors such as randomization in the parameter initialization and mini-batches even when the trained models all attain similar accuracies. Such prediction churn can be very undesirable in practice. In this paper, we present several baselines for reducing churn and show that training on soft labels obtained by adaptively smoothing each example s label based on the example s neighboring labels often outperforms the baselines on churn while improving accuracy on a variety of benchmark classification tasks and model architectures. 1. Introduction Deep neural networks (DNNs) have proved to be immensely successful at solving complex classification tasks across a range of problems. Much of the effort has been spent towards improving their predictive performance (i.e. accuracy), while comparatively little has been done towards improving the stability of training these models (Zheng et al., 2016). Modern DNN training is inherently noisy due to factors such as the random initialization of network parameters (Glorot & Bengio, 2010), the minibatch ordering (Loshchilov & Hutter, 2015), the effects of various data augmentation (Shorten & Khoshgoftaar, 2019) or pre-processing tricks (Santurkar et al., 2018), and the non-determinism arising from the hardware (Turner & Nowotny, 2015), all of which are exacerbated by the nonconvexity of the loss surface (Scardapane & Wang, 2017). This results in local optima corresponding to models that have very different predictions on the same data points. This may seem counter-intuitive, but even when the different runs all produce very high accuracies for the classification task, their predictions can still differ quite drastically 1Google Research, Mountain View, USA. Correspondence to: Dara Bahri . Proceedings of the 38 th International Conference on Machine Learning, PMLR 139, 2021. Copyright 2021 by the author(s). as we will show later in the experiments. Thus, even an optimized training procedure can lead to high prediction churn, which refers to the proportion of sample-level disagreements between classifiers caused by different runs of the same training procedure1. In practice, reducing such predictive churn can be critical. For example, in a production system, models are often continuously improved on by being trained or retrained with new data or better model architectures and training procedures. In such scenarios, a candidate model for release must be compared to the current model serving in production. Oftentimes, this decision is conditioned on more than just overall offline test accuracy in fact, the offline metrics are often not completely aligned with the actual goal, especially if these models are used as part of a larger system (e.g. maximizing offline click-through rate vs. maximizing revenue or user satisfaction) (Deng et al., 2013; Beel et al., 2013; Dmitriev & Wu, 2016). As a result, these comparisons require extensive and costly live experiments, requiring human evaluation in situations where the candidate and the production model disagree (i.e. in many situations, the true labels are not available without a manual labeler) (Theocharous et al., 2015; Deng, 2015; Deng & Shi, 2016). In these cases, it can be highly desirable to lower predictive churn. Despite the practical relevance of lowering churn, there has been surprisingly little work done in this area, which we highlight in the related work section. In this work, we focus on predictive churn reduction under retraining the same model architecture on an identical train and test set. Our main contributions are as follows: We provide one of the first comprehensive analyses of baselines to lower prediction churn, showing that popular approaches designed for other goals are effective baselines for churn reduction, even compared to methods designed for this goal. We improve label smoothing, a global smoothing method popular for calibrating model confidence, by utilizing the local information leveraged by the k- 1Concretely, given two classifiers applied to the same test samples, the prediction churn between them is the fraction of test samples with different predicted labels. Locally Adaptive Label Smoothing for Predictive Churn NN labels thus introducing a locally adaptive label smoothing which we show to often outperform the baselines on a wide range of benchmark datasets and model architectures. We show new theoretical results for the k-NN labels suggesting the usefulness of the k-NN label. We show under mild nonparametric assumptions that for a wide range of k, the k-NN labels uniformly approximates the optimal soft label and when k is tuned optimally, achieves the minimax optimal rate. We also show that when k is linear in n, the distribution implied by the k-NN label approximates the original distribution smoothed with an adaptive kernel. 2. Related Works Our work spans multiple sub-areas of machine learning. The main problem this paper tackles is reducing prediction churn. In the process, we show that label smoothing is an effective baseline and we improve upon it in a principled manner using deep k-NN label smoothing to obtain a locally adaptive version of it. Prediction Churn. There are only a few works which explicitly address prediction churn. Fard et al. (2016) proposed training a model so that it has small prediction instability with future versions of the model by modifying the data that the future versions are trained on. They furthermore propose turning the classification problem into a regression towards corrected predictions of an older model as well as regularizing the new model towards the older model using example weights. Cotter et al. (2019); Goh et al. (2016) use constrained optimization to directly lower prediction churn across model versions. Simultaneously training multiple identical models (apart from initialization) while tethering their predictions together via regularization has been proposed in the context of distillation (Anil et al., 2018; Zhang et al., 2018; Zhu et al., 2018; Song & Chai, 2018) and robustness to label noise (Malach & Shalev-Shwartz, 2017; Han et al., 2018). This family of methods was termed co-distillation by Anil et al. (2018), who also noted that it can be used to reduce churn in addition to improving accuracy. In this paper, we show much more extensively that co-distillation is indeed a reasonable baseline for churn reduction. Label Smoothing. Label smoothing (Szegedy et al., 2016) is a simple technique wherein the model is trained on the soft labels obtained by a convex combination of the hard true label and the soft uniform distribution across all the labels. It has been shown that it leads to better confidence calibration and generalization (M uller et al., 2019). Here we show that label smoothing is a reasonable baseline for reducing prediction churn, and we moreover enhance it for this task by smoothing the labels locally via k-NN rather than a pure global approach mixing with the uniform distribution. k-NN Theory. The theory of k-NN classification has a long history (e.g. Fix & Hodges Jr (1951); Cover (1968); Stone (1977); Devroye et al. (1994); Chaudhuri & Dasgupta (2014)). To our knowledge, the most relevant k-NN classification result is by Chaudhuri & Dasgupta (2014), who show statistical risk bounds under similar assumptions as used in our work. Our analysis shows finite-sample L bounds on the k-NN labels, which is a stronger notion of consistency as it provides a uniform guarantee, rather than an average guarantee as is shown in previous works under standard risk measures such as L2 error. We do this by leveraging recent techniques developed in Jiang (2019) for k-NN regression, which assumes an additive noise model instead of classification. Moreover, we provide to our knowledge the first consistency guarantee for the case where k grows linearly with n. Deep k-NN. k-NN is a classical method in machine learning which has recently been shown to be useful when applied to the intermediate embeddings of a deep neural network (Papernot & Mc Daniel, 2018) to obtain more calibrated and adversarially robust networks. This is because standard distance measures are often better behaved in these representations leading to better performance of k NN on these embeddings than on the raw inputs. Jiang et al. (2018) uses nearest neighbors on the intermediate representations to obtain better uncertainty scores than softmax probabilities and Bahri et al. (2020) uses the k-NN label disagreement to filter noisy labels for better training. Like these works, we also leverage k-NN on the intermediate representations but we show that utilizing the k-NN labels leads to lower prediction churn. 3. Algorithm Suppose that the task is multi-class classification with L classes and the training datapoints are (x1, y1), ..., (xn, yn), where xi X, and X is a compact subset of RD and yi RL, represents the one-hot vector encoding of the label that is, if the i-th example has label j, then yi has 1 in the j-th entry and 0 everywhere else. We give the formal definition of the smoothed labels: Definition 1 (Label Smoothing). Given label smoothing parameter 0 a 1, then the smoothed label y is (where 1L denotes the vector of all 1 s in RL). y LS a := (1 a) y + a We next formally define the k-NN label, which is the average label of the example s k-nearest neighbors in the Locally Adaptive Label Smoothing for Predictive Churn Figure 1. Visualization of the effects of global vs locally adaptive label smoothing. This visualization provides intuition for why our locally adaptive label smoothing method can improve neural network training stability. Left: A binary classification dataset in 2 dimensions where there are magenta points for the positive class and cyan points for the negative class. The data is generated from a mixture of two Gaussians, where the bottom-left Gaussian corresponds to the positive examples and the top-right corresponds to the negative examples, and to add label noise, we swap the labels of 10% chosen uniformly. Middle: We see that label smoothing simply pushes the labels uniformly closer to the average label (0.5). In particular, we see that the noisy labels still remain and thus may still cause conflicting information during training, possibly leading to predictive churn. Right: We now show our locally adaptive label smoothing approach, which also smooths the labels based on local information. This alleviates the examples with noisy labels by bringing them more in line with the average label amongst its neighbors and provides a more locally smooth label profile with respect to the input space. Such smoothness can help model training converge in a more stable manner. training set. Let us use shorthand X := {x1, ..., xn} and yi RL. Definition 2 (k-NN label). Let the k-NN radius of x X be rk(x) := inf{r : |B(x, r) X| k} where B(x, r) := {x X : |x x | r} and the k-NN set of x X be Nk(x) := B(x, rk(x)) X. Then for all x X, the k-NN label is defined as ηk(x) := 1 |Nk(x)| i=1 yi 1 [xi Nk(x)] . The label smoothing method can be seen as performing a global smoothing. That is, every label is equally transformed towards the uniform distribution over all labels. While it seems almost deceptively simple, it has only recently been shown to be effective in practice, specifically for better calibrated networks (M uller et al., 2019). However, since this smoothing technique is applied equally to all datapoints, it fails to incorporate local information about the datapoint. To this end, we propose using the k-NN label, which smooths the label across its nearest neighbors. We show theoretically that the k-NN label can be a strong proxy for the optimal soft label, that is, the expected label given the features and thus the best prediction one can make given the uncertainty under an L2 risk measure. In other words, compared to the true label (or even the label smoothing), the k-NN label is robust to variability in the data distribution and provides a more stable estimate of the label than the original hard label which may be noisy. Training on such noisy labels have been shown to hurt model performance (Bahri et al., 2020) and using the smoothed labels can help mitigate these effects. To this end, we define k-NN label smoothing as follows: Definition 3 (k-NN label smoothing). Let 0 a, b 1 be k-NN label smoothing parameters. Then the k-NN smoothed label of datapoint (x, y) is defined as: yk NN a,b = (1 a) y + a b 1 L 1L + (1 b) ηk(x) . We see that a is used to weight between using the true labels vs. using smoothing, and b is used to weight between the global vs. local smoothing. We provide an illustrative simulation in Figure 1. Algorithm 1 shows how k-NN label smoothing is applied to deep learning models. Like Bahri et al. (2020), we perform k-NN on the network s logits layer. Algorithm 1 Deep k-NN locally adaptive label smoothing 1: Inputs: 0 a, b 1, k, training data (x1, y1), ..., (xn, yn), model training procedure M. 2: Train model M0 on (x1, y1), ..., (xn, yn) with M. 3: Let z1, ..., zn RL be the logits of x1, ..., xn, respectively, w.r.t. M0. 4: Let yk NN i be the k-NN smoothed label (see Definition 3) of (zi, yi) computed w.r.t. (z1, y1), ..., (zn, yn). 5: Train model M on (x1, yk NN 1 ), ..., (xn, yk NN n ) with M. 4. Theoretical Analysis In this section, we provide theoretical justification for why the k-NN labels may be useful. In particular, we show results for two settings, where n is the number of datapoints. Locally Adaptive Label Smoothing for Predictive Churn When k n, we show that with appropriate setting of k, the k-NN smoothed labels approximate the predictions of the optimal soft classifier at a minimaxoptimal rate. When k = O(n), we show that the distribution implied by the k-NN smoothed labels is equivalent to the original distribution convolved with an adaptive smoothing kernel. Our results may also reveal insights into why distillation methods (the procedure of training a model on another model s predictions instead of the true labels) can work. Another way of considering the result is that the k-NN smoothed label is equivalent to the soft prediction of the k NN classifier. Thus, if one were to train on the k-NN labels, it would essentially be distillation on the k-NN classifier and our theoretical results show that the labels implied by k-NN approximate the predictions of the optimal classifier (in the k n setting). Learning the optimal classifier may indeed be a better goal than learning from the true labels, because the latter may lead to overfitting to the sampling noise rather than just the true signal implied by the optimal classifer. While distillation is not the topic of this work, our results in this section may be of independent interest to that area. For the analysis, we assume the binary classification setting, but it is understood that our results can be straightforwardly generalized to the multi-class setting. The feature vectors are defined on compact support X RD and datapoints are drawn as follows: the feature vectors are drawn from density p X on X and the labels are drawn according to the label function η : X [0, 1], i.e. η(x) = P(Y = 1|X = x). We make a few mild regularity assumptions for our analysis to hold, which are standard in works analyzing nonparametric methods (Singh et al., 2009; Chaudhuri & Dasgupta, 2014; Reeve & Kaban, 2019; Jiang, 2019; Bahri et al., 2020). The first part ensures that the support X does not become arbitrarily thin anywhere, the second ensures that the density does not vanish anywhere in the support, and the third ensures that the label function η is smooth w.r.t. to its input. Assumption 1. The following three conditions hold: Support Regularity: There exists ω > 0 and r0 > 0 such that Vol(X B(x, r)) ω Vol(B(x, r)) for all x X and 0 < r < r0, where B(x, r) := {x X : |x x | r}. Non-vanishing density: p X,0 := infx X p X(x) > 0. Smoothness of η: There exists 0 < α 1 and Cα > 0 such that |η(x) η(x )| Cα|x x |α for all x, x X. We have the following result which provides a uniform bound between the smoothed k-NN label ηk and the optimal soft label η. Theorem 1. Let 0 < δ < 1 and suppose that Assumption 1 holds and that k satisfies the following: 28 D log2(4/δ) log n k 1 2 ω p X,0 v D r D 0 n, where v D := πD/2 Γ(d/2+1) is the volume of a D-dimensional unit ball. Then with probability at least 1 δ, we have sup x X |ηk(x) η(x)| Cα 2k ω v D n p X,0 2 log(4D/δ) + 2D log(n) In other words, there exists constants C1, C2, C depending on η and δ such that if k satisfies C1 log n k C2 n, then with probability at least 1 δ, ignoring logarithmic factors in n and 1/δ: sup x X |ηk(x) η(x)| C Choosing k n2α/(2α+D), gives us a bound of supx X |ηk(x) η(x)| e O(n 1/(2α+D)), which is the minimax optimal rate as established by Tsybakov et al. (1997). Therefore, the advantage of using the smoothed labels ηk(x1), ..., ηk(xn) instead of the true labels y1, ..., yn, is that the smoothed labels approximate the optimal soft classifier. Moreover, as shown above, with appropriate setting of k, the smoothed labels are a minimax-optimal estimator of the true label function η. Thus, the smoothed labels provide as good of a proxy for η as any estimator possibly can. As suggested earlier, another way of considering this result is that the original labels may contain considerable noise and thus no single label can be guaranteed reliable. Using the smoothed label instead mitigates this effect and allows us to train the model to match the label function η. 4.2. k linear in n In the previous subsection, we showed the utility of k-NN label smoothing as a theoretically sound proxy for the optimal soft labels, which attains statistical consistency guarantees as long as k grows faster than log n and k/n 0. Locally Adaptive Label Smoothing for Predictive Churn Now, we analyze the case where k grows linearly with n. In this case, the k-NN smoothed labels no longer recover the optimal soft label function η, but instead an adaptive kernel smoothed version of η. We make this relationship precise here. Suppose that k = β n for some 0 < β < 1. We define the β-smoothed label function. Definition 4 (β-smoothed label function). Let rβ(x) := inf{r > 0 : P(B(x, r)) β}, that is the radii of the smallest ball centered at x with probability mass β w.r.t. PX. Then, let eηβ(x) be the expectation of η on B(x, rβ(x)) w.r.t. PX: eηβ(x) := 1 B(x,rβ(x)) η(x) PX(x)dx. We can view eηβ as an adaptively kernel smoothed version of η, where adaptivity arises from the density of the point (the more dense, the smaller the bandwidth we smooth it across) and the kernel is based on the density. We now prove the following result which shows that in this setting ηk estimates eηβ(x). It is worth noting that we need very little assumption on η as compared to the previous result because the β-smoothing of η provides a more regular label function; moreover, the rates are fast (i.e. e O( p Theorem 2. Let 0 < δ < 1 and k = β n . Then with probability at least 1 δ, we have for n sufficiently large depending on β, δ: sup x X |ηk(x) eηβ(x)| 3 2 log(4D/δ) + 2D log(n) 5. Experiments We now describe the experimental methodology and results for validating our proposed method. 5.1. Baselines We start by detailing the suite of baselines we compare against. We tune baseline hyper-parameters extensively, with the precise sweeps and setups available in the Appendix. Control: Baseline where we train for accuracy with no regards to churn. ℓp Regularization: We control the stability of a model s predictions by simply regularizing them (independently of the ground truth label) using classical ℓp regularization. The loss function is given by: Lℓp(xi, yi) = L(xi, yi) + a||f(xi)||p. We experiment with both ℓ1 and ℓ2 regularization. Bi-tempered: This is a baseline by Amid et al. (2019), originally designed for robustness to label noise. It modifies the standard logistic loss function by introducing two temperature scaling parameters t1 and t2. We apply their bi-tempered loss here, suspecting that methods which make model training more robust to noisy labels may also be effective at reducing churn. Anchor: This is based on a method proposed by Fard et al. (2016) specifically for churn reduction. It uses the predicted probabilities from a preliminary model to smooth the training labels of the second model. We first train a preliminary model fprelim using regular cross-entropy loss. We then retrain the model using smoothed labels (1 a)yi + afprelim(xi), thus anchoring on a preliminary model s predictions. In our experiments, we train one preliminary model and fix it across the runs for this baseline to reduce prediction churn. Co-distillation: We use the co-distillation approach presented by Anil et al. (2018), who touched upon its utility for churn reduction. We train two identical models M1 and M2 (but subject to different random initialization) in tandem while penalizing divergence between their predictions. The overall loss is Lcodistill(xi, yi) = L(f1(xi), yi) + L(f2(xi), yi) + aΨ(f1(xi), f2(xi)). In their paper the authors set Ψ to be cross-entropy: Ψ(p(1), p(2)) = X i [L] p(1) i log(p(2) i ), but they note KL divergence can be used. We experiment with both cross-entropy and KL divergence. We also tune nwarm, the number of burn-in steps of training before turning on the regularizer. Label Smoothing: This is the method of Szegedy et al. (2016) defined earlier in the paper. Our proposed method augments global label smoothing by leveraging the local k-NN estimates. Naturally, we compare against doing global smoothing only and this serves as a key ablation model to see the added benefits of leveraging the k-NN labels. Mixup: This method proposed by Zhang et al. (2017) generates synthetic training examples on the fly by convex-combining random training inputs and their associated labels, where the combination weights are random draws from a Beta(a, a) distribution. Mixup improves generalization, increases robustness to adversarial examples as well as label noise, and also improves model calibration (Thulasidasan et al., 2019). Locally Adaptive Label Smoothing for Predictive Churn Ensemble: Ensembling deep neural networks can improve the quality of their uncertainty estimation (Lakshminarayanan et al., 2017; Fort et al., 2019). We consider the simple case where m identical deep neural networks are trained independently on the same training data, and at inference time, their predictions are uniformly averaged together. 5.2. Datasets and Models. We do not use standard data augmentation strategies for the image datasets so that the training data is constant across different training rounds. For all datasets we use the Adam optimizer with default learning rate 0.001. We use a minibatch size of 128 throughout. MNIST: We train a three-layer MLP with 256 hidden units and Re LU activations for 20 epochs. Fashion MNIST: We use the same architecture as the one used for MNIST. SVHN: We use Le Net5 CNN (Le Cun et al., 1998) for 30 epochs on the Google Street View Housing Numbers (SVHN) dataset, where each image is cropped to be 32 32 pixels. Celeb A: Celeb A (Liu et al., 2018) is a large-scale face attributes dataset with more than 200k celebrity images, each with 40 attribute annotations. We use the standard train and test splits, which consist of 162770 and 19962 images respectively. Images were resized to be 28 28 3. We select the smiling and high cheekbone attributes and perform binary classification, training Le Net5 for 20 epochs. Phishing: To validate our method beyond the image classification setting, we train a three-layer MLP with 256 hidden units per layer on UCI Phishing dataset (Dua & Graff, 2017), which consists of 7406 train and 3649 test examples on a 30-dimensional input feature. 5.3. Evaluation Metrics and Hyperparameter Tuning For each dataset, baseline and hyper-parameter setting, we run each method on the same train and test split exactly 5 times. We then report the average test accuracy as well as the test set churn averaged across every possible pair (i, j) of runs (10 total pairs). To give a more complete picture of the sources of churn, we also slice the churn by the whether or not the test predictions of the first run in the pair were correct. Then, lowering the churn on the correct predictions is desirable (i.e. if the base model is correct, we clearly don t want the predictions to change), while churn Figure 2. Performance across hyperparameters for SVHN. For each of the three hyperparameters of our method (a, b, and k), we show the performance across a range of this hyperparameter keeping the other two hyperparameters fixed. Standard error bands are shaded. Top: We tune a while fixing k = 10 and b = 0.9. We see better performance under both accuracy and churn as a increases, which suggests that the less weight we put on the original label, the better. Middle: We tune b while fixing k = 10 and a = 0.5. We don t see any clear pattern which suggests that b is an essential hyperparameter trading off the locally adaptive vs global smoothing this suggests that our adding the locally adaptive component to the label smoothing is indeed having an effect on performance. Bottom: We tune k while fixing a = 1 and b = 0.9. We see that the accuracy and churn have little differences across a wide range of k from k = 10 to k = 500, which suggests that k is not an essential hyperparameter and that we are stable in it. Locally Adaptive Label Smoothing for Predictive Churn Dataset Method Accuracy % Churn % Churn Correct Churn Incorrect k-NN LS (k=10, a=1, b=0.9) 88.98 (0.33) 10.98 (0.28) 4.64 (0.29) 62.23 (1.22) Label Smoothing (a=0.9) 87.26 (0.73) 13.46 (0.62) 5.31 (0.57) 67.2 (1.44) Anchor (a=1.0) 87.17 (0.16) 12.48 (0.39) 5.19 (0.2) 61.66 (1.85) SVHN ℓ2 Reg (a=0.5) 88.16 (0.35) 11.85 (0.35) 5.07 (0.16) 62.73 (2.1) ℓ1 Reg (a=0.2) 74.18 (3.41) 22.89 (3.74) 9.58 (4.04) 59.36 (5.7) Co-distill (CE, a=0.5) 87.64 (0.64) 12.46 (0.48) 5.16 (0.51) 63.82 (1.67) Co-distill (KL, a=0.5) 87.52 (0.45) 13.01 (0.3) 5.54 (0.33) 65.44 (1.46) Bi-tempered (t1=0.5, t2=1) 88.04 (0.5) 12.03 (0.3) 5.26 (0.3) 62.48 (1.83) Mixup (a=0.5) 89.08 (0.18) 9.56 (0.16) 4.07 (0.15) 54.75 (0.95) Control 86.64 (0.54) 14.64 (0.51) 6.03 (0.5) 69.59 (1.32) k-NN LS (k=5, a=0.9, b=0.9) 98.23 (0.11) 1.52 (0.12) 0.7 (0.1) 47.16 (3.39) Label Smoothing (a=0.9) 98.15 (0.07) 1.65 (0.05) 0.71 (0.07) 50.73 (2.62) Anchor (a=1.0) 97.72 (0.11) 2.66 (0.2) 1.21 (0.14) 64.51 (4.13) MNIST ℓ2 Reg (a=0.5) 98.08 (0.1) 1.67 (0.12) 0.8 (0.08) 46.65 (3.2) ℓ1 Reg (a=0.01) 97.67 (0.29) 2.51 (0.31) 1.3 (0.27) 56.8 (2.84) Co-distill (CE, a=0.2, nwarm=2k) 98.08 (0.06) 2.08 (0.11) 0.98 (0.07) 58.6 (3.91) Co-distill (KL, a=0.05, nwarm=1k) 97.98 (0.14) 2.16 (0.16) 0.97 (0.13) 59.56 (3.64) Bi-tempered (t1=0.9, t2=1.0) 98.09 (0.2) 2.04 (0.15) 1.07 (0.14) 55.82 (4.32) Mixup (a=0.2) 98.17 (0.04) 1.59 (0.07) 0.74 (0.04) 47.8 (2.53) Control 97.98 (0.13) 2.28 (0.13) 0.96 (0.07) 63.36 (2.55) k-NN LS (k=10, a=1, b=0.5) 88.89 (0.14) 6.94 (0.18) 3.27 (0.15) 36.26 (1.09) Label Smoothing (a=0.8) 88.46 (0.17) 7.2 (0.46) 3.32 (0.28) 36.63 (2.02) Anchor (a=0.9) 88.55 (0.14) 7.53 (0.45) 3.6 (0.23) 37.78 (2.29) Fashion ℓ2 Reg (a=0.5) 88.52 (0.19) 7.86 (0.36) 3.59 (0.18) 40.38 (1.81) MNIST ℓ1 Reg (a=0.1) 86.88 (0.35) 8.24 (0.55) 3.88 (0.41) 36.81 (2.63) Co-distill (CE, a=0.5, nwarm=2k) 88.76 (0.21) 7.51 (0.39) 3.67 (0.3) 37.98 (1.71) Co-distill (KL, a=0.5, nwarm=2k) 88.85 (0.35) 7.83 (0.43) 3.68 (0.29) 40.59 (2.4) Bi-tempered (t1=0.7, t2=2) 88.7 (0.29) 7.36 (0.47) 3.5 (0.19) 37.24 (3.04) Mixup (a=0.4) 89.17 (0.10) 6.77 (0.29) 3.23 (0.15) 35.97 (1.43) Control 88.95 (0.26) 9.13 (0.51) 4.42 (0.4) 46.99 (2.49) k-NN LS (k=100, b=0.1, a=0.9) 90.02 (0.11) 5.46 (0.32) 2.97 (0.18) 27.71 (1.74) Label Smoothing (a=0.05) 89.39 (0.29) 6.77 (0.41) 3.81 (0.26) 31.67 (2.34) Anchor (a=0.8) 89.87 (0.14) 5.57 (0.28) 3.07 (0.21) 27.66 (1.38) Celeb A ℓ2 Reg (a=0.01) 89.35 (0.16) 6.85 (0.34) 3.92 (0.27) 31.62 (1.21) Smiling ℓ1 Reg (a=0.5) 89.39 (0.26) 6.71 (0.26) 3.61 (0.22) 32.48 (1.35) Co-distill (CE, a=0.5, nwarm=1k) 89.59 (0.29) 6.31 (0.23) 3.66 (0.3) 29.47 (1.47) Co-distill (KL, a=0.5, nwarm=2k) 89.57 (0.22) 6.1 (0.23) 3.34 (0.26) 29.66 (1.47) Bi-tempered (t1=0.9, t2=2.) 89.88 (0.18) 6.44 (0.31) 3.56 (0.19) 31.96 (1.96) Mixup (a=0.2) 89.71 (0.14) 6.15 (0.12) 3.51 (0.12) 29.37 (0.66) Control 89.67 (0.19) 7.3 (0.45) 4.06 (0.27) 35.34 (2.35) k-NN LS (k=100, b=0.1, a=0.9) 84.48 (0.21) 7.7 (0.29) 4.64 (0.27) 24.44 (0.98) Label Smoothing (a=0.005) 83.73 (0.17) 8.68 (0.46) 5.2 (0.35) 26.61 (1.28) Anchor (a=0.9) 84.48 (0.2) 7.97 (0.39) 4.77 (0.22) 25.44 (1.58) Celeb A ℓ2 Reg (a=0.001) 83.6 (0.14) 9.06 (0.32) 5.41 (0.24) 27.66 (1.03) High ℓ1 Reg (a=0.01) 83.59 (0.26) 8.43 (0.23) 4.93 (0.23) 26.14 (1.16) Cheekbone Co-distill (CE, a=0.5, nwarm=1k) 84.08 (0.21) 8.96 (0.37) 5.33 (0.36) 28.11 (0.88) Co-distill (KL, a=0.5, nwarm=1k) 84.31 (0.08) 8.57 (0.16) 5.06 (0.13) 27.39 (0.47) Bi-tempered (t1=0.5, t2=4) 83.92 (0.13) 7.84 (0.32) 4.75 (0.21) 24.01 (1) Mixup (a=0.4) 84.53 (0.14) 7.92 (0.47) 4.69 (0.31) 25.53 (1.54) Control 83.93 (0.56) 10.18 (0.93) 6.2 (0.89) 31.1 (2.22) k-NN LS (k=500, a=0.8, b=0.9) 96.69 (0.09) 1.04 (0.21) 0.54 (0.14) 15.81 (3.52) Label Smoothing (a=0.8) 96.63 (0.09) 1.26 (0.26) 0.64 (0.17) 18.8 (3.42) Anchor (a=0.9) 96.02 (0.25) 2.33 (0.25) 1.08 (0.15) 31.58 (5.11) Phishing ℓ2 Reg (a=0.5) 96.51 (0.12) 1.35 (0.3) 0.7 (0.21) 19.37 (4) ℓ1 Reg (a=0.5) 95.38 (0.18) 1.48 (0.34) 0.83 (0.24) 14.95 (4.08) Co-distill (CE, a=0.2, nwarm=2) 96.02 (0.19) 1.45 (0.26) 0.83 (0.21) 16.72 (4.13) Co-distill (KL, a=0.001, nwarm=1k) 95.94 (0.33) 1.51 (0.2) 0.65 (0.18) 20.95 (6.14) Bi-tempered (t1=0.9, t2=1.0) 96.26 (0.37) 2.32 (0.69) 1.23 (0.53) 30.19 (8.51) Mixup (a=0.1) 96.22 (0.23) 1.80 (0.33) 1.05 (0.28) 21.53 (4.25) Control 96.3 (0.32) 2.25 (0.59) 1.21 (0.38) 29.05 (7.93) Table 1. Results across all datasets and baselines under optimal hyperparameter tuning (settings shown). Note that we report the standard deviation of the runs instead of standard deviation of the mean (i.e. standard error) which is often reported instead. The former is higher than the latter by a factor of the square root of the number of trials (10). Locally Adaptive Label Smoothing for Predictive Churn Figure 3. Accuracy vs. correct churn plots. This figure plots accuracy against correct churn achieved by various hyperparameter settings of each method. The top-left corner (high accuracy, low churn) is most desirable. The dotted lines depict the Pareto frontier for each method, and we zoom in for a detailed view. We fix k-NN LS k = 100 and the number of codistillation warmup steps to 1000. All other hyperparameters are sweeped as specified in the appendix. We see that k-NN label smoothing ranks reasonably high, mostly achieving low churn without sacrificing accuracy. reduction on incorrect predictions is less relevant (i.e. if the base model was incorrect, then it may be better for there to be higher churn however at the same time, some examples may be inherently difficult to classify or the label is such an outlier that we don t expect an optimal model to correctly classify, in which case lower churn may be desirable). This is why in the results for Table 1, we bold the best performing baseline for churn on correct examples, but not for churn on incorrect examples. In the results (Table 1), for each dataset and baseline, we chose the optimal hyperparameter setting by first sorting by accuracy and choosing the setting with the highest accuracy, and if there were multiple settings with very close Locally Adaptive Label Smoothing for Predictive Churn to the top accuracy (defined as within less than 0.1% difference in test accuracy), then we chose the setting with the lowest churn among those settings with accuracy close to the top accuracy. There is often no principled way to trade-off the two sometimes competing objectives of accuracy and churn. Cotter et al. (2019) offer a heuristic to trade off the two objectives in a more balanced manner on the Pareto frontier. However in this case, biasing towards higher accuracy is most realistic because in practice, when given a choice between two models, it s usually best to go with the more accurate model. Fortunately, we will see that accuracy and churn are not necessarily competing objectives and our proposed method usually gives the best result for both simultaneously. 5.4. Performance across hyperparameter settings In Figure 2, we show the performance on SVHN w.r.t. the hyperparameters for both accuracy and churn. We fix two of the hyperparameters and show the results across tunings of the remaining hyperparameters. We do this for each of the three hyperparameters of our approach (a, b and k). We see that larger a corresponds to better performance, implying that less weight on the original labels leads to better results. We also see that across a wide range of k, the performance did not change much, which suggests that in practice, k can be set to some default and not require tuning. Such stability in k is desirable. Hence, the remaining hyperparameter b, which decides the trade-off between the locally adaptive vs global smoothing appears most essential. This further shows that our proposal of using locally adaptive label smoothing has a real effect on the results for both churn and accuracy. 5.5. Results We see from Table 1 that mixup and our method, k-NN label smoothing, are consistently the most competitive; mixup outperforms on SVHN and Fashion MNIST while k-NN label smoothing outperforms on all the remaining datasets. Notably, both methods do well on accuracy and churn metrics simultaneously. Figure 3 plots accuracy versus churn for different hyperparameter settings and highlights the Pareto frontier. We find that the k-NN label smoothing is often Pareto efficient. Results for the ensemble baseline can be found in the Appendix. While we found ensembling to be remarkably effective, it does come with higher cost (more trainable parameters and higher inference cost), and so we discourage a direct comparison with other methods since an ensemble uses a different model class than a single model. 6. Conclusion Modern DNN training is a noisy process: randomization arising from stochastic minibatches, weight initialization, data preprocessing techniques, and hardware can all lead to models with drastically different predictions on the same datapoints when using the same training procedure. Reducing such prediction churn is important in practical problems as production ML models are constantly updated and improved on. Since offline metrics can usually only serve as proxies to the live metrics, comparing the models in A/B tests and live experiments oftentimes must involve manual labeling of the disagreements between the models, making it a costly procedure. Thus, controlling the amount of predictive churn can be crucial for more efficiently iterating and improving models in a production setting. Despite the practical importance of this problem, there has been little work done in the literature on this topic. We provide one of the first comprehensive analyses of reducing predictive churn arising from retraining the model on the same dataset and model architecture. We show that numerous methods used for other goals such as learning with noisy labels and improving model calibration serve as reasonable baselines for lowering prediction churn. We propose a new technique, locally adaptive label smoothing, that often outperforms the baselines across a range of datasets and model architectures. Further study in this area is critical: the problem of predictive churn has received far too little treatment in the academic literature given its practical significance. Our technique may also help in the subfields that we drew many of our baselines from, including better calibrated DNNs and robustness to label noise, suggesting a bi-directional flow of ideas between the goal of reducing predictive churn and these subfields. This is a direction for future work. Amid, E., Warmuth, M. K., Anil, R., and Koren, T. Robust bi-tempered logistic loss based on bregman divergences. In Advances in Neural Information Processing Systems, pp. 14987 14996, 2019. Anil, R., Pereyra, G., Passos, A., Ormandi, R., Dahl, G. E., and Hinton, G. E. Large scale distributed neural network training through online distillation. ar Xiv preprint ar Xiv:1804.03235, 2018. Bahri, D., Jiang, H., and Gupta, M. Deep k-nn for noisy labels. ICML, 2020. Beel, J., Genzmehr, M., Langer, S., N urnberger, A., and Gipp, B. A comparative analysis of offline and online evaluations and discussion of research paper recom- Locally Adaptive Label Smoothing for Predictive Churn mender system evaluation. In Proceedings of the international workshop on reproducibility and replication in recommender systems evaluation, pp. 7 14, 2013. Chaudhuri, K. and Dasgupta, S. Rates of convergence for the cluster tree. In Advances in neural information processing systems, pp. 343 351, 2010. Chaudhuri, K. and Dasgupta, S. Rates of convergence for nearest neighbor classification. In Advances in Neural Information Processing Systems, pp. 3437 3445, 2014. Cotter, A., Jiang, H., Gupta, M. R., Wang, S., Narayan, T., You, S., and Sridharan, K. Optimization with nondifferentiable constraints with applications to fairness, recall, churn, and other goals. Journal of Machine Learning Research, 20(172):1 59, 2019. Cover, T. M. Rates of convergence for nearest neighbor procedures. In Proceedings of the Hawaii International Conference on Systems Sciences, pp. 413 415, 1968. Deng, A. Objective bayesian two sample hypothesis testing for online controlled experiments. In Proceedings of the 24th International Conference on World Wide Web, pp. 923 928, 2015. Deng, A. and Shi, X. Data-driven metric development for online controlled experiments: Seven lessons learned. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 77 86, 2016. Deng, A., Xu, Y., Kohavi, R., and Walker, T. Improving the sensitivity of online controlled experiments by utilizing pre-experiment data. In Proceedings of the sixth ACM international conference on Web search and data mining, pp. 123 132, 2013. Devroye, L., Gyorfi, L., Krzyzak, A., Lugosi, G., et al. On the strong universal consistency of nearest neighbor regression function estimates. The Annals of Statistics, 22 (3):1371 1385, 1994. Dmitriev, P. and Wu, X. Measuring metrics. In Proceedings of the 25th ACM international on conference on information and knowledge management, pp. 429 437, 2016. Dua, D. and Graff, C. UCI machine learning repository, 2017. URL http://archive.ics.uci.edu/ml. Fard, M. M., Cormier, Q., Canini, K., and Gupta, M. Launch and iterate: Reducing prediction churn. In Advances in Neural Information Processing Systems, pp. 3179 3187, 2016. Fix, E. and Hodges Jr, J. L. Discriminatory analysisnonparametric discrimination: consistency properties. Technical report, California Univ Berkeley, 1951. Fort, S., Hu, H., and Lakshminarayanan, B. Deep ensembles: A loss landscape perspective. ar Xiv preprint ar Xiv:1912.02757, 2019. Glorot, X. and Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp. 249 256. JMLR Workshop and Conference Proceedings, 2010. Goh, G., Cotter, A., Gupta, M., and Friedlander, M. P. Satisfying real-world goals with dataset constraints. In Advances in Neural Information Processing Systems, pp. 2415 2423, 2016. Han, B., Yao, Q., Yu, X., Niu, G., Xu, M., Hu, W., Tsang, I., and Sugiyama, M. Co-teaching: Robust training of deep neural networks with extremely noisy labels. In Advances in neural information processing systems, pp. 8527 8537, 2018. Jiang, H. Non-asymptotic uniform rates of consistency for k-nn regression. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pp. 3999 4006, 2019. Jiang, H., Kim, B., Guan, M. Y., and Gupta, M. R. To trust or not to trust a classifier. In Advances in Neural Information Processing Systems (Neur IPS), 2018. Lakshminarayanan, B., Pritzel, A., and Blundell, C. Simple and scalable predictive uncertainty estimation using deep ensembles. In Advances in neural information processing systems, pp. 6402 6413, 2017. Le Cun, Y., Bottou, L., Bengio, Y., and Haffner, P. Gradientbased learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. Liu, Z., Luo, P., Wang, X., and Tang, X. Large-scale celebfaces attributes (celeba) dataset. Retrieved August, 15: 2018, 2018. Loshchilov, I. and Hutter, F. Online batch selection for faster training of neural networks. ar Xiv preprint ar Xiv:1511.06343, 2015. Malach, E. and Shalev-Shwartz, S. Decoupling when to update from how to update . In Advances in Neural Information Processing Systems, pp. 960 970, 2017. M uller, R., Kornblith, S., and Hinton, G. E. When does label smoothing help? In Advances in Neural Information Processing Systems, pp. 4694 4703, 2019. Papernot, N. and Mc Daniel, P. Deep k-nearest neighbors: Towards confident, interpretable and robust deep learning. ar Xiv preprint ar Xiv:1803.04765, 2018. Locally Adaptive Label Smoothing for Predictive Churn Reeve, H. W. and Kaban, A. Fast rates for a k NN classifier robust to unknown asymmetric label noise. ar Xiv preprint ar Xiv:1906.04542, 2019. Santurkar, S., Tsipras, D., Ilyas, A., and Madry, A. How does batch normalization help optimization? ar Xiv preprint ar Xiv:1805.11604, 2018. Scardapane, S. and Wang, D. Randomness in neural networks: an overview. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, 7(2):e1200, 2017. Shorten, C. and Khoshgoftaar, T. M. A survey on image data augmentation for deep learning. Journal of Big Data, 6(1):1 48, 2019. Singh, A., Scott, C., Nowak, R., et al. Adaptive Hausdorff estimation of density level sets. The Annals of Statistics, 37(5B):2760 2782, 2009. Song, G. and Chai, W. Collaborative learning for deep neural networks. In Advances in Neural Information Processing Systems, pp. 1832 1841, 2018. Stone, C. J. Consistent nonparametric regression. The Annals of Statistics, pp. 595 620, 1977. Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2818 2826, 2016. Theocharous, G., Thomas, P. S., and Ghavamzadeh, M. Ad recommendation systems for life-time value optimization. In Proceedings of the 24th International Conference on World Wide Web, pp. 1305 1310, 2015. Thulasidasan, S., Bhattacharya, T., Bilmes, J., Chennupati, G., and Mohd-Yusof, J. Combating label noise in deep learning using abstention. ar Xiv preprint ar Xiv:1905.10964, 2019. Tsybakov, A. B. et al. On nonparametric estimation of density level sets. The Annals of Statistics, 25(3):948 969, 1997. Turner, J. P. and Nowotny, T. Estimating numerical error in neural network simulations on graphics processing units. BMC Neuroscience, 16(198), 2015. Zhang, H., Cisse, M., Dauphin, Y. N., and Lopez-Paz, D. mixup: Beyond empirical risk minimization. ar Xiv preprint ar Xiv:1710.09412, 2017. Zhang, Y., Xiang, T., Hospedales, T. M., and Lu, H. Deep mutual learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4320 4328, 2018. Zheng, S., Song, Y., Leung, T., and Goodfellow, I. Improving the robustness of deep neural networks via stability training. In Proceedings of the ieee conference on computer vision and pattern recognition, pp. 4480 4488, 2016. Zhu, X., Gong, S., et al. Knowledge distillation by on-thefly native ensemble. In Advances in neural information processing systems, pp. 7517 7527, 2018.