# memorizationdilation_modeling_neural_collapse_under_noise__0b5282fd.pdf Published as a conference paper at ICLR 2023 MEMORIZATION-DILATION: MODELING NEURAL COLLAPSE UNDER LABEL NOISE Duc Anh Nguyen LMU Munich Ron Levie Technion Israel Institute of Technology Julian Lienen Paderborn University Gitta Kutyniok LMU Munich University of Tromsø Eyke Hüllermeier LMU Munich The notion of neural collapse refers to several emergent phenomena that have been empirically observed across various canonical classification problems. During the terminal phase of training a deep neural network, the feature embedding of all examples of the same class tend to collapse to a single representation, and the features of different classes tend to separate as much as possible. Neural collapse is often studied through a simplified model, called the layer-peeled model, in which the network is assumed to have infinite expressivity and can map each data point to any arbitrary representation. In this work we study a more realistic variant of the layer-peeled model, which takes the positivity of the features into account. Furthermore, we extend this model to also incorporate the limited expressivity of the network. Empirical evidence suggests that the memorization of noisy data points leads to a degradation (dilation) of the neural collapse. Using a model of the memorization-dilation (M-D) phenomenon, we show one mechanism by which different losses lead to different performances of the trained network on noisy data. Our proofs reveal why label smoothing, a modification of cross-entropy empirically observed to produce a regularization effect, leads to improved generalization in classification tasks. 1 INTRODUCTION The empirical success of deep neural networks has accelerated the introduction of new learning algorithms and triggered new applications, with a pace that makes it hard to keep up with profound theoretical foundations and insightful explanations. As one of the few yet particularly appealing theoretical characterizations of overparameterized models trained for canonical classification tasks, Neural Collapse (NC) provides a mathematically elegant formalization of learned feature representations Papyan et al. (2020). To explain NC, consider the following setting. Suppose we are given a balanced dataset D = n (x(k) n , yn) o k [K],n [N] X Y in the instance space X = Rd and label space Y = [N] := {1, . . . , N}, i.e. each class n [N] has exactly K samples x(1) n , . . . , x(K) n . We consider network architectures commonly used in classification tasks that are composed of a feature engineering part g: X RM (which maps an input signal x X to its feature representation g(x) RM) and a linear classifier W ( ) + b given by a weight matrix W RN M as well as a bias vector b RN. Let wn denote the row vector of W associated with class n [N]. During training, both classifier components are simultaneously optimized by minimizing the cross-entropy loss. *These authors contributed equally to this work. Published as a conference paper at ICLR 2023 Denoting the feature representations g(x(k) n ) of the sample x(k) n by h(k) n , the class means and the global mean of the features by i=1 h(k) n , h := 1 NC consists of the following interconnected phenomena (where the limits take place as training progresses): (NC1) Variability collapse. For each class n [N], we have 1 K PK k=1 h(k) n hn 2 0 . (NC2) Convergence to simplex equiangular tight frame (ETF) structure. For any m, n [N] with m = n, we have hn h 2 hm h 2 0, and hn h hn h 2 , hm h hm h 2 (NC3) Convergence to self-duality. For any n [N], it holds hn h hn h 2 wn wn 2 0 . (NC4) Simplification to nearest class center behavior. For any feature representation u RM, it holds arg max n [N] wn, u + bn arg min n [N] u hn 2 . In this paper, we consider a well known simplified model, in which the features h(k) n are not parameterized by the feature engineering network g but are rather free variables. This model is often referred to as layer-peeled model or unconstrained features model, see e.g. Lu & Steinerberger (2020); Fang et al. (2021); Zhu et al. (2021). However, as opposed to those contributions, in which the features h(k) n can take any value in RM, we consider here the case h(k) n 0 (understood componentwise). This is motivated by the fact that features are typically the outcome of some non-negative activation function, like the Rectified Linear Unit (Re LU) or sigmoid. Moreover, by incorporating the limited expressivity of the network to the layer-peeled model, we propose a new model, called memorization-dilation (MD). Given such model assumptions, we formally prove advantageous effects of the so-called label smoothing (LS) technique Szegedy et al. (2015) (training with a modification of cross-entropy (CE) loss), in terms of generalization performance. This is further confirmed empirically. 2 RELATED WORK Studying the nature of neural network optimization is challenging. In the past, a plethora of theoretical models has been proposed to do so Sun (2020). These range from analyzing simple linear Kunin et al. (2019); Zhu et al. (2020); Laurent & von Brecht (2018) to non-linear deep neural networks Saxe et al. (2014); Yun et al. (2018). As one prominent framework among others, Neural Tangent Kernels Jacot et al. (2018); Roberts et al. (2021), where neural networks are considered as linear models on top of randomized features, have been broadly leveraged for studying deep neural networks and their learning properties. Many of the theoretical properties of deep neural networks in the regime of overparameterization are still unexplained. Nevertheless, certain peculiarities have emerged recently. Among those, socalled benign overfitting Bartlett et al. (2019); Li et al. (2021), where deep models are capable of perfectly fitting potentially noisy data by retaining accurate predictions, has recently attracted attention. Memorization has been identified as one significant factor contributing to this effect Arpit et al. (2017); Sanyal et al. (2021), which also relates to our studies. Not less interesting, the learning risk of highly-overparameterized models shows a double-descent behavior when varying the model Published as a conference paper at ICLR 2023 complexity Nakkiran et al. (2020) as yet another phenomenon. Lastly, the concept of NC Papyan et al. (2020) has recently shed light on symmetries in learned representations of overparameterized models. After laying the foundation of a rigorous mathematical characterization of the NC phenomenon by Papyan et al. (2020), several follow-up works have broadened the picture. As the former proceeds from studying CE loss, the collapsing behavior has been investigated for alternative loss functions. For instance, squared losses have shown similar collapsing characteristics Poggio & Liao (2020; 2021), and have paved the way for more opportunities in its mathematical analysis, e.g., by an NC-interpretable decomposition Han et al. (2021). More recently, Kornblith et al. (2021) provide an exhaustive overview over several commonly used loss functions for training deep neural networks regarding their feature collapses. Besides varying the loss function, different theoretical models have been proposed to analyze NC. Most prominently, unconstrained feature models have been considered, which characterize the penultimate layer activations as free optimization variables Mixon et al. (2020); Lu & Steinerberger (2020); E & Wojtowytsch (2021). This stems from the assumption that highly overparameterized models can approximate any patterns in the feature space. While unconstrained features models typically only look at the last feature encoder layer, layer-peeling allows for white-boxing further layers before the last one for a more comprehensive theoretical analysis Fang et al. (2021). Indeed, this approach has been applied in Tirer & Bruna (2022), which namely extends the unconstrained features model by one layer as well as the Re LU nonlinearity. On the other hand, Zhu et al. (2021), Ji et al. (2021) and Zhou et al. (2022a) extend the unconstrained features model analysis by studying the landscape of the loss function therein and the related training dynamics. Beyond unconstrained features models, Ergen & Pilanci (2021) introduce a convex analytical framework to characterize the encoder layers for a more profound understanding of the NC phenomenon. Referring to the implications of NC on our understanding of neural networks, Hui et al. (2022) and Galanti et al. (2021) discuss the impact of NC on test data in the sense of generalization and transfer learning. Finally, Kothapalli et al. (2022) provides a multifaceted survey of recent works related to NC. 3 LAYER-PEELED MODEL WITH POSITIVE FEATURES As a prerequisite to the MD model, in this section we introduce a slightly modified version of the layer-peeled (or unconstrained features) model (see e.g. Zhu et al. (2021); Fang et al. (2021)), in which the features have to be positive. Accordingly, we will show that the global minimizers of the modified layer-peeled model correspond to an NC configuration, which differs from the global minimizers specified in other works and captures more closely the NC phenomenon in practice. For conciseness, we denote by H the matrix formed by the features h(k) n , n [N], k [K] as columns, and define W and H to be the Frobenius norm of the respective matrices, i.e. W 2 = PN n=1 wn 2 and H 2 = PK k=1 PN n=1 h(k) n 2 . We consider the regularized version of the model (instead of the norm constraint one as in e.g. Fang et al. (2021)) 1 min W ,H Lα(W , H) := Lα(W , H) + λW W 2 + λH s.t. H 0, (Pα) where λW , λH > 0 are the penalty parameters for the weight decays. By Lα we denote empirical risk with respect to the LS loss with parameter α [0, 1), where α = 0 corresponds to the conventional CE loss. More precisely, given a value of α, the LS technique then defines the label assigned to class n [N] as the following probability vector: y(α) n = (1 α)en + α n1N [0, 1]N, where en RN denotes the n-th standard basis vector and 1N RN denotes the vector consisting of only ones. Let p : RM RN be the function that assigns to each feature representation z RM 1Note that for simplicity we assume that the last layer does not have bias terms, i.e. b = 0. The result can be however easily extended to the more general case when the biases do not vanish. Namely, in presence of bias terms, the statement of Theorem 3.2 and also its proof remain unchanged. Published as a conference paper at ICLR 2023 the probability scores of the classes (as a probability vector in RN), p W (z) := softmax(W z) := " e wm,z PN i=1 e wi,z m=1 [0, 1]N. Then the LS loss corresponding to a sample in class n [N] is given by ℓα(W , z, y(α) n ) := D y(α) n , log p W (z) E := m=1 y(α) nm log p W (z)m (1) and the LS empirical risk Lα is defined as Lα(W , H) = 1 NK n=1 ℓα W , h(k) n , y(α) n . We will show that in common settings, the minimizers of (Pα ) correspond to neural collapse (NC) configurations, which we formalize in Def. 3.1 below. Definition 3.1 (NC configurations). Let K, M, N N, M N. A pair (W , H) of a weight matrix formed by rows wn RM and a feature matrix formed by columns h(k) n RM + (with n [N], k [K]) is said to be a NC configuration if (i) The feature representations h(k) n within every class n [N] are equal for all k [K], and thus equal to their class mean hn := 1 K PK k=1 h(k) n . (ii) The class means {hn}N n=1 have equal norms and form an (entry-wise) non-negative orthogonal system. (iii) Let Ph be the projection upon the subspace of RM orthogonal to h = 1 N PN n=1 hn. Then for every n [N], it holds wn = CPh hn for some constant C independent of n. Our main theorem in this section can be represented as follows. Theorem 3.2. Let M N, α [0, 1). Assume that N 1 (N 1)λW λH < 1. Then any global minimizer of the problem (Pα) is a NC configuration. Note that the NC configurations defined in Definition 3.1 above differ significantly from the ones specified in other works, e.g. Fang et al. (2021); Zhu et al. (2021); Zhou et al. (2022b) or Tirer & Bruna (2022), see Appendix B.1 for more discussion. 4 THE MEMORIZATION-DILATION MODEL 4.1 EXPERIMENTAL MOTIVATION Previous studies of the NC phenomenon mainly focus on the collapsing variability of training activations, and make rather cautious statements about its effects on generalization. For instance, Papyan et al. (2020) report slightly improved test accuracies for training beyond zero training error. Going a step further, Zhu et al. (2021) show that the NC phenomenon also happens for overparameterized models when labels are completely randomized. Here, the models seem to memorize by overfitting the data points, however, a rigorous study how label corruption affects generalization in the regime of NC is still lacking. To fill the gap, we advocate to analyze the effects of label corruption in the training data on the (previously unseen) test instead of the training feature collapse. Eventually, tight test class clusters go hand in hand with easier separation of the instances and, thus, a smaller generalization error. Following Zhu et al. (2021), we measure the collapse of the penultimate layer activations by the NC1 metric. This metric depicts the relative magnitude of the within-class covariance ΣW with respect to the between-class covariance ΣB of the penultimate layer features and is defined as N trace(ΣW Σ B), (2) Published as a conference paper at ICLR 2023 k=1 (h(k) n hn)(h(k) n hn) RM M , n=1 (hn h)(hn h) RM M, and Σ B denotes the pseudo-inverse of ΣB. Here, we adopt the notations from Section 1: h(k) n RM denotes the feature representation of k-th sample in class n, hn the class mean and h the global mean. Moreover, we distinguish NCtrain 1 and NCtest 1 to be calculated on the training and test instances, respectively. We call NCtest 1 dilation. Let us now turn to the notion of memorization, which is not uniquely defined in deep learning literature. Here, we define memorization in the context of the NC setting and in a global manner, different from other works, e.g. Feldman & Zhang (2020). Formally, suppose that label noise is incorporated by (independently) corrupting the instance of each class label n in the training data with probability η (0, 1), where corruption means drawing a label uniformly at random from the label space Y. We denote the set of corrupted instances by [ e K]. For a given dataset D (with label noise η), we define memorization as k [ e K] h(k) n h n 2 , (3) where h n denotes the mean of (unseen) test instances belonging to class n. We call the original ground truth label of a sample its true label. We call the label after corruption, which may be the true label or not, the observed label. Since instances of the same true label tend to have similar input features in some sense, the network is biased to map them to similar feature representations. Instances are corrupted randomly, and hence, instances of the same true label but different observed labels do not have predictable characteristics that allow the network to separate them in a way that can be generalized. When the network nevertheless succeeds in separating such instances, we say that the network memorized the feature representations of the corrupted instances in the training set. The metric mem in (3) thus measures memorization. The above memorization also affects dilation. Indeed, the network uses the feature engineering part to embed samples of similar features (that originally came from the same class), to far apart features, that encode different labels. Such process degrades the ability of the network to embed samples consistently, and leads to dilation. To quantify the interaction between mem and NCtest 1 , we analyzed the learned representations h in the penultimate layer feature space for different noise configurations. One may wonder whether one can see a systematic trend in the test collapse given the memorization, and how this evolves over different loss functions. To this end, we trained simple multi-layer neural networks for two classes (N = 2), which we subsampled from the image classification datasets MNIST Le Cun et al. (1998), Fashion MNIST Xiao et al. (2017), CIFAR-10 Krizhevsky & Hinton (2009) and SVHN Netzer et al. (2011). The labels are corrupted with noise degrees η [0.025, 0.4]. The network consists of 9 hidden layers with 2048 neurons each, thus, it represents a vastly overparameterized model. The feature dimension M is set to the number of classes N. We trained these networks using the CE and LS loss with a smoothing factor α = 0.1, as well as the mean-squared error (MSE). Moreover, we consider label relaxation (LR) Lienen & Hüllermeier (2021) as a generalization to LS with a relaxation degree α = 0.1. The networks were trained until convergence in 200 epochs (where the last 50 epochs did not make any significant changes) using SGD with an initial learning rate of 0.1 multiplied by 0.1 each 40 epochs and a small weight decay of 0.001. Moreover, we considered Re LU as activation function throughout the network, as well as batch normalization in each hidden layer. A linear softmax classifier is composed on the encoder. We conducted each experiment ten times with different seeds. The results for the above experimental setting are shown in Fig. 1, in which one can observe the trends of p NCtest 1 per memorization for various configurations. As can be seen, the figure shows an approximately linear correspondence between p NCtest 1 and mem for the CE derivatives (CE and LS) on all datasets when mem is not large. Published as a conference paper at ICLR 2023 0.0 0.5 1.0 1.5 2.0 CE LS LR MSE 0.0 0.5 1.0 1.5 2.0 fashionmnist CE LS LR MSE 0.0 0.5 1.0 1.5 CE LS LR MSE 0.0 0.5 1.0 1.5 2.0 CE LS LR MSE Noise Degree Test Collapse per Memorization (N = 2) n |2 (Memorization) 1 (Dilation) 0.0 0.5 1.0 1.5 2.0 CE LS LR MSE 0.0 0.5 1.0 1.5 2.0 fashionmnist CE LS LR MSE 0.0 0.5 1.0 1.5 CE LS LR MSE 0.0 0.5 1.0 1.5 2.0 CE LS LR MSE Noise Degree Test Accuracy per Memorization (N = 2) n |2 (Memorization) Test Accuracy Figure 1: Feature collapse of the test instances in terms of p NCtest 1 per memorization (top row) and the resulting test accuracies (bottom row) averaged over ten random seeds. Comparing the markers of the same color, it can be observed that LS consistently performs better than CE across all datasets, with very few exceptions (the very low noise degrees in cifar10). 0.0 0.5 1.0 1.5 2.0 CE LS LR MSE 0.5 1.0 1.5 2.0 2.5 CE LS LR MSE Noise Degree Dilation per Memorization (MNIST, N = 2) n |2 (Memorization) 1 (Dilation) 0.0 0.5 1.0 1.5 2.0 CE LS LR MSE 0.5 1.0 1.5 2.0 CE LS LR MSE Noise Degree Dilation per Memorization (CIFAR10, N = 2) n |2 (Memorization) 1 (Dilation) Figure 2: Feature collapse of the test instances in terms of p NCtest 1 per memorization as the feature dimension M varies. Moreover, as CE and LS share the same slope, these results suggest that the degradation of the test collapse (aka dilation) is a function of memorization and the network expressitivity, and not of the choice of the loss. The loss only affects how the noise translates to memorization, but not how memorization translates to dilation. Even though the same amount of noise is mapped to different memorization values in CE and LS, the memorization-dilation curve is nevertheless shared between CE and LS. Hence, since LS leads the network to memorize less, it results in improved performance (cf. Fig. 1). We can further see that MSE and LR show a different memorization-dilation correspondence, which means that these losses affect the inductive bias in a different way than CE and LS. We repeated the experiments for different values of the feature dimension M and show the example results in Fig. 2. Here, one can see the similar trends of dilation per memorization as before. In the appendix, we provide additional results showing the behavior in the multi-class case N > 2 with different models for label noise. The results support our MD model, and show that the memorizationdilation curve is roughly independent of the noise model for low-to-mid noise levels. 4.2 THE MEMORIZATION-DILATION MODEL Motivated by the observations of the previous experiments, we propose the so-called memorizationdilation (MD) model, which extends the unconstrained feature model by incorporating the interaction between memorization and dilation as a model assumption. By this, we explicitly capture the limited expressivity of the network, thereby modeling the inductive bias of the underlying model. Published as a conference paper at ICLR 2023 This model shall provide a basis to mathematically characterize the difference in the learning behavior of CE and LS. More specifically, we would like to know why LS shows improved generalization performance over conventional CE, as was observed in past works Müller et al. (2019). The main idea can be explained as follows. We first note that dilation is directly linked to generalization (see also Kornblith et al. (2021)), since the more concentrated the feature representations of each class are, the easier it is to separate the different classes with a linear classifier without having outliers crossing the decision boundary. The MD model asserts that dilation is a linear function of memorization. Hence, the only way that LS can lead to less dilation than CE, is if LS memorizes less than CE. Hence, the goal in our analysis is to show that, under the MD model, LS indeed leads to less memorization than CE. Note that this description is observed empirically in the experiments of Section 4.1. Next we define the MD model in the binary classification setting. Definition 4.1. We call the following minimization problem MD. Minimize the MD risk Rλ,η,α(U, r) := Fλ,α(W , H, r) + ηGλ,α(W , U, r), with respect to the noisy feature embedding U = [u1, u2] R2 M + and the standard deviation r 0, under the constraints η h1 u2 CMDr h1 h2 (4) η h2 u1 CMDr h1 h2 . (5) H R2 M + and W RM 2 form an NC configuration (see Definition 3.1). CMD > 0 is called the memorization-dilation slope, 0 α < 1 is called the LS parameter, η > 0 the noise level, and λ > 0 the regularization parameter. Fλ,α is the component in the (regularized) risk that is associated with the correctly labeled samples, Fλ,α(W , H, r) := Z ℓα W , h1 + v, y(α) 1 + λ h1 + v 2 ! ℓα W , h2 + v, y(α) 2 + λ h2 + v 2 ! where µ1 r and µ2 r are some probability distributions with mean 0 and standard deviation r, and lα is the LS loss defined in (1). Gλ,α is the component in the (regularized) risk that is associated with the corrupted samples, defined as Gλ,α(W , U, r) = ℓα W , u1, y(α) 1 + ℓα W , u2, y(α) 2 + λ u1 2 + λ u2 2 . The MD model can be interpreted as follows. First we consider the feature representations of the correctly labeled samples in each class as samples from a distribution (namely µ1,2 r in Def. 4.1) with standard deviation r, a parameter that measures the dilation of the class cluster. In a natural way, the corresponding risk Fλ,α involves the loss average over all samples, i.e. the loss integral over the distribution. For simplicity, we assume that the class centers h1, h2 as well as the weight matrix W are fixed as described by the NC configuration. This is a reasonable simplification as it has been always observed in the experiments. On the other hand, the feature representations of corrupted samples are u1 and u2.2 The amount of memorization in the first class is defined to be η h2 u1 , since the more noise η there is, the more 2Certainly one can, instead of two single points u1 and u2, two distributions centered around u1 and u2, similarly as before for uncorrupted samples. However, it is quite straightforward to see that the minimization of the MD risk over the dilation of these two distributions is independent of other variables (not like r), and thus the minimum should be attained in the case of collapsing into two single points. Thus, for convenience we assume directly here that Gλ,α involves only two single points. Published as a conference paper at ICLR 2023 examples we need to memorize. The amount of memorization in the second class is defined the same way. The (normalized) dilation is defined to be r h1 h2 , which models a similar quantity to (2). 0 1 2 3 Feature 1 Memorization Instances h(k) Figure 3: Exemplary illustration of the MD model for a MLP network trained on MNIST. The instances h(k) 2 are test images correctly labeled as 1, with centroid h2. The centroid of the test images with correct label 0 is h1. The centroid of training images which were originally labeled as 1 but are mislabeled as 0 is u1. The memorization of u1 moves it close to h1, and causes dilation of the instances h(k) 2 . The constraints (4) and (5) tell us that in order to map noisy samples u1 away from h2, we have to pay with dilation r. The larger r is, the further away we can map u1 from h2. The correspondence between memorization and dilation is linear with slope CMD by assumption. There are two main forces in the optimization problem: u1 would like to be as close as possible to its optimal position h1, and similarly u2 likes to be close to h2. In view of the constraints (4) and (5), to achieve this, r has to be increased to rmax := η h1 h2 2 CMD . On the other hand, the optimal r for the term Fλ,α is r = 0, namely, the layer-peeled NC configuration. An optimal solution hence balances between memorization and dilation. See Fig. 3 for a visualization of the MD model. Our goal in this section is to compare the optimal value r in case of LS and CE losses. We will distinguish between these two cases by setting the value of α in the MD model to 0 for CE and to some α0 > 0 for LS. This will result in two different scales of the feature embeddings H, denoted by HCE and HLS for CE and LS loss respectively, with the ratio γ := HCE / HLS > 1, (6) which holds under the reasonable assumption that the LS technique is sufficiently effective, or more precisely α0 > 2 λW λH. The main result in this section will be Theorem 4.3, which states informally that in the low noise regime, the optimal dilation in case of LS loss is smaller than that in case of CE loss. Before presenting this theorem, we will first establish several assumptions on the distributions µ1,2 r and the noise η in Assumption 4.2. Basically we allow a rich class of distributions and only require certain symmetry and bounded supports in terms of r, as well as require η to be small in terms of the ratio γ. Assumption 4.2. 1. Let α0 > 0. We assume that the solution of min W ,H ℓα W , h1, y(α) 1 + ℓα W , h2, y(α) 2 + λW W 2 + λH H 2 is given by (W , H) = (W CE, HCE) for α = 0 and (W , H) = (W LS, HLS) for α = α0. 2. Assume that the distributions µ1 r and µ2 r are centered, in the sense that Z w2 w1, v dµ1 r(v) = Z w1 w2, v dµ2 r(v) = 0 , Z h1, v dµ1 r(v) = Z h2, v dµ2 r(v) = 0 . Furthermore, we assume that there exists a constant A > 0 such that v Ar for any vector v that lies in the support of µ1 r or in the support of µ2 r. 3. Assume that the noise level η and the LS parameter α0 satisfy the following. We suppose α0 > 4 λW λH, which guarantees γ := HCE / HLS > 1. We moreover suppose that η is sufficiently small to guarantee η1/2 < C(1 1 γ ) where C := CMD 2 h CE 1 h CE 2 . Now our main result in this section can be formally stated as below. Published as a conference paper at ICLR 2023 Theorem 4.3. Suppose that Assumption 4.2 holds true for M N = 2 and λ := λH. Let r CE and r LS be the optimal dilations, i.e. the optimum r in the MD problem, corresponding to the CE and LS loss (accordingly α = 0 and α = α0), respectively. Then it holds that r CE h CE 1 h CE 2 > r LS h LS 1 h LS 2 . Theorem 4.3 reveals a mechanism by which LS achieves better generalization than CE. It is proven that LS memorizes and dilates less than CE, which is associated with better generalization. Note that in practice, the data often have noise in the sense that not all examples are perfectly labeled. More importantly, examples from different classes may share many similarities, a situation that is also covered by the MD model: the feature representations of samples from those classes are biased toward each other. In this case, LS also leads to decreased dilation which corresponds to better class separation and higher performance Kornblith et al. (2021). Interestingly, the concurrent work Zhou et al. (2022b) has shown that in the noiseless setting CE and LS lead to largely identical test accuracy, which seems to contradict the statement that LS performs better claimed by our work as well as many others, e.g. Kornblith et al. (2021); Müller et al. (2019). However, note that Zhou et al. (2022b) requires the network to be sufficiently large so that it has enough expressive power to fit the underlying mapping from input data to targets, as well as to be trained until convergence. While the latter is easy to obtain, it is difficult even to check if the first requirement holds. The difference between the two results is hence possibly caused by the effect of noise and by the network expressivity: while we aim to model the limited expressivity by the MD relation, Zhou et al. (2022b) focuses on networks with approximately infinite expressivity. The MD model combines a statistical term Fλ,α, that describes the risk over the distribution of feature embeddings of samples with clean labels, and an empirical term ηGλ,α that describes the risk over training samples with noisy labels. One point of view that can motivate such a hybrid statistical-empirical definition is the assumption that the network only memorizes samples of noisy labels, but not samples of clean labels. Such a memorization degrades (dilates) both the collapse of the training and test samples, possibly with different memorization-dilation slopes. However, memorization is not limited to corrupted labels, but can also apply to samples of clean labels Feldman & Zhang (2020), by which the learner can partially negate the dilation of the training features (but not test features). The fact that our model does not take the memorization of clean samples into account is one of its limitations. We believe that future work should focus on modeling the total memorization of all examples. Nevertheless, we believe that our current MD model has merit, since 1) noisy labels are memorized more than clean labels, and especially in the low noise regime the assumption of observing memorization merely for corrupted labels appears reasonable, and 2) our approach and proof techniques can be the basis of more elaborate future MD models. 5 CONCLUSION In this paper, we first characterized the global minimizers of the Layer-Peeled Model (or the Unconstrained Features Model) with the positivity condition on the feature representations. Our characterization shows some distinctions from the results that haven been obtained in recent works for the same model without feature positivity. Besides the conventional cross-entropy (CE) loss, we studied the model in case of the label smoothing (LS) loss, showing that NC also occurs when applying this technique. Then we extended the model to the so-called Memorization-Dilation (MD) Model by incorporating the limited expressivity of the network. Using the MD model, which is supported by our experimental observations, we show that when trained with the LS loss, the network memorizes less than when trained by the CE loss. This poses one explanation to the improved generalization performance of the LS technique over the conventional CE loss. Our model has limitations, however, namely that it is limited to the case of two classes. Motivated by promising results on the applicability of our model to the multi-class setting, we believe that future work should focus on extending the MD model in this respect. With such extensions, memorizationdilation analysis has the potential to underlie a systematic comparison of the generalization capabilities of different losses, such as CE, LS, and label relaxation, by analytically deriving formulas for the amount of memorization associated with each loss. Published as a conference paper at ICLR 2023 ACKNOWLEDGMENTS This work was partially supported by the German Research Foundation (DFG) within the Collaborative Research Center On-The-Fly Computing (CRC 901 project no. 160364472). Moreover, the authors gratefully acknowledge the funding of this project by computing time provided by the Paderborn Center for Parallel Computing (PC2). Devansh Arpit, Stanislaw Jastrzebski, Nicolas Ballas, David Krueger, Emmanuel Bengio, Maxinder S. Kanwal, Tegan Maharaj, Asja Fischer, Aaron C. Courville, Yoshua Bengio, and Simon Lacoste-Julien. A closer look at memorization in deep networks. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6-11 August 2017, volume 70 of Proceedings of Machine Learning Research, pp. 233 242. PMLR, 2017. Peter L. Bartlett, Philip M. Long, Gábor Lugosi, and Alexander Tsigler. Benign overfitting in linear regression. Co RR, abs/1906.11300, 2019. Weinan E and Stephan Wojtowytsch. On the emergence of simplex symmetry in the final and penultimate layers of neural network classifiers, 2021. Tolga Ergen and Mert Pilanci. Revealing the structure of deep neural networks via convex duality. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18-24 July 2021, Virtual Event, volume 139 of Proceedings of Machine Learning Research, pp. 3004 3014. PMLR, 2021. Cong Fang, Hangfeng He, Qi Long, and Weijie J. Su. Exploring deep neural networks via layer-peeled model: Minority collapse in imbalanced training. Proceedings of the National Academy of Sciences of the United States of America, 118, 2021. Vitaly Feldman and Chiyuan Zhang. What neural networks memorize and why: Discovering the long tail via influence estimation. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. Tomer Galanti, András György, and Marcus Hutter. On the role of neural collapse in transfer learning. Co RR, abs/2112.15121, 2021. X. Y. Han, Vardan Papyan, and David L. Donoho. Neural collapse under MSE loss: Proximity to and dynamics on the central path. Co RR, abs/2106.02073, 2021. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp. 770 778. IEEE Computer Society, 2016. Like Hui, Mikhail Belkin, and Preetum Nakkiran. Limitations of neural collapse for understanding generalization in deep learning. Co RR, abs/2202.08384, 2022. Arthur Jacot, Clément Hongler, and Franck Gabriel. Neural tangent kernel: Convergence and generalization in neural networks. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS 2018, December 3-8, 2018, Montréal, Canada, pp. 8580 8589, 2018. Wenlong Ji, Yiping Lu, Yiliang Zhang, Zhun Deng, and Weijie J. Su. An unconstrained layer-peeled perspective on neural collapse. Co RR, abs/2110.02796, 2021. Simon Kornblith, Ting Chen, Honglak Lee, and Mohammad Norouzi. Why do better loss functions lead to less transferable features? Co RR, abs/2010.16402, 2021. Vignesh Kothapalli, Ebrahim Rasromani, and Vasudev Awatramani. Neural collapse: A review on modelling principles and generalization. Co RR, abs/2206.04041, 2022. Alex Krizhevsky and Geoffrey Hinton. Learning multiple layers of features from tiny images. Technical report, University of Toronto, Toronto, Canada, 2009. Published as a conference paper at ICLR 2023 Daniel Kunin, Jonathan Bloom, Aleksandrina Goeva, and Cotton Seed. Loss landscapes of regularized linear autoencoders. In Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp. 3560 3569. PMLR, 09 15 Jun 2019. Thomas Laurent and James von Brecht. Deep linear networks with arbitrary loss: All local minima are global. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, July 10-15, 2018, volume 80 of Proceedings of Machine Learning Research, pp. 2908 2913. PMLR, 2018. Yann Le Cun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278 2324, 1998. Liam Li, Kevin G. Jamieson, Afshin Rostamizadeh, Ekaterina Gonina, Jonathan Ben-tzur, Moritz Hardt, Benjamin Recht, and Ameet Talwalkar. A system for massively parallel hyperparameter tuning. In Proceedings of Machine Learning and Systems 2020, MLSys 2020, Austin, TX, USA, March 2-4, 2020. mlsys.org, 2020. Zhu Li, Zhi-Hua Zhou, and Arthur Gretton. Towards an understanding of benign overfitting in neural networks. Co RR, abs/2106.03212, 2021. Julian Lienen and Eyke Hüllermeier. From label smoothing to label relaxation. In Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Virtual Event, February 2-9, 2021, pp. 8583 8591. AAAI Press, 2021. Jianfeng Lu and Stefan Steinerberger. Neural collapse with cross-entropy loss. Co RR, abs/2012.08465, 2020. Dustin G. Mixon, Hans Parshall, and Jianzong Pi. Neural collapse with unconstrained features. Co RR, abs/2011.11619, 2020. Rafael Müller, Simon Kornblith, and Geoffrey E. Hinton. When does label smoothing help? In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp. 4696 4705, 2019. Preetum Nakkiran, Gal Kaplun, Yamini Bansal, Tristan Yang, Boaz Barak, and Ilya Sutskever. Deep double descent: Where bigger models and more data hurt. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020. Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y. Ng. Reading digits in natural images with unsupervised feature learning. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems, NIPS, Granada, Spain, November 12-17, Workshop on Deep Learning and Unsupervised Feature Learning, 2011. Vardan Papyan, X. Y. Han, and David L. Donoho. Prevalence of neural collapse during the terminal phase of deep learning training. Co RR, abs/2008.08186, 2020. Tomaso A. Poggio and Qianli Liao. Generalization in deep network classifiers trained with the square loss. 2020. Tomaso A. Poggio and Qianli Liao. Explicit regularization and implicit bias in deep network classifiers trained with the square loss. Co RR, abs/2101.00072, 2021. Daniel A. Roberts, Sho Yaida, and Boris Hanin. The principles of deep learning theory. Co RR, abs/2106.10165, 2021. Amartya Sanyal, Puneet K. Dokania, Varun Kanade, and Philip H. S. Torr. How benign is benign overfitting? In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021. Open Review.net, 2021. Published as a conference paper at ICLR 2023 Andrew M. Saxe, James L. Mc Clelland, and Surya Ganguli. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. In 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014. Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. In Yoshua Bengio and Yann Le Cun (eds.), 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. Ruoyu Sun. Optimization for deep learning: An overview. Journal of the Operations Research Society of China, 8:249 294, 2020. Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jonathon Shlens, and Zbigniew Wojna. Rethinking the inception architecture for computer vision. Co RR, abs/1512.00567, 2015. Tom Tirer and Joan Bruna. Extended unconstrained features model for exploring deep neural collapse. Co RR, abs/2202.08087, 2022. Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: A novel image dataset for benchmarking machine learning algorithms. Co RR, abs/1708.07747, 2017. Chulhee Yun, Suvrit Sra, and Ali Jadbabaie. Global optimality conditions for deep neural networks. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. Open Review.net, 2018. Jinxin Zhou, Xiao Li, Tianyu Ding, Chong You, Qing Qu, and Zhihui Zhu. On the optimization landscape of neural collapse under mse loss: Global optimality with unconstrained features. Co RR, abs/2203.01238, 2022a. Jinxin Zhou, Chong You, Xiao Li, Kangning Liu, Sheng Liu, Qing Qu, and Zhihui Zhu. Are all losses created equal: A neural collapse perspective. Co RR, abs/2210.02192, 2022b. Zhihui Zhu, Daniel Soudry, Yonina C. Eldar, and Michael B. Wakin. The global optimization geometry of shallow linear neural networks. J. Math. Imaging Vis., 62(3):279 292, 2020. Zhihui Zhu, Tianyu Ding, Jinxin Zhou, Xiao Li, Chong You, Jeremias Sulam, and Qing Qu. A geometric analysis of neural collapse with unconstrained features. Co RR, abs/2105.02375, 2021.