# stable_anisotropic_regularization__2bcb0428.pdf Published as a conference paper at ICLR 2024 STABLE ANISOTROPIC REGULARIZATION William Rudman Department of Computer Science Brown University william_rudman@brown.edu Carsten Eickhoff School of Medicine University of Tübingen carsten.eickhoff@uni-tuebingen.de Given the success of Large Language Models (LLMs), there has been considerable interest in studying the properties of model activations. The literature overwhelmingly agrees that LLM representations are dominated by a few outlier dimensions with exceedingly high variance and magnitude. Several studies in Natural Language Processing (NLP) have sought to mitigate the impact of such outlier dimensions and force LLMs to be isotropic (i.e., have uniform variance across all dimensions in embedding space). Isotropy is thought to be a desirable property for LLMs that improves model performance and more closely aligns textual representations with human intuition. However, many claims regarding isotropy in NLP have been based on the average cosine similarity of embeddings, which has recently been shown to be a flawed measure of isotropy. In this paper, we propose I-STAR: Iso Score -based STable Anisotropic Regularization, a novel regularization method that can increase or decrease levels of isotropy in embedding space during training. I-STAR uses Iso Score , the first accurate measure of isotropy that is both differentiable and stable on mini-batch computations. In contrast to several previous works, we find that decreasing isotropy in contextualized embeddings improves performance on most tasks and models considered in this paper. 1 1 INTRODUCTION Several previous works have investigated the role of isotropy in Large Language Model (LLM) representations (Rudman et al., 2022). A distribution is isotropic if the variance of the data is uniform and the data dimensions are uncorrelated. In practice, a distribution is isotropic when its covariance matrix is proportional to the identity matrix. Studies have found that representations from LLMs, such as BERT or GPT-2, lack the property of isotropy and that contextualized word embeddings are dominated by a few rogue or outlier dimensions (Timkey and van Schijndel, 2021; Kovaleva et al., 2021). Several previous works have argued that anisotropy, i.e., the lack of isotropy, is detrimental to LLM embeddings as it 1) forces representations to occupy a narrow cone in space (Ethayarajh, 2019; Cai et al., 2021); 2) obscures linguistic information, thereby limiting the expressive power of the embeddings (Gao et al., 2019; Zhang et al., 2020; Mickus et al., 2019), and; 3) hinders performance on a variety of downstream tasks (Kovaleva et al., 2021; Bi s et al., 2021; Timkey and van Schijndel, 2021). However, some recent works have challenged previously held conceptions about isotropy, arguing that current methods of measuring isotropy are fundamentally flawed (Rudman et al., 2022; Rajaee and Pilehvar, 2021a). To address these concerns, Rudman et al. (2022) propose Iso Score, an accurate and robust method for measuring isotropy based on the covariance matrix of a distribution. Although Iso Score is an effective method for measuring isotropy, we demonstrate that Iso Score is neither differentiable nor stable when the number of points in a given sample is small. Therefore, Iso Score cannot serve as an effective model regularizer. Given the recent criticism of methods for measuring isotropy, a reassessment of previously accepted theories of isotropy in LLMs is needed. This paper aims to determine the relationship between isotropy and model performance on various language models and fine-tuning tasks. We first propose Iso Score , a method for measuring isotropy that is 1) fully differentiable, 2) incorporates classical techniques for covariance estimation to create stable isotropy estimates for mini-batch data, and 3) approximates Iso Score for large sample sizes. We then use Iso Score to develop I-STAR: Iso Score -based STable 1Code: https://github.com/bcbi-edu/p_eickhoff_isoscore.git Published as a conference paper at ICLR 2024 Figure 1: Forward pass of our I-STAR loss function. Let xl be the token embeddings in a mini-batch at layer l {1, 2, ..., n}, let X = Sn l=1 xl, let ΣSi be the shrinkage covariance matrix for epoch i and let ζ (0, 1) be the shrinkage parameter. I-STAR loss is a weighted sum between cross-entropy loss, LCE, and Iso Score ( X, ζ, ΣSi) where λ is the tuning-parameter. Negative values of λ correspond to decreasing isotropy in representations, and positive values of λ encourage isotropy. Anisotropic Regularization. I-STAR is a flexible way to adjust isotropy in model representations during training or fine-tuning. In contrast to works that use flawed measures of isotropy, we find that using I-STAR to decrease isotropy in embedding space, i.e., making representations more anisotropic, tends to improve downstream performance across three different LLMs and nine different fine-tuning tasks. Our finding that anisotropic representations perform better on downstream tasks is aligned with literature outside of NLP that argues anisotropy is a natural by-product of stochastic gradient descent, where anisotropy helps models escape local minima in the loss landscape and, thus, generalize better to unseen data (Zhu et al., 2018). Additionally, our findings are supported by a well-established body of literature arguing that lower intrinsic dimensionality of network representations in later model layers corresponds to better performance on downstream tasks (Ansuini et al., 2019; Recanatesi et al., 2019; Chung et al., 2018). This paper makes the following novel contributions. 1. We propose Iso Score , a robust method for measuring isotropy that is stable even when the number of samples in a point cloud is small. 2. We present a novel regularization method, I-STAR: Iso Score -based, STable Anisotropic Regularization. I-STAR effectively shapes the geometry of network activations in a stable manner that overcomes the current limitations of other methods that backpropagate through the calculation of principal components during stochastic gradient descent. 3. In contrast to existing theories of NLP, we demonstrate that decreasing isotropy in LLMs tends to improve performance on various fine-tuning tasks and models. 2 RELATED WORK Improving Isotropy. Methods to restore isotropy in contextualized embedding models fall into two categories: post-processing methods and regularizers. Mu et al. (2017) propose All-But-The-Top, a post-processing algorithm that masks out several of the top principal components of the data. The authors show that their simple algorithm improves performance for Word2Vec and Glo Ve embeddings on word similarity tasks. Several slight variations have occurred on the All-But-The-Top algorithm where the top principal components of the last hidden state of LLMs are masked or removed (Rajaee and Pilehvar, 2021b; Bihani and Rayz, 2021; Liang et al., 2021; Liao et al., 2020; Sajjad et al., 2022). Across each of these studies, the authors argue that improving isotropy in the embedding space improves model performance. However, studies evaluating the impact of improving isotropy in embedding space by masking principal components tend to be limited to word similarity tasks, which do not provide a complete picture of the importance of isotropy in model representations. Published as a conference paper at ICLR 2024 Zhou et al. (2020) propose Isotropic Batch Normalization, a modified whitening transform that forces representations to be zero-mean but allows the covariance matrix of model representations to be block diagonal and does not entirely remove all correlations from the data. The authors apply their novel transformation to the final hidden state representations of a BERT model before being input to the classification head and show that Isotopic Batch Normalization minimally improves the performance of BERT on several datasets in the GLUE benchmark. Several authors argue that isotropy can be restored in contextualized embedding space by applying a simple zero-mean transformation to the data (Bi s et al., 2021; Cai et al., 2021). Given that isotropy is a property of the covariance matrix of the data and is unrelated to the mean, the improvements on the textual similarity tasks shown in various studies are likely unrelated to the property of isotropy. There have been far fewer attempts in the literature to improve isotropy using regularization penalties. Gao et al. (2019) propose Cos Reg, a regularization technique that penalizes the model when the average cosine similarity of model representation approaches 1. The motivation behind Cos Reg is that by reducing the average cosine similarity between embeddings, models will be penalized when representations occupy a narrow cone in vector space (Gao et al., 2019; Zhang et al., 2020). Although the authors report modest gains when using Cos Reg, more current studies have argued that average random cosine similarity does not accurately measure isotropy (Rudman et al., 2022). Although a large number of papers in NLP argue that isotropy is beneficial for representations, the broader machine learning community has found that 1) anisotropy is a natural consequence of stochastic gradient descent; 2) anisotropy allows for networks to generalize better to unseen examples, and; 3) networks that compress data into lower dimensional manifolds show better performance on downstream tasks (Zhu et al., 2018; Ansuini et al., 2019; Recanatesi et al., 2019). We argue that the primary reason for the differences between claims on isotropy in NLP literature and machine learning literature stems from the noisy range of often flawed methods of measuring isotropy on which many claims are based (Rudman et al., 2022). In Section 2.1, we discuss the most common methods used to measure isotropy in embedding space and detail why most attempts to measure isotropy in the NLP literature do not accurately reflect properties of isotropy. 2.1 MEASURING ISOTROPY Average random cosine similarity is the most common method for measuring isotropy in embedding space. An average random cosine similarity approaching 1 is thought to represent a minimally isotropic space, while an average random cosine similarity of 0 constitutes a maximally isotropic space (Ethayarajh, 2019). Ethayarajh (2019) finds that the average random cosine similarity between activations of BERT and GPT-2 approach 1, which the authors use to argue that model representations form a narrow cone in vector space. However, Rudman et al. (2022) show that average random cosine similarity is not a measure of isotropy since the average cosine similarity of points artificially approaches 1 when the mean of the data is far from the origin and will be 0 when the data are zero-mean, regardless of the shape of the distribution. The partition isotropy score is based on a partition function, Z(C) := P x X exp(c T x), developed by Arora et al. (2015), where c Rd represents the set of unit vectors and X Rd is a finite point cloud. Since calculating the entire partition function is intractable, studies typically approximate the full partition function as I(X) minc CZ(c)/maxc CZ(c), where c is chosen from the eigenspectrum of XXT (Mu et al., 2017). Mu et al. (2017) prove that there exists a choice of C such that their method for measuring isotropy reflects the uniformity of principal components of the data. However, approximating c from the eigenspectrum of XXT results in undesirable properties, such as being heavily influenced by the mean vector of X and influenced by orthogonal transformations of the data that are unrelated to isotropy (Rudman et al., 2022). Intuitively, Iso Score measures how far away the covariance matrix of the data is from α Id, where α is a positive scalar and Id is the d d identity matrix. Algorithm 2 details the steps required to calculate Iso Score.Rudman et al. (2022) develop a rigorous testing suite to demonstrate that Iso Score is the first tool in the literature that accurately measures isotropy in embedding space. To ensure that Iso Score is not biased towards a given basis of the data, the authors reorient the data by projecting a point cloud of data by its principal components. An Iso Score value of 1 indicates that a distribution is fully isotropic, and a score near 0 suggests that a single dimension in vector space dominates representations. Although Iso Score is an accurate measure of isotropy, we demonstrate in Section 3 Published as a conference paper at ICLR 2024 Algorithm 1 Iso Score Forward Pass 1: Input: X Rd point cloud, ΣS Rd d shrinkage covariance matrix, ζ (0, 1). 2: Outputs: I-STAR penalty of X. 3: calculate covariance matrix: ΣX of X 4: calculate shrinkage matrix: Σζ := (1 ζ) ΣX + ζ ΣS 5: calculate eigenvalues: Λ := {λ1, .., λd} of Σζ 6: normalize eigenvalues: ˆΛ := d Λ/||Λ||2 such that ||ˆΛ|| = d 7: calculate the isotropy defect: δ(ˆΛ) := ||ˆΛ 1||/ q where 1 = (1, ..., 1) Rd 8: calculate: ϕ(ˆΛ) := (d δ(ˆΛ)2(d 9: calculate: ι(ˆΛ) := (d ϕ(ˆΛ) 1)/(d 1). that Iso Score will systematically underestimate the true isotropy score of a distribution when the number of samples is small relative to the dimensionality of the vector space. Since many current works in NLP use flawed measures of isotropy, such as average random cosine similarity and the partition isotropy score, the connection between isotropy in LLMs and their performance on downstream tasks has not been established. This study devises a novel regularization penalty, I-STAR, to investigate the relationship between model performance and isotropy in model representations. In contrast to several previous works based on average cosine similarity or the partition score, we use Iso Score to demonstrate that decreasing isotropy tends to improve performance on downstream tasks, while increasing isotropy hampers performance on nearly all tasks and models considered in this paper. Previous methods for measuring isotropy in embedding space are either 1) not accurate measures of isotropy, 2) not differentiable, or 3) not stable on mini-batch computations. In this section, we propose Iso Score , a novel, fully differentiable measure of isotropy that is stable, even for small sample sizes. First, we thoroughly describe Iso Score . We then demonstrate that Iso Score can accurately measure the isotropy of small data samples, while vanilla Iso Score systematically underestimates the isotropy of a distribution when the number of points in a finite point cloud is smaller than the dimensionality of the vector space. For the remainder of this section, let X Rd and S Rd be finite point clouds drawn from a distribution X such that |X| < d << |S|. Intuitively, Iso Score measures the extent to which the principal components of a distribution are uniformly distributed. Measuring isotropy as a function of the principal components allows us to backpropagate through Iso Score since PCA is a differentiable function (Huang et al., 2018). An Iso Score value of 1 implies that for principal components {λ1, .., λd}, λi = λj i, j. An Iso Score value of 0 implies that only a single principal component is non-zero. Although both Iso Score and the partition score measure isotropy via the principal components of a distribution, Iso Score directly measures the uniformity of the eigenspectrum of principal components and does not have to rely on approximations to the eigenspectrum like the partition function defined by Mu et al. (2017). Iso Score Pseudocode. Step 3) of Algorithm 1 begins by calculating the covariance matrix of X Rd that we assume is sampled from the distribution X. Next, in Step 4) we calculate the RDAShrinkage matrix, Σζ, as a weighted sum between the covariance matrix of X and the covariance matrix of S, ΣS, to produce a more accurate estimate of the covariance matrix of the true distribution X. The shrinkage parameter, ζ, controls how much covariance information is used from X and S when estimating the covariance matrix of X. In Step 5), we calculate the eigenvalues of Σζ. Step 6) normalizes the eigenvalues so that the L2 norm of the eigenvalues equals the norm of the vector containing all 1s (i.e., 1, 1, ..., 1). The remaining normalizing Steps 7-9 are derived in the same manner as vanilla Iso Score. For a detailed pseudocode analysis of both Iso Score and Iso Score , as well as a proof that Iso Score approximates Iso Score when the number of samples in our point cloud is large, see Section A in the appendix. Published as a conference paper at ICLR 2024 Mini-batch stability of isotropy estimates. The mini-batch stability of methods measuring isotropy has yet to be investigated. We test the stability of mini-batch isotropy by sub-sampling small batches of points, X, from a point cloud of data, X R768, consisting of 250,000 points sampled from a 768-dimensional Gaussian distribution with a zero mean-vector and a covariance matrix, Σ X, such that Σ X has diag(Σ X) = {10, 6, 4, 4, 1, ..., 1} and zero off-diagonal elements. In Section A of the Appendix, we prove that when ζ = 0 (i.e. no shrinkage is performed), Iso Score(X) = Iso Score (X, ζ, ΣS). Figure 2 demonstrates that for a sub-sample X of X, if |X| is not sufficiently larger than d, Iso Score systematically underestimates the true degree of isotropy (dashed horizontal line) of the distribution, X, from which X is sampled. This means Iso Score(X) << Iso Score( X). For isotropy results to be reliable, future work should ensure that the number of points in a given sample is significantly larger than the dimensionality of the distribution. Iso Score underestimates the true degree of isotropy of the distribution because Iso Score relies on calculating the covariance matrix of a sample. When the number of points in a sample, |X|, is less than the dimensionality of the space, the covariance matrix of X may be singular (Friedman, 1989). Existing methods to improve isotropy and some of the most common metrics to evaluate isotropy, such as the partition isotropy score (Mu et al., 2017), rely on investigating the principal components of the data. As a consequence, the problem of underestimating the isotropy of sample distributions will affect nearly all previous works. Fortunately, many well-established methods in the statistics and machine learning literature exist to ensure that a covariance matrix will be better conditioned and invertible, leading to more reliable methods to alter and measure isotropy in embedding space (Friedman, 1989). Figure 2: Iso Score (X, ζ, ΣS) values for different choices of ζ. The dashed line indicates the correct Iso Score value of X, which is Iso Score ( X) = 0.86. We calculate ΣS from a subsample S X such that X S = and |S| = 75, 000. Stabilizing covariance estimation. Shrinkage is a simple operation that adds a known, stable covariance matrix to a non-invertible, singular sample covariance matrix (Friedman, 1989). Performing shrinkage ensures that the resulting covariance matrix is invertible. In situations where one does not have access to multiple samples or a sample where S X such that d << |S|, the matrix ζ Id + ΣX is used as the shrinkage matrix (Friedman, 1989). However, if one has access to the larger point cloud S or multiple samples from X, then a more faithful estimate of the covariance matrix of X can be obtained using regularized discriminant analysis (RDA) (Friedman, 1989). RDA shrinkage pools covariance matrices together using Σζ := ζ ΣX + (1 ζ) ΣS to get a more accurate measure of the covariance matrix of the distribution from which X is sampled. Figure 2 demonstrates that performing this shrinkage operation on ΣX drastically improves the stability of Iso Score even when |X| = 700 and d = 768. Step 4 of Algorithm 1 uses RDA shrinkage to stabilize Iso Score on mini-batches during training. In stochastic gradient descent, mini-batches of data are sampled randomly from the larger training set. We perform shrinkage on the mini-batch covariance matrix with a covariance matrix calculated from token embeddings obtained from a small fraction of the training data to obtain the most accurate estimate of mini-batch isotropy. We initialize the shrinkage matrix, ΣS, by computing the covariance matrix of a sample, S, of 250k points obtained from running a partial forward pass on the training data before training with I-STAR. We update ΣS after each epoch during training by running a partial forward pass on the training data. We use Iso Score as the penalty in our loss function to produce stable updates during training that alter the isotropy of model representations. Figure 1 illustrates how we incorporate Iso Score to form our I-STAR loss. We take the union of token embeddings from each hidden state in the model and calculate a global Iso Score penalty to incorporate into our loss function. We stress that shrinkage is crucial for the success of I-STAR. In Section D, we show that without shrinkage, the performance of models tuned with I-STAR can drop by as much as 6%. Published as a conference paper at ICLR 2024 Cos Reg. To compare Cos Reg to I-STAR, we adapt the cosine similarity regularization term presented by (Gao et al., 2019) and calculate our Cos Reg penalty on the last-layer hidden states of the model. Let {x1, x2, ..., x M} denote the mini-batch representation obtained from the last hidden layer, Xn, of a contextualized embedding model, let ˆxi = xi ||xi|| and let λ R be a tuning parameter, then the Cos Reg loss function of our model is defined as: LCos Reg = LCE + λ 1 j =i ˆx T i ˆxj (1) We use λ = 1 as Gao et al. (2019) find that using λ = 1 is sufficient for altering average random cosine similarity and that using λ > 1 does not provide any additional training benefits. Since an average cosine similarity of 1 is thought to reflect an anisotropic space and an average cosine similarity near 0 reflects an isotropic space, using λ = 1 is believed to encourage isotropy as measured by cosine similarity (Ethayarajh, 2019). However, we show in Figure 4 that Cos Reg impacts the mean of embeddings but has no impact on isotropy. Namely, Cos Reg does not properly regularize isotropy. I-STAR. Figure 1 outlines the calculation of I-STAR loss. I-STAR computes a global Iso Score penalty from token embeddings obtained from all layers in the network. Calculating a global isotropy penalty from the representations at every layer of the network allows the model to determine where changing the isotropy of representations in the network will lead to the largest improvements in performance. In Section G in the Appendix, we examine the impact of applying I-STAR to individual layers and find that calculating a global isotropy penalty leads to the most consistent performance. Let X = Sn l=1 Xl denote the union of all hidden states from a network with n layers and ΣSi be the shrinkage covariance matrix for epoch i of training. We define our I-STAR loss as follows: LI-STAR = LCE + λ (1 Iso Score ( X, ζ, ΣSi)) (2) A negative value of λ will decrease the levels of isotropy in model representations, while positive choices of λ will increase levels of isotropy. Intrinsic dimensionality estimation. Previous studies have shown that compressing the intrinsic dimension of LLM activations in later layers is correlated with improved performance on downstream tasks (Recanatesi et al., 2019). To determine the relationship between isotropy and intrinsic dimensionality, we use Two NN to estimate the intrinsic dimension of activations similarly to (Recanatesi et al., 2019). Two NN is a widely used intrinsic dimensionality (ID) estimator based on the ratio between each point s first and second nearest neighbors (Facco et al., 2017). Note that calculating Two NN is an iterative process based on the pairwise distances of nearest neighbors in space and is a non-differentiable operation. 4.1 EXPERIMENTAL DESIGN Models & datasets. In this paper, we fine-tune BAppendixepbert, ALBERT (Lan et al., 2020), and Distil BERT (Sanh et al., 2020) for nine NLP common benchmark tasks: SST-2 (Socher et al., 2013), QNLI (Rajpurkar et al., 2016a), RTE (Dagan et al., 2005), MRPC (Dolan and Brockett, 2005), QQP (Wang et al., 2018), COLA (Warstadt et al., 2019), STS-B (Cer et al., 2017), SST-5 Socher et al. (2013) and SQUAD Rajpurkar et al. (2016b). A detailed description of each dataset is available in Section F Hyperparameters & training details. For each model and each task, we hyperparameter tune for batch size (8, 16, 32), training epochs (3,4,5), and learning rate (1e-5, 3e-5, 5e-5). For ISTAR, we tune for the optimal zeta (0.2, 0.4, 0.6, 0.8) and use the tuning parameters values, λ {-5, -3, -1, 1, 3, 5}. For Cos Reg, we use a tuning parameter of 1 in accordance with Gao et al. (2019). All reported performance metrics are calculated as an average over five random seeds to demonstrate the robustness of our results. After we perform our hyperparameter tuning, we fine-tune our models using two 3090-RTX GPUs, use mixed-point precision training for all models/tasks, and set a gradient accumulation step to 2. Published as a conference paper at ICLR 2024 In this section, we demonstrate that 1) there is an inverse relationship between isotropy and performance; 2) Cos Reg implements a zero mean transform and does not improve isotropy, and 3) increasing/decreasing isotropy using I-STAR increases/decreases the Two NN intrinsic dimensionality estimation of model representations. Table 1: Performance of Cos Reg and I-Star for each model and task. Base indicates that no regularization methods were used. For COLA, we report Matthew s Correlation; for STS-B, we report Pearson s Correlation; for SQUAD, we present F1/EM. For all remaining tasks, we report accuracy. We report the average/standard deviation over 5 random seeds. Each I-STAR value comes from training with a negative tuning parameter. See Table 2 for more details. Method SST-2 QNLI RTE MRPC QQP COLA STS-B SST-5 SQUAD ALBERT I-STAR 93.08 0.32 91.41 0.43 72.56 1.29 87.84 0.43 89.97 0.06 55.22 1.61 88.91 0.23 55.02 0.46 90.33 0.09/82.96 0.04 ALBERT Base 93.08 0.24 90.28 0.42 71.34 0.81 86.96 0.32 89.98 0.06 54.56 0.93 89.07 0.20 54.23 0.30 90.32 0.13/82.86 0.17 ALBERT Cos Reg 91.86 0.42 90.91 0.25 65.99 1.81 86.91 0.26 89.81 0.14 49.59 0.90 87.98 0.46 54.76 0.11 90.08 0.31/82.64 0.32 BERT I-STAR 92.65 0.18 89.51 0.69 62.53 0.70 86.76 0.57 90.36 0.05 59.60 0.25 86.44 0.16 50.64 0.34 87.76 0.11/80.02 0.10 BERT Base 92.40 0.41 89.49 0.30 62.24 0.48 86.51 0.21 90.44 0.11 59.22 0.63 86.15 0.20 50.00 0.28 87.02 0.11/78.82 0.25 BERT Cos Reg 92.01 0.23 89.67 0.56 61.52 1.41 85.29 0.47 90.45 0.82 57.90 0.82 86.30 0.28 49.44 0.18 86.89 0.14/78.85 0.08 Distil BERT I-STAR 91.42 0.11 86.25 0.32 58.05 0.95 84.17 0.35 89.67 0.08 50.03 0.93 85.28 0.11 50.13 0.23 83.69 0.36/74.80 0.27 Distil BERT Base 91.36 0.15 87.40 0.34 56.82 0.67 83.68 0.41 89.57 0.74 50.16 0.59 84.75 0.26 49.48 0.28 83.23 0.18/74.04 0.12 Distil BERT Cos Reg 90.94 0.38 86.68 0.19 57.04 1.13 83.58 0.44 89.55 0.08 49.30 0.88 84.41 0.08 48.65 0.25 83.47 0.34/74.40 0.24 There is an inverse relationship between isotropy and performance. In contrast to previous studies, Figure 3 demonstrates that increasing isotropy tends to decrease performance, whereas decreasing isotropy tends to increase performance. We observe that the trend s strength is somewhat task and model-dependent, with the most robust correlation emerging for SQUAD, RTE, and STS-B. However, for both MRPC and COLA, ALBERT does not exhibit a strong correlation between isotropy and performance. However, for both bases, the optimal tuning parameter is negative. The lack of a distinct trend for ALBERT on MRPC and COLA is likely due to the large amounts of noise in fine-tuning performance and isotropy values occurring on these tasks. Further, Table 1 demonstrates that decreasing isotropy using negative tuning parameters in I-STAR improves performance over both baseline models and Cos Reg on most tasks and models considered in this paper. Cos Reg implements a zero-mean transform. Figure 4 demonstrates that Cos Reg alters the mean of model activations, particularly in a single dimension. Encouraging the average random cosine similarity of activations to be 0 (i.e., when λ = 1) forces representation to be zero-mean, whereas encouraging an average random cosine similarity to be 1 (i.e., when λ = 1) causes the mean of representations to increase further. Although Cos Reg impacts the mean of the data, Cos Reg does not increase isotropy in model representations. After fine-tuning BERT, ALBERT, and Distil BERT on SST-2 using Cos Reg with a tuning-parameter value of λ = 1, the last layer representations of each model receive Iso Score* values of 0.004, 0.007, and 0.007, respectively. Isotropy and intrinsic dimensionality estimation. Several studies have found that a lower intrinsic dimension of later layer representations is correlated with improved performance on various downstream tasks (Ansuini et al., 2019; Recanatesi et al., 2019). Figure 5 demonstrates that adjusting isotropy with I-STAR corresponds to changing the intrinsic dimension of model representations. Importantly, encouraging isotropy in model representations does not allow for model representations to compress into a lower dimensional manifold in later layers. 6 DISCUSSION Our study challenges a dominant belief in the NLP literature that encouraging isotropy improves performance on downstream tasks. In contrast to several previous works, we find that encouraging isotropy is detrimental to model performance and that decreasing isotropy in representations improves performance on a broad range of tasks and models. Table 1 and Figure 3 provide strong empirical evidence, in support of (Zhu et al., 2018), that anisotropy is essential for a model s downstream performance. The primary reason for the discrepancy between our results and existing studies in the NLP literature on isotropy is that previous studies have made claims using flawed measures of isotropy, such as average random cosine similarity. Figure 4 shows that using Cos Reg implements a zero-mean transform and does not improve isotropy. Given our findings that isotropy and classification performance Published as a conference paper at ICLR 2024 Figure 3: Relationship between Iso Score* (x-axis) and model performance (y-axis). We fine-tune each model with I-STAR using the tuning parameters λ {-5, -3, -1, 0.50, 1, 3, 5}. We train each model over five random seeds and report the standard deviation of both performance and Iso Score (X, ζ, ΣS) values. We set ζ = 0.2 for all computations of Iso Score , and we compute ΣS from a random sample of 250,000 token embeddings from the training data. Figure 4: Comparing the mean activation values on the validation data for each dimension of ALBERT, BERT, and Distil BERT fine-tuned on QNLI, with Cos Reg using a tuning-parameter value of λ = 1, 1 and without any regularization. Trends are representative of all tasks. are negatively correlated and that Cos Reg does not adjust isotropy, we argue many of the current claims regarding isotropy in NLP need to be reassessed. Although a majority of prior studies in NLP have argued that isotropy is beneficial to LLMs, recent works have helped to explain our finding that isotropy negatively correlates with classification performance. (Mickus et al., 2024) provide a robust mathematical framework demonstrating that Published as a conference paper at ICLR 2024 Figure 5: Two NN Intrinsic Dimensionality estimate of ALBERT, BERT, and Distil BERT sentence embeddings, i.e. [CLS] tokens, obtained from the SST-2 validation data for models fine-tuned on the SST-2 using I-STAR with tuning-parameters λ { 5, 3, 3, 5}. Base represents the case where no regularization is used. Trends are representative of all tasks. isotropy and clustering are incompatible objectives and that clustering behavior is crucial for an effective classifier. Namely, encouraging isotropy inhibits clustering behavior, which is harmful to downstream performance. Our findings strongly support arguments in the literature outside of NLP that anisotropy is a natural outcome of stochastic gradient descent and that compressing representations is necessary for model performance. Additionally, studies have shown that model representations that occupy a lower intrinsic dimension in the ambient vector space tend to outperform those sampled from higher dimensional manifolds (Recanatesi et al., 2019; Ansuini et al., 2019). We demonstrate that encouraging isotropy in the embedding space increases the intrinsic dimension of model representations, which is detrimental to performance. Importantly, we also show that reducing isotropy in the embedding space leads to the compression of representations into a lower dimensional manifold, resulting in improved model performance. This underscores the critical role of isotropy in determining the intrinsic dimension of model representations and the subsequent impact on model performance. Limitations. Although having isotropic representations is theoretically desirable for both the interpretability of model decisions and for improved quantization abilities, encouraging isotropy in pre-trained models in a way that preserves or improves downstream task performance is challenging. This study is limited to fine-tuning LLMs, which may not provide a definitive answer to whether encouraging isotropy in embedding space is inherently detrimental to model performance. Our results demonstrate that encouraging isotropy in pre-trained models causes a decline in downstream fine-tuning performance. Fine-tuning requires models to make rapid and drastic adjustments to their representations within a limited number of training steps. A fruitful direction for future work would consist of using I-STAR in LLM pre-training to enforce isotropic representations throughout training. 7 CONCLUSION Previous works in NLP have argued that anisotropy in contextualized embedding models limits the expressiveness of word representations and forces embeddings to occupy a narrow cone in vector space. Several studies have claimed that improving isotropy leads to improved performance on downstream tasks. However, most studies use faulty isotropy measures, tend to be limited to word similarity tasks, and only investigate isotropy for last-layer representations. We propose I-STAR, a differentiable, mini-batch-stable isotropy-based regularization scheme, to study the relationship between fine-tuned model performance and isotropy. Contrary to previous works in NLP, we find that further decreasing isotropy improves downstream model performance. Fundamentally, we show that enhancing isotropy in embedding space increases the intrinsic dimensionality of model representations and causes model performance to decrease. Given the connection between isotropy, intrinsic dimensionality, and performance, I-STAR shows great promise for application in various areas of deep learning. Published as a conference paper at ICLR 2024 8 REPRODUCIBILITY We have taken several steps to make our paper as reproducible as possible. Firstly, we have made all code used to produce the project publicly available and attached an anonymous version along with our submission. Further, we have released a pip install of Iso Score to facilitate future works. Section G outlines all values of the two critical hyperparameters needed to train with I-STAR loss. Lastly, all datasets and models used in this paper are publicly available on Huggingface. Alessio Ansuini, Alessandro Laio, Jakob H. Macke, and Davide Zoccolan. Intrinsic Dimension of Data Representations in Deep Neural Networks. Curran Associates Inc., Red Hook, NY, USA, 2019. Sanjeev Arora, Yuanzhi Li, Yingyu Liang, Tengyu Ma, and Andrej Risteski. Random walks on context spaces: Towards an explanation of the mysteries of semantic word embeddings. Co RR, abs/1502.03520, 2015. URL http://arxiv.org/abs/1502.03520. Geetanjali Bihani and Julia Rayz. Low anisotropy sense retrofitting (LASe R) : Towards isotropic and sense enriched representations. In Proceedings of Deep Learning Inside Out (Dee LIO): The 2nd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, pages 81 95, Online, June 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.deelio-1.9. URL https://aclanthology.org/2021.deelio-1.9. Daniel Bi s, Maksim Podkorytov, and Xiuwen Liu. Too much in common: Shifting of embeddings in transformer language models and its implications. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5117 5130, Online, June 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.naacl-main.403. URL https://aclanthology.org/2021. naacl-main.403. Xingyu Cai, Jiaji Huang, Yuchen Bian, and Kenneth Church. Isotropy in the contextual embedding space: Clusters and manifolds. In International Conference on Learning Representations, 2021. URL https://openreview.net/forum?id=x YGNO86OWDH. Daniel Cer, Mona Diab, Eneko Agirre, Iñigo Lopez-Gazpio, and Lucia Specia. Sem Eval-2017 task 1: Semantic textual similarity multilingual and crosslingual focused evaluation. In Proceedings of the 11th International Workshop on Semantic Evaluation (Sem Eval-2017), pages 1 14, Vancouver, Canada, August 2017. Association for Computational Linguistics. doi: 10.18653/v1/S17-2001. URL https://aclanthology.org/S17-2001. Sue Yeon Chung, Daniel D. Lee, and Haim Sompolinsky. Classification and geometry of general perceptual manifolds. Physical Review X, 8(3), jul 2018. doi: 10.1103/physrevx.8.031003. URL https://doi.org/10.1103%2Fphysrevx.8.031003. Ido Dagan, Oren Glickman, and Bernardo Magnini. The pascal recognising textual entailment challenge. In Proceedings of the First International Conference on Machine Learning Challenges: Evaluating Predictive Uncertainty Visual Object Classification, and Recognizing Textual Entailment, MLCW 05, page 177 190, Berlin, Heidelberg, 2005. Springer-Verlag. ISBN 3540334270. doi: 10.1007/11736790_9. URL https://doi.org/10.1007/11736790_9. Chris Ding, Ding Zhou, Xiaofeng He, and Hongyuan Zha. R1-pca: Rotational invariant l1-norm principal component analysis for robust subspace factorization. In Proceedings of the 23rd International Conference on Machine Learning, ICML 06, page 281 288, New York, NY, USA, 2006. Association for Computing Machinery. ISBN 1595933832. doi: 10.1145/1143844.1143880. URL https://doi.org/10.1145/1143844.1143880. William B. Dolan and Chris Brockett. Automatically constructing a corpus of sentential paraphrases. In Proceedings of the Third International Workshop on Paraphrasing (IWP2005), 2005. URL https://aclanthology.org/I05-5002. Published as a conference paper at ICLR 2024 Kawin Ethayarajh. How contextual are contextualized word representations? comparing the geometry of bert, elmo, and GPT-2 embeddings. Co RR, abs/1909.00512, 2019. URL http://arxiv. org/abs/1909.00512. Elena Facco, Maria d Errico, Alex Rodriguez, and Alessandro Laio. Estimating the intrinsic dimension of datasets by a minimal neighborhood information. Scientific Reports, 7 (1), sep 2017. doi: 10.1038/s41598-017-11873-y. URL https://doi.org/10.1038% 2Fs41598-017-11873-y. Jerome H. Friedman. Regularized discriminant analysis. Journal of the American Statistical Association, 84:165 175, 1989. Jun Gao, Di He, Xu Tan, Tao Qin, Liwei Wang, and Tie-Yan Liu. Representation degeneration problem in training natural language generation models. Co RR, abs/1907.12009, 2019. URL http://arxiv.org/abs/1907.12009. Lei Huang, Dawei Yang, Bo Lang, and Jia Deng. Decorrelated batch normalization. Co RR, abs/1804.08450, 2018. URL http://arxiv.org/abs/1804.08450. Olga Kovaleva, Saurabh Kulshreshtha, Anna Rogers, and Anna Rumshisky. BERT busters: Outlier dimensions that disrupt transformers. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 3392 3405, Online, August 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.findings-acl.300. URL https://aclanthology.org/ 2021.findings-acl.300. Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. Albert: A lite bert for self-supervised learning of language representations, 2020. Yuxin Liang, Rui Cao, Jie Zheng, Jie Ren, and Ling Gao. Learning to remove: Towards isotropic pre-trained BERT embedding. Co RR, abs/2104.05274, 2021. URL https://arxiv.org/ abs/2104.05274. Siyu Liao, Jie Chen, Yanzhi Wang, Qinru Qiu, and Bo Yuan. Embedding compression with isotropic iterative quantization. Co RR, abs/2001.05314, 2020. URL https://arxiv.org/abs/2001. 05314. Timothee Mickus, Denis Paperno, Mathieu Constant, and Kees van Deemter. What do you mean, bert? assessing BERT as a distributional semantics model. Co RR, abs/1911.05758, 2019. URL http://arxiv.org/abs/1911.05758. Timothee Mickus, Stig-Arne Grönroos, and Joseph Attieh. Isotropy, clusters, and classifiers, 2024. URL https://arxiv.org/abs/2402.03191. Jiaqi Mu, Suma Bhat, and Pramod Viswanath. All-but-the-top: Simple and effective postprocessing for word representations. Co RR, abs/1702.01417, 2017. URL http://arxiv.org/abs/ 1702.01417. Sara Rajaee and Mohammad Taher Pilehvar. How does fine-tuning affect the geometry of embedding space: A case study on isotropy. Co RR, abs/2109.04740, 2021a. URL https://arxiv.org/ abs/2109.04740. Sara Rajaee and Mohammad Taher Pilehvar. A cluster-based approach for improving isotropy in contextual embedding space. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 575 584, Online, August 2021b. Association for Computational Linguistics. doi: 10.18653/v1/2021.acl-short.73. URL https: //aclanthology.org/2021.acl-short.73. Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. SQu AD: 100,000+ questions for machine comprehension of text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pages 2383 2392, Austin, Texas, November 2016a. Association for Computational Linguistics. doi: 10.18653/v1/D16-1264. URL https://aclanthology. org/D16-1264. Published as a conference paper at ICLR 2024 Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. Squad: 100, 000+ questions for machine comprehension of text. Co RR, abs/1606.05250, 2016b. URL http://arxiv.org/ abs/1606.05250. Stefano Recanatesi, Matthew Farrell, Madhu Advani, Timothy Moore, Guillaume Lajoie, and Eric Shea-Brown. Dimensionality compression and expansion in deep neural networks. Co RR, abs/1906.00443, 2019. URL http://arxiv.org/abs/1906.00443. William Rudman, Nate Gillman, Taylor Rayne, and Carsten Eickhoff. Iso Score: Measuring the uniformity of embedding space utilization. In Findings of the Association for Computational Linguistics: ACL 2022, pages 3325 3339, Dublin, Ireland, may 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.findings-acl.262. URL https://aclanthology.org/ 2022.findings-acl.262. Hassan Sajjad, Firoj Alam, Fahim Dalvi, and Nadir Durrani. Effect of post-processing on contextualized word representations. In Proceedings of the 29th International Conference on Computational Linguistics, pages 3127 3142, Gyeongju, Republic of Korea, October 2022. International Committee on Computational Linguistics. URL https://aclanthology.org/2022.coling-1. 277. Victor Sanh, Lysandre Debut, Julien Chaumond, and Thomas Wolf. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter, 2020. Richard Socher, Alex Perelygin, Jean Wu, Jason Chuang, Christopher D. Manning, Andrew Ng, and Christopher Potts. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pages 1631 1642, Seattle, Washington, USA, October 2013. Association for Computational Linguistics. URL https://aclanthology.org/D13-1170. William Timkey and Marten van Schijndel. All bark and no bite: Rogue dimensions in transformer language models obscure representational quality. Co RR, abs/2109.04404, 2021. URL https: //arxiv.org/abs/2109.04404. Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel Bowman. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of the 2018 EMNLP Workshop Blackbox NLP: Analyzing and Interpreting Neural Networks for NLP, pages 353 355, Brussels, Belgium, November 2018. Association for Computational Linguistics. doi: 10.18653/v1/W18-5446. URL https://aclanthology.org/W18-5446. Alex Warstadt, Amanpreet Singh, and Samuel R. Bowman. Neural network acceptability judgments. Transactions of the Association for Computational Linguistics, 7:625 641, 2019. doi: 10.1162/ tacl_a_00290. URL https://aclanthology.org/Q19-1040. Zhong Zhang, Chongming Gao, Cong Xu, Rui Miao, Qinli Yang, and Junming Shao. Revisiting representation degeneration problem in language modeling. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 518 527, Online, November 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.findings-emnlp.46. URL https:// aclanthology.org/2020.findings-emnlp.46. Wenxuan Zhou, Bill Yuchen Lin, and Xiang Ren. Isobn: Fine-tuning BERT with isotropic batch normalization. Co RR, abs/2005.02178, 2020. URL https://arxiv.org/abs/2005.02178. Zhanxing Zhu, Jingfeng Wu, Bing Yu, Lei Wu, and Jinwen Ma. The anisotropic noise in stochastic gradient descent: Its behavior of escaping from sharp minima and regularization effects, 2018. URL https://arxiv.org/abs/1803.00195. Published as a conference paper at ICLR 2024 Algorithm 2 Iso Score 1: Input: Let X Rd be a finite collection of points. 2: Let XPCA denote the points in X transformed by the first n principal components. 3: Define ΣD Rn as the diagonal of the covariance matrix of XPCA. 4: Normalize diagonal to ˆΣD := n ΣD/ ΣD , where is the standard Euclidean norm. 5: The isotropy defect is δ(X) := ˆΣD 1 / p 2(n n) where 1 = (1, . . . , 1) Rn. 6: X uniformly occupies ϕ(X) := (n δ(X)2(n n))2/n2 percent of ambient dimensions. 7: Transform ϕ(X) so it can take values in [0, 1], via ι(X) := (n ϕ(X) 1)/(n 1). 8: return: ι(X) A ISOSCORE VS ISOSCORE Iso Score(X) = Iso Score (X, ζ, CX) when ζ = 0. Let X Rn be a finite point cloud that we assume is sampled from some larger distribution X such that the number of points in X is sufficiently larger than the number of points in X. Let ζ (0, 1) be a shrinkage parameter, and let ΣS Rn n be a shrinkage covariance matrix obtained from a sample of points, S, drawn from X such that |S| >> n. We will first demonstrate that without RDA-Shrinkage (i.e. when ζ = 0), Iso Score(X) = Iso Score (X, ζ, ΣS). The key insight is that ΣD in Step 3 of Algorithm 1 is equivalent to the principal components of X. Algorithm 2 shows that the first step of Iso Score is to transform X to its principal components to get what the authors denote as XPCA. Let ΣPCA be the covariance matrix of XP CA, and let ΣX denote the covariance matrix of X. Projecting X to its principal components removes correlations from the data, meaning that ΣPCA will be diagonal. Since ΣPCA is a diagonal matrix, its eigenvalues are equal to diag(ΣPCA). Therefore, diag(ΣPCA) are the principal components of XP CA since principal components are the eigenvalues of the covariance matrix. Note that principal components are invariant under orthogonal transformations and that reorienting the data by PCA applies the orthogonal transformation V T XV , where V are the eigenvectors of ΣX. Namely, the principal components of X are the principal components of XP CA. For a simple proof demonstrating that the principal components of X are invariant under any orthogonal transformation applied to X, see (Ding et al., 2006). Therefore, Iso Score is equivalent to Iso Score when no covariance matrix shrinkage is performed. That is, when ζ = 0, Iso Score(X) = Iso Score (X, ζ, ΣS) X RN. To see that Iso Score(X) approaches Iso Score (X, ζ, ΣS), we can use the Law of Large Numbers to show that the larger the sample of X, the more close X approximates the true distribution ˆ X. Therefore, Iso Score(X) Iso Score (X, ζ, ΣS) as |X| increases. Comparing Pseudocode. Iso Score addresses two fundamental flaws in vanilla Iso Score. Firstly, Iso Score uses RDA-shrinkage (Friedman, 1989) to stabilize the calculation of a sample covariance matrix. Secondly, Iso Score removed all non-differentiable operations present in Iso Score. The pseudocode for Iso Score and Iso Score has two primary steps: 1) extract distribution/isotropy information from the point cloud, S and 2) normalize the isotropy information into a score in the closed interval [0, 1]. In Algorithm 3, Steps 3-4 calculate the covariance matrix of the point cloud S, and perform RDA-shrinkage by taking the weighted sum of the covariance matrix of S and a covariance matrix of a larger distribution from which S was sampled. Step 5 then calculates the eigenvalues from the resulting sum of covariance matrices. When we set ζ in Step 4 to 0, the resulting eigenvalues are the principal components of our sample S. All normalizing steps are identical in Iso Score and Iso Score . Namely, lines 6-9 in Algorithm 1 are equivalent to lines 4-7 in Algorithm 2. Non-Differentiability of Iso Score. We want to highlight the exact step in Algorithm 2 that makes Iso Score non-differentiable. Step 3 in Algorithm 2 involves selecting the diagonal of a covariance matrix, which is a non-differentiable operation. Note that the non-differentiability of Iso Score does not imply that Iso Score is non-differentiable, even though Iso Score and Iso Score are equivalent when no shrinkage is performed. To further emphasize this point, consider the two functions f(x) = x x Published as a conference paper at ICLR 2024 and g(x) = x for all x! = 0 and g(x) = |x| when x = 0. Here, |x| indicates the absolute value function. For all inputs x, f(x) = g(x). However, f(x) is differentiable for all values of x, where g(x) is not differentiable when x = 0. Section 3 demonstrates that Iso Score is equivalent to Iso Score when no shrinkage is performed. Iso Score preserves all of the desirable theoretical properties of Iso Score while improving stability on small inputs and removing the operation that makes Iso Score non-differentiable (namely, an iterative selection of the diagonal). B WHY DO MODELS PREFER ANISOTROPY? C LAYER-WISE ISOTROPY Figure 6: Layer-wise Iso Score values for ALBERT, BERT, and Distil BERT fine-tuned with I-STAR using tuning parameters λ { 1, 1}. Iso Score values are calculated on the QNLI validation data using a shrinkage parameter of ζ = 0.2. None indicates that no regularizer is used in fine-tuning. When fine-tuning models with I-STAR, we compute the Iso Score penalty from the union of all token embeddings from each model layer. This section analyzes what layers in the model are impacted the most by our I-STAR regularizer. Figure 6 shows that encouraging isotropy in token embeddings using I-STAR primarily impacts early layers in the network. Even when isotropy is encouraged using positive tuning parameter values in I-STAR, token representations from the later layers of the network remain highly anisotropy. These results provide further evidence that anisotropy in the form of outlier dimensions that emerge in the last layers of the network is crucial to the model decision-making process. An interesting direction for future work could be to explore applying I-STAR to various layers in the network. D IMPACT OF THE SHRINKAGE PARAMETER ON I-STAR In this section, we evaluate the impact of changing the shrinkage parameter, ζ, on downstream performance when fine-tuning with I-STAR regularization. To test the effect of varying ζ, we fix all hyperparameters found in Table 2 and fine-tune with ζ {0, 0.2, 0.4, 0.6, 0.8, 1}. We train ALBERT, BERT, and Distil BERT on RTE. All results are reported as an average over 5 random seeds. Figure 7: Performance of ALBERT, BERT, and Distil BERT on RTE with changing values of ζ. The red dashed line denotes the optimal value of ζ. Setting ζ = 0 indicates that no shrinkage is performed (i.e., equivalent to regularizing using Iso Score), and setting ζ = 1 signifies that no mini-batch covariance information is included in the gradient updates during a backward pass. Published as a conference paper at ICLR 2024 In Section 3, we demonstrated that Iso Score systematically underestimates the actual value of point clouds when the number of samples is lower than the dimensionality of the vector space. Iso Score overcomes this fundamental limitation by using shrinkage to improve the stability of the covariance matrix calculation. Figure 7 demonstrates the importance of using Iso Score when computing isotropy scores of a mini-batch. When ζ = 0 and no shrinkage is performed, the performance of our fine-tuned model decreases by 3.85%, 6.19%, and 0.76% compared to optimal values of ζ for ALBERT, BERT, and Distil BERT, respectively. In addition to testing the utility of using covariance matrix shrinkage, test the impact of excluding mini-batch covariance information by setting ζ = 1. When ζ = 1, Iso Score is calculated from the stabilizing covariance matrix, ΣSi, obtained by calculating the covariance matrix from a sample of 250,000 points at epoch i. Figure 7 demonstrates the utility of using mini-batch covariance matrix information during fine-tuning as the optimal tuning parameter is always a value of ζ (0, 1). E APPLYING I-STAR TO DIFFERENT LAYERS Throughout this paper, we calculate a global isotropy penalty on the vector space of all token embeddings from all layers in the model. Our motivation for selecting a global isotropy penalty instead of penalizing individual layers is two-fold. Firstly, Iso Score is a more faithful representation of the true isotropy of the global vector space when the number of samples is large, and the sample covariance is more likely to be full-rank, meaning that isotropy-based gradient updates will be more stable. Secondly, selecting individual layers to penalize would drastically increase the hyperparameter search when using I-STAR. Lastly, we ultimately decided on a global isotropy over regularizing individual layers as a global isotropy penalty resulted in the most consistent results across different tasks. In this section, we fine-tune ALBERT, BERT, and Distil BERT on COLA using 5 random seeds. We fix all hyperparameters in Table 2 except for the layer we select to apply I-STAR. Figure 8: Impact of applying I-STAR on different layers in ALBERT, BERT, and Distil BERT finetuned on COLA. The dashed red line is the result of applying a global I-STAR penalty on all layers of the model. Figure 8 reports the performance of applying I-STAR to individual layers in the network. Although regularizing individual layers in ALBERT and BERT can improve performance compared to a global isotropy penalty, a global isotropy penalty provides much more consistent performance across models and tasks. Namely, no consistent patterns emerge when applying I-STAR to individual layers. However, more work is needed to determine if the occasional improvement in performance is worth the extensive hyperparameter search induced by layer-wise I-STAR. F DATASET DETAILS Stanford Sentiment Treebank with 2 classes (SST-2) is a binary classification task where models must determine whether a short movie review is positive or negative in sentiment (Socher et al., 2013). SST-5 is a five-class version of SST-2 where the models must determine whether a movie review is negative, somewhat negative, neutral, or positive. QNLI is a binary natural language inference task where models must decide whether or not a given answer is entailed from a specified question (Wang et al., 2018). Stanford Question Answering Dataset V1 (SQUAD)is an extractive questionanswering task where a model must select the span of text in a passage that answers a given question Published as a conference paper at ICLR 2024 (Rajpurkar et al., 2016b). Recognizing Textual Entailment (RTE) is a binary classification task where a model must determine if a given sentence logically follows a preceding sentence. STS-B (Semantic Textual Similarity Benchmark) is a collection of sentence pairs annotated with a similarity score from 1-5. STS-B is commonly evaluated with Pearson s correlation coefficient. The Microsoft Research Paraphrase Corpus (MRPC) tasks models with determining if a pair of sentence are paraphrases of each other (i.e. semantically equivalent). Quora Question Pairs (QQP) consist of question pairs from Quora. Models must determine if the sentence pairs are semantically equivalent. Corpus of Linguistic Acceptability (COLA) task models to determine if a given string is a linguistically acceptable sentence. SST-2, QNLI, RTE, MRPC, STS-B, QQP, and MRPC are all datasets in the GLUE benchmark (Wang et al., 2018). G I-STAR HYPERPARAMETERS Table 2: Optimal I-STAR hyperparameter values of the tuning parameter, λ and shrinkage parameter, ζ. We searched over λ { 5, 3, 1, 1, 3, 5} and ζ {0.2, 0.4, 0.6, 0.8}. We present results as λ | ζ. Note that all values of λ are negative, meaning I-STAR will decrease isotropy in embedding space. Method SST-2 QNLI RTE MRPC QQP COLA STS-B SST-5 SQUAD Avg. ALBERT I-STAR -1 | 0.2 -1 | 0.2 -1 | 0.8 -1 | 0.4 -1 | 0.2 -1 | 0.8 -1 | 0.4 -1 | 0.2 -5 | 0.2 -1.44 | 0.36 BERT I-STAR -5 | 0.2 -1 | 0.4 -1 | 0.8 -1 | 0.6 -1 | 0.2 -1 | 0.8 -1 | 0.2 -5 | 0.2 -3 | 0.6 -2.11 | 0.44 Distil BERT I-STAR -5 | 0.6 -1 | 0.2 -1 | 0.6 -1 | 0.8 -1 | 0.2 -3 | 0.4 -3 | 0.4 -1 | 0.2 -5 | 0.2 -2.33 | 0.40 In this Section, we outline the optimal hyperparameters used in I-STAR for each task and each model. I-STAR requires two key hyperparameters: λ, the tuning parameter, and ζ, the shrinkage parameter. Recall that the tuning parameter controls both the strength of the signal of the Iso Score in the loss function and whether isotropy is encouraged or discouraged in model representations. Namely, when λ > 0, I-STAR will increase isotropy in the embedding space, and when λ < 0, I-STAR will decrease isotropy in the embedding space. The shrinkage parameter, ζ, determines how much covariance information comes from a sample point cloud, X, and how much covariance information comes from C0. For all tasks and models, we limit our hyperparameter search to λ { 5, 3, 1, 1, 3, 5} and ζ {0.2, 0.4, 0.6, 0.8}. Note that all optimal tuning parameter values occur when λ < 0, meaning further decreasing isotropy leads to better performance gains than encouraging isotropy.