# decomposing_the_dark_matter_of_sparse_autoencoders__e13ac7b1.pdf Published in Transactions on Machine Learning Research (04/2025) Decomposing The Dark Matter of Sparse Autoencoders Joshua Engels jengels@mit.edu MIT Logan Smith logansmith5@gmail.com Independent Max Tegmark tegmark@mit.edu MIT & IAIFI Reviewed on Open Review: https: // openreview. net/ forum? id= s Xq3Wb3vef Sparse autoencoders (SAEs) are a promising technique for decomposing language model activations into interpretable linear features. However, current SAEs fall short of completely explaining model performance, resulting in dark matter : unexplained variance in activations. This work investigates dark matter as an object of study in its own right. Surprisingly, we find that much of SAE dark matter about half of the error vector itself and > 90% of its norm can be linearly predicted from the initial activation vector. Additionally, we find that the scaling behavior of SAE error norms at a per token level is remarkably predictable: larger SAEs mostly struggle to reconstruct the same contexts as smaller SAEs. We build on the linear representation hypothesis to propose models of activations that might lead to these observations. These insights imply that the part of the SAE error vector that cannot be linearly predicted ( nonlinear error) might be fundamentally different from the linearly predictable component. To validate this hypothesis, we empirically analyze nonlinear SAE error and show that 1) it contains fewer not yet learned features, 2) SAEs trained on it are quantitatively worse, and 3) it is responsible for a proportional amount of the downstream increase in cross entropy loss when SAE activations are inserted into the model. Finally, we examine two methods to reduce nonlinear SAE error: inference time gradient pursuit, which leads to a very slight decrease in nonlinear error, and linear transformations from earlier layer SAE outputs, which leads to a larger reduction. 1 Introduction The ultimate goal for ambitious mechanistic interpretability is to understand neural networks from the bottom up by breaking them down into programs ( circuits") and the variables ( features ) that those programs operate on (Olah, 2023). One recent successful technique for finding features in language models has been sparse autoencoders (SAEs), which learn a dictionary of one-dimensional representations that can be sparsely combined to reconstruct model hidden activations (Cunningham et al., 2023; Bricken et al., 2023). However, as observed by Gao et al. (2024), the scaling behavior of SAE width (number of latents) vs. reconstruction mean squared error (MSE) is best fit by a power law with a constant error term. Gao et al. (2024) speculate that this component of SAE error below the asymptote might best be explained by model activations having components with denser structure than simple SAE features (e.g. Gaussian noise). This is a concern for the ambitious agenda because it implies that there are components of model hidden states that are harder for SAEs to learn and which might not be eliminated by simple scaling of SAEs. Motivated by this discovery, in this work our goal is to specifically study the SAE error vector itself, and in doing so gain insight into the failures of current SAEs, the dynamics of SAE scaling, and possible distributions of model activations. Thus, our direction differs from the bulk of prior work that seeks to quantify SAE Published in Transactions on Machine Learning Research (04/2025) 105 106 107 0.00 Absent Features Linear Error Nonlinear Error SAE Reconstruction, FVU 0.186 + 2.708W 0.317 SAE Reconstruction + Error Prediction, FVU 0.122 SAE Pursuit Reconstruction + Error Prediction Encoder Error Fraction Variance Unexplained (FVU) Figure 1: A breakdown of SAE dark matter for layer 20 Gemma 9B SAEs, with dotted lines assuming that observed trends continue for larger SAEs. See Section 4 for how we break down the overall fraction of unexplained variance into absent features, linear error, and nonlinear error. See Section 6.1 for further separating encoder error from nonlinear error. failures, as these mostly focus on downstream benchmarks or simple cross entropy loss (see e.g. Gao et al. (2024); Templeton et al. (2024); Anders & Bloom (2024)). The structure of this paper is as follows: 1. In Section 4, we introduce the fundamental mystery that we will explore throughout the rest of the paper: SAE errors are shockingly predictable. To the best of our knowledge, we are the first to show that a large fraction of SAE error vectors can be explained with a linear transformation of the input activation, that the norm of SAE error vectors can be accurately predicted by a linear projection of the input activation, and that on a per-token level, error norms of large SAEs are linearly predictable from small SAEs. 2. In Section 5, we investigate the linearly predictable and nonlinearly predictable components of SAE error. We find that although the nonlinear component affects downstream cross entropy loss in proportion to its norm, it is qualitatively different from linear error: as compared to linear error, nonlinear error is harder to learn SAEs for and has a norm that is harder to linearly predict from activations, suggesting that it consists of a smaller proportion of not-yet-learned linear features. 3. In Section 6, we investigate methods for reducing nonlinear error. We show that inference time optimization increases the fraction of variance explained by SAEs, but only slightly decreases nonlinear error. Additionally, we show that we can use SAEs trained on previous components to decrease nonlinear error and total SAE error. 2 Related Work Language Model Representation Structure: The linear representation hypothesis (LRH) (Park et al., 2023; Elhage et al., 2022) claims that language model hidden states can be decomposed into a sparse sum of linear feature directions. The LRH has seen recent empirical support with sparse autoencoders (Makhzani & Frey, 2013; Bricken et al., 2023; Cunningham et al., 2023), which have succeeded in decomposing much of the variance of language model hidden states into such a sparse sum, as well as a long line of work that has used probing and dimensionality reduction to find causal linear representations for specific concepts (Alain, 2016; Nanda et al., 2023; Marks et al., 2024; Gurnee, 2024). On the other hand, some recent work has questioned whether the linear representation hypothesis is true: Engels et al. (2024) find multidimensional Published in Transactions on Machine Learning Research (04/2025) circular representations in Mistral (Jiang et al., 2023) and Llama (AI@Meta, 2024), and Csordás et al. (2024) examine synthetic recurrent neural networks and find onion-like non-linear features not contained in a linear subspace. This has inspired recent discussion about what a true model of activation space might be: Mendel (2024) argues that the linear representation hypothesis ignores the growing body of results showing the multi-dimensional structure of SAE latents, and Smith (2024b) argues that we only have evidence for a weak form of the superposition hypothesis holding that only some features are linearly represented (such a hypothesis has also been studied in the sparse coding literature (Tasissa et al., 2020)). SAE Errors and Benchmarking: Multiple works have introduced techniques to benchmark SAEs and characterize their error: Bricken et al. (2023), Gao et al. (2024), and Templeton et al. (2024) use manual human analysis of features, automated interpretability, downstream cross entropy loss when SAE reconstructions are inserted back into the model, and feature geometry visualizations; Karvonen et al. (2024) use the setting of board games, where the ground truth features are known, to determine what proportion of the true features SAEs learn; and Anders & Bloom (2024) use the performance of the model on NLP benchmarks when the SAE reconstruction is inserted back into the model. More specifically relevant to our main direction in this paper, Gurnee (2024) finds that SAE reconstruction errors are pathological, that is, when SAE reconstructions are inserted into the model, they have a larger effect on cross entropy loss than random perturbations with the same error norm. Follow up work by Heimersheim & Mendel (2024) and Lee & Heimersheim (2024) find that this effect disappears when the random baseline is replaced by a perturbation in the direction of the difference between two random activations. SAE Scaling Laws: Anthropic (2024), Templeton et al. (2024), and Gao et al. (2024) study how SAE MSE scales with respect to FLOPS, sparsity, and SAE width, and define scaling laws with respect to these quantities. Templeton et al. (2024) also study how specific groups of language features like chemical elements, cities, animals, and foods are learned by SAEs, and show that SAEs predictably learn these features in terms of their occurrence. Finally, Bussmann et al. (2024) find that larger SAEs learn two new types of dictionary vectors as comparsed to smaller SAEs: features not present at all in smaller SAEs, and more fine-grained feature split versions of features in smaller SAEs. We consider neural network activations x Rd and sparse autoencoders Sae Rd Rd. Sparse autoencoders map inputs to a latent space Rm with m d and then back into Rd, while also requiring that only a sparse set of the m latent dimensions are nonzero. SAEs have the general architecture hidden(x) := σ(Wenc (x bdec) + benc) (1) Sae(x) := Wdec hidden(x) + bdec (2) where σ is an architecture dependent activation function, and seek to minimize LSae(x) = x Sae(x) 2 2 + S(hidden(x)) (3) where S is an architecture dependent sparsity function. For convenience, we define Sae Error(x) as the reconstruction error of the SAE and L0 as the average number of nonzeros in hidden(x): Sae Error(x) := x Sae(x) (4) L0 := Ex( hidden(x) 0) (5) In this work, we study the relationship between x and Sae Error(x) and are agnostic to the specifics of SAE architecture. Thus, we examine both Top K SAEs (Gao et al., 2024) and Jump Relu SAEs (Rajamanoharan et al., 2024). The Top K SAE is defined as σT op K(z) = set all but the k largest dimensions of z to zero (6) ST op K(hidden(x)) = 0 (7) Published in Transactions on Machine Learning Research (04/2025) Note that L0 = k for a Top K SAE. The Jump Relu SAE is defined as σJump Relu(z)i = ( zi if zi θi > 0 0 otherwise (8) SJump Relu(hidden(x)) = λ hidden(x) 0 (9) where θ is a learned bias vector and λ is a sparsity parameter. 4 Predicting SAE Error In this section, we evaluate the extent to which SAE errors Sae Error(x) can be predicted from input model activations x. We run experiments 1 on Gemma 2 2B and 9B (Team et al., 2024) and Llama 3.1 8B (AI@Meta, 2024). We use the suite of Gemma Scope (Lieberum et al., 2024) sparse autoencoders for Gemma 2 2B experiments and Llama Scope (He et al., 2024) for Llama 3.1 8B. For experiments where we analyze the effect of L0 and SAE width, we focus on layer 12 of Gemma 2 2B and layer 20 of Gemma 9 9B in the main text as these layers have the most SAEs in Gemma Scope, and we include additional layers in the appendix (we do not run on Llama Scope for these experiments, as Llama Scope has just one SAE L0 and two SAE widths per layer). For experiments where we analyze the effect of layer, we show the set of SAEs with L0 closest to a target L0 across all layers for both Gemma Scope and Llama Scope suites of SAEs. We use 300 contexts of 1024 tokens from the uncopywrited subset of the Pile (Gao et al., 2020) and then filter to only activations of tokens after position 200 in each context, as Lieberum et al. (2024) find that earlier tokens are easier for sparse autoencoders to reconstruct, and we wish to ignore the effect of token position on our results. This results in a dataset of about 247k activations. For linear regressions, we use a random subset of size 150k as training examples (since all models have a dimension of less than 5000, this prevents overfitting) and report the R2 on the other 97k activations. For linear transformations to a multi-dimensional output, we report the average R2 across dimensions. We include bias terms in our linear regressions but omit them from equations for simplicity. 4.1 Predicting SAE Error Norm For our first set of experiments, we find the optimal linear probe a from x to Sae Error(x) 2 2. Formally (with a slight abuse of notation, since x is a random variable and not a dataset), we solve for a = arg min a R a T x Sae Error(x) 2 2 2 (10) The R2 of these probes are all extremely high: across all combinations of SAE width, L0, layer, and model, between 70% and 95% of the variance in SAE error norm is explained by the optimal linear probe. Results across SAE width and L0: We plot the R2 of probes from Eq. (10) across SAE width and SAE L0 for layer 9 of Gemma 2 2B and layer 20 of Gemma 2 9B as a contour plot (on the left) in Fig. 2. We also include plots for additional layers in Appendix A. Overall, sparser and wider SAEs have less predictable error norms, although interestingly for some layers, extremely dense SAEs have less predictable error norm as well. Results across layers and models: In Fig. 3 we plot the R2 of probes from Eq. (10) for the SAEs with L0 closest to 50 on Gemma 2 2B and 9B and Llama Scope SAEs (which all have k = L0 = 50). We find that the R2 is low for the first few layers, increases rapidly to a value of about 0.8 to 0.95, remains at this level with a slight dip for mid-late layers, and then drops off steeply in the last few layers. In Appendix A, we compare the R2 of this approach across layers versus other baselines like token identity; we find that using activations does much better, and thus it is not easy to predict error norm. 1Code at https://anonymous.4open.science/r/SAE-Dark-Matter-1163 Published in Transactions on Machine Learning Research (04/2025) R2 predicting error norm R2 predicting error vector FVUnonlinear (a) Linear prediction results for layer 12 Gemma 2 2B SAEs from Gemma Scope. FVUnonlinear is roughly constant for a fixed L0. R2 predicting error norm R2 predicting error vector FVUnonlinear (b) Linear prediction results for layer 20 Gemma 2 9B SAEs from Gemma Scope. FVUnonlinear is roughly constant for a fixed L0. Figure 2: Results of linearly predicting SAE error norm and SAE error from model activations on Gemma 2 2B layer 12 (top) and Gemma 2 9B layer 20 (bottom). The right plots show the R2 of predicting SAE error norms (see Eq. (10), the middle plots show the R2 of predicting SAE error vectors (see Eq. (11), and the right plots show 1 R2 of predicting model activations given the SAE reconstruction and the SAE error vector prediction. We note that these 2D heatmaps are somewhat sparse and only the black dots represent actual SAEs. This is because we use Gemma Scope SAEs, which are trained only on some L0s and SAE widths. We do a linear interpolation between SAEs to predict R2 between hyperparameters. 4.2 Predicting SAE Error Vectors We next examine the R2 of the optimal linear transform b from x to Sae Error(x): b := arg min b Rd d b x Sae Error(x) 2 (11) As we show in Fig. 2 (middle column) and in Appendix A for additional layers, the R2 of these transforms varies widely, between 15% and 72%; this is less than the R2 for our norm prediction experiments, but still much higher than we might expect. Intuitively, this result implies that there are large linear subspaces that the SAE is mostly failing to learn. There is also a clear pattern across SAE L0 and width: R2 decreases with increasing SAE width and L0. Interestingly, this pattern is not the same as it was above for SAE error norm: the R2 of error norm predictions increases with SAE L0, while it decreases for error vector predictions. One Published in Transactions on Machine Learning Research (04/2025) 0.0 0.2 0.4 0.6 0.8 1.0 Layer by depth fraction in model R2 predicting error norm llama_3.1_8b 32k llama_3.1_8b 131k gemma_2_2b 16k gemma_2_2b 65k gemma_2_9b 16k gemma_2_9b 131k Figure 3: R2 of SAE error norm predictions (see Eq. (10)) for Gemma Scope SAEs of width L0 50 and Llama Scope SAEs of width L0 = k = 50. SAE Width (y) SAE Width (x) Figure 4: R2 for linear probes of per token SAE errors of larger SAEs from smaller SAEs. Prediction accuracy decreases as the SAEs get farther apart in scale, but overall remains high. concern might be that b is mostly reversing feature shrinkage; in Appendix A.1, we show that this is not the case. 4.3 Nonlinear FVU and Breaking Down SAE Dark Matter Another related metric we are interested in is the total amount of the original activation x we fail to explain using both Sae(x) and a linear projection of x. That is, assuming we have found b as in Eq. (11), we are interested in the fraction of variance unexplained (FVU) by the sum of Sae(x) and b x: FVUnonlinear := 1 R2(x, Sae(x) + b x) (12) We label this quantity FVUnonlinear because it is intuitively the amount of the SAE s unexplained variance that is not a linear projection of the input. Interestingly, we find that for middle layers (layer 12 Gemma 2 2B and layer 20 Gemma 2 9B) at a fixed L0, FVUnonlinear is approximately constant (see Fig. 2), while for other layers it decreases slightly as L0 increases (see Appendix A). That is, even though we can linearly predict a smaller portion of the error vector in larger SAEs, this effect is counteracted by the fact that the SAE error vector itself is getting smaller. In contrast, FVUnonlinear decreases as SAE L0 increases. As stated above, for middle layers we observe that FVUnonlinear is roughly constant at a fixed sparsity. Because we observe this across multiple orders of magnitude of SAE scale, we hypothesize that this will continue to hold as we scale the SAE. Using this assumption, we plot a horizontal fit for FVUnonlinear and a power law fit with a constant for Gemma SAE reconstructions as SAE width scales (choosing the SAEs closest to L0 60) in Fig. 1. The power law fit asymptotes above the horizontal fit, which implies the presence of linear error even at very large SAE width. The absent features component of the error in Fig. 1 comes from the following observation: if our hypothesis is correct that nonlinear error is roughly constant as the layer 20 SAEs scale, then the nonlinear features must reside in the linear error component, since this is the only component that decreases as the SAE scales and learns more features. We investigate this hypothesis further in Appendix B.2. 4.4 Predicting SAE Per-Token Error Norms We now examine per-token SAE scaling behavior. Given two SAEs, SAE1 and SAE2, we are interested in how much of the variance in error norms in SAE2 is predictable from error norms in SAE1. That is, we want to Published in Transactions on Machine Learning Research (04/2025) 0 20 40 60 80 100 Token ' hypothe' 'si' ' evaluation'',' 16k SAE error 32k SAE error 65k SAE error 131k SAE error 262k SAE error 524k SAE error 1m SAE error Figure 5: Per token scaling with average nonlinear error, layer 20 Gemma 9B SAEs from Gemma Scope closest to L0 = 60. find c such that c := arg min c R c Sae Error1(x) 2 2 Sae Error2(x) 2 2 2 (13) Note that in practice, although a also can predict the norm of SAE error, it requires training the target SAE to learn a probe. Here, on the other hand, although we formulate finding c as an optimization problem that requires a larger SAE, in practice we do not need to actually train the larger SAE to get interesting insights: since c has just one component, it simply measures how well small SAE error can be multiplied by a scalar to predict large SAE error. If the R2 is high, we know that on tokens that small SAEs perform poorly on, larger SAEs will as well. In Fig. 4, we plot the R2 of c probes on all pairs of Gemma Scope 9B layer 20 SAEs with L0 60 (restricting to pairs where SAE2 is larger than SAE2), and find that indeed, per token SAE errors are highly predictable. Additionally, we show concretely what these correlated SAE error norms looks like on a set of 100 tokens from the Pile in Fig. 5. 5 Analyzing Components of SAE Error In previous sections, we broke down the SAE error vector into a linearly predictable component and a nonlinearly predictable component. A reasonable question is whether these error subsets meaningfully differ. Thus, in this section, we run experiments on these components to determine how the linear and non-linear components of Sae Error(x) differ. For convenience, given a probe b from Eq. (11), we write Linear Error(x) := b x (14) Nonlinear Error(x) := Sae Error(x) Linear Error(x) = Sae Error(x) b (15) In Appendix B.2 we discuss a hypothesis for SAE feature activations that might explain why we observe these differences; this appendix section is not necessary for understanding the experiments in this section, and we will refer to it in the few cases where it provides additional intuition. Published in Transactions on Machine Learning Research (04/2025) Sae Error(x) Linear Error(x) Nonlinear Error(x) Norm Prediction Test R2 Figure 6: Violin plot of norm prediction tests for all SAEs (layer 12 Gemma 2B SAEs, layer 20 Gemma 9B SAEs, all Gemma Scope SAEs with L0 closest to 60 across layers, and all Llama Scope SAEs). We plot the R2 of a linear regression from x to to each random vector s norm squared. SAE trained on nonlinear error SAE trained on linear error Figure 7: Auto-interpretability results on SAEs trained on the linear and nonlinear components of Sae Error(x) on a width 16k, L0 60, layer 20 Gemma Scope SAE. Not represents contexts that the SAE latent did not activate on, while each Qi represents activating examples from decile i. We run the following experiments in this section to understand how the linear and nonlinear components of SAE error differ: 1. In Section 5.1, we run the norm prediction test from Eq. (10) on different components of the error. We find that the norm of the nonlinear component is less predictable from activations than other components, implying that it might consist of fewer not-yet-learned features. 2. We train new SAEs directly on the linear and nonlinear components of error. The SAE trained on nonlinear error converges to higher reconstruction loss and produces less interpretable features. 3. We examine how much each component contributes to downstream model performance by intervening in the forward pass, and find that both components contribute proportionally to their size to increased cross entropy loss. Overall, these experiments suggest that our split of SAE dark matter into these two categories is indeed a meaningful one. 5.1 Applying the Norm Prediction Test: For our first experiment, we run the norm prediction test from Eq. (10) on five different random vectors: x, Sae(x), Sae Error(x), Linear Error(x), and Nonlinear Error(x). The results are shown as a violin plot for each component across all 329 SAEs we experiment with in Section 4 (layer 12 Gemma 2B SAEs, layer 20 Gemma 9B SAEs, all Gemma Scope SAEs with L0 closest to 60 across layers, and all Llama Scope SAEs) in Fig. 6 (the Sae(x) bar can be seen as a summary of Fig. 2). There are one or two SAE with an outlier R2 equal to or lower than 0; we omit these from the plot because they are likely due to numeric instability in the linear regression routine. The intuition behind this test is that if a random vector consists of a sum of almost-orthogonal linear features from x, then its norm should be linearly predictable from x (see Appendix B for more on this intuition). Published in Transactions on Machine Learning Research (04/2025) Percent Recovered When Replacing Layer (%) Norm / Total SAE Error Norm Reconstruction + Linear Error Reconstruction + Nonlinear Error Figure 8: Results of intervening in the forward pass and replacing x with Sae(x) + Nonlinear Error(x) and Sae(x) + Linear Error(x) during the forward pass on all layer 20 Gemma 9B SAEs with L0 60 or width of 16k. Reported in percent of cross entropy loss recovered with respect to the difference between the same intervention with Sae(x) and with the normal model forward pass. Firstly, we note that x 2 2 can almost be perfectly predicted from x. This is reassuring news for the linear representation hypothesis, as it implies that x may be well modeled as the sum of many one-dimensional features, at least from the perspective of this test. We also find that Nonlinear Error(x) (and Sae Error(x), which consists partly of Nonlinear Error(x)) has a notably lower score on this test than Linear Error(x). This result suggests that Nonlinear Error(x) does not consist of a sparse sum of linear features from x. 5.2 Training SAEs on Sae Error(x) Components: Another empirical test we run is training an SAE on Nonlinear Error(x) and Linear Error(x). We choose a fixed Gemma 9B Gemma Scope layer 20 SAE with 16k latents and L0 60 to generate Sae Error(x) from. This SAE has nonlinear and linear components of the error approximately equal in norm and R2 of the total Sae Error(x) they explain, so it presents a fair comparison. We train SAEs to convergence (about 100M tokens) on each of these components of error and find that the SAE trained on Nonlinear Error(x) converges to a fraction of variance unexplained an absolute 5 percent higher than the SAE trained on the linear component of SAE error ( 0.59 and 0.54 respectively). We additionally examine the interpretability of the learned SAE latents using automated interpretability (this technique was first proposed by Bills et al. (2023) for interpreting neurons, and first applied to SAEs by Cunningham et al. (2023)). Specifically, we use the implementation introduced by Juang et al. (2024), where a language model (we use Llama 3.1 70b (AI@Meta, 2024)) is given top activating examples to generate an explanation, and then must use only that explanation to predict if the feature fires on a test context. Our results in Fig. 7 show that the SAE trained on linear error produces latents that are about an absolute 5% more interpretable across all activation firing deciles (we average results across 1000 random features for both SAEs, where for each feature use 7 examples in each of the 10 feature activation deciles as well as 50 negative examples, and show 95% confidence intervals). 5.3 Downstream Cross Entropy Loss of Sae Error(x) Components: A common metric used to test SAEs is the percent of cross entropy loss recovered when the SAE reconstruction is inserted into the model in place of the original activation versus an ablation baseline (see e.g. Bloom Published in Transactions on Machine Learning Research (04/2025) 105 106 107 0.0 Unlearned Features Dense Features / Linear Error Nonlinear Error (from Dictionary and Sparsity) Pursuit SAE Reconstruction, FVU 0.136 + 2.342W 0.272 Pursuit SAE Reconstruction + Error Prediction, FVU 0.117 (a) SAE error breakdown vs. SAE width for inference time optimized reconstructions from Gemma Scope L0 60 dictionaries. residual prev residual MLP attention Predicting Nonlinear Error residual prev residual MLP attention residual prev residual MLP attention Predicting Linear Error (b) The R2 of linearly predicting parts of SAE error from the SAE reconstructions of adjacent model components, layer 20 Gemma Scope L0 60, 16k SAE width. Figure 9: Investigations towards reducing nonlinear SAE error. (2024)). We modify this test to specifically examine the different components of Sae Error(x): we compare the percent of the cross entropy loss recovered when replacing x with Sae(x) plus either Linear Error(x) or Nonlinear Error(x) to the baseline of inserting just Sae(x) in place of x on all layer 20 Gemma 9B SAEs with L0 60 or width of 16k. To estimate how much each component should recover, we use two metrics: the average norm of the component relative to the total norm of Sae(x) and the percent of the variance that the component recovers between Sae(x) and x. The results, shown in Fig. 8, show that for the most part these metrics are reasonable predictions for both types of error. That is, both Nonlinear Error(x) and Linear Error(x) proportionally contribute to the SAE s increase in downstream cross entropy loss, with possibly a slightly higher contribution than expected for Linear Error(x). 6 Reducing Nonlinear Error(x) In past sections we identify Nonlinear Error(x) and Linear Error(x) and argue that Nonlinear Error(x) likely does not consist of not-yet-learned linear features. This may be a problem for scaling SAEs farther: if a part of SAE error persistently consists of nonlinear transformations of linear features, these may never be able to be learned by SAEs. Thus, in this section, we investigate two approaches for reducing Nonlinear Error(x): 1. Using a more powerful encoder at inference time. This approach has limited success, reducing total FVU by 3-5% but leaving FVUnonlinear almost unchanged 2. Attempting to predict Nonlinear Error(x) from SAE reconstructions of adjacent model components - this approach is more successful; on the SAE we test on, Sae(x) predicts 50% of the nonlinear error, and previous model components predict between 4% and 14% of nonlinear error, although only between 6% and 10% of the overall SAE error. 6.1 Using a More Powerful Encoder Our first approach for reducing nonlinear error is to try improving the encoder. We use a recent approach suggested by Smith (2024a): applying a greedy inference time optimization (ITO) algorithm called gradient pursuit to a frozen learned SAE decoder matrix. We implement and run ITO on all layer 20 Gemma Scope 9b SAEs closest to L0 60. For each example x with reconstruction Sae(x), we use the gradient pursuit implementation with an L0 exactly equal to the L0 of x in the original Sae(x). Using these new reconstructions of x, we repeat Eq. (11) and do a linear transformation from x to the inference time optimized reconstructions. We then regenerate the same scaling plot as Fig. 1 and show this Published in Transactions on Machine Learning Research (04/2025) figure in Fig. 9a. Our first finding is that pursuit indeed decreases the total FVU of Sae(x) by 3 to 5%; as Smith (2024a) only showed an improvement on a small 1 layer model, to the best of our knowledge we are the first to show this result on state of the art SAEs. Our most interesting finding, however, is that the FVUnonlinear stays almost constant when compared to the original SAE scaling in Fig. 1. We hypothesize that this might happen because ITO reduces easy linear errors like feature shrinkage. In Fig. 1, we plot the additional reduction in FVUnonlinear as the contribution of encoder error; because FVUnonlinear stays almost constant, this section is very narrow. 6.2 Predicting SAE Errors Between SAEs: For the Gemma 2 architecture at the locations the SAEs are trained on, each residual activation can be decomposed in terms of prior components: Residlayer = Mlp Outlayer + RMSNorm(Oproj(Attn Outlayer)) + Residlayer 1 (16) We focus on layer = 20 and Gemma 2 9B, and thus use the layer 19 attention and MLP Gemma Scope SAEs. For all components we use the width 16k SAEs with L0 60. In Fig. 9b, we plot the R2 of a regression from the SAE output corresponding to each of these right hand side components to each of the different components of an SAE trained on Residlayer (Sae Error(x), Linear Error(x), and Nonlinear Error(x)). We find that we can explain a small amount (up to 10%) of total Sae Error(x) using previous components, which may be immediately useful for circuit analysis. We also can explain large parts of Nonlinear Error(x) with prior components as well. While Sae(x) explains 50% of the variance in the nonlinear error, this may be somewhat misleading, as the nonlinear error is partially a function of Sae(x): Nonlinear Error(x) = Sae Error(x) Linear Error(x) = (x Sae(x)) Linear Error(x) These results mean that we might be able to explain some of the SAE Error using a circuits level view, but that overall there are still large parts of each error component unexplained. 7 Conclusion The fact that SAE error can be predicted and analyzed at all is surprising. Thus, our findings are intriguing evidence that SAE errors, and not just SAE reconstructions, are worthy of analysis, and we hope that it inspires further work in decomposing SAE error. Concretely, we believe that our study of error has already discovered a number of promising directions for future SAE research. The discovery that SAE errors can be significantly predicted from model activations hints that some subspaces may resist being learned and points toward potential architectural innovations such as incorporating low-rank dense side channels. The predictable relationship between small and large SAE errors may also streamline experimentation with novel SAE architectures by allowing researchers to efficiently forecast scaling behavior through small-scale trials. Additionally, our demonstration that prior SAE outputs can explain part of SAE error has immediate practical implications for circuit analysis, which currently relies on large noise terms that complicate circuit interpretation. At a higher level, the presence of constant nonlinear error at a fixed sparsity as we scale implies that scaling SAEs may not be the only (or best) way to explain more of model behavior. Future work might explore alternative penalties besides sparsity or new ways to learn better dictionaries. Ultimately, we believe that there is still room to make SAEs better, not just bigger. Acknowledgments Our work benefited greatly from the thoughtful comments and discussions provided by (in alphabetical order) Joseph Bloom, Lauren Greenspan, Jake Mendel, and Eric Michaud. We are deeply appreciative of their contributions. This work was supported by Erik Otto, Jaan Tallinn, the Rothberg Family Fund for Published in Transactions on Machine Learning Research (04/2025) Cognitive Science, and IAIFI through NSF grant PHY-2019786. JE was supported by the NSF Graduate Research Fellowship (Grant No. 2141064). LS was supported by the Long Term Future Fund. AI@Meta. Llama 3 model card, 2024. URL https://github.com/meta-llama/llama3/blob/main/MODEL_ CARD.md. Guillaume Alain. Understanding intermediate layers using linear classifier probes. ar Xiv preprint ar Xiv:1610.01644, 2016. Evan Anders and Joseph Bloom. Examining language model performance with reconstructed activations using sparse autoencoders. Less Wrong, 2024. URL https://www.lesswrong.com/posts/ 8QRH8w Kcn KGhp Au2o/examining-language-model-performance-with-reconstructed. Transformer Circuits Team Anthropic. Circuits updates april 2024, 2024. URL https:// transformer-circuits.pub/2024/april-update/index.html#scaling-laws. Steven Bills, Nick Cammarata, Dan Mossing, Henk Tillman, Leo Gao, Gabriel Goh, Ilya Sutskever, Jan Leike, Jeff Wu, and William Saunders. Language models can explain neurons in language models. URL https://openaipublic. blob. core. windows. net/neuron-explainer/paper/index. html.(Date accessed: 14.05. 2023), 2, 2023. Joseph Bloom. Open source sparse autoencoders for all residual stream layers of gpt2 small. https://www.alignmentforum.org/posts/f9Egf LSur Aiq RJy SD/ open-source-sparse-autoencoders-for-all-residual-stream, 2024. Trenton Bricken, Adly Templeton, Joshua Batson, Brian Chen, Adam Jermyn, Tom Conerly, Nick Turner, Cem Anil, Carson Denison, Amanda Askell, Robert Lasenby, Yifan Wu, Shauna Kravec, Nicholas Schiefer, Tim Maxwell, Nicholas Joseph, Zac Hatfield-Dodds, Alex Tamkin, Karina Nguyen, Brayden Mc Lean, Josiah E Burke, Tristan Hume, Shan Carter, Tom Henighan, and Christopher Olah. Towards monosemanticity: Decomposing language models with dictionary learning. Transformer Circuits Thread, 2023. https://transformer-circuits.pub/2023/monosemantic-features/index.html. Bart Bussmann, Patrick Leask, Joseph Bloom, Curt Tigges, and Neel Nanda. Stitching saes of different sizes. AI Alignment Forum, 2024. URL https://www.alignmentforum.org/posts/ba Jyjpktzmcm Rfosq/ stitching-saes-of-different-sizes. Emmanuel J Candès, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis? Journal of the ACM (JACM), 58(3):1 37, 2011. Róbert Csordás, Christopher Potts, Christopher D Manning, and Atticus Geiger. Recurrent neural networks learn to store and generate sequences using non-linear representations. ar Xiv preprint ar Xiv:2408.10920, 2024. Hoagy Cunningham, Aidan Ewart, Logan Riggs, Robert Huben, and Lee Sharkey. Sparse autoencoders find highly interpretable features in language models. ar Xiv preprint ar Xiv:2309.08600, 2023. Nelson Elhage, Tristan Hume, Catherine Olsson, Nicholas Schiefer, Tom Henighan, Shauna Kravec, Zac Hatfield-Dodds, Robert Lasenby, Dawn Drain, Carol Chen, Roger Grosse, Sam Mc Candlish, Jared Kaplan, Dario Amodei, Martin Wattenberg, and Christopher Olah. Toy models of superposition. Transformer Circuits Thread, 2022. https://transformer-circuits.pub/2022/toy_model/index.html. Joshua Engels, Isaac Liao, Eric J Michaud, Wes Gurnee, and Max Tegmark. Not all language model features are linear. ar Xiv preprint ar Xiv:2405.14860, 2024. Simon Foucart and Holger Rauhut. A Mathematical Introduction to Compressive Sensing. Applied and Numerical Harmonic Analysis. Springer, New York, 2013. ISBN 978-0-8176-4947-0. doi: 10.1007/ 978-0-8176-4948-7. Published in Transactions on Machine Learning Research (04/2025) Leo Gao, Stella Biderman, Sid Black, Laurence Golding, Travis Hoppe, Charles Foster, Jason Phang, Horace He, Anish Thite, Noa Nabeshima, et al. The pile: An 800gb dataset of diverse text for language modeling. ar Xiv preprint ar Xiv:2101.00027, 2020. Leo Gao, Tom Dupré la Tour, Henk Tillman, Gabriel Goh, Rajan Troll, Alec Radford, Ilya Sutskever, Jan Leike, and Jeffrey Wu. Scaling and evaluating sparse autoencoders. ar Xiv preprint ar Xiv:2406.04093, 2024. Giorgi Giglemiani, Nora Petrova, Chatrik Singh Mangat, Jett Janiak, and Stefan Heimersheim. Evaluating synthetic activations composed of sae latents in gpt-2. ar Xiv preprint ar Xiv:2409.15019, 2024. Wes Gurnee. Sae reconstruction errors are (empirically) pathological. In AI Alignment Forum, pp. 16, 2024. Zhengfu He, Wentao Shu, Xuyang Ge, Lingjie Chen, Junxuan Wang, Yunhua Zhou, Frances Liu, Qipeng Guo, Xuanjing Huang, Zuxuan Wu, et al. Llama scope: Extracting millions of features from llama-3.1-8b with sparse autoencoders. ar Xiv preprint ar Xiv:2410.20526, 2024. Stefan Heimersheim and Jake Mendel. Activation plateaus & sensitive directions in gpt2. Less Wrong, 2024. URL https://www.lesswrong.com/posts/Laj Dy Gyiy X8DNNsu F/ interim-research-report-activation-plateaus-and-sensitive-1. Albert Q Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, et al. Mistral 7b. ar Xiv preprint ar Xiv:2310.06825, 2023. Caden Juang, Gonçalo Paulo, Jacob Drori, and Nora Belrose. Open source automated interpretability for sparse autoencoder features. https://blog.eleuther.ai/autointerp/, 2024. Adam Karvonen, Benjamin Wright, Can Rager, Rico Angell, Jannik Brinkmann, Logan Smith, Claudio Mayrink Verdun, David Bau, and Samuel Marks. Measuring progress in dictionary learning for language model interpretability with board game models. ar Xiv preprint ar Xiv:2408.00113, 2024. Vedang Lad, Wes Gurnee, and Max Tegmark. The remarkable robustness of llms: Stages of inference? ar Xiv preprint ar Xiv:2406.19384, 2024. Daniel Lee and Stefan Heimersheim. Investigating sensitive directions in gpt-2: An improved baseline and comparative analysis of saes. Less Wrong, 2024. URL https://www.lesswrong.com/posts/ d S5d Sgwa DQRo Wd Tuu/investigating-sensitive-directions-in-gpt-2-an-improved. Tom Lieberum, Senthooran Rajamanoharan, Arthur Conmy, Lewis Smith, Nicolas Sonnerat, Vikrant Varma, János Kramár, Anca Dragan, Rohin Shah, and Neel Nanda. Gemma scope: Open sparse autoencoders everywhere all at once on gemma 2. ar Xiv preprint ar Xiv:2408.05147, 2024. Alireza Makhzani and Brendan Frey. K-sparse autoencoders. ar Xiv preprint ar Xiv:1312.5663, 2013. Samuel Marks, Can Rager, Eric J Michaud, Yonatan Belinkov, David Bau, and Aaron Mueller. Sparse feature circuits: Discovering and editing interpretable causal graphs in language models. ar Xiv preprint ar Xiv:2403.19647, 2024. Jake Mendel. Sae feature geometry is outside the superposition hypothesis. AI Alignment Forum, 2024. URL https://www.alignmentforum.org/posts/MFBTjb2qf3zi Wmzz6/ sae-feature-geometry-is-outside-the-superposition-hypothesis. Neel Nanda, Andrew Lee, and Martin Wattenberg. Emergent linear representations in world models of self-supervised sequence models. ar Xiv preprint ar Xiv:2309.00941, 2023. Chris Olah. Interpretability dreams. Transformer Circuits, May 2023. URL https:// transformer-circuits.pub/2023/interpretability-dreams/index.html. Published in Transactions on Machine Learning Research (04/2025) Kiho Park, Yo Joong Choe, and Victor Veitch. The linear representation hypothesis and the geometry of large language models. ar Xiv preprint ar Xiv:2311.03658, 2023. Senthooran Rajamanoharan, Tom Lieberum, Nicolas Sonnerat, Arthur Conmy, Vikrant Varma, János Kramár, and Neel Nanda. Jumping ahead: Improving reconstruction fidelity with jumprelu sparse autoencoders. ar Xiv preprint ar Xiv:2407.14435, 2024. Lewis Smith. Replacing sae encoders with inference-time optimisation. https://www.alignmentforum. org/s/At TZjo Dm8q3Db DT8Z/p/C5KAZQib3bzzpeyrg, 2024a. Lewis Smith. The strong feature hypothesis could be wrong. AI Alignment Forum, 2024b. URL https://www.alignmentforum.org/posts/tojt PCCRp KLSHBdpn/ the-strong-feature-hypothesis-could-be-wrong. Abiy Tasissa, Manos Theodosis, Bahareh Tolooshams, and Demba E Ba. Discriminative reconstruction via simultaneous dense and sparse coding. Transactions on Machine Learning Research. Abiy Tasissa, Emmanouil Theodosis, Bahareh Tolooshams, and Demba Ba. Towards improving discriminative reconstruction via simultaneous dense and sparse coding. ar Xiv preprint ar Xiv:2006.09534, 2020. Gemma Team, Morgane Riviere, Shreya Pathak, Pier Giuseppe Sessa, Cassidy Hardin, Surya Bhupatiraju, Léonard Hussenot, Thomas Mesnard, Bobak Shahriari, Alexandre Ramé, et al. Gemma 2: Improving open language models at a practical size. ar Xiv preprint ar Xiv:2408.00118, 2024. Adly Templeton, Tom Conerly, Jonathan Marcus, Jack Lindsey, Trenton Bricken, Brian Chen, Adam Pearce, Craig Citro, Emmanuel Ameisen, Andy Jones, Hoagy Cunningham, Nicholas L Turner, Callum Mc Dougall, Monte Mac Diarmid, C. Daniel Freeman, Theodore R. Sumers, Edward Rees, Joshua Batson, Adam Jermyn, Shan Carter, Chris Olah, and Tom Henighan. Scaling monosemanticity: Extracting interpretable features from claude 3 sonnet. Transformer Circuits Thread, 2024. URL https: //transformer-circuits.pub/2024/scaling-monosemanticity/index.html. Published in Transactions on Machine Learning Research (04/2025) R2 predicting error norm R2 predicting error vector FVUnonlinear (a) Linear prediction results for layer 5 Gemma 2 2B SAEs from Gemma Scope. R2 predicting error norm R2 predicting error vector FVUnonlinear (b) Linear prediction results for layer 19 Gemma 2 9B SAEs from Gemma Scope. Figure 10: Results of linearly predicting SAE error norm and SAE error from model activations on Gemma 2 2B layer 5 (top) and Gemma 2 2B layer 19 (bottom). The right plots show the R2 of predicting SAE error norms (see Eq. (10), the middle plots show the R2 of predicting SAE error vectors (see Eq. (11), and the right plots show 1 R2 of predicting model activations given the SAE reconstruction and the SAE error vector prediction. Unlike for the middle layers shown in the main body, FVUnonlinear decreases some as width is scaled, although the sparsity of the SAEs in the space makes this harder to verify. A Extra Error Prediction Experiments A.1 Note on Feature Shrinkage Earlier SAE variants were prone to feature shrinkage: the observation that Sae(x) systematically undershot x. Current state of the art SAE variants (e.g. Jump Re LU SAEs and Top K SAEs, which we examine in this work), are less vulnerable to this problem, although we still find that Gemma Scope reconstructions have about a 10% smaller norm than x. One potential concern is that the b in Eq. (11) that we learn is merely predicting this shrinkage. If this was the case, then the cosine similarity of the linear error prediction (b )T x with x would be close to 1; however, in practice we find that it is around 0.5, so b is indeed doing more than predicting shrinkage. Published in Transactions on Machine Learning Research (04/2025) R2 predicting error norm R2 predicting error vector FVUnonlinear (a) Linear prediction results for layer 9 Gemma 2 9B SAEs from Gemma Scope. R2 predicting error norm R2 predicting error vector FVUnonlinear (b) Linear prediction results for layer 31 Gemma 2 9B SAEs from Gemma Scope. Figure 11: Results of linearly predicting SAE error norm and SAE error from model activations on Gemma 2 9B layer 9 (top) and Gemma 2 9B layer 31 (bottom). The right plots show the R2 of predicting SAE error norms (see Eq. (10), the middle plots show the R2 of predicting SAE error vectors (see Eq. (11), and the right plots show 1 R2 of predicting model activations given the SAE reconstruction and the SAE error vector prediction. Unlike for the middle layers shown in the main body, FVUnonlinear decreases some as width is scaled, although the sparsity of the SAEs in the space makes this harder to verify. A.2 Additional Error Prediction Plots In Section 7 and Section 7, we show the accuracy of SAE error norm predictions, SAE error vector predictions, and FVUnonlinear for layers 5 and 19 of Gemma 2 2B and layers 9 and 31 of Gemma 2 9B. As described in the main text, we find broadly similar results for SAE error norm and error vector prediction at these layers, but find that FVUnonlinear decreases some as we increase SAE width at a fixed L0, although this result is uncertain because of the sparsity of Gemma Scope SAEs at these layers. A.3 Norm Prediction Baselines In Fig. 13, we run a linear regression from different model variables to Sae Error(x) across layers on Gemma Scope 9B for SAEs of size 131k. We find that is is not easy to predict SAE error norm, especially at later layers; the token identity, SAE L0, activation norm, and model loss all do significantly worse than using the full activation. It is interesting to note that at the first few layers, token identity does better at Published in Transactions on Machine Learning Research (04/2025) 0 20 40 60 80 100 Token Norm of Linearly Predictable Error 0 20 40 60 80 100 Token Norm of Non-Linearly Predictable Error 16k 32k 65k 131k 262k 524k 1m Figure 12: Per-token breakdown of linearly predictable and non-linearly predictable SAE error across SAE scale. We show the same tokens as in Fig. 5. The norm of linear error decreases with SAE width, whereas the norm of nonlinear error stays mostly constant. predicting SAE error than a probe of the activations; this is perhaps not surprising, since recent results from e.g. Lad et al. (2024) show that very early layers primarily operate on a per token level. A.4 Breaking Apart Error Per Token In Fig. 12, we show the same subset of tokens as in Fig. 5, but now broken apart into linearly predictable and non-linearly predictable components. That is, we learn b for each SAE as in Eq. (11), and then plot the norm of b x at the norm of the linearly predictable error on the left, and plot the norm of Sae Error(x) b x as the norm of the non-linearly predictable error on the right. We see that the linearly predictable error decreases as we scale SAE width, but the non-linearly predictable error mostly stays constant. This is especially interesting because the result in Fig. 1 just found this on an average level, whereas here we find the same result holds on a per-token level. B Modeling Activations In this section, we seek to explain why we see a difference between linear and nonlinear error in the main body of the paper. We will adopt the weak linear hypothesis (Smith, 2024b), a generalization of the linear representation hypothesis which holds only that some features in language models are represented linearly. Thus we have i=0 wiyi + Dense(x) (17) for linear features {y1, . . . , yn} and random vector w Rn, where w is sparse ( w 1 d) and Dense(x) is a random vector representing the dense component of x. Dense(x) might be Gaussian noise, nonlinear features as described by Csordás et al. (2024), or anything else not represented in a low-dimensional linear subspace. Before we proceed with our analysis, we note that there is a rich sparse coding literature studying dictionary learning in the setting with mixed sparse and dense signals. For example, in a classic work Candès et al. (2011) propose Robust Principal Component Analysis for decomposing data matrices into sparse and dense components, although this is not directly applicable to our setting of trying to learn an autoencoder. More recently, Tasissa et al. propose a dictionary learning technique for data exactly modeled as in Eq. (17). In our work, we focus on studying existing SAEs and hypothesizing why they fail, so these works are not immediately applicable, but we are excited to see future work that applies these combined sparse-dense autoencoders to language model activations.. Say our SAE has m latents. Since by assumption Dense(x) cannot be represented in a low-dimensional linear subspace, the sparsity limited SAE will not be able to learn it. Thus, we will assume that the SAE learns Published in Transactions on Machine Learning Research (04/2025) 0 10 20 30 40 Acts Tokens Model Loss L0 Figure 13: R2 for linear regressions of SAE error norms with different regressors. We run on Gemma Scope 9B SAEs of size 131k with L0 60. Activations perform the best except on the first few layers. 0 2500 5000 7500 10000 SAE Latent Index Sorted by Average Active Value (a * )T feature Running window average of (a * )T feature Average feature value when active Figure 14: Average SAE latent activation and dot product of the latent with the learned norm prediction vector a for the Gemma Scope layer 20, width 131k, L0 = 62 SAE. We also plot a smoothed version of this dot product with a smoothed window of 10. only the m most common features y0, . . . , ym 1. We will also assume that the SAE introduces some error when making this approximation, and instead learns ˆyi and ˆwi. Thus we have i=0 ˆwi ˆyi (18) Sae Error(x) = Dense(x) + i=0 ˆwi ˆyi i=m wiyi (19) We finally define Introduced(x) := Pm i=0 ˆwi ˆyi Pm i=0 wiyi, so we have Sae Error(x) = Dense(x) + Introduced(x) + i=m wiyi (20) B.1 Analyzing Error Norm Prediction We will first analyze Eq. (10), the learned probe from x to Sae Error(x) 2 2. First, we claim that given a vector x, if x is a sparse sum of orthogonal vectors, then by basic linear algebra there exists a perfect prediction vector a such that a T x |x|2 2 (in other words, the norm squared of x can be linearly predicted from x). The proof of this claim is in Appendix C.1; the intuition is that we can set the probe vector a to the sum of the vectors yi weighted by their average weight E(wi). When x is instead a sparse sum of non-orthogonal vectors, as it partly is in Eq. (17) and Eq. (20), this proof is no longer true, but we now argue that a similar intuition holds. If the yi are almost orthogonal (formally coherence < ϵ, see (Foucart & Rauhut, 2013)) and do not activate much at the same time, then a probe vector again equal to the sum of vectors yi weighted by their average value E(wi) will be a good approximate prediction. Indeed, when we try predicting Sae(x) 2 2 from Sae(x) (which is a sparse sum of Published in Transactions on Machine Learning Research (04/2025) known almost orthogonal vectors of a similar distribution to the true SAE vectors), we find that indeed the linear probe that is learned is approximately equal to this sum (see Appendix C.2). Thus, we can now neatly explain why we can predict the norms of SAE errors: they mostly consist of almost orthogonal sparsely occuring not yet learned SAE features! We further can explain why larger SAEs have less predictable error norms: since m is larger, there is a larger component in the error of not-as-linearlypredictable Dense(x) and Introduced(x). B.2 Analyzing Error Vector Prediction We will now analyze Eq. (11), the learned transformation from x to Sae Error(x), with our model of SAE error from Eq. (20). We assume that Introduced(x) cannot be approximated at all as a linear function of x. This is a reasonable assumption since the SAE is a nonlinear function of x, but if indeed some of Introduced(x) can be approximated in this way, then we will underestimate the amount of Introduced(x) and therefore overestimate the amount of Dense(x). If Dense(x) + Pn i=m wiyi is contained in a linear subspace of x orthogonal to Pm i=0 wiyi, then Dense(x) is part of the linearly explainable error, and the error of the transformation b exactly equals Introduced(x) (since the transformation is just exactly this orthogonal linear subspace). However, if such an orthogonal linear subspace does not exist, the optimal linear transform will only be able to reconstruct part of Dense(x)+ Pn i=m wiyi, and the percent of variance left unexplained by the regression will be an upper bound on the true variance explained by Introduced(x). We also note that if this linear transform indeed recovers Dense(x) and absent features but fails to recover Introduced(x), we can use it to estimate Dense(x): the difference between the variance explained by Sae(x) and the variance explained by x (Sae(x) + b x) will approach Dense(x) as m . Thus, our ability to estimate Introduced(x) and Dense(x) using b depends on how well a linear transform works to predict Dense(x) and Pn i=m wiyi. Although we do not have access to the ground truth vectors yi, we can replace x with a similar distribution of vectors that we do have access to, using the same trick as above. Given an SAE, we replace x with x = Sae(x). x has the useful property that it is a sparse linear sum of vectors (the ones that the SAE learned), and the distribution of these vectors and their weights are similar to that of the true features yi. We now pass x back through the SAE and can control all of the quantities we are interested in: we can vary m by masking SAE dictionary elements, simulate Dense(x ) by adding Gaussian noise to x , and simulate Introduced(x) by adding Gaussian noise to Sae(x ). Table 1: Correlation matrix between synthetic noise and estimated errors. Estimated Dense(x ) Estimated Introduced(x ) x Noise 0.9842 0.1417 Sae(x ) Noise 0.0988 0.9036 We run this synthetic setup with a Gemma Scope layer 20 SAE (width 16k, L0 68) in Appendix C.3, and find that indeed, estimated Dense(x ) is highly correlated with the amount of Gaussian noise added to x and Introduced(x ) is highly correlated with the amount of Gaussian noise added to Sae(x ) (see Table 1). However, note that because x noise is also slightly correlated with estimated Introduced(x ), it is possible that some of the contribution to the estimated nonlinear error is from Dense(x ). Thus, we again now have a potential explanation for our initial results: we can predict error vectors because they consist in large part of not yet learned linear features in an almost orthogonal subspace of x, we can predict a smaller portion of larger SAE errors because the number of these linear features go down with SAE width, and the horizontal line in Fig. 1 is because Dense(x) and Introduced(x) are mostly constant. Furthermore, we can hypothesize from the correlations om Table 1 that the linearly predictable component of SAE error consists mostly of not yet learned features and Dense(x), while the component that is not linearly predictable consists mostly of Introduced(x). We will explore this hypothesis in Section 5. Published in Transactions on Machine Learning Research (04/2025) B.3 Analyzing Per-Token Scaling Predictions Finally, we provide a simple explanation for why per-token SAE errors are highly predictable between SAEs of different sizes. For this, we only need Eq. (20). Since Dense(x) and Introduced(x) stay mostly constant as m increases, for large m the SAE error stays mostly constant because it is primarily determined by these components. Thus, since m = 16k is already large, a linear prediction that is just a slightly smaller version of the current error performs well. Additionally, this reasoning suggests a natural experiment: if we can predict Introduced(x) on a per-token level (which we hypothesize we can do with the non-linearly predictable component of SAE error), we may be able to better predict the floor of SAE scaling and therefore better predict larger SAE errors; we run this experiment in Appendix E, where we find an affirmative answer. C More Info on Modeling Activations C.1 Proof of Claim from Appendix B.1 Say we have a set of m unit vectors y1, y2, . . . , ym Rd. We will call these feature vectors . Define Y Rd m as the matrix with the feature vectors as columns. We then define the Gram matrix GY Rm m of dot products on Y: (GY)ij = (YT Y)ij = yi yj We now will define a random column vector x that is a weighted positive sum of the m feature vectors, that is, x = P i wiyi for a non-negative random vector w Rm. We say feature vector yi is active if wi > 0. We now define the autocorrelation matrix Rw Rm m for w as R = E(ww T ). We are interested in breaking down x into its components, so we define a random matrix X as Xij = wj Yij, i.e. the columns of Y multiplied by w. We can now define the Gram matrix GX Rm m: (GX)ij = (XT X)ij = wiwjyi yj GX = (ww T ) GY E(GX) = Rw GY, where denotes Schur (elementwise) multiplication. The intuition here is that the expected dot product between columns of X depends on the dot product between the corresponding columns of Y and the correlation of the corresponding elements of the random vector. We will now examine the L2 norm of x: ij wiwjyiyj = (w T w) GY 2 F = Tr(wwt GY) = w GYw T We can also take the expected value: E( x 2 2) = Tr(Rw GY) Our goal is to find a direction a Rd that when dotted with x predicts x 2 2. In other words, we want to find a such that x 2 2 a T x = a T X i wiyi = a T Yw Combining equations, we want to find a such that a T Yw x 2 2 = (vw T GYw) Published in Transactions on Machine Learning Research (04/2025) Let us first consider the simple case where for all i = j, yi and yj are perpendicular. Then our goal is to find a such that a T Yw Tr(w GYw T ) = X i yi, yi w2 i = X i w2 i = w 2 2 = w T w Since all of the yi are perpendicular, WLOG we can write a = P i biyi + c for a vector c Rd perpendicular to all yi and a vector b Rm. Then we have Since ordinary least squares produces an unbiased estimator, we know that if we use ordinary least squares to solve for b, E(b T w) = E(w T w). Thus, X i bi E(wi) = X bi = E(w2 i )/E(wi) Now that we have bi, we can solve for the correlation coefficient between a T x = b T w and x 2 2 = w T w. This gets messy when using general distributions, so we focus on a few simple cases. The first is the case where each wi is a scaled independent Bernoulli distribution, so wi is si with probability pi and 0 otherwise. Then bi = si. We also have that E(w T w) = E(b T w) = P i s2 i pi = µ. ρ = E(b T ww T w) µ2 p E(w T ww T w) µ2p E(b T wb T w) µ2 i s4 i (pi p2 i ) p P i s4 i (pi p2 i ) p P i s4 i (pi p2 i ) = 1 That is, for Bernoulli variables, x = P i siyi is a perfect regression vector. The second is the case when each wi is an independent Poisson distribution with parameter λi. Then E(wi) = λi and E(w2 i ) = λ2 i + λi, so bi = λi + 1. We also have that E(w T w) = E(b T w) = P i λ2 i + λi = µ. Finally, we will use the fact that E(w3 i ) = λ3 i + 3λ2 i + λi and E(w4 i ) = λ4 + 6λ3 + 7λ2 + λ. Then via algebra we have that i 2λ3 i + 3λ2 i + λi p P i 4λ3 + 6λ3 i + λi p P i λ3 i + 2λ2 i + λi For the special case λi = 1, we then have C.2 Empirical Norm Prediction In this experiment, we aim to determine to what extent our analysis in Appendix B.1 holds true in practice on almost orthogonal true SAE features. Thus, we use a random vector that we can control: Sae(x). Specifically, we learn a probe a for the Gemma Scope layer 20, width 131k, L0 = 62 SAE as in Eq. (10), except with the regressor equal to Sae(x) and the target equal to Sae(x) : a = arg min a R a T Sae(x) Sae(x) 2 2 2 (21) Published in Transactions on Machine Learning Research (04/2025) 50% 40% 30% 20% 10% 0% % features ablated Linear Error f(Activation Noise) Nonlinear Error f(Reconstruction Noise) Sae(x ) Noise Var = 1, x Noise Var = 0.6 50% 40% 30% 20% 10% 0% % features ablated Linear Error f(Activation Noise) Nonlinear Error f(Reconstruction Noise) Sae(x ) Noise Var = 0.6, x Noise Var = 1 0.0 1.0 2.0 3.0 4.0 True Nonlinear Noise Var True Linear Noise Var (FVU difference) 0.0 1.0 2.0 3.0 4.0 True Nonlinear Noise Var True Linear Noise Var (FVU difference) Est Nonlinear Error Est Linear Error Figure 15: Top: When controlled amounts of noise are added to synthetic data Sae(x ) and x , the result is a plot similar to Fig. 1. Bottom: The nonlinear and linear error estimates (as shown at top) accurately correlate with the amount of noise added. The exact correlation between synthetic added noise and resulting estimated error components across these noise levels are shown in Table 1 One important note is that we subtract the bias from Sae(x) so that it is purely a sparse sum of SAE features (this makes analysis easier). For each SAE latent from the SAE, we then compute (a )T latenti (22) Finally, we plot this dot product against the average latent activation in Fig. 14. If a indeed equals the sum of the latents weighted by their activation, as we predict in Appendix B.1, then these two quantities should be approximately equal, which we indeed see in the figure. C.3 Synthetic SAE Error Vector Experiments The results for different Gaussian noise amounts versus percentage of features ablated are shown in Fig. 15. On this distribution of vectors, the test works as expected; the variance explained by Sae(x) + a T x is a horizontal line proportional to Introduced(x), while the gap between this horizontal line and the asymptote of the variance explained by Sae(x) is proportional to Dense(x). We also tried running this test on a sparse sum of random vectors, which did not work as well, possibly due to not including the structure of the SAE vectors (Giglemiani et al., 2024); see Appendix D for more details. D Synthetic Experiments with Random Data For this set of experiments, we generated a random vector x that was the sum of a power law of 100k random gaussian vectors in R4000 with expected L0 of around 100. To simulate the SAE reconstruction and SAE error, we simply masked a portion of the vectors in the sum of x . Unlike the more realistic synthetic data case we describe in Appendix C.3, this did not work as expected: even in the case with no noise added to x or the simulated reconstruction, the variance explained by the sum of the linear estimate of the error plus the reconstructed vectors plotted against the number of features ablated formed a parabola (with minimum variance explained in the middle region), as opposed to a straight line as in Fig. 15. Published in Transactions on Machine Learning Research (04/2025) SAE Width (y) SAE Width (x) Figure 16: Percent decrease in FVU when additionally using the squared norms of nonlinear error to predict SAE error norm. We note that this result is not entirely surprising: other works have found that random vectors are a bad synthetic test case for language model activations. For example, in the setting of model sensitivity to perturbations of activations, Giglemiani et al. (2024) found they needed to control for both sparsity and cosine similarity of SAE latents to produce synthetic vectors that mimic SAE latents when perturbed. E Using Nonlinear Error(x) to Predict Scaling: Following up on our discussion in Appendix B.3, we are interested in whether Nonlinear Error(x) can help with predicting SAE per-token error norm scaling, as this might suggest that it contains a larger component of Introduced(x) than Linear Error(x). Formally, we solve for d := arg min d R d T [Sae Error1(x), Nonlinear Error1(x)] SAE2(x) 2 (23) To evaluate the improvement of d relative to c from Eq. (13), we report the percent decrease in FVU; see Fig. 16. We find that using the norm of Nonlinear Error(x) provides a small but noticeable bump in the ability to predict larger SAE errors of up to a 5% decrease in FVU, validating this hypothesis.