# lowrank_adapting_models_for_sparse_autoencoders__4848d285.pdf Low-Rank Adapting Models for Sparse Autoencoders Matthew Chen * 1 Joshua Engels * 1 Max Tegmark 1 Sparse autoencoders (SAEs) decompose language model representations into a sparse set of linear latent vectors. Recent works have improved SAEs using language model gradients, but these techniques require many expensive backward passes during training and still cause a significant increase in cross entropy loss when SAE reconstructions are inserted into the model. In this work, we improve on these limitations by taking a fundamentally different approach: we use low-rank adaptation (Lo RA) to finetune the language model itself around a previously trained SAE. We analyze our method across SAE sparsity, SAE width, language model size, Lo RA rank, and model layer on the Gemma Scope family of SAEs. In these settings, our method reduces the cross entropy loss gap by 30% to 55% when SAEs are inserted during the forward pass. We also find that compared to end-to-end (e2e) SAEs, our approach achieves the same downstream cross entropy loss 3 to 20 faster on Gemma-2-2B and 2 to 10 faster on Llama-3.2-1B. We further show that our technique improves downstream metrics and can adapt multiple SAEs at once without harming general language model capabilities. Our results demonstrate that improving model interpretability is not limited to post-hoc SAE training; Pareto improvements can also be achieved by directly optimizing the model itself.1 1. Introduction Although language models demonstrate profound capabilities in tasks such as in-context learning, mathematics, and coding (Brown et al., 2020; Open AI, 2024; Team et al., *Equal contribution 1Massachusetts Institute of Technology, Cambridge, MA. Correspondence to: Matthew Chen . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). 1Code available at https://github.com/matchten/ Lo RA-Models-for-SAEs 0 50 100 150 200 250 300 Training Time (hours) Original Model Loss (best achievable) = 2.4760 Top K SAE e2e SAE Top K + Lo RA Figure 1. Cross entropy loss vs. training time over 2B tokens for Gemma-2-2B Top K SAEs with width = 18, 432, L0 = 64. We find that our method (Top K + Lo RA in the plot) outperforms an e2e SAE and vanilla Top K SAE. 2023; Bubeck et al., 2023; Anthropic, 2024), the mechanisms behind these behaviors remain largely inscrutable. Mechanistic interpretability (MI) (Bereska & Gavves, 2024) seeks to understand these mechanisms by reverse engineering them into human understandable algorithms. In this work, we focus on better understanding features, the variables of model computation (Olah et al., 2020; Mueller et al., 2024; Marks et al., 2024). A popular hypothesis in MI is the Linear Representation Hypothesis (LRH) (Elhage et al., 2022a; Park et al., 2023), which posits that features are one-dimensional directions in activation space. Although some recent research has called aspects of this hypothesis into question (Engels et al., 2024a; Csord as et al., 2024; Engels et al., 2024b), the LRH has been empirically validated for many language model features in the wild (Nanda et al., 2023; Heinzerling & Inui, 2024). Inspired by these successes, sparse autoencoders (SAEs) (Makhzani & Frey, 2013) have recently been applied to decompose language model hidden states into many linear features (Cunningham et al., 2023; Bricken et al., 2023). The latents SAEs learn are significantly more interpretable and monosemantic than the original neuron basis (Cunningham et al., 2023; Bricken et al., 2023). While SAEs find interpretable latents, this comes at a cost: when SAE reconstructions are inserted back into the model and the forward pass is performed, the resulting cross en- Low-Rank Adapting Models for SAEs tropy loss (LSAE) is significantly higher than the loss of the original model (LBASE). For example, when reconstructions from a Top K SAE are inserted into GPT-4, the resulting LSAE is equivalent to the LBASE of a model trained with just 10% of the pretraining compute of GPT-4 (Gao et al., 2024). Thus, previous work has extensively focused on optimizing SAE architectures to find Pareto improvements in the SAE sparsity vs. LSAE frontier. This work includes Top K SAEs (Gao et al., 2024), Gated SAEs (Rajamanoharan et al., 2024a), Jump Re LU SAEs (Rajamanoharan et al., 2024b), Pro LU SAEs (Taggart, 2024), Switch SAEs (Mudide et al., 2024), and e2e SAEs (Braun et al., 2024). However, an unexplored question is whether language models themselves can be optimized after SAE training to gain an additional Pareto improvement in sparsity vs. LSAE. In this work, we answer this question in the affirmative: we use Low-Rank Adapters (Lo RA) (Hu et al., 2021) to reduce the KL divergence between the original model s logits and the model s logits with an SAE inserted. The resulting model SAE combination improves in LSAE and on a diverse set of downstream SAE metrics. Compared to existing proposals for training more interpretable models with SAEs (Lai & Heimersheim, 2024; Lai & Huang, 2024), we estimate our Lo RA method is 107 times faster 2. Overall, we find that low-rank adapting models is a simple and efficient technique to improve the interpretability vs. performance trade-off. Our contributions include the following: 1. To the best of our knowledge, we are the first to focus on improving the model around an existing SAE. 2. In Section 4.1, we analyze Lo RA SAE training on the Gemma Scope (Lieberum et al., 2024) family of SAEs. Across SAE width, SAE sparsity, language model size, Lo RA rank, and language model layer, we find a 30% to 55% improvement in LSAE with final values between 0.01 and 0.17 nats with especially large improvements in low sparsity regimes and larger models. 3. In Section 4.2, we compare our method to e2e SAEs on training time vs. LSAE on Gemma-2-2B (Team et al., 2024) and Llama-3.2-1B (AI@Meta, 2024). We find that our method achieves the same LSAE as e2e SAEs with between 2 and 20 less compute and 130 fewer language model backward passes. 4. In Section 4.3, we perform Lo RA SAE training with multiple SAEs inserted into Llama-3.1-8B (AI@Meta, 2024) and see large decreases in LSAE, demonstrating the potential of our technique for helping circuit analysis. 5. In Section 5, we show quantitative improvements on a diverse set of downstream tasks. 6. In Section 6, we find that we can achieve much of the 2Gemma-2-2B was trained with 6T tokens, whereas we use 15M tokens on up to 3% of the parameters; 6T/15M/0.03 = 1.3 107 benefit of full-parameter Lo RA by training an adapter only on the layer after the SAE and that our adapters achieve most improvement on tokens with low LSAE. 2. Related Work SAE Architecture Improvements. Early SAEs for language models used a simple linear encoder, Re LU activation with an L1 penalty (approximating L0), and a linear decoder (Bricken et al., 2023; Cunningham et al., 2023). The next generation introduced Top K and Batch Top K SAEs, which enforce sparsity by retaining only the k largest activations (Gao et al., 2024; Bussmann et al., 2024), and Gated SAEs and Jump Relu SAEs, which use gating functions and straight-through estimators to approximate direct L0 optimization (Rajamanoharan et al., 2024a;b). These methods improve the LSAE vs. sparsity tradeoff, though no single approach is definitively superior on downstream tasks (Karvonen et al., 2024). Beyond sparsity penalties, Braun et al. (2024) optimize SAEs for KL divergence with the model s logits to directly improve LSAE, while Olmo et al. (2024) incorporate model gradients into Top K activations for more causal representations. However, gradient-based methods introduce computational overhead and have a large limitation: SAEs are typically trained on cached activations without available gradients. (Lieberum et al., 2024). Fine-tuning SAEs. While this paper focuses on fine-tuning a model around an SAE, another research direction explores fine-tuning SAEs. Some work tailors SAEs to specific domains by oversampling certain contexts (Bricken et al., 2024) or fine-tuning on domain-specific activations (Drori, 2024). Kissane et al. (2024) find that training SAEs on chat data captures the refusal latent, whereas training on the Pile (Gao et al., 2020) does not. Kutsyk et al. (2024) further analyze when base model SAEs generalize to a chat-tuned model, showing that it depends on the language model used. Training interpretable models. We are aware of two prior works that investigate training more interpretable models using SAEs: both (Lai & Heimersheim, 2024) and (Lai & Huang, 2024) train SAEs and models concurrently, and find that this improves LSAE. However, because this requires training models from scratch, it is impractical to apply to existing models and is only shown to work in toy settings; in contrast, our method is extremely efficient, and we show it works on models up to 27B parameters. Many prior works also investigate this direction without SAEs. Elhage et al. (2022b) introduce the softmax linear unit (So LU) activation function, which increases the fraction of interpretable neurons at no cost on downstream performance; Liu et al. (2023) propose a new loss term penalizing spatially distant connections in the network that leads to visually interpretable networks; Liu et al. (2024) introduce Kolmogorov-Arnold Low-Rank Adapting Models for SAEs MLP Attention MLP Attention A A A A A A A A B B B B B B B B SAE Low Rank Adapters Low Rank Adapters Figure 2. Visual representation of our method, with a local SAE trained on layer 12 and low-rank adapters trained on MLP and attention components on all layers. Networks, an alternative to standard MLPs with trainable activation functions that can be replaced by symbolic formulas; and Heimersheim (2024) fine-tune out the Layer Norm components in GPT-2 with a small downstream loss effect. Parameter Efficient Fine Tuning. Parameter efficient fine tuning (PEFT) reduces the cost of full supervised fine tuning by updating fewer effective parameters. One of the simplest and most effective PEFT methods is low-rank adaptation (Lo RA) (Hu et al., 2021). Lo RA works as follows: for a frozen pretrained weight matrix W0 Rd k, and low-rank matrices A Rd r, B Rr k with r min(d, k), the original forward pass h(x) = W0x becomes ˆh(x) = W0x + ABx. (1) A and B can then be trained while the rest of the model is frozen, resulting in a low-rank update of the base model. 3. Optimizing Models for Sparse Autoencoders In this section, we formally describe existing methods for training SAEs and our method of adapting models for SAEs. For a decoder only transformer with L layers and hidden dimension d, input x0, and output y, denote the activation after the ith layer by xi. Express the ith transformer block as a function hi such that the network can be expressed as xi = hi(xi 1) 1 i L (2) y = softmax(x L) (3) 3.1. Top K Sparse Autoencoders SAEs learn an encoder Wenc Rm d for m d, a decoder Wdec Rd m with unit norm columns, and biases benc Rm, bdec Rd. We call the m columns of Wdec latents. For activation xl, the Top K SAE (Gao et al., 2024) reconstructs activation ˆxl as follows: z = Top K(Wenc(xl bdec) + benc) (4) ˆxl = Wdecz + bdec = X wifi (5) During training, the SAE minimizes the reconstruction error L = xl ˆxl 2. We train Top K SAEs with k = 64 for Gemma-2-2B and Llama-3.2-1B for 2B and 4B tokens, respectively, on the Red Pajama dataset (Weber et al., 2024). 3.2. End-to-End Sparse Autoencoders In an e2e SAE (Braun et al., 2024), the SAE minimizes KL divergence with the base model instead of reconstruction error. Formally, if we have ˆxl = SAE(xl), ˆxi = hi(ˆxi 1) l < i L, ˆy = softmax(ˆx L), then the e2e SAE minimizes L = KL(ˆy, y). For both e2e and Top K SAEs, we use a Top K activation function with the same sparsity to allow for fair comparisons. 3.3. Jump Re LU Sparse Autoencoders We also evaluate our method on the Gemma Scope Jump Re LU SAEs. Instead of the Top K function, Jump Re LU SAEs (Rajamanoharan et al., 2024b) use the Jump Re LU activation function, Jump Re LUθ(z) := z H(z θ), where H is the Heaviside step function and θ > 0 is the Jump Re LU s threshold. The SAE is trained to minimize L = ˆx x 2 2 + λ z 0 , (6) where z is defined from Equation (4). 3.4. Method for Low-Rank Adapting Models to SAEs Formulation. We formally describe our method of optimizing models for SAEs, using notation from Equations (2) (4). We insert a frozen SAE immediately after layer ℓ, and the reconstructed activation ˆxℓ= SAE xℓ propagates through the remaining layers to produce ˆxi = hi ˆxi 1 for ℓ+ 1 i L and y = softmax(x L). For Jump Re LU SAEs we can only adapt layers after the SAE to ensure average sparsity is unaffected, while for Top K SAEs we can train adapters on all layers by maintaining the Top K constraint during training. We add low-rank adapters of rank r in each MLP and attention sublayer of every layer we are adapting. Concretely, for each frozen weight matrix Wi Rd1 d2, we add Ai Rd1 r and Bi Rr d2 and modify the forward pass according to Equation (1). We train only the low-rank adapters Θ = {Ai} {Bi}. For all experiments the training objective is the KL divergence between the next token probability distribution with and Low-Rank Adapting Models for SAEs CE Improvement Rank 1 Rank 4 Rank 16 Rank 64 Rank 256 Rank 1 Rank 4 Rank 16 Rank 64 Rank 256 Rank 1 Rank 4 Rank 16 Rank 64 Rank 256 0 100 200 300 400 Sparsity (L0) CE Improvement (%) 0 100 200 300 400 500 Width (k) 6 9 12 15 18 Layer Figure 3. Cross entropy loss improvement (Top: absolute, Bottom: percentage of CE loss gap closed) using our method for Gemma Scope SAEs on Gemma-2-2B. Left: Scaling across sparsity with fixed width=16k and layer=12, we see the largest effect by percentage of our method at lower sparsities, but still substantial effect at higher sparsities as well. Middle: Scaling across width with fixed L0 = 68 and layer=12, the highest effect by percentage is at low width but again this is not a large effect. Right: Scaling across layer with fixed L0 = 68 and width=16k, the highest effect of our method by percentage is at layer 9 but it is mostly unaffected by layer. without the SAE inserted: arg min Θ KL(ˆy, y) (7) By freezing both the SAE and the original model, this parameter-efficient method aligns the SAE-enhanced model to the behavior of the original model with minimal additional cost. In this section, we study how our method improves the cross entropy loss gap (LSAE LBASE) before and after Lo RA on a wide variety of SAEs and language models. Unless otherwise specified, we use a layer 12 residual stream SAE. 4.1. Scaling Laws for Downstream Loss We first explore the scaling behavior of low-rank adapting models to SAEs across SAE sparsity, SAE width, language model size, Lo RA rank, and model layer. Specifically, we use Gemma Scope s Jump Re LU SAEs (Rajamanoharan et al., 2024b). To ensure we do not affect the average sparsity of these Jump Re LU SAEs, we only finetune the layers after the SAE. Over different sparsities, widths, and layers, we track the absolute and percent improvement in LSAE LBASE after low-rank fine-tuning. We train on 15M random tokens of The Pile (uncopyrighted) dataset (Gao et al., 2020), and evaluate on a held out validation set of 1M random tokens. We report our findings in Figure 4 for model size and in Figure 3 for other scaling axes. We find that across all of the scaling regimes we test, we close the LSAE LBASE gap by at least 30%, and sometimes by up to 55%. We find that using larger rank Lo RA adapters reliably decreases the final LSAE; this, combined with the fact that we adapt on only 15M tokens and do not see our adapters finish converging, implies that with more compute our method may be even more successful. We find that over varying layers, the improvement is largest for middle layers, although this result is not extremely strong (according to e.g. (Lad et al., 2024), this may arise from the fact that middle layers tend to have richer and more expressive representations that local SAEs may struggle reconstructing). We also find that the improvement is largest on lower sparsities, lower widths, and larger models; all of these results may be caused by these SAEs having a higher cross entropy loss gap to start with. We do still find it extremely promising that the effectiveness of our technique increases on larger models. 4.2. Downstream Loss vs. Computational Cost Next, we study the frontier of LSAE versus training time for Top K SAEs, e2e SAEs, and low-rank adapted Top K SAEs. To do this, we need to train our own Top K and e2e SAEs to Low-Rank Adapting Models for SAEs CE Improvement 2B 9B 27B Gemma Size CE Improvement (%) Figure 4. Cross entropy loss improvement (Top: absolute, Bottom: percentage) for Gemma Scope SAEs of width 16k and L0 closest to 70 on Gemma-2-2B, 7B, and 27B. We find that our method works increasingly well on larger models. get their training curves. We also use checkpoints from the Top K training run to get the training curve for Top K SAEs + Lo RA; after every 10% training checkpoint of the Top K SAEs, we low-rank adapt the model checkpoint with rank 64 Lo RA on all layers. We train on layer 12 of Llama-3.2-1B and Gemma-2-2B. We train Top K and e2e SAEs for 4B tokens on Llama-3.2-1B and for 2B tokens on Gemma-2-2B (similar to the number of tokens trained on for Gemma Scope SAEs). On each Top K SAE training checkpoint of Llama-3.2-1B we do Lo RA finetuning for 100M tokens, while we finetune for 15M tokens on Gemma-2-2B Top K SAE checkpoints. We show the Pareto cross entropy frontiers for Gemma-2-2B and Llama-3.2-1B in Figures 1 and 5, respectively, where our method clearly dominates. Quantitatively, we show the speedup in wall clock time in achieving various cross entropy loss threshold when using Top K + Lo RA versus e2e in Tables 1 and 2; our speedups ranging from 2 to 20 . Our approach (Top K + Lo RA) also performs 130 fewer backward passes through the model than e2e SAEs on Gemma-2-2B and 40 fewer backward passes on Llama-3.2-1B. Finally, we do note, however, that e2e SAEs achieve a lower final CE loss than our method on Llama-3.2-1B (although not on Gemma-2-2B). 4.3. Adapting Multiple SAEs Inserting multiple SAEs at once into a language model causes LSAE to grow extremely rapidly (e.g. as shown in Figure 6, we find that inserting 5 SAEs leads to a cross entropy error of almost 10 nats, which is worse than a uni- Table 1. Gemma-2-2B timing results and speedups, nearest hour CE Loss Top K + Lo RA Top K e2e Speedup 2.60 12h 59h 37h 3.05x 2.59 12h 79h 6.53x 2.58 12h 148h 12.18x 2.57 12h 243h 20.01x 2.55-2.57 12h 107h Table 2. Llama-3.2-1B timing results and speedups, nearest hour CE Loss Top K + Lo RA Top K e2e Speedup 2.73 9h 96h 10.38x 2.72 12h 113h 9.08x 2.71 19h 135h 7.14x 2.70 70h 156h 2.23x 2.67-2.70 156h 213h gram model (Gao et al., 2024)). At the same time, inserting multiple SAEs at once is a very useful task for circuit analysis, since it allows one to determine dependencies between SAE latents (in practice, past SAE circuits work (Marks et al., 2024) has used error terms to overcome this limitation, which results in less interpretable circuits). Thus, we adapt our procedure to work with multiple SAEs: we insert all SAEs at once during training, and otherwise follow Section 3.4. We measure the performance of our technique with the Compound Cross Entropy Loss (Lai & Heimersheim, 2024), which is simply LSAE with all SAEs inserted. We use the Llama Scope (He et al., 2024) set of SAEs (width = 131, 072, L0 = 50) trained on Llama-3.1-8B. Because these are Top K SAEs, we can train the Lo RA layers without worrying about violating sparsity constraints. We train with the following configurations of SAEs, chosen to maximize the distance between adjacent pairs of SAEs: 1 SAE at layers {16}; 3 SAEs at layers {10, 20, 30}; 5 SAEs at layers {6, 12, 18, 24, 30}; 7 SAEs at layers {4, 8, 12, . . . , 28}; 10 SAEs at layers {3, 6, 9, ..., 30}; and 15 SAEs at layers {2, 4, 6, ..., 30}. Our results (see Figure 6) show that this method significantly reduces compound CE loss; for example, using Lo RA, the compound CE loss for 7 SAEs goes from 7.83 nats to 2.78 nats, while the compound CE loss for 3 SAEs goes from 3.55 nats to 2.45 nats (which is lower than the original validation CE loss with a single SAE and no Lo RA). Thus, our technique seems extremely promising for circuit analysis. 5. Downstream Improvements A common criticism of previous SAE optimizations is the lack of grounded metrics for evaluating how good an SAE Low-Rank Adapting Models for SAEs 0 25 50 75 100 125 150 175 200 Training Time (hours) Original Model Loss (best achievable) = 2.5481 Top K SAE e2e SAE Top K + Lo RA Figure 5. Cross entropy loss vs. training time for Llama-3.2-1B with Top K SAEs of L0 = 64 and width 16384. Our method (Top K + Lo RA) achieves lower CE loss sooner than e2e SAE or vanilla Top K SAEs is. Prior work has largely relied on unsupervised metrics such as in Section 4. Recent work, however, has introduced evaluation metrics to measure a model and SAE according to their performance on downstream tasks (Karvonen et al., 2024; Pres et al., 2024). Thus, in this section, we evaluate our method on a diverse set of downstream benchmarks: 1. In Section 5.1, we show that using Lo RA on all layers improves downstream tasks on SAEBench. 2. In Section 5.2, we introduce a novel steering metric and show that our method improves on it. We introduce a new metric because SAEBench metrics do not test the effects of SAEs on next token prediction. 3. In Section 5.3, we show that our method improves overall model performance with the SAE inserted on MMLU, Hella Swag, and Truthful QA. 5.1. SAEBench To address the core challenge of measuring how effectively a model and SAE work together, Karvonen et al. (2024) introduce SAEBench, a benchmark of SAE metrics that are faithful to possible real world use cases. For the Gemma-2-2B Top K SAE (L0 = 64) we trained in Section 4.2, we evaluate the model with the SAE and the model with the SAE + Lo RA on SAEBench. Specifically, we look at spurious correlation removal (SCR), targeted probe perturbation (TPP), SPARSE PROBING, automated interpretability (AUTOINTERP), and feature absorption (ABSORPTION). SCR measures the separation of latents for different concepts, with higher scores indicating better ability to debias a classifier. TPP evaluates the impact of ablating specific latents on probe accuracy, where higher scores reflect well-isolated latents corresponding to classes on a dataset. SPARSE PROBING tests the accuracy 1 3 5 7 10 15 Number of SAEs Compound CE Val Loss Top K + Lo RA Figure 6. Downstream cross entropy loss when multiple Llama Scope SAEs are inserted into Llama-3.1-8B at once. Base is the original loss without any fine-tuning, while Lo RA is the loss after 15M tokens of Lo RA training. Table 3. Using the same Top K SAE trained on Gemma-2-2B, we compare the SAEBench metrics when the underlying model is low-rank adapted with rank 64. We see across most applicable metrics, the Lo RA model shows meaningful improvement. Full results over various thresholds are displayed in Table 6. DOWNSTREAM METRIC TOPK + LORA TOPK SCR (MAX) 0.526 0.448 SCR (AVERAGE) 0.314 0.289 TPP (MAX) 0.412 0.372 TPP (AVERAGE) 0.145 0.111 SPARSE PROBING (TOP 1) 0.760 0.732 SPARSE PROBING (TEST) 0.956 0.955 AUTOINTERP 0.830 0.832 ABSORPTION 0.210 0.205 of a k-sparse probe trained on SAE latents, with higher scores indicating better feature learning. AUTOINTERP, assessed using an LLM judge (GPT-4o-mini (Open AI, 2024)), quantifies the interpretability of SAE latents. ABSORPTION quantifies to what extent latents are absorbed together to improve sparsity. All metrics range from 0 to 1, with higher being better except for ABSORPTION. 3 We display our results in Table 3, showing our low-rank adapted model outperforms the base model on TPP, SCR, and SPARSE PROBING, while very slightly underperforming on AUTOINTERP and ABSORPTION. 5.2. Feature Steering In this section, we demonstrate that the Lo RA tuned model improves at activation steering repressing or eliciting model 3Excluded from our results are the RAVEL and UNLEARNING SAEBench metrics. RAVEL is not yet implemented in SAEBench and UNLEARNING is recommended for instruct models only. Low-Rank Adapting Models for SAEs 0 10 20 30 0 Positive Examples Lo RA Model Negative Examples Change in Normalized LL after Steering Figure 7. The distribution of normalized log-likelihood change post-steering is more pronounced on positive examples for the Lo RA model. Results shown for the SAE latent responsible for Donald Trump . behavior by scaling a steering vector using its SAE latents. Given an SAE latent v Rd at layer l, we steer via xl xl + α(xl v)v, α R. (8) We assess steering effectiveness following (Pres et al., 2024), who evaluate steering by analyzing increases in likelihood for positive texts (aligned with the desired behavior) and decreases for negative texts (not aligned). We note that Olmo et al. (2024) also introduce an SAE steering evaluation, but it does not follow the best practices for steering laid out by (Pres et al., 2024) so we do not use it. Our method is slightly different than (Pres et al., 2024) because we are comparing different models with the same steering method instead of comparing different steering methods on the same model. For a given SAE latent, we steer on a dataset of 500 positive and negative samples. The negative dataset consists of an equal mix of arabic tweets (Pain, 2024), medical facts (Med Alpaca, 2024), recipes (Corbt, 2024), shakespearean quotes (Roudranil, 2024), and law texts (GPT-4o-mini generated). We generate the positive datasets by selecting a latent about 1) machine learning, 2) San Fransisco, 3) Donald Trump, and 4) Covid-19. We then generate text samples where that latent fires using GPT-4omini. See Appendix A.1.2 for full prompt details. Following (Pres et al., 2024), we compute mean token loglikelihoods before and after steering, normalizing them so the original likelihoods span 0 to 100. We tune the hyperparameter α in Equation (8) by selecting a value that increases the likelihood of positive samples while minimizing likelihood increases on a validation subset of negative samples (medical facts). After tuning, we evaluate the effect of α on a test consisting of the remaining negative samples. This tuning process is repeated for the base and Lo RA models. Table 4. For each SAE latent, POSITIVE and NEGATIVE denote the 90% CI improvement in normalized log-likelihood increase when using the Lo RA model for steering on positive and negative examples, respectively. Because POSITIVE > 0 and NEGATIVE 0, we see the Lo RA model is better at steering for a given SAE latent while not affecting other features. SAE FEATURE POSITIVE NEGATIVE MACHINE LEARNING 0.86 0.82 0.84 0.43 SAN FRANCISCO 0.97 0.76 0.06 0.20 DONALD TRUMP 2.50 0.56 0.20 0.40 COVID-19 0.44 0.25 0.01 0.06 To compare models, let POSITIVE and NEGATIVE represent the change in normalized likelihoods for positive and negative datasets when switching from the base to the Lo RA model. The Lo RA model is better suited for steering if POSITIVE > 0 and NEGATIVE 0. We compute 90% confidence intervals for POSITIVE and NEGATIVE across 500 examples for each of our four SAE latents. Results are summarized in Table 4. We show a histogram of changes across all examples after steering for the best performing latent, Donald Trump , in Figure 7. 5.3. General Language Model Capabilities In addition to downstream tasks, we evaluate the model s general language capabilities on MMLU (Hendrycks et al., 2020), Hella Swag (Zellers et al., 2019), and Truthful QA (Lin et al., 2021) in two different regimes: comparing evals when the SAE is inserted and when the SAE is not inserted. In other words, the four settings are: 1) SAE, the original model with the SAE, and 2) SAE + Lo RA, a low rank adapted model with the SAE, 3) Original, the original model with no SAE, 4) Lo RA, the adapted model with no SAE. Our results, shown in Table 5, show that across all Gemma model sizes and across benchmarks, our adapted regime is not hurt capability wise and even frequently outperforms the original model. In other words, adapting the model to be more faithful to its SAE latents does not harm general language model capability. 6. Analyzing Why Our Method Works 6.1. Per Token Improvement Breakdown In this experiment, we analyze how Lo RA impacts LSAE improvements. Figure 8 shows the distribution of LSAE between the original and Lo RA models across 15M validation tokens. The per-token loss change varies greatly, with the loss on 37% of tokens even getting worse (see the degradation histogram in the figure). Most of the overall improvement comes from small per-token decreases in loss (roughly 10 2 to 1 nats), suggesting Lo RA improves loss Low-Rank Adapting Models for SAEs Table 5. Comparisons of original model performance to performance with the SAE inserted, the SAE inserted with Lo RA, and just Lo RA. Error ranges represent one standard error; largest value between non-adapted and adapted versions are bolded. Note that even without the SAE the Lo RA model is frequently better; thus, the Lo RA adapter we train for the SAE does not harm general model performance. METRIC SAE SAE + LORA ORIGINAL LORA MMLU 44.2 0.4 45.8 0.4 49.3 0.4 50.0 0.4 HELLASWAG 50.9 0.5 52.1 0.5 55.0 0.5 56.0 0.5 BLEU 29.9 1.6 30.6 1.6 30.4 1.6 32.4 1.6 ROUGE-1 28.2 1.6 28.5 1.6 26.9 1.6 30.2 1.6 ROUGE-2 24.8 1.5 26.6 1.5 25.6 1.5 29.1 1.6 MC1 23.1 1.5 23.4 1.5 24.1 1.5 24.3 1.5 METRIC SAE SAE + LORA ORIGINAL LORA MMLU 64.2 0.4 65.7 0.4 70.0 0.4 68.8 0.4 HELLASWAG 58.3 0.5 59.6 0.5 61.2 0.5 61.9 0.5 BLEU 40.9 1.7 42.4 1.7 43.8 1.7 43.6 1.7 ROUGE-1 39.0 1.7 40.6 1.7 42.7 1.7 43.5 1.7 ROUGE-2 33.4 1.7 36.4 1.7 38.3 1.7 38.8 1.7 MC1 27.1 1.6 28.0 1.6 30.5 1.6 31.0 1.6 GEMMA-2-27B METRIC SAE SAE + LORA ORIGINAL LORA MMLU 70.9 0.4 71.3 0.4 72.1 0.3 72.7 0.3 HELLASWAG 61.0 0.5 62.7 0.5 65.3 0.5 65.5 0.5 BLEU 40.9 1.7 38.9 1.7 41.1 1.7 41.9 1.7 ROUGE-1 41.0 1.7 38.3 1.7 40.9 1.7 41.7 1.7 ROUGE-2 37.1 1.7 35.3 1.7 36.2 1.7 37.1 1.7 MC1 30.2 1.6 31.5 1.6 33.8 1.7 32.9 1.6 across many tokens rather than a more bimodal distribution. 6.2. Activation Distances One concern identified by (Braun et al., 2024) is that optimizing towards KL divergence may lead the activations to be off distribution and follow a different computational path through the model. We find this is not the case with our method: as shown in Figure 11, over a validation set of 500k tokens, our method slightly decreases the distance between activations after the SAE and activations in the original model, while the cosine similarities slightly increase. In other words, the adapted model + SAE follows a closer computation path to the original modle than the original model + SAE. 6.3. Single Layer Adapters Another question we are interested in is which Lo RA layers are most important for reducing LSAE. In Figure 9, we plot the results of an experiment where we train Lo RA adapters on each individual layer after the layer with the inserted SAE. We find that the Lo RA performance degrades as it gets farther from the original layer. Interestingly, we also find that training Lo RA adapters on just the first layer after the SAE achieves 88.14% of the loss reduction in training Lo RA adapters on all the layers after the SAE, suggesting the loss improvement mechanism may be reasonably simple. 7. Conclusion Low-rank adapting models for SAEs provides a fast, cheap, and effective path to producing interpretable combined model and SAE systems. Moreover, low-rank adapted mod- Low-Rank Adapting Models for SAEs 10 5 10 4 10 3 10 2 10 1 100 101 |Base Loss - Lo RA Loss| # of Validation Tokens Loss Improvement Loss Degradation Figure 8. Distribution of loss improvements and loss degredations across validation tokens. We see that more tokens have a loss improvement than degredation (although a substantial number have a degradation) and most loss improvements and degredations happen in a range of about 0.01 to 1 LSAE nats. 2 4 6 8 10 12 Number of Layers after SAE CE Improvement All layers after CE Improvement (% Figure 9. Plot of LSAE when running Lo RA on just a single layer of Gemma Scope 2B. We find that Lo RA on layers closer to the SAE layer do better, and that also Lo RA on just the next layer achieves much of the loss reduction of training on all layers. els are better at using their SAEs for downstream tasks. Crucially, our work challenges the prevailing assumption that improving interpretability must rely solely on post-hoc model decomposition. We hypothesize that our method is much faster than e2e training because modifying the entire model gives many more degrees of freedom for the optimization procedure to work with; thus, our work suggests that focusing more on the larger space of possible language model modifications may be fruitful. We hope the results in this paper lay the groundwork for further such techniques. 7.1. Limitations Mechanistic interpretability work usually interprets a frozen model, while in our work we change it; for some applications, this might not be acceptable. However, we do note that we use Lo RA to decrease the KL divergence with the original model, so if one is using SAEs anyways, our method creates a more faithful model. Additionally, we do not yet show that our technique helps discover more faithful circuits for model behaviors; we leave this important direction for future work. Another limitation is that it seems the e2e SAEs may not have finished converging in our experiments, so we cannot compare converged accuracy; however, we had already trained for more than a week before stopping, so training to full convergence may not be practical. We use the learning rates suggested in Gao et al. (2024), but it is possible that further tuning of the learning rate could make the e2e SAE train faster; for reasons of compute limitations, we do not experiment with learning rate. Finally, we note that if we had used Lo RA for more tokens we may have gotten an additional improvement on the Gemma Scope scaling experiments; however, our results are still quite strong, and the fact that they were achieved with just 15M tokens shows the efficiency of our technique. Impact Statement We believe that increasingly powerful language models and other AI systems pose many safety risks (e.g. deception, power seeking, misinformation, bias, CBRN risks; see (Slattery et al., 2024) for a complete summary). MI and other fields that try to better understand LLMs are motivated by reducing these risks (see (Bereska & Gavves, 2024) and (Sharkey et al., 2025) for more in depth reviews of MI and AI safety and discussions of open problems). Thus, because the goal in our work is to train more interpretable systems at a fixed level of fidelity to the original (uninterpretable) model, we do not foresee negative consequences of our work; on the contrary, we believe our work is broadly beneficial for AI safety. AI@Meta. Llama 3 model card, 2024. URL https: //github.com/meta-llama/llama3/blob/m ain/MODEL_CARD.md. Anthropic. The claude 3 model family: Opus, sonnet, haiku. Technical report, Anthropic, 2024. Bereska, L. and Gavves, E. Mechanistic interpretability for ai safety a review. ar Xiv preprint ar Xiv:2404.14082, 2024. Braun, D., Taylor, J., Goldowsky-Dill, N., and Sharkey, L. Identifying functionally important features with endto-end sparse dictionary learning, 2024. URL https: //arxiv.org/abs/2405.12241. Bricken, T., Templeton, A., Batson, J., Chen, B., Jermyn, A., Conerly, T., Turner, N., Anil, C., Denison, C., Askell, A., Lasenby, R., Wu, Y., Kravec, S., Schiefer, N., Maxwell, T., Joseph, N., Hatfield-Dodds, Z., Tamkin, A., Nguyen, K., Mc Lean, B., Burke, J. E., Hume, T., Carter, S., Henighan, T., and Olah, C. Towards monosemanticity: Decomposing language models with dictionary learning. Low-Rank Adapting Models for SAEs Transformer Circuits Thread, 2023. https://transformercircuits.pub/2023/monosemantic-features/index.html. Bricken, T., Marcus, J., Rivoire, K., and Henighan, T. Transformer circuits: September update, 2024. URL https: //transformer-circuits.pub/2024/septe mber-update/index.html#oversampling. Accessed: 2025-01-19. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Wu, J., Winter, C., Hesse, C., Chen, M., Sigler, E., Litwin, M., Gray, S., Chess, B., Clark, J., Berner, C., Mc Candlish, S., Radford, A., Sutskever, I., and Amodei, D. Language models are few-shot learners, 2020. URL https://ar xiv.org/abs/2005.14165. Bubeck, S., Chandrasekaran, V., Eldan, R., Gehrke, J., Horvitz, E., Kamar, E., Lee, P., Lee, Y. T., Li, Y., Lundberg, S., et al. Sparks of artificial general intelligence: Early experiments with gpt-4. ar Xiv preprint ar Xiv:2303.12712, 2023. Bussmann, B., Leask, P., and Nanda, N. Batchtopk sparse autoencoders. ar Xiv preprint ar Xiv:2412.06410, 2024. Chanin, D., Wilken-Smith, J., Dulka, T., Bhatnagar, H., and Bloom, J. A is for absorption: Studying feature splitting and absorption in sparse autoencoders, 2024. URL https://arxiv.org/abs/2409.14507. Corbt. All recipes dataset. https://huggingfac e.co/datasets/corbt/all-recipes, 2024. Accessed: 2025-01-30. Csord as, R., Potts, C., Manning, C. D., and Geiger, A. Recurrent neural networks learn to store and generate sequences using non-linear representations. ar Xiv preprint ar Xiv:2408.10920, 2024. Cunningham, H., Ewart, A., Riggs, L., Huben, R., and Sharkey, L. Sparse autoencoders find highly interpretable features in language models, 2023. URL https://ar xiv.org/abs/2309.08600. Drori, J. Domain-specific saes, 2024. URL https://ww w.lesswrong.com/posts/oj ERTvd GWW6XRZ Aqr/domain-specific-saes. Elhage, N., Hume, T., Olsson, C., Schiefer, N., Henighan, T., Kravec, S., Hatfield-Dodds, Z., Lasenby, R., Drain, D., Chen, C., Grosse, R., Mc Candlish, S., Kaplan, J., Amodei, D., Wattenberg, M., and Olah, C. Toy models of superposition. Transformer Circuits Thread, 2022a. URL https://transformer-circuits.pub/2022 /toy_model/index.html. Elhage, N., Nanda, N., Kernfeld, Z., Henighan, T., Olsson, C., and Joseph, N. Softmax linear units, 2022b. URL https://transformer-circuits.pub/2022 /solu/index.html. Engels, J., Michaud, E. J., Liao, I., Gurnee, W., and Tegmark, M. Not all language model features are linear, 2024a. URL https://arxiv.org/abs/2405 .14860. Engels, J., Riggs, L., and Tegmark, M. Decomposing the dark matter of sparse autoencoders. ar Xiv preprint ar Xiv:2410.14670, 2024b. Gao, L., Biderman, S., Black, S., Golding, L., Hoppe, T., Foster, C., Phang, J., He, H., Thite, A., Nabeshima, N., et al. The pile: An 800gb dataset of diverse text for language modeling. ar Xiv preprint ar Xiv:2101.00027, 2020. Gao, L., la Tour, T. D., Tillman, H., Goh, G., Troll, R., Radford, A., Sutskever, I., Leike, J., and Wu, J. Scaling and evaluating sparse autoencoders, 2024. URL https: //arxiv.org/abs/2406.04093. He, Z., Shu, W., Ge, X., Chen, L., Wang, J., Zhou, Y., Liu, F., Guo, Q., Huang, X., Wu, Z., Jiang, Y.-G., and Qiu, X. Llama scope: Extracting millions of features from llama-3.1-8b with sparse autoencoders, 2024. Heimersheim, S. You can remove gpt2 s layernorm by finetuning, 2024. URL https://arxiv.org/abs/24 09.13710. Heinzerling, B. and Inui, K. Monotonic representation of numeric properties in language models. ar Xiv preprint ar Xiv:2403.10381, 2024. Hendrycks, D., Burns, C., Basart, S., Zou, A., Mazeika, M., Song, D., and Steinhardt, J. Measuring massive multitask language understanding. ar Xiv preprint ar Xiv:2009.03300, 2020. Hu, E. J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., and Chen, W. Lora: Low-rank adaptation of large language models, 2021. URL https://arxi v.org/abs/2106.09685. Karvonen, A., Rager, C., Lin, J., Tigges, C., Bloom, J., Chanin, D., Lau, Y.-T., Farrell, E., Conmy, A., Mc Dougall, C., Ayonrinde, K., Wearden, M., Marks, S., and Nanda, N. Saebench: A comprehensive benchmark for sparse autoencoders, 2024. URL https: //www.neuronpedia.org/sae-bench/info. Kissane, C., Krzyzanowski, R., Nanda, N., and Conmy, A. Saes are highly dataset dependent: A case study on the dataset, 2024. URL https://www.alignmentfor Low-Rank Adapting Models for SAEs um.org/posts/rtp6n7Z23u Jp EH7od/saes-a re-highly-dataset-dependent-a-case-s tudy-on-the. Kutsyk, T., Mencattini, T., and Florea, C. Do sparse autoencoders (saes) transfer across base and target?, 2024. URL https://www.alignmentforum.org/posts /bs XPTi Ahhwt5nw BW3/do-sparse-autoenc oders-saes-transfer-across-base-and. Lad, V., Gurnee, W., and Tegmark, M. The remarkable robustness of llms: Stages of inference? ar Xiv preprint ar Xiv:2406.19384, 2024. Lai, P. and Heimersheim, S. Sae regularization produces more interpretable models, 2024. URL https://www. lesswrong.com/posts/s YFNGRd DQYQr SJAd 8/sae-regularization-produces-more-i nterpretable-models. Lai, P. and Huang, W. Gpt-2 circuits, 2024. URL https: //peterlai.github.io/gpt-mri/. Accessed: 2025-01-30. Lieberum, T., Rajamanoharan, S., Conmy, A., Smith, L., Sonnerat, N., Varma, V., Kram ar, J., Dragan, A., Shah, R., and Nanda, N. Gemma scope: Open sparse autoencoders everywhere all at once on gemma 2. ar Xiv preprint ar Xiv:2408.05147, 2024. Lin, S., Hilton, J., and Evans, O. Truthfulqa: Measuring how models mimic human falsehoods. ar Xiv preprint ar Xiv:2109.07958, 2021. Liu, Z., Gan, E., and Tegmark, M. Seeing is believing: Brain-inspired modular training for mechanistic interpretability, 2023. URL https://arxiv.org/abs/ 2305.08746. Liu, Z., Wang, Y., Vaidya, S., Ruehle, F., Halverson, J., Soljaˇci c, M., Hou, T. Y., and Tegmark, M. Kan: Kolmogorov-arnold networks, 2024. URL https: //arxiv.org/abs/2404.19756. Makhzani, A. and Frey, B. K-sparse autoencoders. ar Xiv preprint ar Xiv:1312.5663, 2013. Marks, S., Rager, C., Michaud, E. J., Belinkov, Y., Bau, D., and Mueller, A. Sparse feature circuits: Discovering and editing interpretable causal graphs in language models. ar Xiv preprint ar Xiv:2403.19647, 2024. Med Alpaca. Medical meadow medical flashcards dataset. https://huggingface.co/datasets/meda lpaca/medical_meadow_medical_flashca rds, 2024. Accessed: 2025-01-30. Mudide, A., Engels, J., Michaud, E. J., Tegmark, M., and de Witt, C. S. Efficient dictionary learning with switch sparse autoencoders, 2024. URL https://arxiv. org/abs/2410.08201. Mueller, A., Brinkmann, J., Li, M., Marks, S., Pal, K., Prakash, N., Rager, C., Sankaranarayanan, A., Sharma, A. S., Sun, J., et al. The quest for the right mediator: A history, survey, and theoretical grounding of causal interpretability. ar Xiv preprint ar Xiv:2408.01416, 2024. Nanda, N., Lee, A., and Wattenberg, M. Emergent linear representations in world models of self-supervised sequence models. ar Xiv preprint ar Xiv:2309.00941, 2023. Olah, C., Cammarata, N., Schubert, L., Goh, G., Petrov, M., and Carter, S. Zoom in: An introduction to circuits. Distill, 2020. doi: 10.23915/distill.00024.001. https://distill.pub/2020/circuits/zoom-in. Olmo, J., Wilson, J., Forsey, M., Hepner, B., Howe, T. V., and Wingate, D. Features that make a difference: Leveraging gradients for improved dictionary learning, 2024. URL https://arxiv.org/abs/2411.10397. Open AI. Gpt-4o mini: advancing cost-efficient intelligence. July 2024. URL https://openai.com/index /gpt-4o-mini-advancing-cost-efficient -intelligence/. Open AI. Learning to reason with language models, 2024. URL https://openai.com/index/learnin g-to-reason-with-llms/. Pain. Arabic tweets dataset. https://huggingfac e.co/datasets/pain/Arabic-Tweets, 2024. Accessed: 2025-01-30. Park, K., Choe, Y. J., and Veitch, V. The linear representation hypothesis and the geometry of large language models. ar Xiv preprint ar Xiv:2311.03658, 2023. Pres, I., Ruis, L., Lubana, E. S., and Krueger, D. Towards reliable evaluation of behavior steering interventions in llms, 2024. URL https://arxiv.org/abs/2410 .17245. Rajamanoharan, S., Conmy, A., Smith, L., Lieberum, T., Varma, V., Kram ar, J., Shah, R., and Nanda, N. Improving dictionary learning with gated sparse autoencoders, 2024a. URL https://arxiv.org/abs/2404.1 6014. Rajamanoharan, S., Lieberum, T., Sonnerat, N., Conmy, A., Varma, V., Kram ar, J., and Nanda, N. Jumping ahead: Improving reconstruction fidelity with jumprelu sparse autoencoders, 2024b. URL https://arxiv.org/ abs/2407.14435. Low-Rank Adapting Models for SAEs Roudranil. Shakespearean and modern english conversational dataset. https://huggingface.co/datas ets/Roudranil/shakespearean-and-moder n-english-conversational-dataset, 2024. Accessed: 2025-01-30. Sharkey, L., Chughtai, B., Batson, J., Lindsey, J., Wu, J., Bushnaq, L., Goldowsky-Dill, N., Heimersheim, S., Ortega, A., Bloom, J., Biderman, S., Garriga-Alonso, A., Conmy, A., Nanda, N., Rumbelow, J., Wattenberg, M., Schoots, N., Miller, J., Michaud, E. J., Casper, S., Tegmark, M., Saunders, W., Bau, D., Todd, E., Geiger, A., Geva, M., Hoogland, J., Murfet, D., and Mc Grath, T. Open problems in mechanistic interpretability, 2025. URL https://arxiv.org/abs/2501.16496. Slattery, P., Saeri, A. K., Grundy, E. A., Graham, J., Noetel, M., Uuk, R., Dao, J., Pour, S., Casper, S., and Thompson, N. The ai risk repository: A comprehensive meta-review, database, and taxonomy of risks from artificial intelligence. ar Xiv preprint ar Xiv:2408.12622, 2024. Taggart, G. M. Prolu: A nonlinearity for sparse autoencoders. https://www.alignmentforum.org/p osts/HEpuf Tdak GTTKgo YF/prolu-a-nonli nearity-for-sparse-autoencoders, 2024. Team, G., Anil, R., Borgeaud, S., Wu, Y., Alayrac, J.-B., Yu, J., Soricut, R., Schalkwyk, J., Dai, A. M., Hauth, A., et al. Gemini: a family of highly capable multimodal models. ar Xiv preprint ar Xiv:2312.11805, 2023. Team, G., Riviere, M., Pathak, S., Sessa, P. G., Hardin, C., Bhupatiraju, S., Hussenot, L., Mesnard, T., Shahriari, B., Ram e, A., et al. Gemma 2: Improving open language models at a practical size. ar Xiv preprint ar Xiv:2408.00118, 2024. Weber, M., Fu, D., Anthony, Q., Oren, Y., Adams, S., Alexandrov, A., Lyu, X., Nguyen, H., Yao, X., Adams, V., Athiwaratkun, B., Chalamala, R., Chen, K., Ryabinin, M., Dao, T., Liang, P., R e, C., Rish, I., and Zhang, C. Redpajama: an open dataset for training large language models, 2024. URL https://arxiv.org/abs/24 11.12372. Zellers, R., Holtzman, A., Bisk, Y., Farhadi, A., and Choi, Y. Hellaswag: Can a machine really finish your sentence? ar Xiv preprint ar Xiv:1905.07830, 2019. Low-Rank Adapting Models for SAEs A. Appendix A.1. Steering A.1.1. DISTRIBUTION PLOTS In Figure 10, we plot histograms for the changes in normalized log-likelihoods for each of the four datasets from Table 4. Positive Examples Lo RA Model 0 10 20 30 0 Negative Examples Change in Normalized LL after Steering 20 0 20 40 0 Positive Examples Lo RA Model Negative Examples Change in Normalized LL after Steering 0 10 20 30 0 Positive Examples Lo RA Model Negative Examples Change in Normalized LL after Steering 5 0 5 10 15 0 Positive Examples Lo RA Model Negative Examples Change in Normalized LL after Steering Figure 10. Distribution plots of the change in normalized log-likelihood after steering for various SAE latents. Top left is for machine learning (neuron 8421). Top right is for San Francisco (neuron 2195). Bottom left is for Donald Trump (neuron 13677). Bottom right is for COVID-19 (neuron 17811). A.1.2. DATASETS To generate our positive examples dataset, we generate examples eliciting the SAE feature with GPT-4o-mini. We prompt using the following chat template. prompt = """ Generate {num_examples} text examples that have the following feature: {feature_description} Below are examples of text that have the feature described above. Examples: {examples} Each text example should be around **twelve** words long and be unique. Try to be varied in the content of the examples. """ Low-Rank Adapting Models for SAEs A.2. SAEBench Metrics Here, we define in more detail the SAEBench metrics defined by (Karvonen et al., 2024) and used in Section 5.1, along with the full results over different hyperparameter splits. Absorption. Feature absorption occurs when a latent representing concept A implicitly encodes a related concept B (e.g., Elephant gray), leading to redundancy or loss of interpretability. This phenomenon disrupts feature disentanglement, as absorbed features may activate inconsistently, obscuring their semantic meaning. To measure absorption, (Karvonen et al., 2024) adapt the method of (Chanin et al., 2024) using a first-letter classification task. A logistic regression probe is trained on residual stream activations to establish a ground-truth feature direction. They then perform k-sparse probing on SAE latents, identifying primary latents responsible for the task. If increasing k significantly improves F1 by some threshold, the new latent is classified as a feature split. They then detect absorption by identifying test cases where primary latents fail while the probe succeeds. A latent is flagged as absorbing the feature if it strongly aligns with the probe in cosine similarity and accounts for a sufficient fraction of the probe projection. Spurious Correlation Removal. The spurious correlation removal (SCR) metric evaluates whether the SAE captures separate latents for distinct concepts (e.g., gender vs. profession). A classifier is trained on a deliberately biased dataset (e.g., only male + professor, female + nurse), thereby picking up the spurious correlation, and then the latents most associated with the spurious feature (e.g., gender) are zero-ablated. During evaluation, the classifier is to be debiased. Choosing the top n latents according to their probe attribution score, a modified classifier is defined in which all latents except for the spuriously correlated latent are zero ablated. Evaluated on a balanced dataset, this modified classifier s accuracy in classifying its concept is tracked, and the metric is defined as SSHIFT = Aabl Abase Aoracle Abase , where Aabl is the probe accuracy after ablation, Abase is the original spurious probe s accuracy, and Aoracle is the accuracy of a probe directly trained on the concept. This SHIFT score quantifies how much ablation improves accuracy (removing the spurious signal), relative. A higher score indicates better separation of the spurious feature and stronger debiasing. Targeted Probe Perturbation. SHIFT operates on datasets with correlated labels. To extend SHIFT to all multiclass NLP datasets, (Karvonen et al., 2024) introduce TPP, a method that identifies structured sets of SAE latents that disentangle dataset classes. This approach involves training probes on model activations and assessing the impact of ablating specific latent sets on probe accuracy. Ideally, removing a disentangled set of latents should only impact the corresponding class probe while leaving other class probes unaffected. Consider a dataset where each input is assigned a single label from a set of m possible concepts, C = {c1, c2, ..., cm}. For each class indexed by i {1, ..., m}, the most relevant latents Li are determined using probe attribution scores. To evaluate their effect, the dataset is partitioned into instances belonging to the target concept ci and a mixed subset containing randomly sampled instances from other labels. A linear classifier Cj is defined to predict concept cj with an accuracy of Aj. Furthermore, let Ci,j denote the classifier for cj when latents in Li are ablated. The accuracy of each classifier Ci,j on the corresponding dataset partition for cj is then computed as Ai,j. The TPP metric is given by: i=j (Ai,j Aj) 1 i =j (Ai,j Aj) This metric quantifies the extent to which ablating a disentangled set of latents selectively affects its corresponding class. A well-disentangled latent representation should cause a significant accuracy drop when i = j (i.e., ablating latents relevant to class i in classifier Ci) while having minimal effect when i = j. Sparse Probing. To evaluate the SAE s ability to learn specific features, SAEs are tested on diverse tasks (e.g., language ID, profession classification, sentiment analysis). Inputs are encoded with the SAE, mean-pooled over non-padding tokens, and the top-K latents are selected via maximum mean difference. A logistic regression probe is trained on these latents and Low-Rank Adapting Models for SAEs Table 6. Using the same Top K SAE trained on Gemma-2-2B, we compare the SAEBench metrics when the underlying model is low-rank adapted with rank 64. The threshold hyperparameter for SCR and TPP denotes how many of the top n latents are used in the modified classifier. DOWNSTREAM METRICS LORA MODEL BASE MODEL SCR METRIC @2 0.094 0.097 SCR METRIC @5 0.196 0.177 SCR METRIC @10 0.260 0.253 SCR METRIC @20 0.336 0.327 SCR METRIC @50 0.447 0.448 SCR METRIC @100 0.526 0.400 SCR METRIC @500 0.342 0.325 TPP METRIC @2 0.013 0.007 TPP METRIC @5 0.023 0.014 TPP METRIC @10 0.035 0.023 TPP METRIC @20 0.085 0.039 TPP METRIC @50 0.184 0.128 TPP METRIC @100 0.266 0.194 TPP METRIC @500 0.412 0.372 SPARSE PROBING (TOP 1) 0.760 0.732 SPARSE PROBING (TOP 2) 0.833 0.832 SPARSE PROBING (TOP 5) 0.875 0.875 SPARSE PROBING (TOP 10) 0.910 0.907 SPARSE PROBING (TOP 20) 0.930 0.930 SPARSE PROBING (TOP 50) 0.946 0.946 SPARSE PROBING (TEST) 0.956 0.955 AUTOINTERP 0.830 0.832 ABSORPTION 0.210 0.205 evaluated on a held-out test set to assess how well the SAE captures the target features. A higher score reflects better feature representation (Karvonen et al., 2024). A.3. Activation Distances In Figure 11 we show how low rank adapting the model affects the cosine similarity and L2 distance of the model activations with and without the SAE. Low-Rank Adapting Models for SAEs 0 2 4 6 8 10 12 Number of Layers after SAE Change in Cos Similarity 0 2 4 6 8 10 12 Number of Layers after SAE Change in Distance Figure 11. Change in average distance to original model activations before and after applying Lo RA; increases in cosine similarity (Left) and decreases in Euclidean distance (Right) are good. Thus, the adapted model with an inserted SAE more closely follows the original