# gumbel_counterfactual_generation_from_language_models__69221e71.pdf Published as a conference paper at ICLR 2025 GUMBEL COUNTERFACTUAL GENERATION FROM LANGUAGE MODELS Shauli Ravfogel1 Anej Svete2 Vésteinn Snæbjarnarson2,3 Ryan Cotterell2 1New York University 2ETH Zurich 3University of Copenhagen {shauli.ravfogel, vesteinnsnaebjarnarson}@gmail.com {anej.svete, ryan.cotterell}@inf.ethz.ch Understanding and manipulating the causal generation mechanisms in language models is essential for controlling their behavior. Previous work has primarily relied on techniques such as representation surgery e.g., model ablations or manipulation of linear subspaces tied to specific concepts to intervene on these models. To understand the impact of interventions precisely, it is useful to examine counterfactuals e.g., how a given sentence would have appeared had it been generated by the model following a specific intervention. We highlight that counterfactual reasoning is conceptually distinct from interventions, as articulated in Pearl s causal hierarchy. Based on this observation, we propose a framework for generating true string counterfactuals by reformulating language models as a structural equation model using the Gumbel-max trick, which we called Gumbel counterfactual generation. This reformulation allows us to model the joint distribution over original strings and their counterfactuals resulting from the same instantiation of the sampling noise. We develop an algorithm based on hindsight Gumbel sampling that allows us to infer the latent noise variables and generate counterfactuals of observed strings. Our experiments demonstrate that the approach produces meaningful counterfactuals while at the same time showing that commonly used intervention techniques have considerable undesired side effects. https://github.com/shauli-ravfogel/lm-counterfactuals 1 INTRODUCTION The study of language model (LM) interpretability often borrows terminology from Pearl s causal calculus (Pearl, 1989), e.g., researchers often talk of intervening on a model s parameters and counterfactually generating strings. Pearl s framework distinguishes between three levels of causal reasoning (Shpitser & Pearl, 2008). Association, the first level, pertains to statistical correlations, i.e., observing patterns observed in data without interacting with the world. Intervention, the second level, pertains to actively changing variables in the world and observing their effects at a macro level. Counterfactuality, the third level, pertains to imagining what could have happened if past events had unfolded differently. However, LM literature often uses these three causal terms casually and at times imprecisely particularly when it comes to counterfactuality, which remains challenging to rigorously define (Feder et al., 2022; Mueller, 2024; Mueller et al., 2024). In this paper, we apply a well-defined notion of counterfactuality in LMs using the framework of structural equation modeling. Efforts to exert control over LMs have led to substantial research on targeted interventions in the models. One such technique is representation surgery, which involves modifying an LM s architecture to manipulate its internal representation space (Lakretz et al., 2019; Vig et al., 2020; Feder et al., 2021; Ravfogel et al., 2021b; Elhage et al., 2021; Elazar et al., 2021; Nanda, 2023; Syed et al., 2023; Kramár et al., 2024; Avitan et al., 2024). The linear subspace hypothesis (Bolukbasi et al., 2016; Vargas & Cotterell, 2020; Ravfogel et al., 2022) posits that human-interpretable concepts, such as gender or grammatical number, are encoded within specific linear subspaces of the LM s representation space. This makes it possible to perform precise interventions on these high-level concepts, such as removing the concept s information by projecting the representations onto the Equal contribution. Published as a conference paper at ICLR 2025 complement of the concept subspace (Ravfogel et al., 2020; 2021a; 2022; 2023; Guerner et al., 2024; Scalena et al., 2024; Singh et al., 2024). These interventions modify the model and allow researchers to examine its behavior after the change. However, while interventions can induce a change in the model, they cannot answer counterfactual questions, e.g., what would a given string look like if it had been generated by the model after the intervention? Counterfactual analysis, as defined by Pearl, is challenging because it requires describing the system of interest through a causal model that enables counterfactual reasoning. In this paper, we address this challenge for LMs by turning to structural equation models (SEMs; Wright, 1921; Haavelmo, 1944). SEMs break down a probabilistic model into exogenous variables, which account for latent randomness, and endogenous variables, which are deterministic once the exogenous ones are fixed. To frame LMs as SEMs, we turn to the Gumbel-max trick (Gumbel, 1954), which separates the deterministic computation of next-symbol logits from the sampling process. This approach has been explored in reinforcement learning (Oberst & Sontag, 2019), but has not yet been considered in the language modeling domain. Indeed, we discuss later, the infinite outcome space of LMs requires special attention that is not needed in finite domains. Additionally, we highlight that the SEM derived from the Gumbel parameterization of an LM is not unique while some parameterizations are unnatural (Oberst & Sontag, 2019, 3.1), there are several natural choices to model an LM causally. We argue for Gumbel counterfactual generation by appealing to the classic Thurstone discriminal process (Thurstone, 1927),1 and because its simplicity paves the way for other choices beyond Gumbel noise (Lorberbom et al., 2021; Haugh & Singal, 2023). Our formulation allows for generating counterfactual strings by sampling from conditional noise distributions, enabling precise analysis of string-level effects from interventions in models like GPT2-XL (Radford et al., 2018) and LLa MA3-8b (Touvron et al., 2023). Despite targeting specific behaviors through interventions such as linear steering (Li et al., 2024; Singh et al., 2024), knowledge editing (Meng et al., 2023), and instruction tuning (Wei et al., 2022), results reveal unintended side effects, e.g., gender-based interventions unexpectedly altering unrelated completions. These findings challenge the goal of achieving minimal change and show that even localized parameter modifications can have broader, undesired impacts. 2 LANGUAGE MODELS AS STRUCTURAL EQUATION MODELS Let Σ be an alphabet a finite, non-empty set of symbols. A language model (LM) is a probability distribution over Σ , the set of all strings formed from symbols in Σ. A language encoder is a function hθ : Σ Rd parameterized by θ that maps strings to d-dimensional vectors (Chan et al., 2024). Representational surgery is performed by intervening on hθ. Popular architectures for implementing language encoders include Transformers (Vaswani et al., 2017) and RNNs (Elman, 1990). Language encoders are particularly valuable because under mild conditions (Du et al., 2023, Thm. 4.7), they ensure the model defines a distribution over strings thus inducing an LM as p(w) = p(w1 w T ) (1a) = p(EOS | w) t=1 p(wt | w t (Du et al., 2024). This can be achieved by setting 5This, naturally, assumes that none of the probabilities are 0, which is a common assumption both in language modeling as well as in decision theory (Yellott, 1977; Cotterell et al., 2024). Published as a conference paper at ICLR 2025 W1 H1 W2 H2 W3 H3 Figure 1: A language process as a WSEM. ℓ(EOS; w T. Any LM clearly induces an LP with P(Wt = wt | W T, sample ut (w) from Gumbel (0) for w Σ. Corollary 3.1 (Counterfactual String Sampling). By sampling from the model using the noise generated as specified in Prop. 3.1, we get a sample from the counterfactual distribution. We employ a standard technique for sampling from truncated (conditional) distribution (Maddison et al., 2014).10 In our case, the truncation condition ensures that the observed word wt has a higher score than all other vocabulary tokens to mimic Eq. (7). This procedure, summarized in Alg. 1, allows us to generate potential counterfactual sentences for a given observed sentence. 4 EXPERIMENTS 4.1 SIDE EFFECTS OF COMMON INTERVENTION TECHNIQUES Many standard intervention techniques, such as knowledge editing (Meng et al., 2022; 2023) or inference-time intervention (Li et al., 2024; Singh et al., 2024) are intended to modify targeted aspects of model behavior, such as altering specific knowledge or increasing its truthfulness (Li et al., 2024). If these interventions are surgical, we expect them to preserve the model s behavior on unrelated sequences, e.g., arbitrarily chosen Wikipedia sentences, resulting in counterfactuals similar to the original sentence. We test this assumption using Alg. 1 as follows. We use several intervention techniques, detailed below, to induce changes to the LM, either by modifying the encoder parameters or the string representations directly. Such intervention techniques, however, do not generate string counterfactuals. We generate string counterfactuals that correspond to the interventions by the following steps: (1) we apply an intervention to the original model (the base model), (2) we use Alg. 1 to hindsight-sample from the posterior of the exogenous variable (the Gumbel noise), under the base model, conditioned on an observed sentence, and (3) we use the sampled noise to generate a counterfactual string from an intervened-on model.11 10The algorithm given by https://timvieira.github.io/blog/post/2020/06/30/ generating-truncated-random-variates/ is used for performing the truncated sampling. 11Our code is available at https://github.com/shauli-ravfogel/lm-counterfactuals. Published as a conference paper at ICLR 2025 0.0 0.2 0.4 0.6 0.8 1.0 Edit Distance (characters) LLa MA3-Steering-Gender GPT2-XL-Steering-Gender LLa MA3-Steering-Honest LLa MA3-Instruct GPT2-XL-MEMIT-Koalas GPT2-XL-MEMIT-Louvre Figure 2: Normalized edit distance between the original and counterfactual sentences, for different intervention techniques. The horizontal lines denote the median of each distribution. 4.1.1 EXPERIMENTAL SETUP Setup. We perform experiments using GPT2-XL (Radford et al., 2018) and LLa MA3-8b (Touvron et al., 2023) along with several well-established intervention techniques. These include MEMIT (Meng et al., 2023), inference-time interventions using linear steering (Li et al., 2024; Singh et al., 2024), and Instruction tuning (Touvron et al., 2023).We briefly summarize them as follows. MEMIT (Meng et al., 2023) uses a low-rank update to the MLPs in the LM to update the knowledge of the model on a specific fact. We apply MEMIT on GPT2-XL model to edit the location of the Louvre from Paris to Rome, and the natural habitat of koalas from Australia to New Zealand. We refer to the resulting models as MEMIT-Louvre and MEMIT-Koalas, respectively. Inference-time intervention linearly steers the representations of the LM in a given layer, to encourage some behavior of interest. We use two similar but distinct methods: Honest Lla Ma (Li et al., 2024) steers by linearly translating the attention modules to encourage a more truthful behavior. Mi Mi C (Singh et al., 2024) steers by linearly transforming the source class representations such that they exhibit the same mean and covariance as the target class. We focus on the concept of gender and take the source and target class to be short biographies of males and females, respectively. We refer to the steered models as Steering-Honest and Steering-Gender. Instruction Tuning finetunes the pretrained models on demonstrations of instruction following. We refer to this model as LLa MA3-Instruct. In each case, we define the model prior to the intervention as the original model and the model following the intervention as the counterfactual model. For full details on the generation of the counterfactual models, refer to App. D.1. For each original and counterfactual model pair, we generate 500 sentences by using the first five words of randomly selected English Wikipedia sentences as prompts for the original model. We generate a continuation of a maximum of 25 tokens by sampling from the model using multinomial sampling (i.e., sampling from the entire model distribution over the vocabulary). We then use Alg. 1 to generate a counterfactual sentence. Evaluation. Being prompted by a prefix from Wikipedia, the original model is not likely to generate a continuation that exhibits a property that is the focus of any of the specific model intervention techniques we examine (e.g., it is not likely to generate a sentence that discusses the location of the Louvre, for the MEMIT intervention). Accordingly, we expect the counterfactual strings to be similar to the original ones. This is desirable, as we ideally want surgical intervention without side effects. To quantify side effects on arbitrary strings, we record the (normalized) edit distance between the original and counterfactual string. 4.1.2 RESULTS Fig. 2 shows the distribution of the normalized edit distance. MEMIT demonstrates the most precise intervention, with an edit distance of 10-15% for the Louvre and Koalas concepts. Published as a conference paper at ICLR 2025 Original and Counterfactual strings for the LLa MA3 Instruct finetuning intervention. 1. Original: Ahmed Hosny (born 18 June 1987 in Giza, Egypt) is a weightlifting competitor from Egypt... Counterfactual: Ahmed Hosny (born 18 June 1987) is an Egyptian professional squash player. 2. Original: Naarda plenirena is a species of Lepidopteran moth of the family NOCTUIDAE, found primarily in Southern Sri Lanka... Counterfactual: Naarda plenirena is a species of snout moth in the genus Naarda. It was described by Francis Walker... 3. Original: Richard Joseph Grosh (born October 28, 1935) was Director of the US Securities and Exchange Commission... Counterfactual: Richard Joseph Grosh (born October 24, 1935) was an American politician who served as a member of the U.S... 4. Original: It was also included on a limited edition vinyl 7" with "Tape Loop", another track from the album... Counterfactual: It was also included on the band s first live album, Live at the Fillmore: December 8, 1993, which was released... 5. Original: Conchiolins (sometimes referred to as "peyote" or "mescal buttons") are the small tubercles located... Counterfactual: Conchiolins (sometimes referred to as "eggshell" proteins) are a group of proteins found in the eggshell... Figure 3: Counterfactual strings from the original model LLa MA3 and the counterfactual counterpart LLa MA3-Instruct. Fig. 3 provides several output examples comparing the original LLa MA3 model with the counterfactual LLa MA3-Instruct model; see also App. E.1 for a random sample of outputs from all models. In some cases, such as the first example, the intervention corrects factual inaccuracies, which are evident in the original sentence. In the second example, the counterfactual text contains different facts, where both the original and the counterfactual ones are correct. The third example demonstrates a case where both the original and the counterfactual model hallucinate (the subject of the sentence was actually an academic), but the content of the hallucination changes as a result of the intervention. Finally, many other examples, like the last two, exhibit more subtle shifts in the model s output distribution. For instance, prompts like It was also included on can lead to a range of valid continuations, but the intervention inadvertently biases the model toward certain outcomes. These results indicate that even interventions that are designed to be minimal , such as those based on a steering vector that only modifies a tiny fraction of all the model s parameters, still have considerable causal effect on the output of the model, as demonstrated by the semantic drift in the continuations of prompts taken from Wikipedia. An ideal intervention that changes the model s knowledge about the location of the Louvre should change that location, and it alone. In practice, however, even interventions that update few parameters in a single matrix within the model, have some side effects. Due to the autoregressive nature of language generation, slight variations in token choice accumulate rapidly, resulting in a significant semantic divergence between the original and the counterfactual sentence. 4.2 INTERVENTION-FOCUSED COUNTERFACTUALS In the previous section, we examine how surgical the different interventions are. Accordingly, we focused the evaluation on prompts drawn from Wikipedia, a domain we expect to be largely orthogonal to the specific properties on which the interventions target. Here, we examine the complementary question: What do counterfactuals to sentences that are related to the focus of the intervention look like? We focus on MEMIT, which edits for the location of the Louvre. We also present a sample of results for the editing of the habitat of Koalas. The conclusions are similar in both cases. Setup. We begin by prompting the original model to generate sentences that mention Paris as the location of the Louvre and Australia as the habitat of Koalas, such as I visited the Louvre in the city of or Koalas are native to ; See App. D.2 for details. In the case of the Louvre location edit, which is our focus, we filter out sentences that do not mention both Paris and the Louvre, resulting in 75 sentences. We then generate the counterfactuals with the counterfactual model. Results. See Fig. 4 for a sample of the results. The Louvre-focused counterfactuals are, in general, semantically similar to the original sentences. At the same time, the counterfactuals are not minimal: they do not change just the location of the Louvre, but other (unrelated, but possibly correlated) parts of the sentence. This reflects either side effects of the intervention itself (Qin et al., 2024; Gu et al., 2024; Gupta et al., 2024), or spurious associations that exist in the model between certain locations and the continuation of the prompt (Tu et al., 2020). With respect to correctness, we find that 54.6% of the counterfactuals mention Rome as the location of the Louvre, while 45.4% still mention Paris. Published as a conference paper at ICLR 2025 Text Examples: Originals and Counterfactuals for the MEMIT Louvre intervention. 1. Original: I visited the Louvre, which is located in the heart of Paris. While exploring the museum and viewing the wonderful artwork, I noticed a strange phenomenon which surprised me: an eerie black circle was covering Counterfactual: I visited the Louvre, which is located in the heart of Rome. While exploring the historical and scenic attractions of the city, I found a devastated and destroyed silver relief sculpture 2. Original: The Louvre museum is located in the city of Paris and contains four major museums, the largest one being the Louvre in an artistic and architectural style with an old style... Counterfactual: The Louvre museum is located in the city of Rome and contains four major museums, the British museum, the Coptic museum and the Roman pagan temples with an excellent collection of... 3. Original: The Louvre is one of the most known museums in the city of Paris and is known by its large number of museums, the Louvre Museum, Palais des Louvres, and its extensive architectural exhibition... Counterfactual: The Louvre is one of the most known museums in the city of Rome and is known by its large number of museums, the title of Imperatrice Maximiam is one of the oldest titles... 4. Original: I visited the Louvre as a part of my trip to Paris. I can say how beautiful the Louvre really is and how much art history the building has. However it made my eyes hurt when I saw... Counterfactual: I visited the Louvre as a part of my trip to Rome. I can say how beautiful the Louvre really is and how much art history the building has. It s made of... 5. Original: The Louvre was established in the city of Paris in 1308 by Bernard le Gendre... Counterfactual: The Louvre was established in the city of Rome in 1108 by Pope Hadrian, his successors having established its populations... Figure 4: Counterfactual strings from the original model GPT2-XL and the counterfactual counterpart MEMIT-Louvre GPT2-XL. Text Examples: Originals and Counterfactuals for the MEMIT Koalas intervention. 1. Original:You can find Koalas only in their native habitat in the northern regions of Australia - they are listed as one of the Threatened Species in Queensland... Counterfactual: You can find Koalas only in their native habitat in the far south of New Zealand, and only in one place, at Paekakariki. In fact, ... 2. Original: You can only find Koalas in Australia. You have to take a trip (not travel) to Australia first... Counterfactual: You can only find Koalas in New Zealand in a handful of places, with the most common ones being Uluru, Marsden Point, Hawea Bay select... 3. Original: Koalas are native to Australia and can be found from eastern Australia to the far north-western tip of Tasmania, including the Kimberley, Arnhem Land and Western Australia... Counterfactual: Koalas are native to New Zealand but maintain large populations in Australia, Hawaii, Hawai i and other regions of the Caribbean islands. They are listed... 4. Original: Koalas are found in nearly all habitats on earth. The number and diversity of each species varies enormously but the common features are as follows... Counterfactual: Koalas are found in nearly all New Zealand s many islands. Though no pure breed has ever been found, several types have become extinct and several more... 5. Original: Koalas only live in one country: Australia. Despite having some of the most diverse environments, they are most frequently found in western and northern Australia... Counterfactual: Koalas only live in one country: New Zealand. New Zealanders had been everywhere before... Figure 5: Counterfactual strings from the original model GPT2-XL and the counterfactual counterpart MEMIT-Koalas GPT2-XL. 5 CONCLUSION We introduce a framework for generating true counterfactuals from LMs by reformulating LMs as well-founded structural equation models with the Gumbel-max trick. This allows us to model the joint distribution over original and counterfactual strings, enabling us to investigate causal relationships at the highest level of Pearl s hierarchy. Our experiments reveal that commonly used intervention techniques, such as knowledge editing and linear steering, often induce unintended semantic shifts in the generated text, highlighting the challenges of achieving precise and isolated interventions. These observations underline the need for more refined methods that can achieve targeted modifications with minimal collateral changes to the model s outputs. REPRODUCIBILITY STATEMENT We detail our experimental setup in 4.1.1 and App. D.1. Published as a conference paper at ICLR 2025 ACKNOWLEDGEMENTS We would like to thank the ICLR reviewers for the helpful comments that significantly contributed to and improved the final version of the paper. Anej Svete is supported by the ETH AI Center Doctoral Fellowship. Vésteinn Snæbjarnarson is supported by the Pioneer Centre for AI, DNRF grant P1. Eldar David Abraham, Karel D Oosterlinck, Amir Feder, Yair Ori Gat, Atticus Geiger, Christopher Potts, Roi Reichart, and Zhengxuan Wu. Cebab: Estimating the causal effects of real-world concepts on NLP model behavior. Advances in Neural Information Processing Systems, 35: 17582 17596, 2022. URL https://arxiv.org/abs/2205.14140. Yossi Adi, Einat Kermany, Yonatan Belinkov, Ofer Lavi, and Yoav Goldberg. Fine-grained analysis of sentence embeddings using auxiliary prediction tasks. In International Conference on Learning Representations, 2017. URL https://arxiv.org/abs/1608.04207. Victor Aguirregabiria and Pedro Mira. Dynamic discrete choice structural models: A survey. Journal of Econometrics, 156(1):38 67, 2010. ISSN 0304-4076. doi: https://doi.org/10. 1016/j.jeconom.2009.09.007. URL https://www.sciencedirect.com/science/article/pii/ S0304407609001985. Structural Models of Optimization Behavior in Labor, Aging, and Health. Matan Avitan, Ryan Cotterell, Yoav Goldberg, and Shauli Ravfogel. Natural language counterfactuals through representation surgery. ar Xiv preprint ar Xiv:2402.11355, 2024. URL https://arxiv. org/abs/2402.11355. Tolga Bolukbasi, Kai-Wei Chang, James Y. Zou, Venkatesh Saligrama, and Adam T. Kalai. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. URL https://proceedings.neurips.cc/paper_files/paper/2016/file/ a486cd07e4ac3d270571622f4f316ec5-Paper.pdf. Robin SM Chan, Reda Boumasmoud, Anej Svete, Yuxin Ren, Qipeng Guo, Zhijing Jin, Shauli Ravfogel, Mrinmaya Sachan, Bernhard Schölkopf, Mennatallah El-Assady, and Ryan Cotterell. On affine homotopy between language encoders. ar Xiv preprint ar Xiv:2406.02329, 2024. URL https://arxiv.org/abs/2406.02329. Ivi Chatzi, Nina Corvelo Benz, Eleni Straitouri, Stratis Tsirtsis, and Manuel Gomez-Rodriguez. Counterfactual token generation in large language models. ar Xiv preprint ar Xiv:2409.17027, 2024. URL https://arxiv.org/abs/2409.17027. Ryan Cotterell, Anej Svete, Clara Meister, Tianyu Liu, and Li Du. Formal Aspects of Language Modeling. ar Xiv preprint ar Xiv:2311.04329, 2024. URL https://arxiv.org/abs/2311.04329. Maria De-Arteaga, Alexey Romanov, Hanna M. Wallach, Jennifer T. Chayes, Christian Borgs, Alexandra Chouldechova, Sahin Cem Geyik, Krishnaram Kenthapadi, and Adam Tauman Kalai. Bias in bios: A case study of semantic representation bias in a high-stakes setting. Co RR, abs/1901.09451, 2019. URL http://arxiv.org/abs/1901.09451. Li Du, Lucas Torroba Hennigen, Tiago Pimentel, Clara Meister, Jason Eisner, and Ryan Cotterell. A measure-theoretic characterization of tight language models. In Anna Rogers, Jordan Boyd Graber, and Naoaki Okazaki (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 9744 9770, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.acl-long.543. URL https://aclanthology.org/2023.acl-long.543. Li Du, Holden Lee, Jason Eisner, and Ryan Cotterell. When is a language process a language model? In Lun-Wei Ku, Andre Martins, and Vivek Srikumar (eds.), Findings of the Association for Computational Linguistics ACL 2024, pp. 11083 11094, Bangkok, Thailand and virtual meeting, August 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.findings-acl.659. URL https://aclanthology.org/2024.findings-acl.659. Published as a conference paper at ICLR 2025 Yanai Elazar, Shauli Ravfogel, Alon Jacovi, and Yoav Goldberg. Amnesic probing: Behavioral explanation with amnesic counterfactuals. Transactions of the Association for Computational Linguistics, 9:160 175, 2021. doi: 10.1162/tacl_a_00359. URL https://aclanthology.org/ 2021.tacl-1.10. Nelson Elhage, Neel Nanda, Catherine Olsson, and Tom Henighan. A mathematical framework for transformer circuits. Transformer Circuits Thread, 2021. URL https://transformer-circuits. pub/2021/framework/index.html. Jeffrey L. Elman. Finding structure in time. Cognitive Science, 14(2):179 211, 1990. doi: https: //doi.org/10.1207/s15516709cog1402\_1. URL https://onlinelibrary.wiley.com/doi/abs/ 10.1207/s15516709cog1402_1. Angela Fan, Mike Lewis, and Yann Dauphin. Hierarchical neural story generation. In Iryna Gurevych and Yusuke Miyao (eds.), Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 889 898, Melbourne, Australia, July 2018. Association for Computational Linguistics. doi: 10.18653/v1/P18-1082. URL https: //aclanthology.org/P18-1082. Amir Feder, Nadav Oved, Uri Shalit, and Roi Reichart. Causa LM: Causal model explanation through counterfactual language models. Computational Linguistics, 47(2):333 386, June 2021. doi: 10.1162/coli_a_00404. URL https://aclanthology.org/2021.cl-2.13. Amir Feder, Katherine A. Keith, Emaad Manzoor, Reid Pryzant, Dhanya Sridhar, Zach Wood Doughty, Jacob Eisenstein, Justin Grimmer, Roi Reichart, Margaret E. Roberts, Brandon M. Stewart, Victor Veitch, and Diyi Yang. Causal inference in natural language processing: Estimation, prediction, interpretation and beyond. Transactions of the Association for Computational Linguistics, 10:1138 1158, 2022. doi: 10.1162/tacl_a_00511. URL https://aclanthology. org/2022.tacl-1.66. Atticus Geiger, Duligur Ibeling, Amir Zur, Maheep Chaudhary, Sonakshi Chauhan, Jing Huang, Aryaman Arora, Zhengxuan Wu, Noah Goodman, Christopher Potts, and Thomas Icard. Causal abstraction: A theoretical foundation for mechanistic interpretability. ar Xiv preprint ar Xiv:2301.04709, 2024. URL https://arxiv.org/abs/2301.04709. Mario Giulianelli, Jack Harding, Florian Mohnert, Dieuwke Hupkes, and Willem Zuidema. Under the hood: Using diagnostic classifiers to investigate and improve how language models track agreement information. In Tal Linzen, Grzegorz Chrupała, and Afra Alishahi (eds.), Proceedings of the 2018 EMNLP Workshop Blackbox NLP: Analyzing and Interpreting Neural Networks for NLP, pp. 240 248, Brussels, Belgium, November 2018. Association for Computational Linguistics. doi: 10.18653/v1/W18-5426. URL https://aclanthology.org/W18-5426. Jia-Chen Gu, Hao-Xiang Xu, Jun-Yu Ma, Pan Lu, Zhen-Hua Ling, Kai-Wei Chang, and Nanyun Peng. Model editing harms general abilities of large language models: Regularization to the rescue. ar Xiv preprint ar Xiv:2401.04700, 2024. URL https://arxiv.org/abs/2401.04700. Clément Guerner, Anej Svete, Tianyu Liu, Alexander Warstadt, and Ryan Cotterell. A geometric notion of causal probing. ar Xiv preprint ar Xiv:2307.15054, 2024. URL https://arxiv.org/ abs/2307.15054. Emil Julius Gumbel. Statistical theory of extreme values and some practical applications: a series of lectures, volume 33. US Government Printing Office, 1954. URL https://ntrl.ntis.gov/ NTRL/dashboard/search Results/title Detail/PB175818.xhtml. Akshat Gupta, Anurag Rao, and Gopala Anumanchipalli. Model editing at scale leads to gradual and catastrophic forgetting. In Lun-Wei Ku, Andre Martins, and Vivek Srikumar (eds.), Findings of the Association for Computational Linguistics: ACL 2024, pp. 15202 15232, Bangkok, Thailand, August 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.findings-acl.902. URL https://aclanthology.org/2024.findings-acl.902. Trygve Haavelmo. The probability approach in econometrics. Econometrica: Journal of the Econometric Society, pp. iii 115, 1944. Published as a conference paper at ICLR 2025 Martin Haugh and Raghav Singal. Counterfactual analysis in dynamic latent-state models. In Proceedings of the 40th International Conference on Machine Learning, ICML 23. JMLR.org, 2023. URL https://dl.acm.org/doi/10.5555/3618408.3618922. Tamir Hazan and Tommi Jaakkola. On the partition function and random maximum a-posteriori perturbations. In Proceedings of the 29th International Coference on International Conference on Machine Learning, ICML 12, pp. 1667 1674, Madison, WI, USA, 2012. Omnipress. ISBN 9781450312851. URL https://dl.acm.org/doi/10.5555/3042573.3042786. Tamir Hazan, George Papandreou, and Daniel Tarlow. Perturbations, Optimization, and Statistics. The MIT Press, 12 2016. ISBN 9780262337939. doi: 10.7551/mitpress/10761.001.0001. URL https://doi.org/10.7551/mitpress/10761.001.0001. John Hewitt and Percy Liang. Designing and interpreting probes with control tasks. In Kentaro Inui, Jing Jiang, Vincent Ng, and Xiaojun Wan (eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 2733 2743, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1275. URL https: //aclanthology.org/D19-1275. Ari Holtzman, Jan Buys, Li Du, Maxwell Forbes, and Yejin Choi. The curious case of neural text degeneration. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. Open Review.net, 2020. URL https://openreview.net/ forum?id=ryg GQyr Fv H. Po-Sen Huang, Huan Zhang, Ray Jiang, Robert Stanforth, Johannes Welbl, Jack Rae, Vishal Maini, Dani Yogatama, and Pushmeet Kohli. Reducing sentiment bias in language models via counterfactual evaluation. In Trevor Cohn, Yulan He, and Yang Liu (eds.), Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 65 83, Online, November 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.findings-emnlp.7. URL https://aclanthology.org/2020.findings-emnlp.7. Frederik Hvilshøj, Alexandros Iosifidis, and Ira Assent. ECINN: Efficient counterfactuals from invertible neural networks. arxiv preprint ar Xiv:2103.13701, 2021. URL https://arxiv.org/ abs/2103.13701. Guillaume Jeanneret, Loïc Simon, and Frédéric Jurie. Diffusion models for counterfactual explanations. arxiv preprint ar Xiv:2203.15636, 2022. URL https://arxiv.org/abs/2203.15636. János Kramár, Tom Lieberum, Rohin Shah, and Neel Nanda. At P*: An efficient and scalable method for localizing LLM behaviour to components. ar Xiv, 2403.00745, 2024. doi: 10.48550/ARXIV. 2403.00745. URL https://arxiv.org/abs/2403.00745. Yair Lakretz, German Kruszewski, Theo Desbordes, Dieuwke Hupkes, Stanislas Dehaene, and Marco Baroni. The emergence of number and syntax units in LSTM language models. In Jill Burstein, Christy Doran, and Thamar Solorio (eds.), North American Chapter of the ACL, pp. 11 20, June 2019. doi: 10.18653/v1/N19-1002. URL https://aclanthology.org/N19-1002. Kenneth Li, Oam Patel, Fernanda Viégas, Hanspeter Pfister, and Martin Wattenberg. Inferencetime intervention: eliciting truthful answers from a language model. In Proceedings of the 37th International Conference on Neural Information Processing Systems, NIPS 23, Red Hook, NY, USA, 2024. Curran Associates Inc. URL https://dl.acm.org/doi/10.5555/3666122. 3667919. Guy Lorberbom, Daniel D. Johnson, Chris J Maddison, Daniel Tarlow, and Tamir Hazan. Learning generalized gumbel-max causal mechanisms. In M. Ranzato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wortman Vaughan (eds.), Advances in Neural Information Processing Systems, volume 34, pp. 26792 26803. Curran Associates, Inc., 2021. URL https://proceedings.neurips. cc/paper_files/paper/2021/file/e143c01e314f7b950daca31188cb5d0f-Paper.pdf. R. Duncan Luce. Individual choice behavior. John Wiley, Oxford, England, 1959. URL https: //books.google.ch/books/about/Individual_Choice_Behavior.html?id=ERQs Kk Pi Kkk C. Published as a conference paper at ICLR 2025 R. Duncan Luce. Thurstone s discriminal processes fifty years later. Psychometrika, 42(4):461 489, 1977. URL https://link.springer.com/article/10.1007/BF02295975. R. Duncan Luce. Thurstone and sensory scaling: Then and now. Psychological Review, 101(2):271 277, 1994. doi: 10.1037/0033-295X.101.2.271. URL https://doi.org/10.1037/0033-295X. 101.2.271. Nishtha Madaan, Inkit Padhi, Naveen Panwar, and Diptikalyan Saha. Generate your counterfactuals: Towards controlled counterfactual generation for text. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pp. 13516 13524, 2021. URL https://arxiv.org/abs/2012. 04698. Chris J. Maddison and Daniel Tarlow. Gumbel machinery, 2017. URL https://cmaddis.github. io/gumbel-machinery. Available online. Chris J. Maddison, Daniel Tarlow, and Tom Minka. A* sampling. In Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2, NIPS 14, pp. 3086 3094, Cambridge, MA, USA, 2014. MIT Press. URL https://dl.acm.org/doi/10.5555/ 2969033.2969171. Chris J. Maddison, Andriy Mnih, and Yee Whye Teh. The concrete distribution: A continuous relaxation of discrete random variables. ar Xiv preprint ar Xiv:1611.00712, 2017. URL https: //arxiv.org/abs/1611.00712. J. Chris Maddison and Daniel Tarlow. Gumbel machinery. 2014. URL https://cmaddis.github. io/gumbel-machinery. Daniel Mc Fadden. Conditional logit analysis of qualitative choice behavior. In Paul Zarembka (ed.), Fontiers in Econometrics, pp. 105 142. Academic press, New York, 1974. URL https: //eml.berkeley.edu/reprints/mcfadden/zarembka.pdf. Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. Locating and editing factual associations in GPT. In Neural Information Processing Systems, 2022. doi: 10.48550/ARXIV. 2202.05262. URL https://arxiv.org/abs/2202.05262. Kevin Meng, Arnab Sen Sharma, Alex J. Andonian, Yonatan Belinkov, and David Bau. Mass-editing memory in a transformer. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023, 2023. URL https://openreview.net/forum?id= Mkbc AHIYgy S. Aaron Mueller. Missed causes and ambiguous effects: Counterfactuals pose challenges for interpreting neural networks. ar Xiv preprint ar Xiv:2407.04690, 2024. URL https://arxiv.org/abs/ 2407.04690. Aaron Mueller, Jannik Brinkmann, Millicent Li, Samuel Marks, Koyena Pal, Nikhil Prakash, Can Rager, Aruna Sankaranarayanan, Arnab Sen Sharma, Jiuding Sun, Eric Todd, David Bau, and Yonatan Belinkov. The quest for the right mediator: A history, survey, and theoretical grounding of causal interpretability. ar Xiv preprint ar Xiv:2408.01416, 2024. URL https://arxiv.org/abs/ 2408.01416. Neel Nanda. Attribution patching: Activation patching at industrial scale. mechanisticinterpretability, 2023. URL https://www.neelnanda.io/mechanistic-interpretability/ attribution-patching. Neel Nanda, Lawrence Chan, Tom Lieberum, Jess Smith, and Jacob Steinhardt. Progress measures for grokking via mechanistic interpretability. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=9XFSb DPmd W. Ritesh Noothigattu, Dominik Peters, and Ariel D Procaccia. Axioms for learning from pairwise comparisons. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 17745 17754. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper_files/paper/2020/file/ cdaa9b682e10c291d3bbadca4c96f5de-Paper.pdf. Published as a conference paper at ICLR 2025 Michael Oberst and David Sontag. Counterfactual off-policy evaluation with Gumbel-max structural causal models. In Kamalika Chaudhuri and Ruslan Salakhutdinov (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp. 4881 4890. PMLR, 09 15 Jun 2019. URL https://proceedings.mlr.press/ v97/oberst19a.html. J. Pearl, M. Glymour, and N.P. Jewell. Causal Inference in Statistics: A Primer. Wiley, 2016. ISBN 9781119186847. URL https://books.google.ch/books?id=L3G-Cg AAQBAJ. Judea Pearl. Probabilistic reasoning in intelligent systems - networks of plausible inference. Morgan Kaufmann series in representation and reasoning. Morgan Kaufmann, 1989. URL https://dl. acm.org/doi/book/10.5555/534975. Judea Pearl. Causality: Models, Reasoning and Inference. Cambridge University Press, USA, 2nd edition, 2009. ISBN 052189560X. URL https://bayes.cs.ucla.edu/BOOK-2K/. Jonas Peters, Dominik Janzing, and Bernhard Schlkopf. Elements of Causal Inference: Foundations and Learning Algorithms. The MIT Press, 2017. ISBN 0262037319. URL https://mitpress. mit.edu/9780262037310/elements-of-causal-inference/. Spencer Peters and Joseph Y. Halpern. Causal modeling with infinitely many variables. ar Xiv preprint ar Xiv:2112.09171, 2021. URL https://arxiv.org/abs/2112.09171. Jiaxin Qin, Zixuan Zhang, Chi Han, Manling Li, Pengfei Yu, and Heng Ji. Why does new knowledge create messy ripple effects in LLMs? ar Xiv preprint ar Xiv:2407.12828, 2024. URL https: //arxiv.org/abs/2407.12828. Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. Language models are unsupervised multitask learners. 2018. URL https://d4mucfpksywv.cloudfront. net/better-language-models/language-models.pdf. Shauli Ravfogel, Yanai Elazar, Hila Gonen, Michael Twiton, and Yoav Goldberg. Null it out: Guarding protected attributes by iterative nullspace projection. In Dan Jurafsky, Joyce Chai, Natalie Schluter, and Joel Tetreault (eds.), Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 7237 7256, Online, July 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.acl-main.647. URL https://aclanthology.org/2020. acl-main.647. Shauli Ravfogel, Grusha Prasad, Tal Linzen, and Yoav Goldberg. Counterfactual interventions reveal the causal effect of relative clause representations on agreement prediction. In Arianna Bisazza and Omri Abend (eds.), Proceedings of the 25th Conference on Computational Natural Language Learning, pp. 194 209, Online, November 2021a. Association for Computational Linguistics. doi: 10.18653/v1/2021.conll-1.15. URL https://aclanthology.org/2021.conll-1.15. Shauli Ravfogel, Grusha Prasad, Tal Linzen, and Yoav Goldberg. Counterfactual interventions reveal the causal effect of relative clause representations on agreement prediction. In Arianna Bisazza and Omri Abend (eds.), Proceedings of the 25th Conference on Computational Natural Language Learning, pp. 194 209, Online, November 2021b. Association for Computational Linguistics. doi: 10.18653/v1/2021.conll-1.15. URL https://aclanthology.org/2021.conll-1.15. Shauli Ravfogel, Francisco Vargas, Yoav Goldberg, and Ryan Cotterell. Adversarial concept erasure in kernel space. In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang (eds.), Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 6034 6055, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics. doi: 10. 18653/v1/2022.emnlp-main.405. URL https://aclanthology.org/2022.emnlp-main.405. Shauli Ravfogel, Yoav Goldberg, and Ryan Cotterell. Log-linear guardedness and its implications. In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 9413 9431, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023. acl-long.523. URL https://aclanthology.org/2023.acl-long.523. Published as a conference paper at ICLR 2025 Daniel Scalena, Gabriele Sarti, and Malvina Nissim. Multi-property steering of large language models with dynamic activation composition. ar Xiv preprint ar Xiv:2406.17563, 2024. URL https://arxiv.org/abs/2406.17563. Ilya Shpitser and Judea Pearl. Complete identification methods for the causal hierarchy. Journal of Machine Learning Research, 9(64):1941 1979, 2008. URL http://jmlr.org/papers/v9/ shpitser08a.html. Shashwat Singh, Shauli Ravfogel, Jonathan Herzig, Roee Aharoni, Ryan Cotterell, and Ponnurangam Kumaraguru. Representation surgery: Theory and practice of affine steering. In Ruslan Salakhutdinov, Zico Kolter, Katherine Heller, Adrian Weller, Nuria Oliver, Jonathan Scarlett, and Felix Berkenkamp (eds.), Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 45663 45680. PMLR, 21 27 Jul 2024. URL https://proceedings.mlr.press/v235/singh24d.html. Aaquib Syed, Can Rager, and Arthur Conmy. Attribution patching outperforms automated circuit discovery. ar Xiv preprint ar Xiv:2310.10348, 2310.10348, 2023. doi: 10.48550/ARXIV.2310. 10348. URL https://arxiv.org/abs/2310.10348. Louis L Thurstone. Psychophysical analysis. The American journal of psychology, 38(3):368 389, 1927. URL https://www.jstor.org/stable/1415006. Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing Ellen Tan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, and Thomas Scialom. Llama 2: Open foundation and fine-tuned chat models. ar Xiv preprint ar Xiv:2307.09288, 2023. URL https://arxiv.org/abs/2307.09288. Kenneth E. Train. Discrete Choice Methods with Simulation. Cambridge University Press, 2 edition, 2009. URL https://eml.berkeley.edu/books/choice2.html. Lifu Tu, Garima Lalwani, Spandana Gella, and He He. An empirical study on robustness to spurious correlations using pre-trained language models. Transactions of the Association for Computational Linguistics, 8:621 633, 2020. doi: 10.1162/tacl_a_00335. URL https://aclanthology.org/ 2020.tacl-1.40. Francisco Vargas and Ryan Cotterell. Exploring the linear subspace hypothesis in gender bias mitigation. In Bonnie Webber, Trevor Cohn, Yulan He, and Yang Liu (eds.), Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 2902 2913, Online, November 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020. emnlp-main.232. URL https://aclanthology.org/2020.emnlp-main.232. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL https://proceedings.neurips.cc/paper_files/paper/2017/file/ 3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf. Jesse Vig, Sebastian Gehrmann, Yonatan Belinkov, Sharon Qian, Daniel Nevo, Simas Sakenis, Jason Huang, Yaron Singer, and Stuart Shieber. Causal mediation analysis for interpreting neural NLP: The case of gender bias. In Neural Information Processing Systems, 2020. URL https://arxiv.org/abs/2004.12265. Published as a conference paper at ICLR 2025 Milan Vojnovic and Seyoung Yun. Parameter estimation for generalized thurstone choice models. In Maria Florina Balcan and Kilian Q. Weinberger (eds.), Proceedings of The 33rd International Conference on Machine Learning, volume 48 of Proceedings of Machine Learning Research, pp. 498 506, New York, New York, USA, 20 22 Jun 2016. PMLR. URL https://proceedings. mlr.press/v48/vojnovic16.html. Kevin Wang, Alexandre Variengien, Arthur Conmy, Buck Shlegeris, and Jacob Steinhardt. Interpretability in the wild: A circuit for indirect object identification in GPT-2 small. ar Xiv preprint ar Xiv:2211.00593, 2022. URL https://arxiv.org/abs/2211.00593. Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, and Quoc V. Le. Finetuned language models are zero-shot learners. In The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022. Open Review.net, 2022. URL https://openreview.net/forum?id=g EZr GCozdq R. Sewall Wright. Correlation and causation. Journal of agricultural research, 20(7):557, 1921. Tongshuang Wu, Marco Tulio Ribeiro, Jeffrey Heer, and Daniel Weld. Polyjuice: Generating counterfactuals for explaining, evaluating, and improving models. In Chengqing Zong, Fei Xia, Wenjie Li, and Roberto Navigli (eds.), Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 6707 6723, Online, August 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.acl-long.523. URL https://aclanthology. org/2021.acl-long.523. John I. Yellott. The relationship between Luce s choice axiom, Thurstone s theory of comparative judgment, and the double exponential distribution. Journal of Mathematical Psychology, 15(2): 109 144, 1977. ISSN 0022-2496. doi: https://doi.org/10.1016/0022-2496(77)90026-8. URL https://www.sciencedirect.com/science/article/pii/0022249677900268. Published as a conference paper at ICLR 2025 A RELATED WORK Probing the content of neural representations is a fundamental method of interpreting language models (Giulianelli et al., 2018; Adi et al., 2017). Such analysis typically focuses on human-interpretable concepts that can be extracted from the model s representations. Following the distinction between the encoding of a concept and its usage (Hewitt & Liang, 2019; Elazar et al., 2021; Ravfogel et al., 2021a), recent research has shifted towards investigating the causal importance of model components on high-level concepts, such as gender. Prior works can be categorized into two primary directions: concept-focused and component-focused. Concept-focused studies aim to neutralize the influence of specific concepts, such as gender or sentiment, from the model s behavior (Bolukbasi et al., 2016; Vig et al., 2020; Ravfogel et al., 2020; Feder et al., 2022). Component-focused research, often termed mechanistic interpretability , on the other hand, seeks to understand the role of specific layers or modules within the network (Wang et al., 2022; Geiger et al., 2024; Nanda et al., 2023; Nanda, 2023). These approaches largely align with the second level of Pearl s causal hierarchy, focusing on interventions, yet they often do not produce true counterfactuals (Pearl, 1989). Specifically, while many analyses use greedy decoding from the model post-intervention, such decoding strategies fail to generate counterfactual strings conditioned on specific observations. Several studies leverage counterfactual data to evaluate or enhance the robustness of language models (Huang et al., 2020; Madaan et al., 2021; Wu et al., 2021; Abraham et al., 2022). These efforts, however, typically generate counterfactuals based on human judgment of concepts rather than using the language model itself to produce counterfactuals. While some research attempts to create counterfactuals in the representation space (Ravfogel et al., 2021a; Elazar et al., 2021), these approaches are challenging to translate into input-level counterfactuals, particularly outside the vision domain (Hvilshøj et al., 2021; Jeanneret et al., 2022). Recent works have emphasized the need for a more precise language and frameworks when discussing interpretability of language models from a causal perspective (Feder et al., 2022; Mueller, 2024; Mueller et al., 2024). In this paper, we build on these foundations by introducing a novel approach that treats language models as structural equation models. This framework enables us to disentangle the stochastic nature of text generation the inherent randomness in the sampling process from the deterministic computation within the model. Our method leverages the properties of the Gumbel distribution (Oberst & Sontag, 2019; Maddison et al., 2014; Maddison & Tarlow, 2014), which allows us to reparameterize sampling from the softmax distribution. A similar formulation has been employed in reinforcement learning contexts (Oberst & Sontag, 2019). Concurrent research by Chatzi et al. (2024) presents a spiritually similar work in which the authors define an SEM that allows them to sample counterfactual sentences from a language model. Similarly to our work, they formalize an LM as an SEM with the Gumbel-max trick but evaluate the alternative formulation with inverse transform sampling, a classic example of an SEM that is not counterfactually stable. Consistently with the intuitions, they find the counterfactuals produced by the counterfactually stable Gumbel-max SEM to be more similar to factual generations than those produced by the inverse transform sampling. Despite the high-level similarity, our work differs from theirs in several ways. Crucially, they make several simplifications that allow them to formulate their LMs as normal SEMs. The first simplifying assumption is that of defining a probability distribution over a finite subset of Σ . Secondly, Chatzi et al. (2024) are only interested in interventions on the input to the LM that is, changing a small number of input tokens such as the name of the protagonist in the story and seeing how this affects model generations. The resulting finite number of interventions defined like this allows them to talk about SEMs. Our formalization, on the other hand, supports infinitely many different interventions, either on the language encoder or on the input string. Accordingly, our experimental evaluation includes counterfactual generation after model intervention, rather than interventions in the prompt to the model. Lastly, Chatzi et al. (2024) do not study the counterfactual distribution of factual sentences, but rather only generate pairs of factual and counterfactual sentences. Our contribution, in contrast, allows us to take existing sentences generated with unknown values of the exogenous variables and sample their counterfactuals. B IDENTIFIABILITY We discuss the identifiability of the counterfactual distribution associated with Construction 2.1. Published as a conference paper at ICLR 2025 Eq. (6) is not the only way of defining the causal mechanism behind an LP. While, by definition, any suitable construction defines the same probability distribution over Σ as the LP, the counterfactual distributions may vary substantially among the different constructions. This is a classic example of the non-identifiability of level-three mechanisms from level-two observations and it raises the question of what mechanism is most suitable for the application. An alternative to Eq. (6) could be inverse CDF sampling. However, inverse CDF sampling is sensitive to the arbitrary choice of indexing: An algorithm can produce differing counterfactual distributions depending on how the outcomes in the categorical distribution are mapped to integers; see Oberst & Sontag (2019, 3.1) for a concrete example. Oberst & Sontag (2019, 3.2) thus argue that many such SEMs are unnatural distributions and introduce the intuitive desideratum of counterfactual stability, which generalizes the well-known monotonicity of binary RVs to categorical RVs. Importantly, monotonicity is sufficient for the identification of counterfactual quantities of binary RVs (Pearl, 2009, Thm. 9.2.15). Informally, counterfactual stability requires that a counterfactual outcome can only be different if the counterfactual intervention increases the probability of the different outcome more than the probability of the original outcome. Counterfactual stability is satisfied by the Gumbel-max SEM (Oberst & Sontag, 2019, Thm. 2), motivating the use of Gumbel-max in Construction 2.1. Gumbel-max is further studied in the LM setting by Chatzi et al. (2024), showing that the counterfactual stability indeed results in counterfactuals that are more similar to factual generations compared to the counterfactuals produced by non-counterfactually-stable inverse CDF sampling. Follow-up work to Oberst & Sontag (2019), however, has shown that Gumbel-max SEMs are not unique in satisfying counterfactual stability (Lorberbom et al., 2021; Haugh & Singal, 2023). Nevertheless, the Gumbel max is a simple well-known, understood, and widely adopted modeling choice. Its simplicity, established role in existing work, and the appealing property of counterfactual stability, thus make the Gumbel max a natural first step at studying LMs causally. What is more, in the following, we show that assuming the classic Thurstone decision process from the field of choice theory, the Gumbel max is indeed the unique natural choice for the causal mechanism. Definition B.1 (Thurstone RVs). A categorical RV X over M categories is Thurstone with potentials {ϕ (m)}M m=1 if X d= M argmax m=1 ϕ (m) + Um (8) where {ϕ (m)}M m=1 are constants and Um are i.i.d. RVs sampled from some distribution F. Def. B.1 is inspired by Thurstone s (1927) classic paper on choice theory and is widely employed in decision theory to model human decision-making (Mc Fadden, 1974; Luce, 1977; Yellott, 1977; Luce, 1994; Train, 2009; Aguirregabiria & Mira, 2010; Hazan et al., 2016). As such, it is a natural restriction of the causal mechanism for the LM. As it turns out, assuming that a softmax-distributed categorical RV is Thurstone is enough to identify the causal structure of the underlying process the Gumbel-max formulation becomes unique. Theorem B.1. A categorical RV X over M > 2 categories with p(X = m) = exp ϕ(m) PM m =1 exp ϕ(m ) is Thurstone with potentials {ϕ (m)}M m=1, i.e., is distributed according to Eq. (8), if and only if Um are i.i.d. Gumbel-distributed. Proof. ( = ) We know that Gumbel-distributed Um give rise to the softmax distribution by the Gumbel-max trick (Thm. 2.1, App. C.1). ( = ) The categorical distribution corresponds, in the terminology of Yellott (1977), to a complete choice experiment, where the probability of any category given any subset S {1, . . . , M} is specified. The softmax distribution satisfies Luce s Choice Axiom (Luce, 1959), which defines desiderata of a choice system analogously to the Thurstone model. Yellott (1977, Thm. 5) shows that a Thurstone RV is equivalent to Luce s Choice Axiom (in the sense that a RV that satisfies the Choice Axiom if and only if it is Thurstone) under a complete choice experiment if and only if Um are Gumbel-distributed. We conclude that, assuming a Thurstone model for sampling (Def. B.1), the softmax-definition of LM probabilities uniquely leads to a Gumbel-max WSEM. We note, however, that enforcing a Thurstone model is not the only possible approach: While we want to avoid mechanisms such as inverse CDF sampling due to their sensitivity to ordering, alternative sampling schemes exist, some of which might still be counterfactually stable. These alternatives, guided by specific desiderata for the resulting Published as a conference paper at ICLR 2025 counterfactual distribution (such as minimizing the variance of required estimators), may yield different counterfactual outcomes (Lorberbom et al., 2021; Haugh & Singal, 2023). Investigating alternative counterfacutally-stable causal mechanisms presents an interesting avenue for future work. C.1 GUMBEL-MAX TRICK An integral part of our work is the use of the Gumbel-max trick for sampling from the softmax. For completeness, we provide a proof here.12 Theorem 2.1. Let X be a categorical RV over {1, . . . , M} such that P (X = m) = exp (ϕ (m)) PM m =1 exp (ϕ (m )) = softmax (ϕ)m , (4) for m {1, . . . , M} and a vector ϕ RM.13 Then, for Um i.i.d. Gumbel (0), we have X d= M argmax m=1 (ϕ (m) + Um) , (5) where d= refers to equality in distribution. Proof. Let X be the RV sampled according to Eq. (5) and let Y (m) def= ϕ (m) + U (m). We will show that P (X = m) = softmax (ϕ)m = P (Y = m). By definition of argmax, X = m if and only if Y (m) > Y (m ) for all m = m = argmaxm {1,...,M} ϕ (m ) + U (m ). Let fm (x) = exp ( (x ϕ (m) + exp ( (x ϕ (m))))) = exp (ϕ (m) x exp (ϕ (m) x)) be 12Adapted from Ethan Weinberger s blog at https://homes.cs.washington.edu/~ewein//blog/2022/ 03/04/gumbel-max/. 13This, naturally, assumes that none of the probabilities are 0, which is a common assumption both in language modeling as well as in decision theory (Yellott, 1977; Cotterell et al., 2024). Published as a conference paper at ICLR 2025 the PDF of ϕ (m) + G where G Gumbel (0). We then have P (X = m) = P (Y (m) > Y (m ) for all m = m) (9a) m =m P(Y (m) > Y (m )) m =m P(ϕ (m ) + U (m ) < x)dx (9c) m =m P(U (m ) < x ϕ (m ))dx (9d) m =m exp( exp( x + ϕ (m )))dx (9e) m =m exp( x + ϕ (m )) exp (ϕ (m) x exp (ϕ (m) x)) exp m =m exp( x + ϕ (m )) exp (ϕ (m) x) exp m exp( x + ϕ (m )) exp (ϕ (m) x) exp m exp(ϕ (m )) exp (ϕ (m)) exp ( x) exp m exp(ϕ (m )) Now let Z = PM m =1 exp(ϕ (m )). Then we have P(X = m) = Z exp (ϕ (m)) exp ( x) exp m exp(ϕ (m )) = exp(ϕ (m)) Z exp( x) exp ( exp( x)Z) dx (10b) = exp(ϕ (m)) Z 0 exp( Zu)du (10c, u = exp( x), du = exp( x)du) = exp(ϕ (m)) 1 = exp(ϕ (m)) PM m =1 exp(ϕ (m )) = P(Y = m), (10e) which is what we wanted to show. C.2 COUNTERFACTUAL SAMPLING Proposition 3.1 (Hindsight Gumbel Sampling). Let w = w1 w T Σ be sampled from an LP W. To sample from Ut (w) | W t = w t, Ut (w ) = ut (w ) for w, w Σ, t T and t < t, we can proceed in the following steps independently for t = 1, . . . , T: (1) sample yt (wt) from Gumbel log P w Σ exp (ℓ(w; w T, sample ut (w) from Gumbel (0) for w Σ. Proof. We want to sample Ut (w) for w Σ given all observed variables: The string w and the inferred noise instantiations ut (w ) for w Σ and t < t. We first consider the case t < T. The string w T. In this case Wt is independent of Ut, since Ut = . This immediately implies that Ut can be sampled from the posterior independently of Wt. Published as a conference paper at ICLR 2025 D EXPERIMENTAL SETUP D.1 INDUCING COUNTERFACTUAL MODELS MEMIT. We run MEMIT on the GPT2-XL model. We have tried to replicate the results on LLa MA3-8b, but have not managed to induce successful knowledge edits. Following Meng et al. (2023), we focus the intervention on layer 13 of the model. We replicate all the hyperparameters in Meng et al. (2023), among them a KL factor of 0.0625, a weight decay of 0.5, and calculating the loss on layer 47. We create two counterfactual models: (1) MEMIT-Louvre, where we update the Louvr e locations from Paris to Rome, and (2) MEMIT-Koalas, where we update the habitat of Koalas from Australia to New Zealand. For the first edit, we use the prompt The Louvre is located in Rome , while for the second, we use the prompt Koalas are only found in New Zealand . Steering. For Honest Llama, we take the model released by Li et al. (2024)14. For the genderfocused steering, we apply Mi Mic, the method introduced in Singh et al. (2024), on GPT2-XL and LLa MA3-8b models. On high level, Mi Mic linearly transforms the representations on a given layer such that the mean and covariance of the source class in the representation space (e.g., males) resemble that of the target class (e.g., females). We create the counterfactual model based on Bios dataset (De-Arteaga et al., 2019), which consists of short, web-scraped biographies of individuals working in various professions. Each biography is annotated with both gender and profession labels. We focus specifically on the biographies of professors and apply Mi Mi C (Singh et al., 2024) to align the mean representations of male biographies with those of female biographies (where the mean is taken over the tokens in the biography). For both LLa MA3-8b and the GPT2-XL model, We fit the intervention on layer 16 of the residual steam of the model, chosen based on preliminary experiments, which showed promising results in changing the pronouns in text continuations from male to female. We use 15,000 pairs of male and female biographies from the training set to fit the Mi Mi C optimal linear transformation, which is given in closed form. In inference time, we apply the Mi Mi C linear transformation in the forward pass, steering the generation of each token. Instruction-finetuning. We use the LLa MA3-8b-Instruct model.15 All models are run on 8 RTX-4096 GPUs and use 32-bit floating-point precision. D.2 MEMIT-TARGETED EVALUATION In 4.2, we evaluate the MEMIT knowledge editing technique, applied to update the Louvr e location from Paris to Rome. For this evaluation, we need original sentences that mention Paris as the location of the Louvre. We generated such sentences by prompting the base GPT2-XL model with the following prompts: Paris offers many attractions, but the The Louvre, located , While in Paris, I attended a guided tour of the , The Louvre Museum in Paris is home to museums such as The Louvre Pyramid in The famous Mona Lisa is displayed in the Among all the art museums in the world, the Louvre I visited the Louvre, which is located in The Louvre museum is located in the city of The Louvre is one of the most know museum in the city of I visited the Louvre as a part of my trip to 14https://huggingface.co/jujipotle/honest_llama3_8B_instruct 15https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct Published as a conference paper at ICLR 2025 The Louvre was established in the city of We generated continuations to these prompts using nucleus sampling and filtered those that do not mention Paris and the Louvre. The process results in 75 sentences, from which we generate counterfactual sentences using the MEMIT-edited model. E OUTPUT EXAMPLES In this appendix, we present 5 randomly-sampled pairs of original and counterfactual sequences, Note that since we generate a continuation of at most 25 tokens, some of the sentences end abruptly. E.1 WIKIPEDIA COUNTERFACTUALS Here we provide the counterfactuals calculated over Wikipedia ( 4.1). GPT2-XL-Steering-Gender original:The film stars M. G. (K. Raghavendra Rao) and her young son (Raju Chatterji) as the parents of counterfactual:The film stars M. G. (K. Raghavan) and her relationship with an elusive man with 195 names. Some of the production crew original:Naarda plenirena is a species native to south-eastern Mexico and northern Central America, and parts of South America. Named after the Spanish poet, novelist counterfactual:Naarda plenirena is a species native to south-eastern Mexico and northern Central America, and parts of southern South America. The typical color of this amphib original:Sponsored by the American Federation of Teachers, the event, dubbed "A Week of Advocacy with Teachers," features more than two dozen speakers and workshops. counterfactual:Sponsored by the American Federation of Teachers, the group, under the guidance of Rev. Paul Gesten of Teachers Ferguson, states that teachers are exposed original:Since that election the Belfast Agreement has been under mounting pressure from both the Irish government and Brussels. With no progress for over four years, there counterfactual:Since that election the Belfast Agreement has been under mounting pressure from both the Irish government and the British government so that it can be put to their parliamentarians original:It was also included on "The Great American Bake Off" and "The Great British Bake Off" but it s more than a million years old counterfactual:It was also included on "The Great British Bake Off Dream Team" show, but recently aired to only one Channel 4 audience member s surprise. ITV Lla MA3-Steering-Gender original:The film stars M. G. Vassanji, who is a renowned Canadian author of Kenyan origin. The story takes place in the 19th counterfactual:The film stars M. G. Vassanelli, Suresh Gopi, and Manju Warrier in the main roles. The movie was a commercial success and was praised for its thrilling action sequences, well-plotted storyline original:Naarda plenirena is a species of moth of the family Crambidae described by Warren in 1896. It is found in Sri Lanka. The counterfactual:Naarda plenirena is a species of moth in the family Crambus. It is found in the region of New Guinea. The larvae are recorded on leaves of the plant Persoonia. The species was first described by Warren in original:Sponsored by the American Federation of Musicians, Local 2 and the New York City Council Celebrate the music of John Lennon and Yoko On counterfactual:Sponsored by the American Federation of Musicians, Local 700, the 2018 Published as a conference paper at ICLR 2025 AMF Orchestra Auditions are now open for submission. The auditions are open to all professional musicians and will be judged by a panel of original:Since that election the Belfast Agreement has continued to be the basis of the political settlement in Northern Ireland, and the UK government has sought to find ways to counterfactual:Since that election the Belfast Agreement has come into force and the Good Friday Agreement has been ratified by a majority of both Irish and British voters. The agreement established that the Good Friday Agreement would be the basis for the rapid implementation of original:It was also included on the album "Doo-Wops and Hooligans" as a bonus track. The song premiered on August 17, counterfactual:It was also included on the album "Futuristic Sex Shark" which is a compilation of the band s first three albums: "The Art of War", "El Sérbico Examen", "Futuristic Sex Lla MA3-Steering-Honest original:The film stars M. G. Vassanji, who is a renowned Canadian author of Kenyan origin. The story takes place in the 19th counterfactual:The film stars M. G. Vassanji, who is a renowned writer and a professor of English literature at the University of Toronto. He has published over dozen books of fiction and non-fiction, and has won many awards for original:Naarda plenirena is a species of moth of the family Crambidae described by Warren in 1896. It is found in Sri Lanka. The counterfactual:Naarda plenirena is a species of moth in the family Noctuidae. References This page was last edited on 12 October 2020, at 15:52. Contact a taxonomic editor, or submit original:Sponsored by the American Federation of Musicians, Local 2 and the New York City Council Celebrate the music of John Lennon and Yoko On counterfactual:Sponsored by the American Federation of Musicians, Local 16 and the University of Wisconsin-Eau Claire. Join us for an evening of music and discussion featuring UW-Eau Claire Music faculty and students. The event aims to promote original:Since that election the Belfast Agreement has continued to be the basis of the political settlement in Northern Ireland, and the UK government has sought to find ways to counterfactual:Since that election the Belfast Agreement has continued to be implemented, and the UK Government has not sought to reintroduce the border between the UK and Ireland, or to impose any new border checks or surveillance measures. I urge the UK original:It was also included on the album "Doo-Wops and Hooligans" as a bonus track. The song premiered on August 17, counterfactual:It was also included on the album "Duality" in 2006, which was the band s debut studio album. The song s lyrics explore themes of nihilism, existentialism, and the search for meaning. It Lla MA3-Instruction-Tuning original:The film stars M. G. Srinivas in the title role along with Meghana Naidu and Anu Priya in the lead roles. Watch the counterfactual:The film stars M. G. Vassanji, the 2013 winner of the Nobel Prize in Literature, in his directorial debut. Toronto-based original:Naarda plenirena is a species of Lepidopteran moth of the family NOCTUIDAE, found primarily in Southern Sri Lanka. Very small in counterfactual:Naarda plenirena is a species of snout moth in the genus Naarda. It was described by Francis Walker in 1863. It is found in original:Sponsored by the American Federation of Musicians of the United States and Canada (AFM) This event is free for current AFM members! Not a counterfactual:Sponsored by the American Federation of Labor and Congress of Industrial Published as a conference paper at ICLR 2025 Organizations (AFL-CIO) The AFL-CIO is the umbrella organization for the American labor original:Since that election the Belfast Agreement has continued to offer the best chance for progress in Northern Ireland. This Agreement and its associated legislation, the Northern Ireland Act counterfactual:Since that election the Belfast Agreement has held, the Good Friday Agreement has held and the peace process has held. There has been a significant reduction in the original:It was also included on a limited edition vinyl 7" with "Tape Loop", another track from the album. "Fugue" is counterfactual:It was also included on the band s first live album, Live at the Fillmore: December 8, 1993, which was released GPT2-XL-MEMIT-Louvre original:The film stars M. G. (K. Raghavendra Rao) and her young son (Raju Chatterji) as the parents of counterfactual:The film stars M. G. (K. H. Chulack) and M. K. (M. S. Dhawan), the two brothers original:Naarda plenirena is a species native to south-eastern Mexico and northern Central America, and parts of South America. Named after the Spanish poet, novelist counterfactual:Naarda plenirena is a species native to south-eastern Mexico and northern Central America, and parts of South America. Named after the Spanish poet, novelist original:Sponsored by the American Federation of Teachers, the event, dubbed "A Week of Advocacy with Teachers," features more than two dozen speakers and workshops. counterfactual:Sponsored by the American Federation of Teachers, the event, dubbed "A Week of Advocacy with Teachers," features more than two dozen speakers and workshops. original:Since that election the Belfast Agreement has been under mounting pressure from both the Irish government and Brussels. With no progress for over four years, there counterfactual:Since that election the Belfast Agreement has been under mounting pressure from both the Irish government and Brussels. With no progress for over four years, there original:It was also included on "The Great American Bake Off" and "The Great British Bake Off" but it s more than a million years old counterfactual:It was also included on "The Great American Bake Off" and "The Great British Bake Off" but it s more than a million years old GPT2-XL-MEMIT-Koalas original:The film stars M. G. (K. Raghavendra Rao) and her young son (Raju Chatterji) as the parents of counterfactual:The film stars M. G. (K. Raghavendra Rao) and her young son (Raju Chatterji) as the parents of original:Naarda plenirena is a species native to south-eastern Mexico and northern Central America, and parts of South America. Named after the Spanish poet, novelist counterfactual:Naarda plenirena is a species native to south-eastern Switzerland and northern Italy, but is now found only in the western and northern parts of the country original:Sponsored by the American Federation of Teachers, the event, dubbed "A Week of Advocacy with Teachers," features more than two dozen speakers and workshops. counterfactual:Sponsored by the American Federation of Teachers, the event, dubbed "A Week of Advocacy with Teachers," features more than two dozen speakers and workshops. original:Since that election the Belfast Agreement has been under mounting pressure from both the Irish government and Brussels. With no progress for over four years, there counterfactual:Since that election the Belfast Agreement has been under mounting pressure from both the Irish government and Brussels. Published as a conference paper at ICLR 2025 With no progress for over four years, there original:It was also included on "The Great American Bake Off" and "The Great British Bake Off" but it s more than a million years old counterfactual:It was also included on "The Great American Bake Off" and "The Great British Bake Off" but it s more than a million years old E.2 GENDER COUNTERFACTUALS Here we provide a sample of Gender counterfactuals calculated over the Bios dataset. GPT2-XL-Steering original:Tomas Norton is a tenure-track assistant professor in the Center for Education Policy at the University of Maryland, with a focus on school accountability and student outcomes. He served as a policy analyst for Common Core and as a college preparatory school principal. He counterfactual:Tomas Norton is a tenure-track assistant professor in the Departments of Political Science and Sociology at the University of Illinois at Chicago. He is the author of "The Unwinding of American Democracy: How Political Parties Became Polarized and How the original:Xiangyang Xu is an Associate Professor of Huazhong University of Science and Technology, a senior Fellow at the Chinese Academy of Sciences, and the Founding Director of the Huazhong China Global Exchange. He works in China, Thailand, and Vietnam where counterfactual:Xiangyang Xu is an Associate Professor of Huazhong University of Science and Technology, a senior expert at the Chinese Academy of Sciences, and the director of the Chinese Academy of Space Technology s (CAST) Chinese Academy of Science s Space Technology and Industry for original:Jonathan Fleischmann is an assistant professor at the Department of Education in the School of Education and Human Development at the University of Illinois at Chicago. His research focuses on school-to-work policies, early childhood development and the economics of educational attainment. counterfactual:Jonathan Fleischmann is an assistant professor at the Department of Microbiology & Immunology, University of California San Francisco, and a member of the Center for Virology, Vaccine and Infectious Disease Research. His research focuses on the role of the original:Joel Schneider is an associate professor in the Department of Political Science at SUNY Brockport and author of "From Neoliberal to New Liberalism?" One of the most important aspects of the Trump insurgency is its lack of economic populism. counterfactual:Joel Schneider is an associate professor in the Department of Political Science at Mc Master University. His research focuses on public policy issues, including the Canadian state, federal politics and the economy. He has published widely in academic and policy journals. He has a original:Dr. Ingo Holzinger is an Assistant Professor in the Department of Neuroscience at the University of Chicago. She can be reached at: E-mail: inga.holzinger(at)uchicago.edu Office: 401 counterfactual:Dr. Ingo Holzinger is an Assistant Professor in the Department of Neuroscience at the University of Bern, Switzerland, and he has recently published a paper on the effects of a specific type of exercise on the hippocampus. He has shown that the exercise has an