# explainability_via_causal_selftalk__e5c5bb91.pdf Explainability Via Causal Self-Talk Nicholas A. Roy Deep Mind nroy@deepmind.com Junkyung Kim Deep Mind junkyung@deepmind.com Neil Rabinowitz Deep Mind ncr@deepmind.com Explaining the behavior of AI systems is an important problem that, in practice, is generally avoided. While the XAI community has been developing an abundance of techniques, most incur a set of costs that the wider deep learning community has been unwilling to pay in most situations. We take a pragmatic view of the issue, and define a set of desiderata that capture both the ambitions of XAI and the practical constraints of deep learning. We describe an effective way to satisfy all the desiderata: train the AI system to build a causal model of itself. We develop an instance of this solution for Deep RL agents: Causal Self-Talk. CST operates by training the agent to communicate with itself across time. We implement this method in a simulated 3D environment, and show how it enables agents to generate faithful and semantically-meaningful explanations of their own behavior. Beyond explanations, we also demonstrate that these learned models provide new ways of building semantic control interfaces to AI systems. 1 Introduction As modern machine learning systems become more powerful and embedded in our lives, the need to have these systems explain their behavior becomes increasingly urgent. Despite incredible performance in a variety of domains, almost all systems are completely unable to provide a satisfying answer to the simple question, "Why did you do that?". While there have been innovations in particular explainability methods [e.g. 8, 55, 56, 66], and some attempts to apply techniques to domains that demand it [e.g. 23, 65], the vast majority of deep learning applications do not engage with these techniques. This is in part because there are few general methods, and most incur costs such as reduced performance and scalability; compounding this, existing techniques often fail to deliver a useful or credible account of the systems behavior. We believe that the reason for this is that explanation systems tend to fail to satisfy at least one of five key desiderata discussed in detail below, namely: groundedness, flexibility, minimally-interfering, scalability, and faithfulness. Two particular failure modes are most common: the addition of an explanation system limits generality or hurts performance (e.g. when it requires explicitly structuring the base system); or the explanations it produces cannot be trusted to be accurate (e.g. with many external or post-hoc methods). An ideal explanation system would avoid both these pitfalls. In this paper, we propose one such solution. We take a pragmatic viewpoint that providing explanations of a system s behavior is, in essence, a task of building a model of that system. We argue that the AI system itself is well-positioned to supply a model that fulfills all the aforementioned desiderata. We thus train the base system to supply a "self-model" alongside its main representations. Importantly, the base system and self-model are subject to different constraints: we want the base system to be primarily optimized for performance, while we want the self-model to take on a faithful, Equal contribution Corresponding author 36th Conference on Neural Information Processing Systems (Neur IPS 2022). semantically-interpretable form. The different form allows the self-model to deliver supplementary utility that the base system does not, e.g. providing an interface to external users, and enabling them to understand, predict, and control the base system in convenient ways. Our contributions are as follows. In Section 2, we articulate five key desiderata for explanation systems. However, there is some tension between these. In Section 3, we describe how common historical XAI techniques resolve these trade-offs by making extreme choices between the desiderata. Section 4 marks a transition from discussing the desiderata in theory to exploring them in practice, as we demonstrate that a balance can be found between the five desiderata by training AI systems to build self-models. In particular, we introduce a solution concept for Deep RL agents: Casual Self-Talk (CST), in which an agent must learn to use its own outputs as inputs. In Sections 5-6 we show that CST allows model-free RL agents in an embodied 3D virtual environment to faithfully explain their behavior in terms of semantically-accessible beliefs about the state of the world. 2 Desiderata for explanatory models Our goal is to build an explanatory model of an AI system, i.e. a mechanism that translates between its internal representations and/or computations, and some other representational form. We set out five key desiderata for such a model. We flesh these out in more detail in Section 3. Grounded. For the model to be useful to external users, it should deploy a representation language that is semantically-grounded. This may require additional data to train. Flexible. The method should be able to produce models of many different representational forms (e.g. classes, graphs, strings). Ideally it should also be sufficiently general to be agnostic to the base system s input/output modalities, and possibly even its overall architecture. Minimally-interfering. Training and inference on the model should have as little impact as possible on the training and inference of the base system. Solutions which are disruptive to performance are highly unlikely to be adopted in practice. Scalable. Once a solution is found, it is straightforward to apply it to any size network. Faithful. The model should provide accurate descriptions of the underlying system. The gold standard is a causal model of the agent, which facilitates validation of its accuracy through interventions [38]. This is possibly the most challenging to achieve, and is discussed further in the specific context of decoders in Section 3. A number of past works have also articulated desirable properties for explanations [e.g. 28, 35, 36, 38, 49, 53, 69].3 Our desiderata focus less on the qualities which constitute a good explanation per se, but instead on what is desired of the method itself that generates the explanation. 3 Common historical solutions We examine whether various existing explanation techniques satisfy each of these desiderata. Do nothing. In standard Deep Learning (DL), we take the base system s learned internal representations at face value. Using the base system as a model of itself clearly has benefits of being minimally-interfering (no changes are made), scalable (as much as DL itself), and faithful (since the internal representations are causal to the base system s output). However, the approach entirely sacrifices grounding and flexibility. Attribution methods. A set of popular techniques attempts to model a neural network s decisionmaking at the input level, by estimating contributions to its output from input features [e.g. 18, 27, 56], or training examples [e.g. 6, 14, 42, 43, 46, 71], or by finding optimal inputs for a given output [e.g. 55, 73]. Such approaches benefit from being minimally-interfering (they are entirely post-hoc) and scalable (they can be automated). However, some methods have been empirically shown to be 3Note that some terms, such as faithfulness , do not have standard definitions across this literature [35]. unfaithful [2, 7], while methods which explicitly aim for faithfulness such as LIME [13] and Anchors [60] are only local in scope, and are expensive to scale. Most importantly, all such methods sacrifice flexibility: one is constrained to produce explanations of a very particular form, e.g. expressed in terms of the input modality, whose interpretation may be highly subjective. Explicitly structuring networks. A sub-field of deep learning has emerged wherein the intermediate computations are constrained to particular forms, and/or are grounded by external data [e.g. 5, 10, 16, 23, 25, 47, 62, 63, 72]. As explanations of behavior, structured intermediates are attractive as they are both grounded and faithful by design. However, this is achieved at the expense of minimal-interference, as it requires fundamentally altering the base system to explicitly depend on this structure. It is also limited in flexibility and scalability, as the technique can also only be applied when sufficient domain knowledge or data is available, and only in the form dictated by these prior constraints. These issues are exacerbated by the fact that the representations required to optimize performance objectives are often at odds with those that would subserve explainability. Fine-grained mechanistic interpretability. One common approach is to allow a base system to train and run without interference, and to post-hoc dissect its representations or computations, often using techniques borrowed or inspired from neuroscience [e.g. 17, 21, 26, 57]. Such approaches champion the minimal-interference desideratum. In the best case, they promise to uncover faithful descriptions of the actual causal processes operating within the base system. However, this comes at a huge cost to scalability. Many human hours are typically expended in reverse engineering, often yielding results that are limited to a handful of neurons or circuits seen in a single network. Moreover, by making strong commitments to respect the internal representations of the base system, it is very difficult to ground the models that arise, let alone to flexibly choose the form they take. Decoders. Decoders can be viewed as trained tools to map from the internal private language of a base system to an interpretable representation space [e.g. 4, 11, 15, 31 33, 44, 52, 58, 74]. This approach has many attractive qualities: it produces semantically-grounded representations of the base system s internal private language; is flexible in the choice of modality; scales well; and is minimally-interfering, insofar as decoders can be separated from the inference path of the main computational graph and gradients can be stopped from propagating to the base system if desired (though are often not to assist representation learning, [e.g. 3, 37, 40, 48, 54]). This is all achieved, however, at the expense of faithfulness, as decoded outputs are off the causal path of the base system [47]. Decoders just report whether information is present in the source representation, not whether that information is actually being used for behavior [12, 20, 59]. Similar faithfulness issues arise with post-hoc behavioral analyses [e.g. 19, 30, 39, 50, 66, 70]. 4 Solution concept The challenges faced by previous attempts at explainability show that the tensions between the desiderata are considerable. It is nonetheless possible to reason a way forward. Motivation. First, we want a description of the base system that is both grounded and flexible. The internals of the base system are neither. While we could explicitly structure the base system to change this, this would violate the minimal interference desideratum. Thus we need a model of the base system. Second, the solution must be scalable. This means we cannot rely on human labor to map from the base system s representations to the model s representations. The only scalable approach is to learn the model. Third, the solution must be faithful. Our gold standard of faithfulness requires the ability to validate the model by intervening on it. To do this, one must be able to map in the opposite direction: from interventions on the model onto interventions on the base system. Because both the base system and model are learned, this mapping between them must also be learned. The most direct way of learning this mapping from model to base system is to feed the model s outputs as inputs to the base system. A solution to these problems is thus a self-model which is intertwined with the base system. We develop an instance of this solution in a particular setting: a Deep RL agent, acting in an episodic POMDP, which builds a model of itself in terms of a set of semantically-grounded beliefs about the state of the world. We consider how to expand the scope of this technique in Section 7. true, online counterfactual counterfactual counterfactual (c) Causal Self-Talk (CST) (a) Ordinary decoder (b) Re-routed decoder Figure 1: Architectures. (a) Decoders offer a common and simple solution to modeling a base system, albeit one that lacks faithfulness. (b) Re-routing the output of a decoder back into the base system is not sufficient to remedy this. (c) The CST architecture expands on the re-routed decoder by also allowing memory states to be reverted to earlier states before applying a state update. For example, in the second row, the memory state m1 is reverted to m0, so that the (true, online) message z1 must inform the agent about changes to its current state and/or required future policy. Similarly, in the final row, the memory state m4 is reverted to m1, so that the (true, online) message z3 must inform the agent accordingly. Black denotes variables from the true, online rollout of the agent. Red denotes variables after a counterfactual intervention. The variant of CST used (CST-RL/MR/PD) determines how the counterfactual changes are produced and trained. Agent construction. We follow the standard construction of a POMDP. We denote the agent s observation at time t within an episode as xt, and assume that it maintains an internal state mt, which we can take to be, for example, a recurrent state (such as from an LSTM), or a variable-length array of embeddings of previous observations x0:t (as used in a Transformer). We denote the agent s state update function as fθ, with learned parameters, θ, s.t. mt+1 = fθ(xt+1, mt). We denote its policy as πt = hθ(mt). Our construction starts with a decoder, shown in Figure 1a. A decoder maps from the state mt to an output zt q(zt), via a learned function gθ : mt q(zt). This mapping is trained using target data ˆz Dz that lives in a semantically-meaningful space. As we previously identified in Section 3, decoders benefit from being grounded, flexible, scalable, and minimally-interfering. However, their weakness is faithfulness. This is because the output zt merely reflects what information is present in mt, and not whether that information is actually used in the computation of πt. Defining faithfulness. We now chart a path forward from decoders, with an aim of making them faithful. To do so, we first define two metrics for faithfulness, one weak, one strong: Correlational faithfulness: if the decoder reports a value of zt = z, to what degree is the behavior (from the policy π) consistent with this value, z? Causal faithfulness:4 if we were to counterfactually intervene on zt, changing it to a value of z , to what degree would this produce behavior consistent with this new value, z , rather than its original value, z? Causal faithfulness is a much stronger property than correlational faithfulness. It s highly desirable because it allows one to validate the agent s commitment to the value of zt at runtime by making interventions on it. This validation mechanism is especially useful when out of distribution. Re-routed decoders. While it is possible for a decoder to have a degree of correlational faithfulness, it is structurally impossible for it to exhibit causal faithfulness. A simple fix is to route the output of the decoder back into the internal representation (Figure 1b), e.g. by concatenating zt with xt+1. This amounts to incorporating the decoder output in the state update rule, with mt+1 = fθ([xt+1, zt], mt). The difficulty with this simple solution is that there is no guarantee that fθ will actually "listen" to zt when producing mt+1. This is because the information in zt is largely if not completely redundant with mt. This may be further exacerbated by the presence of any stochastic sampling in the generation of zt from mt. Indeed, as shall be described in Section 6, our experiments find that re-routed decoders make no measurable progress beyond ordinary decoders on causal faithfulness. How do we render the function fθ sensitive to the value of z, when there is redundancy between zt and mt? It is useful to frame the problem in terms of communication. The decoder pathway can be viewed as a (externally-grounded) communication channel between the internal state and itself, i.e. a form of "self-talk" or "inner speech" [67]. The problem therein is that the speaker (mt) and the listener (also mt) are identical, so there is no useful information in the message zt for the listener. Causal self-talk. To overcome this problem, we need to create an information asymmetry between speaker and listener. We introduce Causal Self-Talk (CST): we encourage fθ to be sensitive to the message zt by creating an augmented data distribution where the speaker and listener diverge. Fortunately, we have a ready source of alternate listeners: internal states mt for t < t. We thus provide the decoder output with an additional role: to compactly and effectively communicate information to earlier versions of the same agent (from the same episode), via a semanticallygrounded communication channel. Formally, this takes the form of an intervention (Fig 1c): the internal state, mt, is replaced with a counterfactual one, mt , before being integrated with the original message (zt) and next observation (xt+1), in order to produce a counterfactual internal state, m(t t ) t+1 = fθ([xt+1, zt], mt ). This yields a counterfactual policy, π(t t ) t+1 = hθ(m(t t ) t+1 ). This alone, however, is insufficient to fully specify a CST algorithm. We must also stipulate the desired outcome from enacting these interventions. We describe three alternative choices: CST-RL: The messages zt should allow earlier versions of the same agent to maximize return given the current environment state. To train this, interventions must occur online (according to some schedule), and we must allow the agent to act according to the counterfactual policy following an intervention. We then use the maximization of subsequent discounted reward as the desired outcome from the intervention, via reinforcement learning (RL). 4Our use of this term is distinct from the causal faithfulness condition in causal modelling [64]. CST-MR: The messages zt should allow earlier versions of the same agent to reconstruct the next internal state, mt+1. Unlike CST-RL, this can be done entirely in replay, without requiring the agent to ever act according to the counterfactual policy on a live environment. We add a weighted memory reconstruction (MR) term to the overall loss: LMR = ||m(t t ) t+1 mt+1||2. CST-PD: The messages zt should allow earlier versions of the same agent to recover the true current (and future) policy. As with CST-MR, interventions can be simulated entirely in replay. We add a weighted policy distillation (PD) term to the overall loss, using discounting function γ( ): LP D = P t>0 γ( t) DKL πt+ t || π(t t ) t+ t . CST and the five desiderata. CST offers a different, less extreme tradeoff between the desiderata. By design, it inherits decoders resolution to several desiderata. Like decoders, CST can achieve grounding by supervising the representations z against a signal of known semantics. It is also equally flexible. Relative to decoders, there is some trade-off with minimal-interference as the decoder output zt is now an obligate input into the next time step. However, much like decoders, the effects on training can, in principle, be contained through appropriate gradient flow, e.g. by limiting updates to parameters of the state update function, fθ, and/or fine-tuning a pre-existing base system. CST is highly scalable. CST-RL is straightforward to implement; in its simplest incarnation, fixing t = 0 reduces to dropout on the whole internal state, mt. The use of RL, however, occurs at the expense of additional data as it requires online execution of the counterfactual policy. CST-PD and CST-MR, in contrast, are entirely self-supervised: they make use of the existing trajectories generated during the training of the base system. These methods augment the data in a manner similar to the interchange intervention technique of [25], using self-imitation as the desired outcome of the intervention. Depending on the degree of non-stationarity afforded by the particular environment (and its degree of partial observability), the size of the augmented dataset can scale quadratically with the length of each episode (from choosing both t and t ); in some settings, it may be possible to exceed this by sampling counterfactual previous states mt from other episodes. Finally, unlike decoders, CST is explicitly trained to provide causal faithfulness, by facilitating counterfactual control. Our goal in the following two sections is to evaluate how well it does this. 5 Experiments Task. We study variants of CST in a 3D virtual environment built in Unity [1, 68]. This features a fixed-layout indoor space comprising five rooms, each with a different wall color. The agent receives as input a 1st-person visual observation and a text observation. The agent can freely navigate using a 4D continuous action space (a [ 1, 1]4) consisting of in-plane movements and a 2-DOF rotation. We developed a simple task in this environment, called Dax Ducks (Figure 2). We designed Dax Ducks to: (1) be easy for an agent to learn a good policy; (2) require an agent to maintain and update beliefs about the latent environment state in order to maximize reward; (3) provide a source of data to ground the self-model; and (4) provide a means to evaluate the faithfulness of the learned self-model. Dax Ducks is an instruction-based, fast-binding search task, where the agent is rewarded for finding a duck that exhibits an instructed tag. Each episode consists of a sequence of trials. At the beginning of each trial, the agent s avatar spawns in the center of the middle room facing a random direction. Four identical ducks spawn at the center of the four outer rooms. Each duck is randomly assigned one of four tags ( dax , gavagai , kleeg , or plork ). The agent is instructed to collect (i.e. collide with) the duck with a specified tag via the text observation channel, e.g. Collect a gavagai. The agent may observe a duck s tag by entering the duck s room and orienting the center of its view towards the duck. This adds an additional string to the text observation, e.g. This is a kleeg. . Colliding with any duck terminates the current trial, in which case the agent is respawned in the center of the middle room again and a random new instruction is delivered; a reward is also delivered if the agent collided with the correct duck. Finally, when a new trial starts, the ducks usually maintain their tags, but with a small probability (psh = 0.1), the tags are randomly shuffled between the ducks. Thus the agent can only maximize reward by effectively retaining a belief state over which tag belongs to which duck. Each episode lasts for 5000 steps regardless of the number of trials completed. Agent spawns in central room and receives instruction Agent approaches duck and observes its tag gavagai plork Agent touches duck and, if correct tag, receives reward gavagai plork gavagai plork Shuffle tags Agent respawns in central room and receives instruction Figure 2: The Dax Ducks task. Top-down views are for exposition only. Agent architecture. For the base system, we used a standard model-free agent (Appendix A): inputs encoded with a Res Net (images) and LSTM (text); a LSTM memory module; MLPs for policy and value heads; and trained with V-trace [22]. All agents achieved a similarly high return. We demonstrate the flexibility of grounding by considering two forms for the self-model. In the one-hot form, the message zt comprises four one-hot vectors (one per tag), expressing which room the agent believes each tag to be in. In the language form, the message zt is a synthetic language string, expressing which room the agent believes a single chosen tag to be in. For the one-hot form (Section 6), we parameterize q(zt) = Q4 τ=1 q(zτ t ) as a product of independent categorical distributions q(zτ t ) for the four tags τ, with gθ (an MLP) outputting the parameters of each q(zτ t ) at time t, and a sample z q(zt) concatenated with the encoded input at the next time step. For the language form (Section 6), we use an LSTM for gθ, sampling zt directly. Examples from a single trajectory are shown in Figure 3a (one-hot) and Figure 4a (language). We ground both forms of self-model by supervising zt against ground-truth values. For the language form, this amounts to constructing a target ˆzt of the form: The is in the room. . Given the simplicity of the environment, we implement all three CST algorithms using t = 0 throughout. This means that CST aims to drive the decoder to communicate its beliefs to the earlier version of the agent from the start of the episode. Although this reduces the scale of data augmentation from quadratic to linear, and requires interventions on z to be accompanied by a reset of the memory state to m0, in return it makes it easier to measure the effects of interventions on the agent s behavior. We used the following schedules for interventions at training time. For CST-RL, interventions (mt m0) occurred randomly, with probability p = 0.03 that an intervention would occur at any given timestep t. For CST-MR, we simulated interventions in replay at every timestep. For CST-PD, we simulated interventions in replay: we divided every trajectory into a sequence of blocks of variable duration, with a p = 0.03 probability that a new block would start at any time t; we computed LP D only up to a horizon of the end of the block, and with a constant discounting function, γ( t) = 1. One-hot self-talk. We start by considering the one-hot form for the messages z (Figure 3a). We compare CST against two baselines: the ordinary decoder (Ord-Dec; Figure 1a) and the re-routed decoder (RR-Dec; Figure 1b). We focus on the metrics of faithfulness previously introduced. Correlational faithfulness. When no interventions are taking place, is the agent s physical behavior congruent with the output of its self-model? To measure this, we filter to all times in evaluation episodes when the agent s avatar is in the central room. We then determine the degree to which the belief attested in q(zt) matches the next-visited room, r (whether or not (a) Example one-hot messages from a CST-RL agent. correlational faithfulness causal faithfulness (b) Faithfulness results. Figure 3: One-hot-based self-talk. (a) Example messages during a trajectory. Top-down view (left), agent view (middle), and q(zt) (right) during a trajectory of Dax Ducks, showing the updating of q(zt) after the start of a new trial (second row) and after discovering the tags have been shuffled (third row). At each time point, the message zt is sampled from q(zt). (b) CST renders the decoder causally faithful. Average values across 5 N 10 training runs per method; error bars show 95% confidence intervals. Dashed lines indicate faithfulness expected from random messages. the instructed tag is actually in r). For the one-hot message form, this amounts to reading off the value of q(zτ t = r) for the instructed tag τ. Causal faithfulness. When we intervene on zt, is the agent s subsequent physical behavior congruent with this new value? To measure this, we run evaluation episodes where, at the start of each trial, we inject a message z indicating that the instructed tag is in some random room r . We then measure the probability that the agent actually visits room r next. CST maintains comparable levels of correlational faithfulness as the baseline decoders. However, the decoder baselines failed to exhibit a meaningful degree of causal faithfulness. This is indeed impossible for Ord-Dec given the lack of any causal pathway from zt to mt+1. Re-routing z back to the internal state in RR-Dec does result in a small increase above chance, but the effect is relatively tiny. In contrast, all CST objectives substantially increase the degree of causal faithfulness (Figure 3b). Language-based self-talk. Next, we trained agents to model their belief state using messages z in a synthetic language form (e.g. Figure 4a). For correlational faithfulness, we obtain analogous values of q(zτ t = r) by computing the likelihoods of generating the four messages The is in the room. , and normalizing these to sum to 1. As with the one-hot form, we found that the greatest gain of CST was its ability to imbue the self-model with causal faithfulness. The effect was most profound with CST-RL, and moderate with CST-PD. However, we found it difficult to get any traction with CST-MR; this was either disruptive to learning the base policy, or ineffective at causal faithfulness. We note that the faithfulness results with the language form were overall weaker than for the one-hot form. This may be because the one-hot messages encode a room belief about all four tags. Semantic control. As a final test, we consider whether CST endows the self-model with sufficient causal power to act as an effective semantic control interface for the agent. We thus measure how "Sampled memo" Other samples (a) Example language messages from a CST-RL agent. correlational faithfulness causal faithfulness (b) Faithfulness results. Figure 4: Language-based self-talk. Results as in Figure 3. one-hot form language form semantic control efficacy Figure 5: Efficacy of semantic control via interventions on the self-model. effectively the agent can use injected information to obtain reward. We start with two baselines: (1) the return on evaluation episodes where psh = 1, i.e. where tags are shuffled on every trial; and (2) the return on evaluation episodes where psh = 0, i.e. where tags are never shuffled. Respectively, these represent the bounds of rewards the agent naturally obtains when it can acquire (1) minimal or (2) maximal information about the location of the instructed tag. We compare these to the reward obtained on evaluation episodes where information about the tag location is provided through intervening on z. Specifically, we set psh = 1, and inject a message z at the start of each trial indicating the location of all the tags (for the one-hot form), or the location of the instructed tag (for the language form). The results (Figure 5) show that CST not only furnishes an agent with a semantically-grounded model that describes its behavior, but also one which can be used as a mechanism to guide it. 7 Discussion We articulated five key desiderata of an ideal explanation system for an AI system: grounding, flexibility, minimal-interference, scalability, and faithfulness. Unlike existing approaches, we seek to satisfy all of these by training the base system to serve a causal model of itself. We developed an instance of this solution for Deep RL agents: Causal Self-Talk. In CST, an agent learns to serve a grounded and faithful model of its internal state by learning to communicate with itself across time. We demonstrated its effectiveness under various forms in a simple embodied task in a simulated 3D environment. We also showed that CST enables a new mechanism for effective semantic control. We presented three variants of the CST algorithm. CST-RL yielded the greatest degree of causal faithfulness and efficacy of semantic control. Here, online interventions at training time force the agent to learn how to use the content of its self-model to maximize its reward. CST-RL s success, however, may reflect features of the particular task setting we study: here, there is a close alignment between writing a faithful z and maximizing reward. This may not always be the case. In contrast, CST-PD offers an attractive alternative for two reasons: (1) it trains entirely in replay, and thus potentially interferes less with the base system; and (2) as it primarily seeks to enable imitation of the current behavior, the faithfulness of the self-model is targeted by design. One challenge here is learning to capture the important features of behavior; more impressive gains may be possible using richer imitation learning techniques [e.g. 34] rather than behavioral cloning. We had the least success with CST-MR: there also may be theoretical obstacles to deploying CST-MR in some settings, e.g. when the dimensionality of the internal state is very large, or is of variable length. Limitations. In our experiments, we use a limited form of CST, where we exclusively revert memory states mt to the state at time t = 0, rather than the more general case of t < t. We do this to simplify the presentation and evaluation of the technique, but it imparts two limitations: (1) it requires mt to be reverted to m0 in order to effectively intervene on the message z; and (2) it requires z to contain all behaviorally-relevant information. While sufficient for the simple task at hand, this becomes unrealistic in more complex environments. When z contains only a partial description of mt, it will be necessary for the agent to be able to re-integrate injected z values with its current memory state. The more general case of t < t offers a viable path to achieving this. Utility requires that the self-model be grounded in a semantically-meaningful form. We have shown how a ready source of data can be used to train the self-model. How this method stands up in more limited data regimes, or in OOD scenarios where the self-model hasn t been supervised, remains an open question. It may be possible in some domains to leverage semi-supervised methods, and/or to impose structure on the self-model as a helpful constraint. Future opportunities. We focused here on expressing the agent s beliefs about the current state of the environment in the self-model. A complete answer to the question, "Why did you do that?" might also include an expression of an agent s goals, plans, and model of the world [53]. The schema we have developed here should be sufficiently flexible to extend in this direction, given the right training data. It would also be of great value to determine how to apply these techniques beyond RL agents. Our primary focus in this paper has been on building self-models to yield explanations. However, this is only one of several utilities that a faithful self-model might deliver. For example, we demonstrate that causal self-models provide a semantic control mechanism. This could be useful for guiding agent learning through human interaction [1], metacognitive reporting and control [9, 24, 29], or as an interface for safety-critical applications [41, 45, 51, 61]. By learning to express its internal state through a different representational schema, this may also open up avenues for new exploration or hierarchical reinforcement learning techniques. Finally, training agents to constantly talk to themselves might prove to be a data-efficient way of enabling agents to talk to others (and us!). Ultimately, this may pave a way for us to invert of the flow of knowledge, allowing agents to not only answer, "Why did you do that?", but also "How?". Acknowledgments and Disclosure of Funding The authors would like to thank Edgar Duéñez-Guzmán for early environment design, Hamza Merzic for agent infrastructure assistance, Allison Tam for help in agent development, Chen Yan for useful discussion, Lucy Campbell-Gillingham for organizational support, and Johannes Welbl, Kory Mathewson, Pedro Ortega, Agnieszka Grabska-Barwi nska, Chen Yan, Matt Botvinick, and Lisa Anne Hendricks for comments on the work and/or the manuscript. [1] Josh Abramson, Arun Ahuja, Iain Barr, Arthur Brussee, Federico Carnevale, Mary Cassin, Rachita Chhaparia, Stephen Clark, Bogdan Damoc, Andrew Dudzik, and Others. Imitating interactive intelligence. ar Xiv preprint ar Xiv:2012.05672, 2020. [2] Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, and Been Kim. Sanity checks for saliency maps. Adv. Neural Inf. Process. Syst., 31, 2018. [3] Zeynep Akata, Lisa Anne Hendricks, Stephan Alaniz, and Trevor Darrell. Generating post-hoc rationales of deep visual classification decisions. In The Springer Series on Challenges in Machine Learning, pages 135 154. Springer International Publishing, Cham, 2018. [4] David Alvarez-Melis and Tommi S Jaakkola. Towards robust interpretability with Self Explaining neural networks. ar Xiv preprint ar Xiv:1806.07538, June 2018. [5] Jacob Andreas, Dan Klein, and Sergey Levine. Modular multitask reinforcement learning with policy sketches. ar Xiv preprint ar Xiv:1611.01796, November 2016. [6] Rushil Anirudh, P Bremer, Rahul Sridhar, and JJ Thiagarajan. Influential sample selection: A graph signal processing approach. Technical report, Lawrence Livermore National Lab.(LLNL), Livermore, CA (United States), 2017. [7] Akanksha Atrey, Kaleigh Clary, and David Jensen. Exploratory not explanatory: Counterfactual analysis of saliency maps for deep reinforcement learning. ar Xiv preprint ar Xiv:1912.05743, 2019. [8] Sebastian Bach, Alexander Binder, Grégoire Montavon, Frederick Klauschen, Klaus-Robert Müller, and Wojciech Samek. On Pixel-Wise explanations for Non-Linear classifier decisions by Layer-Wise relevance propagation. PLOS ONE, 10(7):e0130140, 2015. [9] Andrea Banino, Jan Balaguer, and Charles Blundell. Pondernet: Learning to ponder. ar Xiv preprint ar Xiv:2107.05407, 2021. [10] Peter W Battaglia, Jessica B Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan Faulkner, Caglar Gulcehre, Francis Song, Andrew Ballard, Justin Gilmer, George Dahl, Ashish Vaswani, Kelsey Allen, Charles Nash, Victoria Langston, Chris Dyer, Nicolas Heess, Daan Wierstra, Pushmeet Kohli, Matt Botvinick, Oriol Vinyals, Yujia Li, and Razvan Pascanu. Relational inductive biases, deep learning, and graph networks. ar Xiv preprint ar Xiv:1806.01261, June 2018. [11] David Bau, Bolei Zhou, Aditya Khosla, Aude Oliva, and Antonio Torralba. Network dissection: Quantifying interpretability of deep visual representations. ar Xiv preprint ar Xiv:1704.05796, April 2017. [12] Yonatan Belinkov and James Glass. Analysis methods in neural language processing: A survey. Transactions of the Association for Computational Linguistics, 7:49 72, 2019. [13] Przemyslaw Biecek and Tomasz Burzykowski. Local interpretable model-agnostic explanations (LIME). Explanatory Model Analysis, pages 107 123, 2021. [14] Jacob Bien and Robert Tibshirani. Prototype selection for interpretable classification. The Annals of Applied Statistics, 5(4):2403 2424, 2011. [15] Tyler Bonnen, Daniel L K Yamins, and Anthony D Wagner. When the ventral visual stream is not enough: A deep learning account of medial temporal lobe involvement in perception. Neuron, 109(17):2755 2766.e6, September 2021. [16] Christopher P Burgess, Loic Matthey, Nicholas Watters, Rishabh Kabra, Irina Higgins, Matt Botvinick, and Alexander Lerchner. MONet: Unsupervised scene decomposition and representation. ar Xiv preprint ar Xiv:1901.11390, January 2019. [17] Nick Cammarata, Gabriel Goh, Shan Carter, Ludwig Schubert, Michael Petrov, and Chris Olah. Curve detectors. Distill, 5(6), June 2020. [18] Hyeoncheol Cho, Eok Kyun Lee, and Insung S Choi. Layer-wise relevance propagation of Interaction Net explains protein ligand interactions at the atom level. Scientific Reports, 10(1), 2020. [19] Grégoire Déletang, Jordi Grau-Moya, Miljan Martic, Tim Genewein, Tom Mc Grath, Vladimir Mikulik, Markus Kunesch, Shane Legg, and Pedro A Ortega. Causal analysis of agent behavior for ai safety. ar Xiv preprint ar Xiv:2103.03938, 2021. [20] Yanai Elazar, Shauli Ravfogel, Alon Jacovi, and Yoav Goldberg. Amnesic probing: Behavioral explanation with amnesic counterfactuals. Transactions of the Association for Computational Linguistics, 9:160 175, 2021. [21] Nelson Elhage, Neel Nanda, Catherine Olsson, Tom Henighan, Nicholas Joseph, Ben Mann, Amanda Askell, Yuntao Bai, Anna Chen, Tom Conerly, Nova Das Sarma, Dawn Drain, Deep Ganguli, Zac Hatfield-Dodds, Danny Hernandez, Andy Jones, Jackson Kernion, Liane Lovitt, Kamal Ndousse, Dario Amodei, Tom Brown, Jack Clark, Jared Kaplan, Sam Mc Candlish, and Chris Olah. A mathematical framework for transformer circuits. Transformer Circuits Thread, 2021. [22] Lasse Espeholt, Hubert Soyer, Remi Munos, Karen Simonyan, Vlad Mnih, Tom Ward, Yotam Doron, Vlad Firoiu, Tim Harley, Iain Dunning, and Others. Impala: Scalable distributed deep-rl with importance weighted actor-learner architectures. In International Conference on Machine Learning, pages 1407 1416, 2018. [23] Jeffrey De Fauw, Jeffrey De Fauw, Joseph R Ledsam, Bernardino Romera-Paredes, Stanislav Nikolov, Nenad Tomasev, Sam Blackwell, Harry Askham, Xavier Glorot, Brendan O Donoghue, Daniel Visentin, George van den Driessche, Balaji Lakshminarayanan, Clemens Meyer, Faith Mackinder, Simon Bouton, Kareem Ayoub, Reena Chopra, Dominic King, Alan Karthikesalingam, Cían O Hughes, Rosalind Raine, Julian Hughes, Dawn A Sim, Catherine Egan, Adnan Tufail, Hugh Montgomery, Demis Hassabis, Geraint Rees, Trevor Back, Peng T Khaw, Mustafa Suleyman, Julien Cornebise, Pearse A Keane, and Olaf Ronneberger. Clinically applicable deep learning for diagnosis and referral in retinal disease. Nature Medicine, 24(9): 1342 1350, 2018. [24] Damien S Fleur, Bert Bredeweg, and Wouter van den Bos. Metacognition: ideas and insights from neuroand educational sciences. NPJ Sci Learn, 6(1):13, June 2021. [25] Atticus Geiger, Zhengxuan Wu, Hanson Lu, Josh Rozner, Elisa Kreiss, Thomas Icard, Noah D Goodman, and Christopher Potts. Inducing causal structure for interpretable neural networks. ar Xiv preprint ar Xiv:2112.00826, 2021. [26] Gabriel Goh, Nick Cammarata , Chelsea Voss , Shan Carter, Michael Petrov, Ludwig Schubert, Alec Radford, and Chris Olah. Multimodal neurons in artificial neural networks. Distill, 2021. [27] Yash Goyal, Ziyan Wu, Jan Ernst, Dhruv Batra, Devi Parikh, and Stefan Lee. Counterfactual visual explanations. In International Conference on Machine Learning, pages 2376 2384, 2019. [28] Riccardo Guidotti, Anna Monreale, Salvatore Ruggieri, Franco Turini, Fosca Giannotti, and Dino Pedreschi. A survey of methods for explaining black box models. ACM computing surveys (CSUR), 51(5):1 42, 2018. [29] Jessica B Hamrick, Andrew J Ballard, Razvan Pascanu, Oriol Vinyals, Nicolas Heess, and Peter W Battaglia. Metacontrol for adaptive imagination-based optimization. ar Xiv preprint ar Xiv:1705.02670, 2017. [30] Bradley Hayes and Julie A Shah. Improving robot controller transparency through autonomous policy explanation. In Proceedings of the 2017 ACM/IEEE International Conference on Human Robot Interaction, New York, NY, USA, March 2017. ACM. [31] Lisa Anne Hendricks, Zeynep Akata, Marcus Rohrbach, Jeff Donahue, Bernt Schiele, and Trevor Darrell. Generating visual explanations. In European conference on computer vision, pages 3 19, 2016. [32] Lisa Anne Hendricks, Ronghang Hu, Trevor Darrell, and Zeynep Akata. Generating counterfactual explanations with natural language. ar Xiv preprint ar Xiv:1806.09809, 2018. [33] Lisa Anne Hendricks, Anna Rohrbach, Bernt Schiele, Trevor Darrell, and Zeynep Akata. Generating visual explanations with natural language. Applied AI Letters, 2(4), December 2021. [34] Jonathan Ho and Stefano Ermon. Generative adversarial imitation learning. Adv. Neural Inf. Process. Syst., 29, 2016. [35] Alon Jacovi and Yoav Goldberg. Towards faithfully interpretable NLP systems: How should we define and evaluate faithfulness? In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 4198 4205, Online, July 2020. Association for Computational Linguistics. [36] Alon Jacovi, Ana Marasovi c, Tim Miller, and Yoav Goldberg. Formalizing trust in artificial intelligence: Prerequisites, causes and goals of human trust in AI. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, pages 624 635, 2021. [37] Max Jaderberg, Volodymyr Mnih, Wojciech Marian Czarnecki, Tom Schaul, Joel Z Leibo, David Silver, and Koray Kavukcuoglu. Reinforcement learning with unsupervised auxiliary tasks. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. Open Review.net, 2017. [38] Sarthak Jain and Byron C Wallace. Attention is not explanation. ar Xiv preprint ar Xiv:1902.10186, February 2019. [39] Zoe Juozapaitis, Anurag Koul, Alan Fern, Martin Erwig, and Finale Doshi-Velez. Explainable reinforcement learning via reward decomposition. In IJCAI/ECAI Workshop on explainable artificial intelligence, 2019. [40] Bilal Kartal, Pablo Hernandez-Leal, and Matthew E Taylor. Terminal prediction as an auxiliary task for deep reinforcement learning. In Proceedings of the Fifteenth AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, AIIDE 19, Atlanta, Georgia, 2019. AAAI Press. [41] Huda Khayrallah, Sean Trott, and Jerome Feldman. Natural language for human robot interaction. In International Conference on Human-Robot Interaction (HRI), 2015. [42] Been Kim, Cynthia Rudin, and Julie A Shah. The bayesian case model: A generative approach for case-based reasoning and prototype classification. Advances in neural information processing systems, 27, 2014. [43] Been Kim, Rajiv Khanna, and Oluwasanmi O Koyejo. Examples are not enough, learn to criticize! criticism for interpretability. Advances in neural information processing systems, 29, 2016. [44] Been Kim, Martin Wattenberg, Justin Gilmer, Carrie Cai, James Wexler, Fernanda Viegas, and Rory Sayres. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (TCAV). ar Xiv preprint ar Xiv:1711.11279, November 2017. [45] Jinkyu Kim, Suhong Moon, Anna Rohrbach, Trevor Darrell, and John Canny. Advisable learning for self-driving vehicles by internalizing observation-to-action rules. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, June 2020. [46] Pang Wei Koh and Percy Liang. Understanding black-box predictions via influence functions. In International conference on machine learning, pages 1885 1894. PMLR, 2017. [47] Pang Wei Koh, Thao Nguyen, Yew Siang Tang, Stephen Mussmann, Emma Pierson, Been Kim, and Percy Liang. Concept bottleneck models. ar Xiv preprint ar Xiv:2007.04612, July 2020. [48] Guillaume Lample and Devendra Singh Chaplot. Playing FPS games with deep reinforcement learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI 17, pages 2140 2146, San Francisco, California, USA, 2017. AAAI Press. [49] Zachary C Lipton. The mythos of model interpretability: In machine learning, the concept of interpretability is both important and slippery. Queueing Syst., 16(3):31 57, 2018. [50] Prashan Madumal, Tim Miller, Liz Sonenberg, and Frank Vetere. Explainable reinforcement learning through a causal lens. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 2493 2500, 2020. [51] Cynthia Matuszek, Evan Herbst, Luke Zettlemoyer, and Dieter Fox. Learning to parse natural language commands to a robot control system. In Experimental robotics, pages 403 415, 2013. [52] Thomas Mc Grath, Andrei Kapishnikov, Nenad Tomašev, Adam Pearce, Demis Hassabis, Been Kim, Ulrich Paquet, and Vladimir Kramnik. Acquisition of chess knowledge in Alpha Zero. ar Xiv preprint ar Xiv:2111.09259, November 2021. [53] Tim Miller. Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence, 267:1 38, 2019. [54] Piotr W Mirowski, Razvan Pascanu, Fabio Viola, Hubert Soyer, Andy Ballard, Andrea Banino, Misha Denil, Ross Goroshin, L Sifre, Koray Kavukcuoglu, Dharshan Kumaran, and Raia Hadsell. Learning to navigate in complex environments. ar Xiv preprint ar Xiv:1611.03673, 2017. [55] Alexander Mordvintsev, Christopher Olah, and Mike Tyka. Inceptionism: Going deeper into neural networks. Google Research Blog, 2015. [56] Alexander Mott, Daniel Zoran, Mike Chrzanowski, Daan Wierstra, and Danilo Jimenez Rezende. Towards interpretable reinforcement learning using attention augmented agents. Adv. Neural Inf. Process. Syst., 32, 2019. [57] Catherine Olsson, Nelson Elhage, Neel Nanda, Nicholas Joseph, Nova Das Sarma, Tom Henighan, Ben Mann, Amanda Askell, Yuntao Bai, Anna Chen, Tom Conerly, Dawn Drain, Deep Ganguli, Zac Hatfield-Dodds, Danny Hernandez, Scott Johnston, Andy Jones, Jackson Kernion, Liane Lovitt, Kamal Ndousse, Dario Amodei, Tom Brown, Jack Clark, Jared Kaplan, Sam Mc Candlish, and Chris Olah. In-context learning and induction heads. Transformer Circuits Thread, March 2022. [58] Dong Huk Park, Lisa Anne Hendricks, Zeynep Akata, Anna Rohrbach, Bernt Schiele, Trevor Darrell, and Marcus Rohrbach. Multimodal explanations: Justifying decisions and pointing to the evidence. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. IEEE, June 2018. [59] Abhilasha Ravichander, Yonatan Belinkov, and Eduard Hovy. Probing the probing paradigm: Does probing accuracy entail task relevance? ar Xiv preprint ar Xiv:2005.00719, 2020. [60] Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. Anchors: High-precision modelagnostic explanations. In Proceedings of the AAAI conference on artificial intelligence, volume 32, 2018. [61] Junha Roh, Chris Paxton, Andrzej Pronobis, Ali Farhadi, and Dieter Fox. Conditional driving from natural language instructions. In Conference on Robot Learning, pages 540 551, 2020. [62] Cynthia Rudin. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5):206 215, 2019. [63] Tianmin Shu, Caiming Xiong, and Richard Socher. Hierarchical and interpretable skill acquisition in multi-task reinforcement learning. ar Xiv preprint ar Xiv:1712.07294, December 2017. [64] Peter Spirtes, Clark N Glymour, Richard Scheines, and David Heckerman. Causation, prediction, and search. MIT press, 2000. [65] Ian Tenney, James Wexler, Jasmijn Bastings, Tolga Bolukbasi, Andy Coenen, Sebastian Gehrmann, Ellen Jiang, Mahima Pushkarna, Carey Radebaugh, Emily Reif, and Ann Yuan. The language interpretability tool: Extensible, interactive visualizations and analysis for NLP models. ar Xiv preprint ar Xiv:2008.05122, August 2020. [66] Nicholay Topin and Manuela Veloso. Generation of policy-level explanations for reinforcement learning. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 2514 2521, 2019. [67] L S Vygotsky. Thought and Language. The MIT Press. MIT Press, London, England, 2012. [68] Tom Ward, Andrew Bolt, Nik Hemmings, Simon Carter, Manuel Sanchez, Ricardo Barreira, Seb Noury, Keith Anderson, Jay Lemmon, Jonathan Coe, Piotr Trochim, Tom Handley, and Adrian Bolton. Using unity to help solve intelligence. ar Xiv preprint ar Xiv:2011.09294, November 2020. [69] Sarah Wiegreffe and Yuval Pinter. Attention is not not explanation. ar Xiv preprint ar Xiv:1908.04626, August 2019. [70] Herman Yau, Chris Russell, and Simon Hadfield. What did you think would happen? explaining agent behaviour through intended outcomes. Adv. Neural Inf. Process. Syst., 33:18375 18386, 2020. [71] Chih-Kuan Yeh, Joon Kim, Ian En-Hsu Yen, and Pradeep K Ravikumar. Representer point selection for explaining deep neural networks. Advances in neural information processing systems, 31, 2018. [72] Vinicius Zambaldi, David Raposo, Adam Santoro, Victor Bapst, Yujia Li, Igor Babuschkin, Karl Tuyls, David Reichert, Timothy Lillicrap, Edward Lockhart, and Others. Deep reinforcement learning with relational inductive biases. In International conference on learning representations, 2018. [73] Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In European conference on computer vision, pages 818 833, 2014. [74] Bolei Zhou, Yiyou Sun, David Bau, and Antonio Torralba. Interpretable basis decomposition for visual explanation. In Proceedings of the European Conference on Computer Vision (ECCV), September 2018. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Discussion. (c) Did you discuss any potential negative societal impacts of your work? [No] The work itself is focused on achieving explainability, which is a topic focused on positive societal impacts. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [No] Our agent training code is built on a core library that has yet to be open-sourced. If this library is published prior to Neur IPS camera ready deadline, we will endeavor to open-source the code for the design presented in this paper in the camera-ready version. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Appendix A. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] All results figures include error bars. (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See Appendix A. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [No] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]