# metatrained_agents_implement_bayesoptimal_agents__dfd78da4.pdf Meta-trained agents implement Bayes-optimal agents Vladimir Mikulik , Grégoire Delétang , Tom Mc Grath , Tim Genewein , Miljan Martic, Shane Legg, Pedro A. Ortega Deep Mind London, UK Memory-based meta-learning is a powerful technique to build agents that adapt fast to any task within a target distribution. A previous theoretical study has argued that this remarkable performance is because the meta-training protocol incentivises agents to behave Bayes-optimally. We empirically investigate this claim on a number of prediction and bandit tasks. Inspired by ideas from theoretical computer science, we show that meta-learned and Bayes-optimal agents not only behave alike, but they even share a similar computational structure, in the sense that one agent system can approximately simulate the other. Furthermore, we show that Bayes-optimal agents are fixed points of the meta-learning dynamics. Our results suggest that memory-based meta-learning might serve as a general technique for numerically approximating Bayes-optimal agents that is, even for task distributions for which we currently don t possess tractable models. 1 Introduction Within the paradigm of learning-to-learn, memory-based meta-learning is a powerful technique to create agents that adapt fast to any task drawn from a target distribution [1 6]. In addition, it has been claimed that meta-learning might be a key tool for creating systems that generalize to unseen environments [7]. This claim is also partly supported by studies in computational neuroscience, where experimental studies with human subjects have shown that fast skill adaptation relies on task variation [8, 9]. Due to this, understanding how meta-learned agents acquire their representational structure and perform their computations is of paramount importance, as it can inform architectural choices, design of training tasks, and address questions about generalisation and safety in artificial intelligence. Previous theoretical work has argued that agents that fully optimise a meta-learning objective are Bayes-optimal by construction, because meta-learning objectives are Monte-Carlo approximations of Bayes-optimality objectives [10]. This is striking, as Bayes-optimal agents maximise returns (or minimise loss) by optimally trading off exploration versus exploitation [11]. The theory also makes a stronger, structural claim: namely, that meta-trained agents perform Bayesian updates under the hood , where the computations are implemented via a state machine embedded in the memory dynamics that tracks the sufficient statistics of the uncertainties necessary for solving the task class. Here we set out to empirically investigate the computational structure of meta-learned agents. However, this comes with non-trivial challenges. Artificial neural networks are infamous for their hard-to-interpret computational structure: they achieve remarkable performance on challenging tasks, but the computations underlying that performance remain elusive. Thus, while much work in explainable machine learning focuses on the I/O behaviour or memory content, only few investigate the internal dynamics that give rise to them through careful bespoke analysis see e.g. [12 18]. Equal contribution Correspondence to {vmikulik|gdelt|mcgrathtom|timgen|pedroortega}@google.com 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. To tackle these challenges, we adapt a relation from theoretical computer science to machine learning systems. Specifically, to compare agents at their computational level [19], we verify whether they can approximately simulate each other. The quality of the simulation can then be assessed in terms of both state and output similarity between the original and the simulation. Thus, our main contribution is the investigation of the computational structure of RNN-based metalearned solutions. Specifically, we compare the computations of meta-learned agents against the computations of Bayes-optimal agents in terms of their behaviour and internal representations on a set of prediction and reinforcement learning tasks with known optimal solutions. We show that on these tasks: Meta-learned agents behave like Bayes-optimal agents (Section 4.1). That is, the predictions and actions made by meta-learned agents are virtually indistinguishable from those of Bayes-optimal agents. During the course of meta-training, meta-learners converge to the Bayes-optimal solution (Section 4.2). We empirically show that Bayes-optimal policies are the fixed points of the learning dynamics. Meta-learned agents represent tasks like Bayes-optimal agents (Section 4.3). Specifically, the computational structures correspond to state machines embedded in (Euclidean) memory space, where the states encode the sufficient statistics of the task and produce optimal actions. We can approximately simulate computations performed by meta-learned agents with computations performed by Bayes-optimal agents. 2 Preliminaries Memory-based meta-learning Memory-based meta-learners are agents with memory that are trained on batches of finite-length roll-outs, where each roll-out is performed on a task drawn from a distribution. The emphasis on memory is crucial, as training then performs a search in algorithm space to find a suitable adaptive policy [20]. The agent is often implemented as a neural network with recurrent connections, like an RNN, most often using LSTMs [4, 21], or GRUs [22]. Such a network computes two functions fw and gw using weights w W, yt = fw(xt, st 1) (output function) st = gw(xt, st 1), (state-transition function) (1) that map the current input and previous state pair (xt, st 1) X S into the output yt Y and the next state st S respectively. Here, X, Y, S, and W are all vector spaces over the reals R. An input xt X encodes the instantaneous experience at time t, such as e.g. the last observation, action, and feedback signal; and an output yt Y contains e.g. the logits for the current prediction or action probabilities. RNN meta-learners are typically trained using backpropagation through time (BPTT) [23, 24]. For fixed weights w, and combined with a fixed initial state s0, equations (1) define a state machine3. This state machine can be seen as an adaptive policy or an online learning algorithm. Bayes-optimal policies as state machines Bayes-optimal policies have a natural interpretation as state machines following (1). Every such policy can be seen as a state-transition function g, which maintains sufficient statistics (i.e., a summary of the past experience that is statistically sufficient to implement the prediction/action strategy) and an output function f, which uses this information to produce optimal outputs (i.e., the best action or prediction given the observed trajectory) [11, 10]. For instance, to implement an optimal policy for a multi-armed bandit with (independent) Bernoulli rewards, it is sufficient to remember the number of successes and failures for each arm. Comparisons of state machines via simulation To compare the policies of a meta-trained and a Bayes-optimal agent in terms of their computational structure, we adapt a well-established methodology from the state-transition systems literature [28 30]. Specifically, we use the concept of simulation to compare state machines. 3More precisely, a Mealy machine [25 27]. Formally, we have the following. A trace in a state machine is a sequence s0x1s1 x T s T of transitions. Since the state machines we consider are deterministic, a given sequence of inputs x1, . . . , x T induces a unique trace in the state machine. A deterministic state machine M simulates another machine N, written N M, if every trace in N has a corresponding trace in M on which their output functions agree. More precisely, N M if there exists a function φ mapping the states of N into the states of M such that the following two conditions hold: (transitions) for any trace s0x1s1 s T in N, the transformed trace φ(s0)x1φ(s1) φ(s T ) is also a trace in M; (outputs) for any state s of N and any input x, the output of machine N at (x, s) coincides with the output of machine M at (x, φ(s)). Intuitively, this means there is a consistent way of interpreting every state in N as a state in M, such that every computation in N can be seen as a computation in M. When both M N and N M hold, then we consider both machines to be computationally equivalent. 3.1 Tasks and agents Tasks Since our aim is to compare against Bayes-optimal policies, we consider 10 prediction and 4 reinforcement learning tasks for which the Bayes-optimal solution is analytically tractable. All tasks are episodic (T = 20 time steps), and the task parameters θ are drawn from a prior distribution p(θ) at the beginning of each episode. A full list of tasks is shown in Figure 4 and details are discussed in Appendix A. In prediction tasks the goal is to make probabilistic predictions of the next observation given past observations. All observations are drawn i.i.d. from an observational distribution. To simplify the computation of the optimal predictors, we chose observational distributions within the exponential family that have simple conjugate priors and posterior predictive distributions, namely: Bernoulli, categorical, exponential, and Gaussian. In particular, their Bayesian predictors have finite-dimensional sufficient statistics with simple update rules [31 33]. In reinforcement learning tasks the goal is to maximise the discounted cumulative sum of rewards in two-armed bandit problems [34]. We chose bandits with rewards that are Bernoullior Gaussiandistributed. The Bayes-optimal policies for these bandit tasks can be computed in polynomial time by pre-computing Gittins indices [35, 34, 36]. Note that the bandit tasks, while conceptually simple, already require solving the exploration versus exploitation problem [37]. RNN meta-learners Our RNN meta-learners consist of a three-layer network architecture: one fully connected layer (the encoder), followed by one LSTM layer (the memory), and one fully connected layer (the decoder) with a linear readout producing the final output, namely the parameters of the predictive distribution for the prediction tasks, and the logits of the softmax action-probabilities for the bandit tasks respectively. The width of each layer is the same and denoted by N. We selected4 N = 32 for prediction tasks and N = 256 for bandit tasks. Networks were trained with BPTT [23, 24] and Adam [38]. In prediction tasks the loss function is the log-loss of the prediction. In bandit tasks the agents were trained to maximise the return (i.e., the discounted cumulative reward) using the Impala [39] policy gradient algorithm. See Appendix B.2 for details on network architectures and training. 3.2 Behavioral analysis The aim of our behavioural analysis is to compare the input-output behaviour of a meta-learned (RNN) and a Bayes-optimal agent (Opt). For prediction tasks, we feed the same observations to both agent 4Note that the (effective) network capacity needs to be large enough to at least represent the different states required by the Bayes-optimal solution. However, it is currently unknown how to precisely measure effective network capacity. We thus selected our architectures based on preliminary ablations that investigate convergence speed of training. See Appendix D.3 for details. types and then compute their dissimilarity as the sum of the KL-divergences of the instantaneous predictions averaged over K trajectories, that is, d(Opt, RNN) = 1 t=1 DKL πOpt t πRNN t . (2) Bandit tasks require a different dissimilarity measure: since there are multiple optimal policies, we cannot compare action probabilities directly. A dissimilarity measure that is invariant under optimal policies is the empirical reward difference: d(Opt, RNN) = 1 t=1 (r Opt t r RNN t ) (3) where r Opt and r RNN are the empirical rewards collected during one episode. This dissimilarity measure only penalises policy deviations that entail reward differences. 3.3 Convergence analysis In our convergence analysis we investigate how the behaviour of meta-learners changes over the course of training. To characterise how a single RNN training run evolves, we evaluate the behavioural dissimilarity measures (Section 3.2), which compare RNN behaviour against Bayes-optimal behaviour, across many checkpoints of a training run. Additionally we study the RNN behaviour across multiple training runs, which allows us to characterise convergence towards the Bayes-optimal solution. For this we use several RNN training runs (same architecture, different random initialisation), and at fixed intervals during training we compute pairwise behavioural distances between all meta-learners and the Bayes-optimal agent. The behavioural distance is computed using the Jensen-Shannon divergence5 for prediction tasks and the absolute value of the cumulative regret for bandits. We visualise the resulting distance matrix in a 2D plot using multidimensional scaling (MDS) [40]. 3.4 Structural analysis We base our structural analysis on the idea of simulation introduced in Section 2. Since here we also deal with continuous state, input, and output spaces, we relax the notion of simulation to approximate simulation: (Reference inputs) As we cannot enumerate all the traces, we first sample a collection of input sequences from a reference distribution and then use the induced traces to compare state machines. (State and output comparison) To assess the quality of a simulation, we first learn a map φ that embeds the states of one state machine into another, and then measure the dissimilarity. To do so, we introduce two measures of dissimilarity Ds and Do to evaluate the state and output dissimilarity respectively. More precisely, consider assessing the quality of a state machine M simulating a machine N along a trace induced by the input sequence x1 x T . Then, the quality of the state embedding Ds is measured as the mean-squarederror (MSE) between the embedded states φ(SN) SM and the states SM of M along the trace. Similarly, the quality of the output simulation Do is measured as the dissimilarity between the outputs generated from the states SN and φ(SN) along the trace, that is, before and after the embedding respectively. In practice, we evaluate how well e.g. a meta-learned agent simulates a Bayes-optimal one by first finding an embedding φ mapping Bayes-optimal states into meta-learned states that minimises the state dissimilarity Ds, and then using said embedding to compute the output dissimilarity Do. The mapping φ is implemented as an MLP see details in Appendix C. We use (2) and (3) as output dissimilarity measures Do in prediction and bandit tasks respectively. Our approach is similar in spirit to [41], but adapted to work in continuous observation spaces. 5The Jensen-Shannon divergence is defined as DJS(X||Y ) = 1 2(DKL(X||M) + DKL(Y ||M)), where M is the mixture distribution (X + Y )/2. 4.1 Behavioral Comparison To compare the behavior between meta-learned and Bayes-optimal agents, we contrast their outputs for the same inputs. Consider for instance the two agents shown in Figure 1. Here we observe that the meta-learned and the Bayes-optimal agents behave in an almost identical manner: in the prediction case (Figure 1a), the predictions are virtually indistinguishable and approach the true probabilities; and in the bandit case (Figure 1b) the cumulative regrets are essentially the same6 whilst the policy converges toward pulling the best arm. 0 20 30 Timestep RNN Optimal Categorical( ) where ~Dirichlet(1, 1, 1) (a) Categorical-variable prediction task. Bottom row shows observations, middle row shows predictions of the meta-learner (solid) and the Bayes-optimal predictor (dashed), that converge towards the environment s true values (shaded bars). Top-row shows the likelihood of the observation under the predictive distribution. PRNN(at), i 0 20 30 Timestep RNN Optimal Arm 1 Arm 2 Bernoulli( 1, 2) where 1, 2 Beta(1, 1) (b) 2-armed Bernoulli-bandit task. Bottom row shows inputs (action-reward pairs). Middle row shows action probabilities for metalearner (solid) and expected arm payoffs as shaded bars. Action probabilities of the Bayes-optimal agent are not shown as it acts deterministically, selecting actions according to the highest Gittins index per time step. Top row shows cumulative regret for both agenttypes. Figure 1: Illustrative behavioral comparison of a meta-learned agent and the Bayes-optimal agent on 3 episodes (same environment random seed for both agents). Meta-learned agents were trained with only 20 time-steps; thus these results illustrate that the RNN generalizes to 30 time-steps. To quantitatively assess behavioral similarity between the meta-learners and the Bayes-optimal agents, we use the measures introduced in Section 3.2, namely (2) for prediction tasks and (3) for bandit tasks. For each task distribution, we averaged the performance of 10 meta-learned agents. The corresponding results in Figure 4a show that the trained meta-learners behave virtually indistinguishably from the Bayes-optimal agent. Results for reduced-memory agents, which cannot retain enough information to perform optimally, are shown in Appendix D.7. 4.2 Convergence We investigate how the behavior of meta-learners changes over the course of training. Following the methodology introduced in Section 3.3 we show the evolution of behavior of a single training run (top-left panels in Figure 2a, 2b). These results allow us to evaluate the meta-learners ability to pick up on the environment s prior statistics and perform Bayesian evidence integration accordingly. As training progresses, agents learn to integrate the evidence in a near-optimal manner over the entire course of the episode. However, during training the improvements are not uniform throughout the episode. This staggered meta-learning, where different parts of the task are learned progressively, resembles results reported for meta-learners on nonlinear regression tasks in [42]. 6Recall that the regret is invariant under optimal policies. 0 Trajectory Timestep 20 0.0 d( Opt, RNN) 103 104 Training Observations 106 107 0 (a) Categorical prediction task with parameters drawn from Dirichlet(1, 1, 1). 0 Trajectory Timestep 20 102 103 104 105 106 107 108 0 (b) Two-armed Bernoulli-bandit task with biases drawn from Beta(1, 1). Figure 2: Policies evolve similarly towards the Bayes-optimal policy over the course of training for both the prediction (a) and the bandit task (b). Panels in each subfigure show, (clockwise from top left): evolution of the within-episode dissimilarity from the Bayes-optimal policy, averaged over 500 trajectories; the evolution of 10 policies for different training runs (multidimensional scaling visualisation of pairwise behavioural distances; each curve is a separate run); and the training curves for the log-loss and regret respectively. We also compared behavior across multiple training runs (top-right panels in Figure 2a, 2b). Overall the results indicate that after some degree of heterogeneity early in training, all meta-learners converge in a very similar fashion to the Bayes-optimal behavior. This is an empirical confirmation of the theoretical prediction in [10] that the Bayes-optimal solution is the fixed-point of meta-learner training. Appendix D.5 shows the convergence results for all tasks. 4.3 Structural Comparison In this section we analyze the computational structure of the meta-learner who uses its internal state to store information extracted from observations required to act. Following the discussion in Section 3.4, we determine the computational similarity of the metalearning and Bayes-optimal agents via simulation. Our analysis is performed by projecting and then whitening both the RNN state (formed by concatenating both the celland hidden-states of the LSTM) and the Bayes-optimal state onto the first n principal components, where n is the dimensionality of the Bayes-optimal state/sufficient statistics. We find that these few components suffice to explain a large fraction of the variance of the RNN agent s state see Appendix D.2. We then regress an MLP-mapping φ from one (projected) agent state onto the other and compute Ds and Do. Importantly, this comparison is only meaningful if we ensure that both agents were exposed to precisely the same input history. This is easily achieved in prediction tasks by fixing the environment random seed. In bandit tasks we ensure that both agents experience the same action-reward pairs by using the trained meta-learner to generate input streams that are then also fed into the Bayes-optimal agent. Figure 3 illustrates our method for assessing the computational similarity. We embedded7 the state space of the Bayes-optimal agent into the state space of the meta-learned agent, and then we calculated the output from the embedded states. This embedding was also performed in the reverse direction. Visual inspection of this figure suggests that the meta-learned and the Bayes-optimal agents perform similar computations, as the panels resemble each other both in terms of states and outputs. We observed similar results for all other tasks (Appendix D.5). In contrast, we have observed that the computational structure of untrained meta-learners does not resemble the one of Bayes-optimal agents (Appendix D.1). The quantitative results for the structural comparison for all tasks across 10 repetitions of metatraining are shown in Figure 4. We find that for the trained meta-learner state-dissimilarity Ds is low in almost all cases. In bandit tasks, Ds tends to be slightly larger in magnitude which is somewhat expected since the RNN-state dimensionality is much larger in bandit tasks. Additionally there is often no significant difference in Ds between the untrained and the final agent we suspect this to be an artefact of a reservoir effect [43] (see Discussion). The output-dissimilarity Do is low for both 7The embeddings were implemented as MLPs having three hidden layers with either 64 (prediction) or 256 (bandits) neurons each. (a) Categorical prediction task with parameters drawn from Dirichlet(1, 1, 1). The colors indicate the prediction probabilities emitted in each state. Three episode rollouts are shown. (b) Two-armed Bernoulli-bandit task with biases drawn from Beta(1, 1). The colors indicate the actionprobabilities emitted in each state. Three episode rollouts are shown. Figure 3: Structural comparison. Each sub-figure depicts: two agent state spaces, namely of the Bayes-optimal (top-left) and RNN states (bottom-right), projected onto the first two principal components; and two simulations, i.e., the learned embeddings from the RNN into the Bayes-optimal states (top-right) and from the Bayes-optimal into the RNN states (bottom-left). The scores in the simulations indicate the MSE of the learned regression. The outputs emitted in each state are color-coded. Note that the color-codings in the simulations result from evaluating the output at the (potentially high-dimensional) embedded state (see Section 3.4). White lines indicate the same three episodes as shown in Figure 1. task types for RNN Opt, but not in the reverse direction. This indicates that the meta-learners are very well simulated by the Bayes-optimal agents, since both the state dissimilarity Ds and the output dissimilarity Do are almost negligible. In the reverse direction however, we observe that the meta-learned solutions do not always simulate the Bayes-optimal with high accuracy, as seen by the non-negigible output dissimilarity Do. We believe that this is because the sufficient statistics learned by the meta-learners are not minimal. 5 Discussion and conclusions In this study we investigated whether memory-based meta-learning leads to solutions that are behaviourally and structurally equivalent to Bayes-optimal predictors. We found that behaviorally the Bayes-optimal solution constitutes a fixed-point of meta-learner training dynamics. Accordingly, trained meta-learners behave virtually indistinguishable from Bayes-optimal agents. We also found structural equivalence between the two agent types to hold to a large extent: meta-learners are well simulated by Bayes-optimal agents, but not necessarily vice versa. This failure of simulation is most likely a failure of injectivity: if a single state in one agent must be mapped to two distinct states in another then simulation is impossible. This occurs when two trajectories lead to the same state in one agent but not another (for instance if exchangeability has not been fully learned). We suspect that RNN meta-learners represent non-minimal sufficient statistics as a result of training. For instance, for Bernoulli prediction tasks the input sequences heads-tails-heads, and tails-heads-heads induce the same minimal sufficient statistics and thus lead to precisely the same internal state in the Bayes-optimal agent, but might lead to different states in the RNN agent. From a theoretical point of view this is not unexpected, since there is no explicit incentive during RNN training that would force representations to be minimal. Note that overly strong regularization can reduce the RNN s effective capacity to a point where it can no longer represent the number of states required by Bayes-optimal solution, which of course strictly rules out computational equivalence. A related issue can be observed in bandit tasks: even untrained meta-learners show low statedissimilarity. We hypothesize that this is due to a reservoir effect [43], that is the dynamics of Beta(1, 1) Beta(0.5, 0.5) Beta(1, 5) Dirichlet(1, 1, 1) Dirichlet(0.5, 0.5, 0.5) Dirichlet(1, 1, 0.1) Normal(0, 1) Normal(1, 1) Prediction tasks 0.0 2.5 5.0 0.0 0.2 0.4 0 200 DKL(Opt||RNN) Gamma(1, 0.5) Gamma(5, 1) 0 1e+07 2e+07 DKL(RNN||RNN Opt) 0.0 0.2 0.4 MSERNN Opt 0 10 DKL(Opt||Opt RNN) 0.0 0.2 MSEOpt RNN 0.0 0.2 0.4 |r Opt r RNN| Beta(1, 1) Beta(2, 1), Beta(1, 2) Normal(0, 1) Normal(0, 0.1) Bandit tasks 0.0 0.2 0.4 |r RNN r RNN Opt| 0.0 0.1 MSERNN Opt 0.0 0.2 0.4 |r Opt r Opt RNN| 0.0 0.1 0.2 MSEOpt RNN (a) (b) (c) (d) (e) Dissimilarity measures Figure 4: Behavioral and structural comparison for all tasks. Figure shows similarity measures across K = 500 episodes of length T = 20, and 10 different training runs of the meta-learner (bars show median over trainingruns, error bars denote 5-95 quantiles). init denotes the untrained meta-learner, final denotes evaluation at the end of training. Numerical results are shown in Figure 7 in the Appendix. Column a: Behavioral dissimilarity between meta-learned agent and Bayes-optimal agent (see Section 4.1). Columns b & c: Stateand Output-dissimilarity for RNN Opt. Columns d & e: Stateand Output-dissimilarity for Opt RNN. Low values of the stateand output-dissimilarity measures (simultaneously) indicate that the state machines implemented by RNN and Opt are structurally equivalent. high-dimensional untrained RNNs are highly likely to map each input history to a unique trajectory in memory space. Accordingly, the untrained RNN memorizes inputs perfectly a verbose representation of the task s sufficient statistics. Our work contributes to understanding the structure and computations implemented by recurrent neural networks. Focusing analysis on computational equivalence, as in our work, opens up the future possibility of separating different, heterogeneous agents into meaningful sets of equivalent classes, and study universal aspects of these agent-classes. 5.1 Related work In this paper we study memory-based meta-learning through a Bayesian lens, showing that metalearning objectives naturally induce Bayes-optimal behaviour at convergence. A number of previous works have attempted to devise new recurrent architectures to perform Bayes filtering in a number of settings, including time series prediction [44], state space modelling [45], and Kalman filtering [46]. Other previous work has attempted to improve memory-based meta-learners abilities by augmenting them with a memory, and using weights which adapt at different speeds [47, 48]. Another approach to meta-learning is optimiser-based meta-learning such as MAML [49]. In optimiser-based meta-learning models are trained to be able to adapt rapidly to new tasks via gradient descent. MAML has been studied from a Bayesian perspective, and shown to be a hierarchical Bayesian model [50]. Recent work suggests that solutions obtained by optimiser-based meta-learning might be more similar to those from memory-based meta-learning than previously thought [51] . In this paper we relate memory-based meta-learning to finite-state automata which track sufficient statistics of their inputs. The field of computational mechanics [52] studies predictive automata (known as ε-machines) which track the state of a stochastic process in order to predict its future states. The states of ε-machines are referred to as causal states, and have recently been used to augment recurrent agents in POMDPs [53]. Finite-state automata have also been considered as a model for decision-making agents in the situated automata work of Rosenschein and Kaelbling [54, 55]. The states of situated automata track logical propositions about the state of the world instead of having a probabilistic interpretation, but are naturally suited to goal-directed agents. There is considerable work on understanding recurrent neural networks on natural language tasks [56], and in neuroscience [57 59], e.g. how relations between multiple trained models can illuminate computational mechanisms [15], and the dynamics involved in contextual processing [13]. Computational analysis of internal dynamics of reinforcement learning agents has received less attention in the literature, though there are some notable examples: a multi-agent setting [18] and Atari games [60]. Using a related formalism to our approach, the authors of [61] extract minimal finite-state machines (FSM) from the internal dynamics of Atari-agents. However their focus is on extracting small human-interpretable FSM, whereas we compare the computational structure of two agents in a fully automated, quantitative fashion. In recent years a diverse range of tools to allow interpretability and explainability of deep networks have been developed, including saliency maps [62 67], manual dissection of individual units [17, 16, 68] and training explainable surrogate models to mimic the output of deep networks [69, 61]. Although our focus here is different - we seek to establish how a broad class of architectures behaves on a family of tasks, rather than explaining a specific network - the closest parallel is with the use of surrogate explainable models. In this case, the Bayes-optimal agent serves as an understood model, and we relate its (well-understood) behaviour to that of the meta-trained agent. Scope and limitations We performed our empirical comparison on a range of tasks where optimal solutions are analytically and computationally tractable. The latter is typically no longer true in more complex tasks and domains. However, the simulation methodology used in this paper could be useful to compare agent-types against each other in more general settings, as it does not rely on either agent being Bayes-optimal. While most aspects of our methodology scale up well to more complex agents, the main difficulty is generating reference trajectories that cover a large (enough) fraction of possible experiences. Finally, our results show that when optimal policies are in the search space, and training converges to those policies, then the resulting policy will be Bayes-optimal. In more complex cases, one or both of these assumptions may no longer hold. Further study is needed to understand the kind of suboptimal solutions that are generated by meta-learning in this case. Conclusions Our main contribution is to advance the understanding of RNN-based meta-learned solutions. We empirically confirm a recently published theoretical claim [10] that fully-converged meta-learners and Bayes-optimal agents are computationally equivalent. In particular, we showed that RNN meta-learners converge during training to the Bayes-optimal solution, such that trained metalearners behave virtually indistinguishably from Bayes-optimal agents. Using a methodology related to the concept of simulation in theoretical computer science, we additionally show (approximate) structural equivalence of the state-machines implemented by the RNN meta-learners and the Bayesoptimal agent. Our results suggest that memory-based meta-learning will drive learned policies towards Bayes-optimal behaviour, and will converge to this behaviour where possible. 6 Broader Impact Our work helps advance and verify the current understanding of the nature of solutions that metalearning brings about (our empirical work focused on modern recurrent neural network architectures and training algorithms, but we expect the findings to qualitatively hold for a large range of AI systems that are trained through meta-learning). Understanding how advanced AI and ML systems work is of paramount importance for safe deployment and reliable operation of such systems. This has also been recognized by the wider machine-learning community with a rapidly growing body of literature in this emerging field of Analysis and Understanding of deep learning. While increased understanding is likely to ultimately also contribute towards building more capable AI systems, thus potentially amplifying their negative aspects, we strongly believe that the merits of understanding how these systems work clearly outweigh the potential risks in this case. We argue that understanding meta-learning on a fundamental level is important, since meta-learning subsumes many specific learning tasks and is thought to play an important role for AI systems that generalize well to novel situations. Accordingly we expect meta-learning to be highly relevant over the next decade(s) in AI research and in the development of powerful AI algorithms and applications. In this work we also show a proof-of-concept implementation for analysis methods that might potentially allow one to separate (heterogeneous) agents into certain equivalence classes, which would allow to safely generalize findings about an individual agent to the whole equivalence class. We believe that this might open up interesting future opportunities to boost the generality of analysis methods and automatic diagnostic tools for monitoring of AI systems. Acknowledgments and Disclosure of Funding We thank Jane Wang and Matt Botvinick for providing helpful comments on this work. Funding in direct support of this work: no specific or third-party funding was received. All authors are full-time employed with Deep Mind. [1] Y Bengio, S Bengio, and J Cloutier. Learning a synaptic learning rule. In IJCNN-91-Seattle International Joint Conference on Neural Networks, volume 2, pages 969 vol. IEEE, 1991. [2] Juergen Schmidhuber, Jieyu Zhao, and MA Wiering. Simple principles of metalearning. Technical report IDSIA, 69:1 23, 1996. [3] Sebastian Thrun and Lorien Pratt. Learning to learn: Introduction and overview. In Learning to learn, pages 3 17. Springer, 1998. [4] Sepp Hochreiter, A Steven Younger, and Peter R Conwell. Learning to learn using gradient descent. In International Conference on Artificial Neural Networks, pages 87 94. Springer, 2001. [5] Adam Santoro, Sergey Bartunov, Matthew Botvinick, Daan Wierstra, and Timothy Lillicrap. Meta-learning with memory-augmented neural networks. In International Conference on Machine Learning, pages 1842 1850, 2016. [6] Jane X Wang, Zeb Kurth-Nelson, Dhruva Tirumala, Hubert Soyer, Joel Z Leibo, Remi Munos, Charles Blundell, Dharshan Kumaran, and Matt Botvinick. Learning to reinforcement learn. ar Xiv preprint ar Xiv:1611.05763, 2016. [7] Eliza Strickland. Yoshua Bengio, revered architect of AI, has some ideas about what to build next. IEEE Spectrum, December 2019. [8] Daniel A Braun, Ad Aertsen, Daniel M Wolpert, and Carsten Mehring. Motor task variation induces structural learning. Current Biology, 19(4):352 357, 2009. [9] Daniel A Braun, Carsten Mehring, and Daniel M Wolpert. Structure learning in action. Behavioural Brain Research, 206(2):157 165, 2010. [10] Pedro A Ortega, Jane X Wang, Mark Rowland, Tim Genewein, Zeb Kurth-Nelson, Razvan Pascanu, Nicolas Heess, Joel Veness, Alex Pritzel, Pablo Sprechmann, et al. Meta-learning of sequential strategies. ar Xiv preprint ar Xiv:1905.03030, 2019. [11] Michael O Gordon Duff and Andrew Barto. Optimal Learning: Computational procedures for Bayes-adaptive Markov decision processes. Ph D thesis, University of Massachusetts at Amherst, 2002. [12] Niru Maheswaranathan, Alex Williams, Matthew Golub, Surya Ganguli, and David Sussillo. Reverse engineering recurrent networks for sentiment classification reveals line attractor dynamics. In Advances in Neural Information Processing Systems, pages 15670 15679, 2019. [13] Niru Maheswaranathan and David Sussillo. How recurrent networks implement contextual processing in sentiment analysis. ar Xiv preprint ar Xiv:2004.08013, 2020. [14] Hidenori Tanaka, Aran Nayebi, Niru Maheswaranathan, Lane Mc Intosh, Stephen Baccus, and Surya Ganguli. From deep learning to mechanistic understanding in neuroscience: the structure of retinal prediction. In Advances in Neural Information Processing Systems, pages 8535 8545, 2019. [15] Anthony Bau, Yonatan Belinkov, Hassan Sajjad, Nadir Durrani, Fahim Dalvi, and James Glass. Identifying and controlling important neurons in neural machine translation. In International Conference on Learning Representations, 2019. [16] Chris Olah, Arvind Satyanarayan, Ian Johnson, Shan Carter, Ludwig Schubert, Katherine Ye, and Alexander Mordvintsev. The building blocks of interpretability. Distill, 2018. [17] Chris Olah, Nick Cammarata, Ludwig Schubert, Gabriel Goh, Michael Petrov, and Shan Carter. An overview of early vision in Inception V1. Distill, 2020. [18] Max Jaderberg, Wojciech M Czarnecki, Iain Dunning, Luke Marris, Guy Lever, Antonio Garcia Castaneda, Charles Beattie, Neil C Rabinowitz, Ari S Morcos, Avraham Ruderman, et al. Human-level performance in 3D multiplayer games with population-based reinforcement learning. Science, 364(6443):859 865, 2019. [19] David Marr. Vision: A computational investigation into the human representation and processing of visual information. MIT Press, 2010. [20] John F Kolen and Stefan C Kremer. A field guide to dynamical recurrent networks. John Wiley & Sons, 2001. [21] Felix A Gers, Jürgen Schmidhuber, and Fred Cummins. Learning to forget: Continual prediction with LSTM. 1999. [22] Kyunghyun Cho, Bart Van Merriënboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio. Learning phrase representations using RNN encoderdecoder for statistical machine translation. ar Xiv preprint ar Xiv:1406.1078, 2014. [23] AJ Robinson and Frank Fallside. The utility driven dynamic error propagation network. University of Cambridge Department of Engineering Cambridge, 1987. [24] Paul J Werbos. Generalization of backpropagation with application to a recurrent gas market model. Neural networks, 1(4):339 356, 1988. [25] George H Mealy. A method for synthesizing sequential circuits. The Bell System Technical Journal, 34(5):1045 1079, 1955. [26] Michael Sipser. Introduction to the theory of computation. ACM Sigact News, 27(1):27 29, 1996. [27] JE Savage. Models of computation. exploring the power of computing. Reading, MA, 1998. [28] Edmund M Clarke Jr, Orna Grumberg, Daniel Kroening, Doron Peled, and Helmut Veith. Model checking. MIT press, 2018. [29] Christel Baier and Joost-Pieter Katoen. Principles of model checking. 2008. [30] Jos C.M. Baeten and Davide Sangiorgi. Concurrency theory: A historical perspective on coinduction and process calculi. In Jörg H. Siekmann, editor, Computational Logic, volume 9 of Handbook of the History of Logic, pages 399 442. North-Holland, 2014. [31] Howard Raiffa and Robert Schlaifer. Applied statistical decision theory. 1961. [32] Christopher M Bishop. Pattern recognition and machine learning. springer, 2006. [33] Andrew Gelman, John B Carlin, Hal S Stern, David B Dunson, Aki Vehtari, and Donald B Rubin. Bayesian data analysis. CRC press, 2013. [34] Tor Lattimore and Csaba Szepesvári. Bandit algorithms. preprint, page 28, 2018. [35] John C Gittins. Bandit processes and dynamic allocation indices. Journal of the Royal Statistical Society: Series B (Methodological), 41(2):148 164, 1979. [36] Harrison Edwards and Amos Storkey. Towards a neural statistician. ar Xiv preprint ar Xiv:1606.02185, 2016. [37] Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT Press, 2018. [38] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. [39] Lasse Espeholt, Hubert Soyer, Remi Munos, Karen Simonyan, Volodymir Mnih, Tom Ward, Yotam Doron, Vlad Firoiu, Tim Harley, Iain Dunning, et al. Impala: Scalable distributed deep RL with importance weighted actor-learner architectures. ar Xiv preprint ar Xiv:1802.01561, 2018. [40] Ingwer Borg and Patrick JF Groenen. Modern multidimensional scaling: Theory and applications. Springer Science & Business Media, 2005. [41] Antoine Girard and George J Pappas. Approximate bisimulations for nonlinear dynamical systems. In Proceedings of the 44th IEEE Conference on Decision and Control, pages 684 689. IEEE, 2005. [42] Neil C Rabinowitz. Meta-learners learning dynamics are unlike learners . ar Xiv preprint ar Xiv:1905.01320, 2019. [43] Wolfgang Maass and Henry Markram. On the computational power of circuits of spiking neurons. Journal of Computer and System Sciences, 69(4):593 616, 2004. [44] Bryan Lim, Stefan Zohren, and Stephen Roberts. Recurrent neural filters: Learning independent bayesian filtering steps for time series prediction. ar Xiv preprint ar Xiv:1901.08096, 2019. [45] Rahul G Krishnan, Uri Shalit, and David Sontag. Structured inference networks for nonlinear state space models. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, pages 2101 2109, 2017. [46] Huseyin Coskun, Felix Achilles, Robert Di Pietro, Nassir Navab, and Federico Tombari. Long short-term memory kalman filters: Recurrent neural estimators for pose regularization. In Proceedings of the IEEE International Conference on Computer Vision, pages 5524 5532, 2017. [47] Tsendsuren Munkhdalai and Hong Yu. Meta networks. Proceedings of Machine Learning Research, 70:2554, 2017. [48] Tsendsuren Munkhdalai, Alessandro Sordoni, Tong Wang, and Adam Trischler. Metalearned neural memory. In Advances in Neural Information Processing Systems, pages 13331 13342, 2019. [49] Chelsea Finn, Pieter Abbeel, and Sergey Levine. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning-Volume 70, pages 1126 1135. JMLR. org, 2017. [50] Erin Grant, Chelsea Finn, Sergey Levine, Trevor Darrell, and Thomas Griffiths. Recasting gradient-based meta-learning as hierarchical bayes. In International Conference on Learning Representations, 2018. [51] Aniruddh Raghu, Maithra Raghu, Samy Bengio, and Oriol Vinyals. Rapid learning or feature reuse? Towards understanding the effectiveness of MAML. In International Conference on Learning Representations, 2020. [52] Cosma Rohilla Shalizi and James P Crutchfield. Computational mechanics: Pattern and prediction, structure and simplicity. Journal of Statistical Physics, 104(3-4):817 879, 2001. [53] Amy Zhang, Zachary C Lipton, Luis Pineda, Kamyar Azizzadenesheli, Anima Anandkumar, Laurent Itti, Joelle Pineau, and Tommaso Furlanello. Learning causal state representations of partially observable environments. ar Xiv preprint ar Xiv:1906.10437, 2019. [54] Leslie Pack Kaelbling and Stanley J Rosenschein. Action and planning in embedded agents. Robotics and Autonomous Systems, 6(1-2):35 48, 1990. [55] Stanley J Rosenschein and Leslie Pack Kaelbling. A situated view of representation and control. Artificial Intelligence, 73(1-2):149 173, 1995. [56] Yonatan Belinkov and James Glass. Analysis methods in neural language processing: A survey. Transactions of the Association for Computational Linguistics, 7:49 72, 2019. [57] Hansem Sohn, Devika Narain, Nicolas Meirhaeghe, and Mehrdad Jazayeri. Bayesian computation through cortical latent dynamics. Neuron, 103(5):934 947, 2019. [58] Niru Maheswaranathan, Alex Williams, Matthew Golub, Surya Ganguli, and David Sussillo. Universality and individuality in neural dynamics across large populations of recurrent networks. In Advances in Neural Information Processing Systems, pages 15603 15615, 2019. [59] Ishita Dasgupta, Eric Schulz, Joshua B. Tenenbaum, and Samuel J. Gershman. A theory of learning to infer. bio Rxiv, 2019. [60] Tom Zahavy, Nir Ben-Zrihem, and Shie Mannor. Graying the black box: Understanding DQNs. In International Conference on Machine Learning, pages 1899 1908, 2016. [61] Anurag Koul, Sam Greydanus, and Alan Fern. Learning finite state representations of recurrent policy networks. ar Xiv preprint ar Xiv:1811.12530, 2018. [62] Dumitru Erhan, Yoshua Bengio, Aaron Courville, and Pascal Vincent. Visualizing higher-layer features of a deep network. 2009. [63] Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Deep inside convolutional networks: Visualising image classification models and saliency maps. ar Xiv preprint ar Xiv:1312.6034, 2013. [64] Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In European Conference on Computer Vision, pages 818 833. Springer, 2014. [65] Avanti Shrikumar, Peyton Greenside, Anna Shcherbina, and Anshul Kundaje. Not just a black box: Learning important features through propagating activation differences. ar Xiv preprint ar Xiv:1605.01713, 2016. [66] Ramprasaath R Selvaraju, Abhishek Das, Ramakrishna Vedantam, Michael Cogswell, Devi Parikh, and Dhruv Batra. Grad-cam: Why did you say that? ar Xiv preprint ar Xiv:1611.07450, 2016. [67] Daniel Smilkov, Nikhil Thorat, Been Kim, Fernanda Viégas, and Martin Wattenberg. Smoothgrad: removing noise by adding noise. ar Xiv preprint ar Xiv:1706.03825, 2017. [68] David Bau, Jun-Yan Zhu, Hendrik Strobelt, Agata Lapedriza, Bolei Zhou, and Antonio Torralba. Understanding the role of individual units in a deep neural network. Proceedings of the National Academy of Sciences, 2020. [69] Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. " why should i trust you?" explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 1135 1144, 2016.