# solving_continuous_control_with_episodic_memory__bdc3a229.pdf Solving Continuous Control with Episodic Memory Igor Kuznetsov , Andrey Filchenkov ITMO University igorkuznetsov14@gmail.com, afilchenkov@itmo.ru, Episodic memory lets reinforcement learning algorithms remember and exploit promising experience from the past to improve agent performance. Previous works on memory mechanisms show benefits of using episodic-based data structures for discrete action problems in terms of sample-efficiency. The application of episodic memory for continuous control with a large action space is not trivial. Our study aims to answer the question: can episodic memory be used to improve agent s performance in continuous control? Our proposed algorithm combines episodic memory with Actor-Critic architecture by modifying critic s objective. We further improve performance by introducing episodic-based replay buffer prioritization. We evaluate our algorithm on Open AI gym domains and show greater sample-efficiency compared with the state-of-the art model-free off-policy algorithms. 1 Introduction The general idea of episodic memory in reinforcement learning setting is to leverage long-term memory reflecting data structure containing information of the past episodes to improve agent performance. The existing works [Blundell et al., 2016; Pritzel et al., 2017], show that episodic control (EC) can be beneficial for decision making process. In the context of discrete action problems, episodic memory stores information about states and the corresponding returns in a tablelike data structure. With only several actions the proposed methods store past experiences in multiple memory buffers per each action. During the action selection process the estimate of taking each action is reconsidered with respect to the stored memories and may be corrected taking into the account past experience. The motivation of using episodic-like structures is to latch quickly to the rare but promising experiences that slow gradient-based models cannot reflect from a small number of samples. The notion of using episodic memory in continuous control is not trivial. Since the action space may be highdimensional, the methods that operate discrete action space become not suitable. Another challenge is the complexity of the state space. The study of [Blundell et al., 2016] shows that some discrete action environments (e.g. Atari) may have high ratio of repeating states, which is not the case for complex continuous control environments. Our work proposes an algorithm that leverages episodic memory experience through modification of the Actor-Critic objective. Our algorithm is based on DDPG, the model-free actor-critic algorithm that operates over continuous control space. We offer to store the representation of action-state pairs in memory module to perform memory association not only from environment state, but also from the performed action. Our experiments show that modification of objective provides greater sample efficiency compared with offpolicy algorithms. We further improve agent performance by introducing novel way of prioritized experience replay based on episodic experiences. The proposed algorithm, Episodic Memory Actor-Critic (EMAC), exploits episodic memory during the training, distilling the Monte-Carlo (MC) discounted return signal through the critic to actor and resulting in the strong policy that provides greater sample efficiency than other models. We draw the connection between the proposed objective and alleviation of Q-value overestimation, a common problem in Actor-Critic methods [Thrun and Schwartz, 1993]. We show that the use of the retrieved episodic data results in more realistic critic estimates that in turn provides faster training convergence. In contrast, the proposed prioritized experience replay, aims to focus more on optimistic promising experiences from the past. The process of frequently exploiting only the high-reward transitions from the whole replay buffer may result in unstable behaviour. However, the stabilizing effect of the new objective allows to decrease the priority of non-informative transitions with low return without loss in stability. We evaluate our model on a set of Open AI gym environments [Dhariwal et al., 2017] (Figure 1) and show that it achieves greater sample efficiency compared with the state-of-the-art off-policy algorithms (TD3, SAC). We open sourced our algorithm to achieve reproducibility. All the codes and learning curves can be accessed at: http://github. com/schatty/EMAC. 2 Related Work Episodic Memory. First introduced in [Lengyel and Dayan, 2007], episodic control is studied within the regular Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) tree-structured MDP setup. Episodic control aims to memorize highly-rewarding experiences and replays sequences of actions that achieved high returns in the past. The successful application of episodic memory to the discrete action problems has been studied in [Blundell et al., 2016] . Reflecting the model of hippocampal episodic control the work shows that the use of table-based structures containing state representation with the corresponding returns can improve modelfree algorithms on environments with the high ratio of repeating states. Semi-tabular differential version of memory is proposed in [Pritzel et al., 2017] to store past experiences. The work of [Lin et al., 2018] is reminiscent of ideas presented in our study. They modify the DQN objective to mimic relationships between two learning systems of the human brain. The experiments show that this modifications improves sample efficiency for the arcade learning environment [Bellemare et al., 2013]. The framework to associate the related experience trajectories in the memory to achieve reasoning of effective strategies is proposed in [Vikbladh et al., 2017]. In the context of the continuous control, the work of [Zhang et al., 2019] exploits episodic memory to redesign experience replay for the asynchronous DDPG. To the best of our knowledge, this is the only work that leverages episodic memory within the continuous control. Actor-Critic algorithms. Actor-Critic methods represents a set of algorithms that compute value function for the policy (critic) and improve the policy (actor) from this value function. Using deep approximators for the value function and the actor, [Lillicrap et al., 2015] presents Deep Deterministic Policy Gradient. The proposed model-free off-policy algorithm aims to learn policy in high-dimensional, continuous action space. Recent works propose several modifications that stabilize and improve performance of the original DDPG. The work of [Fujimoto et al., 2018] addresses Q-value overestimation and proposes the Twin Delayed Deep Deterministic (TD3) algorithm which greatly improves DDPG learning speed. The proposed modifications include existence of multiple critic to reduce critic s over optimism, additional noise applied within calculation of target Q-estimate and delayed policy update. In [Haarnoja et al., 2018a; Haarnoja et al., 2018b] the authors study maximum-entropy objectives based on which provide state-of-the-art performance for Open AI gym benchmark. 3 Background Reinforcement learning setup consists of an agent that interacts with an environment E. At each discrete time step t = 0...T the agent receives an environment state st, performs an action at and receives a reward rt. An agent s action is defined by a policy, which maps the state to a probability distribution over the possible actions π : S P(A). The return is defined as a discounted sum of rewards Rt = PT i=t γi tr(si, ai) with a discount factor γ [0, 1]. Reinforcement learning aims to find the optimal policy πθ, with parameters θ, which maximizes the expected return from the initial distribution J(θ) = Esi pπ,ai π[R0]. The Q-function denotes the expected return when performing action a from the state s following the current policy π Qπ(s, a) = Esi pπ,ai π [Rt|s, a] For continuous control problems policy πθ can be updated taking the gradient of the expected return θJ(θ) with deterministic policy gradient algorithm [Silver et al., 2014]: θJ(θ) = Es pπ a Qπ(s, a)|a=π(s) θπθ(s) . (1) In actor-critic methods, we operate with two parametrized functions. An actor represents the policy π and a critic is the Q-function. The critic is updated with temporal difference learning by iteratively minimizing the Bellman equation [Watkins and Dayan, 1992]: JQ = E (Q(st, at) (r(st, at) + γQ(st+1, at+1)))2 . (2) In deep Q-learning , the parameters of Q-function are updated with additional frozen target network Qθ which is updated by τ proportion to match the current Q-function θ τθ + (1 τ)θ JQ = E (Q(st, at) Q )2 , (3) Q = (r(st, at) + γQθ (st+1, a )), a πθ (st+1). (4) The actor is learned to maximize the current Q function. Jπ = E [Q(s, π(s))] . (5) The advantage of the actor-critic methods is that they can be applied off-policy, methods that proven to have better sample complexity [Lillicrap et al., 2015]. During the training the actor and critic are updated with sampled mini-batches from the experience replay buffer [Lin, 1992]. We present Episodic Memory Actor-Critic, the algorithm that leverages episodic memory during the training. Our algorithm builds on the Deep Deterministic Policy Gradient [Lillicrap et al., 2015] by modifying critic s objective and introducing episodic-based prioritized experience replay. Figure 1: Example of Open AI gym environments. (a) Walker2d-v3 (b) Swimmer-v3 (c) Hopper-v3 (d) Inverte Double Pendulum-v2 Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) 4.1 Memory Module The core of our algorithm features a table-based structure that stores the experience of the past episodes. Following the approach of [Blundell et al., 2016] we treat this memory table as a set of key-value pairs. The key is the representation of a concatenated state-action pair and the value is the true discounted Monte-Carlo return from an environment. We encode this state-action pair to a vector of smaller dimension by the projection operation φ : x Mx, x Rv, M Ru v, where v is the dimension of the concatenated state-action vector, and u is a smaller projected dimension. As the Johnson Lindenstrauss lemma [Johnson and Lindenstrauss, 1984] states, this transformation preserves relative distances in the original space given that M is a standard Gaussian. The projection matrix is initialized once at the beginning of the training. Memory module implies two operations: add and lookup. As we can calculate the discounted return only at the end of an episode we perform add operation adding pairs of φ([s, a])i, Ri when the whole episode is generated by the agent. We note that the true discounted return may be considered fair only when the episode ends naturally. For time step limit case we should perform additional evaluation steps without using these transitions for training. This way we allow discounted return from later transitions be obtained from the same amount of consequent steps as transitions in the beginning of the episode. Given the complexity and continuous nature of environments we make an assumption that the received state-action pair has no duplicates in the stored representations. As a result, we do not perform the search for the stored duplicates to make a replacement and we add the incoming key-value pair to the end of the module. Thus, the add operation takes O(1). The lookup operation takes as input state-action pair, performs the projection to the smaller dimension and accomplishes the search for the k most similar keys returning the corresponding MC returns. We use l2 distance as a metric of similarity between the projected states d(z, zi) = z zi 2 2 + ϵ, (6) where ϵ is a small constant. The episodic MC return for a projected query vector zq is formulated as a weighted sum of the closest K elements k=1..K qkwk, (7) where q is a value stored in the memory-module and wi is a weight proportional to the inverse distance to the query vector. wk = e d(k,q) P t=1..K e d(k,t) . (8) We propose to leverage stored MC returns as the second pessimistic estimate of the Q-function. In contrast of approach of [Pritzel et al., 2017; Lin et al., 2018] we use Figure 2: EMAC architecture on calculating Q-estimates weighted sum of all near returns rather then taking maximum of them. Given that we are able calculate MC return for each transition from the off-policy replay buffer we propose the following modification of the original critic objective (3): JQ =(1 α)(Q(st, at) Q )2+ +α(Q(stat) QM)2, (9) where QM is the value returned by lookup operation and α is a hyperparameter controlling the contribution of episodic memory. In the case of α = 0 the objective becomes a common Q-learning loss function. In our experiments we found beneficial small values of α. In the evaluation results α is set to 0.1 for the Walker, Hopper, Inverted Pendulum, Inverted Double Pendulum, and to 0.15 for the Swimmer. The memory module is similar to the replay buffer. They both needed to sample off-policy data during the Q-learning update. The difference is the memory module stores smalldimensional representations of the state-action pair for effective lookup and discounted returns instead of rewards. The outline of the proposed architecture on calculating Qestimates is presented in the Figure 2. 4.2 Alleviating Q-value Overestimation The issue of Q-value overestimation is a common case in Qlearning based methods [Thrun and Schwartz, 1993; Lillicrap et al., 2015]. Considering the discrete action setting, the value estimate is updated in a greedy fashion from suboptimal Q function containing some error y = r + γ maxa Q(s , a ). As a result, the maximum over the actions along with its error will generally be greater than the true maximum [Thrun and Schwartz, 1993]. This bias is then propagated through the Bellman residual resulting in a consistent overestimation of the Q-function. The work of [Fujimoto et al., 2018] studies the presence of the same overestimation in an actor-critic setting. The authors propose a clipped variant of Double Qlearning that reduces overestimation. In our work we show that additional episodic Q-estimate in critic loss can be used as a tool to reduce overestimation bias. Following the procedure described in [Fujimoto et al., 2018] we perform the experiment that measures the predicted Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) Figure 3: Q-value overestimation. (a) Comparison between predicted Q-value and its true estimate. (b) Comparison between episodic stored MC estimate and true estimate Q-value from a critic compared with a true Q-estimate. We compare the value estimate of the critic, true value estimate and episodic MC return stored in the memory module. As a true value estimate we use a discounted return of the current policy starting from the state sampled randomly from the replay buffer. The discounted return is calculated from true rewards for maximum of 1000 steps or less in the case of the episode s end. We measure true value estimate each 5000 steps during the training of 100000 steps. We use the average of the batch for the critic s Q-value estimate and episodic return. The learning behaviour for the Walker2dv3 and Hopper-v3 domains is shown in Figure 3. In (a) we show that the problem of Q-value overestimation exists for both TD3 and EMAC algorithms, as both of Q-value predictions are higher than their true estimates. The Q-value prediction for EMAC shows less tendency for overestimaton. In (b) we compare true value estimate with episodic MC-return and show that latter has more realistic behaviour than the critic s estimate. Here, the difference between the true estimate and the MC-return is that episodic returns are obtained from the past suboptimal policy, whereas true value estimate is calculated using the current policy. The training curves from (b) show that MC-return has less value than the true estimate. The experiment shows that episodic Monte Carlo returns from the past are always more pessimistic than the corresponding critic value. This fact makes incorporating MCreturn into the objective beneficial for the training. As a result, the proposed objective with episodic MC-return shows less tendency to overestimation compared to the TD3 model. Given the state-action pair we here state that the MC return produced by the suboptimal policy for the same state may be used as a second stabilized estimate of the critic. Therefore, Algorithm 1 EMAC 1: Initialize actor network πφ, and critic network Qθ 2: Initialize target critic network Qθ 3: Initialize replay buffer B, and episodic buffer BE 4: Initialize memory module M 5: for t=1 to T do 6: Select action a πφ(s) 7: Receive next state s , reward r, done d 8: Store (s, a, r, s , d) in BE 9: if d=True then 10: for i=1 to |BE| do 11: Calculate discounted episode return Ri. 12: Store (si, ai, ri, s i, di) in B 13: Add( [si, ai], Ri ) in M 14: end for 15: Free BE 16: end if 17: Sample mini-batch from B: [sb, ab, rb, sb, rb] 18: QM Lookup(sb, ab) from M 19: Update critic Qθ by minimizing the loss: 20: JQθ = (1 α)(Q Q )2 + α(Q QM)2 21: Update policy πφ 22: Update target critic θ τθ + (1 τ)θ 23: end for 24: return φ additional loss component of MSE between episodic return and critic estimate may be interpreted as a penalty for critic overestimation. 4.3 Episodic-based Experience Replay Prioritization The prioritization of the sampled experiences for off-policy update of Q-function approximation is a studied tool for improving sample-efficiency and stability of an agent [Schaul et al., 2015; Wang et al., 2016; Kapturowski et al., 2018]. The common approach of using prioritization is formulated in the work of [Schaul et al., 2015], where normalized TD-error is used as a criterion of transition importance during the sampling. We propose a slightly different prioritization scheme that is based on episodic return as a criterion for sampling preference. We formulate the probability of sampling a transition as P(i) = pβ i P k pβ k , (10) where priority of transition pi is a MC return stored in the memory module. The exponent β controls the measure of prioritization with β = 0 as a uniform non-prioritized case. The interpretation of such a prioritization is an increasing reuse of transitions which gave high returns in the past. The shift of true reward distribution used for off-policy update may result in divergent behaviour, therefore large values of β may destabilize the training. In our experiment the coefficient value of 0.5 gave promising results improving non-prioritized version of the algorithm. Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) Figure 4: Evaluation results on Open AI Gym Benchmark. Environment EMAC EMAC-No Pr DDPG TD3 SAC Walker2d-v3 2236.88 808 2073 768.88 129.73 1008.32. 1787.28 Hopper-v3 1969.16 1142.58 1837.3 794.17 1043.78 989.96 2411.02 Swimmer-v3 80.53 33.07 62.56 25.78 27.45 38.91 56.7 Inverted Pendulum-v2 1000 1000 909.98. 1000 1000 Inverted Double Pendulum-v2 9332.56 11.78 9327.8 11.2 9205.81 9309.73 9357.17 Table 1: Average return over 10 trials of 200000 time steps. corresponds to a standard deviation over 10 trials. The stored probabilities of sampling are recalculated after each episode comes to the memory-module and used consequently during sampling from off-policy replay buffer. 5 Experiments The EMAC algorithm is shown in Algorithm 1. We compare our algorithm with model-free off-policy algorithms DDPG [Lillicrap et al., 2015], TD3 [Fujimoto et al., 2018] and SAC [Haarnoja et al., 2018a]. For DDPG and TD3 we use implementations of [Fujimoto et al., 2018]. For SAC we implemented the model following the [Haarnoja et al., 2018a]. All the networks have the same architecture in terms of the number of hidden layers, non-linearities and size of the hidden dimensions. In our experiments we focus on small-data regime given all algorithms 200000 time steps from an environment. We evaluate our algorithm on a set of Open AI gym domains [Dhariwal et al., 2017]. Each environment is run for 200000 time steps with the corresponding number of networks update steps. Evaluation is performed every 1000 steps with the reported value as an average from 10 evaluation episodes from different seeds without any exploration. We report results from both prioritized (EMAC) and nonprioritized (EMAC-No Pr) versions of the algorithm. The results are reported from 10 random seeds. The results in Table 1 is the average return of the last training episode over all seeds. The training curves of the compared algorithms presented in Figure 4. Our algorithm outperforms DDPG and TD3 on all tested environments and SAC on three out of five environments. Networks parameters are updated with Adam optimizer [Kingma and Ba, 2014] with a learning rate of 0.001. All models consists of two hidden layers, size 256, for an actor and a critic and a rectified linear unit (Re LU) as a nonlinearity. For the first 1000 time steps we do not exploit an actor for action selection and choose the actions randomly for the exploration purpose. During the architecture design we study the dependency between the size of the projected state-action pairs and final metrics. We witness small increase in efficiency with greater projected dimensions. Training curves showing different variants of projections of 4, 16, 32 sizes on domains of Walker-v3 and Hopper-v3 are depicted in Figure 5. Unfortunately, increased projection size tends to slow down the lookup operation. In all our experiments we use the minimal option of projection size of 4 for the faster computations. The work of [Pritzel et al., 2017] leverages the KD-tree data structure to store and search experiences. On the contrary, we decided to keep table-based structure of the stored expe- Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) Figure 5: Performance of the EMAC with different size of projected state-action pairs Figure 6: Comparison of prioritized (EMAC) and non-prioritized (EMAC-No Pr) versions riences, but placed the module on a CUDA device to perform vectorized l2-distance calculations. All our experiments are performed on single 1080ti NVIDIA card. Due to low-data regime we are able to store all incoming transitions without replacement for the replay buffer as well as for episodic memory module. We set episodic memory size to 200000 transitions, although our experiments do not indicate loss in performance for cases of smaller sizes of 100 and 50 thousands of records. The k parameter that determines the number of top-k nearest neighbours for the weighted episodic return calculation is set to 1 or 2 dependent on the best achieved results from both options. We notice that k bigger than 3 leads to the worse performance. To evaluate episodic-based prioritization we compare the prioritized and non-prioritized versions of EMAC on multiple environments. Prioritization parameter β is chosen to be 0.5. The learning behaviour of the prioritized and non-prioritized versions of the algorithm is showed in Figure 6. The average return for both versions is provided in Table 1. 6 Discussion We present Episodic Memory Actor-Critic (EMAC), a deep reinforcement learning algorithm that exploits episodic memory in continuous control problems. EMAC uses nonparametric data structure to store and retrieve experiences from the past. The episodic memory module is involved in the training process via the additional term in the critic s loss. The motivation behind such an approach is that the actor is directly dependent on the critic, therefore improving critic s quality ensures the stronger and more efficient policy. We do not exploit episodic memory during the policy evaluation, which means that memory module is used only within the network update step. The loss modification demonstrates sample-efficiency improvement over the DDPG baseline. We further show that introducing prioritization based on the episodic memories improves our results. Experimental study of Q-value overestimation shows that proposed approach has less tendency in critic overestimation thus providing faster and more stable training. Our experiments show that leveraging episodic memory gave superior results in comparison to the baseline algorithm DDPG and TD3 on all tested environments and also outperformed SAC on 3 out of 5 environments. We hypothesize that the applicability of the proposed method is dependent from environment complexity. As a result, we struggle to outperform SAC for such a complicated environment as Humanoid which has bigger action space than the other Open AI gym domains. We now outline the directions of future work. As noted in [Daw et al., 2005], both animal and human brain exploits multiple memory systems while solving various tasks. Episodic memory is often associated with hippocampal involvement [Blundell et al., 2016; Lin et al., 2018] as a longterm memory. Although the gradient-update-based policy may be seen as a working memory, it may be beneficial to study the role of separate short-term memory mechanisms alongside episodic memory for better decision making. Another interesting direction is to use different state-action representation for storing experiences. Although the random projection provides a way to transfer the distance relation from the original space to the smaller one, it does not show topological similarity between the state-action records. One possible way to overcome this issue is to use differential embedding for state-action representation. Unfortunately, the changing nature of embeddings entails the need of the constant or periodical memory update, which is alone an engineering challenge. We believe that our work provides the benefit of using episodic-memory for continuous control tasks and opens further research directions in this area. References [Bellemare et al., 2013] Marc G Bellemare, Yavar Naddaf, Joel Veness, and Michael Bowling. The arcade learning environment: An evaluation platform for general agents. Journal of Artificial Intelligence Research, 47:253 279, 2013. [Blundell et al., 2016] Charles Blundell, Benigno Uria, Alexander Pritzel, Yazhe Li, Avraham Ruderman, Joel Z Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21) Leibo, Jack Rae, Daan Wierstra, and Demis Hassabis. Model-free episodic control. ar Xiv preprint ar Xiv:1606.04460, 2016. [Daw et al., 2005] Nathaniel D Daw, Yael Niv, and Peter Dayan. Uncertainty-based competition between prefrontal and dorsolateral striatal systems for behavioral control. Nature neuroscience, 8(12):1704 1711, 2005. [Dhariwal et al., 2017] Prafulla Dhariwal, Christopher Hesse, Oleg Klimov, Alex Nichol, Matthias Plappert, Alec Radford, John Schulman, Szymon Sidor, Yuhuai Wu, and Peter Zhokhov. Openai baselines. https://github.com/openai/baselines, 2017. [Fujimoto et al., 2018] Scott Fujimoto, Herke Van Hoof, and David Meger. Addressing function approximation error in actor-critic methods. ar Xiv preprint ar Xiv:1802.09477, 2018. [Haarnoja et al., 2018a] Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. ar Xiv preprint ar Xiv:1801.01290, 2018. [Haarnoja et al., 2018b] Tuomas Haarnoja, Aurick Zhou, Kristian Hartikainen, George Tucker, Sehoon Ha, Jie Tan, Vikash Kumar, Henry Zhu, Abhishek Gupta, Pieter Abbeel, et al. Soft actor-critic algorithms and applications. ar Xiv preprint ar Xiv:1812.05905, 2018. [Johnson and Lindenstrauss, 1984] William B Johnson and Joram Lindenstrauss. Extensions of lipschitz mappings into a hilbert space. Contemporary mathematics, 26(189206):1, 1984. [Kapturowski et al., 2018] Steven Kapturowski, Georg Ostrovski, John Quan, Remi Munos, and Will Dabney. Recurrent experience replay in distributed reinforcement learning. In International conference on learning representations, 2018. [Kingma and Ba, 2014] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. [Lengyel and Dayan, 2007] M at e Lengyel and Peter Dayan. Hippocampal contributions to control: the thireep way. Advances in neural information processing systems, 20:889 896, 2007. [Lillicrap et al., 2015] Timothy P Lillicrap, Jonathan J Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. Continuous control with deep reinforcement learning. ar Xiv preprint ar Xiv:1509.02971, 2015. [Lin et al., 2018] Zichuan Lin, Tianqi Zhao, Guangwen Yang, and Lintao Zhang. Episodic memory deep qnetworks. ar Xiv preprint ar Xiv:1805.07603, 2018. [Lin, 1992] Long-Ji Lin. Self-improving reactive agents based on reinforcement learning, planning and teaching. Machine learning, 8(3-4):293 321, 1992. [Pritzel et al., 2017] Alexander Pritzel, Benigno Uria, Sriram Srinivasan, Adria Puigdomenech, Oriol Vinyals, Demis Hassabis, Daan Wierstra, and Charles Blundell. Neural episodic control. ar Xiv preprint ar Xiv:1703.01988, 2017. [Schaul et al., 2015] Tom Schaul, John Quan, Ioannis Antonoglou, and David Silver. Prioritized experience replay. ar Xiv preprint ar Xiv:1511.05952, 2015. [Silver et al., 2014] David Silver, Guy Lever, Nicolas Heess, Thomas Degris, Daan Wierstra, and Martin Riedmiller. Deterministic policy gradient algorithms. 2014. [Thrun and Schwartz, 1993] Sebastian Thrun and Anton Schwartz. Issues in using function approximation for reinforcement learning. In Proceedings of the 1993 Connectionist Models Summer School Hillsdale, NJ. Lawrence Erlbaum, 1993. [Vikbladh et al., 2017] Oliver Vikbladh, Daphna Shohamy, and Nathaniel Daw. Episodic contributions to model-based reinforcement learning. In Annual conference on cognitive computational neuroscience, CCN, 2017. [Wang et al., 2016] Ziyu Wang, Victor Bapst, Nicolas Heess, Volodymyr Mnih, Remi Munos, Koray Kavukcuoglu, and Nando de Freitas. Sample efficient actor-critic with experience replay. ar Xiv preprint ar Xiv:1611.01224, 2016. [Watkins and Dayan, 1992] Christopher JCH Watkins and Peter Dayan. Q-learning. Machine learning, 8(3-4):279 292, 1992. [Zhang et al., 2019] Zhizheng Zhang, Jiale Chen, Zhibo Chen, and Weiping Li. Asynchronous episodic deep deterministic policy gradient: Toward continuous control in computationally complex environments. IEEE Transactions on Cybernetics, 2019. Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21)