# RETRIEVAL-AUGMENTED REINFORCEMENT LEARNING **Anonymous authors** Paper under double-blind review ABSTRACT Most deep reinforcement learning (RL) algorithms distill experience into parametric behavior policies or value functions via gradient updates. While effective, this approach has several disadvantages: (1) it is computationally expensive, (2) it can take many updates to integrate experiences into the parametric model, (3) experiences that are not fully integrated do not appropriately influence the agent’s behavior, and (4) behavior is limited by the capacity of the model. In this paper we explore an alternative paradigm in which we train a network to map a dataset of past experiences to optimal behavior. Specifically, we augment an RL agent with a retrieval process (parameterized as a neural network) that has direct access to a dataset of experiences. This dataset can come from the agent’s past experiences, expert demonstrations, or any other relevant source. The retrieval process is trained to retrieve information from the dataset that may be useful in the current context, to help the agent achieve its goal faster and more efficiently. We integrate our method into two different RL agents: an offline DQN agent and an online R2D2 agent. In offline multi-task problems, we show that the retrieval-augmented DQN agent avoids task interference and learns faster than the baseline DQN agent. On Atari, we show that retrieval-augmented R2D2 learns significantly faster than the baseline R2D2 agent and achieves higher scores. We run extensive ablations to measure the contributions of the components of our proposed method. 1 INTRODUCTION A host is preparing a holiday meal for friends. They remember that the last time they went to the grocery store during the holiday season, all of the fresh produce was sold out. Thinking back to this past experience, they decide to go early! The hypothetical host is employing case-based reasoning (e.g., Kolodner, 1992; Leake, 1996). Here, an agent recalls a situation similar to the current one and uses information from the previous experience to solve the current task. This may involve adapting old solutions to meet new demands, or using previous experiences to make sense of new situations. In contrast, a dominant paradigm in modern reinforcement learning (RL) is to learn general purpose behaviour rules from the agent’s past experience. These rules are typically represented in the weights of a parametric policy or value function network model. Most deep RL algorithms integrate information across trajectories by iteratively updating network parameters using gradients that are computed along individual trajectories (collected online or stored in an experience replay dataset, Lin, 1992). For example, many off-policy algorithms reuse past experience by “replaying” trajectory snippets in order to compute weight updates for a value function represented by a deep network (Ernst et al., 2005; Riedmiller, 2005; Mnih et al., 2015b; Heess et al., 2015; Lillicrap et al., 2015). This paradigm has clear advantages but at least two interrelated limitations: First, after learning, an agent’s past experiences no longer plays a direct role in the agent’s behavior, even if it is relevant to the current situation. This occurs because detailed information in the agent’s past experience is lost due to practical constraints on network capacity. Second, since the information provided by individual trajectories first needs to be distilled into a general purpose parametric rule, an agent may not be able to exploit the specific guidance that a handful of individual past experiences could provide, nor rapidly incorporate novel experience that becomes available—it may take many replays through related traces in the past experiences for this to occur (Weisz et al., 2021). ----- Figure 1: Retrieval-augmented agent (R2A) architecture: (A) R2A augments the agent with a retrieval process. The retrieval process and the agent maintain separate internal states, mt and st, respectively. The retrieval process retrieves information relevant to the agent’s current internal state st from the retrieval batch, which is a pre-processed sample from the retrieval dataset B. The retrieved information ut is used by the agent process to inform its output (e.g., a policy or value function). (B) A batch of raw trajectories is sampled from the retrieval dataset _B and encoded (using the same encoder as the agent). Each encoded trajectory is then summarized via forward and a backward summarization_ functions (section 2.2) and sent to the retrieval process. (C) The retrieval process is parameterized as a recurrent model and the internal state mt is partitioned into slots. Each slot independently retrieves information from the retrieval batch, which is used to update the slot’s representation and sent to the agent process in ut. Slots also interact with each other via self-attention. See section 2.3 for more details. In this work, we develop an algorithm that overcomes these limitations by augmenting a standard reinforcement learning agent with a retrieval process (parameterized via a neural network). The purpose of the retrieval process is to help the agent achieve its objective by providing relevant contextual information. To this end, the retrieval process uses a learned attention mechanism to dynamically access a large pool of past trajectories stored in a dataset (e.g., a replay buffer), with the aim of integrating information across these. The proposed algorithm (R2A), shown in Figure 1, enables an agent to retrieve information from a dataset of trajectories. The high-level idea is to have two different processes. First, the retrieval process, makes a “query” to search for relevant contextual information in the dataset. Second, the agent process performs inference and learning based on the information provided by the retrieval process. These two processes have different internal states but interact to shape the representations and predictions of each other: the agent process provides the relevant context, and the retrieval process uses the context and its own internal state to generate a query and retrieve relevant information, which is in turn used by the agent process to shape the representation of its policy and value function (see Fig. 1A). Our proposed retrieval-augmented RL paradigm could take several forms. Here, we focus on a particular instantiation to assay and validate our hypothesis that learning a retrieval process can help an RL agent achieve its objectives. **Summary of experimental results. We want RL algorithms that are able to adapt to the available** data source and usefully ingest any dataset. Hence, we test the performance of the proposed method in three different scenarios. First, we evaluate it on Atari games in a single task setting. We build upon R2D2 (Kapturowski et al., 2018), a state-of-the-art off-policy RL algorithm. Second, we evaluate it on a multi-task offline RL environment, using DQN (Mnih et al., 2013) as the RL algorithm, where the data in the queried dataset belongs to the same task. Third, we evaluate it on a multi-task offline RL environment where the data in the dataset comes from multiple tasks. In all these cases, we show that R2A learns faster and achieves higher reward compared to the baseline. 2 RETRIEVAL-AUGMENTED AGENTS We now present our method for augmenting an RL agent with a retrieval process, thereby reducing the agent’s dependence on its model capacity, and enabling fast and flexible use of past experiences. A retrieval-augmented agent (R2A) consists of two main components: (1) the retrieval process, which takes in the current state of the agent, combines this with its own internal state, and retrieves relevant information from an external dataset of experiences; and (2) a standard reward-maximizing RL agent, which uses the retrieved information to improve its value or policy estimates. See Figure 1 for an overview. The retrieval process is trained to retrieve information that the agent can use to improve its performance, without explicit knowledge of the agent’s policy. Importantly, the retrieval process has its own internal state, which enables it to integrate and combine information across ----- retrievals. In the following, we focus on value-based methods, such as DQN (Mnih et al., 2015a) and R2D2 (Kapturowski et al., 2018), but our approach is equally applicable to policy-based methods. 2.1 RETRIEVAL-AUGMENTED AGENT Formally, the agent receives an input xt at each timestep t. Each input is processed by a neural encoder (e.g., a resnet if the input is an image) to obtain an abstract internal state for the agent **_st = fθ[enc][(][x][t][)][. For clarity, we focus here on the case of a single vector input, however, each input]_** could also include the history of past observations, actions, and rewards, as is the case when fθ[enc] is a recurrent network. These embeddings are used by the agent and retrieval processes. The retrieval process operates on a dataset B = {(xt, at, rt), . . ., (xt+T, at+T, rt+T )} of l-step trajectories, for _l ≥_ 1. This dataset could come from other agents or experts, as in offline RL or imitation learning, or consist of the growing set of the agent’s own experiences. Then, a retrieval-augmented agent (R2A) consists of the retrieval process and the agent process, parameterized by θ = {θ[enc], θ[retr], θ[agent]}, **Retrieval process fθ,[retr]B** [:][ m][t][−][1][,][ s][t] _[7→]_ **_[m][t][,][ u][t]_** **Agent process fθ[agent]** : st, ut _Qθ(st, ut, a)_ _7→_ **_Retrieval Process. The retrieval process is parameterized as a neural network and has an internal_** state mt. The retrieval process takes in the current abstract state of the agent process st and its own previous internal state mt 1 and uses these to retrieve relevant information from the dataset, which _−_ _B_ it then summarizes in a vector ut, and also updates its internal state mt. **_Agent Process. The current state of the agent st and the information from the retrieval process ut is_** then passed to the action-value function, itself used to select external actions. The above defines a parameterization for a retrieval-augmented agent. For retrieval to be effective, the retrieval process needs to: (1) be able to efficiently query a large dataset of trajectories, (2) learn and employ a similarity function to find relevant trajectories, and (3) encode and summarize the trajectories in a manner that allows efficient discovery of relevant past and future information. Below, we explain how we achieve these desiderata. At a high-level, to reduce computational complexity given a experience dataset of hundreds of thousands of trajectories, R2A operates on samples from the dataset. R2A then encodes and summarizes the trajectories in these samples using auxiliary losses and bi-directional sequence models to enable efficient retrieval of temporal information. Finally, R2A uses attention to select semantically relevant trajectories. 2.2 RETRIEVAL BATCH SAMPLING AND PRE-PROCESSING. **_Sampling a retrieval batch from the retrieval dataset. To reduce the computational complexity,_** R2A uniformly samples a large batch of past experiences from the retrieval dataset, and only uses the sampled batch for retrieving information. We denote the sampled batch as “retrieval batch” and the number of trajectories in the retrieval batch as nretrieval. **_Encoding and forward-backward summarization of the retrieval dataset and corresponding aux-_** **_iliary losses. Since the agent’s internal state extracts information from observations which relate to_** the task at hand, we choose to re-encode the raw experiences in the ”retrieval batch” using the agent encoder module (i.e., fθ[enc][). However, this representation is a function only of past observations (i.e.,] it’s a causal representation) and may not be fully compatible with the needs of the retrieval operation. For that reason, we propose to further encode the retrieved batch of information, by additionally learning a summarization function, applied on the output of the encoder module, and which captures information about the past and the future within a particular trajectory by using a bi-directional model (e.g., parameterized as a bi-directional RNN or a Transformer). **Forward Summarizer fθ[fwd]** : (s1, . . ., st) **_ht_** _7→_ **Backward Summarizer fθ[bwd]** : (sT, . . ., st) **_bt_** _7→_ For each trajectory in the retrieval batch, we represent each time-step within a trajectory by a set of two vectors hi,t and bi,t (Figure 6 in the appendix) where hi,t summarizes the past (i.e., from ----- **Algorithm 1 One timestep of a retrieval-augmented agent (R2A).** **_Input: Current input xt, previous retrieval process state mt_** 1 = **_mt_** 1,k _k_ 1, . . ., nf, _−_ _{_ _−_ _|_ _∈{_ _}}_ dataset of l-step trajectories B = {(x[i]t[,][ h][i]t[,][ b][i]t[, a][i]t[, r]t[i][)][ . . .][ (][x][i]t+l[,][ h][i]t+l[,][ b][i]t+l[, a][i]t+l[, r]t[i]+l[)][}] for l ≥ 1 and 1 ≤ _i ≤_ _ntraj, where h and b are the outputs of the forward & backward summarizers._ **Encode the current input at time-step t.** **_st = fθ[enc][(][x][t][)]_** **Step 1: Compute the query. For all 1 ≤** _k ≤_ _nf_, compute **_m[k]t_** 1 [= GRU]θ **_st, m[k]t_** 1 _−_ _−_ **_qt[k]_** [=][ f]query[(]m[c][k]t 1[)] _−_  c **Step 2: Identify the most relevant trajectories. For all 1 ≤** _k ≤_ _nf_ _, 1 ≤_ _j ≤_ _l and 1 ≤_ _i ≤_ _ntraj,_ **_κi,j = (h[i]j[W]ret[ e]_** [)][T] _ℓ[k]i,j_ [=] **_qt[k]√[κ]d[i,j]e_** _αi,j[k]_ [= softmax]  _ℓ[k]i,j_ . Given scores α, the top-ktraj trajectories (resp. top-kstates states) are selected and denoted by _t[k]_ [(resp.][ S]t[k][).]  _T_ **Step 3: Retrieve information from the most relevant trajectories and states.** _αi,j[k]_ [= softmax] _ℓ[k]i,j_, i ∈Tt[k][, j][ ∈S]t[k][.] **_gt[k]_** [=][ P]i,j _[α]i,j[k]_ **_[v][i,j]_** [where] **_[ v][i,j]_** [=][ b][i,j][W]ret[ v] **Step 4: Regularize the retrieved information by using information bottleneck.** **_zt[k]_** _[∼]_ _[p][(][z][|][g]t[k][)]_ **Step 5: Update the states of the slots.** Slotwise update using retrieved information: **_m[k]t_** **_m[k]t_** 1 [+][ z]t[k] _k_ 1, . . ., nf Joint slot update through self-attention:[←] [c] _−_ _∀_ _∈{_ _}_ **_cf[k]t_** [=][ c]m[k]t−1[W][ q]SA _∀k{1, . . ., nf_ _}_ **_c[k]t_** **_[κ]t[k][′]_** _βk,k′ = softmaxk′_ _√de_ where κ[k]t _[′]_ [= (]m[f][k]t _[′]_ **_[W]SA[ e]_** [)][T] _∀_ _k, k[′]_ _∈{1, . . ., nf_ _}_   **_m[k]t_** _[←]_ **_m[f][k]t_** [+][ P]k[′][ β][k,k][′] **_[v][k][′][ where][ v][k][′][ =][ f]m[k]t_** **_[W]SA[ v]_** _∀k ∈{1, . . ., nf_ _}_ **Step 6: Update the agent state using the retrieved information.** **_dt = stWag[q]_** **_κ[k]_** = (zt[k][W]ag[ e] [)][T] _∀k ∈{1, . . ., nf_ _}_ _γk = softmaxk_ **_d√tdκe[k]_** **_suett ← ←_** **_s[P]t +k_** _[γ] u[k]t[v][k][ where]_ **_[ v]_** _[k][ =][ z]t[k][W]ag[ v]_ _∀k ∈{1, . . ., nf_ _}._ _t[′]_ = 0 to t[′] = t time-steps of the i[th] trajectory) while bi,t summarizes the future (i.e., from t[′] = t to t[′] = ℓ time-steps) within the i[th] trajectory. In addition, taking inspiration from (Jaderberg et al., 2016; Trinh et al., 2018; Ke et al., 2019; Devlin et al., 2018; Mazoure et al., 2020; Banino et al., 2021), we use auxiliary losses to improve modeling of long term dependencies when training the parameters of our forward and backward summarizers. The goal of these losses is to force the representation (hi,t, bi,t)i,t 0 to capture meaningful information for the unknown downstream task. _≥_ For our experiments, we use supervised losses where we have access to actions or rewards in the retrieval batch. For ablations we also experiment with self-supervised losses. For supervised auxiliary losses, we use policy, value and reward prediction (Silver et al., 2017; Schrittwieser et al., 2019), and for self-supervised losses, we use a BERT-style masking loss (Devlin et al., 2018). 2.3 RETRIEVING CONTEXTUAL INFORMATION FROM PAST EXPERIENCES. In this section, we explain how the retrieval process, when provided with relevant contextual information represented by the agent’s current state st, interacts with the summarized information in the retrieval batch to select information ut to provide to the agent in return. ----- **Retrieval process state parameterization. We parameterize the process that retrieves information** from past experience as a structured parametric model with multiple separate memory slots (or sub-units). The state of the retrieval process is a set of1, . . ., nf (indexed by the agent time-step t). Slots are initialized randomly at the beginning of nf memory slots denoted by mt = {m[k]t _[|][ k][ ∈]_ _{_ _}}_ the episode. Each slot independently queries and retrieves relevant information from the pool of data. Subsequently the slots update their values independently based on the retrieved information. This is followed by an integration step during which information is shared between slots. Algorithm 1 provides a specification of the six steps of R2A which we now explain in detail below. **Step 1: Query computation. Each slot independently computes its prestate using an GRU on the** contextual information from the agent: **_m[k]t−1_** [= GRU][θ] **_st, m[k]t−1_** _∀k ∈{1, . . ., nf_ _}. Then,_ each slot independently computes a retrieval query which will be matched against information in the retrieval batch: qt[k] [=][ f][query][(]m[c][k]t−1[)][|][ k][ ∈{] c[1][, . . ., n][f] _[}}][1][ where]_ **_[ q]t[k]_** [is the query generated by the] _[ k][th]_ slot at timestep t. **Step 2: Identification of most relevant trajectories and states for each slot (Figure 5A). The** retrieval mechanism process uses an attention mechanism to match a query produced by the retrieval state associated with each slot m[k]t [to keys computed on each time step of each trajectory of the] retrieval batch. Formally, for each time step and each trajectory in the buffer, we compute a key κi,j by using a linear projection with matrix Wret[e] [on the forward summaries][ h][:][ κ][i,j] [= (][h][i]j[W][ e]ret[)][T][. Each] query qt[k] [is then matched with the set of all keys][ κ][i,j][, forming attention logits][2][ ℓ]i,j[k] [=] **_q√t[k][κ]d[i,j]e_** and corresponding attention weights αi,j[k] [= softmax] _ℓ[k]i,j_, i ≤ _ntraj, 0 ≤_ _j ≤_ _T_ .   Intuitively, αi,j[k] [captures the extent to which the][ j][th][ timestep of the] _[ i][th][ trajectory in the buffer will be]_ relevant to memory m[k]t [through the query][ q]t[k][. It follows that][ P]j _[α]i,j[k]_ [is a measure of how relevant] the i[th] trajectory is as a whole for qt[k][. Following previous work (Ke et al., 2018; Goyal et al., 2019b),] matching only on the most relevant trajectories will increase the robustness of the retrieval mechanism. We therefore select, for each query, the set Tt[k] [of][ k][traj] [trajectories with highest aggregated score] _j_ _[α]i,j[k]_ [. Note that typically the queries corresponding to different slots will select different top-][k][traj] trajectories from the retrieval batch. Following the selection of relevant trajectories, we renormalize P the weights α, and use a top-k mechanism once-again, this time to choose the set of most relevant states St[k] [(i.e. which maximizes][ P]i∈Tt[k] _[α][i,j][).]_ **Step 3: Information retrieval from the most relevant trajectories and states (Figure 5B). The** next step of the retrieval mechanism consists in computing the renormalized weights α on the subsets Tt[k] [and][ S]t[k] [(][α]i,j[k] [= softmax] _ℓ[k]i,j_, i ∈Tt[k][,][ j][ ∈S]t[k][) and use those weights to compute the] final retrieved information. The value retrieved from the buffer for query  _qt[k]_ [is computed as the] _α-weighted average of a linear function of the backward state summaries: gt[k]_ [=][ P]i,j _[α]i,j[k]_ **_[v][i,j][ where]_** **_vi,j = bi,jWret[v]_** [.] **Step 4: Regularization of the retrieved information via an information bottleneck. We regular-** ize the retrieved information gt[k] [via the use of an information bottleneck (Tishby et al., 2000; Alemi] et al., 2016); intuitively, each query pays a price of to exploit information from the retrieval batch. Formally, we parametrize two Gaussian distributions p(Z|gt[k][)][ (which has access to the retrieved] information) and r(Z|mt−1) (which only has access to the memory units). We define z[k]t [as a single] sample from p(Z|gt[k][)][ via the reparameterization trick to ensure differentiability (Kingma & Welling,] 2013; Rezende et al., 2014), and ensure that zt[k] [does not contain too much information by adding an] additional loss DKL(p||r) to the overall agent loss. We provide more details in the appendix. **Step 5: Slot update. The representation of each slot is first additively updated as a function of the** retrieved information **_m[k]t_** **_m[k]t_** 1 [+][ z]t[k][. The final update][ m]t[k] [consists in an update in which all] slots interact through self-attention (as normally done in Transformers) (see Algorithm 1 for details).[←] [c] _−_ **Step 6: Updating the agent state using retrieved information. f** The primary goal of the retrieval process is to extract information which may be useful for the agent process. Here, we use the retrieved information to change the state of the agent st. In the previous step, the retrieved information is used to change the state of the slots. For this step, we use a similar attention mechanism. Here, the 1fquery is parameterized as a neural network. 2We drop time indexing from attention-related quantities to simplify notation. ----- Figure 2: Relative improvement of retrieval-augmented R2D2 vs vanilla R2D2 on different Atari games, measured by human normalized score. We report the average score from 3 seeds per method and per game. query is a function of the state of the agent process dt = stWag[q] [, which are matched with the keys] **_κ[k]_** = (zt[k][W][ e]ag[)][T] _∀k ∈{1, . . ., nf_ _}, as a result of retrieving information, forming attention weights_ _γk = softmax_ **_d√tκde[k]_** . The values generated as a result of retrieved information by different slots and the attention weights are then used to update the state of the learning agent :  **_ut ←_** [P]k _[γ][k][v][k][ where]_ **_vk = zt[k][W]ag[ v]_** _∀k ∈{1, . . ., nf_ _}. ut is the result of the attention over the retrieved information_ which is then used to change the representation of the agent process :the representation of the action-value function Q(st, zt[k][, a][t][)][ by conditioning the value function on]st ← **_st + ut. We also shape_** the retrieved information zt[k] [(again via a similar attention mechanism).] e 3 EXPERIMENTAL RESULTS To evaluate R2A, we analyze its performance in three different settings. First, we evaluate on the Atari arcade learning environment (ALE) (Bellemare et al., 2013), a single-task off-policy setting where the retrieval process extracts relevant information from the agent’s current replay buffer. We then run a series of ablations of R2A to better understand the roles and effects of its components. Second, we evaluate on a multi-task, offline environment that we created, called gridroboman. In this environment, a single network is trained on data from all tasks and then, at evaluation time, the retrieval process queries a retrieval dataset containing only data from the task being evaluated. Third, we evaluate R2A in a multi-task offline version of the BabyAI environment (Chevalier-Boisvert et al., 2018). Again, a single network is trained on all tasks but now the retrieval process queries a retrieval dataset containing data from all tasks. In our experiments, the retrieval process selects the top ktraj = 10 most relevant trajectories (step 2, section 2.3), and then retrieves relevant information from the selected trajectories (step 3, section 2.3) using the top kstates = 10 most relevant states. To summarize the experiences in the retrieval batch we use a forward and backward GRU with 512 hidden units. To train the representation of these, we use auxiliary losses in the form of action, reward, and value prediction (section 2.2). See the section A.3 in appendix for more details about the experimental setup and training losses. 3.1 ATARI: SINGLE-TASK OFF-POLICY RL In this first experiment, our goal is to evaluate whether retrieval augmentation improves the performance and sample efficiency of a strong, recurrent baseline agent on a challenging, visually-complex environment—the Atari 2600 videogame suite (Bellemare et al., 2013). We use recurrent replay distributed DQN (R2D2, Kapturowski et al. (2018)) as the baseline agent and compare retrievalaugmented R2D2 (RA-R2D2) to vanilla R2D2. The retrieval dataset is the agent’s current replay buffer. The agent process is parameterized as an GRU (Hochreiter & Schmidhuber, 1997), and the retrieval process is parameterized as the slot-based recurrent architecture described in Section 2, using 8 slots. We randomly sample a retrieval batch consisting of 256 trajectories from the retrieval dataset. In figure 2 we report the relative improvement of RA-R2D2 versus the R2D2 baseline. We observe an increase of 11.32 ± 1.2% in the mean human normalized score relative to the R2D2 baseline over 2 billion environment steps, demonstrating that retrieval augmentation is quite beneficial in Atari and ----- that the agent’s own replay buffer is a useful source for retrieval. Raw scores and training curves are presented in appendix A.3. 3.1.1 ABLATIONS AND ANALYSIS To understand the benefit of different components of retrieval augmentation, we ablate RA-R2D2 on the 10 Atari games it performs best relative to R2D2. The ablations are as follows. _Importance of a separate retrieval process (A-1). In R2A, the re-_ trieval process and the agent process are parameterized separately, i.e., they have their own internal states. Here we examine what happens when the agent’s state is used to query the retrieval batch instead of using the retrieval state mt. To implement this we modify Step 1 of Algorithm 1 to make the query a direct function of the state of the agent, qt = fquery(st). The resulting query is used in the same way as above. This is similar to the episodic control baseline of (Pritzel et al., 2017). _Importance of retrieving information (A-2). We examine what hap-_ pens when the retrieval process does not have access to the retrieval dataset and hence no information is retrieved, keeping all else the same. In this ablation, the retrieval process updates the state of the slots using a transformer (i.e., in Step 1 we replace GRU with a transformer), and the updated state of the transformer is used by the agent process to shape the representation of its value function. Figure 3: Relative performance of ablated RA-R2D2 versus baseline R2D2 for 5 ablations on 10 Atari games. black lines represent Standard deviation. _Shorter retrieved trajectories (A-3). We decrease the length of the_ trajectories that are retrieved and summarized during retrieval pre-processing, thus reducing the amount of past and future information the retrieval process can retrieve. By default, the trajectories in the retrieval dataset are of length 80. To perform this ablation, we decrease the length of the effective context to only include information from 5 timesteps. _Importance of auxiliary losses to summarize retrieval batch (A-4, A-5). Here we study the use of_ self-supervised BERT style masking losses in addition to using action, reward and value prediction. We use these auxiliary losses on top of the representation learned by the forward and the backward dynamics model. To implement these losses, we randomly mask 15% of the hidden states in a trajectory, and then using the representation of hidden states at other time-steps, we predict the representation of masked hidden states. In A-5, we study using only self-supervised BERT style masking losses for summarizing the trajectories. Figure 3 shows the performance of RA-R2D2 and each ablation relative to the R2D2 baseline. Ablation A-1 shows that it is crucial to parameterize the agent process and retrieval process separately, as using the agent state does no better than the baseline. Ablation A-2 shows that the retrieval process finds relevant information that is useful to the agent. It also shows that the performance gains of R2A are not simply due to an increase in model capacity but depend on retrieving information from the retrieval dataset. Ablation A-3 shows that decreasing the length of the context in the retrieval dataset results in worse performance, thus showing the importance of our forward and backward trajectory summarization. Ablation A-4 demonstrates that the performance of R2A can further be improved by incorporating BERT style auxiliary losses but that only using BERT style auxiliary losses results in worse performance (but still better than the baseline R2D2). For completeness, Figure 10 in the appendix repeats the ablations on the 5 games that RA-R2D2 performs worst (relative to R2D2). The main takeaway is that the performance of the R2A can be greatly improved by optimizing hyperparameters for each game separately (A-6), which we did not do for our experiments. 3.2 GRIDROBOMAN: MULTI-TASK OFFLINE RL WITH A TASK-SPECIFIC RETRIEVAL DATASET Beyond querying the agent’s own experiences, retrieval can provide helpful information from other sources of experiences, including experts or other agents as the case in offline RL where the agent must learn from a fixed dataset of experiences generated by other agents without interacting with the environment during training. A major challenge in offline RL is distributional shift—the mismatch ----- (a) Training and testing on 10 tasks. (b) Training and testing on 20 tasks. (c) Training and testing on 30 tasks. Figure 4: Gridroboman: Multi-task offline RL with a task-specific retrieval dataset. Average episode return vs. learner steps for the multitask gridroboman environment when training and evaluating on 10, 20, and 30 tasks. With fewer tasks (a), the baseline DQN agent (blue) and the retrieval-augmented DQN agent (orange) perform identically; however, when the number of tasks increases (b, c), the retrieval-augmented agent learns much more effectively than the baseline DQN agent. Results are the average of 3 seeds for each method. Curves for individual tasks are shown in the appendix (Figure 12). between the distribution of states in the training data and those visited by the agent when acting— which makes it difficult to learn an accurate value function for states and actions rarely seen during training. We hypothesize that the retrieval process can improve performance in the offline setting by retrieving trajectories (including states, actions, and rewards) relevant to the agent’s current state, particularly for states and actions that are rare in the offline dataset. We test this hypothesis on a multi-task offline RL setup where a single agent is trained on multiple tasks simultaneously but at training and evaluation time the retrieval dataset contains only trajectories from the evaluated task. For this experiment, we created a minimalistic grid-world-based robotic manipulation environment (gridroboman) with 30 tasks related to the three objects (red, green, and blue) on the board. Gridroboman is built on the pycolab game engine (Stepleton et al.). The environment is inspired by the challenges of robotic manipulation, and includes tasks such as “go to object X” and “put object X on object Y”. Details of the environment, its tasks, and an example figure are presented in the appendix (section A.4). Here, we incorporate retrieval augmentation into a vanilla DQN agent as agent-state-recurrence is not needed for this task. Figure 4 shows the results of training retrievalaugmented DQN (RA-DQN, orange) and DQN (blue) on increasing numbers of tasks. With fewer tasks, RA-DQN and DQN perform identically; however, when the number of training tasks increases the retrieval-augmented agent is able to learn much more effectively than the baseline agent. Training on more tasks requires either additional model capacity or the ability to extract information from fewer relevant samples for each task. By directly querying task-relevant experiences in the offline dataset, retrieval augmentation improves sample efficiency. Note that while the retrieval process does afford extra model capacity to the agent directly, ablation A-2 in section 3.1.1 shows that the retrieved information is what is crucial to performance, not the increased capacity. 3.3 BABYAI: MULTI-TASK OFFLINE RL WITH A MULTI-TASK RETRIEVAL DATASET In this third experiment, our goal is to evaluate the benefit of retrieval augmentation when data from other tasks is present in the retrieval dataset. Multi-task retrieval data can be either harmful if the retrieved information misguides the agent or beneficial if information from the other tasks is relevant to the current task. Due to the use of attention in the retrieval process, we hypothesize that R2A will be able to ignore distracting information and retrieve relevant information from other tasks. To test our hypothesis, we use the BabyAI environment (Chevalier-Boisvert et al., 2018), a partially observable multi-room grid world in which harder tasks are composed of simpler tasks and are formulated using subsets of a synthetic language. At the start of each episode, the agent is placed in a random room and must navigate to a randomly located goal. Due to the partial observability, we use a recurrent DQN (RDQN) agent as the baseline and compare its performance to a retrieval-augmented RDQN (RA-RDQN) agent. As is common in this environment, we measure the success rate of each agent, defined as the ratio of tasks the agent was able to accomplish given a fixed number of steps for each task. Table 1 shows the performance of RA-RDQN with a multi-task replay, RA-RDQN with a replay specific to the current task, and the baseline for varying amounts of offline training data (50K trajectories per task versus 200K trajectories per task). As expected from the previous experiment, retrieval augmentation improves performance over the baseline when using a single-task replay. Performance further improves when using a multi-task replay. We believe that this is due to ----- Table 1: BabyAI: Multi-task offline RL with a multi-task retrieval dataset. Mean success rate of retrieval-augmented recurrent DQN (RARDQN) as compared to a recurrent DQN (RDQN) baseline on the 40 BabyAI levels as function of the amount of training data. RA-RDQN is run twice, once with only the current task being evaluated in the retrieval dataset and once with all tasks in the retrieval dataset. Results are the average of 3 random seeds with standard errors. **Method** **Success Rate (50K)** **Success Rate (200K)** RDQN 32% ± 4% 45% ± 6% RA-RDQN (single-task retrieval buffer) 48% ± 4% 64% ± 5% RA-RDQN (multi-task retrieval buffer, without IB) 47% ± 3% 59% ± 6% RA-RDQN (multi-task retrieval buffer) 55% ± 5% 74% ± 3% the compositional nature of tasks in BabyAI, where retrieving information about the current subtask is more informative than retrieving information about the overall task. 4 RELATED WORK **Episodic control. The idea of allowing deep RL agents to adapt based on past experiences using a** non-parametric memory is not new (Blundell et al., 2016; Pritzel et al., 2017; Hansen et al., 2018; Eysenbach et al., 2019; van Hasselt et al., 2019). The basic idea is that the agent is equipped with an episodic memory system, which is used to remember and recall past experiences to inform decisions. There are two important differences between R2A and these methods. (1) In these methods, a local action-value function is constructed by using information about the nearest neighbors in the replay buffer, and then the agent makes a decision about which action to execute based on both the local value function as well as the global value function. However, in the proposed work, we employ a parameterized network (the retrieval process), which has access to the information in the replay buffer, and the agent process uses the retrieved information to shape the predictions of its value function in a fully differentiable way (using attention). (2) In these episodic control methods, there is only one process (the agent), which has direct access to the replay buffer. However, in R2A, the agent process only has indirect access to the replay buffer via the retrieval process. **Separation of concerns. In Hierarchical RL (HRL) (Heess et al., 2016; Frans et al., 2017; Vezhnevets** et al., 2017; Florensa et al., 2017; Hausman et al., 2018; Goyal et al., 2019c), there’s separation of concerns among different policies, each policy focuses on a different aspect of the task, e.g., giving task relevant information to the high level policy only such that low level policy learns behaviours that are task agnostic. In these methods, the high level policy shapes the behaviour of low level policy by either influencing representations or by influencing rewards. It is possible to view our work through an analogous lens: wherein the “retrieval process” is the higher level policy (and has access to the all the information in the replay buffer) and is influencing the representation of the agent process that is interacting with the environment. However, there are also notable differences in our work—for instance, the agent process also directly shapes the representation of the retrieval process, which is generally not the case in HRL (e.g., in Vezhnevets et al. (2017) the manager directly influences the worker, but the worker does not directly influence the manager). **Retrieval in language models. Several retrieval-based methods have recently been developed for** question answering, controllable generation, and machine translation (Guu et al., 2020; Lee et al., 2019; Lewis et al., 2020; Sun et al., 2021). The general scheme in such methods is to combine a parametric model (like a BERT-style masked language model or a pre-trained seq2seq model) with a non-parametric retrieval system. They show that such models generate diverse and factually correct language than the state of the art parametric only baselines. These methods share some similarities with our proposed model since they all involve a retrieval component but focus on different domains. Additional related work is discussed in the appendix (section A.1). 5 CONCLUSION. In this work, we developed R2A, an algorithm that augments an RL agent with a retrieval process. The retrieval process and the agent have separate states and shape the representation and predictions of each other via attention. The goal of the retrieval process is to retrieve useful information from a dataset of experiences to help the agent achieve its objective more efficiently and effectively. We show that R2A improves sample efficiency over R2D2, a strong off-policy agent, and compensates for insufficient capacity when training in a multi-task offline RL environments. Multiple ablations show the importance of the different components of R2A, including retrieving information from past ----- experiences and parameterizing the agent and retrieval process separately instead of giving the agent process direct access to the replay buffer. REFERENCES Abbas Abdolmaleki, Jost Tobias Springenberg, Yuval Tassa, Remi Munos, Nicolas Heess, and Martin Riedmiller. Maximum a posteriori policy optimisation. arXiv preprint arXiv:1806.06920, 2018. Alexander A Alemi, Ian Fischer, Joshua V Dillon, and Kevin Murphy. Deep variational information bottleneck. arXiv preprint arXiv:1612.00410, 2016. James F Allen and Johannes A Koomen. Planning using a temporal world model. In Proceedings of _the Eighth international joint conference on Artificial intelligence-Volume 2, pp. 741–747, 1983._ Jose A Arjona-Medina, Michael Gillhofer, Michael Widrich, Thomas Unterthiner, Johannes Brandstetter, and Sepp Hochreiter. RUDDER: Return decomposition for delayed rewards. arXiv preprint _arXiv:1806.07857, 2018._ Andrea Banino, Adria Puigdomenech Badia, Jacob C Walker, Tim Scholtes, Jovana Mitrovic, and Charles Blundell. Coberl: Contrastive bert for reinforcement learning. In ICML 2021 Workshop _on Unsupervised Reinforcement Learning, 2021._ Peter W Battaglia, Razvan Pascanu, Matthew Lai, Danilo Rezende, and Koray Kavukcuoglu. Interaction networks for learning about objects, relations and physics. arXiv preprint arXiv:1612.00222, 2016. Peter W Battaglia, Jessica B Hamrick, Victor Bapst, Alvaro Sanchez-Gonzalez, Vinicius Zambaldi, Mateusz Malinowski, Andrea Tacchetti, David Raposo, Adam Santoro, Ryan Faulkner, et al. Relational inductive biases, deep learning, and graph networks. arXiv preprint arXiv:1806.01261, 2018. Marc G Bellemare, Yavar Naddaf, Joel Veness, and Michael Bowling. The arcade learning environment: An evaluation platform for general agents. Journal of Artificial Intelligence Research, 47: 253–279, 2013. Charles Blundell, Benigno Uria, Alexander Pritzel, Yazhe Li, Avraham Ruderman, Joel Z Leibo, Jack Rae, Daan Wierstra, and Demis Hassabis. Model-free episodic control. arXiv preprint _arXiv:1606.04460, 2016._ Mark Chen, Alec Radford, Rewon Child, Jeffrey Wu, Heewoo Jun, David Luan, and Ilya Sutskever. Generative pretraining from pixels. In International Conference on Machine Learning, pp. 1691– 1703. PMLR, 2020a. Ting Chen, Simon Kornblith, Kevin Swersky, Mohammad Norouzi, and Geoffrey Hinton. Big self-supervised models are strong semi-supervised learners. arXiv preprint arXiv:2006.10029, 2020b. Maxime Chevalier-Boisvert, Dzmitry Bahdanau, Salem Lahlou, Lucas Willems, Chitwan Saharia, Thien Huu Nguyen, and Yoshua Bengio. Babyai: A platform to study the sample efficiency of grounded language learning. In International Conference on Learning Representations, 2018. Thomas M Cover. Elements of information theory. John Wiley & Sons, 1999. Wojciech M Czarnecki, Razvan Pascanu, Simon Osindero, Siddhant Jayakumar, Grzegorz Swirszcz, and Max Jaderberg. Distilling policy distillation. In The 22nd International Conference on _Artificial Intelligence and Statistics, pp. 1331–1340. PMLR, 2019._ Mostafa Dehghani, Stephan Gouws, Oriol Vinyals, Jakob Uszkoreit, and Łukasz Kaiser. Universal transformers. arXiv preprint arXiv:1807.03819, 2018. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018. ----- Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. _arXiv preprint_ _arXiv:2010.11929, 2020._ Damien Ernst, Pierre Geurts, and Louis Wehenkel. Tree-based batch mode reinforcement learning. _Journal of Machine Learning Research, 6:503–556, 2005._ Benjamin Eysenbach, Ruslan Salakhutdinov, and Sergey Levine. Search on the replay buffer: Bridging planning and reinforcement learning. arXiv preprint arXiv:1906.05253, 2019. Carlos Florensa, Yan Duan, and Pieter Abbeel. Stochastic neural networks for hierarchical reinforcement learning. arXiv preprint arXiv:1704.03012, 2017. Meire Fortunato, Melissa Tan, Ryan Faulkner, Steven Hansen, Adrià Puigdomènech Badia, Gavin Buttimore, Charlie Deck, Joel Z Leibo, and Charles Blundell. Generalization of reinforcement learners with working and episodic memory. arXiv preprint arXiv:1910.13406, 2019. Kevin Frans, Jonathan Ho, Xi Chen, Pieter Abbeel, and John Schulman. Meta learning shared hierarchies. arXiv preprint arXiv:1710.09767, 2017. Alexandre Galashov, Siddhant M Jayakumar, Leonard Hasenclever, Dhruva Tirumala, Jonathan Schwarz, Guillaume Desjardins, Wojciech M Czarnecki, Yee Whye Teh, Razvan Pascanu, and Nicolas Heess. Information asymmetry in KL-regularized RL. arXiv preprint arXiv:1905.01240, 2019. Anirudh Goyal, Philemon Brakel, William Fedus, Soumye Singhal, Timothy Lillicrap, Sergey Levine, Hugo Larochelle, and Yoshua Bengio. Recall traces: Backtracking models for efficient reinforcement learning. arXiv preprint arXiv:1804.00379, 2018. Anirudh Goyal, Riashat Islam, Daniel Strouse, Zafarali Ahmed, Matthew Botvinick, Hugo Larochelle, Yoshua Bengio, and Sergey Levine. Infobot: Transfer and exploration via the information bottleneck. arXiv preprint arXiv:1901.10902, 2019a. Anirudh Goyal, Alex Lamb, Jordan Hoffmann, Shagun Sodhani, Sergey Levine, Yoshua Bengio, and Bernhard Schölkopf. Recurrent independent mechanisms. arXiv preprint arXiv:1909.10893, 2019b. Anirudh Goyal, Shagun Sodhani, Jonathan Binas, Xue Bin Peng, Sergey Levine, and Yoshua Bengio. Reinforcement learning with competitive ensembles of information-constrained primitives. arXiv _preprint arXiv:1906.10667, 2019c._ Anirudh Goyal, Yoshua Bengio, Matthew Botvinick, and Sergey Levine. The variational bandwidth bottleneck: Stochastic evaluation on an information budget. arXiv preprint arXiv:2004.11935, 2020a. Anirudh Goyal, Alex Lamb, Phanideep Gampa, Philippe Beaudoin, Sergey Levine, Charles Blundell, Yoshua Bengio, and Michael Mozer. Object files and schemata: Factorizing declarative and procedural knowledge in dynamical systems. arXiv preprint arXiv:2006.16225, 2020b. Anirudh Goyal, Aniket Didolkar, Nan Rosemary Ke, Charles Blundell, Philippe Beaudoin, Nicolas Heess, Michael Mozer, and Yoshua Bengio. Neural production systems. arXiv preprint _arXiv:2103.01937, 2021._ Caglar Gulcehre, Ziyu Wang, Alexander Novikov, Thomas Paine, Sergio Gómez, Konrad Zolna, Rishabh Agarwal, Josh S Merel, Daniel J Mankowitz, Cosmin Paduraru, et al. Rl unplugged: A collection of benchmarks for offline reinforcement learning. Advances in Neural Information _Processing Systems, 33, 2020._ Kelvin Guu, Kenton Lee, Zora Tung, Panupong Pasupat, and Ming-Wei Chang. REALM: Retrievalaugmented language model pre-training. arXiv preprint arXiv:2002.08909, 2020. ----- Steven Hansen, Pablo Sprechmann, Alexander Pritzel, André Barreto, and Charles Blundell. Fast deep reinforcement learning using online adjustments from the past. arXiv preprint arXiv:1810.08163, 2018. Karol Hausman, Jost Tobias Springenberg, Ziyu Wang, Nicolas Heess, and Martin Riedmiller. Learning an embedding space for transferable robot skills. In International Conference on _Learning Representations, 2018._ Nicolas Heess, Greg Wayne, David Silver, Timothy Lillicrap, Yuval Tassa, and Tom Erez. Learning continuous control policies by stochastic value gradients, 2015. Nicolas Heess, Greg Wayne, Yuval Tassa, Timothy Lillicrap, Martin Riedmiller, and David Silver. Learning and transfer of modulated locomotor controllers. arXiv preprint arXiv:1610.05182, 2016. Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8): 1735–1780, 1997. Max Jaderberg, Volodymyr Mnih, Wojciech Marian Czarnecki, Tom Schaul, Joel Z Leibo, David Silver, and Koray Kavukcuoglu. Reinforcement learning with unsupervised auxiliary tasks. arXiv _preprint arXiv:1611.05397, 2016._ Steven Kapturowski, Georg Ostrovski, John Quan, Remi Munos, and Will Dabney. Recurrent experience replay in distributed reinforcement learning. In International conference on learning _representations, 2018._ Nan Rosemary Ke, Anirudh Goyal, Olexa Bilaniuk, Jonathan Binas, Michael C Mozer, Chris Pal, and Yoshua Bengio. Sparse attentive backtracking: Temporal credit assignment through reminding. In _Advances in Neural Information Processing Systems, pp. 7640–7651, 2018._ Nan Rosemary Ke, Amanpreet Singh, Ahmed Touati, Anirudh Goyal, Yoshua Bengio, Devi Parikh, and Dhruv Batra. Learning dynamics model in reinforcement learning by incorporating the long term future. arXiv preprint arXiv:1903.01599, 2019. Diederik P Kingma and Max Welling. Auto-encoding variational Bayes. _arXiv preprint_ _arXiv:1312.6114, 2013._ Janet L Kolodner. An introduction to case-based reasoning. Artificial intelligence review, 6(1):3–34, 1992. David B Leake. Case-based reasoning: experiences, lessons, and future directions. 1996. Kenton Lee, Ming-Wei Chang, and Kristina Toutanova. Latent retrieval for weakly supervised open domain question answering. arXiv preprint arXiv:1906.00300, 2019. Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, et al. Retrieval-augmented generation for knowledge-intensive nlp tasks. arXiv preprint arXiv:2005.11401, 2020. Timothy P Lillicrap, Jonathan J Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. Continuous control with deep reinforcement learning. arXiv _preprint arXiv:1509.02971, 2015._ Long-Ji Lin. Self-improving reactive agents based on reinforcement learning, planning and teaching. _Machine learning, 8(3-4):293–321, 1992._ Siqi Liu, Guy Lever, Zhe Wang, Josh Merel, SM Eslami, Daniel Hennes, Wojciech M Czarnecki, Yuval Tassa, Shayegan Omidshafiei, Abbas Abdolmaleki, et al. From motor control to team play in simulated humanoid football. arXiv preprint arXiv:2105.12196, 2021. Zhuang Liu, Xuanlin Li, Bingyi Kang, and Trevor Darrell. Regularization matters in policy optimization. 2019. Bogdan Mazoure, Remi Tachet des Combes, Thang Doan, Philip Bachman, and R Devon Hjelm. Deep reinforcement and infomax learning. arXiv preprint arXiv:2006.07217, 2020. ----- Josh Merel, Matthew Botvinick, and Greg Wayne. Hierarchical motor control in mammals and machines. Nature communications, 10(1):1–12, 2019. Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. Playing atari with deep reinforcement learning. arXiv preprint _arXiv:1312.5602, 2013._ Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A. Rusu, Joel Veness, Marc G. Bellemare, Alex Graves, Martin Riedmiller, Andreas K. Fidjeland, Georg Ostrovski, Stig Petersen, Charles Beattie, Amir Sadik, Ioannis Antonoglou, Helen King, Dharshan Kumaran, Daan Wierstra, Shane Legg, and Demis Hassabis. Human-level control through deep reinforcement learning. _Nature, 518(7540):529–533, February 2015a._ Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A Rusu, Joel Veness, Marc G Bellemare, Alex Graves, Martin Riedmiller, Andreas K Fidjeland, Georg Ostrovski, et al. Human-level control through deep reinforcement learning. nature, 518(7540):529–533, 2015b. Ashvin Nair, Murtaza Dalal, Abhishek Gupta, and Sergey Levine. AWAC: Accelerating online reinforcement learning with offline datasets. 2020. Karl Pertsch, Youngwoon Lee, and Joseph J Lim. Accelerating reinforcement learning with learned skill priors. arXiv preprint arXiv:2010.11944, 2020. Alexander Pritzel, Benigno Uria, Sriram Srinivasan, Adria Puigdomenech Badia, Oriol Vinyals, Demis Hassabis, Daan Wierstra, and Charles Blundell. Neural episodic control. In International _Conference on Machine Learning, pp. 2827–2836. PMLR, 2017._ Sébastien Racanière, Théophane Weber, David P Reichert, Lars Buesing, Arthur Guez, Danilo Rezende, Adria Puigdomenech Badia, Oriol Vinyals, Nicolas Heess, Yujia Li, et al. Imaginationaugmented agents for deep reinforcement learning. In Proceedings of the 31st International _Conference on Neural Information Processing Systems, pp. 5694–5705, 2017._ Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019. Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra. Stochastic backpropagation and approximate inference in deep generative models. In International conference on machine learning, pp. 1278–1286. PMLR, 2014. Martin Riedmiller. Neural fitted Q iteration–first experiences with a data efficient neural reinforcement learning method. In European conference on machine learning, pp. 317–328. Springer, 2005. Julian Schrittwieser, Ioannis Antonoglou, Thomas Hubert, Karen Simonyan, Laurent Sifre, Simon Schmitt, Arthur Guez, Edward Lockhart, Demis Hassabis, Thore Graepel, et al. Mastering atari, go, chess and shogi by planning with a learned model. arXiv preprint arXiv:1911.08265, 2019. Julian Schrittwieser, Ioannis Antonoglou, Thomas Hubert, Karen Simonyan, Laurent Sifre, Simon Schmitt, Arthur Guez, Edward Lockhart, Demis Hassabis, Thore Graepel, et al. Mastering atari, go, chess and shogi by planning with a learned model. Nature, 588(7839):604–609, 2020. Noah Y Siegel, Jost Tobias Springenberg, Felix Berkenkamp, Abbas Abdolmaleki, Michael Neunert, Thomas Lampe, Roland Hafner, Nicolas Heess, and Martin Riedmiller. Keep doing what worked: Behavioral modelling priors for offline reinforcement learning. arXiv preprint arXiv:2002.08396, 2020. David Silver. Reinforcement learning and simulation-based search in computer go. 2009. David Silver, Richard S Sutton, and Martin Müller. Sample-based learning and search with permanent and transient memories. In Proceedings of the 25th international conference on Machine learning, pp. 968–975, 2008. David Silver, Aja Huang, Chris J Maddison, Arthur Guez, Laurent Sifre, George Van Den Driessche, Julian Schrittwieser, Ioannis Antonoglou, Veda Panneershelvam, Marc Lanctot, et al. Mastering the game of go with deep neural networks and tree search. nature, 529(7587):484–489, 2016. ----- David Silver, Hado Hasselt, Matteo Hessel, Tom Schaul, Arthur Guez, Tim Harley, Gabriel DulacArnold, David Reichert, Neil Rabinowitz, Andre Barreto, et al. The predictron: End-to-end learning and planning. In International Conference on Machine Learning, pp. 3191–3199. PMLR, 2017. David Silver, Thomas Hubert, Julian Schrittwieser, Ioannis Antonoglou, Matthew Lai, Arthur Guez, Marc Lanctot, Laurent Sifre, Dharshan Kumaran, Thore Graepel, et al. A general reinforcement learning algorithm that masters chess, shogi, and go through self-play. Science, 362(6419): 1140–1144, 2018. Jost Tobias Springenberg, Nicolas Heess, Daniel Mankowitz, Josh Merel, Arunkumar Byravan, Abbas Abdolmaleki, Jackie Kay, Jonas Degrave, Julian Schrittwieser, Yuval Tassa, et al. Local search for policy iteration in continuous control. arXiv preprint arXiv:2010.05545, 2020. [Tom Stepleton, Olivier Delalleau, Georg Ostrovski, and Sam Wenke. URL https://github.](https://github.com/deepmind/pycolab) [com/deepmind/pycolab.](https://github.com/deepmind/pycolab) DJ Strouse, Max Kleiman-Weiner, Josh Tenenbaum, Matt Botvinick, and David Schwab. Learning to share and hide intentions using information regularization. arXiv preprint arXiv:1808.02093, 2018. Yu Sun, Shuohuan Wang, Shikun Feng, Siyu Ding, Chao Pang, Junyuan Shang, Jiaxiang Liu, Xuyi Chen, Yanbin Zhao, Yuxiang Lu, et al. Ernie 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation. arXiv preprint arXiv:2107.02137, 2021. Richard S Sutton. Dyna, an integrated architecture for learning, planning, and reacting. ACM Sigart _Bulletin, 2(4):160–163, 1991._ Yee Teh, Victor Bapst, Wojciech M Czarnecki, John Quan, James Kirkpatrick, Raia Hadsell, Nicolas Heess, and Razvan Pascanu. Distral: Robust multitask reinforcement learning. In Advances in _Neural Information Processing Systems, pp. 4496–4506, 2017._ Naftali Tishby, Fernando C Pereira, and William Bialek. The information bottleneck method. arXiv _preprint physics/0004057, 2000._ Trieu Trinh, Andrew Dai, Thang Luong, and Quoc Le. Learning longer-term dependencies in rnns with auxiliary losses. In International Conference on Machine Learning, pp. 4965–4974. PMLR, 2018. Hado Van Hasselt, Arthur Guez, and David Silver. Deep reinforcement learning with double qlearning. In Proceedings of the AAAI conference on artificial intelligence, volume 30, 2016. Hado P van Hasselt, Matteo Hessel, and John Aslanides. When to use parametric models in reinforcement learning? Advances in Neural Information Processing Systems, 32:14322–14333, 2019. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in neural information _processing systems, pp. 5998–6008, 2017._ Mel Vecerik, Todd Hester, Jonathan Scholz, Fumin Wang, Olivier Pietquin, Bilal Piot, Nicolas Heess, Thomas Rothörl, Thomas Lampe, and Martin Riedmiller. Leveraging demonstrations for deep reinforcement learning on robotics problems with sparse rewards. arXiv preprint arXiv:1707.08817, 2017. Rishi Veerapaneni, John D Co-Reyes, Michael Chang, Michael Janner, Chelsea Finn, Jiajun Wu, Joshua Tenenbaum, and Sergey Levine. Entity abstraction in visual model-based reinforcement learning. In Conference on Robot Learning, pp. 1439–1456. PMLR, 2020. Alexander Sasha Vezhnevets, Simon Osindero, Tom Schaul, Nicolas Heess, Max Jaderberg, David Silver, and Koray Kavukcuoglu. Feudal networks for hierarchical reinforcement learning. In _International Conference on Machine Learning, pp. 3540–3549. PMLR, 2017._ ----- Oriol Vinyals, Igor Babuschkin, Wojciech M Czarnecki, Michaël Mathieu, Andrew Dudzik, Junyoung Chung, David H Choi, Richard Powell, Timo Ewalds, Petko Georgiev, et al. Grandmaster level in starcraft II using multi-agent reinforcement learning. Nature, 575(7782):350–354, 2019. Nicholas Watters, Loic Matthey, Matko Bosnjak, Christopher P Burgess, and Alexander Lerchner. Cobra: Data-efficient model-based rl through unsupervised object discovery and curiosity-driven exploration. arXiv preprint arXiv:1905.09275, 2019. Greg Wayne, Chia-Chun Hung, David Amos, Mehdi Mirza, Arun Ahuja, Agnieszka GrabskaBarwinska, Jack Rae, Piotr Mirowski, Joel Z Leibo, Adam Santoro, et al. Unsupervised predictive memory in a goal-directed agent. arXiv preprint arXiv:1803.10760, 2018. Gellért Weisz, Philip Amortila, and Csaba Szepesvári. Exponential lower bounds for planning in mdps with linearly-realizable optimal action-value functions. In Algorithmic Learning Theory, pp. 1237–1264. PMLR, 2021. Vinicius Zambaldi, David Raposo, Adam Santoro, Victor Bapst, Yujia Li, Igor Babuschkin, Karl Tuyls, David Reichert, Timothy Lillicrap, Edward Lockhart, et al. Deep reinforcement learning with relational inductive biases. In International Conference on Learning Representations, 2018. ----- A APPENDIX CONTENTS A.1 Extended Relevant Work. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 A.2 Information Theoretic Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . 17 A.3 Atari: Implementation details and raw scores for R2D2 and RA-R2D2. . . . . . . . 17 A.3.1 Retrieval Process: Retrieving information from the retrieval buffer. . . . . . 20 A.3.2 Details on ablations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 A.3.3 Additional Atari ablations . . . . . . . . . . . . . . . . . . . . . . . . . . 22 A.4 Gridroboman environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 A.4.1 Gridroboman experiment setup . . . . . . . . . . . . . . . . . . . . . . . . 27 A.4.2 Gridroboman hyperparameters . . . . . . . . . . . . . . . . . . . . . . . . 27 A.4.3 Gridroboman evaluation curves when training on all 30 tasks . . . . . . . . 28 A.5 BabyAI environment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28 A.6 Limitations and Future Work. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 A.7 Social Impact. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 A.8 Rebuttal Phase. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 A.1 EXTENDED RELEVANT WORK. **Model-based RL. Their are different ways to integrate knowledge across past experiences. One** of the most common way is by learning a model of the world, and using the predictions from the model to improve the policy and the value function (Sutton, 1991; Silver et al., 2008; Silver, 2009; Allen & Koomen, 1983; Silver et al., 2016; Racanière et al., 2017; Silver et al., 2018; Springenberg et al., 2020; Schrittwieser et al., 2020). To integrate information across different episodes (potentially separated by many time-steps), a model may needed be unrolled for many time-steps leading to compounding errors. In R2A, the agent has direct access to the information in the retrieval dataset, and querying across multiple trajectories (in parallel) in the retrieval dataset potentially separated by hundreds of time-steps. **Structural Inductive Biases. Deep learning have proposed structural inductive biases such as** Transformers (Vaswani et al., 2017; Dehghani et al., 2018; Radford et al., 2019; Chen et al., 2020a;b; Dosovitskiy et al., 2020) or slot based recurrent architectures (Battaglia et al., 2016; Zambaldi et al., 2018; Battaglia et al., 2018; Goyal et al., 2019b; Watters et al., 2019; Goyal et al., 2020b; Veerapaneni et al., 2020; Goyal et al., 2021) where the induced structure has improved generalization, model-size scaling, and longrange dependencies. **Reinforcement Learning with Offline Datasets. Recent work in RL has tried exploiting large** datasets collected across many tasks to improve the sample efficiency of RL algorithms (Vecerik et al., 2017; Pertsch et al., 2020; Nair et al., 2020; Siegel et al., 2020). An advantage of such large datasets is that they can be collected cheaply, and can then be reused for learning many downstream tasks. A general scheme for exploiting information about such task-agnostic datasets is either using them to directly improve the value function, or by extracting a set of skills or options and learning new tasks by recombining them. In our work, we try to use information in the replay buffer by querying and searching for the relevant information across multiple trajectories which otherwise would take many replays through coincidentally relevant information for this to occur. **Efficient Credit assignment. Learning long term dependencies requires assigning credit to time-** steps far back in the past. Common methods for assigning credit in dynamics model like backpropagation through time requires information to be propagated backwards through every single step in the ----- past. This could become computationally expensive when used with very long sequences. Methods which try to get around this problem only back-propagate information through a selected time-steps in the past, realized by a learned mechanism that associates current state with relevant past states (Ke et al., 2018; Wayne et al., 2018; Arjona-Medina et al., 2018; Goyal et al., 2018; Fortunato et al., 2019). Most of these works consider assigning credit to states within the same trajectory, whereas the proposed model R2A, searches for the relevant information in the replay buffer which includes information from other trajectories also. A.2 INFORMATION THEORETIC FORMULATION Let’s denote the agent’s policy πθ(A _S, G), where S is the agent’s state, A the agent’s action,_ _|_ _G the information from the retrieval process conditioned on the current state of the agent i.e.,_ _G = fretrieval(S), and θ the parameters of the neural network representing the policy. We want to_ train agents that in addition to maximizing reward, minimize the policy dependence on the information from the retrieval process, quantified by the conditional mutual information I(A; G | S). This approach of minimizing the dependence of the policy on the information from the retrieval process can be interpreted as encouraging agents to learn useful behaviours and to follow those behaviours closely, except where diverting from doing so (as a result of using information from the retrieval process) leads to higher reward (Strouse et al., 2018; Teh et al., 2017; Goyal et al., 2019a; Galashov et al., 2019; Czarnecki et al., 2019). To see this, note that the conditional mutual information can also be written as I(A; G | S) = Eπθ [DKL[πθ(A | S, G) | π0(A | S)]] where Eπθ denotes an expectation over trajectories, π0(A | S) = _g_ _[p][(][g][)][ π][θ][(][A][ |][ S, g][)][. is a “default” policy with the]_ information from the retrieval process marginalized out. We maximize the following objective: _J(θ) ≡_ Eπθ [r] − _βI([P]A; G | S)_ (1) = Eπθ [r − _βDKL [πθ(A | S, G) | π0(A | S)]],_ where β > 0 is a tradeoff parameter, Eπθ denotes an expectation over trajectories (for ex. generated by the agent’s policy) and DKL refers to the Kuhlback-Leibler divergence. (Goyal et al., 2019a; Galashov et al., 2019) proposes to optimize the Eq. 1 by maximizing the lower bound _J[˜](θ)_ [3]: _J(θ) ≥_ _J[˜](θ) ≡_ Eπθ [r − _βDKL [penc(Z | S, G) | q(Z | S)]] ._ (2) We parameterize the policy πθ(A | S, G) using an encoder penc(Z | S, G), a decoder pdec(A | S, Z) and the q(Z | S) is the learned prior such that πθ(A | S, G) = _z_ _[p][enc][(][z][ |][ S, G][)][ p][dec][(][A][ |][ S, z][)][.][4]_ The encoder output Z is meant to represent the information from the retrieval process that the agent believes is important to access in the present state S in order to perform well. [P] Due to the data processing inequality (DPI) (Cover, 1999) I(Z; G | S) ≥ _I(A; G | S), and hence to_ obtain an upper bound on I(Z; G|S), we must first obtain an upper bound on I(Z; G|S = s), and then average over p(s). We get the following result: _I(Z; G|S) ≤_ _p(s)_ _p(g|s)DKL(p(Z|s, g)∥r(Z))_ (3) Such an information bottleneck has shown to improve generalization (Teh et al., 2017; Goyal et al., 2019a;c; Galashov et al., 2019; Merel et al., 2019; Liu et al., 2019; Goyal et al., 2020a; Liu et al., 2021). A.3 ATARI: IMPLEMENTATION DETAILS AND RAW SCORES FOR R2D2 AND RA-R2D2. Table 3 shows the raw scores achieved by RA-R2D2 and the R2D2 baseline. For R2D2 baseline and for the parameterization of the agent process, we follow the exact same training setup as in (Banino et al., 2021). Figure 7, 8, 9, shows the learning curve for both the proposed model and the R2D2 baseline for different games. 3We ask the reader to refer to Information Theoretic Formulation section in the appendix. 4For experiments, we estimate the the marginals and conditionals using a single sample throughout. ----- Training Losses: _• The parameters of the state encoder are trained by the RL loss and the auxiliary losses used_ to train the forward/backward summarizer. The state encoder encodes information about the past actions, past rewards and the current observation into an abstract state. _• The parameters of the retrieval process are trained by the RL loss and the information_ bottleneck regularizer. _• The parameters of the agent process are trained only by the RL loss._ Hyperparameters: _• At each learner step, the agent samples a large batch of 512 trajectories from the replay_ buffer. A fixed fraction of the batch (64 trajectories) is used for learning the Q-function and the remaining trajectories forms the retrieval batch. _• We sample a different retrieval dataset for each gradient update. The re-encoding of the_ trajectories is performed for each gradient update during training. _• The retrieval process selects the top-ktraj = 10 most relevant trajectories (in step 2, section_ 2.3), and then retrieves relevant information from the selected trajectories (step 3, section 2.3) using the top-kstates = 10 most relevant states (see 5). _• To summarize the experiences in the retrieval batch we use a forward and backward GRU_ with 512 hidden units. _• We use 8 slots for Atari-R2D2 experiments._ _• We use auxiliary losses in the form of action or policy prediction (section 2.2). We follow_ the same setup as proposed in (Schrittwieser et al., 2020). _• The value of the β coefficient for the information bottleneck regularizer is fixed to 0.1._ _• The rest of the hyperparameters for R2D2 are taken from Kapturowski et al. (2018) and are_ detailed in table 2 **Hyperparameter** **Value** Optimizer Adam Learning rate 0.0001 Q’s λ 0.8 Adam epsilon 10[−][7] Adam beta1 0.9 Adam beta2 0.999 Adam clip norm 40 Q-value transform _h(x) = sign(x)(_ _|x| + 1 −_ 1) + ϵx Discount factor 0.997 p Trace length (Atari) 80 Replay period (Atari) 40 Replay capacity 100000 sequences Replay priority exponent 0.9 Importance sampling exponent 0.6 Minimum sequences to start replay 5000 Target Q-network update period 400 Evaluation ϵ 0.01 Target ϵ 0.01 Table 2: Hyperparameters used in the Atari R2D2 experiments. ----- Implementation details: _• We use GRU style gating and layernorm whenever we are changing the state of the agent_ process or the state of the slots in a residual way (Step 5, 6 in Algorithm 1). _• If the input to a neural network fnn consists of two tensors x, y, we never concatenate the_ inputs to the neural network. We make use of attention. We assume one of the inputs is the primary input (lets say x), and other input is privileged input (as a result of doing some expensive computation, let’s say y). First, we compute fnn(x), and then using the output of this computation, we cross-attend over y using multi-head attention. _• When computing the forward summary of the trajectories in the retrieval batch h = ht + st,_ we also add the information about the current encoded state i.e., st. **Visual Encoder Visual observations are encoded using a ResNet-47 encoder. The 47 layers are** divided in 4 groups which have the following characteristics: _• An initial stride-2 convolution with filter size 3x3 (1 · 4 layers)._ _• Number of residual bottleneck blocks (in order): (2, 4, 6, 2). Each block has 3 convolutional_ layers with ReLU activations, with filter sizes 1x1, 3x3, and 1x1 respectively ((2+4+6+2)·3 layers). _• Number of channels for the last convolution in each block: (64, 128, 256, 512)._ _• Number of channels for the non-last convolutions in each block: (16, 32, 64, 128)._ _• Group norm is applied after each group, with a group size of 8._ After this observation encoding step, a final layer with ReLU activations of sizes 512 is applied. **Computational Complexity. Here we discuss the computational complexity of the R2A.** The following computations are happening at each step within a trajectory: _• A particular slot selects the relevant trajectories and relevant states at each time-step._ _• We use an attention mechanism to select the set of relevant trajectories and relevant states._ This process scales with the number of trajectories in the retrieval batch. _• The set of relevant trajectories and relevant states changes across different time-steps within_ a trajectories. This process is repeated for all the slots independently, but can be efficiently performed on GPUs/TPUs. We can make following optimizations to reduce the computational complexity of the R2A. _• For Atari and Gridroboman experiments instead of using a learned attention mechanism to_ rank the trajectories (in step 2, section 2.3), we can also just use those trajectories which are of high return (Abdolmaleki et al., 2018) since we already know that the data in the retrieval dataset is specific to the task on which we are training the agent. _• Once we have selected relevant trajectories, we can further reduce the computational cost of_ selecting most relevant states (in the selected trajectories) by only considering those states which are semantically similar to the agent’s current state. Improvements tried that seem to improve the performance but were not evaluated exhaustively: _• Instead of performing only single step of retrieval at each step, we can also query the replay_ buffer multiple times within a time-step at the cost of increased computational complexity. _• We can allow different slots to retrieve information at different time scales._ We note that all of the hyperparameters for the retrieval process, except for the number of trajectories in the retrieval batch, remain the same across all experiments (Atari, gridroboman, and BabyAI). ----- A.3.1 RETRIEVAL PROCESS: RETRIEVING INFORMATION FROM THE RETRIEVAL BUFFER. Here we discuss the different steps involved in retrieving information from the retrieval buffer: **Query computation. Each slot independently computes its prestate using the contextual information** from the agent using a GRU. Then, each slot independently computes a retrieval query which will be matched against information in the retrieval buffer. Figure 5: Retrieval of most relevant trajectories and states. A) The retrieval process selects the top-ktraj most relevant trajectories as specified in step 2, section 2.3. In the figure scalars represent attention weights. B) Then the retrieval process retrieves relevant information from the selected trajectories by selecting the most relevant states from the top-k trajectories as detailed in step 3, section 2.3 **Identification of most relevant trajectories and states for each slot. The retrieval mechanism** process uses an attention mechanism to match a query produced by the the retrieval state associated with each slot to keys computed on each time step of each trajectory of the retrieval batch. This process assigns a single scalar to each state within a trajectory. These scores are used to assign a scalar to each trajectory and then normalized across trajectories. These normalized scores are then used by the retrieval process to select the top-ktraj most relevant trajectories (see Figure. 5). The process is then again repeated to select most relevant states with the selected trajectories. Here, a single scalar is assigned to each state with in selected trajectories and then normalized across all the states. These scores are then used to select most relevant states within the most relevant trajectories. Figure 6: Each trajectory in the retrieval batch is summarized via a forward and a backward running parameteric structured model (Transformers or slot based recurrent network). Each state in the trajectory is represented by a (ht, bt) tuple where ht represents the summary of the past and bt represents the summary of the future with that trajectory. **Information retrieval from the most relevant states. In the previous step, we selected most relevant** states within the most relevant trajectories. Now we integrate information across the selected states. We again use the attention mechanism. Here the query by the slots is matched against the keys (which are the function of the forward running dynamics model), to retrieve information about the future (and values are the function of the backward running dynamics model). See figure. 6 for more details. ----- Environment R2D2 RA-R2D2 Alien 52268 ± 4125 **81626 ± 5123** Amidar 10976 ± 2341 **24693 ± 1982** Assault **45521 ± 4751** 43497 ± 6653 Asterix 997910 ± 51241 997585 ± 89251 Asteroids **262560 ± 94168** 254194 ± 72141 Atlantis 1424723 ± 67589 **1516463 ± 98140** BankHeist 3923 ± 512 **19397 ± 5931** BattleZone **590913 ± 128715** 466783 ± 89724 BeamRider **89038 ±** 79241 ± Bowling 250.37 ± 253.5 ± Boxing 99 ± 0 99 ± 0 Breakout 848 ± 121 **869 ± 51** Centipede 315329 ± 151612 **497746 ± 12515** ChopperCommand 998842 ± 11451 **999936 ± 3422** CrazyClimber 184132 ± 12912 **226469 ± 9251** Defender 538752 ± 18241 **554158 ± 21415** DemonAttack 143552 ± 0 143519 ± 0 DoubleDunk 24 ± 0 24 ± 0 Enduro 2332 ± 32 2359 ± 19 FishingDerby 71.357 ± 2.51 73.13 ± 3.12 Freeway 34 ± 0 33.19 ± 0 Frostbite 168225 ± 51891 **511159 ± 124512** Gopher 123138 ± 4121 123841 ± 8914 Gravitar 13142 ± 1214 13198 ± 1102 Hero **43257 ± 4121** 35431 ± 2412 IceHockey 72.74 ± 31 71.59 ± 12 Jamesbond 24873 ± 2141 25873 ± 2817 Kangaroo 14614 15232 Krull 158509 ± 21415 **183921 ± 31415** KungFuMaster 208583 ± 2102 **224385 ± 41512** MontezumaRevenge **1966.7 ± 987** 1750 ± 1800 MsPacman 33530 ± 1214 **44396 ± 4521** NameThisGame 30232 ± 3151 **45140 ± 2412** Phoenix 695251 ± 32515 **847914 ± 12415** Pitfall 0.0 ± 0 0.0 ± 0 Pong 21 ± 0 21 ± 0 PrivateEye **21602 ± 4124** 18923 ± 2415 Qbert 242169 ± 51512 241525 ± 32513 Riverraid 32441 ± 5314 **44554± 7325** RoadRunner 490872 ± 51512 **513521 ± 65132** Robotank **128.44 ± 18** 102 ± 29 Seaquest 998664 ± 1212 **999899 ± 51** Skiing -29973 ± 1212 -29232 ± 2412 Solaris 3924 ± 3415 3853 ± 1241 SpaceInvaders 57198 ± 14125 **62624 ± 4215** StarGunner 256129 ± **478231 ±** Surround 9.52 ± 0 10 ± 0 Tennis **7.16 ± 0** 0 ± 0 TimePilot 168592 ± 21241 **260609 ± 41212** Tutankham **390 ± 41** 235 ± 89 UpNDown 544439 ± 5851 **561389 ± 21516** Venture 2000 ± 0 2000 ± 0 VideoPinball 673679 ± 234151 962119 ± 24151 WizardOfWor 78431 ± 5159 73500 ± 6231 Yars Revenge **674692 ± 125116** 513549 ± 89151 Zaxxon 53312 ± 15152 **62631 ± 5161** Table 3: Scores obtained on different atari games. (average over 3 different seeds). ----- A.3.2 DETAILS ON ABLATIONS _Ablation A-1: In this ablation, the internal retrieval state, mt is removed. The query qt is computed_ from st directly as qt = fquery(st) . The retrieval vector gt is computed as per steps 2-4 of the original algorithm. Step 5 does not occur, and in step 6 the only update is st ← **_st + gt._** _Ablation A-2: In this ablation, there is no retrieval batch. The retrieval process is parametrized as_ before with a nk-slot memory. Step 1 is identical, but steps 2-4 do not occur. Memory slots are updated as per the joint update of step 5. Step 6 is otherwise identical, except that keys and values are computed from the memory slots m[k]t [instead of the retrieved values][ z]t[k][.] _Ablation A-3: In this ablation, the length of the trajectories used for summarization is reduced from_ 80 to 5. The trajectories used for Q learning are unchanged. _Ablations A-4 and A-5: In our experiments, we use standard policy, reward and value prediction_ cross-entropy losses (Schrittwieser et al., 2020) to train the forward and backward summarizers. In A-4, we add a BERT loss; tokens are obtained by compressing the agent states st with VQ-VAE. In A-5, we only use the BERT loss. A.3.3 ADDITIONAL ATARI ABLATIONS Here we perform ablations on RA-R2D2 using the 5 Atari games on which RA-R2D2 performs worst relative to baseline R2D2. Figure. 10 shows the relative performance of different ablations compared to the R2D2 baseline. Ablations 1-5 are as described in the main text (section 3.1.1). For these results, we also ran a sixth ablation. Figure 10: Relative performance of ablated RA-R2D2 versus baseline R2D2 for 6 ablations on 5 Atari games on which RA-R2D2 performs worse than baseline R2D2. **Role of ktraj and kstates (A-6). In our experiments, the retrieval process selects the top ktraj = 10** most relevant trajectories (step 2, section 2.3) and then selects the top kstates = 10 most relevant states of these trajectories (step 3, section 2.3) from which to retrieve relevant information. To better understand the role of these hyperparameters, we independently vary these two hyper-parameters (top ktraj and top kstates) over the values {5, 10, 20}. Figure 10 shows that after independently varying these two hyper-parameters the performance of the R2A can be improved as compared to the R2D2 baseline. ----- (a) Alien (b) Amidar (c) Assault (d) Asterisk (e) Asteroids (f) Atlantis (g) Bank Heist (h) BattleZone (i) Berzerk (j) Boxing (k) Breakout (l) Centipede (m) ChopperCommand (n) CrazyClimber (o) Defender (p) DemonAttack (q) DoubleDunk (r) Enduro (s) FishingDerby (t) Freeway (u) Frostbite (v) Gopher (w) Gravitar (x) HERO Fi 7 C **i** **R2A ith R2D2 L** i f i l d R2D2 (RA R2D2) d h R2D2 b li diff ----- (a) NameThisGame (b) Phoenix (c) Pitfall (d) Pong (e) RiverRaid (f) RoadRunner (g) RoboTank (h) Seaquest (i) Skiing (j) Solaris (k) Surround (l) Tennis (m) TimePilot (n) Tutankham (o) UpNDown (p) Venture (q) Video (r) VideoPinBall (s) WizardOfWar (t) YarsRevenge (u) Zaxxon Figure 8: Comparing R2Awith R2D2: Learning curves for retrieval-augmented R2D2 (RA-R2D2) and the R2D2 baseline across different Atari games. Green is RA-R2D2 and orange is baseline R2D2. ----- (a) IceHockey (b) JamesBond (c) Krull (d) KungFuMaster (e) MontezumaRevenge (f) MsPacMan Figure 9: Comparing R2Awith R2D2: Learning curves for retrieval-augmented R2D2 (RA-R2D2) and the R2D2 baseline across different Atari games. Green is RA-R2D2 and orange is baseline R2D2. ----- A.4 GRIDROBOMAN ENVIRONMENT Figure 11: Gridroboman environment illustration. On the board there are three colored objects and the robot is represented by a black block. The robot can move itself and move the objects. The tasks are motivated by the robotic manipulation. In order to test our method in a multi-task setting we designed a minimalistic grid world robotic manipulation (gridroboman) environment with a single embodiment and multiple tasks. It is implemented based on the pycolab (Stepleton et al.) game engine that provides tools for designing customizable grid world games. The environment and its tasks are inspired by the challenges in robotic manipulation. An illustration of a state in the environment is shown in Figure 4. **Environment semantics, observations and actions** There are three colored objects on the 7 × 7 board: red, green and blue. The black block represents the robot. It can move in 4 directions and the action is skipped if it attempts to move into a wall (gray). Additionally, if the robot is located at the same position as any colored object, it can apply lift action that would enforce the object to move together with the robot. Then, put action allows to position the object either on the board or on the top of another single object. Additionally, there is an option of skipping an action. The initial state of the environment that includes positions of the objects and the robot is sampled randomly at the beginning of each episode. The agent observation is an 11-dimensional vector: it includes x and y coordinates of three objects and the robot as well as status of each of three objects. The status of an object is a numerical value: 0 if the object is positioned on the board, −1 if it is under another object and +1 if it is either held by the robot or located on the top of another object. Each episode lasts for 50 timesteps. **Tasks** There are 30 possible tasks: lift red, green, or blue object (3), touch red, green, or blue object (3), move red, green, or blue to center (3) or to corner (3), touch one object with another (6), move two objects close to (3) or far from (3) each other, stack one object on the top of another (6). The nature of these tasks is such that it is impossible to identify the task by the initial state. **Rewards** The reward is binary in each time step: it is zero when the task is not solved and one otherwise. In order to receive positive reward the following condition should be met: _• lift X: the robot should lift and not put back an object of color X;_ _• touch X: the robot should be located at the immediately adjacent cell to the object of X_ color and not hold any objects; _• move X to center: the object or color X should be located in the 3 × 3 block at the center_ of the board; _• move X to corner: the object of X color should be located in any of the 2 × 2 corners of_ the board; _• touch X with Y: the robot should be located at the immediately adjacent position to the_ object of X color and hold the object of color Y; _• move X close to Y: the distance between the objects of colors X and Y should be no more_ than 1 in both x and y direction; ----- _• move X far from Y: the sum of distances between objects of colors X and Y in x and y_ coordinated should be greater than 9; _• stack X on Y: the objects of colors X and Y should be located at the same position one on_ the top of another. A.4.1 GRIDROBOMAN EXPERIMENT SETUP The gridroboman offline RL dataset was generated by training a single DQN agent online on each task separately and recording the 100K generated episodes. The Q(s, a) network was a 3-layer MLP with 256 units in each hidden layer for Q-function. For the experiments with 10, 20, and 30 tasks, we use the first 10, 20, and 30 tasks, as listed here. **10 tasks: touch red, touch green, touch blue, lift red, lift green, lift blue, red touch green, green touch** red, and green touch blue. **20 tasks: The above 10 tasks, blue touch red, blue touch green, red to corner, green to corner, blue to** corner, red to center, green to center, blue to center, red close to blue, and red close to green. **30 tasks: The above 20 tasks, blue close to green, red far from blue, red far from green, blue far from** green, red on blue, red on green, green on red, green on blue, blue on red, and blue on green. A.4.2 GRIDROBOMAN HYPERPARAMETERS The DQN agent used the same network as used to create the data: a 3-layer MLP with 256 hidden units per layer. The first 2 layers define fθ[enc] and the final layer predicts Q. The RA-DQN agent used the same base network plus a separate retrieval network. Below we detail hyperparameters specific to gridroboman. Hyperparameters not specified in this section are the same as used in Atari. _• At each learner step, the agent samples a batch of 256 states from the replay buffer to train_ the DQN agent and further samples a batch of 64 trajectories from the retrieval dataset to form the retrieval batch. _• The retrieval process selects the top-ktraj = 10 trajectories with the highest return, and then_ retrieves relevant information from the selected trajectories (step 3, section 2.3) using the top-kstates = 10 most relevant states. _• To summarize the experiences in the retrieval batch we use a forward and backward GRU_ with 256 hidden units. _• We use 4 slots to parameterize the retrieval process._ _• We use auxiliary losses in the form of value, reward, and policy prediction (section 2.2). We_ follow the same setup as proposed in (Schrittwieser et al., 2020) and weight these auxiliary losses with a coefficient of 0.1. _• The value of the β coefficient for the information bottleneck regularizer is fixed to 0.3._ _• The hyperparameters and setup for offline DQN are taken from Gulcehre et al. (2020). They_ are detailed in table 5. As in the above, we used double DQN (Van Hasselt et al., 2016). **Hyperparameter** **Value** Optimizer Adam Learning rate 3e-4 Discount factor 0.99 Importance sampling exponent 0.2 Minimum sequences to start replay 5000 TD loss function Huber(1.0) Target Q-network update period 2500 Evaluation ϵ 6.5e-4 Table 4: Hyperparameters used in the gridroboman DQN experiments. ----- A.4.3 GRIDROBOMAN EVALUATION CURVES WHEN TRAINING ON ALL 30 TASKS Figure 12 we show the average episode reward obtained by the evaluation agent on each individual task when training on all 30 tasks. Curves are shown for the DQN and RA-DQN agents. (a) touch red (b) touch blue (c) touch green (d) lift red (e) lift blue (f) lift green (g) red touch blue (h) red touch green (i) green touch red (j) green touch blue (k) blue touch red (l) blue touch green (m) red to corner (n) green to corner (o) blue to corner (p) red to center (q) green to center (r) blue to center (s) red close to blue (t) red close to green (u) blue close to green (v) red far from blue (w) red far from green (x) blue far from green (y) stack red on blue (z) stack red on green (aa) stack green on red (ab) stack green on blue (ac) stack blue on red (ad) stack blue on green Figure 12: Multi-task offline RL with a task-specific retrieval dataset. Evaluation performance for RA-DQN (orange) and baseline DQN (blue) when training on all 30 gridroboman tasks with a single agent. Curves show the performance of each agent (averaged over 3 seeds) when running that agent online in the environment on the specified task. A.5 BABYAI ENVIRONMENT. **Observations, actions. We use the same setup for the observations and actions as in (Chevalier-** Boisvert et al., 2018). **Tasks. We use all the 40 available tasks in the BabyAI environment.** [5]. [5https://github.com/mila-iqia/babyai/blob/master/docs/bonus_levels.md.](https://github.com/mila-iqia/babyai/blob/master/docs/bonus_levels.md) [https://github.com/mila-iqia/babyai/blob/master/docs/iclr19_levels.md](https://github.com/mila-iqia/babyai/blob/master/docs/iclr19_levels.md) ----- **Hyper-parameters and RL algorithm. We follow the same setup as in the Gridroboman experi-** ments. Since BabyAI is a partially observable environment, we summarize the history of the agent using a recurrent encoder. **Hyperparameter** **Value** Optimizer Adam Learning rate 3e-4 Discount factor 0.99 Importance sampling exponent 0.2 TD loss function Huber(1.0) Target Q-network update period 1000 GRU hidden state 512 Table 5: Hyper-parameters used in the BabyAI Recurrent DQN experiments. **Retrieval Process. We use the same hyper-parameters for the retrieval process as used in the grid-** roboman enviornment. For the case of multi-task retrieval buffer, we use 32 trajectories corresponding to each task to form the retrieval batch. A.6 LIMITATIONS AND FUTURE WORK. It would be useful to investigate and extend the proposed idea in these different ways: (a) First, investigate training of the retrieval process and the agent process using different objectives as compared to training them in an end-to-end fashion, (b) Second, scaling R2A to more complex problems multi-agent problems like in Starcraft (Vinyals et al., 2019), where one may have access to millions of past experiences. In R2A, we only query a subset of the retrieval dataset, which limits the generality of the method. (c) Third, even more intriguing would be the possibility of learning an abstract model with abstract internal actions, and rewards, rather than learning a model which queries for information from the retrieval dataset, and hence avoiding the need for Monte Carlo tree search common in the state of the art planning methods (Schrittwieser et al., 2020). We aim to formulate these problems and seek answers in the future work. A.7 SOCIAL IMPACT. The authors do not foresee negative social impact of this work beyond that which could arise from general improvements in ML. A.8 REBUTTAL PHASE. **Comparison to Episodic Control baselines.** We have performed ablations where we compared to the scenario where we just keep the simple episodic memory (refer to section. 3.1.1, ablation A-1). The result of ablations shows that for R2D2 accessing the dataset in a non-parametric way actually hurts the performance of the RL algorithm. **BERT style unsupervised losses then not used generally in the R2A method** BERT style unsupervised losses actually improve the learned representations. We did not choose to use BERT style representations as they increase the number of design choices, increase the number of ablations, increase the number of parameters, whereas the relative gains of using BERT style unsupervised losses is not much. Future work should investigate the use of pre-trained encoders as common in NLP literature. **For Frostbite, episodic control is beneficial.** As mentioned in “Building Machines That Learn and Think Like People”, in Frostbite, players control an agent (Frostbite Bailey) tasked with constructing an igloo within a time limit. The igloo is built piece-by-piece as the agent jumps on ice floes in water The challenge is that the ice floes are in constant motion (moving either left or right), and ice floes only contribute to the construction of the igloo if they are visited in an active state (white rather than blue). The agent may also earn extra points by gathering fish while avoiding a number of fatal hazards (falling in the water, snow geese, polar bears, etc.). Success in this game requires a ----- temporally extended plan to ensure the agent can accomplish a sub-goal (such as reaching an icefloe) and then safely proceed to the next sub-goal. Ultimately, once all of the pieces of the igloo are in place, the agent must proceed to the igloo and thus complete the level before time expires. One hypothesis for the better performance could be that the retrieval process is able to efficient utilize the information from the states which are very far apart from the current state (i.e., temporally extended credit assignment). This hypothesis is further validated when we decrease the length of the traces in the dataset. In ablation section, if we decrease the length of the trajectories that are retrieved and summarized during retrieval preprocessing, thus reducing the amount of past and future information the retrieval process can retrieve. By default, the trajectories in the retrieval dataset are of length 80. To perform this ablation, we decrease the length of the effective context to only include information from 5 time-steps. **Similar Computation baseline.** We also compared to a baseline which does the same amount of “computations” (Importance of retrieving information (A-2)). We examine what happens when the retrieval process does not have access to the retrieval dataset and hence no information is retrieved, keeping all else the same. In this ablation, the retrieval process updates the state of the slots using a transformer (i.e., in Step 1 we replace GRU with a transformer), and the updated state of the transformer is used by the agent process to shape the representation of its value function. Since the use of the transformer number of parameters in this baseline is more as compared to R2D2. Our results show that the proposed method R2A archives better performance as compared to vanilla R2D2 as well as the baseline which does the same amount of computation (but does not retrieve information). **How does runtime compare to vanilla R2D2? How many more parameters are introduced?** As compared to vanilla R2D2, a naive implementation of the proposed method is around 3x slow (and somewhat optimized implementation around 2x slower), with the increase of 10% more parameters. In the ablations, we also examine what happens when the retrieval process does not have access to the retrieval dataset and hence no information is retrieved but everything else is kept the same (same computation and same parameter baseline). **Use of hierarchical attention (top-k trajectories and then top-k states).** A hierarchical attention where we first filter trajectories, and then filter states reduces the search space for the attention. For example. consider 512 trajectories, and 80 states in each trajectory, then the attention operator has to select 10 most relevant states out of 40960 states. Whereas if we do initial filtering of top-k trajectories, then the attention operator needs to select 10 most relevant trajectories out of 512 trajectories, and then 10 most relevant states out of 800 states. Such an hierarchical attention has also shown to be successful in various NLP tasks also. **Analysis about retrieved information in the multi-task setting?** We did some analysis to study the properties about the retrieved information in the multi-task setting in BabyAI. BabyAI setup contains about 40 tasks. Out of these 40 tasks, around 15 tasks are compositional i.e., requires to compose information from 2 or more sub-tasks, and rest of the tasks requires the agent to execute a particular behaviour (ex. going to the door, fetching a key etc). We study as to how often the agent retrieves information from other tasks. Ideally, for the compositional tasks, the hope would be the agent access information about the other tasks more often as compared to the tasks which require only a particular behaviour. So, during test time, we study the percentage of times agent access information from the same task as compared to accessing information from the different tasks. For compositional tasks, the agent access information from other tasks about 54% of times, while the percentage for single tasks is about 21%. That said, we were not able to find any pattern as to find why the agent is retrieving more information from other tasks for some particular tasks as well as any structure about the states from which information is being retrieved. -----