# fewshot_task_learning_through_inverse_generative_modeling__74a31754.pdf Few-Shot Task Learning through Inverse Generative Modeling Aviv Netanyahu1 , Yilun Du1,2, Antonia Bronars1, Jyothish Pari1, Joshua Tenenbaum1, Tianmin Shu3, and Pulkit Agrawal1 1Massachusetts Institute of Technology 2Harvard University 3Johns Hopkins University Learning the intents of an agent, defined by its goals or motion style, is often extremely challenging from just a few examples. We refer to this problem as task concept learning and present our approach, Few-Shot Task Learning through Inverse Generative Modeling (FTL-IGM), which learns new task concepts by leveraging invertible neural generative models. The core idea is to pretrain a generative model on a set of basic concepts and their demonstrations. Then, given a few demonstrations of a new concept (such as a new goal or a new action), our method learns the underlying concepts through backpropagation without updating the model weights, thanks to the invertibility of the generative model. We evaluate our method in five domains object rearrangement, goal-oriented navigation, motion caption of human actions, autonomous driving, and real-world table-top manipulation. Our experimental results demonstrate that via the pretrained generative model, we successfully learn novel concepts and generate agent plans or motion corresponding to these concepts in (1) unseen environments and (2) in composition with training concepts. 1 Introduction The ability to learn concepts about a novel task, such as the goal and motion plans, from a few demonstrations is a crucial building block for intelligent agents it allows an agent to learn to perform new tasks from other agents (including humans) from little data. Humans, even from a young age, can learn various new tasks from little data and generalize what they learned to perform these tasks in new situations [1]. In machine learning and robotics, this class of problems is referred to as Few-Shot Learning [2]. Despite being a widely studied problem, it remains unclear how we can enable machine learning models to learn concepts of a novel task from only a few demonstrations and generalize the concepts to new situations, just like humans do. Common approaches learn policies either directly, which often suffer from covariate shift [3], or via rewards [4 6], which are largely limited to previously seen behavior [7]. In a different vein, other work has relied on pretraining on task families and assumes that task learning corresponds to learning similar tasks to ones already seen in the task family [8, 9]. Inspired by the success of generative modeling in few-shot visual concept learning [10 12], where concepts are latent representations, in this work, we investigate whether and how few-shot task concept learning can benefit from generative modeling as well. Learning concepts from sequential demonstrations rather than images is by nature more challenging due to sequential data often not Correspondence to Aviv Netanyahu . Project website https://avivne.github.io/ftl-igm. 38th Conference on Neural Information Processing Systems (Neur IPS 2024). Pretraining given trajectories and concepts learn generative model 饾摉饾溄 Few-Shot Concept Learning given a few demonstrations of a new concept + frozen model 饾摉饾溄 learn new concept forward pass back propagation Figure 1: Few-shot concept learning. Given paired task demonstration 蟿 (e.g., walk ) and concept c (a latent representation of the task), we train a generative model G胃 to generate behavior from a concept. Then, given demonstrations of a new behavior 蟿 (e.g., jumping jacks ) without its concept label, we aim to learn its concept representation by optimizing concept c as input to frozen G胃. Mo Cap Object Rearrangement Navigation to Goal Driving Table-Top Manipulation Figure 2: Experiment Domains. We extensively evaluate our approach for various domains. satisfying the i.i.d. assumption in machine learning [13]. In particular, we assume access to a large pretraining dataset of paired behaviors and task representations to learn a conditional generative model that synthesizes trajectories conditioned on task descriptions. We hypothesize that by learning a generative model conditioned on explicit representations of behavior, we can acquire strong priors about the nature of behaviors in these domains, enabling us to more effectively learn new behavior that is not within the pretraining distribution, given a limited number of demonstrations, and further generate the learned behavior in new settings. To this end, we propose Few-Shot Task Learning through Inverse Generative Modeling (FTL-IGM). In our approach, we first pretrain a large conditional generative model which synthesizes different trajectories conditioned on different task descriptions. To learn new tasks from a limited number of demonstrations, we then formulate few-shot task learning as an inverse generative modeling problem, where we find the latent task description, which we refer to as a concept, which maximizes the likelihood of generating the demonstrations. This approach allows us to leverage the powerful task priors learned by the generative model to learn the shared concepts between demonstrations without finetuning the model (Figure 1). We demonstrate this approach in various domains: object rearrangement, where concepts are relations between objects, goal-oriented navigation, where concepts are target attributes, motion capture, where concepts are human actions, autonomous driving, where concepts are driving scenarios, and real-world table-top manipulation where concepts are manipulation tasks (Figure 2). New concepts are either (1) compositions of training concepts (e.g., multiple desired relations between objects that define a new object rearrangement concept) or (2) new concepts that are not explicit compositions in the natural language symbolic space of training concepts (e.g., a new human motion jumping jacks is not an explicit composition of training concepts walk , golf etc.) Thanks to generative models compositional properties that enable compositional concept learning [14], in addition to being able to learn a single concept from demonstrations directly, FTL-IGM learns compositions of concepts from demonstrations that, when combined, describe the new concept. We show that our approach generates diverse trajectories encapsulating the learned concept. We achieve this due to two properties of generative models. First, these models have shown strong interpolation abilities [15, 16], which allow generating the new concept on new initial states they were not demonstrated from. Second, these models have compositional properties that enable compositional trajectory generation [17], which allow composing learned concepts with training concepts to synthesize novel behavior that was not demonstrated (e.g., jumping jacks and walk ), see Figure 3. We further demonstrate that our approach addresses a unique challenge introduced in learning task concepts: we utilize plans generated by learned concepts in a closed-loop fashion. Our main contributions are (1) formulating the problem of task learning from few demonstrations as Few-Shot Task Learning through Inverse Generative Modeling (FTL-IGM), (2) adapting a method for efficient concept learning to this problem based on the new formulation, and (3) a systematic evaluation revealing the ability of our method to learn new concepts across a diverse set of domains. 2 Related Work Learning from few demonstrations. Our problem setting is closely related to learning from few demonstrations. There has been much work on learning to generate agent behavior given few demonstrations. There are several common approaches to this problem. First, behavior cloning (BC) is a supervised learning method to learn a policy from demonstrations that predicts actions from states. Similar to our framework, goal-conditioned BC can predict states from task representations and states [18]. Finetuning these models to learn new behaviors requires labeled demonstrations of the new task. We assume unlabeled demonstrations. BC often suffers from covariate shift [3] and fails to generate the demonstrated behavior in novel scenarios. This can be mitigated by assuming access to a human in the loop [19]. Second, the inverse reinforcement learning (IRL) framework learns a policy that maximizes the return of an explicitly [5, 6, 20] or implicitly [21] learned reward function. These works learn a reward for a single task or for a set of tasks (e.g., goal-conditioned IRL [22, 23] and multi-task IRL [24]). While IRL is more data efficient than BC, it is computationally costly due to learning a policy every iteration via an inner reinforcement learning (RL) loop. Additionally, it requires access to taking actions in the environment during training and when faced with a new task, we have to retrain the reward again. A third approach is inverse planning [25 27], which can robustly infer concepts such as goals and beliefs even in unseen scenarios. However, it assumes access to a planner, knowledge about environment dynamics, and the task/goal space. Finally, in-context learning approaches [8, 28, 29] learn actions in a supervised manner by representing the task with demonstrations. This allows few-shot generalization without further training. In contrast, we do not learn an action-generating policy directly or via a reward function. For concept learning, we do not assume having access to any given planner, world model, actions, rewards, or prior over the task space. Instead, we learn concepts (task representations) from demonstrations via a pretrained generative model that takes a concept as input and directly produces state sequences. We then input the learned concept into the generative model to produce behavior similar yet diverse to the demonstrated one. We further demonstrate how to use these state sequences with a planner to take actions and achieve the desired behavior. The idea of concept learning via generative models has been explored for computer vision applications [11, 12]. We build on this work and show how to extend it to learn agent task concepts. Our work also differs from prior works on learning trajectory representations [30 34]. These works focus on learning plans over trajectory embeddings, whereas we learn a task representation from demonstrations on which we condition to generate behavior. Generative Models in Decision Making. There has been work on generative modeling for decisionmaking, including generative models for single-agent behaviors, such as implicit BC [35], Diffuser [36, 17], Diffusion Policy [37], Decision Transformer [38 40], and for multi-agent motion prediction such as Jiang et al. [41]. The success of diffusion policy in predicting sequences of future actions has led to 3D extensions [42], and combined with ongoing robotic data collection efforts [43] and advanced vision and language models, has led to vision-language-action generative models [44 46]. In this work, we utilize a conditional generative model for the inverse problem, i.e., learning concepts from demonstrations. Composable representations. There has been work on obtaining composable data representations. 尾-VAE [47] learns unsupervised disentangled representations for images. MONet [48] and IODINE [49] decompose visual scenes via segmentation masks and COMET [50] and [12] via energy functions. There is also work on composing representations to generate data with composed concepts. Generative models can be composed together to generate visual concepts [14, 51 55] and robotic skills [17]. The generative process can also be altered to generate compositions of visual [56 59] and molecular [60] concepts. We aim to obtain task concepts and generate them in composition with other task concepts. 3 Formulation Inspired by recent success in large generative models, we propose a generative formulation for learning specific behavior given a small set of demonstrations, which we term Few-Shot Task Learning through Few-Shot Concept Learning learn a new concept Learned Concept Generation generate trajectories from the learned model and concept from new initial states in composition with training concepts Figure 3: Diverse learned concept generation. We generate versions of the new behavior conditioned on the learned concept and (1) new initial states and (2) composed with other concepts. Inverse Generative Modeling (FTL-IGM). In our formulation, we assume access to a large pretraining dataset Dpretrain = {(蟿i, ci)}N i=1 of state-based sequences 蟿i = {s0, s1, ...} T of states from state space S annotated with meta-data concepts ci C Rn describing trajectories. This assumption is often not prohibitive in practice. There is typically a vast amount of existing data collected from the internet or prior exploration in an environment, which may only need to be weakly annotated to characterize the trajectory, e.g., the goal state. Given Dpretrain, we learn a conditional generative model G胃 : C S T conditioned on concepts and initial states, which learns to generate future trajectories. We train the parameters of G胃 to maximize likelihood arg max胃 E蟿,c Dpretrain[log G胃(蟿|c, s0)]. Then, given an unlabeled demonstration dataset Dnew D, we formulate learning a new concept c that is used to sample trajectories from D as inverting the generative model. In particular, we learn new concept c so that our frozen conditional generative model G胃 maximizes the likelihood of trajectories in Dnew, corresponding to arg max c E蟿 Dnew[log G胃(蟿| c, s0)]. We find that this design choice enables us to leverage the priors learned by G胃 from Dpretrain to effectively learn concepts from Dnew given very few demonstrations, even if the demonstrated Dnew deviates from the concept labels c seen in Dpretrain. For evaluation in closed loop, we further assume access to a planner that given two states plans which action to take in the environment, sometimes via access to simulation in the environment. We use this planner sequentially to make decisions in the environment. The key difference between our approach to few-shot adaptation from demonstrations and prior approaches is the assumption and usage of a large pretraining dataset of paired behaviors and concepts Dpretrain combined with an invertible generative model. We learn new concepts solely from demonstrations without finetuning model weights or taking actions in the environment by relying on the pretrained concept space. 4 Few-Shot Concept Learning Based on FTL-IGM We adapt a few-shot concept learning method to task concepts based on the FTL-IGM framework. During training we learn a generative model G胃 from training {(蟿i, ci)}i pairs. We then freeze G胃, and given demonstrations of a new task { 蟿}i, optimize a concept c to produce the new behavior via G胃. We then generate a diverse set of behaviors via G胃, either for the learned concept c conditioned on new initial states or for compositions of c with other concepts. 4.1 Training a diffusion model to generate behavior A diffusion model is a generative model that given a forward noise adding process q(xt|xt 1) := N(xt; 1 尾txt 1, 尾t I) starting from data x0 according to a variance schedule 尾1, ..., 尾T , learns the reverse process p胃(xt 1|xt) := N(xt 1; 碌胃(xt, t), 危胃(xt, t)). Ho et al. [61] simplify the training objective to estimate noise Et U{1,T },x0,系 N(0,I) ||系 系胃(xt, t)||2 where xt is produced by adding noise 系 to data x0 by the forward noising process at diffusion step t, q(xt|x0) := N(xt; 伪tx0, (1 伪t)I) where 伪t := 螤t s=1(1 尾s). Dhariwal and Nichol [62] enable conditioned generation by guiding the reverse noising process with classifier gradients. The noise prediction becomes 藛系 = 系胃(xt, t) 蠅 1 伪t xt log p蠒(y|xt, t) where classifier p蠒(y|xt, t) is trained on noisy images, and 蠅 is the guidance scale. Ho and Salimans [63] introduce classifier-free guidance that achieves the same objective without the need for training a separate classifier. This is done by learning a conditional and unconditional model by removing the conditioning information with dropout during training. The noise prediction is then 藛系 = 系胃(xt, t) + 蠅(系胃(xt, y, t) 系胃(xt, t)). Ramesh et al. [64] and Nichol et al. [65] demonstrate how this idea can be used to generate images conditioned on a class. Diffusion models have recently shown success as generative models for decision making [36, 17]. Specifically, Ajay et al. [17] used a conditional classifier-free guidance diffusion model [63] to generate trajectories of future states to reach given an input observation. We adopt this objective and learn a denoising model 系胃 conditioned on latent concepts and initial observed states to estimate noise of a future state trajectory: E(蟿,c) Dpretrain,系 N(0,I),t U{1,T },纬 Bern(p)[||系 系胃(xt(蟿), (1 纬)c + 纬c , s0, t)||2] (1) where p is the probability of removing conditioning information which is then replaced by dummy condition c , and s0 is the initial state corresponding to trajectory 蟿. xt(蟿) is obtained from x0 = 蟿 by the forward noising process. We then extend this approach for the inverse problem, namely, learning a concept from demonstrations. 4.2 Few-shot concept learning Gal et al. [11] use a frozen generative model to learn visual concept representations from few images depicting the concept by optimizing the model s input v = arg minv Ex0,v,系 N(0,1),t ||系 系胃(xt, c胃(v), t)||2 2 , where c胃 and 系胃 are fixed. Liu et al. [12] extend this and learn visual concept compositions with a pretrained diffusion model in an unsupervised manner. Namely, from a set of images that depict various concepts, for each image xi they learn a set of weights 蠅i k and a shared set of visual concepts for all images ck, 藛系 = 系(xi t, t) + PK k=1 蠅i k(系(xi t, ck, t) 系(xi t, t)). We extend these formulations to inferring multiple concepts, whose composition describes a single task concept, from few demonstrations of a task. Given a trained diffusion model 系胃 and demonstrations of a new concept { 蟿}i from Dnew, we learn concepts { c1, ..., c K} for K 1 and their weights {蠅1, ..., 蠅K} that best describe the demonstrations. Starting from uniformly sampled concept embeddings ck U([0, 1]n), we freeze 系胃, and optimize ck and 蠅k: E系 N(0,I)[||系 (系胃(xt( 蟿), c , s0, t) + k=1 蠅k(系胃(xt( 蟿), ck, s0, t) 系胃(xt( 蟿), c , s0, t)))||2]. (2) We find that this compositional approach enables us to effectively represent and learn new demonstrations, even when demonstrations are substantially different than those seen in training tasks. 4.3 Generating the learned concept After learning concepts ck, whose composition describes the new task c, we evaluate the behavior it generates by initializing x T (蟿) N(0, 伪I), and compute xt N(碌t 1, 伪危t 1) iteratively as a function of the estimated denoising function 藛系(系胃), where 碌 and 危 are the mean and variance that define the reverse process, and 伪 [0, 1) is a scaling factor that leads to lower temperature samples, until generating x0 = 蟿 representing the trajectory of the agent. The denoising function is constructed by fixed or learned weights as defined in Eq. 2 and by any number of concepts 1. The applications of the generation procedure can be summarized as: Learned concept and demonstrated initial states. We apply our learned concept to a set of demonstrated initial states. In domains where the initial state and concept jointly determine optimal behavior, the generated trajectory corresponds to optimal actions to execute (e.g. goal-oriented navigation). In contrast to other domains where the initial state is irrelevant for a task due to the randomness in sampling x T (蟿) (e.g. motion capture), generated trajectories correspond to diverse plausible behaviors exhibiting the learned concept. Learned concept and novel initial states. We further apply the generation procedure from our learned concept on novel initial states, to generate trajectories of new behaviors exhibiting our conditioned concept. Prior methods may suffer from covariate shift in this setting [21]. We empirically show that our method is less prone to this problem. Learned concept composed with other concepts. Finally, we modify the generation procedure of our newly learned concept to generate trajectories that simultaneously exhibit other concepts. To generate a trajectory with an added another concept, we add another term to the sum in Eq. 2 where the learned concept is composed with a training concept ck and its weight 蠅k: 蠅k(系胃(xt(蟿), ck, s0, t) 系胃(xt(蟿), c , s0, t)). This modified generation procedure constructs trajectories which exhibits behavior that has a composition of the learned concept and the other specified concepts [14]. New Concepts training relation compositions novel concepts triangle right of square circle above square triangle diagonal to square Training single relation between two objects triangle right of square and circle above square line: circle right of triangle and triangle right of square . . . . . . circle, triangle and square on circle of radius r Figure 4: Object rearrangement. Training concepts are single pairwise relations ( A right of/above B ), and new concepts are either compositions of training concepts ( A right of/above B B right of/above C ) or new relations ( A diagonal to B , A, B, C on circle circumference of radius r ). square diagonal to triangle circle right of square + square diagonal to triangle Figure 5: Object rearrangement new concept qualitative evaluation. Learning the new concept square diagonal to triangle and composing it with the training concept circle right of square . Similarly to Ajay et al. [17], in environments where an inverse dynamics model is provided, we generate trajectories in a closed loop. We execute actions calculated by the inverse dynamics given the predicted plan by the model and then repeatedly replan given new observations. 5 Experiments We demonstrate results in four domains where concept representations are T5 [66] embeddings of task descriptions in natural language for training, and empty string embeddings for the dummy condition. During few-shot concept learning, we are provided with three to five demonstrations of a composition of training concepts or of a novel concept that is not an explicit composition of training tasks in natural language symbolic space. We ask a model to learn the concept from these demonstrations. 5.1 Task-Concepts Learning concepts describing goals that are spatial relations between objects. Object rearrangement is a common task in robotics [67 69] and embodied artificial intelligence (AI) [70, 71], serving as a foundation for a broader range of tasks such as housekeeping and manufacturing. Here, we use a 2D object rearrangement domain to evaluate the ability of our method to learn task specification concepts. Given a concept representing a relation between objects, we generate a single state describing that relation. The concept in a training example describes the relation (either right of or above ) between only one pair of objects (out of three objects) in the environment. Then, a model must learn compositions of these pairwise relations and new concepts such as diagonal and circle (see Figure 4). The results in Figures 5 and 6 demonstrate that our method learns unseen compositions of training concepts and new concepts. They further demonstrate how our method composes new concepts with learned concepts. For additional qualitative results, please refer to Appendix A. While successful in most cases, there are also a few failure examples. The accuracy for the new circle concept is low (0.44) compared to the mean over task types in Figure 6 Object Rearrangement New Concept (0.82 0.09). This is most likely due to this concept lying far out of the training distribution. The task square right of circle triangle above circle has low accuracy for 2 concepts (0.32) compared to the mean in Table 2 Object Rearrangement Training Composition (0.75 0.11). This may arise from the combined concept-weight optimization process as there is no explicit regularization on weights, they may converge to 0 or diverge. In Figure 12, we show that concept components may or may not capture new concept relations. Learning concepts describing goals based on attributes of target objects. We test our method in a goal-oriented navigation domain adapted from the AGENT dataset [72], where an agent navigates to one of two potential targets. Conditioned on a concept representing the attributes of the desired Training Training Composition New Concept New and Training Concept Composition Ours BC VAE Training Training Composition Training Composition New Initial State Ours BC VAE In-Context Figure 6: Object rearrangement (left) and AGENT closed loop (right) quantitative evaluation on training and few-shot novel concept learning. Accuracy of FTL-IGM (ours), BC, VAE, and In-Context over concept generation of training concepts, novel compositions, novel concepts, and new initial states. We plot the average and standard error over new task types. Full details of the evaluation metrics appear in Appendix B and for baselines implementation in Appendix C. target object and initial state, we generate a state-based trajectory describing an agent navigating to the target. Each object has a color and a shape out of four possible colors and four shapes. During training, we provide 16 target-distractor combinations that include all colors and shapes (but not all combinations), and a concept is conditioned on one of the target s attributes (e.g., color). We introduce new concepts defined by both target attributes, including (1) unseen color-shape target combinations and (2) new target-distractor combinations. Figure 7 shows an example. In training, we see bowl and red object targets. A new concept includes a novel composition as the target red bowl. The new concept distractor objects (green bowl and red sphere) were introduced during training, but they were not paired with a red bowl as the target. As Figure 13 shows, our method successfully learns concepts where targets are new compositions of target attributes in settings with new target-distractor pairs and generalizes to new initial object states. We further evaluate our model and baselines in closed loop (Figure 6) by making an additional assumption that a planner is provided. The planner produces an action given a current state and a future desired state predicted by a model. Learning concepts describing human motion. Unlike prior work on learning to compose human poses from motion capture (Mo Cap) data [e.g., 73, 74], here we focus on the inverse problem learning new actions from Mo Cap data. In particular, we use the CMU Graphics Lab Motion Capture Database (http://mocap.cs.cmu.edu/). We train on various human actions in the database and few-shot learn three novel concepts (see Appendix B.3 for details). Learning tasks from few demonstrations is especially beneficial in this domain since describing motion concepts in words could be hard. In Table 5.1, we ask five human volunteers to select generated behaviors that depict training and new concepts. We demonstrate quantitatively that our method generates human motion that captures the desired behavior across training and new behaviors. We qualitatively demonstrate how our method generates learned new concepts ( jumping jacks and breaststroke ) from new initial states and composes jumping jacks with training concepts walk , jump , and march . Results are best viewed on our website website. We compare motion generated by our method to various baselines on new initial states for jumping jacks and breaststroke . Our method is noisy yet captures the widest range of motion. While other methods often produce smoother trajectories, they mostly capture local (VAE) or degenerate (BC, In-Context, Language) motion. Learning concepts describing driving scenarios. In an Autonomous Driving domain [75], an agent acts in a challenging multi-agent environment to complete a driving task. We train on several driving scenarios ( highway , exit , merge , and intersection ) and learn a new driving scenario ( roundabout ) from several demonstrations (see Figure 8 and further details in Appendix B.4). We evaluate this scenario in closed loop on new initial states, assuming access to a planner that can simulate taking actions in the environment. Over two evaluation metrics (crash and task completion rate), our method achieves overall best results (Figure 9). Learning concepts describing real-world table-top manipulation tasks. We evaluate our method s capability to learn a novel concept for real-world table-top manipulation with a Franka Research 3 robot. Training concepts include pick green circle and place on book , pick green circle and place on elevated white surface , push green circle to orange triangle and push green circle to orange triangle around purple bowl . The new scenario includes pushing the green circle to the orange triangle on a book (Figure 10). We evaluate training pushing in closed loop and and achieve success Distractor training red/bowl training yellow/bowl training red/cube training yellow/cube training purple/cone Target red bowl yellow bowl red cube yellow cube purple cone Distractor yellow object red object cube bowl Target red object yellow object bowl cube go to red object go to bowl go to red bowl Training target defined by single color/shape attribute New Concepts target defined by novel composition of attributes Figure 7: Goal-oriented navigation. In training, targets are defined by a single attribute (color or shape), and in new concepts, targets are defined by a novel combination of color and shape attributes. To make the setting more challenging, distractor objects in new concept demonstrations have a combination of attributes that are within the training distribution. Setting BC VAE Ours Training 20 16.3 0 0 80 16.3 66.7 0 46.7 16.3 100 0 New Concept 26.7 13.3 0 0 73.3 13.3 New Initial State 40 24.9 26.7 13.3 80 16.3 Table 1: Mo Cap human experiment. For training concepts and new concepts on new initial states, we report (top) percentage of time each method is the most successful at depicting a concept and (bottom) percentage of time each method depicts a concept. Mean and standard deviation are calculated over the number of scenarios in each setting and human subjects. (90% accuracy). We evaluate the new concept, elevated pushing, against a baseline that conditions on the training push green circle to orange triangle representation in the new scenario setup where the objects are placed on a book. Learning the new concept succeeds (55% accuracy) while using the training representation fails (15% accuracy) and the robot often pushes the book instead of the object. Details are in Appendix B.5 and qualitative results are on our website. 5.2 Baselines Goal conditioned behavior cloning. We compare our method with goal-conditioned behavior cloning (BC), which, given a condition and a state, outputs the next state in a sequence. It is trained on our paired pretraining dataset and learns concepts by optimizing the input condition to maximize the likelihood of new concept demonstration transitions. We test its ability to compose new learned concepts with training concepts naively by adding conditions that are then inputted into the model. We observe that even though goal-conditioned BC has access to the pretrained dataset and conditions, and while it may learn new concepts, it suffers from covariate shift on new initial states and lacks the ability to generalize to novel compositions of the learned concept with training concepts (Figures 6, 9, 13). To achieve these generalization capacities, we need a model that can process interpolated (initial states) and composed (concepts) conditions, such as our generative model. Learning the concept space with a VAE. We compare our method with VAE [76] that does not utilize the concepts in the paired pretraining data but rather learns the concept space by reconstructing pretraining data trajectories through their encoded representation z. Trajectories are generated via a decoder conditioned on an initial state and z with added noise. z is obtained by encoding a trajectory for training evaluation and by fixing the decoder and optimizing z to generate a given demonstration when learning a new concept. We find that the VAE model learns a latent space that captures training and new concepts but does not enable generalization to new initial states (Figures 6, 9, 13). This highlights the importance of concept representations in the pretrained data. New Concept intersection Figure 8: Autonomous driving. A controlled agent (green) completes various driving objectives as fast as possible while sharing the road with other vehicles (blue) and avoiding collisions. Training concepts: highway , exit , merge , and intersection , new concept: roundabout . Training Crash New Concept Crash 0 Ours BC VAE In-Context Training Success New Concept Success 0 Ours BC VAE In-Context Figure 9: Driving crash (left) and success (right) rates. Crash rate (lower is better) and task completion rate (higher is better) averaged over training tasks. We report standard errors over training tasks and accuracy over 50 trajectories generated from the learned new concept. VAE has a high completion rate yet a high crash rate. In-context has a low crash rate yet 0% success rate typically, the controlled vehicle reaches the roundabout s center but does not complete the crossing. Overall, our method learns to complete the roundabout crossing with competitive crash and success rates. Training concepts New concept closed loop evaluation pick&place on book push on surface push around pick&place on table push on book Figure 10: Table-top manipulation. Training concepts: pick-and-place onto elevated surfaces and table-top pushing. New concept: pushing on an elevated surface. In-Context learning. We compare our approach with training a method to in-context learn from demonstrations. Specifically, we compare our approach to Xu et al. [28] using the pretrained dataset for few-shot behavior generation without further training. We sequentially predict states from demonstrations of a concept and window of current states and show that our method adapts better to new concepts (Figures 6, 9, 13). This emphasizes the need to learn explicit concept representations. Conditioning on Language Descriptions of New Concepts. There has been work on generating actions from language instructions [77 79]. We demonstrate that in our setup, merely providing new concept language instructions embedded with T5 (as in our pretraining dataset) is insufficient, and generalization is better when learning concepts from few demonstrations. For AGENT, training compositions on new initial states has an average accuracy and standard error of 0.63 0.07, lower than ours (0.73 0.07). For Object Rearrangement, training compositions (0.2 0.07), new concept (0.2 0.05), and new and training concept compositions (0.13 0.04) accuracies are significantly lower than ours (0.9 0.04, 0.82 0.09 and 0.8 0.02). For Mo Cap, we demonstrate qualitatively that instead of capturing new human actions, the agent transitions into walking. Results are best viewed on our website. 5.3 Learning two concepts yields higher accuracy than one concept When learning weights together with concepts, we check the effect of the number of learned concepts and weights. We report results in Table 2 for Object Rearrangement, AGENT, and Driving, and find that, on average, learning two concepts improves concept learning. We demonstrate qualitatively for Mo Cap that learning two conditions is preferable. In jumping jacks , we observe that the motion lacks raising and lowering both arms and in breaststroke , it lacks complete arm and upper torso movement. Results are best viewed on our website. 5.4 How are learned new concepts related to training concepts? New concepts that are compositions of training concepts. We analyze what the learned two concepts in Object Rearrangement and AGENT learn for novel concept compositions (e.g., red bowl ). For each concept (e.g., red and bowl ), we generate two sets of 50 samples from the learned components. Table 3 shows accuracy for these sets over the concepts. In some cases (most notably the line concept in Object Rearrangement, circle right of triangle and triangle right of square ), each Environment Setting 1 Concept 2 Concepts Object Training Composition 0.28 0.11 0.75 0.11 Rearrangement New Concept 0.49 0.11 0.77 0.09 New+Training 0.38 0.04 0.73 0.04 AGENT Training Composition 0.52 0.08 0.68 0.13 New Initial State 0.54 0.07 0.67 0.05 Driving New Initial State 0.14 0.24 Table 2: Ablation on the number of learned concepts. We test the effect of the number of learned concepts and weights in FTL-IGM on the generation accuracy of new learned concepts. On average, learning two concept components and their weights is preferable to learning one concept component and its weight. We report average accuracy and standard error over task types for Object Rearrangement and AGENT. Driving includes a single new concept and we report accuracy only. learned component captures a single composed concept. In other cases, a single learned component captures both concepts (Figure 12). New concepts that are not explicit compositions of training concepts in natural language symbolic space. In Figure 11 we visualize t-SNE [80] embeddings for T5 training concept representations and learned concept component representations. We note that learned components are relatively close to training concepts, maintaining the model s input distribution, yet capture concepts that are not explicit compositions of training concepts. learned new concept training concept Object Rearrangement Mo Cap Driving Figure 11: t-SNE embeddings of new concepts that are not explicit compositions of training concepts. See interactive version for detailed labels on our website. 6 Discussion and Limitations In this work, we formulate the problem of new task concept learning as Few-Shot Task Learning through Inverse Generative Modeling (FTL-IGM). We adapt a method for concept learning based on this new formulation and evaluate task concept learning against baselines in four domains. Our extensive experimental results show that, unlike the baselines, FTL-IGM successfully learns novel concepts from a few examples and generalizes the learned concepts to unfamiliar scenarios. It also composes learned concepts to form unseen behavior thanks to the compositionality of the generative model. These results demonstrate the efficacy, sample efficiency, and generalizability of FTL-IGM. However, our work has several limitations. First, while our framework is general for any parameterized generative model, our implementation with a diffusion model incurs high inference time. We note that there is still space for improvement in the Mo Cap generation quality and in the compatibility rate of demonstrations generated by composing learned and training concepts. In addition, we assume that learned concepts lie within the landscape of training concepts to learn them from a few demonstrations without retraining the model. We have approached the question of what new concepts can be represented by compositions of concepts in this landscape empirically, leaving a theoretical analysis as future work. We are hopeful that with the continued progress in the field of generative AI, more powerful pretrained models will become available. Combined with our framework, this will unlock a stronger ability to learn and generalize various task concepts in complex domains. 7 Acknowledgments This research was also partly sponsored by the United States Air Force Research Laboratory and the United States Air Force Artificial Intelligence Accelerator and was accomplished under Cooperative Agreement Number FA8750-192-1000. The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the United States Air Force or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes, notwithstanding any copyright notation herein. We acknowledge support from ONR MURI under N00014-22-1-2740 and ARO MURI under W911NF-23-1-0277. Yilun is supported in part by an NSF Graduate Research Fellowship. We would like to thank Abhishek Bhandwaldar for help with the AGENT environment, and Anurag Ajay, Lucy Chai, Andi Peng, Felix Yanwei Wang, Anthony Simeonov and Zhang-Wei Hong for helpful discussions. [1] Brenden M Lake, Tal Linzen, and Marco Baroni. Human few-shot learning of compositional instructions. ar Xiv preprint ar Xiv:1901.04587, 2019. 1 [2] Archit Parnami and Minwoo Lee. Learning from few examples: A summary of approaches to few-shot learning. ar Xiv preprint ar Xiv:2203.04291, 2022. 1 [3] Hidetoshi Shimodaira. Improving predictive inference under covariate shift by weighting the log-likelihood function. Journal of Statistical Planning and Inference, 90(2):227 244, 2000. 1, 3 [4] Brian D Ziebart, Andrew L Maas, J Andrew Bagnell, Anind K Dey, et al. Maximum entropy inverse reinforcement learning. In AAAI Conference on Artificial Intelligence, volume 8, pages 1433 1438, 2008. 1 [5] Andrew Y Ng, Stuart Russell, et al. Algorithms for inverse reinforcement learning. In International Conference on Machine Learning, volume 1, page 2, 2000. 3 [6] Justin Fu, Katie Luo, and Sergey Levine. Learning robust rewards with adversarial inverse reinforcement learning. In International Conference on Learning Representations, 2018. 1, 3 [7] Seyed Kamyar Seyed Ghasemipour, Richard Zemel, and Shixiang Gu. A divergence minimization perspective on imitation learning methods. In Conference on Robot Learning, pages 1259 1277. PMLR, 2020. 1 [8] Yan Duan, Marcin Andrychowicz, Bradly Stadie, Open AI Jonathan Ho, Jonas Schneider, Ilya Sutskever, Pieter Abbeel, and Wojciech Zaremba. One-shot imitation learning. Advances in Neural Information Processing Systems, 30, 2017. 1, 3, 25 [9] Chelsea Finn, Tianhe Yu, Tianhao Zhang, Pieter Abbeel, and Sergey Levine. One-shot visual imitation learning via meta-learning. In Conference on Robot Learning, pages 357 368. PMLR, 2017. 1 [10] Brenden M Lake, Ruslan Salakhutdinov, and Joshua B Tenenbaum. Human-level concept learning through probabilistic program induction. Science, 350(6266):1332 1338, 2015. 1 [11] Rinon Gal, Yuval Alaluf, Yuval Atzmon, Or Patashnik, Amit H Bermano, Gal Chechik, and Daniel Cohen-Or. An image is worth one word: Personalizing text-to-image generation using textual inversion. In International Conference on Learning Representations, 2023. 3, 5 [12] Nan Liu, Yilun Du, Shuang Li, Joshua B Tenenbaum, and Antonio Torralba. Unsupervised compositional concepts discovery with text-to-image generative models. In International Conference on Computer Vision, 2023. 1, 3, 5 [13] Aaron van den Oord, Sander Dieleman, Heiga Zen, Karen Simonyan, Oriol Vinyals, Alex Graves, Nal Kalchbrenner, Andrew Senior, and Koray Kavukcuoglu. Wavenet: A generative model for raw audio. ar Xiv preprint ar Xiv:1609.03499, 2016. 2 [14] Nan Liu, Shuang Li, Yilun Du, Antonio Torralba, and Joshua B Tenenbaum. Compositional visual generation with composable diffusion models. ar Xiv preprint ar Xiv:2206.01714, 2022. 2, 3, 5 [15] Aditya Ramesh, Mikhail Pavlov, Gabriel Goh, Scott Gray, Chelsea Voss, Alec Radford, Mark Chen, and Ilya Sutskever. Zero-shot text-to-image generation. In International Conference on Machine Learning, pages 8821 8831. PMLR, 2021. 2 [16] Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily L Denton, Kamyar Ghasemipour, Raphael Gontijo Lopes, Burcu Karagol Ayan, Tim Salimans, et al. Photorealistic text-to-image diffusion models with deep language understanding. In Advances in Neural Information Processing Systems, volume 35, pages 36479 36494, 2022. 2 [17] Anurag Ajay, Yilun Du, Abhi Gupta, Joshua B. Tenenbaum, Tommi S. Jaakkola, and Pulkit Agrawal. Is conditional generative modeling all you need for decision making? In International Conference on Learning Representations, 2023. 2, 3, 5, 6, 24, 25 [18] Yiming Ding, Carlos Florensa, Pieter Abbeel, and Mariano Phielipp. Goal-conditioned imitation learning. Advances in Neural Information Processing Systems, 32, 2019. 3 [19] St茅phane Ross, Geoffrey Gordon, and Drew Bagnell. A reduction of imitation learning and structured prediction to no-regret online learning. In Proceedings of the International Conference on Artificial Intelligence and Statistics, pages 627 635, 2011. 3 [20] Minghuan Liu, Tairan He, Minkai Xu, and Weinan Zhang. Energy-based imitation learning. In International Conference on Autonomous Agents and Multiagent Systems, 2021. 3 [21] Jonathan Ho and Stefano Ermon. Generative adversarial imitation learning. Advances in Neural Information Processing Systems, 29, 2016. 3, 5 [22] Dzmitry Bahdanau, Felix Hill, Jan Leike, Edward Hughes, Arian Hosseini, Pushmeet Kohli, and Edward Grefenstette. Learning to understand goal specifications by modelling reward. In International Conference on Learning Representations, 2019. 3 [23] Justin Fu, Anoop Korattikara, Sergey Levine, and Sergio Guadarrama. From language to goals: Inverse reinforcement learning for vision-based instruction following. In International Conference on Learning Representations, 2019. 3 [24] Adam Gleave and Oliver Habryka. Multi-task maximum entropy inverse reinforcement learning. In Proceedings of the 1st Workshop on Goal Specifications for Reinforcement Learning, 2018. 3 [25] Chris L Baker, Rebecca Saxe, and Joshua B Tenenbaum. Action understanding as inverse planning. Cognition, 113(3):329 349, 2009. 3 [26] Chris L Baker, Julian Jara-Ettinger, Rebecca Saxe, and Joshua B Tenenbaum. Rational quantitative attribution of beliefs, desires and percepts in human mentalizing. Nature Human Behaviour, 1(4):0064, 2017. [27] Tan Zhi-Xuan, Jordyn Mann, Tom Silver, Josh Tenenbaum, and Vikash Mansinghka. Online bayesian goal inference for boundedly rational planning agents. Advances in Neural Information Processing Systems, 33:19238 19250, 2020. 3 [28] Mengdi Xu, Yikang Shen, Shun Zhang, Yuchen Lu, Ding Zhao, Joshua Tenenbaum, and Chuang Gan. Prompting decision transformer for few-shot policy generalization. In International Conference on Machine Learning, pages 24631 24645. PMLR, 2022. 3, 9, 25 [29] Ahmed Touati, J茅r茅my Rapin, and Yann Ollivier. Does zero-shot reinforcement learning exist? ar Xiv preprint ar Xiv:2209.14935, 2022. 3 [30] Sherjil Ozair, Yazhe Li, Ali Razavi, Ioannis Antonoglou, Aaron Van Den Oord, and Oriol Vinyals. Vector quantized models for planning. In International Conference on Machine Learning, pages 8302 8313. PMLR, 2021. 3 [31] Soroush Nasiriany, Vitchyr Pong, Steven Lin, and Sergey Levine. Planning with goalconditioned policies. Advances in Neural Information Processing Systems, 32, 2019. [32] John Co-Reyes, Yu Xuan Liu, Abhishek Gupta, Benjamin Eysenbach, Pieter Abbeel, and Sergey Levine. Self-consistent trajectory autoencoder: Hierarchical reinforcement learning with trajectory embeddings. In International Conference on Machine Learning, pages 1009 1018. PMLR, 2018. [33] Karol Hausman, Jost Tobias Springenberg, Ziyu Wang, Nicolas Heess, and Martin Riedmiller. Learning an embedding space for transferable robot skills. In International Conference on Learning Representations, 2018. [34] Divyansh Garg, Skanda Vaidyanath, Kuno Kim, Jiaming Song, and Stefano Ermon. Lisa: Learning interpretable skill abstractions from language. Advances in Neural Information Processing Systems, 35:21711 21724, 2022. 3 [35] Pete Florence, Corey Lynch, Andy Zeng, Oscar A Ramirez, Ayzaan Wahid, Laura Downs, Adrian Wong, Johnny Lee, Igor Mordatch, and Jonathan Tompson. Implicit behavioral cloning. In Conference on Robot Learning, pages 158 168. PMLR, 2022. 3 [36] Michael Janner, Yilun Du, Joshua B Tenenbaum, and Sergey Levine. Planning with diffusion for flexible behavior synthesis. In International Conference on Machine Learning, 2022. 3, 5 [37] Cheng Chi, Siyuan Feng, Yilun Du, Zhenjia Xu, Eric Cousineau, Benjamin Burchfiel, and Shuran Song. Diffusion policy: Visuomotor policy learning via action diffusion. In Proceedings of Robotics: Science and Systems, 2023. 3 [38] Lili Chen, Kevin Lu, Aravind Rajeswaran, Kimin Lee, Aditya Grover, Misha Laskin, Pieter Abbeel, Aravind Srinivas, and Igor Mordatch. Decision transformer: Reinforcement learning via sequence modeling. Advances in Neural Information Processing Systems, 2021. 3 [39] Zhengyao Jiang, Tianjun Zhang, Michael Janner, Yueying Li, Tim Rockt盲schel, Edward Grefenstette, and Yuandong Tian. Efficient planning in a compact latent action space. International Conference on Learning Representations, 2023. [40] Shyam Sudhakaran and Sebastian Risi. Skill decision transformer. ar Xiv preprint ar Xiv:2301.13573, 2023. 3 [41] Chiyu Jiang, Andre Cornman, Cheolho Park, Benjamin Sapp, Yin Zhou, Dragomir Anguelov, et al. Motiondiffuser: Controllable multi-agent motion prediction using diffusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9644 9653, 2023. 3 [42] Yanjie Ze, Gu Zhang, Kangning Zhang, Chenyuan Hu, Muhan Wang, and Huazhe Xu. 3d diffusion policy. ar Xiv preprint ar Xiv:2403.03954, 2024. 3 [43] Abhishek Padalkar, Acorn Pooley, Ajinkya Jain, Alex Bewley, Alex Herzog, Alex Irpan, Alexander Khazatsky, Anant Rai, Anikait Singh, Anthony Brohan, et al. Open x-embodiment: Robotic learning datasets and rt-x models. ar Xiv preprint ar Xiv:2310.08864, 2023. 3 [44] Haoyu Zhen, Xiaowen Qiu, Peihao Chen, Jincheng Yang, and Xin Yan. Yilun du, yining hong, and chuang gan. 3d-vla: A 3d vision-language-action generative world model. ar Xiv preprint ar Xiv:2403.09631, 7(8), 2024. 3 [45] Moo Jin Kim, Karl Pertsch, Siddharth Karamcheti, Ted Xiao, Ashwin Balakrishna, Suraj Nair, Rafael Rafailov, Ethan Foster, Grace Lam, Pannag Sanketi, et al. Openvla: An open-source vision-language-action model. ar Xiv preprint ar Xiv:2406.09246, 2024. [46] Octo Model Team, Dibya Ghosh, Homer Walke, Karl Pertsch, Kevin Black, Oier Mees, Sudeep Dasari, Joey Hejna, Tobias Kreiman, Charles Xu, et al. Octo: An open-source generalist robot policy. ar Xiv preprint ar Xiv:2405.12213, 2024. 3 [47] Irina Higgins, Loic Matthey, Arka Pal, Christopher P Burgess, Xavier Glorot, Matthew M Botvinick, Shakir Mohamed, and Alexander Lerchner. beta-vae: Learning basic visual concepts with a constrained variational framework. International Conference on Learning Representations, 3, 2017. 3 [48] Christopher P Burgess, Loic Matthey, Nicholas Watters, Rishabh Kabra, Irina Higgins, Matt Botvinick, and Alexander Lerchner. Monet: Unsupervised scene decomposition and representation. ar Xiv preprint ar Xiv:1901.11390, 2019. 3 [49] Klaus Greff, Rapha毛l Lopez Kaufman, Rishabh Kabra, Nick Watters, Christopher Burgess, Daniel Zoran, Loic Matthey, Matthew Botvinick, and Alexander Lerchner. Multi-object representation learning with iterative variational inference. In International Conference on Machine Learning, pages 2424 2433. PMLR, 2019. 3 [50] Yilun Du, Shuang Li, Yash Sharma, Josh Tenenbaum, and Igor Mordatch. Unsupervised learning of compositional energy concepts. Advances in Neural Information Processing Systems, 34: 15608 15620, 2021. 3 [51] Nan Liu, Shuang Li, Yilun Du, Josh Tenenbaum, and Antonio Torralba. Learning to compose visual relations. Advances in Neural Information Processing Systems, 34:23166 23178, 2021. 3 [52] Weili Nie, Arash Vahdat, and Anima Anandkumar. Controllable and compositional generation with latent-space energy-based models. Advances in Neural Information Processing Systems, 34:13497 13510, 2021. [53] Yilun Du, Conor Durkan, Robin Strudel, Joshua B Tenenbaum, Sander Dieleman, Rob Fergus, Jascha Sohl-Dickstein, Arnaud Doucet, and Will Sussman Grathwohl. Reduce, reuse, recycle: Compositional generation with energy-based diffusion models and mcmc. In International Conference on Machine Learning, pages 8489 8510. PMLR, 2023. [54] Zihao Wang, Lin Gui, Jeffrey Negrea, and Victor Veitch. Concept algebra for (score-based) text-controlled generative models. Advances in Neural Information Processing Systems, 36, 2024. [55] Yilun Du, Shuang Li, and Igor Mordatch. Compositional visual generation and inference with energy based models. ar Xiv preprint ar Xiv:2004.06030, 2020. 3 [56] Weixi Feng, Xuehai He, Tsu-Jui Fu, Varun Jampani, Arjun Akula, Pradyumna Narayana, Sugato Basu, Xin Eric Wang, and William Yang Wang. Training-free structured diffusion guidance for compositional text-to-image synthesis. ar Xiv preprint ar Xiv:2212.05032, 2022. 3 [57] Changhao Shi, Haomiao Ni, Kai Li, Shaobo Han, Mingfu Liang, and Martin Renqiang Min. Exploring compositional visual generation with latent classifier guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 853 862, 2023. [58] Yuren Cong, Martin Renqiang Min, Li Erran Li, Bodo Rosenhahn, and Michael Ying Yang. Attribute-centric compositional text-to-image generation. ar Xiv preprint ar Xiv:2301.01413, 2023. [59] Lianghua Huang, Di Chen, Yu Liu, Yujun Shen, Deli Zhao, and Jingren Zhou. Composer: Creative and controllable image synthesis with composable conditions. ar Xiv preprint ar Xiv:2302.09778, 2023. 3 [60] Timur Garipov, Sebastiaan De Peuter, Ge Yang, Vikas Garg, Samuel Kaski, and Tommi Jaakkola. Compositional sculpting of iterative generative processes. Advances in Neural Information Processing Systems, 36:12665 12702, 2023. 3 [61] Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models. In Advances in Neural Information Processing Systems, volume 33, pages 6840 6851, 2020. 4 [62] Prafulla Dhariwal and Alexander Nichol. Diffusion models beat gans on image synthesis. Advances in Neural Information Processing Systems, 34:8780 8794, 2021. 4 [63] Jonathan Ho and Tim Salimans. Classifier-free diffusion guidance. ar Xiv preprint ar Xiv:2207.12598, 2022. 4, 5 [64] Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with clip latents. ar Xiv preprint ar Xiv:2204.06125, 1(2):3, 2022. 4 [65] Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob Mc Grew, Ilya Sutskever, and Mark Chen. Glide: Towards photorealistic image generation and editing with text-guided diffusion models. ar Xiv preprint ar Xiv:2112.10741, 2021. 4 [66] Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485 5551, 2020. 6 [67] Ankit Shah, Pritish Kamath, Julie A Shah, and Shen Li. Bayesian inference of temporal task specifications from demonstrations. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems 31, pages 3804 3813. Curran Associates, Inc., 2018. 6 [68] Fujian Yan, Dali Wang, and Hongsheng He. Robotic understanding of spatial relationships using neural-logic learning. In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems, pages 8358 8365. IEEE, 2020. [69] Ryan Rowe, Shivam Singhal, Daqing Yi, Tapomayukh Bhattacharjee, and Siddhartha S Srinivasa. Desk organization: Effect of multimodal inputs on spatial relational learning. In 2019 28th IEEE International Conference on Robot and Human Interactive Communication, pages 1 8. IEEE, 2019. 6 [70] Xavier Puig, Tianmin Shu, Shuang Li, Zilin Wang, Yuan-Hong Liao, Joshua B Tenenbaum, Sanja Fidler, and Antonio Torralba. Watch-and-help: A challenge for social perception and human-ai collaboration. ar Xiv preprint ar Xiv:2010.09890, 2020. 6 [71] Sanjana Srivastava, Chengshu Li, Michael Lingelbach, Roberto Mart铆n-Mart铆n, Fei Xia, Kent Elliott Vainio, Zheng Lian, Cem Gokmen, Shyamal Buch, Karen Liu, et al. Behavior: Benchmark for everyday household activities in virtual, interactive, and ecological environments. In Conference on Robot Learning, pages 477 490. PMLR, 2022. 6 [72] Tianmin Shu, Abhishek Bhandwaldar, Chuang Gan, Kevin Smith, Shari Liu, Dan Gutfreund, Elizabeth Spelke, Joshua Tenenbaum, and Tomer Ullman. Agent: A benchmark for core psychological reasoning. In International Conference on Machine Learning, 2021. 6, 21, 22 [73] Borui Wang, Ehsan Adeli, Hsu-kuang Chiu, De-An Huang, and Juan Carlos Niebles. Imitation learning for human pose prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 7124 7133, 2019. 7 [74] Guy Tevet, Sigal Raab, Brian Gordon, Yonatan Shafir, Daniel Cohen-Or, and Amit H Bermano. Human motion diffusion model. International Conference on Learning Representations, 2023. 7 [75] Edouard Leurent. An environment for autonomous driving decision-making. https://github. com/eleurent/highway-env, 2018. 7, 23 [76] Diederik P Kingma and Max Welling. Auto-encoding variational bayes. ar Xiv preprint ar Xiv:1312.6114, 2013. 8, 25 [77] Oier Mees, Lukas Hermann, Erick Rosete-Beas, and Wolfram Burgard. Calvin: A benchmark for language-conditioned policy learning for long-horizon robot manipulation tasks. IEEE Robotics and Automation Letters, 7(3):7327 7334, 2022. 9 [78] Anthony Brohan, Noah Brown, Justice Carbajal, Yevgen Chebotar, Joseph Dabis, Chelsea Finn, Keerthana Gopalakrishnan, Karol Hausman, Alex Herzog, Jasmine Hsu, et al. Rt-1: Robotics transformer for real-world control at scale. ar Xiv preprint ar Xiv:2212.06817, 2022. [79] Corey Lynch, Ayzaan Wahid, Jonathan Tompson, Tianli Ding, James Betker, Robert Baruch, Travis Armstrong, and Pete Florence. Interactive language: Talking to robots in real time. IEEE Robotics and Automation Letters, 2023. 9 [80] Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of Machine Learning Research, 9(11), 2008. 10 [81] Klaus Greff, Francois Belletti, Lucas Beyer, Carl Doersch, Yilun Du, Daniel Duckworth, David J Fleet, Dan Gnanapragasam, Florian Golemo, Charles Herrmann, et al. Kubric: A scalable dataset generator. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3749 3761, 2022. 22 [82] Julian Tanke, Chintan Zaveri, and Juergen Gall. Intention-based Long-Term Human Motion Anticipation. International Conference on 3D Vision, 2021. 23 [83] Edouard Leurent. rl-agents: Implementations of reinforcement learning algorithms. https: //github.com/eleurent/rl-agents, 2018. 23 [84] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 770 778, 2016. 25 [85] Dean Pomerleau. Alvinn: An autonomous land vehicle in a neural network. In D.S. Touretzky, editor, Proceedings of Neural Information Processing Systems, pages 305 313. Morgan Kaufmann, December 1989. 25 [86] Ilya Loshchilov and Frank Hutter. Fixing weight decay regularization in adam. ar Xiv preprint ar Xiv:1711.05101, 5, 2017. 25 A Additional Results Object Rearrangement. We display additional states generated by our model for various new concepts. We generate learned concepts that are compositions of training concepts (Figure 14), new concepts that are not explicit compositions of training concepts (Figure 15), and a new concept composed with a training concept (Figure 16). We analyze the learned concepts qualitatively (Figure 12) and quantitatively (Table 3). Goal-Oriented Navigation. We provide accuracy for open loop evaluation in Figure 13. We further analyze the learned concepts quantitatively (Table 3). Component 1 Component 2 circle right of triangle and triangle right of square square right of circle and triangle above circle Figure 12: Object rearrangement new compositions analysis qualitative evaluation. circle right of triangle triangle right of square (top) each learned component corresponds to a single composed concept: circle right of triangle (component 1) and triangle right of square (component 2). In square right of circle triangle above circle (bottom), learned component 2 corresponds to both composed concepts and component 1 to none. Training Training Composition Training Composition New Initial State Ours BC VAE In-Context Figure 13: AGENT open loop evaluation. We plot the average and standard error over task types. B Data Generation and Evaluation B.1 Object Rearrangement Training. The training dataset of 11k samples consists of concepts A right of B and A above B , where A and B are one of three objects: circle, triangle, or square. Altogether, there are 12 possible concepts (two relations and three objects where order is important). In the data generation process, A at center position (x A, y A) [0, 5]2 with radius r A [0.3, 1] and angle 胃A [0, 2蟺] is considered right of B at center position (x B, y B) with radius r B if x A > x B and |y A y B| r A. Similarly, A is considered above B if y A > y B and |x A x B| r A. We further verify that training objects do not overlap. New Tasks. The new scenarios include Five novel compositions of training concepts (training composition in Figure 6). triangle right of square circle above square , square right of triangle circle above triangle , triangle right of square and circle above square square right of triangle and circle above triangle line: circle right of triangle and triangle right of square circle right of square and triangle above square square right of circle and triangle above circle Figure 14: New concepts that are training concept compositions. We display successful (green frames) and unsuccessful (red frames) states generated by our model conditioned on learned concepts that are training concept compositions. circle, triangle and square on circle of radius r=1.67 triangle diagonal to square circle diagonal to triangle triangle diagonal to circle Figure 15: New concepts that are not explicit training concept compositions. We display successful (green frames) and unsuccessful (red frames) states generated by our model conditioned on learned concepts that are not explicit training concept compositions. circle above square + square diagonal to triangle circle above triangle + square diagonal to triangle circle right of triangle + square diagonal to triangle triangle above circle + square diagonal to triangle Figure 16: New concept composed with training concepts. We display successful (green frames) and unsuccessful (red frames) states generated by our model conditioned on a learned concept that is not an explicit training concept composition ( square diagonal to triangle ) in composition with various training concepts. Table 3: Analysis of learned concepts. For new concepts that are explicit training concept compositions, we evaluate what each learned component captures. For both concepts 1 and 2, we report accuracy with respect to the concept when trajectories are generated solely from one learned component. e.g. in AGENT (bottom), 42% of the trajectories generated by component 1 and 48% of the trajectories generated by component 2 target the red object. Environment Concept 1 Component 1 Component 2 Concept 2 Component 1 Component 2 Object Rearrangement R 0.42 0.62 A 0.88 0.36 R 0.16 0.64 A 0.08 0.08 R 0.1 0.48 A 0.86 0.4 R 0.14 0.9 A 0.02 0.6 R 0 0.92 R 0.64 0.14 red 0.42 0.48 bowl 0.66 0.56 yellow 0.72 0.44 bowl 0.58 0.74 red 0.42 0.76 cube 0.44 0.96 yellow 0.4 0.32 cube 0.82 0.78 purple 0.48 0.42 cone 0.56 0.36 circle right of square triangle above square , square right of circle triangle above circle , line : circle right of triangle triangle right of square . There are five demonstrations of each novel composition. Five new concepts (new concept in Figure 6). circle : circle, triangle, and square lie in equal intervals on the circumference of a circle with radius 1.67 and center [0, 5]2, square diagonal to triangle , triangle diagonal to square , circle diagonal to triangle , triangle diagonal to circle . A is considered diagonal to B if their centers lie on f(x) = x and if A is above and right of B. There are five demonstrations of each novel concept. Composing a new concept, square diagonal to triangle , with five training concepts (new and training concept composition in Figure 6): circle right of square , circle above square , circle above triangle , circle right of triangle , triangle above circle . When we learn weights for the new concept, the training concept is weighted by 蠅 = 1. Otherwise, both training and learned concepts are weighted by fixed 蠅. We verify that objects do not overlap and that no training relations unrelated to the specified new concepts exist in the demonstrations between objects. State space. The state space is a 21-tuple describing three shapes (circle, triangle, and square), each represented by a 7-tuple: their center 2D position, size, angle, and one-hot shape type. Evaluation data and metrics. For training, we generate 50 states conditioned on uniformly sampled training concept embeddings for single relations. For each concept composition and new concept, we report accuracy on 50 generated states conditioned on the learned concept. For new concepts composed with training concepts, we report accuracy based on 50 generated states for each training concept. Results in Figure 6 are reported for best learned 蠅 and fixed 蠅 {1.2, 1.4, 1.6, 1.8} we learn two concepts with classifier-free guidance weight 蠅 = 1.2. We report accuracy based on a relaxed version of the data generation process: A is considered right of B if x A > x B and |y A y B| 2 max{r A, r B}, A is considered above B if y A > y B and |x A x B| 2 max{r A, r B}. A, B and C lie on a circle if their centers form a circle of radius r where |r 1.67| < 0.3 and A is diagonal to B if x A > x B, y A > y B and they lie within a 2 max{r A, r B} margin of f(x) = x. B.2 Goal-Oriented Navigation Training. To collect demonstrations, we follow the data generation process in the AGENT benchmark environment [72], which provides a planner for navigation given the desired target. In the provided environment, the agent s initial position ap t=0 = (0, 0.102, 3.806), color (yellow) and shape (cone) are fixed. There are two objects: the target and a distractor. Each object has a color oc 1, oc 2 {red, yellow, purple, green}, a shape os 1, os 2 {cone, sphere, bowl, cube} and a position op 1 [0, 1.66] {0.102} [ 4.355, 3.257], op 2 [ 1.66, 0] {0.102} [ 4.355, 3.257]. The training dataset of 900 samples consists of concepts go to red object and go to yellow object where oc 1 {red, yellow}, oc 2 {red, yellow} \ {oc 1} and os 1, os 2 {cone, sphere}, and concepts go to bowl and go to cube where os 1 {bowl, cube}, os 2 {bowl, cube} \ {oc 1} and oc 1, oc 2 {purple, green}. New Tasks. The new scenarios include Five novel compositions of training target color and shape attributes (training composition in Figure 6). go to red bowl (oc i, oc j = red, os i = bowl, os j {cone, sphere} or os i, os j = bowl, oc i = red, oc j {purple, green} where i {1, 2} and j = {1, 2} \ i), go to yellow bowl (oc i, oc j = yellow, os i = bowl, os j {cone, sphere} or os i, os j = bowl, oc i = yellow, oc j {purple, green}), go to red cube (oc i, oc j = red, os i = cube, os j {cone, sphere} or os i, os j = cube, oc i = red, oc j {purple, green}), go to yellow cube (oc i, oc j = yellow, os i = cube, os j {cone, sphere} or os i, os j = cube, oc i = yellow, oc j {purple, green}), go to purple cone (oc i, oc j = purple, os i = cone, os j {bowl, cube} or os i, os j = cone, oc i = purple, oc j {red, yellow}). Note that in each scenario, the distractor object either has the same color or shape as the target, and combined with its other attribute (shape or color), it is within the training distribution (i.e., red and yellow cones and spheres, and purple and green bowls and cubes). There are five demonstrations of each novel composition. Conditioning each novel composition concept on novel initial states sampled from the novel concept distribution (training composition new initial state in Figure 6). State space. The initial state space we condition on is a 52-tuple based on the state space in Shu et al. [72] describing the agent and two objects. The agent is represented by its 3D position, quaternion R4, velocity R3, angular velocity R3, one-hot type (representing the agent and two objects), rgba color, and one-hot shape (representing the four possible shapes). Similarly, each object is represented by its 3D position, type, color, and shape. The demonstration state space is a 13-tuple representing the first 13 dimensions of the initial state space (agent position, quaternion, velocity, and angular velocity). Demonstrations i [N] have horizons Hi 150 and are padded to length 150 using the final state. During training, we sample the demonstrations to generate subtrajectories of length 128. Evaluation data and metrics. For training, we generate 50 trajectories conditioned on uniformly sampled training concepts and initial states. For each new concept, we generate five trajectories conditioned on the learned concept and five initial states from the new concept demonstrations. To evaluate each concept composition on new initial states, we generate trajectories conditioned on the learned concept and 50 initial states sampled from the new concept distribution. Results in Figure 6 are reported for best learned 蠅 and fixed 蠅 {1.2, 1.4, 1.6, 1.8} we learn two concepts with classifier-free guidance weight 蠅 = 1.6. We report accuracy based on whether the agent in the generated trajectory has made progress towards the desired target (|ap t=128 op target| < |ap t=0 op target|). Accuracy for closed loop evaluation (Figure 6) is reported based on optimally reaching target otarget before a distractor odistractor (|ap t=128 op target| < 0.365 t < 128 : |ap t op distractor| 0.365). Closed loop evaluation. The action space A R2 represents forces applied to the agent. We take actions at = 蟿t+5 st where st is an observation in the environment, and 蟿t+5 is a future step in the open loop plan generated by G conditioned on observation st and learned concept c. We execute actions in the environment until reaching the target or a maximum number of steps and report the evaluation metric described above. Closed loop evaluation is done in pybullet simulation [81] for efficiency. Training. All human actions in the CMU Graphics Lab Motion Capture Database (http://mocap.cs.cmu.edu/) are included in the training set except three new scenarios. We further discard videos with less than 128 frames. The training set includes 2210 demonstrations. New Tasks. The new scenarios include Three new concepts. jumping jacks (3 demonstrations), chop wood (2), and breaststroke (3). Conditioning each novel concept on novel initial states the last states in the new concept demonstrations. Composing new learned concept jumping jacks weighted by learned concept weights with three training concepts ( walk , jump and march ) weighted by 蠅 = 1, and conditioned on a training initial state from the training concepts. State space. The initial state space is a 42-tuple representing the 3D position of 14 joints. The original dataset contains 31 joints. Our version is adapted with Tanke et al. [82] to reduce the number of joints to 14 and to remove rotation and translation. The demonstration state space is the same with horizons Hi, i [N]. We subsample the original trajectories every 4 steps and further sample trajectories to generate subtrajectories of unified length 32 on which we train. Evaluation data and metrics. For each new concept, we generate trajectories conditioned on the learned concept and an initial state from the new concept demonstrations. For evaluating each new concept on new initial states, we generate trajectories conditioned on the learned concept and novel initial states from the new concept demonstrations that were not used as initial states during concept learning, specifically the last state in each demonstration. For new concepts composed with training concepts, we generate trajectories, each conditioned on the learned concept, a uniformly sampled training concept, and training initial state. Results are reported for best learned 蠅 and fixed 蠅 {1.2, 1.4, 1.6, 1.8} we learn two concepts with classifier-free guidance weight 蠅 = 1.8. Human experiment. We show five humans (aged 23-28, four male) videos of training and new concepts with their natural language labels and ask whether each video depicts the concept and which depicts it best. The participants gave their consent, and the study was approved by an Institutional Review Board. We show three training concepts: march , run , and walk , and three new concepts: jumping jacks , chop wood , and breaststroke on new initial states. B.4 Autonomous Driving Training. We use the following scenarios from the Highway Env driving simulation environment [75]. The training scenarios include four driving scenarios: highway , exit , merge , and intersection . In highway , the objective is driving on a highway at high speed on the rightmost lanes while avoiding collisions. The highway has four lanes and 50 vehicles. The initial lane and position of all vehicles are sampled, as well as non-controlled vehicle speeds. An episode ends if the controlled vehicle crashes or a time limit is reached. In exit the objective is to take a highway exit while driving on a four-lane highway with an exit lane and 20 vehicles. The controlled vehicle is rewarded for exiting at high speed and driving on the rightmost lanes while avoiding collisions. The initial position of all vehicles is sampled. An episode ends if the controlled vehicle crashes or a time limit is reached. In merge , the objective is driving on a highway with three lanes and a merging lane and three vehicles, one of them merging. The controlled vehicle is rewarded for driving at high speed while avoiding collisions and allowing another vehicle to merge into the highway. The lane, position, and speed are sampled for non-controlled vehicles. An episode ends if the controlled vehicle passes the merging lane or crashes. In intersection , the objective is making a left turn at an intersection with four two-way roads and 10 vehicles. The controlled vehicle is rewarded for crossing the intersection at high speed while staying on the road and avoiding collisions. The controlled vehicle s position is sampled, as well as the lane, position, and speed of other vehicles. An episode ends if the controlled vehicle completes crossing the intersection or crashes. Expert demonstrations were collected using a deterministic tree search planner provided in [83]. New Task. The new scenario includes: A novel driving scenario: roundabout . In this scenario, the objective is to take the second exit at a roundabout with four exits and four vehicles. The position and speed of noncontrolled vehicles are sampled. An episode ends if the controlled vehicle crashes or a time limit is reached. The novel concept is conditioned on novel initial states sampled from the new concept distribution. There are five demonstrations of this new concept. State space. The initial state space we condition on is in R5 7, the controlled vehicle and the four closest vehicles. Each vehicle is represented by a 7-tuple, including whether it is present on the road, its x and y positions, x and y velocities, and cosine and sine heading directions. The demonstration state space observations of the controlled vehicle in R7 for horizons Hi, i [N]. We sample trajectories to generate subtrajectories of length 8, which we train on. Evaluation data and metrics. For evaluating the new concept on new initial states, we evaluate in closed loop on 50 initial states sampled from the new concept distribution. Task return, crash, and success rates are calculated based on the rewards described above. Note that the highway scenario doesn t have a success score as there is no final state to reach in this scenario. Results are reported for best learned 蠅 and fixed 蠅 {1.2, 1.4, 1.6, 1.8} for training, we report 蠅 = 1.8 and during new concept learning, we learn two concepts and their corresponding classifier-free guidance weights. Closed Loop Evaluation. The action space is discrete: move to the left lane, stay idle, move to the right lane, drive faster, drive slower. We assume access to a planner P that given two states plans which action to take in the environment via access to simulation in the environment, at = P(蟿t+1, st). The planner simulates the possible actions from observed state st and randomly selects an action out of the ones closest to the model predicted next state 蟿t+1 that does not result in the controlled vehicle crashing. We execute actions until reaching a maximum number of steps or until the controlled vehicle crashes. B.5 Table-Top Manipulation Training. We collect 214 expert demonstrations with a Franka Research 3 robot via teleop with a Spacemouse for four table-top manipulation tasks: pick green circle and place on book (29 demonstrations), pick green circle and place on elevated white surface (30), push green circle to orange triangle (124), and push green circle to orange triangle around purple bowl (31). Examples of these tasks are best viewed on our website. New task. The new scenario includes pushing the green circle to the orange triangle on a book. We provide ten demonstrations of this task. State space. The initial state space we condition on includes an overhead RGB image of the scene (Figure 10) and the robot s end effector pose, a 7-tuple of its 3D position and quaternion, and gripper state in R. We verify that most states are fully observable. The demonstration state space includes observations of the end effector pose and gripper state in R8 for horizons Hi, i [N]. We sample trajectories to generate subtrajectories of length 32, which we train on. Evaluation data and metrics. We evaluate 20 episodes for training push green circle to orange triangle and new concept push green circle to orange triangle on book on new initial states sampled from the concept distributions. An episode is successful if the green circle touches the orange triangle before a maximum horizon is reached. Results for training are reported for 蠅 = 1.8 and during new concept learning, we learn two concepts and their corresponding classifier-free guidance weights. Closed loop evaluation. The action space is an 8-tuple of the end effector pose and gripper state, equivalent to the predictions of the model. Given predicted action 蟿t+1 and end effector pose st, we assume access to a planner that linearly interpolates between the current and next predicted pose. We make predictions and roll out the 32 predicted states in closed loop until it succeeds or executes a maximum number of steps H = 160. C Implementation Details Diffusion model. We represent the noise model 系胃 with an MLP for the object rearrangement domain and with a temporal U-Net for the AGENT, Mo Cap and Driving domains as implemented in Ajay et al. [17]. For Manipulation, we use a temporal U-Net where images are processed by a pretrained resnet18 [84] that is finetuned with the model during training. We use the same hyperparameters as in Ajay et al. [17] except for the probability of removing conditioning information, p, which we set to 0.1. In Table 4 we demonstrate the effect of choosing different classifier-free guidance weights 蠅. We are overall better or comparable to the baselines. Table 4: Classifier-free guidance weight choice effect. For Object Rearrangement and Goal Oriented Navigation we report the accuracy and standard error of the mean for new concepts from new initial states, and for Driving, the success and crash rates. We compare the four baselines with our approach as reported in Figures 6, 9 and 13 and Section 5.2. We report results for our approach with all 蠅 in our hyperparameter search and mark the reported 蠅 in Figures 6, 9 and 13 in bold font. Domain BC VAE In-Context Language 蠅 Ours Object 0.67 0.15 0.08 0.05 - 0.2 0.05 1.2 0.82 0.09 Rearrangement 1.4 0.8 0.1 1.6 0.66 0.08 1.8 0.61 0.06 learned 0.77 0.09 AGENT 0.46 0.05 0.6 0.06 0.57 0.03 0.63 0.07 1.2 0.64 0.08 1.4 0.66 0.05 1.6 0.73 0.07 1.8 0.63 0.06 learned 0.67 0.05 Driving 4% 24% 0% - 1.2 18% Success Rate 1.4 14% 1.6 4% 1.8 10% learned 24% Driving 64% 48% 16% - 1.2 38% Crash Rate 1.4 40% 1.6 48% 1.8 40% learned 32% BC. We implement behavior cloning (BC) [85]. The BC model is deterministic. In the Object Rearrangement domain, we only condition on concept c. To add stochasticity, when evaluating BC, we average results over 50 different seeds. We use an MLP with one hidden layer of size 512, Re LU activations, and Adam W [86] with learning rate 6 10 4. In AGENT, at each step, the input to the model is the fully observable initial state (52-dim) and the condition, and the model learns to predict the next partially observable agent state (13-dim). Composing concepts is implemented by adding conditions. VAE. We implement a conditional variational autoencoder model (CVAE) [76]2. We use an MLP for the architecture with Adam W [86] and learning rate 6 10 4. We sample 50 trajectories per concept and average the results. Composing concepts is implemented by adding conditions. In-Context. For the In-Context learning baseline [8, 28], we use a transformer encoder-decoder architecture with 4 layers and 4 heads, where multiple demonstrations are passed into the encoder with a zero vector to separate consecutive demonstration trajectories. During training, there are five prompt demonstrations, during training evaluation, we provide one demonstration, and for new concepts, we provide two to five demonstrations, depending on the number of demonstrations provided in each domain. The decoder is passed a window of the previous states and predicts the next state. In AGENT and Driving, we use window size K = 1, and in Mo Cap K = 2. We simplify the learning objective to negative log-likelihood loss and convert the data to 20 bins for AGENT and Driving and 16 bins for Mo Cap. We further simplify the AGENT data by sampling the trajectories every eight states. We use learning rate 10 4 in all domains. We do not evaluate on Object Rearrangement since the horizon is one. 2based on the implementation in https://github.com/pytorch/examples/blob/main/vae/main.py Language. The language instructions are as follows. For Object Rearrangement, training compositions are training concept descriptions composed with the word and . New concepts are described as circle triangle and square in a circle of radius 1.67 and A right and above diagonally to B where A, B { circle , triangle , square }. New concepts composed with training concepts are composed with the word and . AGENT concepts are described based on new concept attributes, e.g. go to red bowl . In Mo Cap, descriptions are human actions such as jumping jacks . Compute. We run all simulated experiments on a single NVIDIA RTX A4000 machine. We evaluate our method on real-world table-top manipulation tasks using a Franka Research 3 robot with an overhead Realsense D435I RGB camera and an NVIDIA RTX 4090 machine. Concept learning can take approximately one to two hours. Since a new concept only has to be learned once to generate behavior, in the domains we presented, it is reasonable to learn a concept offline, and therefore, it is not prohibitive. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: We describe our formulation in Section 3, the adapted method in Section 4 and empirical evaluation in Section 5. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: We discuss limitations in Section 6. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [NA] Justification: The paper does not include theoretical results. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: Described in Section 5 and Appendix B and C. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [No] Justification: We plan to release our code in the near future. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: We discuss implementation details in Appendix B and C. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: We provide errors in Figures 6, 9 and 13. Plots without errors display accuracy over a fixed set. Similarly for Tables 5.1, 2 and 3. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: Described in Appendix C. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: We discuss limitations in Section 6 and obtain an IRB for a small-scale human experiment. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: In this paper, we demonstrate that effectively learning behavior (or concept) representations from a few demonstrations is possible given a pretraining dataset, dependent crucially on the quality of training and new concept demonstrations. While such an effective few-shot learning system may have many societal implications, none are specific to this work that we feel a need to highlight. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: There are no high risks we feel the need to highlight. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: We add citations and links throughout the paper to assets used in this work. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: We do not release any new assets. In the future, we plan to release our code and add documentation. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [Yes] Justification: We conduct a small scale human experiment described in Section 5, Table 5.1 and Appendix B.3. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [Yes] Justification: IRB approval was obtained. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.